Use a projected context to access AI glasses hardware

Applicable XR devices
This guidance helps you build experiences for these types of XR devices.
AI Glasses

After you've requested and been granted the necessary permissions, your app can access the AI glasses hardware. The key to accessing the glasses' hardware (instead of the phone's hardware), is to use a projected context.

There are two primary ways to get a projected context, depending on where your code is executing:

Get a projected context if your code is running in an AI glasses activity

If your app's code is running from within your AI glasses activity, its own activity context is already a projected context. In this scenario, calls made within that activity can already access the glasses' hardware.

Get a projected context if your code is running in phone app component

If a part of your app outside of your AI glasses activity (such as a phone activity or a service) needs to access the glasses' hardware, it must explicitly obtain a projected context. To do this, use the createProjectedDeviceContext() method:

// From a phone Activity, get a context for the AI glasses
try {
    val glassesContext = ProjectedContext.createProjectedDeviceContext(this)
    // Now use glassesContext to access glasses' system services
} catch (e: IllegalStateException) {
    // Projected device was not found
}

Check for validity

After creating the projected context, monitor ProjectedContext.isProjectedDeviceConnected. While this method returns true, the projected context remains valid to the connected device, and your phone app activity or service (such as a CameraManager) can access the AI glasses hardware.

Clean up on disconnect

The projected context is tied to the lifecycle of the connected device, so it is destroyed when the device disconnects. When the device disconnects, ProjectedContext.isProjectedDeviceConnected returns false. Your app should listen for this change and clean up any system services (such as a CameraManager) or resources that your app created using that projected context.

Re-initialize on reconnect

When the AI glasses device reconnects, your app can obtain another projected context instance using createProjectedDeviceContext(), and then re-initialize any system services or resources using the new projected context.

Access audio using bluetooth

Currently, AI glasses connect to your phone as a standard Bluetooth audio device. Both the headset and the A2DP (Advanced Audio Distribution Profile) profiles are supported. Using this approach lets any Android app that supports audio input or output work on glasses, even if they were not purposefully built to support glasses. In some cases, using Bluetooth might work better for your app's use case as an alternative to accessing the glasses' hardware using a projected context.

As with any standard Bluetooth audio device, the permission to grant the RECORD_AUDIO permission is controlled by the phone and not the glasses.

Capture an image with the AI glasses' camera

To capture an image with the AI glasses' camera, set up and bind the CameraX's ImageCapture use case to the glasses' camera using the correct context for your app:

private fun startCamera() {
    // Get the CameraProvider using the projected context.

    val cameraProviderFuture = ProcessCameraProvider.getInstance(
        ProjectedContext.createProjectedDeviceContext(this)
    )

    cameraProviderFuture.addListener({
        // Used to bind the lifecycle of cameras to the lifecycle owner
        val cameraProvider: ProcessCameraProvider = cameraProviderFuture.get()

        // Select the camera. When using the projected context, DEFAULT_BACK_CAMERA maps to the AI glasses' camera.
        val cameraSelector = CameraSelector.DEFAULT_BACK_CAMERA
    
        // Check for the presence of a camera before initializing the ImageCapture use case.
       if (!cameraProvider.hasCamera(cameraSelector)) {
            Log.w(TAG, "The selected camera is not available.")
            return@addListener
        }

        // Get supported streaming resolutions.
        val cameraInfo = cameraProvider.getCameraInfo(cameraSelector)
        val camera2CameraInfo = Camera2CameraInfo.from(cameraInfo)
        val cameraCharacteristics = camera2CameraInfo.getCameraCharacteristic(CameraCharacteristics.SCALER_STREAM_CONFIGURATION_MAP)

        // Define the resolution strategy.
        val targetResolution = Size(1920, 1080)
        val resolutionStrategy = ResolutionStrategy(
            targetResolution,
            ResolutionStrategy.FALLBACK_RULE_CLOSEST_LOWER)

        val resolutionSelector = ResolutionSelector.Builder()
            .setResolutionStrategy(resolutionStrategy)
            .build()

        // If you have other continuous use cases bound, such as Preview or ImageAnalysis, you can use  Camera2 Interop's CaptureRequestOptions to set the FPS
        val fpsRange = Range(30, 30)
        val captureRequestOptions = CaptureRequestOptions.Builder()
                .setCaptureRequestOption(CaptureRequest.CONTROL_AE_TARGET_FPS_RANGE,fpsRange)
                .build()

        // Initialize the ImageCapture use case.
        val imageCapture = ImageCapture.Builder()
            // Optional: Configure resolution, format, etc.
            .setResolutionSelector(resolutionSelector)
            .build()

        try {
            // Unbind use cases before rebinding
            cameraProvider.unbindAll()

            // 4. Bind use cases to camera
            cameraProvider.bindToLifecycle(this as LifecycleOwner, cameraSelector, imageCapture)

        } catch(exc: Exception) {
            // This catches exceptions like IllegalStateException if use case binding fails
            Log.e(TAG, "Use case binding failed", exc)
        }

    }, ContextCompat.getMainExecutor(this))
}

Key points about the code

  • Obtains an instance of the ProcessCameraProvider using the projected device context.
  • Within the projected context's scope, the AI glasses' primary, outward-pointing camera maps to the DEFAULT_BACK_CAMERA when selecting a camera.
  • A pre-binding check uses cameraProvider.hasCamera(cameraSelector) to verify the selected camera is available on the device before proceeding.
  • Uses Camera2 Interop with Camera2CameraInfo to read the underlying CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP, which can be useful for advanced checks on supported resolutions.
  • A custom ResolutionSelector is built to precisely control the output image resolution for ImageCapture.
  • Creates an ImageCapture use case that is configured with a custom ResolutionSelector.
  • Binds the ImageCapture use case to the activity's lifecycle. This automatically manages the opening and closing of the camera based on the Activity's state (for example, stopping the camera when the activity is paused).

After the AI glasses' camera is set up, you can capture an image with the CameraX's ImageCapture class. Refer to the CameraX's documentation to learn about using takePicture() to capture an image.

Capture a video with the AI glasses' camera

To capture a video instead of an image with the AI glasses' camera, replace the ImageCapture components with the corresponding VideoCapture components and modify the capture execution logic.

The main changes involve using a different use case, creating a different output file, and initiating the capture using the appropriate video recording method. For more information about the VideoCapture API and how to use it, see the CameraX's video capture documentation.

The following table shows the recommended resolution and frame rate depending on your app's use case:

Use case Resolution Frame rate
Video Communication 1280 x 720 15 FPS
Computer Vision 640 x 480 10 FPS
AI Video Streaming 640 x 480 1 FPS

Access a phone's hardware from an AI glasses activity

An AI glasses activity can also access the phone's hardware (such as the camera or microphone) by using createHostDeviceContext(context) to get the host device's (phone) context:

// From an AI glasses Activity, get a context for the phone
val phoneContext = ProjectedContext.createHostDeviceContext(this)
// Now use phoneContext to access the phone's hardware

When accessing hardware or resources that are specific to the host device (phone) in a hybrid app (an app containing both mobile and AI glasses experiences), you must explicitly select the correct context to make sure your app can access the correct hardware: