After you've requested and been granted the necessary permissions, your app can access the AI glasses hardware. The key to accessing the glasses' hardware (instead of the phone's hardware), is to use a projected context.
There are two primary ways to get a projected context, depending on where your code is executing:
Get a projected context if your code is running in an AI glasses activity
If your app's code is running from within your AI glasses activity, its own activity context is already a projected context. In this scenario, calls made within that activity can already access the glasses' hardware.
Get a projected context if your code is running in phone app component
If a part of your app outside of your AI glasses activity (such as a phone
activity or a service) needs to access the glasses' hardware, it must explicitly
obtain a projected context. To do this, use the
createProjectedDeviceContext() method:
// From a phone Activity, get a context for the AI glasses
try {
val glassesContext = ProjectedContext.createProjectedDeviceContext(this)
// Now use glassesContext to access glasses' system services
} catch (e: IllegalStateException) {
// Projected device was not found
}
Check for validity
After creating the projected context, monitor
ProjectedContext.isProjectedDeviceConnected. While this method returns
true, the projected context remains valid to the connected device, and your
phone app activity or service (such as a CameraManager) can access the
AI glasses hardware.
Clean up on disconnect
The projected context is tied to the lifecycle of the connected device, so it is
destroyed when the device disconnects. When the device disconnects,
ProjectedContext.isProjectedDeviceConnected returns false. Your app
should listen for this change and clean up any system services (such as a
CameraManager) or resources that your app created using that projected
context.
Re-initialize on reconnect
When the AI glasses device reconnects, your app can obtain another projected
context instance using createProjectedDeviceContext(), and then
re-initialize any system services or resources using the new projected context.
Access audio using bluetooth
Currently, AI glasses connect to your phone as a standard Bluetooth audio device. Both the headset and the A2DP (Advanced Audio Distribution Profile) profiles are supported. Using this approach lets any Android app that supports audio input or output work on glasses, even if they were not purposefully built to support glasses. In some cases, using Bluetooth might work better for your app's use case as an alternative to accessing the glasses' hardware using a projected context.
As with any standard Bluetooth audio device, the permission to grant the
RECORD_AUDIO permission is controlled by the phone and not the glasses.
Capture an image with the AI glasses' camera
To capture an image with the AI glasses' camera, set up and bind the CameraX's
ImageCapture use case to the glasses' camera using the correct
context for your app:
private fun startCamera() {
// Get the CameraProvider using the projected context.
val cameraProviderFuture = ProcessCameraProvider.getInstance(
ProjectedContext.createProjectedDeviceContext(this)
)
cameraProviderFuture.addListener({
// Used to bind the lifecycle of cameras to the lifecycle owner
val cameraProvider: ProcessCameraProvider = cameraProviderFuture.get()
// Select the camera. When using the projected context, DEFAULT_BACK_CAMERA maps to the AI glasses' camera.
val cameraSelector = CameraSelector.DEFAULT_BACK_CAMERA
// Check for the presence of a camera before initializing the ImageCapture use case.
if (!cameraProvider.hasCamera(cameraSelector)) {
Log.w(TAG, "The selected camera is not available.")
return@addListener
}
// Get supported streaming resolutions.
val cameraInfo = cameraProvider.getCameraInfo(cameraSelector)
val camera2CameraInfo = Camera2CameraInfo.from(cameraInfo)
val cameraCharacteristics = camera2CameraInfo.getCameraCharacteristic(CameraCharacteristics.SCALER_STREAM_CONFIGURATION_MAP)
// Define the resolution strategy.
val targetResolution = Size(1920, 1080)
val resolutionStrategy = ResolutionStrategy(
targetResolution,
ResolutionStrategy.FALLBACK_RULE_CLOSEST_LOWER)
val resolutionSelector = ResolutionSelector.Builder()
.setResolutionStrategy(resolutionStrategy)
.build()
// If you have other continuous use cases bound, such as Preview or ImageAnalysis, you can use Camera2 Interop's CaptureRequestOptions to set the FPS
val fpsRange = Range(30, 30)
val captureRequestOptions = CaptureRequestOptions.Builder()
.setCaptureRequestOption(CaptureRequest.CONTROL_AE_TARGET_FPS_RANGE,fpsRange)
.build()
// Initialize the ImageCapture use case.
val imageCapture = ImageCapture.Builder()
// Optional: Configure resolution, format, etc.
.setResolutionSelector(resolutionSelector)
.build()
try {
// Unbind use cases before rebinding
cameraProvider.unbindAll()
// 4. Bind use cases to camera
cameraProvider.bindToLifecycle(this as LifecycleOwner, cameraSelector, imageCapture)
} catch(exc: Exception) {
// This catches exceptions like IllegalStateException if use case binding fails
Log.e(TAG, "Use case binding failed", exc)
}
}, ContextCompat.getMainExecutor(this))
}
Key points about the code
- Obtains an instance of the
ProcessCameraProviderusing the projected device context. - Within the projected context's scope, the AI glasses' primary,
outward-pointing camera maps to the
DEFAULT_BACK_CAMERAwhen selecting a camera. - A pre-binding check uses
cameraProvider.hasCamera(cameraSelector)to verify the selected camera is available on the device before proceeding. - Uses Camera2 Interop with
Camera2CameraInfoto read the underlyingCameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP, which can be useful for advanced checks on supported resolutions. - A custom
ResolutionSelectoris built to precisely control the output image resolution forImageCapture. - Creates an
ImageCaptureuse case that is configured with a customResolutionSelector. - Binds the
ImageCaptureuse case to the activity's lifecycle. This automatically manages the opening and closing of the camera based on the Activity's state (for example, stopping the camera when the activity is paused).
After the AI glasses' camera is set up, you can capture an image with the
CameraX's ImageCapture class. Refer to the CameraX's documentation to learn
about using takePicture() to capture an image.
Capture a video with the AI glasses' camera
To capture a video instead of an image with the AI glasses' camera, replace the
ImageCapture components with the corresponding VideoCapture components
and modify the capture execution logic.
The main changes involve using a different use case, creating a different output
file, and initiating the capture using the appropriate video recording method.
For more information about the VideoCapture API and how to use it, see the
CameraX's video capture documentation.
The following table shows the recommended resolution and frame rate depending on your app's use case:
| Use case | Resolution | Frame rate |
|---|---|---|
| Video Communication | 1280 x 720 | 15 FPS |
| Computer Vision | 640 x 480 | 10 FPS |
| AI Video Streaming | 640 x 480 | 1 FPS |
Access a phone's hardware from an AI glasses activity
An AI glasses activity can also access the phone's hardware (such as the camera
or microphone) by using createHostDeviceContext(context) to get the host
device's (phone) context:
// From an AI glasses Activity, get a context for the phone
val phoneContext = ProjectedContext.createHostDeviceContext(this)
// Now use phoneContext to access the phone's hardware
When accessing hardware or resources that are specific to the host device (phone) in a hybrid app (an app containing both mobile and AI glasses experiences), you must explicitly select the correct context to make sure your app can access the correct hardware:
- Use the
Activitycontext from the phoneActivityor theProjectedContext.createHostDeviceContext()to get the phone's context. - Don't use
getApplicationContext()because the application context can incorrectly return the AI glasses' context if a glasses activity was the most-recently-launched component.