The user's device has cameras (usually both front and back) and various sensors,
such as an accelerometer, that provide data to your AR app for interpreting the
real world. The camera itself may have a depth sensor and be able to determine
the range, dimensions, and other useful data about the targets it detects.
For your AR app, you configure the camera for optimal performance. You can also
use camera image metadata, frame buffering, and shared camera access to tune
performance. The guides above describe some of the AR camera capabilities for
each platform.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2024-10-31 UTC."],[[["Access comprehensive guides for configuring and optimizing camera usage in your AR applications across Android, Android NDK, Unity, and Unreal Engine platforms."],["Leverage device cameras and sensors, including depth sensors, to capture real-world data for enhanced AR experiences."],["Fine-tune camera performance by adjusting configurations, utilizing image metadata, implementing frame buffering, and managing shared camera access."],["Explore platform-specific documentation and resources to effectively integrate AR camera capabilities into your app development workflow."],["Utilize the provided links to access detailed guides on specific camera features and functionalities for each supported platform."]]],[]]