Microsoft.MixedReality.OpenXR Namespace
Important
Some information relates to prerelease product that may be substantially modified before it’s released. Microsoft makes no warranties, express or implied, with respect to the information provided here.
Classes
AnchorConverter |
Provides helper functions to convert an Unity Anchor object to underlying OpenXR anchor handle or SpatialAnchor COM object. |
ARMarker |
Represents a marker detected by an AR device. |
ARMarkerManager |
A manager for ARMarkers. Creates, updates, and removes
|
ARMarkerScale |
Monobehavior to help scale detected markers. |
ControllerModel |
Provides access to a byte stream representing a glTF model of the current controller. |
ControllerModelArticulator |
Handles articulation of the animatable parts of a controller model. |
EyeLevelSceneOrigin |
Add the EyeLevelSceneOrigin component to the scene, it will automatically switch the Unity's scene origin to an eye level experiences. It will try to use "Unbounded" origin mode when it's supported. |
GestureRecognizer |
A gesture recognizer interprets user interactions from hands, motion controllers, and system voice commands to surface spatial gesture events, which users target using their gaze or hand's pointing ray. |
HandMeshTracker |
Represents a user's hand and the ability to render a hand mesh representation of it. |
HandTracker |
Represents a user's hand and the ability to track hand joints from it. |
MeshSettings |
Static entry point for updating the mesh compute settings. |
OpenXRContext |
Retrieve the current OpenXR instance, session handles and states. |
OpenXRTime |
Provides methods to interact with OpenXR time, including retrieving the predicting display times, and converting XR time to Query Performance Counter (QPC) time. |
PerceptionInterop |
Interop functions for Windows Perception APIs |
RequiresNativePluginDLLsAttribute | |
SelectKeywordRecognizer |
A keyword recognizer listening to the "select" keyword localized in the system display language of HoloLens 2. |
SpatialGraphNode |
A spatial graph node represents a spatially tracked point provided by the driver. |
TrackingMapManager |
TrackingMapManager class allows an application to opt into running in an Application-Exclusive tracking mode instead of default shared environment. |
ViewConfiguration |
Observe and manage a view configuration of the current XR session. |
XRAnchorStore |
Handles persisting anchors from the scene to the anchor store, loading anchors from the anchor store to the scene, and managing anchors persisted in the anchor store. |
XRAnchorTransferBatch |
Provides the ability to build up a batch of anchors and export them to a binary stream for transfer. Typically on a second device, it then supports importing the transfer stream and loading in the original batch of anchors. |
Structs
ARMarkersChangedEventArgs |
Event arguments for the markersChanged event. Following design pattern set by UnityEngine.XR.ARFoundation.ARPlanesChangedEventArgs |
GestureEventData |
The data of a gesture event, include the event type, handedness, poses etc. |
HandJointLocation |
Represents locational data for a hand joint. |
ManipulationEventData |
The data of a manipulation gesture event |
MeshComputeSettings |
Settings describing the quality and type of mesh to be provided. |
NavigationEventData |
The data of a navigation gesture event |
QRCodeProperties |
Represents properties of detected QRCodes. |
ReprojectionSettings |
The settings to control the reprojection of current rendering frame, including the reprojection mode and optional stabilization plane override. |
TappedEventData |
The data of a tap gesture event |
Enums
ARMarkerType |
Types of markers that can be tracked. |
FrameTime |
Choose the predicted display time of a frame in pipelined rendering. |
GestureEventType |
Represents the type of gesture recognizer event |
GestureHandedness |
Represents the hand that initiated the gesture. |
GestureSettings |
Represents the set of gestures that may be recognized by a GestureRecognizer. |
Handedness |
Describes which hand the current hand tracker represents. |
HandJoint |
The supported tracked hand joints in OpenXR. |
HandJointsMotionRange |
The requested hand joints' range of motion from a controller. |
HandPoseType |
Represents different possible hand poses. |
MeshComputeConsistency |
The compute consistency to request from the XRMeshSubsystem. |
MeshType |
The type of mesh to request from the XRMeshSubsystem. |
QRCodeType |
Represents types of QRCodes that can be detected. |
ReprojectionMode |
The ReprojectionMode describes the reprojection mode of a projection composition layer. |
TrackingMapType |
Type of tracking maps |
TransformMode |
Types of transforms that can be applied to markers. |
ViewConfigurationType |
A view configuration is a semantically meaningful set of one or more views for which an application can render images. |
VisualMeshLevelOfDetail |
The level of detail of visual mesh to request from the XRMeshSubsystem. |