Similar to
clause 6.5.3, the shared AR conversational experience may be instantiated in various 5G systems:
-
The MTSI architecture (TS 26.114) supports audio and 2D video conversational services.
-
Extending the 5GMS architecture (TS 26.501) to support AR conversational services by combining live uplink and live downlink.
-
An architecture based on something different than MTSI / IMS or 5GMS, for example WebRTC.
For the comparison between different instantiations, please refer to
Table 6.5.3-1.
Based on the use cases, the following formats, codecs and packaging formats are of relevance for Media Streaming of AR:
-
Scene Graph/Description
-
2D Video Formats
-
3D Formats such as static and dynamic point clouds or meshes
-
Animated 3D meshes
-
2D Video Formats with depth
-
Mono, stereo, and spatial audio formats
-
Several video decoding instances
-
Decoding tools for such formats
-
Encoding tools for 2D formats
-
Low-latency downlink and uplink real-time streaming of the above media
-
Uplink streaming of pose information
-
Uplink streaming of media
The list of potential standardization area that has been collected is provided in the following:
-
Immersive media format and simplified media formats with integration into relevant 5G architecture
-
Scene description format, functionality, and profile as an entry point of immersive media
-
Scene description update mechanism
-
Relevant subset of media codecs for different media types and formats
-
CMAF encapsulation of immersive media for 5G media streaming
-
Media payload format to be mapped into RTP streams
-
Capability exchange mechanism and relevant signalling
-
Low-latency uplink streaming of captured AR data
-
Functionalities and session management to support split rendering and network-based media processing allocation with 5G edge/MRF
-
Required QoS and QoE for shared AR/MR conversational experience service