Tech-invite3GPPspaceIETFspace
21222324252627282931323334353637384‑5x

Content for  TR 26.998  Word version:  18.0.0

Top   Top   Up   Prev   Next
0…   4…   4.2…   4.2.2…   4.2.2.2   4.2.2.3   4.2.2.4   4.2.3…   4.3…   4.4…   4.5…   4.6…   4.6.4…   4.6.5…   4.6.8…   5   6…   6.2…   6.2.4…   6.2.4.2   6.2.5…   6.3…   6.3.4…   6.3.4.2   6.3.5…   6.4…   6.4.4   6.4.5…   6.5…   6.5.4   6.5.5   6.5.6…   6.6…   6.6.4   6.6.5…   7…   8…   8.9   9   A…   A.2   A.3…   A.4   A.5   A.6   A.7…

 

6.3.4  Procedures and call flowsp. 64

6.3.4.1  STAR-based interactive immersive servicep. 64

Figure 6.3.4.1-1 illustrates the procedure diagram for interactive immersive services using a STAR-based UE when all essential AR/MR functions in a UE are available without an assist by an edge.
Copy of original 3GPP image for 3GPP TS 26.998, Fig. 6.3.4.1-1: STAR-based procedure for interactive immersive service
Up
Prerequisites and assumptions:
  • The AR/MR Scene Manager includes immersive media rendering and scene graph handling functionalities.
  • The Media Player includes immersive content delivery and immersive media decoding functionalities.
  • The AR/MR Application in the UE is run by the user.
  • The STAR UE initialises AR registration (starts analysing the surroundings where a user/UE is located), it namely:
    1. captures its surroundings via camera(s)
    2. analyses where the device is located
    3. registers the device into the analysed surroundings.
  • AR/MR Application and AR/MR Application Provider have exchanged some information, such as device capability or content configuration, for content rendering. The exchange procedures for device capability and content configuration are FFS.
  • AR/MR Application Provider has established a Provisioning Session and its detailed configurations has been exchanged.
  • AR/MR Application Provider has completed to set up ingesting immersive contents.
Procedures:
Step 1.
The Scene Server context is established, and scene content is ingested by the Media AS.
Step 2.
Service Announcement is triggered by AR/MR Application. Service Access Information including Media Client entry or a reference to the Service Access Information is provided through the M8d interface.
Step 3.
Desired media content is selected.
Step 4.
Optionally, the Service Access information is acquired or updated.
Step 5.
The AR/MR Application initializes the Scene Manager with the entry point (full scene description) URL.
Step 6.
The Media Client establishes the transport session for the scene session between the Scene Manager in the UE and the Scene Server.
Step 7.
The Media Client requests and receives the full scene description. The entry point (scene description) is processed by the AR/MR Scene Manager, and a scene session is created.
Step 8.
The AR/MR Scene Manager requests the creation of a new AR/MR session from the AR Runtime.
Step 9.
The AR Runtime creates a new AR/MR session.
AR Media Delivery Pipeline, steps 10~24, send the interaction and pose information and receives and renders the updated scenes accordingly:
Step 10.
The latest interaction and pose information are acquired by the AR/MR Scene Manager and shared to the Media Client. The Media Client sends this information to the Media AS and Scene Server.
Step 11.
The Scene Server processes the scene according to the interaction and pose information from the UE. Depending on the level of processing, the current scene may be updated or replaced.
Step 12.
When needed, one of the following steps:
Step 12a.
The Scene Server sends a new scene entry points to the AR/MR Scene Manager through the Media AS and Media Client (go to step 7), or
Step 12b.
The Scene Server sends a scene update (updating streams/objects) to the AR/MR Scene Manager through the Media AS and Media Client.
Step 13.
The AR/MR Scene requests to create additional streaming sessions if needed for new media objects in the scene.
Step 14.
The Media Session Handle establishes the additional streaming sessions based on the received request.
Streaming session, steps 15~18 establish the transport sessions for media objects and configure the media pipelines
Step 15.
For the required media content, the Media Client establishes the transport session(s) to acquire delivery manifest(s) information.
Step 16.
The Media Client requests and receives the delivery manifest(s) from the Media AS.
Step 17.
The Media Client processes the delivery manifest(s). It determines for example the number of needed transport sessions for media acquisition. The Media Client is expected to be able to use the delivery manifest(s) information to initialize the media pipelines for each media stream.
Step 18.
The AR/MR Scene Manager and Media Client configures the rendering and delivery media pipelines.
Step 19.
The Media Client establishes the transport session(s) to acquire the media content.
Media session loop includes steps 20~24 which are for streaming, decoding and rendering media components:
Step 20.
The Media Client requests the immersive media data according to the delivery manifest processed, possibly taking into account pose information (e.g., viewport dependent streaming).
Step 21.
The Media Client receives the immersive media data and triggers the media rendering pipeline(s), including the registration of AR content into the real world accordingly.
Step 22.
The Media Client decodes and processes the media data. For encrypted media data, the Media Client may also perform decryption.
Step 23.
The Media Client passes the media data to the AR/MR Scene Manager.
Step 24.
The AR/MR Scene Manager renders the media, and passes the rendered media to the AR Runtime, which performs further processing such as registration of the AR content into the real world, and pose correction.
Up

Up   Top   ToC