The functional architecture enhancements for the AIML enablement service are based on the generic functional model specified in clause 6.2 of TS 23.434. The architecture enhancements are organized into functional entities to describe a functional architecture enhancement which addresses the support for AIML enablement aspects for vertical applications.
Figure 5.2.1-1 illustrates the on-network functional model of AIMLE. In the vertical application layer, the VAL client communicates with the VAL server over VAL-UU reference point. VAL-UU supports both unicast and multicast delivery modes. The AIMLE functional entities on the UE and the server are grouped into AIMLE client(s) and AIMLE server(s) respectively.
The AIMLE includes of a common set of services for comprehensive enablement of AIML functionality, including federated and distributed learning (e.g., FL client registration management, FL client discovery and selection), and reference points. The AIMLE services are offered to the vertical application layer (VAL).
The AIMLE client communicates with the AIMLE server(s) over the AIML-UU reference points. The AIMLE client provides functionality to the VAL client(s) over AIML-C reference point. The VAL server(s) communicate with the AIMLE server(s) over AIML-S reference points. The AIMLE servers communicate with the underlying 3GPP network systems using the respective 3GPP interfaces specified by the 3GPP network system. AIML-E reference point enables interactions between two AIMLE servers (e.g. central and edge AIMLE servers).
The AIMLE server interacts with the ML repository which serves as repository for ML model and ML participants over AIML-R.
Figure 5.2.1.1-1 exhibits the service-based interfaces for providing and consuming AIMLE services. The AIMLE server could provide service to VAL server and AIMLE client through interface SAiml.
The AIMLE server as well as ADAES is deployed as a SEAL server; hence enhancements to SEAL architecture (as specified in TS 23.434) are needed to incorporate the AIMLE service. Figure 5.2.1.1-3 illustrates the service-based representation including AIMLE server as part of the SEAL framework.
Figure 5.2.2-1 illustrates the off-network (UE-to-UE) functional model of AIML enablement. In the vertical application layer, the VAL client communicates with a further VAL client over VAL-PC5 reference point. VAL-PC5 supports both unicast and multicast delivery modes. The UE1, if connected to the network via Uu reference point, can also act as a UE-to-network relay, to enable UE2 to access the VAL server(s) over the VAL-UU reference point.
The AIMLE client communicates with a further AIMLE client(s) over the AIML-PC5 reference points. The AIMLE client provides functionality to the VAL client(s) over AIML-C reference point. Such communication is performed for supporting local ML operations (training, distribution, inference) in a coordinated manner.
The off network functional architecture is similar to SEAL off-network architecture (as specified in TS 23.434).
The AIMLE server functional entity provides AIMLE services supported within the vertical application layer. It interacts with the AIMLE client, the 3GPP network, other SEAL services, and VAL server.
The ML repository is a logical entity that serves both as a registry for ML/FL members and as a repository for application layer ML model related information. It can be accessed by the AIMLE server.
The interactions related to AIML enablement functions between the AIMLE client and AIMLE server are supported by AIML-UU reference point. This reference point utilizes Uu reference point as described in TS 23.401 and TS 23.501.
The interactions related to AIML enablement functions between the VAL client(s) and the AIMLE client within a VAL UE are supported by AIML-C reference point.
This functionality covers the retrieval of ML models by an AIMLE client or a VAL server via the AIMLE server. The AIMLE server provides the functionality required to retrieve ML models stored in a ML repository based on a filtering criteria. The functionality is provided to the ML model consumer via an API following a request/response model or a subscribe/notify model.
This functionality enables the AIMLE server to support ML model training based on requests from a VAL server. The AIMLE server provides the functionality to train a specific ML model or assist the ML model training at VAL server/client(s).
This functionality covers the registration, registration update and de-registration of the candidate FL member to the ML repository which is keeping the FL member registrations. Such candidate member can be a VAL server functionality or an enabler layer functionality (e.g. AIMLE server) which is registering to the ML repository/registry to act as FL member for a given application event (analytics event or event triggered by a VAL layer application server).
This functionality enables a consumer (who can be the AIMLE server or a VAL server e.g. acting as FL server) to subscribe for FL related events and getting notified on changes on the availability of the FL members which are to be used for the FL-related task (e.g., training). This capability at the ML repository acting as an AIML service registry supports the subscription for events related to FL members and the notification to the consumer in case of changes. This feature assumes that such FL members (AIMLE or VAL server or AIMLE clients) have previously registered to this registry their availability and capabilities.
This functionality covers the AIMLE support for ML task transfer, which is applicable to scenarios where an AI/ML member cannot finish the assigned AI/ML task during the performing process. In this scenario, the AIMLE server assists the source AI/ML member by support transferring the intermediate AI/ML information (e.g., the intermediate AI/ML operation status and results) to another AI/ML member (target AI/ML member) for further operations to complete the AI/ML task.
This functionality allows an AIMLE client (e.g., AI/ML capable UEs) to register with an AIMLE server. The AIMLE server stores the client information for future interactions. This functionality is crucial for enabling the AIMLE client to participate in AI/ML operations.
This functionality enables the VAL server to discover available AIMLE clients for AI/ML operations, such as training or inference. The AIMLE server provides the functionality to select suitable AIMLE clients that fulfill the discovery criteria.
This functionality enables the selection of AIMLE clients to participate in AI/ML operations. There are two modes for client selection: VAL server selection and AIMLE server selection. In VAL server selection, the functionality is provided by the AIMLE Server selecting candidate AIMLE clients from the client list provided by the VAL Server. In AIMLE server selection, the functionality is provided by the AIMLE Server selecting candidate AIMLE clients based on the client selection criteria provided by the VAL Server from the AIMLE clients in the ML repository.
This functionality enables the AIMLE server to verify and manage the participation of AIMLE clients in AI/ML operations. The AIMLE client responds with its willingness to perform AI/ML operations based on the information provided by the AIMLE server.
This functionality enables the AIMLE server to manage ML models through interaction with the ML repository. The AIMLE server provides the storage and discovery functionality for ML model information. The storage functionality allows the AIMLE server to store ML model information in the repository. The discovery functionality enables the AIMLE server to retrieve information about available ML models via an API following a request/response model.
This functionality provides the AIMLE server support for horizontal federated learning. This support is applicable to the case where multiple AIMLE clients are expected to locally train the ML model, and the AIMLE server is required to select, configure and coordinate the HFL clients.
This functionality is related to the AIMLE client selection subscription request and notification to enable VAL Servers to subscribe for monitoring AIML members who meet criteria for performing an AIMLE service, selecting AIMLE clients and receiving notification when there is an update on the selected and re-selected AIML client's status when re-selection is performed according to AIML member selection criteria.
Split AI/ML operation is a type of AI/ML operation that allows distributed processing related to ML models into multiple stages on different processing nodes. The intention is to offload the computation-intensive and energy-intensive AI/ML stages to network endpoints, whereas leave the privacy-sensitive and delay-sensitive stage at the end device as described in TS 22.261.
The following functionality is provided to support split AI/ML operation:
An application consuming services from the AI/ML application enablement layer can discover or manage (e.g., create, update, delete) a split operation profile with the AIMLE server for the purpose of consuming results from corresponding instance of a split AI/ML operation pipeline.
A VAL server can register with the AIMLE server to indicate its capabilities for acting as a processing node of an instance of a split AI/ML operation pipeline.
An application consuming services from the AI/ML application enablement layer can subscribe with the AIMLE server to receive event notifications related to an instance of a split AI/ML operation pipeline.
This functionality covers the AIMLE assistance in data management related operations in the ML model lifecycle. AIMLE data management assistance is the process of the AIMLE server assisting AIMLE service consumers with managing data operations (data preparation and processing) performed by VAL clients.
This functionality covers the support for discovering and selecting pre-trained models transfer learning operations in application enablement layer, where the support is based on the request for either an ML task from VAL layer or for an analytics task from ADAES. Transfer Learning enablement allows the consumer to discover the similar ML models to be used as base models for the TL, as well as to support the selection of the best model to be used as pre-trained model.
This functionality covers the AIMLE capability to enable the group management of the entities serving as FL clients at the application enablement layer. Such group management is about the creation, monitoring and update of the FL member groups based on the AI/ML operations, which are based on 1) the analytics event/service by ADAES or 2) the VAL requirement for FL support services.
This functionality is related to the AIMLE support for vertical FL (VFL) among AIMLE clients serving as VFL members. This capability involves the determination of employing VFL based on the ML model training request, the feature alignment and the decision on the AIMLE clients serving as VFL clients, based on the training capability evaluation functionality.
This functionality enables AIMLE server to request AIMLE client for ML model training capability evaluation to support FL training (e.g. HFL, VFL). AIMLE client evaluates its capability and availability to join the FL training process and responds to the AIMLE server with the evaluation result (e.g. join the FL training process with test result, or not join with fail reason). The ML model training capability evaluation result can be used by the AIMLE server to select FL members for FL training process (e.g. HFL, VFL).
This functionality enables the AIMLE server to update trained and deployed ML models by detecting model performance degradation, and triggering ML model re-training and update to fix the observed degradation.
This functionality covers a capability of AIMLE server for monitoring detecting a degradation relation to an ML operation / analytics operation and translating to an ML model degradation (expected or predicted) and performing an action to alleviate this issue (new model training or re-training).
This functionality enables an AIMLE service consumer to request assistance with ML model selection from an AIMLE server. The AIMLE server returns a list of ML models with corresponding model performance.
This functionality provides support for AIMLE operations performed by AIMLE clients spread across multiple edge service areas in edge data networks. The functionality allows the context transfer between edge AIMLE servers when an AIMLE client moves to a different edge service area.
This functionality provides support for assisting hierarchical computing by AIMLE servers. The functionality allows the consumer to request hierarchical computing assistance information to support its operations.
The common identities for SEAL and ADAES refer to TS 23.434 and TS 23.436 respectively. The following clauses list the additional identities and commonly used values for AIMLE Service.
The ML repository ID uniquely identifies the ML-related repository function, which is used for storing the ML models and ML related information, as well as for serving as registry for the ML / FL members.
The FL member ID uniquely identifies the participant entity in a Federated Learning process which is supported by AIMLE service, e.g., AIMLE server ID, VAL server ID or EAS ID.
The AIMLE service area is the area where the AIML Enablement server owner provides its AIML support services. The AIMLE service area can be expressed as a Topological Service Area (e.g., a list of TA), a Geographical Service Area (e.g., geographical coordinates) or both.
The ML model profile ID uniquely identifies a ML model profile that is created by the ML repository as a result of a successful ML model storage request by a AIMLE server.
The AIMLE client set ID is a unique identifier assigned by an AIMLE server that is associated with the set of AIMLE clients that have been selected and have agreed to participate in performing AI/ML related operations, e.g., ML model training.