WO2022032669A1 - 一种数据接收方法、装置及终端设备 - Google Patents

一种数据接收方法、装置及终端设备 Download PDF

Info

Publication number
WO2022032669A1
WO2022032669A1 PCT/CN2020/109323 CN2020109323W WO2022032669A1 WO 2022032669 A1 WO2022032669 A1 WO 2022032669A1 CN 2020109323 W CN2020109323 W CN 2020109323W WO 2022032669 A1 WO2022032669 A1 WO 2022032669A1
Authority
WO
WIPO (PCT)
Prior art keywords
video frame
terminal device
transmission
timing
data
Prior art date
Application number
PCT/CN2020/109323
Other languages
English (en)
French (fr)
Inventor
黄曲芳
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to PCT/CN2020/109323 priority Critical patent/WO2022032669A1/zh
Publication of WO2022032669A1 publication Critical patent/WO2022032669A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W52/00Power management, e.g. TPC [Transmission Power Control], power saving or power classes
    • H04W52/02Power saving arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W74/00Wireless channel access
    • H04W74/04Scheduled access
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Definitions

  • the present application relates to the field of communication technologies, and in particular, to a data receiving method, apparatus, and terminal device.
  • Extended reality refers to a combination of real and virtual and human-computer interaction generated by computer technology and wearable devices. It is augmented reality (AR), virtual reality (VR) , mixed reality (mixed reality, MR) and other forms collectively.
  • AR augmented reality
  • VR virtual reality
  • MR mixed reality
  • the three visual interaction technologies are integrated to achieve an "immersive" experience of seamless transition between the virtual world and the real world.
  • the XR server generates video frames at a rate of 60 frames per second, that is, an original video frame is generated every 16.67ms.
  • the original video frame is encoded to generate encoded video data
  • the access network element such as gNB
  • the access network element receives the encoded video data from the XR server through the user plane function (UPF), and determines it according to the wireless air interface load.
  • Transmission timing video data is transmitted at the transmission timing.
  • the terminal device receives one video data frame every 16.67ms.
  • the network element of the access network is processed, and only transmits the video data in a part of the 16.67ms time period, and does not transmit it in other time periods. However, the end device will always remain listening, causing additional power consumption.
  • Embodiments of the present application provide a data receiving method, apparatus, and terminal device, and the method can prevent the terminal device from continuously monitoring data transmission and reception, which is beneficial to reducing the power consumption of the terminal device.
  • an embodiment of the present application provides a data receiving method, and the method can be executed by a terminal device.
  • the terminal device may determine the timing configuration, and receive the first notification information from the access network element at the notification timing indicated by the timing configuration, so as to determine the start transmission timing and/or transmission duration of the video frame.
  • the terminal device only receives the first notification information of the access network element at the notification timing, and determines the starting transmission timing and/or the transmission duration of the video frame according to the first notification information.
  • the terminal device only monitors data at a specific timing (eg, the initial transmission timing of the video frame), which is beneficial to reduce the power consumption of the terminal device.
  • the terminal device may also monitor the video frame at the start transmission timing of the video frame, and switch the working state to the sleep state at the end transmission timing corresponding to the transmission duration of the video frame.
  • the terminal device starts monitoring data at the initial transmission timing of the video frame, and stops monitoring data at the end transmission timing of the video frame, thereby reducing the power consumption of the terminal device.
  • the terminal device receives the notification timing from the access network element, where the notification timing is determined by the access network element according to the encoding duration of the video frame.
  • the notification timing can be determined by the access network element according to the encoding duration of the video frame. Then, for video frames of different services, the access network element can determine different notification timings to adapt to different services, which is conducive to setting More precise notification timing, thereby better reducing the power consumption of terminal devices.
  • the terminal device may receive multiple video frames, and determine the arrival time interval of the video frames according to the service model of the multiple video frames.
  • the arrival time interval of the video frames is the time difference between the start transmission timings of the two video frames.
  • the terminal device can determine the arrival time interval of adjacent video frames according to the service model of the video frame, so as to determine at which time domain positions the terminal device can be in a dormant state and reduce power consumption of the terminal device.
  • the terminal device can switch from the sleep state to the working state, and monitor the next video frame adjacent to the video frame.
  • the terminal device can maintain the sleep state, and determine the next time to start monitoring data according to the arrival time interval, which is beneficial to reduce the power consumption of the terminal device.
  • the terminal device receives second notification information from the access network element, where the second notification information is used to indicate the next video frame adjacent to the video frame.
  • the start transmission opportunity of a video frame The terminal device monitors the next video frame at the start transmission opportunity of the next video frame.
  • the access network element needs to notify the terminal device again of the timing to start receiving data, so as to avoid the terminal device from being in a dormant state or monitoring data all the time.
  • the start transmission opportunity and/or the end transmission opportunity of the video frame are one or more of subframes, time slots or transmission time intervals.
  • the embodiment of the present application provides another data receiving method, and the method is executed by a terminal device.
  • the terminal device may monitor K video frames in advance, and learn the K video frames to determine the start transmission timing, transmission duration, and arrival time interval of each video frame.
  • the terminal device may determine the start transmission opportunity and/or transmission duration of video frame k+1 according to the start transmission opportunity, transmission duration, and arrival time interval of video frame k among the K video frames.
  • the terminal device learns the start transmission timing and/or end transmission timing of the video frame by itself, thereby determining to monitor the data at a specific timing without continuously monitoring the transmission and reception of data, which is beneficial to reduce the power consumption of the terminal device.
  • the terminal device can also monitor the video frame k+1 at the start transmission timing of the video frame k+1, and change the working state at the end transmission timing corresponding to the transmission duration of the video frame k+1. Switch to sleep state.
  • the terminal device can confirm the transmission end time of each video frame, and stop monitoring data at the transmission end time, which is beneficial to reduce the power consumption of the terminal device.
  • the terminal device determines the transmission duration of the video frame k according to the start transmission timing of the video frame k and the end identifier of the video frame k.
  • the terminal device determines the transmission duration of the video frame k according to the initial transmission timing of the video frame k and the time domain position information.
  • the terminal device does not monitor data in consecutive subframes/slots/transmission time intervals after receiving the first data packet of video frame k.
  • the start transmission timing and the transmission timing when no data is monitored determine the transmission duration of the video frame k.
  • the terminal device determines the transmission duration of the video frame k according to the initial transmission timing of the video frame k and the data amount of the video frame k.
  • the terminal device determines the transmission duration of the video frame k according to the initial transmission timing of the video frame k and the real-time transmission protocol RTP packet header in the data packet of the video frame k.
  • the terminal device sends the start transmission timing and transmission duration of the video frame to the access network element.
  • video frame k is the last video frame transmitted among the K video frames
  • video frame k+1 is the first video frame arriving after video frame k.
  • an embodiment of the present application provides a data receiving apparatus, and the apparatus has a function of implementing the data receiving method provided in the first aspect.
  • This function can be implemented by hardware or by executing corresponding software by hardware.
  • the hardware or software includes one or more modules corresponding to the above functions.
  • an embodiment of the present application provides a data receiving apparatus, and the apparatus has a function of implementing the data receiving method provided in the second aspect.
  • This function can be implemented by hardware or by executing corresponding software by hardware.
  • the hardware or software includes one or more modules corresponding to the above functions.
  • an embodiment of the present application provides a terminal device, where the terminal device includes a processor for implementing the functions or methods involved in the first aspect above, and the terminal device may be, for example, a chip system.
  • the terminal device further includes a memory, and the memory is used for storing necessary program instructions and data to implement the functions of the method described in the first aspect above.
  • an embodiment of the present application provides a terminal device, where the terminal device includes a processor for implementing the functions or methods involved in the second aspect above, and the terminal device may be, for example, a chip system.
  • the terminal device further includes a memory, and the memory is used for storing necessary program instructions and data to implement the functions of the method in the second aspect above.
  • an embodiment of the present application provides a computer-readable storage medium, where the computer-readable storage medium includes a program or an instruction, and when the program or instruction is run on a computer, causes the computer to execute the first aspect or the first aspect method in any of the possible implementations.
  • an embodiment of the present application provides a computer-readable storage medium, where the computer-readable storage medium includes a program or an instruction, and when the program or instruction is run on a computer, the computer executes the second aspect or the second aspect method in any of the possible implementations.
  • the chip system in the above aspects may be a system on chip (system on chip, SOC), or a baseband chip, etc.
  • the baseband chip may include a processor, a channel encoder, a digital signal processor, a modem, an interface module, and the like.
  • an embodiment of the present application provides a chip or a chip system, the chip or chip system includes at least one processor and an interface, the interface and the at least one processor are interconnected through a line, and the at least one processor is used for running a computer program or instruction, to perform the method described in any one of the first aspect or any of the possible implementations of the first aspect.
  • an embodiment of the present application provides a chip or a chip system, the chip or chip system includes at least one processor and an interface, the interface and the at least one processor are interconnected through a line, and the at least one processor is used for running a computer program or instruction, to perform the method described in any one of the second aspect or any of the possible implementations of the second aspect.
  • the interface in the chip may be an input/output interface, a pin or a circuit, or the like.
  • the chip or chip system described above in this application further includes at least one memory, where instructions are stored in the at least one memory.
  • the memory may be a storage unit inside the chip, such as a register, a cache, etc., or a storage unit of the chip (eg, a read-only memory, a random access memory, etc.).
  • FIG. 1 is a schematic diagram of a transmission path between a wearable device and an XR server
  • Figure 2a is a schematic diagram of an XR business model
  • 2b is a schematic diagram of a base station receiving and transmitting XR service model
  • 3a is a schematic diagram of a network system provided by an embodiment of the present application.
  • 3b is a schematic diagram of a detailed network element of a network system provided by an embodiment of the application.
  • FIG. 4 is a schematic flowchart of a data receiving method provided by an embodiment of the present application.
  • FIG. 5 is a schematic flowchart of a terminal device switching from a dormant state to an activated state and starting to monitor data according to an embodiment of the present application;
  • FIG. 6 is a sequence diagram of a terminal device switching from a dormant state to an active state and starting to monitor data according to an embodiment of the present application;
  • FIG. 7 is a schematic flowchart of another data receiving method provided by an embodiment of the present application.
  • 8a is a schematic diagram of an end transmission timing of a video frame according to an embodiment of the present application.
  • 8b is a schematic diagram of a transmission duration of a video frame provided by an embodiment of the present application.
  • FIG. 8c is a schematic diagram of another timing for ending transmission of a video frame provided by an embodiment of the present application.
  • 8d is a schematic diagram of another transmission duration of a video frame provided by an embodiment of the present application.
  • FIG. 9 is a schematic structural diagram of a data receiving apparatus according to an embodiment of the present application.
  • FIG. 10 is a schematic structural diagram of a terminal device according to an embodiment of the present application.
  • FIG. 11 is a schematic structural diagram of another data receiving apparatus provided by an embodiment of the present application.
  • FIG. 12 is a schematic structural diagram of another terminal device provided by an embodiment of the present application.
  • Extended reality refers to a combination of real and virtual and human-computer interaction generated by computer technology and wearable devices. It is augmented reality (AR), virtual reality (VR) , mixed reality (mixed reality, MR) and other forms collectively.
  • AR augmented reality
  • VR virtual reality
  • MR mixed reality
  • the three visual interaction technologies are integrated to achieve an "immersive" experience of seamless transition between the virtual world and the real world.
  • the wearable device may be an application layer user device (such as a helmet, etc.), and the wearable device may be connected to a terminal device (such as a smart phone, etc.), or may be physically integrated with the terminal device.
  • the process of realizing the "immersive" experience of seamless transition between the virtual world and the real world includes: sensors in the helmet sense the user's position and movement changes (such as changes in perspective, line of sight, movement rate, etc.), and generate user information.
  • the user information can be transmitted to the XR server through the path of the wearable device—the operator network (such as the 5G network)—the data network (including the XR server).
  • the XR server According to the user information, the XR server generates new video data and audio data based on the user's scene in the game or in the real scene, and then transmits it back to the helmet, and finally displays it to the user through the helmet, as shown in Figure 1.
  • the service data has strict requirements on the delay. For example, starting from the generation of user location and motion information, if the terminal receives the response data from the XR server within 70 milliseconds (ms), the user feels relatively smooth; if the delay is greater than 70ms (such as 100ms), the user may experience the delay , and then feel dizzy, which affects the experience; if the delay is greater than 150ms, the user feels that the video is stuck and thinks that the XR service is unavailable.
  • ms milliseconds
  • the following is an analysis of the delay of XR service interaction through the path of terminal equipment-5G network-data network, mainly including terminal side delay, network delay and server delay.
  • Terminal-side delay It consists of two parts.
  • the first part includes the delay for the sensor in the helmet to generate the user's position and motion information, the delay for encoding the user information, and the time for delivering the encoded data to the access layer for transmission. extension.
  • the second part includes the delay of buffering, video and audio decoding, display, etc. after receiving the data from the XR server.
  • the total delay of the above two parts is about 25-30ms.
  • Network delay It consists of two parts. The first part is used for uplink transmission. After the terminal receives the data (user location, motion information, etc.) from the upper layer, it transmits the data to the XR server through the 5G network. The second part is used for downlink transmission. After the 5G network receives the response data from the XR server, the delay in transmitting the response data to the terminal. The total delay of the two parts is about 14ms.
  • the processing of the XR server includes generating raw video and audio data, rendering and enhancing virtual objects for the raw video data, video encoding, audio encoding, and encapsulating to form a streaming media file.
  • the above processing flow takes about 26-34ms in total.
  • the XR service data can be divided into two parts: uplink data and downlink data.
  • the uplink data only contains information such as the user's position, line of sight, etc., and the data volume is small; the downlink data contains the video and audio data generated by the XR server, and the data volume is large.
  • the audio data and the video data are encapsulated together, which will not greatly affect the overall service model of the video data. Therefore, the following video data is taken as an example to describe the business model.
  • Fig. 2a is a business model of XR.
  • the XR server generates a video frame at a rate of 60 frames per second, that is, one video frame is generated every 16.67ms.
  • the encoded data stream is divided into multiple data packets according to the standard of 1500 bytes (byte, B) due to IP network limitations. From the perspective of the 5G network, it is reflected in that a group of data packets are received from the XR server every 16.67ms, and the group of data packets includes multiple data packets with a size of about 1500B.
  • a user plane function (UPF) or an access network element (such as a gNB) in the 5G network can receive XR data from the XR server.
  • the gNB receives a group of data packets (ie, one video frame) every 16.67ms, and the group of data packets includes a plurality of data packets with a size of about 1500B.
  • the gNB determines the specific timing for transmitting the set of data packets according to the radio interface load and the priority of the set of data packets and other data to be transmitted, and sends the set of data packets to the terminal device at the specific timing.
  • a set of packets is a set of packets.
  • the UPF/gNB when the XR server continuously sends downlink video data, the UPF/gNB receives a video frame every 16.67ms. After processing, the gNB only transmits the video frame in a part of 16.67ms, and does not transmit data in other time periods, as shown in Figure 2b. However, once the gNB starts to transmit video frames, the terminal device will always keep listening, as shown in Figure 2b. Then the continuous monitoring of the terminal equipment may cause additional power consumption.
  • an embodiment of the present application provides a data receiving method, which can make the terminal device not need to continuously monitor the transmission and reception of data, which is beneficial to reduce the power consumption of the terminal device.
  • FIG. 3a is a schematic diagram of a network system provided by an embodiment of the present application.
  • the network system includes terminal equipment, network element equipment and XR server. Among them, data transmission between the terminal device, the network element and the XR server can be performed in a wireless or wired manner, as shown in Figure 3a.
  • the terminal device may be a device with a wireless transceiver function, or the terminal device may also be a chip.
  • the terminal device may be a user equipment (user equipment, UE), a mobile phone (mobile phone), a tablet computer (Pad), a computer with a wireless transceiver function, a terminal device in virtual reality (virtual reality, VR), augmented reality (augmented reality, AR) terminal equipment, vehicle terminal equipment, remote medical terminal equipment, smart grid (smart grid) terminal equipment, wearable terminal equipment (such as helmets, etc.), Internet of Vehicles, D2D Communication, sensors in machine communication, etc.
  • VR virtual reality
  • AR augmented reality
  • vehicle terminal equipment remote medical terminal equipment
  • smart grid (smart grid) terminal equipment such as helmets, etc.
  • Internet of Vehicles D2D Communication, sensors in machine communication, etc.
  • the network element is an entity that provides network services, including a core network network element, an access network network element, and the like.
  • the core network mainly includes the following key logical network elements: access and mobility management function (AMF), session management function (SMF), UPF, policy control network element ( policy control function, PCF), unified data management network element (unified data management, UDM), as shown in Figure 3b.
  • AMF access and mobility management function
  • SMF session management function
  • UPF policy control network element
  • PCF policy control network element
  • UDM unified data management network element
  • the AMF is mainly responsible for the mobility management in the mobile network, such as user location update, user registration network, user handover and so on.
  • SMF is mainly responsible for session management in the mobile network, such as session establishment, modification and release. Specific functions such as assigning IP addresses to users, choosing to provide packet forwarding functions, etc.
  • PCF is responsible for providing policies to AMF and SMF, such as QoS policies and slice selection policies.
  • UDM is used to store user data, such as subscription information and authentication/authorization information.
  • UPF is mainly responsible for processing user packets, such as forwarding and accounting.
  • the network elements of the access network include: base station (such as gNB, eNB), centralized control unit (central unit control plane, CUCP), centralized unit (central unit, CU), distributed unit (distribution unit, DU), centralized user plane unit (central unit user plane, CUUP), wireless access point (WiFi AP), etc.
  • base station such as gNB, eNB
  • centralized control unit central unit control plane, CUCP
  • centralized unit central unit, CU
  • distributed unit distributed unit
  • DU distributed unit
  • centralized user plane unit central unit user plane, CUUP
  • WiFi AP wireless access point
  • the base station may include CU and DU.
  • the 5G access network can split the functions of the base station in the original access network, deploy some functions of the base station in one CU, and deploy the remaining functions in the DU, and multiple DUs share one CU, which can save costs , and easy network expansion.
  • the segmentation of CU and DU can be divided according to the protocol stack, the radio resource control (radio resource control, RRC), service data adaptation protocol (service data adaptation protocol, SDAP) and packet data convergence protocol (packet data convergence protocol,
  • the PDCP) layer is deployed in the CU, and the rest of the radio link control (radio link control, RLC) layer, the media access control (media access control, MAC) layer and the physical layer (physical, PHY) are deployed in the DU.
  • the CU and DU are connected through the F1 interface.
  • the gNB-CU can represent the gNB to connect to the core network through the NG interface, and the gNB-CU can also represent the gNB to connect to other gNBs through the Xn interface, as shown in Figure 3b.
  • the XR server is an entity that provides XR services, and the XR server is responsible for generating corresponding video frame data according to user instructions.
  • the XR server can send XR data to the network element, or receive feedback information from the network element.
  • the XR server may include an XR encoder for encoding video frames.
  • FIG. 4 is a data receiving method provided by an embodiment of the present application. The method can be executed by a terminal device, and includes the following steps:
  • the terminal device determines a timing configuration, where the timing configuration is used to indicate a notification timing.
  • the terminal device receives first notification information from the network element of the access network at the notification opportunity, where the first notification information is used to indicate the start transmission opportunity and/or the transmission duration of the video frame.
  • the terminal device may determine a timing configuration, and monitor whether a notification message is received at a notification timing indicated by the timing configuration, thereby determining whether to monitor data. That is to say, the terminal device can wake up and monitor data at the notification timing.
  • the method for the terminal device to determine the timing configuration may be that the terminal device receives a notification timing from an access network element.
  • FIG. 5 is a schematic flowchart of a terminal device switching from a sleep state to an activated state and starting to monitor data according to an embodiment of the present application. The process can include the following steps:
  • the access network element configures the notification timing, and sends the notification timing to the terminal device;
  • the terminal device receives the notification timing from the access network element, and determines the timing configuration
  • the terminal device receives the first notification information from the network element of the access network at the notification timing;
  • the terminal device starts to monitor data at the video frame start transmission timing indicated by the first notification information.
  • the terminal device can receive the notification message at the notification timing, and starts monitoring data if the notification message is received. Wherein, that is to say, the terminal device determines the timing configuration according to the received notification timing, switches from the sleep state to the activated state at the notification timing according to the timing configuration, and starts monitoring data.
  • the notification timing configured by the network element of the access network may be determined according to the encoding duration of the XR encoder, or may be determined according to the provisions of the protocol.
  • the access network element eg gNB
  • the access network element eg gNB
  • the access network element eg gNB
  • the XR encoder can send the encoding duration to the gNB, so that the gNB determines the notification timing according to the encoding duration, as shown in Figure 5.
  • the notification timing configured by the gNB is a notification timing every 10ms. It can be understood that this step is optional.
  • the XR encoder can notify the gNB of the encoding duration, and then the encoding duration can be carried in the encoding duration notification message.
  • the specific format of the encoded duration notification message may be a GPRS tunnel protocol (GPRS tunnel protocol, GTP) control packet, or may be a field in the encoded packet header, or a control packet or packet header of other transport layers.
  • the embodiment is not limited.
  • the encoding duration notification message may be transmitted to the gNB via the UPF, and the UPF may decode the notification or transparently transmit the notification (ie, transmit without decoding), which is not limited in this embodiment.
  • the encoded duration notification message may also be transmitted to the gNB via the SMF-AMF, where the SMF decodes the notification message, and the AMF may decode the notification message or transparently transmit the notification message, which is not limited in this embodiment.
  • the gNB can determine the notification timing according to the encoding duration specified in the protocol. That is to say, the gNB can directly determine the notification timing without receiving the encoding duration of the XR encoder to determine the notification timing.
  • the terminal device can also determine the timing configuration by learning. For example, the terminal device first maintains continuous monitoring, and records the timing of receiving the information each time it receives the first notification information sent by the gNB. By learning the timings recorded multiple times, the terminal device can determine the notification timing for the subsequent gNB to send the first notification information to the terminal device, and can also determine the timing configuration.
  • the XR encoder when the XR encoder starts to encode a video frame, the XR encoder can notify the access network element, so that the access network element knows that the XR encoder has started to encode the video frame.
  • the specific content when the XR encoder notifies the network element of the access network may include the following two.
  • the XR encoder may notify the gNB of the information that "the XR encoder starts to encode the video frame", and the information may be carried in the start encoding notification message.
  • the XR encoder may notify the gNB of the information of "encoded data output after 10 ms", and the information may also be carried in the start encoding notification message.
  • the specific format of the start coding notification message may be a GTP control packet, a field in an encoded data packet header, or a control packet or packet header of other transport layers, which is not limited in this embodiment.
  • the start coding notification message may be transmitted to the gNB via the UPF, and the UPF may decode the notification or transparently transmit the notification (ie, transmit without decoding), as shown in FIG. 5 , which is not limited in this embodiment.
  • the start coding notification message can also be transmitted to the gNB via the SMF-AMF, wherein the SMF decodes the notification message, and the AMF can decode the notification message or transparently transmit the notification message, which is not limited in this embodiment.
  • the XR server generates a video frame at a rate of 60 frames per second, that is, one video frame is generated every 16.67 ms.
  • the video frame is encoded and then transmitted. If the XR encoder continuously encodes the video frame, and the encoding time (10ms) is less than the duration of generating the video frame (16.67ms), then the XR encoder does not need to notify the access network element. That is to say, if the XR encoder continuously encodes video frames, the XR encoder does not need to send an encoding start notification message to the access network element.
  • the terminal device may also learn at what timing to start monitoring data according to a pre-obtained semi-static or static notification. That is, the network element of the access network does not need to send the first notification information to the terminal device every time, so that the terminal device starts to monitor video frames. For example, when the base station notifies once, the terminal device starts to monitor the first video frame. When the following video frames arrive, no notification is required, and the terminal device can monitor by itself.
  • the access network element may determine the time when the video frame arrives at the access network element. Then, the network element of the access network can send the first notification information to the terminal device at the notification timing, so that the terminal device wakes up and receives data at the corresponding moment. For example, after the gNB receives the start encoding notification message, it knows that the XR encoder will output encoded video frames after 10ms. Combined with the prior information "XR encoder-UPF-gNB transmission delay is 2ms", gNB determines that "a video frame will arrive soon after 8ms". The gNB notifies the terminal device to wake up and receive video frames through the next notification timing, as shown in Figure 5.
  • the first notification information sent by the access network element to the terminal device may be an RRC message, SDAP layer control signaling, PDCP layer control signaling, RLC layer control signaling, MAC layer control signaling, and PHY layer control signaling
  • RRC message Secure Digital
  • SDAP layer control signaling PDCP layer control signaling
  • RLC layer control signaling RLC layer control signaling
  • MAC layer control signaling PHY layer control signaling
  • PHY layer control signaling PHY layer control signaling
  • the terminal device may determine at which time to start monitoring data according to the first notification information. For example, the first notification information sent by the gNB to the terminal device is "there will be video frame transmission after 5ms", and the terminal device receives the notification from the subframe/slot (slot)/transmission time interval (transmission time interval, TTI), to the Push back 5ms to start monitoring data.
  • the first notification information sent by the gNB to the terminal device is "there will be video frame transmission after 5ms”
  • TTI transmission time interval
  • the first notification information is used to indicate the start transmission timing and/or transmission duration of the video frame.
  • the start transmission timing of the video frame described in this embodiment may also be referred to as the subframe/time slot/transmission time interval where the first data packet of the video frame is located, the data start timing of the video data, etc. The same meaning is not limited in this embodiment.
  • the transmission duration of the video frame described in this embodiment may also be referred to as the total number of subframes/time slots/transmission time intervals occupied by the video frame, etc., which is not limited in this embodiment.
  • the start transmission timing and/or the transmission duration of the video frame may be carried in the first notification information at the same time, or may be carried in different notification information.
  • the first notification information is only used to indicate the start transmission timing of the video frame, and does not carry the transmission duration of the video frame.
  • the terminal device may also determine the timing to start monitoring the video data according to the first notification information.
  • the terminal device monitors the video frame at the start transmission timing of the video frame, and switches the working state to the sleep state at the end transmission timing corresponding to the transmission duration of the video frame. That is to say, the first notification information determines the timing when the terminal device starts monitoring data and ends monitoring data, so that data is only monitored during the transmission duration, and data is not monitored at other timings. It can be understood that the terminal device can determine the end transmission timing of the video frame according to the start transmission timing and the transmission duration of the video frame.
  • the XR server continuously sends multiple video frames to the access network element, and the access network element can transmit the multiple video frames.
  • the terminal device can receive the multiple video frames.
  • the terminal device can learn the multiple video frames to determine the arrival time interval of the video frames.
  • the arrival time interval of the video frames is the time difference between the start transmission timings of the two video frames.
  • the terminal device can determine the next time to monitor the data according to the arrival time interval of the video frame.
  • the terminal device receives K video frames, and the arrival time interval of the K video frames is a fixed value. Then, the terminal device can learn the K video frames, and determine that the arrival time interval between any video frame k-1 in the K video frames and its adjacent video frame k is 10 ms. Then, when the terminal device receives subsequent video frames, it may determine that 10 ms after the start transmission timing of the Kth video frame, start monitoring data to receive the next video frame.
  • the terminal device may receive second notification information from the access network element, where the second notification information is used to indicate the next video frame adjacent to the current video frame start transmission timing.
  • the terminal device monitors the data at the start transmission timing of the next video frame. That is to say, if the video frame is not transmitted according to a certain time pattern, the access network element needs to notify the terminal device when to start monitoring data, thereby reducing the power consumption of the terminal device.
  • the XR encoder does not continuously encode video frames, when the XR encoder continues to encode video frames, it needs to send an encoding start notification message to the gNB again.
  • the specific time length of the period of time may be determined by the XR encoder itself, or may be configured by the network management platform, which is not limited in this embodiment.
  • the network element of the access network buffers the first video frame and then transmits it after receiving the video frame, and directly transmits the adjacent second video frame without buffering, then the two video frames are transmitted directly.
  • the subframe where the first data packet of the frame is located the interval of which will be smaller than the arrival time interval (eg 16ms) of adjacent video frames.
  • the network element of the access network needs to send the second notification information to the terminal device before transmitting the second video frame, where the second notification information is used to instruct the terminal device to switch from the sleep state to the second time domain position of the activated state again, That is, the timing to start monitoring again.
  • the format of the second notification information may be one or more of the following: RRC message, DAP layer control signaling, PDCP layer control signaling, RLC layer control signaling, MAC layer control signaling, and physical layer control signaling Wait.
  • the second notification information may be transmitted at the original notification timing, or the second notification information and the first video frame may be transmitted through the same subframe. Wherein, if the second notification information and the first video frame are transmitted through the same subframe, the terminal device does not need to monitor the notification timing when monitoring the video frame.
  • the network element of the access network may send third notification information to the terminal device after transmitting all the data of one video frame, where the third notification information is used to indicate the end of the transmission of the video frame data.
  • the terminal device stops monitoring data, and switches from the activated state to the dormant state.
  • FIG. 6 is a sequence diagram of a terminal device switching from a sleep state to an activated state and starting to monitor data according to an embodiment of the present application.
  • a square in the timing diagram represents a time domain position, that is, a subframe/slot/TTI.
  • a square represents a subframe, and the time length of a subframe is 1 ms.
  • the following takes a square to represent a subframe as an example, and describes in detail the sequence of the terminal device switching from the sleep state to the activated state and starting to monitor data.
  • the time domain position at which the network element of the access network starts to transmit the video frame to the terminal device through the air interface is subframe N, as shown by the shaded part in FIG. 6 .
  • the video frames sent by the XR server are continuous, that is, the XR encoder will continuously send video frames to the network element of the access network, and the arrival time interval of adjacent video frames is assumed to be 16ms.
  • the transmission of the video frame needs to occupy three subframes, as shown by the shaded part in FIG. 6 .
  • the timing configuration determined by the terminal device indicates that the notification timing of the terminal device is every 10ms, that is, the terminal device wakes up every 10ms to detect whether the first notification message is received.
  • the terminal device wakes up at this subframe N-5 and detects whether the first notification message is received.
  • the above-mentioned time domain position is a time period during which the terminal device is in an activated state.
  • the start coding notification message received by the gNB includes the information of "data output encoded after 10ms", and it is known that the "XR encoder-UPF-gNB transmission delay is 2ms", then the subframe N moves forward, At the time corresponding to subframe N-12, the XR encoder sends a start coding notification message to the gNB. At subframe N-10, the gNB receives the start coding notification message from the XR encoder, and at the next notification timing corresponding to subframe N- At 5, the first notification information is sent to the terminal device, as shown in FIG. 6 .
  • the XR encoder sends a video frame to the UPF at the moment corresponding to subframe N-2, and the gNB receives the video frame at subframe N and transmits the video frame to the terminal device through the air interface.
  • the transmission of the video frame is completed at subframe N+2, and the terminal device can switch from the activated state to the sleep state.
  • the terminal device can record the time domain position of the first data packet of the video frame, that is, subframe N, and switch from the sleep state to the active state again at subframe N+16, and receive the video frame again, as shown in Figure 6 Show.
  • the terminal device before the start of the first video frame, the terminal device only needs to monitor whether there is a first notification message at each notification opportunity. If receiving the first notification message, the terminal device only needs to monitor the first video frame in subframes N, N+1, N+2, and then starts to monitor the second video frame in subframe N+16, and so on.
  • the network element of the access network may also buffer the data, and select an appropriate time to transmit it through the wireless air interface under the premise of meeting the service delay budget.
  • the gNB knows "Data output encoded after 10ms", and knows “XR encoder-UPF-gNB transmission delay of 2ms”. Then, if the gNB buffers the data and then sends it out, the first notification information sent by the gNB to the terminal device may be "data transmission after 6ms", that is to say, the duration of the video frame buffering by the gNB is 2ms.
  • the above solution assumes that the terminal device has only one video service, and as long as there is no video frame transmission in the video service, the terminal device enters a sleep state. However, if the terminal device has other services in parallel at the same time, the terminal device also needs to consider whether other services have data transmission, so as to determine whether to enter the sleep state.
  • the terminal device performs two services at the same time, including service 1 for watching streaming video and service 2 for making video calls, it is assumed that the encoding duration of the video frame of service 1 is 10ms, and the encoding duration of the video frame of service 2 is 7ms, the encoding durations of the two are inconsistent, so the notification timings of the two services are also inconsistent.
  • the network element of the access network may superimpose the notification timing of service 1 and service 2 to determine the notification timing of the terminal device.
  • the terminal device receives the first notification information at the notification timing, and starts monitoring data at the timing indicated by the first notification information. After the transmission of both service 1 and service 2 is completed (for example, the base station may notify the terminal equipment that all service transmissions are completed), the terminal equipment ends monitoring and resumes the sleep state.
  • the embodiment of the present application provides a data receiving method, and the method can be executed by a terminal device.
  • the terminal device may determine the timing configuration, and receive the first notification information from the access network element at the notification timing indicated by the timing configuration, so as to determine the start transmission timing and/or transmission duration of the video frame. It can be seen that the terminal device only receives the first notification information of the access network element at the notification timing, and determines the starting transmission timing and/or the transmission duration of the video frame according to the first notification information.
  • the terminal device only monitors data at a specific timing (eg, the initial transmission timing of the video frame), which is beneficial to reduce the power consumption of the terminal device.
  • FIG. 7 is another data receiving method provided by an embodiment of the present application. The method can be executed by a terminal device and includes the following steps:
  • the terminal device determines the initial transmission timing, transmission duration, and arrival time interval of the video frames of each video frame according to the K video frames monitored in advance;
  • the terminal device determines the initial transmission timing and/or transmission duration of video frame k+1 according to the initial transmission timing, transmission duration, and arrival time interval of video frame k in the K video frames.
  • the terminal device in the initial stage of the video service, can continuously monitor and record the time domain position of each data packet in the video service, thereby determining the time domain position of a video frame corresponding to a cluster of data packets.
  • a terminal device continuously receives video service data (such as including K video frames), and can determine which data packets constitute a cluster of data packets corresponding to a video frame through machine learning, so as to determine the start transmission timing of the video frame, and the transmission time interval between each video frame.
  • the video frame k is the last transmitted video frame among the K video frames received when the terminal device keeps data monitoring.
  • Video frame k+1 is the first video frame that arrives after video frame k.
  • the terminal device can learn to determine the start transmission timing, transmission duration, and arrival time interval of each video frame in the K video frames in the early stage of the video service. Then, after the kth video frame, the terminal device can calculate the initial transmission timing of video frame k+1 according to the initial transmission timing of video frame k and the arrival time interval of the video frame, and calculate the initial transmission timing of video frame k+1.
  • the start transmission timing starts monitoring data. And so on, in the subsequent data monitoring process, it is not necessary to continuously monitor the data, but to monitor the data at a specified time.
  • the terminal device learns K video frames, and determines that the initial transmission timing of video frame k is subframe N, the transmission duration of video frame k is 3 subframes, and the arrival time interval of video frames is 16 subframes. , and the arrival time interval of video frames is a fixed value. Then the terminal device can determine that the starting transmission opportunity of subframe k+1 is subframe N+16, and the terminal device can start monitoring data at subframe N+16. Similarly, the terminal device may also determine that the start transmission timing of subframe k+2 is subframe N+32, and then the terminal device may start monitoring data at subframe N+32.
  • the terminal device may start monitoring at the reference point of the first time domain position where the first data packet of the k+1 th video frame is located in advance. data. For example, if the first time domain position indicates subframe N+16, the terminal device may start monitoring data at subframe N+14.
  • the specific advance time amount may be configured by the network element of the access network, or by the network management, or by the terminal device, which is not limited in this embodiment.
  • the terminal device monitors the video frame k+1 at the start transmission timing of the video frame k+1, and switches the working state to the sleep state at the end transmission timing corresponding to the transmission duration of the video frame k+1. That is to say, the terminal device starts monitoring data at the start of the video frame transmission timing, stops monitoring the data at the end of the transmission timing, and switches to a sleep state.
  • the start transmission timing of video frame k is subframe N
  • the transmission duration of video frame k is 3 subframes
  • the arrival time interval of video frames is 16 subframes
  • the arrival time interval of video frames is a fixed value.
  • the terminal device can determine that the starting transmission timing of subframe k+1 is subframe N+16, and the terminal device can start monitoring data at subframe N+16 and stop monitoring data at subframe N+19.
  • the terminal device does not need to monitor the data any more. is dormant.
  • the terminal device determines the transmission duration of each video frame, which may be implemented through the following manners 1 to 5. That is to say, the terminal device can determine the transmission duration or end transmission timing of the video frame k by using the following methods 1 to 5, stop monitoring the video data at the end transmission timing, and restart until the start transmission timing of the next video frame. Monitor video data.
  • the terminal device may also use the following manners 1 to 5 to determine the duration of transmission of the video frame k+1 or the timing for ending the transmission, which is not limited in this embodiment.
  • the access network element adds a video frame end identifier to the last data packet of each video frame, or adds a video frame end identifier to the transmission block where the last data packet of each video frame is located, as shown in Figure 8a.
  • the base station may add the video frame end flag to the last data packet of each video frame by adding the video frame end flag to the MAC/RLC/PDCP/SDAP subheader.
  • the base station may add a MAC control element (MAC control element, MAC CE), and the MAC CE may only include a MAC subheader.
  • the terminal device can determine the last data packet of the video frame according to the MAC CE, thereby determining the transmission end moment of the video frame.
  • Mode 2 The first subframe/timeslot/TTI occupied by the access network element in each video frame, set the subframe/timeslot/TTI to carry "The current video frame occupies X subframes/timeslots/TTI. ” notification information, as shown in Figure 8b.
  • the terminal device can determine the transmission end moment of the current video frame according to the information.
  • the first subframe occupied by the video frame shown in Figure 8b is subframe N
  • the base station sets the subframe to carry the "current video frame" in the first subframe occupied by the video frame shown in Figure 8b account for 3 TTI" notification information.
  • the terminal device may determine that the transmission end time of the video frame is subframe N+2 according to the position of the first subframe occupied by the video frame and the occupied TTI.
  • the notification information may be carried in the first subframe/slot/TTI corresponding to the video frame, or may be carried in the Nth subframe/slot/TTI corresponding to the video frame, which is not limited in this embodiment. If the notification information is carried in the first subframe/slot/TTI corresponding to the video frame, the base station may transmit the video frame through consecutive or non-consecutive subframes/slots/TTI.
  • the notification information may also be a MAC CE.
  • Mode 3 Multiple data packets corresponding to each video frame occupy consecutive multiple subframes/slots/TTIs during transmission on the wireless interface.
  • the terminal device starts to receive video frames from the start time of each cluster of data packets, and keeps receiving it continuously until no video data is detected in a certain subframe/slot/TTI, then the terminal device considers that the video frame has been transmitted. end, as shown in Figure 8c. It should be noted that, in the application scenario of Mode 3, the network element of the access network needs to ensure that continuous video frames are transmitted.
  • Mode 4 In the first subframe/slot/TTI corresponding to each video frame, the network element of the access network sets the subframe/slot/TTI to carry "the number of data packets/data amount contained in the current video frame. ” notification information, as shown in Figure 8d.
  • the terminal device receives a corresponding number of data packets/data volume, and considers that the transmission of the video frame has ended. For example, in the first subframe occupied by the video frame shown in FIG. 8d, the base station sets the subframe to carry notification information that "the current video frame contains 50 data packets/75000 bytes". Then, the terminal device can determine the transmission end moment of the video frame according to the position of the first subframe occupied by the video frame and the size of the video frame.
  • the notification information may be carried in the first subframe/slot/TTI corresponding to the video frame, or may be carried in the Nth subframe/slot/TTI corresponding to the video frame, which is not limited in this embodiment. If the notification information is carried in the first subframe/slot/TTI corresponding to the video frame, the base station may transmit the video frame through consecutive or non-consecutive subframes/slots/TTI.
  • the notification information may also be a MAC CE.
  • Method 5 The terminal device reads the real-time transport control protocol (RTP) header in the data packet of the video frame, and confirms whether the data packet is the last data packet of the current video frame.
  • RTP real-time transport control protocol
  • the base station may set indication information in the RTP packet header in the data packet of the video frame, and the indication information may be identified by one or more bits.
  • the terminal device can confirm the transmission end time of each video frame, and no longer monitor data at the transmission end time. Further, the terminal device can also combine the transmission start time of each video frame, the transmission end time of each video frame, and the arrival time interval of adjacent video frames to learn and record the transmission process of the received video frames, so as to calculate The start time of the next video frame, start monitoring the next video frame at the start time of the next video frame.
  • the terminal device may notify the access network element of the initial transmission timing and continuous transmission duration of the video frame through RRC signaling or MAC CE.
  • the access network element reconfirms whether the initial transmission opportunity and the continuous transmission duration are the initial transmission when the access network element actually transmits the video frame according to the initial transmission timing and continuous transmission duration of the video frame. timing and duration of continuous transmission to ensure reliable transmission.
  • the base station can also notify the terminal equipment in time according to the initial transmission timing and duration of transmission sent by the terminal Data may not be received within the time period of the initial transmission timing and/or the continuous transmission duration, and data monitoring is suspended.
  • the terminal device notifies the base station of the result of its own learning, and the base station can cooperate with the behavior of the terminal device according to the result. For example, for the video data of the terminal device, the base station will not buffer some video frames in the video data of the terminal device, and some video frames will not be buffered, resulting in abnormal reception of video frames of the terminal device.
  • the start transmission timing of video frames, the transmission duration of video frames, the arrival time interval of video frames, etc. are described in units of subframes or slots. In fact, any other time unit may be used for description without affecting the technical essence of the present application.
  • OFDM orthogonal frequency division multiplexing
  • milliseconds, microseconds, nanoseconds and other time units can also be used to describe the start transmission timing of video frames, the transmission duration of video frames, video
  • the arrival time interval of the frame, etc. is not limited in this embodiment.
  • the embodiment of the present application provides a data receiving method, and the method can be executed by a terminal device.
  • the terminal device can continuously monitor the data in the early stage of the video service, learn the start transmission timing and end transmission timing of the video frame by itself, so as to estimate the start transmission timing and/or end transmission timing of the subsequent video frame, only in the subsequent video frame.
  • the start transmission timing starts monitoring data, and switches to the sleep state at the end transmission timing. It can be seen that the terminal device learns the start transmission timing and the end transmission timing of the video frame by itself, thereby determining to monitor data at a specific timing without continuously monitoring the transmission and reception of data, which is beneficial to reduce the power consumption of the terminal device.
  • the terminal device according to the embodiment of the present application will be described in detail below with reference to FIG. 9 to FIG. 12 .
  • An embodiment of the present application provides a data receiving apparatus. As shown in FIG. 9 , the data receiving apparatus 900 may be used to implement the data receiving method in the embodiment of the present application.
  • the data receiving apparatus 900 may include:
  • a processing unit 901 configured to determine a timing configuration, where the timing configuration is used to indicate a notification timing
  • the transceiver unit 902 is configured to receive the first notification information from the access network element at the notification opportunity; the first notification information is used to indicate the start transmission opportunity and/or the transmission duration of the video frame k.
  • the processing unit 901 is further configured to monitor the video frame at the start transmission timing of the video frame, and switch the working state to the sleep state at the end transmission timing corresponding to the transmission duration of the video frame.
  • the transceiver unit 902 is further configured to receive a notification timing from the access network element, where the notification timing is determined by the access network element according to the encoding duration of the video frame.
  • the transceiver unit 902 is further configured to receive multiple video frames; the processing unit 901 is further configured to determine the arrival time interval of the video frames according to the service model of the multiple video frames. Wherein, the arrival time interval of the video frames is the time difference between the start transmission timings of the two video frames.
  • the processing unit 901 is further configured to switch the working state to the sleep state at the end transmission timing corresponding to the transmission duration of the video frame, after the arrival of the video frame Time interval, switch from sleep state to working state, monitor the next video frame adjacent to the video frame.
  • the transceiver unit 902 is further configured to receive second notification information from the network element of the access network, where the second notification information is used to indicate the same period as the video frame.
  • the processing unit 901 is further configured to monitor the next video frame at the start transmission opportunity of the next video frame.
  • the start transmission opportunity and/or the end transmission opportunity of the video frame is one or more of a subframe, a time slot or a transmission time interval.
  • FIG. 10 is a schematic structural diagram of a terminal device provided by an embodiment of the present application.
  • the terminal device may be a device (such as a chip) having the data receiving function described in the embodiment of the present application.
  • the terminal device 1000 may include a transceiver 1001 , at least one processor 1002 and a memory 1003 .
  • the transceiver 1001, the processor 1002 and the memory 1003 may be connected to each other through one or more communication buses, or may be connected to each other in other ways. In this embodiment, a bus connection is used as an example, as shown in FIG. 10 .
  • the transceiver 1001 may be used to transmit or receive data.
  • the transceiver 1001 may receive the first notification information. It is understood that the transceiver 1001 is a general term and may include a receiver and a transmitter.
  • the processor 1002 may be used to process data.
  • the processor 1002 may include one or more processors, for example, the processor 1002 may be one or more central processing units (CPUs), network processors (NPs), hardware chips, or any combination thereof .
  • the processor 1002 is a CPU, the CPU may be a single-core CPU or a multi-core CPU.
  • the memory 1003 is used for storing program codes and the like.
  • the memory 1003 may include volatile memory, such as random access memory (RAM).
  • the memory 1003 may also include non-volatile memory (non-volatile memory), such as read-only memory (ROM), flash memory (flash memory), hard disk drive (HDD) or solid state hard disk ( solid-state drive, SSD).
  • non-volatile memory such as read-only memory (ROM), flash memory (flash memory), hard disk drive (HDD) or solid state hard disk ( solid-state drive, SSD).
  • ROM read-only memory
  • flash memory flash memory
  • HDD hard disk drive
  • SSD solid state hard disk
  • processor 1002 may be used to implement the data receiving method in the embodiment of the present application, wherein the specific implementation manner is as follows:
  • the timing configuration is used to indicate the notification timing
  • the first notification information from the network element of the access network is received at the notification opportunity; the first notification information is used to indicate the start transmission opportunity and/or the transmission duration of the video frame.
  • the processor 1002 is further configured to monitor the video frame at the start transmission timing of the video frame, and switch the working state to the sleep state at the end transmission timing corresponding to the transmission duration of the video frame.
  • the transceiver 1001 receives a notification timing from an access network element, where the notification timing is determined by the access network element according to the encoding duration of the video frame.
  • the transceiver 1001 is further configured to receive multiple video frames; the processor 1002 is further configured to determine the arrival time interval of the video frames according to the service model of the multiple video frames. Wherein, the arrival time interval of the video frames is the time difference between the start transmission timings of the two video frames.
  • the processor 1002 is further configured to switch the working state to the sleep state at the end transmission timing corresponding to the transmission duration of the video frame, after the arrival of the video frame Time interval, switch from sleep state to working state, monitor the next video frame adjacent to the video frame.
  • the transceiver 1001 is further configured to receive second notification information from the network element of the access network, where the second notification information is used to indicate the same period as the video frame.
  • the processor 1002 is further configured to monitor the next video frame at the start transmission opportunity of the next video frame.
  • the start transmission opportunity and/or the end transmission opportunity of the video frame is one or more of a subframe, a time slot or a transmission time interval.
  • An embodiment of the present application provides another data receiving apparatus.
  • the data receiving apparatus 1100 may be used to implement the data receiving method in the embodiment of the present application.
  • the data receiving apparatus 1100 may include:
  • the processing unit 1101 is used to determine the initial transmission timing, the transmission duration, and the arrival time interval of the video frames of each video frame according to the K video frames monitored in advance;
  • the processing unit 1101 is further configured to determine the initial transmission opportunity and/or the transmission duration of the video frame k+1 according to the initial transmission opportunity of the video frame k in the K video frames, the transmission duration, and the arrival time interval of the video frame. .
  • the processing unit 1101 is further configured to monitor the video frame k+1 at the start transmission timing of the video frame k+1, and change the working state at the end transmission timing corresponding to the transmission duration of the video frame k+1 Switch to sleep state.
  • the processing unit 1101 is further configured to determine the transmission duration of the video frame k according to the start transmission timing of the video frame k and the end identifier of the video frame k.
  • the processing unit 1101 is further configured to determine the transmission duration of the video frame k according to the start transmission timing and the temporal position information of the video frame k.
  • the processing unit 1101 is further configured to The initial transmission timing of frame k and the transmission timing when no data is monitored determine the duration of transmission of video frame k.
  • the processing unit 1101 is further configured to determine the transmission duration of the video frame k according to the start transmission timing of the video frame k and the data amount of the video frame k.
  • the processing unit 1101 is further configured to determine the transmission duration of the video frame k according to the start transmission timing of the video frame k and the real-time transmission protocol RTP packet header in the data packet of the video frame k.
  • the data receiving apparatus 1100 further includes a transceiver unit 1102, and the transceiver unit 1102 is configured to send the start transmission timing and transmission duration of the video frame to the access network element.
  • the video frame k is the last video frame transmitted among the K video frames, and the video frame k+1 is the first video frame that arrives after the video frame k.
  • FIG. 12 is a schematic structural diagram of another terminal device provided by an embodiment of the present application.
  • the terminal device may be a device (eg, a chip) having the function of performing the data receiving described in the embodiment of the present application.
  • the terminal device 1200 may include a transceiver 1201 , at least one processor 1202 and a memory 1203 .
  • the transceiver 1201, the processor 1202 and the memory 1203 may be connected to each other through one or more communication buses, or may be connected to each other in other ways. In this embodiment, a bus connection is used as an example, as shown in FIG. 12 .
  • the transceiver 1201 may be used to transmit or receive data.
  • the transceiver 1201 may send the start transmission timing and transmission duration of the video frame to the access network element. It can be understood that the transceiver 1201 is a general term and may include a receiver and a transmitter.
  • the processor 1202 may be used to process data.
  • the processor 1202 may include one or more processors, for example, the processor 1202 may be one or more central processing units (CPUs), network processors (NPs), hardware chips, or any combination thereof .
  • the processor 1202 is a CPU, the CPU may be a single-core CPU or a multi-core CPU.
  • the memory 1203 is used for storing program codes and the like.
  • the memory 1203 may include volatile memory, such as random access memory (RAM).
  • the memory 1203 may also include non-volatile memory (non-volatile memory), such as read-only memory (ROM), flash memory (flash memory), hard disk drive (HDD) or solid state hard disk ( solid-state drive, SSD).
  • non-volatile memory such as read-only memory (ROM), flash memory (flash memory), hard disk drive (HDD) or solid state hard disk ( solid-state drive, SSD).
  • ROM read-only memory
  • flash memory flash memory
  • HDD hard disk drive
  • SSD solid state hard disk
  • the memory 1203 may also include a combination of the above-described types of memory.
  • processor 1202 may be used to implement the data receiving method in the embodiment of the present application, wherein the specific implementation manner is as follows:
  • K video frames monitored in advance determine the initial transmission timing of each video frame, the transmission duration, and the arrival time interval of the video frames
  • the processor 1202 is further configured to monitor the video frame k+1 at the start transmission timing of the video frame k+1, and change the working state at the end transmission timing corresponding to the transmission duration of the video frame k+1 Switch to sleep state.
  • the processor 1202 is further configured to determine the transmission duration of the video frame k according to the start transmission timing of the video frame k and the end identifier of the video frame k.
  • the processor 1202 is further configured to determine the transmission duration of the video frame k according to the start transmission timing and the temporal position information of the video frame k.
  • the processor 1202 is further configured to The initial transmission timing of frame k and the transmission timing when no data is monitored determine the duration of transmission of video frame k.
  • the processor 1202 is further configured to determine the transmission duration of the video frame k according to the start transmission timing of the video frame k and the data amount of the video frame k.
  • the processor 1202 is further configured to determine the transmission duration of the video frame k according to the initial transmission timing of the video frame k and the real-time transmission protocol RTP packet header in the data packet of the video frame k.
  • the transceiver 1201 is configured to send the start transmission timing and transmission duration of the video frame to the network element of the access network.
  • the video frame k is the last video frame transmitted among the K video frames, and the video frame k+1 is the first video frame that arrives after the video frame k.
  • the embodiments of the present application provide a computer-readable storage medium, where a program or an instruction is stored in the computer-readable storage medium, and when the program or the instruction runs on a computer, the computer executes the data receiving method in the embodiment of the present application.
  • An embodiment of the present application provides a chip or a chip system, the chip or chip system includes at least one processor and an interface, the interface and the at least one processor are interconnected by a line, and the at least one processor is used to run a computer program or instruction to perform the present application The data receiving method in the embodiment.
  • the interface in the chip may be an input/output interface, a pin or a circuit, or the like.
  • the chip system in the above aspects may be a system on chip (system on chip, SOC), or a baseband chip, etc.
  • the baseband chip may include a processor, a channel encoder, a digital signal processor, a modem, an interface module, and the like.
  • the chip or chip system described above in this application further includes at least one memory, where instructions are stored in the at least one memory.
  • the memory may be a storage unit inside the chip, such as a register, a cache, etc., or a storage unit of the chip (eg, a read-only memory, a random access memory, etc.).
  • a computer program product includes one or more computer instructions.
  • the computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable device.
  • Computer instructions may be stored on or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from a website site, computer, server, or data center over a wire (e.g.
  • Coaxial cable, optical fiber, digital subscriber line (Digital Subscriber Line, DSL)) or wireless (such as infrared, wireless, microwave, etc.) means to transmit to another website site, computer, server or data center.
  • a computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device such as a server, a data center, or the like that includes an integration of one or more available media.
  • the available media may be magnetic media (eg, floppy disks, hard disks, magnetic tapes), optical media (eg, high-density digital video discs (DVDs)), or semiconductor media (eg, solid state disks, SSD)) etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

本申请实施例公开了一种数据接收方法、装置及终端设备,其中,该方法中终端设备可以确定时机配置,并在时机配置指示的通知时机接收来自接入网网元的第一通知信息,从而确定视频帧的起始传输时机和/或传输持续时长。可见,终端设备只在通知时机接收接入网网元的第一通知信息,并根据该第一通知信息,确定在视频帧的起始传输时机和/或传输持续时长。终端设备在特定的时机才监听数据,有利于降低终端设备的耗电。

Description

一种数据接收方法、装置及终端设备 技术领域
本申请涉及通信技术领域,尤其涉及一种数据接收方法、装置及终端设备。
背景技术
扩展现实(extended reality,XR)是指通过计算机技术和可穿戴设备产生的一个真实与虚拟组合、可人机交互的环境,是增强现实(augmented reality,AR)、虚拟现实(virtual reality,VR)、混合现实(mixed reality,MR)等多种形式的统称。三者视觉交互技术融合,实现虚拟世界与现实世界之间无缝转换的“沉浸感”体验。
以下行视频数据为例,XR服务器按每秒60帧的速率,生成视频帧,即每16.67ms生成一个原始视频帧。原始视频帧经过编码后生成编码后的视频数据,接入网网元(如gNB)通过用户面网元(user plane function,UPF)从XR服务器接收编码后的视频数据,并根据无线空口负荷确定传输时机,在传输时机传输视频数据。对应的,终端设备每16.67ms接收一个视频数据帧。其中,接入网网元经过处理,只在16.67ms中一部分时间段内传输该视频数据,在其它时间段则不传输。但是,终端设备将始终保持监听状态,造成额外耗电。
发明内容
本申请实施例提供一种数据接收方法、装置及终端设备,该方法可以使终端设备不需要持续监听数据的收发,有利于降低终端设备的耗电。
第一方面,本申请实施例提供一种数据接收方法,该方法可以由终端设备所执行。其中,终端设备可以确定时机配置,并在时机配置指示的通知时机接收来自接入网网元的第一通知信息,从而确定视频帧的起始传输时机和/或传输持续时长。
可见,终端设备只在通知时机接收接入网网元的第一通知信息,并根据该第一通知信息,确定在视频帧的起始传输时机和/或传输持续时长。终端设备在特定的时机(如视频帧的起始传输时机)才监听数据,有利于降低终端设备的耗电。
在一种可能的设计中,终端设备还可以在视频帧的起始传输时机监听该视频帧,并且在该视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
可见,终端设备在视频帧的起始传输时机开始监听数据,并且在视频帧的结束传输时机不再监听数据,从而降低终端设备的耗电。
在一种可能的设计中,终端设备接收来自接入网网元的通知时机,该通知时机是接入网网元根据视频帧的编码时长确定的。
可见,该通知时机可以是接入网网元根据视频帧的编码时长确定的,那么针对不同业务的视频帧,接入网网元可以确定不同的通知时机来适配不同的业务,有利于设置更精确的通知时机,从而更好地降低终端设备的耗电。
在一种可能的设计中,终端设备可以接收多个视频帧,根据多个视频帧的业务模型,确定视频帧的到达时间间隔。其中,视频帧的到达时间间隔为两个视频帧的起始传输时机之间的时间差。
可见,终端设备可以根据视频帧的业务模型确定相邻视频帧的到达时间间隔,从而确定在哪些时域位置处终端设备可以处于休眠状态,降低终端设备的耗电。
在一种可能的设计中,若视频帧的到达时间间隔为固定值,那么终端设备在视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态之后,经过视频帧的到达时间间隔,终端设备可以从休眠状态切换为工作状态,监听与该视频帧相邻的下一视频帧。
可见,当视频帧的到达时间间隔为固定值时,终端设备可以维持休眠状态,并根据该到达时间间隔确定下一次开始监听数据的时机,有利于降低终端设备的耗电。
在一种可能的设计中,若视频帧的到达时间间隔不为固定值,终端设备接收来自接入网网元的第二通知信息,该第二通知信息用于指示与视频帧相邻的下一视频帧的起始传输时机。终端设备在下一视频帧的起始传输时机监听下一视频帧。
可见,若相邻视频帧的到达时间间隔不为固定值,那么接入网网元需要再次通知终端设备开始接收数据的时机,避免终端设备一直处于休眠状态或者一直处于监听数据的状态。
在一种可能的设计中,视频帧的起始传输时机和/或结束传输时机为子帧、时隙或传输时间间隔中的一种或多种。
第二方面,本申请实施例提供另一种数据接收方法,该方法由终端设备所执行。其中,终端设备可以预先监听K个视频帧,并对该K个视频帧进行学习,确定各个视频帧的起始传输时机、传输持续时长、视频帧的到达时间间隔。终端设备可以根据K个视频帧中的视频帧k的起始传输时机、传输持续时长、视频帧的到达时间间隔,确定视频帧k+1的起始传输时机和/或传输持续时长。
可见,终端设备通过自行学习视频帧的起始传输时机和/或结束传输时机,从而确定在特定的时机才监听数据而不需要持续监听数据的收发,有利于降低终端设备的耗电。
在一种可能的设计中,终端设备还可以在视频帧k+1的起始传输时机监听该视频帧k+1,并且在视频帧k+1的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
可见,终端设备可以确认每一个视频帧的传输结束时刻,并在传输结束时刻不再监听数据,有利于降低终端设备的耗电。
在一种可能的设计中,终端设备根据视频帧k的起始传输时机和视频帧k的结束标识,确定所述视频帧k的传输持续时长。
在一种可能的设计中,终端设备根据视频帧k的起始传输时机和时域位置信息,确定视频帧k的传输持续时长。
在一种可能的设计中,若终端设备在接收到视频帧k的第一个数据包之后,在连续的子帧/时隙/传输时间间隔上未监听到数据,终端设备根据视频帧k的起始传输时机和未监听到数据的传输时机,确定视频帧k的传输持续时长。
在一种可能的设计中,终端设备根据视频帧k的起始传输时机和视频帧k的数据量,确定视频帧k的传输持续时长。
在一种可能的设计中,终端设备根据视频帧k的起始传输时机和视频帧k的数据包中的实时传输协议RTP包头,确定视频帧k的传输持续时长。
在一种可能的设计中,终端设备向接入网网元发送该视频帧的起始传输时机和传输持续时长。
在一种可能的设计中,视频帧k为K个视频帧中最后一个传输的视频帧,视频帧k+1为视频帧k之后到达的第一个视频帧。
第三方面,本申请实施例提供一种数据接收装置,该装置具有实现第一方面所提供的数据接收方法的功能。该功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。该硬件或软件包括一个或多个与上述功能相对应的模块。
第四方面,本申请实施例提供一种数据接收装置,该装置具有实现第二方面所提供的数据接收方法的功能。该功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。该硬件或软件包括一个或多个与上述功能相对应的模块。
第五方面,本申请实施例提供一种终端设备,该终端设备包括处理器,用于实现上述第一方面中所涉及的功能或方法,该终端设备例如可以是芯片系统。在一种可行的实现方式中,所述终端设备还包括存储器,所述存储器,用于保存实现上述第一方面所述方法的功能必要的程序指令和数据。
第六方面,本申请实施例提供一种终端设备,该终端设备包括处理器,用于实现上述第二方面中所涉及的功能或方法,该终端设备例如可以是芯片系统。在一种可行的实现方式中,所述终端设备还包括存储器,所述存储器,用于保存实现上述第二方面所述方法的功能必要的程序指令和数据。
第七方面,本申请实施例提供一种计算机可读存储介质,该计算机可读存储介质包括程序或指令,当所述程序或指令在计算机上运行时,使得计算机执行第一方面或第一方面中任一种可能实现方式中的方法。
第八方面,本申请实施例提供一种计算机可读存储介质,该计算机可读存储介质包括程序或指令,当所述程序或指令在计算机上运行时,使得计算机执行第二方面或第二方面中任一种可能实现方式中的方法。
上述方面中的芯片系统可以是片上系统(system on chip,SOC),也可以是基带芯片等,其中基带芯片可以包括处理器、信道编码器、数字信号处理器、调制解调器和接口模块等。
第九方面,本申请实施例提供一种芯片或者芯片系统,该芯片或者芯片系统包括至少一个处理器和接口,接口和至少一个处理器通过线路互联,至少一个处理器用于运行计算机程序或指令,以进行第一方面或第一方面的任一种可能的实现方式中任一项所描述的方法。
第十方面,本申请实施例提供一种芯片或者芯片系统,该芯片或者芯片系统包括至少一个处理器和接口,接口和至少一个处理器通过线路互联,至少一个处理器用于运行计算机程序或指令,以进行第二方面或第二方面的任一种可能的实现方式中任一项所描述的方法。
其中,芯片中的接口可以为输入/输出接口、管脚或电路等。
在一种可能的实现中,本申请中上述描述的芯片或者芯片系统还包括至少一个存储器,该至少一个存储器中存储有指令。该存储器可以为芯片内部的存储单元,例如,寄存器、缓存等,也可以是该芯片的存储单元(例如,只读存储器、随机存取存储器等)。
附图说明
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例中所需要使用的附图 作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。
图1为一种可穿戴设备与XR服务器之间的传输路径的示意图;
图2a为一种XR业务模型的示意图;
图2b为一种基站收发XR业务模型的示意图;
图3a为本申请实施例提供的一种网络系统的示意图;
图3b为本申请实施例提供的一种网络系统的详细网元示意图;
图4为本申请实施例提供的一种数据接收方法的流程示意图;
图5为本申请实施例提供的一种终端设备从休眠状态切换为启动状态并开始监听数据的流程示意图;
图6为本申请实施例提供的一种终端设备从休眠状态切换为启动状态并开始监听数据的时序图;
图7为本申请实施例提供的另一种数据接收方法的流程示意图;
图8a为本申请实施例提供的一种视频帧的结束传输时机的示意图;
图8b为本申请实施例提供的一种视频帧的传输持续时长的示意图;
图8c为本申请实施例提供的另一种视频帧的结束传输时机的示意图;
图8d为本申请实施例提供的另一种视频帧的传输持续时长的示意图;
图9为本申请实施例提供的一种数据接收装置的结构示意图;
图10为本申请实施例提供的一种终端设备的结构示意图;
图11为本申请实施例提供的另一种数据接收装置的结构示意图;
图12为本申请实施例提供的另一种终端设备的结构示意图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行描述。
扩展现实(extended reality,XR)是指通过计算机技术和可穿戴设备产生的一个真实与虚拟组合、可人机交互的环境,是增强现实(augmented reality,AR)、虚拟现实(virtual reality,VR)、混合现实(mixed reality,MR)等多种形式的统称。三者视觉交互技术融合,实现虚拟世界与现实世界之间无缝转换的“沉浸感”体验。
其中,可穿戴设备可以是应用层用户设备(如头盔等),可穿戴设备可以外接终端设备(如智能手机等),也可以与终端设备在物理形态上是一体的。其中,实现虚拟世界与现实世界之间无缝转换的“沉浸感”体验的过程包括:头盔内的传感器感知用户的位置、动作变化(如视角、视线、运动速率变化等),生成用户信息。该用户信息可以通过可穿戴设备—运营商网络(如5G网络)—数据网络(包括XR服务器)的路径传输至XR服务器。XR服务器根据用户信息,结合用户在游戏中或实景中的场景,生成新的视频数据和音频数据,在回传至头盔,最终通过头盔向用户展示,如图1所示。
其中,由于XR业务的交互性,业务数据对时延要求比较严格。例如,从生成用户位置、动作信息开始,如果在70毫秒(ms)内终端接收到来自XR服务器的响应数据,用户感受比较流畅;如果时延大于70ms(如100ms),用户可能感受到时延,进而感到头晕,影响体验;如果时延大于150ms,则用户感到视频卡顿,认为XR业务不可用。
下面对XR业务经过终端设备—5G网络—数据网络的路径进行交互的时延分析,主要包括终端侧时延、网络时延和服务器时延。
终端侧时延:包括两部分,第一部分包括头盔内的传感器生成用户位置、运动信息的时延,对用户信息进行编码的时延,以及将编码后的数据交付至接入层等待传输的时延。第二部分包括接收到来自XR服务器的数据后,进行缓存、视频音频解码、显示等的时延。上述两部分时延总计约25-30ms。
网络时延:包括两部分,第一部分用于上行传输,终端接收到来自上层的数据(用户位置、运动信息等)后,通过5G网络,将该数据传输至XR服务器的时延。第二部分用于下行传输,5G网络接收到来自XR服务器的响应数据后,将该响应数据传输至终端的时延。两部分时延总计约14ms。
服务器时延:包括XR服务器收到来自头盔的请求(包括用户位置、视线信息等)后,经过运算处理,生成响应所需的时间。例如,XR服务器的处理包括生成原始视频音频数据、针对原始视频数据进行渲染增强虚拟物件、视频编码、音频编码、封装形成流媒体文件,上述处理流程总计约26-34ms。
其中,XR业务数据可以分为两部分:上行数据和下行数据。上行数据只包含用户的位置、视线等信息,数据量小;下行数据包含XR服务器所生成的视频音频数据,数据量大。可选的,由于音频数据量相对很少,进行视频音频数据封装时,将音频数据与视频数据封装在一起,不会对视频数据的整体业务模型造成太大影响。所以下面以下行视频数据为例,描述业务模型。
请参见图2a,图2a为一种XR的业务模型。其中,XR服务器按每秒60帧的速率,生成视频帧,即每16.67ms生成一个视频帧。该视频帧经过编码后,由于IP网络限制,按1500字节(byte,B)为标准,将编码后的数据流划分为多个数据包。从5G网络角度,体现为每16.67ms从XR服务器接收到一组数据包,该一组数据包包括多个尺寸约为1500B的数据包。
举例来说,当终端设备通过上行数据发出应用层请求时,5G网络中的用户面网元(user plane function,UPF)或接入网网元(如gNB)可以从XR服务器接收XR数据。例如,gNB每16.67ms接收一组数据包(即一个视频帧),该一组数据包包括多个尺寸约为1500B的数据包。gNB收到该一组数据包后,根据无线接口负荷以及该一组数据包与其它待传输数据的优先级,确定传输该一组数据包的特定时机,并在该特定时机向终端设备发送该一组数据包。
其中,在XR服务器连续发出下行视频数据的情况下,UPF/gNB每隔16.67ms接收一个视频帧。gNB经过处理,只在16.67ms中一部分时间内传输该视频帧,在其它时间段则不传输数据,如图2b所示。但是,一旦gNB开始传输视频帧,终端设备将始终保持监听状态,如图2b所示。那么终端设备的持续监听可能导致造成额外耗电。
为了解决上述问题,本申请实施例提供一种数据接收方法,该方法可以使终端设备不需要持续监听数据的收发,有利于降低终端设备的功耗。
其中,本申请实施例提供的数据接收方法可以应用于如图3a所示的网络系统中。请参见图3a,图3a为本申请实施例提供的一种网络系统的示意图。该网络系统包括终端设备、 网元设备和XR服务器。其中,终端设备、网元和XR服务器之间可以通过无线或有线的方式进行数据传输,如图3a所示。
其中,终端设备可以是一种具有无线收发功能的设备,或者终端设备也可以是一种芯片。所述终端设备可以是用户设备(user equipment,UE)、手机(mobile phone)、平板电脑(Pad)、带无线收发功能的电脑、虚拟现实(virtual reality,VR)中的终端设备、增强现实(augmented reality,AR)中的终端设备、车载终端设备、远程医疗(remote medical)中的终端设备、智能电网(smart grid)中的终端设备、可穿戴终端设备(如头盔等)、车联网、D2D通信、机器通信中的传感器等。
其中,网元是提供网络服务的实体,包括核心网网元、接入网网元等。其中,核心网主要包括以下几个关键逻辑网元:接入和移动管理网元(access and mobility management function,AMF)、会话管理网元(session management function,SMF)、UPF、策略控制网元(policy control function,PCF)、统一数据管理网元(unified data management,UDM),如图3b所示。
其中,AMF主要负责移动网络中的移动性管理,如用户位置更新、用户注册网络、用户切换等。SMF主要负责移动网络中的会话管理,如会话建立、修改、释放。具体功能如为用户分配IP地址、选择提供报文转发功能等。PCF负责向AMF、SMF提供策略,如QoS策略、切片选择策略等。UDM用于存储用户数据,如签约信息、鉴权/授权信息。UPF主要负责对用户报文进行处理,如转发、计费等。
接入网网元包括:基站(如gNB,eNB)、集中控制单元(central unit control plane,CUCP)、集中单元(central unit,CU)、分布式单元(distribution unit,DU)、集中用户面单元(central unit user plane,CUUP)、无线接入点(WiFi AP)等。例如,在第五代移动通信(the 5th generation mobile communication technology,5G)系统中,基站被称为gNB/ng-eNB,本实施例中统一用gNB代表,gNB之间通过Xn接口连接。gNB和5G核心网(5GC)通过NG接口连接,如图3b所示。其中,基站可以包括CU和DU。也就是说,5G接入网可以对原接入网中的基站的功能进行拆分,将基站的部分功能部署在一个CU,将剩余功能部署在DU,多个DU共用一个CU,可以节省成本,以及易于网络扩展。
例如,CU和DU的切分可以按照协议栈切分,将无线资源控制(radio resource control,RRC),服务数据适应协议(service data adaptation protocol,SDAP)以及分组数据汇聚协议(packet data convergence protocol,PDCP)层部署在CU,其余的无线链路控制(radio link control,RLC)层、介质访问控制(media access control,MAC)层以及物理层(physical,PHY)部署在DU。CU和DU之间通过F1接口连接。其中,gNB-CU可以代表gNB通过NG接口和核心网连接,gNB-CU也可以代表gNB通过Xn接口和其他gNB连接,如图3b所示。
其中,XR服务器是提供XR业务的实体,XR服务器负责根据用户指令,生成相应的视频帧数据。XR服务器可以向网元发送XR数据,或者接收网元的反馈信息。XR服务器可以包括XR编码器,用于对视频帧进行编码。
下面将结合具体的实施例进行描述。
请参见图4,图4为本申请实施例提供的一种数据接收方法,该方法可以由终端设备所执行,包括以下步骤:
S401,终端设备确定时机配置,该时机配置用于指示通知时机。
S402,终端设备在通知时机接收来自接入网网元的第一通知信息,该第一通知信息用于指示视频帧的起始传输时机和/或传输持续时长。
为了降低终端设备的功耗,终端设备可以确定时机配置,并且在该时机配置指示的通知时机监听是否接收到通知消息,从而确定是否要监听数据。也就是说,终端设备可以在通知时机醒来监听数据。
在一实施方式中,终端设备确定时机配置的方法可以是终端设备接收来自接入网网元的通知时机。请参见图5,图5为本申请实施例提供的一种终端设备从休眠状态切换为启动状态并开始监听数据的流程示意图。该流程可以包括以下步骤:
接入网网元配置通知时机,并向终端设备发送通知时机;
终端设备接收来自接入网网元的通知时机,确定时机配置;
终端设备在通知时机接收来自接入网网元的第一通知信息;
终端设备在第一通知信息指示的视频帧起始传输时机开始监听数据。
其中,终端设备在通知时机可以接收通知消息,若接收到通知消息,则开始监听数据。其中,也就是说,终端设备根据接收到的通知时机确定时机配置,并根据时机配置在通知时机从休眠状态切换为启动状态,开始监听数据。
具体的,接入网网元配置的通知时机可以是根据XR编码器的编码时长确定,也可以是根据协议规定确定的。
举例来说,若XR编码器的编码时长不为固定值,那么接入网网元(如gNB)需要根据XR编码器的编码时长确定通知时机。假设XR编码器对每个视频帧进行编码的编码时长为10毫秒(ms),XR编码器可以将该编码时长发送给gNB,以使gNB根据该编码时长确定通知时机,如图5所示。例如,若XR编码器编码视频帧所需时长为10ms,那么gNB配置的通知时机为每10ms一次通知时机。可以理解的是,该步骤为可选的步骤。
其中,XR编码器可以将该编码时长通知gNB,那么该编码时长可以携带在编码时长通知消息中。其中,该编码时长通知消息的具体格式可以是GPRS隧道协议(GPRS tunnel protocol,GTP)控制包,也可以是编码后的数据包头中的一个字段,或者是其它传输层的控制包或包头,本实施例不作限定。该编码时长通知消息可以经由UPF传至gNB,UPF可以解码该通知,也可以透传该通知(即不解码传输),本实施例不作限定。该编码时长通知消息也可以经由SMF-AMF传至gNB,其中SMF解码该通知消息,AMF可以解码该通知消息,也可以透传该通知消息,本实施例不作限定。
又例如,若XR编码器的编码时长为固定值(如H.264协议中规定了编码时长),那么gNB可以根据协议中规定的编码时长来确定通知时机。也就是说,gNB可以直接确定通知时机,而不需要接收XR编码器的编码时长来确定通知时机。
另一实施方式中,若接入网网元没有为终端设备配置通知时机,终端设备也可以通过学习的方式确定时机配置。例如,终端设备首先保持持续的监听,在每一次接收到gNB发送的第一通知信息时,记录下接收到该信息的时机。通过对多次记录的时机进行学习,终端设备可以确定后续gNB向终端设备发送第一通知信息的通知时机,也就能确定时机配置。
可选的,当XR编码器开始针对一个视频帧进行编码时,XR编码器可以通知接入网网 元,使接入网网元知道XR编码器已开始对视频帧进行编码。其中,XR编码器通知接入网网元时的具体内容可以包括以下两种。例如,XR编码器可以将“XR编码器开始编码视频帧”的信息通知gNB,该信息可以携带在开始编码通知消息中。又例如,XR编码器可以将“10ms后编码的数据输出”的信息通知gNB,该信息也可以携带在开始编码通知消息中。
其中,该开始编码通知消息的具体格式,可以是GTP控制包,也可以是编码后的数据包头中的一个字段,或者是其它传输层的控制包或包头,本实施例不作限定。该开始编码通知消息可以经由UPF传至gNB,UPF可以解码该通知,也可以透传该通知(即不解码传输),如图5所示,本实施例不作限定。该开始编码通知消息也可以经由SMF-AMF传至gNB,其中SMF解码该通知消息,AMF可以解码该通知消息,也可以透传该通知消息,本实施例不作限定。
可选的,根据前文实施例中的描述,XR服务器按每秒60帧的速率,生成视频帧,即每16.67ms生成一个视频帧。该视频帧经过编码后再进行传输,若XR编码器连续对视频帧进行编码,编码时长(10ms)小于生成视频帧的时长(16.67ms),那么XR编码器无需通知接入网网元。也就是说,若XR编码器连续对视频帧进行编码,则XR编码器无需再向接入网网元发送开始编码通知消息。
可选的,终端设备也可以根据预先获得的半静态或静态通知,获知在哪些时机开始监听数据。也就是说,接入网网元不需要每次都向终端设备发送第一通知信息,以使终端设备开始监听视频帧。例如,基站通知一次,终端设备就开始监听第一个视频帧,后面的视频帧到达时不需要通知,终端设备可以自行监听。
可选的,接入网网元接收到开始编码通知消息后,可以确定视频帧到达接入网网元的时刻。那么接入网网元可以在通知时机向终端设备发送第一通知信息,以使终端设备在对应的时刻醒来并接收数据。例如,gNB接收到开始编码通知消息后,获知10ms后XR编码器将会有编码后的视频帧输出。gNB再结合先验信息“XR编码器-UPF-gNB传输时延为2ms”,确定“8ms后即将有视频帧到达”。gNB通过下一个通知时机,通知终端设备醒来并接收视频帧,如图5所示。其中,接入网网元向终端设备发送的第一通知信息可以是RRC消息、SDAP层控制信令、PDCP层控制信令、RLC层控制信令、MAC层控制信令、PHY层控制信令中的一种或多种承载,本实施例不作限定。
终端设备在接收到第一通知信息后,可以根据第一通知信息,确定在哪个时刻开始监听数据。例如,gNB向终端设备发送的第一通知信息为“5ms后有视频帧传输”,终端设备从接收该通知的子帧/时隙(slot)/传输时间间隔(transmission time interval,TTI),向后推5ms,开始监听数据。
其中,第一通知信息用于指示视频帧的起始传输时机和/或传输持续时长。其中,本实施例所述的视频帧的起始传输时机也可以称为视频帧的第一个数据包所在的子帧/时隙/传输时间间隔,视频数据的数据开始时机等,都是表达相同的含义,本实施例不作限定。类似的,本实施例所述的视频帧的传输持续时长也可以称为视频帧的占用的子帧/时隙/传输时间间隔的总数等,本实施例不作限定。
可选的,视频帧的起始传输时机和/或传输持续时长可以同时携带在第一通知信息中,也可以是携带在不同的通知信息中。例如,第一通知信息仅用于指示视频帧的起始传输时 机,而不携带视频帧的传输持续时长。终端设备也可以根据该第一通知信息确定开始监听视频数据的时机。
可选的,终端设备在视频帧的起始传输时机监听视频帧,并且在视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。也就是说,该第一通知信息确定了终端设备开始监听数据的时机和结束监听数据的时机,从而只在传输持续时长的时间段内监听数据,在其他时机不监听数据。可以理解的是,终端设备根据视频帧的起始传输时机和传输持续时长,可以确定视频帧的结束传输时机。
可选的,假设XR服务器连续向接入网网元发送多个视频帧,接入网网元可以传输该多个视频帧。对应的,终端设备可以接收该多个视频帧。终端设备可以对该多个视频帧进行学习,确定视频帧的到达时间间隔。其中,视频帧的到达时间间隔为两个视频帧的起始传输时机之间的时间差。
其中,若视频帧的到达时间间隔为固定值,那么接入网网元在第一次向终端设备发送第一通知信息后,若一直有视频帧传输,则无需再向终端设备发送第一通知信息,终端设备可以根据视频帧的到达时间间隔确定下一次监听数据的时刻。
举例来说,假设终端设备接收了K个视频帧,并且该K个视频帧的到达时间间隔为固定值。那么终端设备可以对该K个视频帧进行学习,确定该K个视频帧中的任一视频帧k-1与其相邻的视频帧k之间的到达时间间隔为10ms。那么终端设备在接收后续的视频帧时,可以确定在第K个视频帧的起始传输时机之后的10ms之后,开始监听数据以接收下一个视频帧。
其中,若视频帧的到达时间间隔不为固定值,那么终端设备可以接收来自接入网网元的第二通知信息,该第二通知信息用于指示与当前视频帧相邻的下一视频帧的起始传输时机。对应的,终端设备在该下一视频帧的起始传输时机监听数据。也就是说,若视频帧不是按照一定的时间规律传输的,那么接入网网元需要通知终端设备什么时候开始监听数据,从而降低终端设备的耗电。
例如,若XR编码器未连续对视频帧进行编码,则XR编码器继续对视频帧进行编码时,需要再向gNB发送开始编码通知消息。需要注意的是,该一段时间的具体时间长度可以是XR编码器自行确定的,也可以是网管平台配置的,本实施例不作限定。
又例如,若接入网网元在接收到视频帧后,对第一个视频帧进行缓存后再传输,对相邻的第二个视频帧不进行缓存而是直接传输,那么这两个视频帧的第一个数据包所在的子帧,其间隔将小于相邻视频帧的到达时间间隔(如16ms)。那么接入网网元需要在传输第二个视频帧之前,向终端设备发送第二通知信息,该第二通知信息用于指示终端设备再次从休眠状态切换为启动状态的第二时域位置,即再次开始监听的时机。
其中,该第二通知信息的格式可以是以下一种或多种:RRC消息、DAP层控制信令、PDCP层控制信令、RLC层控制信令、MAC层控制信令、物理层控制信令等。该第二通知信息可以在原有的通知时机传输,也可以将该第二通知信息与第一个视频帧通过同一子帧传输。其中,若将该第二通知信息与第一个视频帧通过同一子帧传输,则终端设备监听视频帧时无需监听通知时机。
可选的,接入网网元可以在传输一个视频帧的所有数据后,向终端设备发送第三通知 信息,该第三通知信息用于指示视频帧数据传输结束。对应的,终端设备接收到该第三通知信息后,停止监听数据,由启动状态切换为休眠状态。
请参见图6,图6为本申请实施例提供的一种终端设备从休眠状态切换为启动状态并开始监听数据的时序图。其中,该时序图中的一个方格表示时域位置,即一个子帧/时隙/TTI,例如,一个方格代表一个子帧,一个子帧的时间长度为1ms。为了便于描述,下面以一个方格表示一个子帧为例,对终端设备从休眠状态切换为启动状态并开始监听数据的时序进行详细的介绍。
假设接入网网元通过空口开始向终端设备传输视频帧的时域位置为子帧N,如图6中的阴影部分所示。假设XR服务器发送的视频帧是连续的,即XR编码器将持续向接入网网元发送视频帧,并且相邻视频帧的到达时间间隔假设为16ms。假设传输该视频帧需要占用三个子帧,如图6中的阴影部分所示。假设终端设备确定的时机配置指示终端设备的通知时机每10ms一次,也就是说,终端设备每10ms醒来一次,检测是否接收到第一通知消息。那么由子帧N向前推移,假设终端设备在接收该视频帧之前的一个通知时机为子帧N-5,那么终端设备在该子帧N-5处醒来,检测是否接收到第一通知消息,如图6所示。可以理解的是,上述时域位置为终端设备处于启动状态的时间段。
其中,若gNB接收到的开始编码通知消息中包括“10ms后编码的数据输出”的信息,并且已知“XR编码器-UPF-gNB传输时延为2ms”,那么由子帧N向前推移,在子帧N-12对应的时刻XR编码器向gNB发送开始编码通知消息,子帧N-10处gNB收到来自XR编码器的开始编码通知消息,并在下一个通知时机对应的子帧N-5处向终端设备发送第一通知信息,如图6所示。XR编码器在子帧N-2对应的时刻,向UPF发出视频帧,gNB在子帧N处接收到该视频帧,并通过空口向终端设备传输该视频帧。该视频帧在子帧N+2处传输完成,终端设备可以由启动状态切换为休眠状态。此时终端设备可以记录该视频帧的第一个数据包的时域位置即子帧N,并在子帧N+16处重新由休眠状态切换为启动状态,再次接收视频帧,如图6所示。
也就是说,基于图6所示的时序图,不考虑其它业务共存的情况下,第一个视频帧开始前,终端设备只需要在各个通知时机监听是否有第一通知消息。若收到第一通知消息,终端设备只需要在子帧N、N+1、N+2监听第一个视频帧,然后在子帧N+16开始监听第二个视频帧,依次类推。
可选的,接入网网元在接收到视频帧后,也可以对数据进行缓存,在满足业务时延预算的前提下,选择合适的时机通过无线空口传输。例如,gNB已知“10ms后编码的数据输出”,并且已知“XR编码器-UPF-gNB传输时延为2ms”。那么若gNB对数据进行缓存后再发出,gNB向终端设备发送的第一通知信息可以是“6ms后有数据传输”,也就是说,gNB对视频帧缓存的时长为2ms。
可选的,上述方案假定终端设备只有一个视频业务,只要该视频业务没有视频帧传输,终端设备就进入休眠状态。但是,若终端设备同时有其它业务并行,那么终端设备还需要考虑其它业务是否有数据传输,从而确定是否进入休眠状态。
例如,若终端设备同时进行两项业务,包括观看流媒体视频的业务1和进行视频通话的业务2,其中,假设业务1的视频帧的编码时长为10ms,业务2的视频帧的编码时长为 7ms,两者的编码时长不一致,那么两种业务的通知时机也不一致。接入网网元可以将业务1和业务2的通知时机叠加,确定终端设备的通知时机。终端设备在通知时机接收第一通知信息,并在第一通知信息指示的时机开始监听数据。待业务1和业务2均传输结束后(例如基站可以通知终端设备所有业务传输完成),终端设备再结束监听,恢复休眠状态。
需要注意的是,上文实施例中的举例说明均以每秒60个视频帧为例,即每16.6ms传输一个视频帧。本实施例对于任何其它帧率同样适用,如每秒90个视频帧,即每11.1ms一个视频帧,本实施例不作限定。
本申请实施例提供一种数据接收方法,该方法可以由终端设备所执行。其中,终端设备可以确定时机配置,并在时机配置指示的通知时机接收来自接入网网元的第一通知信息,从而确定视频帧的起始传输时机和/或传输持续时长。可见,终端设备只在通知时机接收接入网网元的第一通知信息,并根据该第一通知信息,确定在视频帧的起始传输时机和/或传输持续时长。终端设备在特定的时机(如视频帧的起始传输时机)才监听数据,有利于降低终端设备的耗电。
由于XR/VR/AR业务的视频业务实时性要求很高,接入网网元通常不会对视频数据进行缓存,而是收到后立即通过无线接口传输。那么终端设备可以通过学习,获知每一个视频帧的起始传输时机和视频帧的到达时间间隔,从而确定监听视频数据的时机。请参见图7,图7为本申请实施例提供的另一种数据接收方法,该方法可以由终端设备所执行,包括以下步骤:
S701,终端设备根据预先监听的K个视频帧,确定各个视频帧的起始传输时机、传输持续时长、视频帧的到达时间间隔;
S702,终端设备根据该K个视频帧中的视频帧k的起始传输时机、传输持续时长、视频帧的到达时间间隔,确定视频帧k+1的起始传输时机和/或传输持续时长。
其中,在视频业务初期,终端设备可以持续监听,记录该视频业务中的每一个数据包所在的时域位置,从而确定一簇数据包对应的一个视频帧所在的时域位置。例如,终端设备连续接收视频业务数据(如包括K个视频帧),通过机器学习可以确定哪些数据包构成的一簇数据包对应的是一个视频帧,从而确定该视频帧的起始传输时机,以及每一个视频帧之间的传输时间间隔。
其中,视频帧k为终端设备保持数据监听时接收到的K个视频帧中,最后一个传输的视频帧。视频帧k+1为视频帧k之后到达的第一个视频帧。本实施例中假设终端设备在视频业务初期,可以学习确定K个视频帧中各个视频帧的起始传输时机、传输持续时长、视频帧的到达时间间隔。那么,在第k个视频帧之后,终端设备可以根据视频帧k的起始传输时机,以及视频帧的到达时间间隔,推算视频帧k+1的起始传输时机,并在视频帧k+1的起始传输时机开始监听数据。并以此类推,在后续的数据监听过程中,可以不用再持续监听数据,而是在指定的时机监听数据。
举例来说,终端设备对K个视频帧进行学习,确定视频帧k的起始传输时机为子帧N,视频帧k的传输持续时长为3个子帧,视频帧的到达时间间隔为16个子帧,且视频帧的到达时间间隔为固定值。那么终端设备可以确定子帧k+1的起始传输时机为子帧N+16,终端 设备可以在子帧N+16处开始监听数据。类似的,终端设备还可以确定子帧k+2的起始传输时机为子帧N+32,那么终端设备可以在子帧N+32处开始监听数据。
可选的,为避免接入网网元调度产生数据抖动,终端设备可以在第k+1个视频帧的第一个数据包所在的第一时域位置的基准点处,提前一段时间开始监听数据。例如,若第一时域位置指示的是子帧N+16,那么终端设备可以在子帧N+14处就开始监听数据。其中,具体的提前时间量可以由接入网网元配置,或由网管配置,或由终端设备配置,本实施例不作限定。
可选的,终端设备在视频帧k+1的起始传输时机监听视频帧k+1,并且在视频帧k+1的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。也就是说,终端设备在视频帧的起始传输时机开始监听数据,并在结束传输时机停止监听数据,切换为休眠状态。
例如,视频帧k的起始传输时机为子帧N,视频帧k的传输持续时长为3个子帧,视频帧的到达时间间隔为16个子帧,且视频帧的到达时间间隔为固定值。那么终端设备可以确定子帧k+1的起始传输时机为子帧N+16,终端设备可以在子帧N+16处开始监听数据,在子帧N+19处停止监听数据。需要注意的是,在视频帧k传输完成,直至视频帧k+1到达之前的前序时间段(如子帧N+3至子帧N+15)内,终端设备都无需再监听数据,可以处于休眠状态。
其中,终端设备确定每一个视频帧的传输持续时长,可以通过下述方式一至方式五来实现。也就是说,终端设备采用下述方式一至方式五可以确定视频帧k的传输持续时长或结束传输时机,在该结束传输时机停止监听视频数据,直到下一个视频帧的起始传输时机再重新开始监听视频数据。可选的,终端设备也可以采用以下方式一至方式五确定视频帧k+1的传输持续时长或结束传输时机,本实施例不作限定。
方式一:接入网网元在每个视频帧的最后一个数据包增加视频帧结束标识,或者在每个视频帧的最后一个数据包所在的传输块中增加视频帧结束标识,如图8a所示。例如,基站可以通过在MAC/RLC/PDCP/SDAP子头中增加视频帧结束标识,实现在每个视频帧的最后一个数据包增加视频帧结束标识。又例如,基站可以增加一个MAC控制单元(MAC control element,MAC CE),该MAC CE可以只包含MAC子头。终端设备可以根据该MAC CE确定视频帧的最后一个数据包,从而确定该视频帧的传输结束时刻。
方式二:接入网网元在每个视频帧所占的第一个子帧/时隙/TTI,设置该子帧/时隙/TTI携带“当前视频帧占X个子帧/时隙/TTI”的通知信息,如图8b所示。对应的,终端设备可以根据该信息确定当前视频帧的传输结束时刻。例如,图8b所示的视频帧所占的第一个子帧为子帧N,基站在图8b所示的视频帧所占的第一个子帧中,设置该子帧携带“当前视频帧占3个TTI”的通知信息。那么终端设备可以根据该视频帧所占的第一个子帧的位置,以及占用的TTI,确定该视频帧的传输结束时刻为子帧N+2。
可选的,该通知信息可以携带在视频帧对应的第一个子帧/时隙/TTI,也可以携带在视频帧对应的第N个子帧/时隙/TTI,本实施例不作限定。如果该通知信息携带在视频帧对应的第一个子帧/时隙/TTI,基站可以通过连续或非连续的子帧/时隙/TTI传输该视频帧。可选的,该通知信息还可以是一个MAC CE。
方式三:每个视频帧所对应的多个数据包,在无线接口传输时,占据连续的多个子帧/ 时隙/TTI。对应的,终端设备从每一簇数据包的起始时刻开始接收视频帧,一直保持连续接收,直到某一个子帧/时隙/TTI没有监听到视频数据,那么终端设备认为该视频帧已经传输结束,如图8c所示。需要注意的是,在方式三的应用场景中,接入网网元需要确保传输的是连续的视频帧。
方式四:接入网网元在每个视频帧所对应的第一个子帧/时隙/TTI,设置该子帧/时隙/TTI携带“当前视频帧所包含的数据包数量/数据量”的通知信息,如图8d所示。对应的,终端设备接收到相应数量的数据包/数据量,认为该视频帧已经传输结束。例如,基站在图8d所示的视频帧所占的第一个子帧中,设置该子帧携带“当前视频帧包含50个数据包/75000字节(bytes)”的通知信息。那么终端设备可以根据该视频帧所占的第一个子帧的位置,该视频帧的大小,确定该视频帧的传输结束时刻。
可选的,该通知信息可以携带在视频帧对应的第一个子帧/时隙/TTI,也可以携带在视频帧对应的第N个子帧/时隙/TTI,本实施例不作限定。如果该通知信息携带在视频帧对应的第一个子帧/时隙/TTI,基站可以通过连续或非连续的子帧/时隙/TTI传输该视频帧。可选的,该通知信息还可以是一个MAC CE。
方式五:终端设备读取视频帧的数据包中的实时传输协议(real-time transport control protocol,RTP)包头,确认该数据包是否为当前视频帧的最后一个数据包。例如,基站可以在视频帧的数据包中的RTP包头设置指示信息,该指示信息可以采用一个或多个比特来标识。
可见,通过上述方式一至方式五,终端设备可以确认每一个视频帧的传输结束时刻,在传输结束时刻不再监听数据。进一步,终端设备还可以结合每一个视频帧的传输起始时刻,每一个视频帧的传输结束时刻,以及相邻视频帧的到达时间间隔,学习记录已接收到的视频帧的传输过程,从而推算下一个视频帧的起始时刻,在下一个视频帧的起始时刻开始监听下一个视频帧。
可选的,终端设备确定视频帧的起始传输时机和持续传输时长后,可以通过RRC信令或MAC CE通知接入网网元该视频帧的起始传输时机和持续传输时长后。对应的,接入网网元根据该视频帧的起始传输时机和持续传输时长后,再次确认该起始传输时机和持续传输时长是否为接入网网元实际传输视频帧时的起始传输时机和持续传输时长,从而保证可靠传输。
在一实施方式中,若出现特殊情况必须改变传输时机,例如基站的收发模块出现问题不能继续传输数据,那么基站也可以根据终端设备发送的起始传输时机和持续传输时长,及时通知终端设备在该起始传输时机和/或持续传输时长的时间段内可能不能收到数据,暂停监听数据。
另一实施方式中,终端设备将自己学习的结果通知基站,基站可以根据该结果,配合终端设备的行为。例如,基站针对该终端设备的视频数据,不会将终端设备的视频数据中的一部分视频帧进行缓存,而一部分视频帧不进行缓存,导致终端设备的视频帧接收异常。
需要注意的是,上述图4至图7所示的实施例中以子帧或slot为单位,来描述视频帧的起始传输时机、视频帧的传输持续时长、视频帧的到达时间间隔等。实际上也可以使用其它任何时间单位进行描述,不影响本申请的技术实质。例如,还可以采用正交频分复用 (orthogonal frequency division multiplexing,OFDM)符号时长、毫秒、微秒、纳秒等时间单位来描述视频帧的起始传输时机、视频帧的传输持续时长、视频帧的到达时间间隔等,本实施例不作限定。
本申请实施例提供一种数据接收方法,该方法可以由终端设备所执行。其中,终端设备可以在视频业务初期持续监听数据,自行学习视频帧的起始传输时机和结束传输时机,从而推算后续视频帧的起始传输时机和/或结束传输时机,只在后续视频帧的起始传输时机开始监听数据,并且在结束传输时机切换为休眠状态。可见,终端设备通过自行学习视频帧的起始传输时机和结束传输时机,从而确定在特定的时机才监听数据而不需要持续监听数据的收发,有利于降低终端设备的耗电。
以下结合图9至图12详细说明本申请实施例的终端设备。
本申请实施例提供一种数据接收装置,如图9所示,该数据接收装置900可用于实现本申请实施例中的数据接收方法。该数据接收装置900可以包括:
处理单元901,用于确定时机配置,该时机配置用于指示通知时机;
收发单元902,用于在通知时机接收来自接入网网元的第一通知信息;该第一通知信息用于指示视频帧k的起始传输时机和/或传输持续时长。
具体实现方式,请参考图4实施例中的S401和S402中的详细描述,在此不再赘述。
在一种实现方式中,处理单元901还用于在视频帧的起始传输时机监听该视频帧,并且在该视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,收发单元902还用于接收来自接入网网元的通知时机,该通知时机是接入网网元根据视频帧的编码时长确定的。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,收发单元902还用于接收多个视频帧;处理单元901还用于根据多个视频帧的业务模型,确定视频帧的到达时间间隔。其中,视频帧的到达时间间隔为两个视频帧的起始传输时机之间的时间差。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,若视频帧的到达时间间隔为固定值,处理单元901还用于在视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态之后,经过视频帧的到达时间间隔,从休眠状态切换为工作状态,监听与该视频帧相邻的下一视频帧。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,若视频帧的到达时间间隔不为固定值,收发单元902还用于接收来自接入网网元的第二通知信息,该第二通知信息用于指示与视频帧相邻的下一视频帧的起始传输时机。处理单元901还用于在下一视频帧的起始传输时机监听下一视频帧。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,视频帧的起始传输时机和/或结束传输时机为子帧、时隙或传输时间间隔中的一种或多种。
在一种实现方式中,图9中的各个单元所实现的相关功能可以通过收发器和处理器来实现。请参见图10,图10是本申请实施例提供的一种终端设备的结构示意图,该终端设 备可以为具有执行本申请实施例所述的数据接收功能的设备(例如芯片)。该终端设备1000可以包括收发器1001、至少一个处理器1002和存储器1003。其中,收发器1001、处理器1002和存储器1003可以通过一条或多条通信总线相互连接,也可以通过其它方式相连接。本实施例中采用总线连接的方式作为一种示例,如图10所示。
其中,收发器1001可以用于发送或者接收数据。例如,收发器1001可以接收第一通知信息。可以理解的是,收发器1001是统称,可以包括接收器和发送器。
其中,处理器1002可以用于对数据进行处理。处理器1002可以包括一个或多个处理器,例如该处理器1002可以是一个或多个中央处理器(central processing unit,CPU),网络处理器(network processor,NP),硬件芯片或者其任意组合。在处理器1002是一个CPU的情况下,该CPU可以是单核CPU,也可以是多核CPU。
其中,存储器1003用于存储程序代码等。存储器1003可以包括易失性存储器(volatile memory),例如随机存取存储器(random access memory,RAM)。存储器1003也可以包括非易失性存储器(non-volatile memory),例如只读存储器(read-only memory,ROM),快闪存储器(flash memory),硬盘(hard disk drive,HDD)或固态硬盘(solid-state drive,SSD)。存储器1003还可以包括上述种类的存储器的组合。
上述处理器1002可以用于实现本申请实施例中的数据接收方法,其中,具体实现方式如下:
确定时机配置,该时机配置用于指示通知时机;
在通知时机接收来自接入网网元的第一通知信息;该第一通知信息用于指示视频帧的起始传输时机和/或传输持续时长。
具体实现方式,请参考图4实施例中的S401和S402中的详细描述,在此不再赘述。
在一种实现方式中,处理器1002还用于在视频帧的起始传输时机监听视频帧,并且在视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,收发器1001接收来自接入网网元的通知时机,该通知时机是接入网网元根据视频帧的编码时长确定的。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,收发器1001还用于接收多个视频帧;处理器1002还用于根据多个视频帧的业务模型,确定视频帧的到达时间间隔。其中,视频帧的到达时间间隔为两个视频帧的起始传输时机之间的时间差。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,若视频帧的到达时间间隔为固定值,处理器1002还用于在视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态之后,经过视频帧的到达时间间隔,从休眠状态切换为工作状态,监听与该视频帧相邻的下一视频帧。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,若视频帧的到达时间间隔不为固定值,收发器1001还用于接收来自接入网网元的第二通知信息,该第二通知信息用于指示与视频帧相邻的下一视频帧的起始传输时机。处理器1002还用于在下一视频帧的起始传输时机监听下一视频帧。
具体实现方式,请参考图4实施例中的详细描述,在此不再赘述。
在一种实现方式中,视频帧的起始传输时机和/或结束传输时机为子帧、时隙或传输时间间隔中的一种或多种。
本申请实施例提供另一种数据接收装置,如图11所示,该数据接收装置1100可用于实现本申请实施例中的数据接收方法。该数据接收装置1100可以包括:
处理单元1101,用于根据预先监听的K个视频帧,确定各个视频帧的起始传输时机、传输持续时长、视频帧的到达时间间隔;
处理单元1101还用于根据K个视频帧中的视频帧k的起始传输时机、传输持续时长、视频帧的到达时间间隔,确定视频帧k+1的起始传输时机和/或传输持续时长。
具体实现方式,请参考图7实施例中的S701和S702中的详细描述,在此不再赘述。
在一种实现方式中,处理单元1101还用于在视频帧k+1的起始传输时机监听视频帧k+1,并且在视频帧k+1的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
具体实现方式,请参考图7实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理单元1101还用于根据视频帧k的起始传输时机和视频帧k的结束标识,确定所述视频帧k的传输持续时长。
具体实现方式,请参考图8a实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理单元1101还用于根据视频帧k的起始传输时机和时域位置信息,确定视频帧k的传输持续时长。
具体实现方式,请参考图8b实施例中的详细描述,在此不再赘述。
在一种实现方式中,若终端设备在接收到视频帧k的第一个数据包之后,在连续的子帧/时隙/传输时间间隔上未监听到数据,处理单元1101还用于根据视频帧k的起始传输时机和未监听到数据的传输时机,确定视频帧k的传输持续时长。
具体实现方式,请参考图8c实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理单元1101还用于根据视频帧k的起始传输时机和视频帧k的数据量,确定视频帧k的传输持续时长。
具体实现方式,请参考图8d实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理单元1101还用于根据视频帧k的起始传输时机和视频帧k的数据包中的实时传输协议RTP包头,确定视频帧k的传输持续时长。
具体实现方式,请参考图7实施例中的详细描述,在此不再赘述。
在一种实现方式中,数据接收装置1100还包括收发单元1102,收发单元1102用于向接入网网元发送该视频帧的起始传输时机和传输持续时长。
具体实现方式,请参考图7实施例中的详细描述,在此不再赘述。
在一种实现方式中,视频帧k为K个视频帧中最后一个传输的视频帧,视频帧k+1为视频帧k之后到达的第一个视频帧。
在一种实现方式中,图11中的各个单元所实现的相关功能可以通过处理器来实现。请参见图12,图12是本申请实施例提供的另一种终端设备的结构示意图,该终端设备可以为具有执行本申请实施例所述的数据接收功能的设备(例如芯片)。该终端设备1200可以包括收发器1201、至少一个处理器1202和存储器1203。其中,收发器1201、处理器1202 和存储器1203可以通过一条或多条通信总线相互连接,也可以通过其它方式相连接。本实施例中采用总线连接的方式作为一种示例,如图12所示。
其中,收发器1201可以用于发送或者接收数据。例如,收发器1201可以向接入网网元发送视频帧的起始传输时机和传输持续时长。可以理解的是,收发器1201是统称,可以包括接收器和发送器。
其中,处理器1202可以用于对数据进行处理。处理器1202可以包括一个或多个处理器,例如该处理器1202可以是一个或多个中央处理器(central processing unit,CPU),网络处理器(network processor,NP),硬件芯片或者其任意组合。在处理器1202是一个CPU的情况下,该CPU可以是单核CPU,也可以是多核CPU。
其中,存储器1203用于存储程序代码等。存储器1203可以包括易失性存储器(volatile memory),例如随机存取存储器(random access memory,RAM)。存储器1203也可以包括非易失性存储器(non-volatile memory),例如只读存储器(read-only memory,ROM),快闪存储器(flash memory),硬盘(hard disk drive,HDD)或固态硬盘(solid-state drive,SSD)。存储器1203还可以包括上述种类的存储器的组合。
上述处理器1202可以用于实现本申请实施例中的数据接收方法,其中,具体实现方式如下:
根据预先监听的K个视频帧,确定各个视频帧的起始传输时机、传输持续时长、视频帧的到达时间间隔;
根据K个视频帧中的视频帧k的起始传输时机、传输持续时长、视频帧的到达时间间隔,确定视频帧k+1的起始传输时机和/或传输持续时长。
具体实现方式,请参考图7实施例中的S701和S702中的详细描述,在此不再赘述。
在一种实现方式中,处理器1202还用于在视频帧k+1的起始传输时机监听视频帧k+1,并且在视频帧k+1的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
具体实现方式,请参考图7实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理器1202还用于根据视频帧k的起始传输时机和视频帧k的结束标识,确定所述视频帧k的传输持续时长。
具体实现方式,请参考图8a实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理器1202还用于根据视频帧k的起始传输时机和时域位置信息,确定视频帧k的传输持续时长。
具体实现方式,请参考图8b实施例中的详细描述,在此不再赘述。
在一种实现方式中,若终端设备在接收到视频帧k的第一个数据包之后,在连续的子帧/时隙/传输时间间隔上未监听到数据,处理器1202还用于根据视频帧k的起始传输时机和未监听到数据的传输时机,确定视频帧k的传输持续时长。
具体实现方式,请参考图8c实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理器1202还用于根据视频帧k的起始传输时机和视频帧k的数据量,确定视频帧k的传输持续时长。
具体实现方式,请参考图8d实施例中的详细描述,在此不再赘述。
在一种实现方式中,处理器1202还用于根据视频帧k的起始传输时机和视频帧k的数 据包中的实时传输协议RTP包头,确定视频帧k的传输持续时长。
具体实现方式,请参考图7实施例中的详细描述,在此不再赘述。
在一种实现方式中,收发器1201用于向接入网网元发送该视频帧的起始传输时机和传输持续时长。
具体实现方式,请参考图7实施例中的详细描述,在此不再赘述。
在一种实现方式中,视频帧k为K个视频帧中最后一个传输的视频帧,视频帧k+1为视频帧k之后到达的第一个视频帧。
本申请实施例提供一种计算机可读存储介质,该计算机可读存储介质存储有程序或指令,当所述程序或指令在计算机上运行时,使得计算机执行本申请实施例中的数据接收方法。
本申请实施例提供一种芯片或者芯片系统,该芯片或者芯片系统包括至少一个处理器和接口,接口和至少一个处理器通过线路互联,至少一个处理器用于运行计算机程序或指令,以进行本申请实施例中的数据接收方法。
其中,芯片中的接口可以为输入/输出接口、管脚或电路等。
上述方面中的芯片系统可以是片上系统(system on chip,SOC),也可以是基带芯片等,其中基带芯片可以包括处理器、信道编码器、数字信号处理器、调制解调器和接口模块等。
在一种实现方式中,本申请中上述描述的芯片或者芯片系统还包括至少一个存储器,该至少一个存储器中存储有指令。该存储器可以为芯片内部的存储单元,例如,寄存器、缓存等,也可以是该芯片的存储单元(例如,只读存储器、随机存取存储器等)。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行所述计算机指令时,全部或部分地产生按照本申请实施例所述的流程或功能。计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线(Digital Subscriber Line,DSL))或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质(例如,软盘、硬盘、磁带)、光介质(例如,高密度数字视频光盘(Digital Video Disc,DVD))、或者半导体介质(例如,固态硬盘(Solid State Disk,SSD))等。
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、计算机软件或者二者的结合来实现,为了清楚地说明硬件和软件的可互换性,在上述说明中已经按照功能一般性地描述了各示例的组成及步骤。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。

Claims (16)

  1. 一种数据接收方法,其特征在于,包括:
    终端设备确定时机配置,该时机配置用于指示通知时机;
    所述终端设备在所述通知时机接收来自接入网网元的第一通知信息;所述第一通知信息用于指示视频帧的起始传输时机和/或传输持续时长。
  2. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    所述终端设备在所述视频帧的起始传输时机开始监听所述视频帧,并且在所述视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
  3. 根据权利要求1或2所述的方法,其特征在于,所述终端设备确定时机配置,包括:
    所述终端设备接收来自接入网网元的通知时机,所述通知时机是所述接入网网元根据所述视频帧的编码时长确定的。
  4. 根据权利要求1至3任一项所述的方法,其特征在于,所述终端设备接收多个视频帧;所述方法还包括:
    所述终端设备根据所述多个视频帧的业务模型,确定视频帧的到达时间间隔;所述视频帧的到达时间间隔为两个视频帧的起始传输时机之间的时间差。
  5. 根据权利要求4所述的方法,其特征在于,所述视频帧的到达时间间隔为固定值;
    所述终端设备在所述视频帧的传输持续时长对应的结束传输时机将工作状态切换为休眠状态之后,所述方法还包括:
    经过所述视频帧的到达时间间隔后,所述终端设备从休眠状态切换为工作状态,监听与所述视频帧相邻的下一视频帧。
  6. 根据权利要求4所述的方法,其特征在于,所述方法还包括:
    若所述视频帧的到达时间间隔不为固定值,所述终端设备接收来自所述接入网网元的第二通知信息,所述第二通知信息用于指示与所述视频帧相邻的下一视频帧的起始传输时机;
    所述终端设备在所述下一视频帧的起始传输时机监听所述下一视频帧。
  7. 一种数据接收方法,其特征在于,包括:
    终端设备根据预先监听的K个视频帧,确定各个视频帧的起始传输时机、传输持续时长、视频帧的到达时间间隔;
    所述终端设备根据所述K个视频帧中的视频帧k的起始传输时机、传输持续时长、所述视频帧的到达时间间隔,确定视频帧k+1的起始传输时机和/或传输持续时长。
  8. 根据权利要求7所述的方法,其特征在于,所述方法还包括:
    所述终端设备在所述视频帧k+1的起始传输时机开始监听所述视频帧k+1,并且在所述视频帧k+1的传输持续时长对应的结束传输时机将工作状态切换为休眠状态。
  9. 根据权利要求7或8所述的方法,其特征在于,所述终端设备确定所述视频帧k的传输持续时长,包括:
    所述终端设备根据所述视频帧k的起始传输时机和所述视频帧k的结束标识,确定所述视频帧k的传输持续时长;或者,
    所述终端设备根据所述视频帧k的起始传输时机和时域位置信息,确定所述视频帧k的传输持续时长;或者,
    若所述终端设备在接收到所述视频帧k的第一个数据包之后,在连续的子帧/时隙/传输时间间隔上未监听到数据,根据所述视频帧k的起始传输时机和所述未监听到数据的传输时机,确定所述视频帧k的传输持续时长;或者,
    所述终端设备根据所述视频帧k的起始传输时机和所述视频帧k的数据量,确定所述视频帧k的传输持续时长;或者,
    所述终端设备根据所述视频帧k的起始传输时机和所述视频帧k的数据包中的实时传输协议RTP包头,确定所述视频帧k的传输持续时长。
  10. 根据权利要求7至9任一项所述的方法,其特征在于,所述视频帧k为所述K个视频帧中最后一个传输的视频帧,所述视频帧k+1为所述视频帧k之后到达的第一个视频帧。
  11. 一种数据接收装置,其特征在于,包括:
    处理单元,用于确定时机配置,该时机配置用于指示通知时机;
    收发单元,用于在所述通知时机接收来自接入网网元的第一通知信息;所述第一通知信息用于指示视频帧k的起始传输时机和/或传输持续时长。
  12. 一种数据接收装置,其特征在于,包括:
    处理单元,用于根据预先监听的K个视频帧,确定各个视频帧的起始传输时机、传输持续时长、视频帧的到达时间间隔;
    所述处理单元还用于根据所述K个视频帧中的视频帧k的起始传输时机、传输持续时长、所述视频帧的到达时间间隔,确定视频帧k+1的起始传输时机和/或传输持续时长。
  13. 一种终端设备,其特征在于,包括处理器,所述处理器与通信接口耦合,使得所述处理器用于读取指令以执行权利要求1至6中任一项所述的方法。
  14. 一种终端设备,其特征在于,包括处理器,所述处理器与通信接口耦合,使得所述处理器用于读取指令以执行权利要求7至10中任一项所述的方法。
  15. 一种芯片,其特征在于,包括处理器和接口;
    所述处理器用于读取指令以执行权利要求1至10中任一项所述的方法。
  16. 一种计算机可读存储介质,其特征在于,包括程序或指令,当所述程序或指令在计算机上运行时,如权利要求1至10中任一项所述的方法被执行。
PCT/CN2020/109323 2020-08-14 2020-08-14 一种数据接收方法、装置及终端设备 WO2022032669A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/109323 WO2022032669A1 (zh) 2020-08-14 2020-08-14 一种数据接收方法、装置及终端设备

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/109323 WO2022032669A1 (zh) 2020-08-14 2020-08-14 一种数据接收方法、装置及终端设备

Publications (1)

Publication Number Publication Date
WO2022032669A1 true WO2022032669A1 (zh) 2022-02-17

Family

ID=80247552

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/109323 WO2022032669A1 (zh) 2020-08-14 2020-08-14 一种数据接收方法、装置及终端设备

Country Status (1)

Country Link
WO (1) WO2022032669A1 (zh)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030148800A1 (en) * 2002-01-17 2003-08-07 Samsung Electronics Co., Ltd. Method and apparatus for reducing power consumption of slotted mode mobile communication terminal
CN101998597A (zh) * 2009-08-14 2011-03-30 华为技术有限公司 广播消息发送指示方法、接收方法及装置和系统
CN108347718A (zh) * 2018-02-02 2018-07-31 北京小米移动软件有限公司 监听通信包的方法、装置以及存储介质
CN109842937A (zh) * 2017-09-20 2019-06-04 维沃移动通信有限公司 信息传输方法、网络设备、终端及计算机可读存储介质

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030148800A1 (en) * 2002-01-17 2003-08-07 Samsung Electronics Co., Ltd. Method and apparatus for reducing power consumption of slotted mode mobile communication terminal
CN101998597A (zh) * 2009-08-14 2011-03-30 华为技术有限公司 广播消息发送指示方法、接收方法及装置和系统
CN109842937A (zh) * 2017-09-20 2019-06-04 维沃移动通信有限公司 信息传输方法、网络设备、终端及计算机可读存储介质
CN108347718A (zh) * 2018-02-02 2018-07-31 北京小米移动软件有限公司 监听通信包的方法、装置以及存储介质

Similar Documents

Publication Publication Date Title
CN107483710A (zh) 在无线通信系统中为终端提供服务的方法和设备
JP5331509B2 (ja) バッファ状態報告を実行する方法及び通信装置
WO2021254335A1 (zh) 一种切换控制方法及通信装置
US20230231787A1 (en) Communication method and an apparatus
EP2658208B1 (en) Method, device, and system for processing streaming media service data
US20240056889A1 (en) Communication processing method for data transmission and related device
WO2022032669A1 (zh) 一种数据接收方法、装置及终端设备
WO2022237527A1 (zh) 一种非连续接收的通信方法及装置
WO2024099159A1 (zh) 确定丢包的方法、装置、通信设备及可读存储介质
WO2023174198A1 (zh) 一种时延控制方法、应用服务器及通信系统
WO2024066419A1 (zh) 业务传输方法及装置
CN109640160B (zh) 基于缓冲控制结构的可持续同步控制系统和方法
WO2023210706A1 (ja) 通信制御方法
WO2023066108A1 (zh) 数据传输方法、装置及终端
US20230371060A1 (en) Resource scheduling method and apparatus
WO2022237473A1 (zh) 应用层测量方法及相关的装置
WO2023066107A1 (zh) 数据传输方法、装置及终端
WO2023066113A1 (zh) 数据传输方法、装置及终端
WO2024012231A1 (zh) 通信方法及装置
WO2023179357A1 (zh) 一种通信方法及装置
WO2024067064A1 (zh) 数据传输方法及装置
WO2023011625A1 (zh) 一种信号收发方法、设备、装置及存储介质
WO2024120328A1 (zh) 数据包处理方法及装置
WO2023217250A1 (en) Method for enabling wireless communication device to access services provided by radio access network and related devices
WO2023210705A1 (ja) 通信制御方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20949167

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20949167

Country of ref document: EP

Kind code of ref document: A1