WO2012076904A1 - Video data transmission system and method - Google Patents
Video data transmission system and method Download PDFInfo
- Publication number
- WO2012076904A1 WO2012076904A1 PCT/GB2011/052451 GB2011052451W WO2012076904A1 WO 2012076904 A1 WO2012076904 A1 WO 2012076904A1 GB 2011052451 W GB2011052451 W GB 2011052451W WO 2012076904 A1 WO2012076904 A1 WO 2012076904A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- datastream
- transmission
- mode
- channel
- Prior art date
Links
- 230000005540 biological transmission Effects 0.000 title claims abstract description 93
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000004891 communication Methods 0.000 claims abstract description 30
- 239000000872 buffer Substances 0.000 claims description 43
- 230000006978 adaptation Effects 0.000 claims description 15
- 230000003247 decreasing effect Effects 0.000 claims description 9
- 230000008569 process Effects 0.000 claims description 5
- 230000003993 interaction Effects 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 230000004044 response Effects 0.000 claims description 3
- 230000001419 dependent effect Effects 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 9
- 238000012795 verification Methods 0.000 description 7
- 230000002452 interceptive effect Effects 0.000 description 6
- 230000006835 compression Effects 0.000 description 4
- 238000007906 compression Methods 0.000 description 4
- 230000033001 locomotion Effects 0.000 description 4
- 230000003139 buffering effect Effects 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000003111 delayed effect Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000012544 monitoring process Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004043 responsiveness Effects 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000001627 detrimental effect Effects 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000008439 repair process Effects 0.000 description 1
- 238000010025 steaming Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/238—Interfacing the downstream path of the transmission network, e.g. adapting the transmission rate of a video stream to network bandwidth; Processing of multiplex streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/238—Interfacing the downstream path of the transmission network, e.g. adapting the transmission rate of a video stream to network bandwidth; Processing of multiplex streams
- H04N21/2381—Adapting the multiplex stream to a specific network, e.g. an Internet Protocol [IP] network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/238—Interfacing the downstream path of the transmission network, e.g. adapting the transmission rate of a video stream to network bandwidth; Processing of multiplex streams
- H04N21/2385—Channel allocation; Bandwidth allocation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/24—Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth, upstream requests
- H04N21/2402—Monitoring of the downstream path of the transmission network, e.g. bandwidth available
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/24—Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth, upstream requests
- H04N21/2405—Monitoring of the internal components or processes of the server, e.g. server load
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/637—Control signals issued by the client directed to the server or network components
- H04N21/6373—Control signals issued by the client directed to the server or network components for rate control, e.g. request to the server to modify its transmission rate
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/637—Control signals issued by the client directed to the server or network components
- H04N21/6377—Control signals issued by the client directed to the server or network components directed to server
- H04N21/6379—Control signals issued by the client directed to the server or network components directed to server directed to encoder, e.g. for requesting a lower encoding rate
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/643—Communication protocols
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/647—Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load, bridging between two different networks, e.g. between IP and wireless
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/65—Transmission of management data between client and server
- H04N21/658—Transmission by the client directed to the server
- H04N21/6587—Control parameters, e.g. trick play commands, viewpoint selection
Definitions
- wireless High Definition (HD) video and content delivery providing low latency so that interactive applications can be supported as well as maintaining high quality in the presence of compression and loss so that the quality expectations of the user are satisfied are complex requirements.
- the joint adaptation of the independent wireless and video data rates is complex and, in some cases, it may not be possible to satisfy desired delay and quality constraints.
- desired delay and quality constraints For example, in gaming, delayed content delivery can compromise the user experience whereas in the case of video, poor quality compression and loss would be undesirable.
- a system which is based on a wireless standard such as 802.11 ⁇ and a video codec such as H.264 can be used to compress and stream video to a client.
- Such a system can allow compression to take place in real time, for example using an encoder or transcoder in a set top box that receives a signal off air and then which transcodes it to a format and rate suitable for redistribution.
- Such systems must cope with time-varying data bandwidth of the communications channel and often incorporate large amounts of buffering to deal with this.
- delay is often a key parameter as it affects session start up times and, in interactive applications, may be visible to the user throughout the session. Poor responsiveness as a result of delay can be detrimental to the user experience.
- Designers therefore often seek techniques for maintaining a balance between high video quality, low probability of playout disruption and low end to end delay.
- An aim of the present invention is to provide a system and method whereby adaptation of the video is possible to allow the same to be suited to specific modes of use of the apparatus and which modes can be selected by the user and/ or in response to another control means.
- a method for controlling a video transmission system comprising: receiving an input indicative of a mode of operation;
- the input is generated by user interaction with the apparatus to select a particular mode.
- the system determines the mode to be used based on detected conditions and then generates the input to reflect the selected mode such that the selection and input is automatically generated.
- the transmission parameters include video encoder parameters.
- the said input is generated by user interaction with the system.
- the video transmission system is operated via a wireless communication medium.
- the method of controlling the wireless video transmission system ensures that the video transmission is suitable for the purpose of use which is reflected in the desired mode of operation and enables the transmission parameters to be set at the most appropriate values for the mode of operation.
- the method further comprises estimating channel conditions for the wireless communications channel, and using such predicted channel conditions in the configuration of the transmission parameters.
- the use of the intended mode of operation in conjunction with the estimated channel conditions allows the configuration of the transmission parameters to be optimised for the desired mode of operation.
- the estimating of channel conditions can be repeated during the datastream transmission.
- a transmission can therefore be optimised on an ongoing basis to refine transmission parameters taking into account the predicted ongoing channel conditions.
- a wireless video transmission system providing switchable modes of operation and typically the modes of operation can be controlled by the user according to a use-case or content type.
- the modes of operation can include, for example, any or any combination of: an interactive low-delay mode, a 'normal' mode, such as for general use, and a movie mode in which quality can be prioritised.
- the interactive mode provides a low latency link for channel hopping and interactive services such as set top box (STB) menus and games and so on
- the movie mode provides increased buffer delay to facilitate reduction of peak data rate over the wireless channel and/or increased retransmission of data and/ or repair of packets.
- a means for a user to control an operating mode can be, for example, a single remote control button to flip between modes.
- the selection of an operating mode causes buffers to be suitably configured and video coding parameters to be selected to support the mode in question.
- the parameters can be selected to support the mode in question such that, for example, a windowed rate adaptation scheme for transmitting delay-constrained video over a wireless channel can be used which uses an H.264 encoder with Hypothetical Reference Decoder (HRD)- compliant rate control and generic 802.11-like radio.
- HRD Hypothetical Reference Decoder
- Such a scheme assumes only loose coupling between systems components which improves implementation feasibility.
- the method may further comprise configuring a buffer of the wireless video transmission system in dependence upon the desired mode of operation indicated by the received user input.
- the method further comprises selecting video coding parameters in dependence upon the desired mode of operation indicated by the received user input.
- the desired mode of operation can be chosen from a high quality mode in which buffer size is increased, and a low latency mode in which buffer size is decreased. Datastream transmission can therefore be optimised so that a probability of playback disruption is decreased.
- Configuration of the transmission parameters includes video source rate adaptation. Datastream transmission can therefore be tailored accordingly so that the probability of playback disruption is minimised.
- a further aspect of the present invention provides a video transmission system comprising:
- a video source to generate and output a video datastream
- a datastream encoder and transmitter unit to receive input to configure transmission parameters in dependence upon a desired mode of operation indicated by a received input, and to encode and transmit a video datastream over a communications channel using such transmission parameters.
- system includes a user input device operable to receive a user input indicative of a desired mode of operation and the datastream encoder and transmitter unit is operable to receive user input information from the user input device.
- the video transmission is achieved over a wireless communication channel.
- the system may further comprise a channel estimator operable to estimate channel conditions for a wireless communications channel, the datastream encoder and transmitter operable to use the estimated channel conditions in the configuration of transmission parameters.
- transmission parameters can be optimised for the desired mode of operation by taking account of the predicted channel conditions for the time of transmission.
- Such a channel estimator can be used to determine channel conditions repeatedly during datastream transmission. This repeated determination allows a transmission configuration to be optimised on an ongoing basis so that the most appropriate transmission parameters can be selected.
- the encoder and transmitter unit include a video data buffer, and it is operable to configure the video data buffer in dependence upon the desired mode of operation indicated by the received user input. Accordingly, the datastream transmission can be optimised so that the probability of playback disruption is decreased.
- the encoder and transmitter unit is operable to select video coding parameters in dependence upon a desired mode of operation indicated by a received user input so that characteristics of the datastream can be controlled to ensure they are appropriated for the desired mode of operation and channel conditions.
- the desired mode of operation may be chosen from a high quality mode in which buffer size is increased, and a low latency mode in which buffer size is decreased.
- the option to choose a high quality mode or low latency mode means that datastream transmission can be optimised for the use, and that the probability of playback disruption is decreased.
- the encoder and transmitter unit is operable to configure the transmission parameters using source rate adaption, therefore further ensuring that datastream transmission can be optimised and the probability of playback disruption is decreased.
- a datastream transmission device to receive input data representing a mode of operation for the device, receive a video datastream to be transmitted, select a set of parameters for transmission of the video datastream on the basis of the mode of operation, modify the video datastream on the basis of the selected parameters, and transmit the modified video datastream over a communications channel.
- Modification of the video datastream can include to encode, transcode, adapt or otherwise process the video datastream.
- To adapt the video datastream can include selecting multiple portions of the video datastream to provide a substream to represent the video datastream with a reduced bit rate.
- Such a device can further include a channel estimator to estimate channel conditions of the communications channel, such that the encoder can use the estimated channel conditions in the selection of the parameters for transmission.
- a communications channel according to the features described above and the examples described below need not be limited to being a wireless channel.
- a wireless channel is predominantly used, but this is not intended to be limiting, and non-wireless or hybrid systems and environments including wireless and non-wireless (such as wired) components are suitable for use.
- Reference to 'wireless' should be construed accordingly.
- the invention, and the system and method described herein may be implemented via wired communication systems based on fibre optic or coaxial cable connections.
- Figure 1 is a schematic block diagram illustrating a wireless video transmission system according to one embodiment of the invention
- Figures 2 and 3 are graphs showing performance parameters of a wireless video transmission system according to an embodiment of the invention.
- Figure 3a is a schematic block diagram of a video transmission system according to an embodiment of the invention
- Figure 4 is a schematic block diagram of a method of controlling a video transmission system according to an embodiment of the invention
- Figure 5 is a schematic block diagram of a video transmission system according to an embodiment of the invention.
- Figure 6 is a schematic block diagram of a video transmission system according to an embodiment of the invention.
- Figure 1 illustrates a wireless video transmission system 1 according to an example which includes a video source 10 which in this case is a video encoder or transcoder.
- the source 10 generates a video datastream.
- unencoded video data input to the video source 10 can be encoded into a coded video bitstream to form a generated video datastream.
- previously encoded video data input to the video source 10 can be transcoded into a coded video bitstream to form a generated video datastream.
- scalably encoded video input to the video source 10 can be adapted, modified or processed in order to produce an alternate suitable coded video bitstream forming a generated video datastream.
- parts of a scalably encoded video datastream can be removed or discarded to provide one or more substreams representing the source content but with a bit rate which is less than that of the complete original video input.
- Such substreams can be suited to current channel and buffer conditions as will be described below.
- the datastream is compliant with suitable video buffer verification model rate control parameters, such as HRD parameters for example (as described in standards such as H.264/AVC). That is to say, there are constraints on the variability of a conforming datastream produced by the video source 10 according to an example.
- An encoded, transcoded or otherwise processed video datastream is supplied to a transmit buffer 12, and then to a radio transmitter 14.
- the transmitter 14 operates in a known manner, and so its detailed operation will not be described here.
- the radio transmitter 14 operates to transmit a wireless radio frequency variable bit rate (VBR) signal 16 containing a packetized form of the video datastream to a radio receiver 20.
- the radio receiver 20 decodes the received radio frequency signal into a received datastream which is supplied to a receiver (rx) buffer 22.
- the receiver buffer 22 then supplies the received datastream to a video sink for example, a player.
- VBR wireless radio frequency variable bit rate
- the system 1 provides a number of modes of operation. Typically there will be three modes of operation which can be selected by the user. According to one example, the three modes of operation are a low delay interactive mode, a normal mode for general use and in which quality and delay are given approximately equal priority and a movie mode in which quality is prioritised. Source rate adaptation is provided in any of these modes to enable the system to deliver the desired video quality at a pre-determined delay over a long period of time.
- a frame work is provided for rate-adaptive streaming whereby the source rate adaptive scheme requires only loosely coupled system components, yet can still benefit from any additional information or control available.
- the video encoder's rate control is compliant with that of a suitable video buffer verification model, such as the H.264/AVC HRD for example.
- data packets are delivered to the transmission buffer 12, in preparation for transmission from transmitter 14. Mismatched source and channel rates are reconciled using buffer 12 as will be described below in more detail.
- Data packets are transmitted by transmitter 14 over VBR channel 16 and received by receiver 20. The process is then reversed with the datastream received by receiver 20 and provided to reception buffer 22 before being provided to the client, in this case video sink 24. The data can then be displayed to a user in the required format. Assuming negligible encode, decode and channel propagation times, the end to end delay of the system 1 is defined by the total amount of buffering delay in transmitter buffer 12 in combination with reception buffer 22.
- the modes which are available utilise a source rate adaptation scheme which controls the characteristics of the encoded bitstream, within certain delay constraints, by specifying video buffer verification model parameters to the video encoder.
- each operating mode maps to an appropriate set of such parameters.
- the rate adaption scheme operates on a windowed basis by predicting channel behaviour then choosing appropriate video buffer verification model parameters for an upcoming transmission window.
- the channel prediction step can use a model of channel behaviour which is updated at the beginning of each window using data collected during the previous period. The nature of these data updates is implementation specific and depends on the channel monitoring capabilities present in system 1.
- the data collected may include, but is not limited to, statistics such as received signal strength indication (RSSI) and/ or retry rates and/ or automatically selected modulation scheme of a wireless channel.
- the channel monitoring capabilities can include the presence of a bandwidth monitoring scheme. It will be appreciated that other statistics or system data may be collected and collated for use in the creation of a channel behaviour prediction.
- the parameters for the rate control are adjusted.
- the rate control adjustment aims to ensure that the decoder buffer does not underflow during the upcoming period.
- the channel behaviour prediction is accurate and the rate control complies with the HRD specified parameters.
- underflow and hence playout disruption is avoided.
- errors in the data bandwidth prediction of channel 16 can occasionally occur and may result in occasional underflow of data in buffer 22.
- worst case channel behaviour prediction can be used to adjust the rate control.
- the use of a worst case adjustment reduces the probability of playout disruption but increases the compression of the video, therefore reducing quality.
- the parameters are set such that bitstream size is reduced and playout resumes as soon as possible.
- a playout curve an example of which is shown in Figure 2
- the curve expresses the total number of bits produced by the encoder at any given time /.
- the curve p ⁇ f) is monotonically non-decreasing and is bounded by p(f) ⁇ pit) ⁇ p J) for all /. These bounds are imposed by virtue of the source's HRD compliance.
- the quantity of data received at the client is represented by r(i), the receiver curve.
- negligible packet loss is assumed. Typically, this can be achievable by means of a link layer retransmission scheme or similar. Changes in network throughput manifest themselves as variation in the gradient of r(i). Regardless of the channel condition, r(i) is bounded by pit) as the coded bitstream is generated in real time so it is not possible to transmit more data than has been generated at a given time.
- the end to end delay of the system, ⁇ is taken as a design parameter, selected by the user or set automatically by the system, and is the time between the insertion of a frame into the transmitter buffer 12 and the removal of that frame from the receiving buffer 22. Such delay is considered to be fixed for the duration of the session.
- a constraint curve is defined as cit)— rit + ⁇ ) and is a time-shifted version of the receiver curve r(i). This constraint curve represents the upper bound on the playout curve at the receiver.
- the aim of the adaptation scheme is therefore to manipulate the encoder such that pit) ⁇ c(t) for the duration of the streaming session.
- a given mode of operation maps to appropriate transmission rate and buffer size parameters for the system. That is to say, for the datastream in question, and for which a user has selected an appropriate mode of operation of the system, each mode represents the selection of an appropriate set of parameters.
- Each set characterises a model including three values— a peak transmission rate buffer capacity b v and encoder-side startup delay (in seconds) d e .
- the upper bound on p ⁇ f) can therefore be derived as p i)— p(f)+b v , where the lower bound is given by p ⁇ f) — R ⁇ t - d ⁇ ).
- the lower bound is not a bound as such; it represents the point at which the encoder buffer empties. If p ⁇ f) falls below p ⁇ f) then both p ⁇ f) and pu ⁇ f) must be recalculated.
- the nature of the channel makes it difficult to know the available bandwidth in advance. To make rate decisions, the system 1 must form an estimate of future channel behaviour.
- the user can select an operation mode to be used.
- the selection of the operation mode can be performed, for example, prior to the start of a movie or gaming ses sion, although it may also be selected during transmission, in which case there may be some disruption to video quality while buffering and associated parameters are reconfigured and the system adjusts.
- the system 1 uses a windowed HRD parameter adaptation scheme which runs with a period known as the window duration.
- a short window can be used as it will enable the system to be more responsive to changes in the environment such as a change in operation mode for example. It will be appreciated that shorter windows are desirable to facilitate responsiveness.
- the encoder's rate control in this case defines the minimum permitted window, w, in a given implementation; for example, the encoder may permit parameter adjustment only at pre-defined instantaneous Decoder Refresh (IDR) frames.
- IDR instantaneous Decoder Refresh
- a suitable window duration and phase is chosen to match the IDR interval.
- the controller performs the following steps:
- Figure 3 shows an example of the information available to the controller at some time, f , during a streaming session.
- the historic source rate is known.
- the achieved rate r(i) is determined to some desired degree of resolution by periodically measuring the changes in buffer occupancy.
- the achieved r(i) may alternatively be determined using other metrics, such as provided by a radio in a particular implementation. It will be appreciated that r(i) is not necessarily identical to the channel bandwidth as its value is restricted by r( ) ⁇ >( ). If the channel rate is higher than video rate for a sufficient period of time, this bound will be met.
- the value of c (/) is known only for / ⁇ f -A. Predictions for the upcoming period are shown by r(i) and c(j) in figure 3.
- r r ⁇ f find r r ⁇ f and c r ⁇ f) for a corresponding probability n.
- the method chosen will depend on information available to the controller. For example, an implementation can use retransmission statistics (if made available by a radio) in estimating r(i) and c(i).
- a suitable controller obtains information from various parts of the system 1 and performs adaptation by modifying parameters elsewhere in the system.
- FIG. 3a is a schematic block diagram of a video transmission system 100 according to an example.
- Controller 200 can receive information from the receiver 20 and/ or buffer 22, via an in- or out-of-band back channel. Controller 200 can receive information from video source 10, tx buffer 12 and/ or tx radio 14. The controller is operable to use information gathered from parts of system 100 to modify parameters for the system 100.
- the lower part of the network stack referred to above can include a physical (PHY) layer and a data link layer as de fined in the open systems interconnection (OSI) model for example.
- the PHY layer is typically part of radio hardware (not shown), whereas the data link layer can exist in the radio hardware in software executing on a processor to which the radio hardware is attached or otherwise communicating.
- reporting mechanisms can provide information from receiving radios or enable probing techniques to increase the quality of prediction.
- b v can be clipped to meet any constraints imposed by the encoder or the encoder level.
- the video buffer verification model parameters for the first window cannot be chosen in the same way as described above as no knowledge of the channel behaviour is available for use in generating the prediction prior to the first window.
- a channel prediction can be obtained during session initialisation.
- Such a channel prediction can be obtained in various different ways. For example, rate estimations can be obtained from statistics of recently-transmitted packets or dummy packets, or a channel prediction can be formed using RSSI figures from a radio.
- inaccuracies in the channel prediction can occasionally result in decoder underflow. If such underflow occurs, the decoder is typically able to decode the frames when they eventually arrive and catch up with the original playout schedule, suffering only jitter in the frame display times. Nevertheless, this jitter may be visible to the user and thus should be avoided if at all possible. If it is found that p ⁇ f f )>c ⁇ f )f ) during the adaption process, one or more frames will be delayed with a probability of at least n according to an example. To minimise the number of delayed frames, an alternative formula for can be used: where t t the recovery time, is the desired time until p i) is first less than or equal to c(t).
- Figure 4 is a block diagram of a method according to an example.
- a system such as system 1 in figure 1 for example, receives a user input indicative of a desired mode of operation.
- the user input can be provided using a dedicated remote control device or certain functions on an existing remote control device (such as using appropriately configured button(s) on a pre-existing remote control).
- the system configures transmission parameters in dependence upon the desired mode of operation as indicated by the received user input.
- encoding, transcoding, processing or otherwise adapting a video datastream for transmission over a communications channel occurs using the configured transmission parameters.
- FIG. 5 is a schematic block diagram of a video transmission system 500 according to an example.
- a video source 501 produces a video datastream 503 for consumption by a user of the system 500.
- the video datastream can be a movie, programme, portion of video, such as a portion used in a videogame, or any other sequence of still images representing scenes in motion for example.
- a user input device 505, such as a remote control device for example, is used to provide user input data 507 indicative of a desired mode of operation 509 of system 500.
- the user input device 505 can be part of the system 500, or can be a device which is part of a consumption apparatus, such as a television, monitor, or other display apparatus for example.
- the mode of operation 509 is a tailored mode for transmission of the datastream 503, and which is designed, in one example, to maximise quality or minimise latency of transmission.
- Video datastream 503 and the mode of operation 509 are input to block 511 which includes a datastream encoder 513 and datastream transmitter 515.
- encoder 513 and transmitter 515 can be distinct modules of the system. However, it will be appreciated that they can be incorporated into the same module without any loss of functionality.
- Encoder 513 is used to encode video stream data 503 into a format suitable for transmission and/or consumption.
- Encoder 513 can also function as a transcoder or module suitable for adapting a video datastream, such as adapting into a suitable video substream for example.
- Transmitter 515 can be a suitable wireless radio frequency transmitter.
- transmission parameters 517 are parameters which are suitable for the mode of operation 509.
- the suitably encoded datastream is transmitted over a communications channel, which can be a wireless communications channel.
- Figure 6 is a schematic block diagram of a video transmission system 600 according to one embodiment.
- the system 600 includes a video source 501 which produces a video datastream 503 for consumption by a user of the system 600.
- a user input device 505, such as a remote control device for example, is used to provide user input data 507 indicative of a desired mode of operation 509 of system 600.
- the mode of operation 509 is a tailored mode for transmission of the datastream 503, and which is designed to maximise quality or minimise latency of transmission for example.
- Video datastream 503 and the mode of operation 509 are input to block 511 which includes a datastream encoder 513 and datastream transmitter 515.
- encoder 513 and transmitter 515 can be distinct modules of the system. However, it will be appreciated that they can be incorporated into the same module without any loss of functionality.
- Encoder 513 is used to encode video stream data 503 into a format suitable for transmission and/or consumption.
- Encoder 513 can also function as a transcoder or module suitable for adapting a video datastream, such as adapting into a suitable video substream for example.
- Transmitter 515 can be a suitable wireless radio frequency transmitter.
- transmission parameters 517 are suitable for the mode of operation 509.
- the suitably encoded datastream is transmitted over a communications channel 601.
- a channel estimator 603 estimates channel conditions for such a communications channel 601.
- the datastream encoder 513 and transmitter 515 are operable to use such predicted channel conditions in the configuration of transmission parameters 517.
- the channel estimator 603 can estimate channel conditions repeatedly during datastream transmission.
- Communications channel 601 can be a wireless communications channel.
- video source 10 has been described as being an encoder or transcoder but may be any source suitable for generating data bitstreams, such as a source for adapting a scalable bitstream to provide substreams (such as in the case of bitstreams conforming to annex G of the H.264/AVC standard for example).
- the system may be operable to self select the mode of operation and create the input to reflect this such that and this self selection may be carried out intelligently in response to indicators provided by the datastream.
- the system can use some form of video content analysis to determine the content of a portion of video and select a mode of operation of the system accordingly and generate an input to reflect this.
- video content analysis can include motion detection to determine the presence of relevant motion in multiple scenes in the video object detection to determine the presence of a type of object or entity; and face recognition to detect the presence and movement of one or more people.
- motion detection to determine the presence of relevant motion in multiple scenes in the video object detection to determine the presence of a type of object or entity
- face recognition to detect the presence and movement of one or more people.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Engineering & Computer Science (AREA)
- Computer Security & Cryptography (AREA)
- Details Of Television Systems (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Mobile Radio Communication Systems (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
Claims
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB1314210.4A GB2501843A (en) | 2010-12-10 | 2011-12-09 | Video data transmission system and method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB1020949.2 | 2010-12-10 | ||
GBGB1020949.2A GB201020949D0 (en) | 2010-12-10 | 2010-12-10 | Multi channel video delivery |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012076904A1 true WO2012076904A1 (en) | 2012-06-14 |
Family
ID=43566974
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/GB2011/052451 WO2012076904A1 (en) | 2010-12-10 | 2011-12-09 | Video data transmission system and method |
Country Status (2)
Country | Link |
---|---|
GB (3) | GB201020949D0 (en) |
WO (1) | WO2012076904A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130243082A1 (en) * | 2012-03-15 | 2013-09-19 | Kabushiki Kaisha Toshiba | Rate optimisation for scalable video transmission |
WO2020074097A1 (en) * | 2018-10-12 | 2020-04-16 | Telefonaktiebolaget Lm Ericsson (Publ) | A monitoring and surveillance system arranged for processing video data associated with a vehicle, as well as corresponding devices and method |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1233622A2 (en) * | 2000-12-07 | 2002-08-21 | Matsushita Electric Industrial Co., Ltd. | Transmission rate control method |
US20070217623A1 (en) * | 2006-03-15 | 2007-09-20 | Matsushita Electric Industrial Co., Ltd. | Apparatus and method for real-time processing |
WO2008027842A2 (en) * | 2006-08-28 | 2008-03-06 | Ortiva Wireless, Inc. | Network adaptation of digital content |
-
2010
- 2010-12-10 GB GBGB1020949.2A patent/GB201020949D0/en not_active Ceased
-
2011
- 2011-02-15 GB GBGB1102616.8A patent/GB201102616D0/en not_active Ceased
- 2011-12-09 GB GB1314210.4A patent/GB2501843A/en not_active Withdrawn
- 2011-12-09 WO PCT/GB2011/052451 patent/WO2012076904A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1233622A2 (en) * | 2000-12-07 | 2002-08-21 | Matsushita Electric Industrial Co., Ltd. | Transmission rate control method |
US20070217623A1 (en) * | 2006-03-15 | 2007-09-20 | Matsushita Electric Industrial Co., Ltd. | Apparatus and method for real-time processing |
WO2008027842A2 (en) * | 2006-08-28 | 2008-03-06 | Ortiva Wireless, Inc. | Network adaptation of digital content |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130243082A1 (en) * | 2012-03-15 | 2013-09-19 | Kabushiki Kaisha Toshiba | Rate optimisation for scalable video transmission |
US9258557B2 (en) * | 2012-03-15 | 2016-02-09 | Kabushiki Kaisha Toshiba | Rate optimization for scalable video transmission |
WO2020074097A1 (en) * | 2018-10-12 | 2020-04-16 | Telefonaktiebolaget Lm Ericsson (Publ) | A monitoring and surveillance system arranged for processing video data associated with a vehicle, as well as corresponding devices and method |
US11665220B2 (en) | 2018-10-12 | 2023-05-30 | Telefonaktiebolaget Lm Ericsson (Publ) | Monitoring and surveillance system arranged for processing video data associated with a vehicle, as well as corresponding devices and method |
Also Published As
Publication number | Publication date |
---|---|
GB201020949D0 (en) | 2011-01-26 |
GB201314210D0 (en) | 2013-09-25 |
GB2501843A (en) | 2013-11-06 |
GB201102616D0 (en) | 2011-03-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10587664B2 (en) | Systems and methods for controlling the encoding of a segmented media stream using segment transmit times | |
US7652993B2 (en) | Multi-stream pro-active rate adaptation for robust video transmission | |
US7652994B2 (en) | Accelerated media coding for robust low-delay video streaming over time-varying and bandwidth limited channels | |
US11206431B2 (en) | Systems and methods for selecting an initial streaming bitrate | |
US7668170B2 (en) | Adaptive packet transmission with explicit deadline adjustment | |
US9544602B2 (en) | Wireless video transmission system | |
US10757481B2 (en) | Class-based intelligent multiplexing over unmanaged networks | |
EP3607724B1 (en) | Systems and methods for achieving optimal network bitrate | |
KR20160040320A (en) | Multipath rate adaptation | |
US20060136970A1 (en) | Joint bit rate control | |
WO2012076904A1 (en) | Video data transmission system and method | |
Stapenhurst et al. | Adaptive HRD parameter selection for fixed delay live wireless video streaming |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11805917 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 1314210 Country of ref document: GB Kind code of ref document: A Free format text: PCT FILING DATE = 20111209 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1314210.4 Country of ref document: GB |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 11805917 Country of ref document: EP Kind code of ref document: A1 |