WO2015169172A1 - 网络视频播放的方法和装置 - Google Patents

网络视频播放的方法和装置 Download PDF

Info

Publication number
WO2015169172A1
WO2015169172A1 PCT/CN2015/077895 CN2015077895W WO2015169172A1 WO 2015169172 A1 WO2015169172 A1 WO 2015169172A1 CN 2015077895 W CN2015077895 W CN 2015077895W WO 2015169172 A1 WO2015169172 A1 WO 2015169172A1
Authority
WO
WIPO (PCT)
Prior art keywords
video frame
independently decodable
fragment
frame
media
Prior art date
Application number
PCT/CN2015/077895
Other languages
English (en)
French (fr)
Inventor
杨晓峰
张仁宙
石腾
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to EP15789417.1A priority Critical patent/EP3142381B1/en
Priority to JP2016566765A priority patent/JP6314252B2/ja
Publication of WO2015169172A1 publication Critical patent/WO2015169172A1/zh
Priority to US15/343,749 priority patent/US10476928B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/764Media network packet handling at the destination 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/61Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/23439Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/262Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
    • H04N21/26258Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists for generating a list of items to be played back in a given order, e.g. playlist, or scheduling item distribution according to such list
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8455Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]

Definitions

  • the present invention relates to the field of communication network technologies, and in particular, to a method and apparatus for network video playback.
  • Perceptual coding is a new coding technology.
  • the coding technology based on background modeling is a typical representative of perceptual coding. It is mainly applied to scenes with basically unchanged backgrounds such as video surveillance, teleconferencing, and news broadcast.
  • the background-based coding technology not only reduces the scene redundancy in the background by the background modeling method, but also adopts a new group of pictures (GOP) structure, so it can be greatly based on the general coding technology. Increase the compression ratio of the code.
  • GOP group of pictures
  • the GOP is the basic structure in the code stream and the smallest unit that can completely decode the video sequence.
  • the traditional GOP includes an independently decodable video frame and multiple non-independent decoded video frames.
  • the number of video frames included in the GOP is called the GOP.
  • the length of the GOP can be appropriately increased, but the length of the GOP is too long, which also causes the defect of the error diffusion effect to be deteriorated; the coding technology based on background modeling improves the GOP structure, and the improved GOP
  • the independently decodable video frame is a background frame (G frame)
  • the non-independent decoded video frame includes a background update frame (S frame), a forward predicted frame (P frame), and a bidirectional predicted frame (B frame).
  • the S frame can only refer to the previous G frame, and the P frame and the B frame can refer to the previous G frame or the nearby S frame.
  • This improved GOP structure can suppress the error diffusion effect, so that the background modeling based coding technique can use a long GOP to further improve the coding efficiency.
  • the Dynamic Adaptive Streaming over HTTP (DASH) technology based on HyperText Transfer Protocol (HTTP) uses HTTP to transmit media content to users. It has become a trend in the development of the network video industry.
  • the media content on the server is sequentially cut into media segments. Each media segment has the same length of time, generally 2 to 10 seconds, and one media segment corresponds to an HTTP network address for the customer. The end obtains the media slice through the address.
  • the server provides a Media Presentation Description (MPD) file for recording the HTTP acquisition mode of the media fragments and the playback time period information of the media fragments.
  • the media slice can be further divided into subsegments, which are composed of one video frame.
  • a segment index (sidx) is defined in the DASH to indicate the starting position of each sub-segment in the media segment in the media segment, and the sidx also includes the playing duration of each sub-slice, and each sub-slice.
  • the access process of the existing DASH is: the terminal first acquires the MPD file, and determines the fragment corresponding to the access time point according to the access time point input by the user and the MPD file, and determines the fragment according to the fragment index sidx of the fragment. Enter the sub-segment corresponding to the time point and the first SAP location of the sub-slice, and start decoding and playing from the video frame corresponding to the first SAP.
  • the access time of the user input is The SAP distance actually accessed may be long, resulting in poor access accuracy and affecting the user's access experience.
  • the method and apparatus for network video playback provided by the embodiments of the present invention can solve the problem of poor access precision when transmitting a video stream based on background modeling and encoding by using the DASH method in the prior art.
  • an embodiment of the present invention provides a terminal for network video playback, including: a transceiver unit, a processing unit, and a playback unit;
  • the transceiver unit is configured to receive a video presentation of the video to describe the MPD file, and obtain an access time point requested by the user;
  • the processing unit is configured to determine, according to the MPD file, a media slice corresponding to the access time point;
  • the transceiver unit is further configured to acquire the media according to the media segment determined by the processing unit. a slice index of the slice, the slice index including location information of the independently decodable video frame and the non-independent decoded video frame of the independently decodable video frame in the media slice;
  • the processing unit is further configured to determine, according to the location information, a video that is closest to the access time point in the independently decodable video frame and the non-independent decoded video frame that references the independently decodable video frame. a frame, and determining that the closest video frame is a non-independent decoded video frame;
  • the transceiver unit is further configured to: after the processing unit determines that the closest video frame is a non-independent decoded video frame, acquire an independently decodable video frame referenced by the closest video frame;
  • the playing unit starts decoding and playing the video from the closest video frame according to the independently decodable video frame referenced by the closest video frame.
  • the processing unit is further configured to determine that the closest video frame is an independently decodable video frame, and the playing unit is further used in the processing unit After determining that the closest video frame is an independently decodable video frame, decoding and playing the video begins with the closest video frame.
  • the fragment index includes a description of a sub-shard of the media slice; the transceiver unit Obtaining a fragment index of the media fragment, where the fragment index includes location information of the independently decodable video frame and the non-independent decoded video frame that references the independently decodable video frame in the media fragment, specifically: obtaining the a fragment index of the media slice, where the description of the sub-slice of the media slice includes the independently decodable video frame in the sub-slice and the non-independent decoded video frame that references the independently decodable video frame location information.
  • the MPD file or the fragment index The frame identifier of the independently decodable video frame referenced by the non-independent decoded video frame is further included; the transceiver unit obtains the independently decodable video frame referenced by the closest video frame, specifically: according to the independently decodable decoding The frame identifier of the video frame acquires the independently decodable video referenced by the closest video frame frame.
  • the frame identifier of the independently decodable video frame includes a fragment identifier of the fragment where the independently decodable video frame is located, and The position information of the video frame in the fragment can be independently decoded; the independently decodable video frame referenced by the transceiver unit to obtain the closest video frame is specifically: according to the fragment of the independently decodable video frame The slice identifier, and location information of the independently decodable video frame in the slice, obtain an independently decodable video frame referenced by the closest video frame.
  • an embodiment of the present invention provides a server for network video playback, including:
  • a receiving unit configured to receive, by the terminal, a request for acquiring a fragment index of the media slice sent by the terminal after determining a media fragment corresponding to the access time point;
  • the sending unit is configured to return, according to the request, the fragment index to the terminal, where the fragment index includes a non-independent decoding of the independently decodable video frame and the reference independently decodable video frame in the media fragment.
  • the location information of the video frame is configured to return, according to the request, the fragment index to the terminal, where the fragment index includes a non-independent decoding of the independently decodable video frame and the reference independently decodable video frame in the media fragment.
  • the embodiment of the present invention provides a method for playing a network video, including: receiving a video presentation of an audio to describe an MPD file, and acquiring an access time point requested by the user; determining, according to the MPD file, the access a media slice corresponding to the time point; acquiring a slice index of the media slice, where the slice index includes the independently decodable video frame in the media slice and the non-independent decoded video frame that references the independently decodable video frame Position information; determining, according to the location information, a video frame that is closest to the access time point in the independently decodable video frame and the non-independent decoded video frame that references the independently decodable video frame; determining the The closest video frame is a non-independent decoded video frame, the independently decodable video frame referenced by the closest video frame is obtained; the independently decodable video frame referenced according to the closest video frame, from the closest The video frame begins to decode and play the video.
  • the method further includes determining that the closest video frame is an independently decodable video frame, and decoding and playing the video from the closest video frame.
  • the fragment index includes a description of a sub-slice of the media slice;
  • the index includes location information of the independently decodable video frame in the media slice and the non-independent decoded video frame that references the independently decodable video frame, specifically: including in the description of the sub-slice of the media slice
  • the position information of the non-independent decoded video frame that can independently decode the video frame and the independently decodable video frame can be independently decoded in the sub-slice.
  • the frame identifier of the independently decodable video frame referenced by the non-independent decoded video frame is further included; the independently decodable video frame that is referenced by the most recent video frame is specifically: according to the independently decodable video frame The frame identifier acquires an independently decodable video frame referenced by the most recent video frame.
  • the frame identifier of the independently decodable video frame includes a fragment identifier of the fragment where the independently decodable video frame is located, and The position information of the video frame in the slice can be independently decoded.
  • the solution can independently decode a video frame and reference location information of a non-independent decoded video frame that can independently decode a video frame in a fragment index including a media slice, thereby Determining a video frame closest to an access time point in the independently decodable video frame and the non-independent decoded video frame, and if the closest video frame is a non-independent decoded video frame, obtaining the closest video frame reference
  • the independently decodable video frame can be decoded and played back from the closest video frame according to the referenced independently decodable video frame, thereby improving the accuracy of terminal access.
  • FIG. 1 is a flowchart of a network video playing method according to an embodiment of the present invention
  • FIG. 2 is a flowchart of a method for playing a network video according to another embodiment of the present invention.
  • FIG. 3 is a schematic diagram of a media slice structure when a separate slice is not encapsulated in the corresponding embodiment of FIG. 2 according to the present invention
  • FIG. 4 is a schematic diagram of a fragment structure when a package is separately sliced in the corresponding embodiment of FIG. 2 according to the present invention
  • FIG. 5 is an example of a slice index in the corresponding embodiment of FIG. 2;
  • FIG. 6 is a flowchart of a method for playing a network video according to another embodiment of the present invention.
  • FIG. 7 is an example of a slice index in the corresponding embodiment of FIG. 6 according to the present invention.
  • FIG. 8 is an example of an MPD file in the corresponding embodiment of FIG. 6 of the present invention.
  • FIG. 9 is a structural diagram of a terminal for network video playback according to an embodiment of the present invention.
  • FIG. 10 is a structural diagram of a server for network video playback according to an embodiment of the present invention.
  • FIG. 11 is a structural diagram of a network video playback terminal based on a computer system according to an embodiment of the present invention.
  • An embodiment of the present invention provides a method for playing a network video.
  • a terminal acquires an MPD file from a server, and obtains a media slice according to the MPD file to play the video.
  • Figure 1 for the method performed by the terminal, including:
  • Step 101 Receive a video presentation of the video to describe the MPD file, and obtain an access time point requested by the user.
  • Step 102 Determine a media slice corresponding to the access time point according to the MPD file.
  • Step 103 Acquire a fragment index of the media fragment, where the fragment index includes the media segment
  • the video frame and the position information of the non-independent decoded video frame that can independently decode the video frame can be independently decoded in the slice.
  • the independently decodable video frame is a background frame
  • the non-independent decoded video frame is a background update frame or a forward predicted frame.
  • the fragment index includes a description of a sub-slice of the media slice, where the fragment index includes a non-independent decoding of the independently decodable video frame and the reference independently decodable video frame in the media slice.
  • the location information of the video frame may include: in the description of the sub-slice of the media slice, the independently decodable video frame in the sub-slice and the non-independent decoded video frame that references the independently decodable video frame. Location information.
  • Step 104 Determine, according to the location information, a video frame that is closest to the access time point in the independently decodable video frame and the non-independent decoded video frame that references the independently decodable video frame.
  • Step 105 Determine that the closest video frame is a non-independent decoded video frame, and obtain an independently decodable video frame referenced by the closest video frame.
  • Step 106 Decode and play the video from the closest video frame according to the independently decodable video frame referenced by the closest video frame.
  • the method may further include determining that the closest video frame is an independently decodable video frame, and decoding and playing the video from the closest video frame.
  • the fragment index in step 103 further includes a frame identifier of the independently decodable video frame referenced by the non-independent decoded video frame
  • obtaining the independently decodable video frame reference in step 105 The video frame is specifically: acquiring the independently decodable video frame referenced by the closest video frame according to the frame identifier of the independently decodable video frame included in the fragment index.
  • the independently decodable video frame referenced by the closest video frame is obtained in step 105.
  • the independently decodable video frame referenced by the closest video frame is obtained according to the frame identifier of the independently decodable video frame included in the MPD file.
  • the frame identifier of the independently decodable video frame includes a fragment identifier of the fragment in which the independently decodable video frame is located, and location information of the independently decodable video frame in the fragment; optionally,
  • the fragment of the independently decodable video frame is a separate fragment including only the independently decodable video frame, where the MPD file includes an acquisition address of the individual fragment; the determining and the access time according to the MPD file.
  • the method further includes: acquiring the individual fragment according to the acquiring address of the individual fragment.
  • the fragment index obtained in step 103 further includes the types of the independently decodable video frame and the non-independent decoded video frame, where the determining the closest video frame is an independently decodable video frame or Specifically, the non-independent decoding of the video frame is: determining, by the fragmentable index, the independently decodable video frame and the non-independent decoded video frame type, that the closest video frame is an independently decodable video frame or is not independently decoded. Video frame.
  • An embodiment of the present invention provides a method for playing a network video.
  • a video stream based on background modeling and encoding is transmitted in a DASH manner, and an independently decodable video frame referenced by a non-independent decoded video frame in a video frame corresponding to the SAP.
  • the frame identifier can be included in the MPD file or in the slice index.
  • Step 201 Receive a video presentation of the video to describe the MPD file, and obtain an access time point requested by the user.
  • the step may be that the terminal requests the MPD file of the video from the server according to the video selected by the user, and receives the MPD file returned by the server, and obtains the start of the user directly inputting in the player interface or by dragging the progress bar.
  • the viewing time that is, the access time point.
  • the step may be that the terminal resumes the video that has not been viewed last time according to the user selection, obtains the server or the locally saved replay time, that is, the access time point, and requests the MPD file of the video from the server. And receiving the MPD file returned by the server, or directly obtaining the MPD file of the video that has been downloaded from the local cache.
  • the MPD file records the acquisition address and the playback duration of each media fragment in the video.
  • the acquisition address is composed of a base address and a fragment identifier of the video, where the fragment identifier is used to indicate the media fragment and the base address.
  • the media address may be obtained from the server according to the obtained address, for example, the base address of the video is: http://www.example.com/server/Segments, and the fragment identifier of the media fragment is determined. For seg1.m4s, the media shard is obtained at http://www.example.com/server/Segments/seg1.m4s.
  • Step 202 Determine, according to the MPD file, a media fragment corresponding to the access time point.
  • the step may be: calculating, according to the playing duration of each media segment recorded in the MPD file, the acquiring address of the media segment corresponding to the access time point, for example, the playing duration of each media segment recorded in the MPD.
  • Both are 2 seconds
  • the fragment identifiers of the included fragment list are seg0.m4s, seg1.m4s, seg2.m4s, seg3.m4s, seg4.m4s, seg5.m4s, ..., and the access time is 7 seconds.
  • the fragment identifier of the media segment corresponding to the access time point is seg3.m4s.
  • Step 203 Acquire a fragment index of the media fragment, where the fragment index includes location information of the SAP in the media fragment, and the video frame corresponding to the SAP is an independently decodable video frame in the media fragment.
  • the fragment index includes location information of the SAP in the media fragment
  • the video frame corresponding to the SAP is an independently decodable video frame in the media fragment.
  • the fragment index further including a frame identifier of the independently decodable video frame referenced by the non-independent decoded video frame in the video frame corresponding to the SAP.
  • the fragment index may have the following two storage modes on the server: mode 1.
  • the fragment index may be located in a separate index fragment, and the fragment identifier of the index fragment is included in the description information of the fragment in the MPD.
  • Method 2 The fragment index may be located at the head of the corresponding media slice.
  • the fragmentation index of the media fragment is obtained by: firstly, determining whether the fragmentation index is stored according to the fragmentation identifier of the index fragment in the description information of the media fragment of the MPD file, if the description information is included.
  • the slice identifier of the index fragment is determined to be the first storage mode, and the index fragment is obtained according to the fragment identifier, and the fragment index is obtained from the index fragment, otherwise determined as the second type.
  • the fragment index is parsed from the head of the media fragment.
  • the fragment identifier of the index fragment is: seg1-index.m4s
  • the fragmentation index further includes a description of the sub-shards of the media shards, where the shard index includes the location information of the SAP in the media shards:
  • the description contains the location information of each SAP in the sub-slice.
  • the description of the sub-slices of the media slice includes the time or position offset of each SAP relative sub-fragment start position in the sub-slice.
  • the video frame that can be independently decoded in the video frame corresponding to the SAP is a background frame (G frame), and the non-independent decoded video frame in the video frame corresponding to the SAP is a forward predicted frame (P frame) Or a background update frame (S frame), the P frame or the S frame refers to the previous G frame, and the P frame or the S frame can be decoded after acquiring the referenced G frame.
  • G frame background frame
  • P frame or the S frame refers to the previous G frame
  • the P frame or the S frame can be decoded after acquiring the referenced G frame.
  • the type of the video frame corresponding to the SAP is further included in the fragment index, and is used to indicate whether the video frame corresponding to the SAP is an independently decodable video frame or a non-independent decoded video frame.
  • the independently decodable video frame and the non-independent decoded video frame of the video are sequentially encapsulated in the media fragment, and the SAP fragment in the media fragment is included in the fragment index of the media fragment.
  • Location information that is, media frame can independently decode video frames and position information of non-independent decoded video frames that can independently decode video frames.
  • the independently decodable video frames of the video can be encapsulated in separate slices.
  • the MPD file includes the acquisition address of the individual fragment; for example, the description of the initialization information of the MPD file includes the fragment identifier of the individual fragment, and the separate fragment is formed by the base address of the video and the fragment identifier. Obtaining an address; the terminal needs to acquire the individual fragment according to the obtained address to the local before determining the media fragment corresponding to the access time point according to the MPD file.
  • FIG. 3 is a schematic diagram showing a structure of a media slice when a separate slice is not encapsulated, wherein the frame marked by the shadow is a video frame corresponding to the SAP in the media slice, for example, a G frame, a P frame, and an S frame.
  • the arrow indicates that the non-independent decoded video frame in the video frame corresponding to the SAP refers to the reference of the independently decodable video frame in the media slice. Relationships, such as P frames and S frames, refer to G frames.
  • the frame identifier of the independently decodable video frame included in the fragmentation index specifically includes: a fragment identifier of the media fragment where the independently decodable video frame is located, and the independently decodable video frame is
  • the location information in the media slice may be a frame sequence number, or may be a storage offset address.
  • FIG. 4 is a schematic diagram showing the structure of a slice when a separate slice is packaged, wherein the frame marked by the shadow is a video frame corresponding to the SAP in the media slice, for example, a G frame, a P frame, and an S frame.
  • the arrow indicates that the non-independent decoded video frame in the video frame corresponding to the SAP refers to the reference relationship of the independently decodable video frame in the independent slice, for example, the P frame and the S frame refer to the G frame.
  • the frame identifier of the independently decodable video frame included in the fragmentation index specifically includes: a fragment identifier of the individual fragment where the independently decodable video frame is located, and the independently decodable video frame is
  • the location information in the separate fragment may be a frame sequence number, or may be a storage offset address.
  • Figure 5 shows an example of a slice index in the two examples corresponding to Figures 3 and 4: the slice index contains a description of each child slice, wherein the description is included in the description of any of the child fragments.
  • the description of each SAP in the sub-slice includes four parameters in the description of any SAP: SAP_type, SAP_delta_time, SegmentURL and Frame_Position, where the parameter SAP_type is used to indicate the type of the video frame corresponding to the SAP, parameter SAP_delta_time For indicating the time offset of the SAP relative to the starting position of the sub-slice, when the parameter SAP_type indicates that the video frame corresponding to the SAP is a non-independent decoded video frame, the parameters SegmentURL and Frame_Position have valid values, and the parameter SegmentURL is used.
  • the fragment identifier of the fragment in which the G frame referenced by the video frame corresponding to the SAP is located, and the parameter Frame_Position is used to indicate the frame number or storage offset address of the G frame in the fragment.
  • the fragment in which the G frame is located refers to the media fragment.
  • the fragment in which the G frame is located refers to the individual fragmentation.
  • the modification cost to the existing system is the lowest; in the example corresponding to FIG. 4, the independently decodable video frames are encapsulated in separate by the server.
  • the terminal determines the media corresponding to the access time point according to the MPD file
  • the fragmentation is downloaded to the local device before fragmentation.
  • the G frame needs to be acquired in real time.
  • Resolving the media shards results in slow access or high cost of downloading all media shards in advance, so the speed of acquiring G frames is accelerated in the case of minimum space overhead, but because of separate packaging and modification of MPDs
  • the file, compared to the first example, is more costly to modify the existing system.
  • Step 204 Determine, according to the location information of the SAP, the SAP that is closest to the access time point in the media slice, where the video frame corresponding to the closest SAP is the first video frame.
  • the step includes: first, determining a sub-segment including the access time point according to a play duration of each sub-segment included in the fragment index, and secondly, acquiring each of the sub-slices included in the fragment index.
  • the time offset of the SAP relative to the starting position of the sub-slice, the actual playing time of each SAP in the sub-slice is calculated, thereby determining which access time point is closest to which SAP in the sub-slice.
  • the fragment index of seg3.m4s is obtained in step 203, and the fragment index describes the media fragment.
  • the sub-segment sub_seg0, sub_seg1, sub_seg2, and sub_seg3 are respectively included, and the play durations are 0.5 seconds, 0.6 seconds, 0.4 seconds, and 0.5 seconds, respectively, so that the access time point is determined to be located in the sub-slice sub_seg1, where sub_seg1 includes three SAP: SAP0, SAP1, and SAP2, whose time offset values are 0.1, 0.2, and 0.4, respectively, so that the actual playback time of SAP0, SAP1, and SAP2 is calculated as 6.6 seconds, 6.7 seconds, and 6.9 seconds, respectively, and SAP2 is determined.
  • the access time point which is the 7th second.
  • the method for determining the SAP closest to the access time point is only an optional manner.
  • the playback duration of each sub-segment included in the fragment index and each SAP in the sub-slice may be acquired.
  • the actual play time of all SAPs is calculated, and the SAP closest to the access time point is determined in all SAPs.
  • the client can select the closest SAP in the SAP before or after the access time point according to its own strategy, and details are not described herein again.
  • Step 205 Determine whether the first video frame is an independently decodable video frame, if the result is determined If the video frame is not independently decoded, step 206 is performed. Otherwise, if the result of the determination is that the video frame can be independently decoded, step 208 is performed to start decoding and playing the video from the first video frame.
  • the determining whether the first video frame is an independently decodable video frame is: determining whether the first video frame is an independently decodable video frame according to a type of the video frame corresponding to the SAP included in the fragment index. For example, it is judged based on the value of the SAP_type parameter in the example of step 203.
  • Step 206 Acquire an independently decodable video frame referenced by the first video frame according to a frame identifier of the independently decodable video frame included in the fragment index.
  • the frame identifier of the independently decodable video frame is a slice identifier of the media slice in which the video frame can be independently decoded, and the independently decodable video frame is in the media slice.
  • the step is specifically: obtaining a base address of the video from the MPD, and forming the media identifier of the media slice by the slice identifier of the media slice and the base address of the video. And then obtaining an independently decodable video frame referenced by the first video frame from the media slice according to the location information.
  • the frame identifier of the independently decodable video frame includes a slice identifier and a frame sequence number, where the value of the slice identifier is test01.m4s, and the value of the frame sequence number is 3, after the media slice is obtained, according to the child
  • the offset address information of each frame included in the fragment header is parsed from the media slice to obtain a third video frame; for example, the frame identifier of the independently decodable video frame includes a slice identifier and a storage offset The address is moved, where the value of the fragment identifier is test01.m4s, and the value of the storage offset address is 0x10. After the media fragment is obtained, the query is obtained according to the offset address information of each frame included in the sub-segment header.
  • the video frame is obtained by parsing the offset address information of the next frame of the video frame, that is, the end address information of the video frame.
  • the media slice may be directly obtained from the local.
  • the frame identifier of the independently decodable video frame includes a slice identifier in a separate slice in which the independently decodable video frame is located, and an independently decodable video frame in the
  • the location information in the fragmentation is specifically as follows: the local identifier is correspondingly obtained from the local fragment. Independently fragmenting, and then obtaining an independently decodable video frame referenced by the first video frame from the independent fragment according to the location information.
  • the only difference from the previous example is that the implementation of this step is to obtain a separate fragment corresponding to the fragment identifier from the local.
  • Step 207 Decode and play the video from the first video frame according to the independently decodable video frame referenced by the first video frame.
  • the step is to sequentially obtain the media segment corresponding to the access time point and the subsequent media segment from the server according to the MPD file, and decode the independently decodable video frame referenced by the first video frame, so that the first video frame can be And subsequent video frames are decoded to play the video from the first video frame.
  • the density of the SAP (including both the G frame and the S frame and the P frame that references the G frame) is increased by extending the definition of the SAP, and is included in the media slice in the slice index of the media slice.
  • the location information of all SAPs and the frame identifier of the G frame referenced by the non-independent decoded video frame in the video frame corresponding to the SAP so that the terminal can more accurately determine the closest to the access time point according to the access time point and the fragment index.
  • the SAP when the type of the closest SAP corresponding video frame is an independently decodable video frame, can directly decode and play from the closest SAP corresponding video frame, when the type of the closest SAP corresponding video frame is
  • the G frame referenced by the closest SAP corresponding video frame may be found by using the frame identifier included in the fragment index, and the closest SAP corresponding video frame is decoded according to the G frame.
  • the close SAP corresponding video frame starts to play, thereby improving the accuracy of terminal access.
  • Step 601 - Step 602 Refer to steps 201-202 of the corresponding embodiment of FIG. 2.
  • Step 603 Obtain a fragment index of the media fragment, where the fragment index includes location information of the SAP in the media fragment, and the video frame corresponding to the SAP is an independently decodable video frame in the media fragment. Or refer to a non-independent decoded video frame that can independently decode a video frame.
  • step 203 of the corresponding embodiment of FIG. 2 refers to step 203 of the corresponding embodiment of FIG. 2, and the difference lies in the fragmentation cable in this embodiment.
  • the frame identifier of the independently decodable video frame referenced by the non-independent decoded video frame in the video frame corresponding to the SAP is not included.
  • Figure 7 shows an example of a slice index: the slice index contains a description of each child slice, wherein the description of each SAP in the sub-slice is included in the description of any sub-slice, in the pair
  • the description of any SAP contains two parameters: SAP_type, SAP_delta_time, the parameter SAP_type is used to indicate the type of video frame corresponding to the SAP, and the parameter SAP_delta_time is used to indicate the time offset of the SAP relative to the starting position of the sub-slice. the amount.
  • Step 604 Determine, in the media segment, the SAP that is closest to the access time point according to the location information of the SAP, where the video frame corresponding to the closest SAP is the first video frame.
  • step 704 can also record the sequence number of the closest SAP in the SAP list of the media slice.
  • the closest SAP is the third SAP in the second sub-slice sub_seg1 of the media slice
  • the total number of SAPs of all the previous sub-slices in the media segment is 8
  • the serial number of the closest SAP obtained in the SAP list of the media slice is 11.
  • Step 605 Determine whether the first video frame is an independently decodable video frame. If the result of the determination is a non-independent decoding video frame, perform step 606. Otherwise, if the result of the determination is that the video frame can be independently decoded, step 608 is performed. The first video frame begins to decode and play the video.
  • Step 606 The frame identifier of the independently decodable video frame referenced by the non-independent decoded video frame in the video frame corresponding to the SAP is included in the MPD file received by the step 601, and the first video frame reference is obtained according to the frame identifier.
  • Video frames can be decoded independently.
  • the independently decodable video frame and the non-independent decoded video frame of the video are sequentially encapsulated in the media fragment, and the SAP fragment in the media fragment is included in the fragment index of the media fragment.
  • Location information that is, media slices can be independently decoded video frames and references can be independently decoded video
  • the position information of the non-independent decoded video frame of the frame; optionally, the independently decodable video frame of the video may be encapsulated in a separate fragment, and the MPD file includes the obtained address of the separate fragment, for example, an MPD file pair.
  • the description of the initialization information includes the fragment identifier of the individual fragment, and the base address of the video and the fragment identifier constitute the acquisition address of the individual fragment, and the terminal needs to determine the media score corresponding to the access time point according to the MPD file.
  • the slice is previously acquired locally according to the acquired address.
  • the frame identifier of the independently decodable video frame included in the MPD file specifically includes: a fragment identifier of the media slice in which the independently decodable video frame is located, and the independent The position information of the video frame in the media slice is decoded, and the location information may be a frame sequence number, or may be a storage offset address.
  • the frame identifier of the independently decodable video frame included in the fragmentation index specifically includes: a fragment identifier of the individual fragment in which the independently decodable video frame is located, and the The position information of the video frame in the separate fragment is independently decoded, and the location information may be a frame sequence number, or may be a storage offset address.
  • Figure 8 shows an example of an MPD file in the two examples corresponding to Figures 3 and 4: the MPD file contains a description of each media slice, wherein the description of any media slice is included
  • the description of the G frame referenced by the SAP in the media fragment includes three parameters in the description of any referenced G frame: SAPCountRange, SegmentURL, Frame_Position, respectively, for indicating the range of the SAP sequence number that references the G frame,
  • SAPCountRange the description of any referenced G frame
  • SegmentURL SegmentURL
  • Frame_Position respectively, for indicating the range of the SAP sequence number that references the G frame
  • the slice identifier of the slice where the G frame is located the frame number of the G frame in the slice, or the storage offset address.
  • the fragment in which the G frame is located refers to the media fragment.
  • the fragment in which the G frame is located refers to the individual fragmentation.
  • the slice identifier of the slice is test01.m4s, and the frame of the G frame in the slice
  • the serial number is 3.
  • the fragment identifier of the fragment where the frame is located is test01.m4s, and the storage offset address of the G frame in the fragment is 0x10.
  • step 206 of the corresponding embodiment of FIG. 2 is obtained after the frame identifier included in the MPD is obtained.
  • the modification cost to the existing system is the lowest; in the example corresponding to FIG. 4, the independently decodable video frames are stored in the server by the server.
  • the terminal downloads the separate fragment to the local before determining the media fragment corresponding to the access time point according to the MPD file, it can be avoided that the P frame or the S frame and the G frame it refers to are not in the video frame corresponding to the SAP.
  • the media fragment of the G frame needs to be obtained in real time, resulting in slow access speed or high cost of downloading all media fragments in advance. Therefore, the G frame is accelerated when the space overhead is minimum. Speed, but because of the need to separately package separate shards and modify MPD files, the cost of making changes to existing systems is greater than the first example.
  • Step 607 Refer to step 207 of the corresponding embodiment of FIG. 2.
  • the density of the SAP (including both the G frame and the S frame and the P frame that references the G frame) is increased by extending the definition of the SAP, and is included in the media slice in the slice index of the media slice.
  • the closest SAP at the time point when the type of the closest SAP corresponding video frame is an independently decodable video frame, can directly decode and play from the closest SAP corresponding video frame, when the closest SAP corresponding video
  • the G frame referenced by the closest SAP corresponding video frame may be found by using the frame identifier included in the fragment index, and the closest SAP corresponding video frame is decoded according to the G frame.
  • the closest SAP corresponding video frame starts playing, thereby improving terminal access Precision.
  • An embodiment of the present invention provides a terminal for network video playback.
  • the terminal 900 includes:
  • the transceiver unit 901 is configured to receive a video presentation of the video to describe the MPD file, and obtain an access time point requested by the user;
  • the processing unit 902 is configured to determine, according to the MPD file, a media slice corresponding to the access time point;
  • the transceiver unit 901 is further configured to acquire, according to the media fragment determined by the processing unit 902, a fragment index of the media fragment, where the fragment index includes independently decodable video frames and references in the media fragment. Decoding location information of the non-independent decoded video frame of the video frame;
  • the processing unit 902 is further configured to determine, according to the location information, a video frame that is closest to the access time point in the independently decodable video frame and the non-independent decoded video frame that references the independently decodable video frame. And determining that the closest video frame is a non-independent decoded video frame;
  • the transceiver unit 901 is further configured to: after the processing unit determines that the closest video frame is a non-independent decoded video frame, acquire an independently decodable video frame referenced by the closest video frame;
  • the playing unit 903 decodes and plays the video from the closest video frame according to the independently decodable video frame referenced by the closest video frame.
  • the processing unit 902 is further configured to determine that the closest video frame is an independently decodable video frame
  • the playing unit 903 is further configured to determine, in the processing unit 902, the closest video frame as an independently decodable video. After the frame, the video is decoded and played starting from the closest video frame.
  • the fragment index includes a description of a sub-shard of the media fragment
  • the transceiver unit 901 obtains a fragment index of the media fragment, where the fragment index includes the media fragment.
  • the location information of the independently decodable video frame and the non-independently decoded video frame that can be independently decoded are specifically: obtaining a fragment index of the media fragment, and describing the sub-slice of the media fragment Included in the sub-slices are independently decodable video frames and non-independent decoded video that references independently decodable video frames The position information of the frame.
  • the MPD file or the fragment index further includes a frame identifier of the independently decodable video frame referenced by the non-independent decoded video frame; and the transceiver unit 901 obtains the reference of the closest video frame.
  • the independently decoding the video frame is specifically: acquiring the independently decodable video frame referenced by the closest video frame according to the frame identifier of the independently decodable video frame.
  • the frame identifier of the independently decodable video frame includes a fragment identifier of the fragment where the independently decodable video frame is located, and location information of the independently decodable video frame in the fragment;
  • the 901 obtains the independently decodable video frame referenced by the closest video frame, specifically: a fragment identifier according to the fragment where the independently decodable video frame is located, and the independently decodable video frame in the fragment.
  • the location information acquires the independently decodable video frames referenced by the closest video frame.
  • the fragment of the independently decodable video frame is a separate fragment that includes only the independently decodable video frame, where the MPD file includes an acquisition address of the individual fragment, and the transceiver unit 901 is configured to Before the determining, by the processing unit, the media fragment corresponding to the access time point, the processing unit is further configured to acquire the individual fragment according to the acquiring address of the individual fragment.
  • the fragment index further includes a type of the independently decodable video frame and the non-independent decoded video frame; and the processing unit 902 is further configured to: according to the independently decodable video frame and the non-independent The type of decoded video frame determines that the closest video frame is an independently decodable video frame or a non-independent decoded video frame.
  • the transceiver unit 901 is further configured to acquire, according to the media fragment determined by the processing unit 902, a fragment index of the media fragment, where the fragment index includes an independently decodable video frame in the media fragment.
  • the location information of the non-independent decoded video frame that references the independently decodable video frame further includes: the independently decodable video frame is a background frame, and the non-independent decoded video frame is a background update frame or a forward predicted frame.
  • the terminal can be a mobile phone, a PC, a tablet or any other network video playback device in the technical field.
  • the server 1000 includes:
  • the receiving unit 1001 is configured to receive, by the terminal, a request for acquiring a fragment index of the media slice sent by the terminal after determining a media fragment corresponding to the access time point;
  • the sending unit 1002 is configured to return, according to the request, the fragment index to the terminal, where the fragment index includes an independently decodable video frame in the media slice and a non-independent decoded video that references an independently decodable video frame. Position information of the frame, so that the terminal determines the video frame closest to the access time point in the independently decodable video frame and the non-independent decoded video frame according to the location information, The closest video frame begins to decode and play the video.
  • the terminal for network video playing in the embodiment of the present invention may be implemented based on a computer system, and the methods shown in FIG. 1, FIG. 2, and FIG. 6 may all be implemented in a terminal for network video playback based on a computer system.
  • Figure 11 illustrates an embodiment of a terminal for network video playback implemented based on a computer system.
  • the terminal for network video playing in this embodiment may include: a processor 1101, a memory 1102, and a communication interface 1103, where:
  • the communication interface 1103 is configured to communicate with a server. Messages exchanged between the terminal of the network video playback and the server are transmitted and received through the communication interface 1103. Specifically, the communication interface 1103 is configured to receive an MPD file of the video returned by the server, and an access time point for acquiring the user request; the memory 1102 is configured to store the program instruction; and the processor 1101 is configured to receive the MPD file of the video by the communication interface 1103.
  • the program instruction stored in the memory 1102 is invoked, and the following operations are performed: determining a media slice corresponding to the access time point according to the MPD file; and acquiring the media slice a slice index, the slice index including location information of the independently decodable video frame and the non-independent decoded video frame of the independently decodable video frame in the media slice; and the independently decodable according to the location information Determining, in the non-independent decoded video frame of the independently decodable video frame, a video frame that is closest to the access time point; determining that the closest video frame is a non-independent solution And a coded video frame, the independently decodable video frame referenced by the closest video frame is obtained; and the video is decoded and played from the closest video frame according to the independently decodable video frame referenced by the closest video frame.
  • the processor 1101 may be a central processing unit (CPU), an application-specific integrated circuit (ASIC), or the like.
  • the terminal for network video playback in this embodiment may include a bus 1104.
  • the processor 1101, the memory 1102, and the communication interface 1103 can be connected and communicated via the bus 1104.
  • the memory 1102 may include: a random access memory (RAM), a read-only memory (ROM), a disk and the like having a storage function.
  • the processor 1101 can also be used to perform the steps of the terminal for performing network video playback as described in FIG. 1 , FIG. 2 and FIG. 6 in the method embodiment, and the embodiments of the present invention are not described in detail herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

本发明提供了一种网络视频播放的方法和装置,所述方法包括:接收视频的MPD文件,获取用户请求的接入时间点;根据MPD文件确定与接入时间点对应的媒体分片;获取该媒体分片的分片索引,分片索引包含该媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息;根据该位置信息,在该可独立解码视频帧和该非独立解码视频帧中确定与接入时间点最接近的视频帧;如果最接近的视频帧为非独立解码视频帧,获取最接近的视频帧引用的可独立解码视频帧;根据该引用的可独立解码视频帧,从最接近的视频帧开始解码并播放视频。使用本发明可提高网络视频播放的接入精度。

Description

网络视频播放的方法和装置 技术领域
本发明涉及通信网络技术领域,尤其涉及一种网络视频播放的方法和装置。
背景技术
感知编码是一种新的编码技术,基于背景建模的编码技术是感知编码的典型代表,其主要应用于视频监控、电话会议、新闻播报等背景基本不变的场景。基于背景建模的编码技术除了通过背景建模的方法有效地减少背景中的场景冗余,还采用了新的图片组(Group of pictures,GOP)结构,因此能在通用编码技术的基础上大幅提高编码压缩率。
GOP是码流中的基本结构,也是视频序列能完整解码的最小单元,传统的GOP包含一个可独立解码视频帧和多个非独立解码视频帧,GOP中包含的视频帧个数称为该GOP的长度,为了提高压缩率,可以适当的增加GOP的长度,但是GOP的长度太长也会带来误差扩散效应变差的缺陷;基于背景建模的编码技术改进了GOP结构,改进的GOP中可独立解码视频帧为背景帧(G帧),非独立解码视频帧包含背景更新帧(S帧)、前向预测帧(P帧)以及双向预测帧(B帧)。其中S帧只能参考前一G帧,P帧和B帧可以引用之前的G帧或附近的S帧。这种改进的GOP结构可以抑制误差扩散效应,从而使得基于背景建模的编码技术可以采用很长的GOP来进一步提高编码效率。
基于超文本传输协议(HyperText Transfer Protocol,HTTP)的动态自适应流媒体(Dynamic Adaptive Streaming over HTTP,DASH)技术采用HTTP的方式向用户传送媒体内容,目前已成为网络视频行业发展的趋势,其关键是将服务器上的媒体内容顺序地切割成媒体分片(segment),每个媒体分片的时间长度相同,一般为2~10秒,一个媒体分片对应于一个HTTP网络地址,以便客户 端通过该地址获取该媒体分片。服务器提供媒体展示描述(Media Presentation Description,MPD)文件,用于记录这些媒体分片的HTTP获取方式以及媒体分片的播放时间段信息。媒体分片可以被继续划分为子分片(subsegment),子分片又由一个个视频帧组成。DASH中定义了分片索引(segment index,sidx)用以指示媒体分片中各个子分片在媒体分片中的起始位置,sidx中还包含各个子分片的播放时长、各个子分片中第一个流接入点(Stream Access Point,SAP)的位置信息。
现有DASH的接入流程是:终端首先获取MPD文件,根据用户输入的接入时间点以及该MPD文件,确定该接入时间点对应的分片,根据该分片的分片索引sidx确定接入时间点对应的子分片以及该子分片的第一个SAP位置,并从第一个SAP对应的视频帧开始解码播放。
当基于背景建模编码的码流采用DASH进行传输时,因为GOP很长,且当前一个GOP只定义了一个SAP、即定义GOP中的可独立解码帧为SAP,因此用户输入的接入时间与实际接入的SAP距离可能很长,导致接入精度差,影响用户的接入体验。
发明内容
本发明的实施例提供的网络视频播放的方法和装置,能够解决现有技术中采用DASH方式传输基于背景建模编码的视频流时,接入精度差的问题。
第一方面,本发明实施例提供了一种网络视频播放的终端,包括:收发单元,处理单元和播放单元;
所述收发单元,用于接收视频的媒体展示描述MPD文件,并获取用户请求的接入时间点;
所述处理单元,用于根据所述MPD文件确定与所述接入时间点对应的媒体分片;
所述收发单元,还用于根据所述处理单元确定的所述媒体分片获取所述媒 体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息;
所述处理单元,还用于根据所述位置信息,在所述可独立解码视频帧和所述引用可独立解码视频帧的非独立解码视频帧中确定与所述接入时间点最接近的视频帧,以及确定所述最接近的视频帧为非独立解码视频帧;
所述收发单元,还用于在所述处理单元确定所述最接近的视频帧为非独立解码视频帧后,获取所述最接近的视频帧引用的可独立解码视频帧;
所述播放单元,根据所述最接近的视频帧引用的可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
在第一方面的第一种可能的实现方式中,所述处理单元,还用于确定所述最接近的视频帧为可独立解码视频帧;所述播放单元,还用于在所述处理单元确定所述最接近的视频帧为可独立解码视频帧后,从所述最接近的视频帧开始解码并播放视频。
结合第一方面或第一方面的第一可能的实现方式,在第二种可能的实现方式中,所述分片索引中包含对所述媒体分片的子分片的描述;所述收发单元获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息具体为:获取所述媒体分片的分片索引,在所述对所述媒体分片的子分片的描述中包含所述子分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
结合第一方面或第一方面的第一种可能的实现方式或第一方面的第二种可能的实现方式,在第三种可能的实现方式中,所述MPD文件中或所述分片索引中还包含所述非独立解码视频帧所引用的可独立解码视频帧的帧标识;所述收发单元获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧的帧标识获取所述最接近的视频帧引用的可独立解码视频 帧。
结合第一方面的第三种可能的实现方式,在第四种可能的实现方式中,所述可独立解码视频帧的帧标识包括所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息;所述收发单元获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息获取所述最接近的视频帧引用的可独立解码视频帧。
第二方面,本发明实施例提供了一种网络视频播放的服务器,包括:
接收单元,用于接收终端在确定与接入时间点对应的媒体分片后发送的获取所述媒体分片的分片索引的请求;
所述发送单元,用于根据所述请求向所述终端返回所述分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
第三方面,本发明实施例提供了一种网络视频播放的方法,包括:接收视频的媒体展示描述MPD文件,并获取用户请求的接入时间点;根据所述MPD文件确定与所述接入时间点对应的媒体分片;获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息;根据所述位置信息,在所述可独立解码视频帧和所述引用可独立解码视频帧的非独立解码视频帧中确定与所述接入时间点最接近的视频帧;确定所述最接近的视频帧为非独立解码视频帧,获取所述最接近的视频帧引用的可独立解码视频帧;根据所述最接近的视频帧引用的可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
在第三方面的第一种可能的实现方式中,所述方法还包括:确定所述最接近的视频帧为可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
结合第三方面或第三方面的第一可能的实现方式,在第二种可能的实现方式中,所述分片索引中包含对所述媒体分片的子分片的描述;所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息具体为:在所述对所述媒体分片的子分片的描述中包含所述子分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
结合第三方面或第三方面的第一种可能的实现方式或第三方面的第二种可能的实现方式,在第三种可能的实现方式中,所述MPD文件中或所述分片索引中还包括所述非独立解码视频帧所引用的可独立解码视频帧的帧标识;所述获取所述最近接的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧的帧标识获取所述最近接的视频帧引用的可独立解码视频帧。
结合第三方面的第三种可能的实现方式,在第四种可能的实现方式中,所述可独立解码视频帧的帧标识包括所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息。
由上述本发明的实施例提供的技术方案可以看出,本方案通过在分片索引包含媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息,从而在所述可独立解码视频帧和所述非独立解码视频帧确定与接入时间点最接近的视频帧,如果该最接近的视频帧为非独立解码视频帧,获取该最接近的视频帧引用的可独立解码视频帧,根据该引用的可独立解码视频帧,从最接近的视频帧开始解码并播放视频,从而提高终端接入的精度。
附图说明
为了更清楚地说明本发明实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下, 还可以根据这些附图获得其他的附图。
图1为本发明实施例提供的网络视频播放方法流程图;
图2为本发明另一实施例提供的网络视频播放方法流程图;
图3为本发明图2对应实施例中当不封装单独分片时的媒体分片结构示意图;
图4为本发明图2对应实施例中当封装单独分片时的分片结构示意图;
图5为本发明图2对应实施例中分片索引的一个示例;
图6为本发明再一实施例提供的网络视频播放方法流程图;
图7为本发明图6对应实施例中分片索引的一个示例;
图8为本发明图6对应实施例中MPD文件的一个示例;
图9为本发明实施例提供的网络视频播放的终端结构图;
图10为本发明实施例提供的网络视频播放的服务器结构图;
图11为本发明实施例提供的基于计算机系统的网络视频播放终端结构图。
具体实施方式
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。
本发明实施例提供了一种网络视频播放的方法,该方法中,终端从服务器获取MPD文件,并根据MPD文件获取媒体分片后播放视频。终端执行的方法请参照图1,具体包括:
步骤101:接收视频的媒体展示描述MPD文件,并获取用户请求的接入时间点。
步骤102:根据所述MPD文件确定与所述接入时间点对应的媒体分片。
步骤103:获取所述媒体分片的分片索引,所述分片索引包含所述媒体分 片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
其中,该可独立解码视频帧为背景帧,该非独立解码视频帧为背景更新帧或前向预测帧。
其中,所述分片索引中包含对所述媒体分片的子分片的描述,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息具体可以为:在所述对所述媒体分片的子分片的描述中包含所述子分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
步骤104:根据所述位置信息,在所述可独立解码视频帧和所述引用可独立解码视频帧的非独立解码视频帧中确定与所述接入时间点最接近的视频帧。
步骤105:确定所述最接近的视频帧为非独立解码视频帧,获取所述最接近的视频帧引用的可独立解码视频帧。
步骤106:根据所述最接近的视频帧引用的可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
可选的,该方法还可以包括:确定所述最接近的视频帧为可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
可选的,当步骤103中的分片索引还包含所述非独立解码视频帧所引用的可独立解码视频帧的帧标识时,步骤105中获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述分片索引包含的所述可独立解码视频帧的帧标识获取所述最接近的视频帧引用的可独立解码视频帧。
可选的,当步骤101中的MPD文件包含所述非独立解码视频帧所引用的可独立解码视频帧的帧标识时,步骤105中获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述MPD文件包含的所述可独立解码视频帧的帧标识获取所述最接近的视频帧引用的可独立解码视频帧。
其中,所述可独立解码视频帧的帧标识包括所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息;可选的,所述可独立解码视频帧所在分片为仅包括可独立解码视频帧的单独分片,所述MPD文件中包含所述单独分片的获取地址;所述根据MPD文件确定与所述接入时间点对应的媒体分片之前,所述方法还包括:根据所述单独分片的获取地址获取所述单独分片。
可选的,步骤103获取到的分片索引中还包含所述可独立解码视频帧和所述非独立解码视频帧的类型,所述确定所述最接近的视频帧为可独立解码视频帧或者为非独立解码视频帧具体为:通过所述分片索引包含的可独立解码视频帧和所述非独立解码视频帧类型确定所述最接近的视频帧为可独立解码视频帧或者为非独立解码视频帧。
本发明实施例提供了网络视频播放的方法,该方法中,基于背景建模编码的视频流采用DASH的方式进行传输,SAP对应的视频帧中非独立解码视频帧所引用的可独立解码视频帧的帧标识可以包含在MPD文件中或分片索引中。
当SAP对应的视频帧中非独立解码视频帧所引用的可独立解码视频帧的帧标识包含在分片索引时,终端接入到视频流的详细流程请参照图2,具体包括:
步骤201:接收视频的媒体展示描述MPD文件,并获取用户请求的接入时间点。
具体的,该步骤可以是终端根据用户选择的视频,向服务器请求该视频的MPD文件,并接收服务器返回的MPD文件,以及获取用户在播放器界面中直接输入或者通过拖动进度条选择的开始观看时间、即接入时间点。
具体的,该步骤还可以是终端根据用户选择对上次未看完的视频进行续播,获取到服务器或者本地保存的续播时间、即接入时间点,向服务器请求该视频的MPD文件,并接收服务器返回的MPD文件,或者直接从本地缓存中获取已经下载过的该视频的MPD文件。
其中,MPD文件记录了该视频中各媒体分片的获取地址以及播放时长,该获取地址由视频的基地址和分片标识组成,其中,分片标识用于指示该媒体分片与基地址的相对存储地址,终端可以根据该获取地址从服务器获取该媒体分片,例如,视频的基地址(BaseURL)为:http://www.example.com/server/Segments,媒体分片的分片标识为seg1.m4s,则该媒体分片的获取地址为:http://www.example.com/server/Segments/seg1.m4s。
步骤202:根据所述MPD文件确定与所述接入时间点对应的媒体分片。
具体的,该步骤可以是根据MPD文件中记录的各媒体分片的播放时长计算出该接入时间点对应的媒体分片的获取地址,例如,MPD中记录的每个媒体分片的播放时长均为2秒,包含的分片列表的分片标识分别为seg0.m4s、seg1.m4s、seg2.m4s、seg3.m4s、seg4.m4s、seg5.m4s、…,接入时间点为7秒处,则接入时间点对应媒体分片的分片标识为seg3.m4s。
步骤203:获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中SAP的位置信息,所述SAP对应的视频帧为所述媒体分片中可独立解码视频帧或引用可独立解码视频帧的非独立解码视频帧,所述分片索引还包含所述SAP对应的视频帧中非独立解码视频帧所引用的可独立解码视频帧的帧标识。
其中,分片索引在服务器可以有以下两种存储方式:方式一、分片索引可以位于单独的索引分片中,并且在MPD对该分片的描述信息中包含该索引分片的分片标识;方式二、分片索引可以位于对应媒体分片的头部。
获取媒体分片的分片索引具体为:首先,根据查找MPD文件的对该媒体分片的描述信息中是否包含索引分片的分片标识来确定分片索引的存储方式,如果描述信息中包含索引分片的分片标识,则确定为第一种存储方式,根据分片标识获取到所述索引分片,并从所述索引分片中获取到该分片索引,否则确定为第二种存储方式,获取该媒体分片后从该媒体分片的头部解析出分片索引。例如,当对媒体分片的描述信息是<SegmentURL media='seg1.m4s'index= ‘seg1-index.m4s'/>,则确定出存在索引分片,且索引分片的分片标识为:seg1-index.m4s;再例如,对媒体分片的描述信息是<SegmentURL media='seg1.m4s'/>,则确定分片索引在媒体分片的头部。该步骤的实现具体过程可以按照现有技术,此处不再赘述。
其中,分片索引中还包含对所述媒体分片的子分片的描述,则所述分片索引包含所述媒体分片中SAP的位置信息可以为:在对媒体分片的子分片的描述中包含该子分片中每个SAP的位置信息。例如,在对媒体分片的子分片的描述中包含该子分片中每个SAP相对子分片起始位置的时间或位置偏移量。
其中,在本实施例中,所述SAP对应的视频帧中可独立解码视频帧为背景帧(G帧),所述SAP对应的视频帧中非独立解码视频帧为前向预测帧(P帧)或者背景更新帧(S帧),所述P帧或S帧引用之前的G帧,获取到该引用的G帧后即可对所述P帧或S帧解码。
可选的,在分片索引中还包括所述SAP对应的视频帧的类型,用于指示SAP对应的视频帧为可独立解码视频帧、还是非独立解码视频帧。
服务器在进行媒体分片的封装时,将视频的可独立解码视频帧和非独立解码视频帧依次封装在媒体分片中,并在该媒体分片的分片索引中包含该媒体分片中SAP的位置信息,即媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息,可选的,还可以将视频的可独立解码视频帧封装在单独分片中,并且MPD文件中包含该单独分片的获取地址;例如,MPD文件对初始化信息的描述中包含该单独分片的分片标识,通过视频的基地址和该分片标识组成该单独分片的获取地址;终端需要在根据MPD文件确定与接入时间点对应的媒体分片之前根据该获取地址将该单独分片获取到本地。
图3给出了当不封装单独分片时的媒体分片结构示意图,其中通过阴影标记的帧为媒体分片中SAP对应的视频帧,例如G帧、P帧、S帧。箭头表示SAP对应的视频帧中非独立解码视频帧引用媒体分片中可独立解码视频帧的引用 关系,例如P帧和S帧引用G帧。采用该封装方式时,分片索引中包含的所述可独立解码视频帧的帧标识具体包括:所述可独立解码视频帧所在媒体分片的分片标识、以及所述可独立解码视频帧在该媒体分片中的位置信息,位置信息可以是帧序号,还可以是存储偏移地址。
图4给出了当封装单独分片时的分片结构示意图,其中通过阴影标记的帧为媒体分片中SAP对应的视频帧,例如G帧、P帧、S帧。箭头表示SAP对应的视频帧中非独立解码视频帧引用独立分片中可独立解码视频帧的引用关系,例如P帧和S帧引用G帧。采用该封装方式时,分片索引中包含的所述可独立解码视频帧的帧标识具体包括:所述可独立解码视频帧所在单独分片的分片标识、以及所述可独立解码视频帧在该单独分片中的位置信息,位置信息可以是帧序号,还可以是存储偏移地址。
图5给出了图3和图4对应的两个例子中,分片索引的一个示例:分片索引包含对每个子分片的描述,其中,在对任一子分片的描述中包含该子分片中每个SAP的描述,在对任一SAP的描述中包含了四个参数:SAP_type,SAP_delta_time,SegmentURL和Frame_Position,其中,参数SAP_type用于表示该SAP对应的视频帧的类型,参数SAP_delta_time用于表示该SAP相对于该子分片起始位置的时间偏移量,当参数SAP_type指示该SAP对应的视频帧为非独立解码视频帧时,参数SegmentURL和Frame_Position才具有有效值,参数SegmentURL用于表示该SAP对应的视频帧引用的G帧所在分片的分片标识,参数Frame_Position用于表示该G帧在该分片中的帧序号或存储偏移地址。在图3对应的例子中,G帧所在分片指媒体分片,在图4对应的例子中,G帧所在分片指单独分片。
在图3对应的例子中,不需要更改服务器现有的分片封装方式,因此对现有系统的改动成本最低;在图4对应的例子中,通过服务器将可独立解码视频帧封装在单独分片中,并且终端在根据MPD文件确定与接入时间点对应的媒体 分片之前将单独分片下载到本地,可以避免第一个例子中当SAP对应的视频帧中P帧或S帧和它引用的G帧不在同一个媒体分片时,需要实时获取该G帧所在媒体分片再解析导致访问速度慢、或者事先下载所有媒体分片开销大的问题,因此在空间开销最小的情况下加快了获取G帧的速度,但是因为要另外封装单独分片和修改MPD文件,和第一个例子相比对现有系统的改动成本更大。
步骤204:根据所述SAP的位置信息,在所述媒体分片中确定与所述接入时间点最接近的SAP,所述最接近的SAP对应的视频帧为第一视频帧。
具体的,该步骤包括:首先,根据分片索引中包含的各个子分片的播放时长确定包含该接入时间点的子分片,其次,获取分片索引中包含的该子分片中各个SAP相对子分片起始位置的时间偏移量,计算该子分片中各个SAP的实际播放时间,从而确定接入时间点与该子分片中哪个SAP最近。
例如,在步骤202中根据接入时间点为第7秒,确定媒体分片为seg3.m4s后,在步骤203中获取到seg3.m4s的分片索引,该分片索引描述了该媒体分片包含4个子分片sub_seg0、sub_seg1、sub_seg2、sub_seg3,它们的播放时长分别为0.5秒、0.6秒、0.4秒、0.5秒,从而确定该接入时间点位于子分片sub_seg1中,其中sub_seg1包含三个SAP:SAP0、SAP1和SAP2,它们的时间偏移量值分别为0.1、0.2和0.4,从而计算出SAP0、SAP1和SAP2的实际播放时间分别为第6.6秒、6.7秒、6.9秒,并确定SAP2距离接入时间点,即第7秒最近。
当然,以上确定与接入时间点最接近的SAP的方法只是一种可选方式,例如,还可以是获取所述分片索引中包含的各个子分片的播放时长和子分片中每个SAP相对子分片起始位置的时间偏移量后,计算所有SAP的实际播放时间,在所有SAP中确定与接入时间点最近的SAP。而且,客户端可以根据自己的策略,只在接入时间点之前或者之后的SAP中选择距离最近的SAP,这里不再赘述。
步骤205:判断所述第一视频帧是否为可独立解码视频帧,如果判断结果 为非独立解码视频帧,则执行步骤206,否则判断结果为可独立解码视频帧,则执行步骤208,从所述第一视频帧开始解码并播放视频。
其中,判断第一视频帧是否为可独立解码视频帧具体为:根据分片索引中包含的SAP对应的视频帧的类型判断所述第一视频帧是否为可独立解码视频帧。例如,根据步骤203的例子中的SAP_type参数值判断。
步骤206:根据分片索引中包含的所述可独立解码视频帧的帧标识获取所述第一视频帧引用的可独立解码视频帧。
在步骤203提到的图3对应的例子中,所述可独立解码视频帧的帧标识是可独立解码视频帧所在媒体分片的分片标识、以及可独立解码视频帧在所述媒体分片中的位置信息,则该步骤具体为:从MPD中获取视频的基地址,将所述媒体分片的分片标识和视频的基地址组成媒体分片的获取地址后,从服务器获取该媒体分片,然后根据所述位置信息从所述媒体分片中获取第一视频帧引用的可独立解码视频帧。
例如,所述可独立解码视频帧的帧标识包含的是分片标识和帧序号,其中分片标识的值是test01.m4s,帧序号的值是3,则获取该媒体分片后,根据子分片头部包含的每个帧的偏移地址信息,从该媒体分片中解析得到第3视频帧;再例如,所述可独立解码视频帧的帧标识包含的是分片标识和存储偏移地址,其中分片标识的值是test01.m4s,存储偏移地址的值是0x10,则获取该媒体分片后,根据子分片头部包含的每个帧的偏移地址信息,查询得到该视频帧下一帧的偏移地址信息、即该视频帧的结束地址信息后解析获得该视频帧。可选的,该步骤中如果第一视频帧和其引用的G帧在同一个媒体分片中,并且该媒体分片已经获取过,则可以直接从本地获取该媒体分片。
在步骤203提到的图4对应的例子中,所述可独立解码视频帧的帧标识包含的是可独立解码视频帧所在单独分片中的分片标识、以及可独立解码视频帧在所述单独分片中的位置信息,则该步骤具体为:从本地获取到该分片标识对应 的独立分片,然后根据所述位置信息从独立分片中获取第一视频帧引用的可独立解码视频帧。和上一个例子对应步骤的区别仅仅在于,该步骤的实现是从本地获取与分片标识对应的单独分片。
步骤207:根据所述第一视频帧引用的可独立解码视频帧,从所述第一视频帧开始解码并播放视频。
具体的,该步骤是根据MPD文件从服务器依次获取接入时间点对应的媒体分片以及之后的媒体分片,对第一视频帧引用的可独立解码视频帧解码,从而可以对第一视频帧及随后的视频帧进行解码,从第一视频帧开始播放视频。
在该实施例中,通过扩展SAP的定义增加SAP的密度(既包含G帧,还包含引用G帧的S帧和P帧),并在媒体分片的分片索引中包含该媒体分片中所有SAP的位置信息、以及该SAP对应的视频帧中非独立解码视频帧所引用G帧的帧标识,使得终端根据接入时间点和分片索引能更精确的确定与接入时间点最接近的SAP,当该最接近的SAP对应视频帧的类型为可独立解码视频帧时、可以直接从该最接近的SAP对应视频帧开始解码和播放,当该最接近的SAP对应视频帧的类型为非独立解码视频帧时,可以通过分片索引包含的所述帧标识找到该最接近的SAP对应视频帧引用的G帧,并根据该G帧对最接近的SAP对应视频帧解码后从该最接近的SAP对应视频帧开始播放,从而提高了终端接入的精度。
当SAP对应的视频帧中非独立解码视频帧所引用的可独立解码视频帧的帧标识包含在MPD文件时,终端接入到视频流的详细流程请参照图6,具体包括:
步骤601-步骤602:参考图2对应实施例的步骤201-202。
步骤603:获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中SAP的位置信息,所述SAP对应的视频帧为所述媒体分片中可独立解码视频帧或引用可独立解码视频帧的非独立解码视频帧。
该步骤的实现参考图2对应实施例的步骤203,区别在于本实施例中分片索 引不包含所述SAP对应的视频帧中非独立解码视频帧所引用的可独立解码视频帧的帧标识。
图7给出了分片索引的一个示例:分片索引包含对每个子分片的描述,其中,在对任一子分片的描述中包含该子分片中每个SAP的描述,在对任一SAP的描述中包含了两个参数:SAP_type,SAP_delta_time,参数SAP_type用于表示该SAP对应的视频帧的类型,参数SAP_delta_time用于表示该SAP相对于该子分片起始位置的时间偏移量。
步骤604:根据所述SAP的位置信息在所述媒体分片中确定与所述接入时间点最接近的SAP,所述最接近的SAP对应的视频帧为第一视频帧。
该步骤的具体实现参考图2对应实施例的步骤204。
进一步的,步骤704还可以记录该最接近的SAP在该媒体分片的SAP列表中的序号。
例如,在计算出最接近的SAP是媒体分片的第二个子分片sub_seg1中的第3个SAP后,进一步的,根据该媒体分片中之前的所有子分片的SAP总数为8,从而得到该最接近的SAP在该媒体分片的SAP列表中的序号是11。
步骤605:判断所述第一视频帧是否为可独立解码视频帧,如果判断结果为非独立解码视频帧,则执行步骤606,否则判断结果为可独立解码视频帧,则执行步骤608,从所述第一视频帧开始解码并播放视频。
该判断操作的具体实现参考图2对应实施例的步骤205。
步骤606:步骤601接收到的MPD文件中包含所述SAP对应的视频帧中非独立解码视频帧所引用的可独立解码视频帧的帧标识,根据该帧标识获取所述第一视频帧引用的可独立解码视频帧。
服务器在进行媒体分片的封装时,将视频的可独立解码视频帧和非独立解码视频帧依次封装在媒体分片中,并在该媒体分片的分片索引中包含该媒体分片中SAP的位置信息,即媒体分片中可独立解码视频帧和引用可独立解码视频 帧的非独立解码视频帧的位置信息;可选的,还可以将视频的可独立解码视频帧封装在单独分片中,并且MPD文件中包含该单独分片的获取地址,例如,MPD文件对初始化信息的描述中包含该单独分片的分片标识,通过视频的基地址和该分片标识组成该单独分片的获取地址,终端需要在根据MPD文件确定与接入时间点对应的媒体分片之前根据该获取地址将该单独分片获取到本地。
这两种封装方式下的分片结构示意图参考图3和图4,以及图2对应实施例中步骤203对该图3和图4的详细描述。
当采用图3所示的封装方式时,MPD文件中包含的所述可独立解码视频帧的帧标识具体包括:所述可独立解码视频帧所在媒体分片的分片标识,以及所述可独立解码视频帧在该媒体分片中的位置信息,位置信息可以是帧序号,还可以是存储偏移地址。
在采用图4所示的封装方式时,分片索引中包含的所述可独立解码视频帧的帧标识具体包括:所述可独立解码视频帧所在单独分片的分片标识,以及所述可独立解码视频帧在该单独分片中的位置信息,位置信息可以是帧序号,还可以是存储偏移地址。
图8给出了图3和图4对应的两个例子中,MPD文件的一个示例:MPD文件中包含对每个媒体分片的描述,其中,在对任一媒体分片的描述中包含该媒体分片中SAP所引用的G帧的描述,在对任一被引用G帧的描述中包含了三个参数:SAPCountRange,SegmentURL,Frame_Position,分别用于表示引用该G帧的SAP序号范围、该G帧所在分片的分片标识、该G帧在分片中的帧序号或存储偏移地址。在图3对应的例子中,G帧所在分片指媒体分片,在图4对应的例子中,G帧所在分片指单独分片。
例如,参数SAPCountRange=”1,15”,参数SegmentURL=test01.m4s,参数Frame_Position指帧序号,且Frame_Position=3,其含义是序号范围为1-15的SAP引用该G帧,且G帧所在分片的分片标识为test01.m4s,该G帧在该分片中的帧 序号为3。
再例如,参数SAPCountRange=”1,15”,参数SegmentURL=test01.m4s,参数Frame_Position指存储偏移地址,且Frame_Position=0x10,其含义是序号范围为1-15的SAP引用该G帧,且G帧所在分片的分片标识为test01.m4s,该G帧在该分片中的存储偏移地址为0x10。
其中,当获取到MPD中包含的帧标识后,根据该帧标识获取所述第一视频帧引用的可独立解码视频帧的详细描述参见图2对应实施例的步骤206的描述。
在图3对应的例子中,不需要更改服务器现有的分片封装方式,因此对现有系统的改动成本最低;在图4对应的例子中,通过服务器将可独立解码视频帧存放在单独分片中,并且终端在根据MPD文件确定与接入时间点对应的媒体分片之前将单独分片下载到本地,可以避免当SAP对应的视频帧中P帧或S帧和它引用的G帧不在同一个媒体分片时,需要实时获取该G帧所在媒体分片在解析导致访问速度慢、或者事先下载所有媒体分片开销大的问题,因此在空间开销最小的情况下加快了获取G帧的速度,但是因为要另外封装单独分片和修改MPD文件,和第一个例子相比对现有系统的改动成本更大。
步骤607:参考图2对应实施例的步骤207。
在该实施例中,通过扩展SAP的定义增加SAP的密度(既包含G帧,还包含引用G帧的S帧和P帧),并在媒体分片的分片索引中包含该媒体分片中所有SAP的位置信息,以及在MPD中包含该SAP对应的视频帧中非独立解码视频帧所引用G帧的帧标识,使得终端根据接入时间点和分片索引能更精确的确定与接入时间点最接近的SAP,当该最接近的SAP对应视频帧的类型为可独立解码视频帧时、可以直接从该最接近的SAP对应视频帧开始解码和播放,当该最接近的SAP对应视频帧的类型为非独立解码视频帧时,可以通过分片索引包含的帧标识找到该最接近的SAP对应视频帧引用的G帧,并根据该G帧对最接近的SAP对应视频帧解码后从该最接近的SAP对应视频帧开始播放,从而提高了终端接入 的精度。
本发明一个实施例提供了网络视频播放的终端,请参照图9,该终端900包括:
收发单元901,用于接收视频的媒体展示描述MPD文件,并获取用户请求的接入时间点;
处理单元902,用于根据所述MPD文件确定与所述接入时间点对应的媒体分片;
收发单元901,还用于根据处理单元902确定的所述媒体分片获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息;
处理单元902,还用于根据所述位置信息,在所述可独立解码视频帧和所述引用可独立解码视频帧的非独立解码视频帧中确定与所述接入时间点最接近的视频帧,以及确定所述最接近的视频帧为非独立解码视频帧;
收发单元901,还用于在所述处理单元确定所述最接近的视频帧为非独立解码视频帧后,获取所述最接近的视频帧引用的可独立解码视频帧;
播放单元903,根据所述最接近的视频帧引用的可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
可选的,处理单元902,还用于确定所述最接近的视频帧为可独立解码视频帧;播放单元903,还用于在处理单元902确定所述最接近的视频帧为可独立解码视频帧后,从所述最接近的视频帧开始解码并播放视频。
可选的,所述分片索引中包含对所述媒体分片的子分片的描述,收发单元901获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息具体为:获取所述媒体分片的分片索引,在所述对所述媒体分片的子分片的描述中包含所述子分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频 帧的位置信息。
可选的,所述MPD文件中或所述分片索引中还包含所述非独立解码视频帧所引用的可独立解码视频帧的帧标识;收发单元901获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧的帧标识获取所述最接近的视频帧引用的可独立解码视频帧。
可选的,所述可独立解码视频帧的帧标识包括所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息;收发单元901获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息获取所述最接近的视频帧引用的可独立解码视频帧。
可选的,所述可独立解码视频帧所在分片为仅包括可独立解码视频帧的单独分片,所述MPD文件中包含所述单独分片的获取地址;收发单元901,用于在所述处理单元根据所述MPD文件确定与所述接入时间点对应的媒体分片之前,进一步用于根据所述单独分片的获取地址获取所述单独分片。
可选的,所述分片索引中还包含所述可独立解码视频帧和所述非独立解码视频帧的类型;处理单元902,还用于根据所述可独立解码视频帧和所述非独立解码视频帧的类型确定所述最接近的视频帧为可独立解码视频帧或者为非独立解码视频帧。
可选的,收发单元901,还用于根据处理单元902确定的所述媒体分片获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息进一步包括:所述可独立解码视频帧为背景帧,所述非独立解码视频帧为背景更新帧或前向预测帧。
本发明的所有实施例中,终端可以是手机、PC、平板电脑或技术领域内任意其他的网络视频播放设备。
本发明一个实施例提供了网络视频播放的服务器,请参照图10,该服务器1000包括:
接收单元1001,用于接收终端在确定与接入时间点对应的媒体分片后发送的获取所述媒体分片的分片索引的请求;
发送单元1002,用于根据所述请求向所述终端返回所述分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息,以便于所述终端根据所述位置信息在所述可独立解码视频帧和所述非独立解码视频帧中确定与所述接入时间点最接近的视频帧后,从所述最接近的视频帧开始解码并播放视频。
本发明实施例中的网络视频播放的终端可以基于计算机系统来实现,图1、图2、图6所示的方法均可在基于计算机系统的网络视频播放的终端来实现。图11示出了基于计算机系统来实现的网络视频播放的终端的实施例。本实施例中网络视频播放的终端可以包括:处理器1101、存储器1102和通信接口1103,其中:
通信接口1103,用于与服务器通信。网络视频播放的终端和服务器之间交互的消息均通过通信接口1103发送和接收。具体地,通信接口1103用于接收服务器返回的视频的MPD文件,以及用于获取用户请求的接入时间点;存储器1102用于存储程序指令;处理器1101用于通信接口1103接收视频的MPD文件并获取用户请求的接入时间点后,调用存储器1102中存储的程序指令,执行以下操作:根据所述MPD文件确定与所述接入时间点对应的媒体分片;获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息;根据所述位置信息,在所述可独立解码视频帧和所述引用可独立解码视频帧的非独立解码视频帧中确定与所述接入时间点最接近的视频帧;确定所述最接近的视频帧为非独立解 码视频帧,获取所述最接近的视频帧引用的可独立解码视频帧;根据所述最接近的视频帧引用的可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
其中,处理器1101可以是中央处理器(central processing unit,CPU)、专用集成电路(application-specific integrated circuit,ASIC)等。其中,本实施例中的网络视频播放的终端可以包括总线1104。处理器1101、存储器1102以及通信接口1103之间可通过总线1104连接并通信。其中,存储器1102可以包括:随机存取存储器(random access memory,RAM),只读存储器(read-only memory,ROM),磁盘等具有存储功能的实体。
处理器1101还可以用于执行方法实施例中图1、图2、图6描述的网络视频播放的终端的各步骤,本发明实施例在此不再详述。
以上所述,仅为本发明较佳的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到的变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应该以权利要求的保护范围为准。

Claims (18)

  1. 一种网络视频播放的终端,其特征在于,所述终端包括:收发单元,处理单元和播放单元;
    所述收发单元,用于接收视频的媒体展示描述MPD文件,并获取用户请求的接入时间点;
    所述处理单元,用于根据所述MPD文件确定与所述接入时间点对应的媒体分片;
    所述收发单元,还用于根据所述处理单元确定的所述媒体分片获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息;
    所述处理单元,还用于根据所述位置信息,在所述可独立解码视频帧和所述引用可独立解码视频帧的非独立解码视频帧中确定与所述接入时间点最接近的视频帧,以及确定所述最接近的视频帧为非独立解码视频帧;
    所述收发单元,还用于在所述处理单元确定所述最接近的视频帧为非独立解码视频帧后,获取所述最接近的视频帧引用的可独立解码视频帧;
    所述播放单元,根据所述最接近的视频帧引用的可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
  2. 根据权利要求1所述的终端,其特征在于:
    所述处理单元,还用于确定所述最接近的视频帧为可独立解码视频帧;
    所述播放单元,还用于在所述处理单元确定所述最接近的视频帧为可独立解码视频帧后,从所述最接近的视频帧开始解码并播放视频。
  3. 根据权利要求1或2所述的终端,其特征在于:
    所述分片索引中包含对所述媒体分片的子分片的描述;
    所述收发单元获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位 置信息具体为:获取所述媒体分片的分片索引,在所述对所述媒体分片的子分片的描述中包含所述子分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
  4. 根据权利要求1-3任一所述的终端,其特征在于:
    所述MPD文件中或所述分片索引中还包含所述非独立解码视频帧所引用的可独立解码视频帧的帧标识;
    所述收发单元获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧的帧标识获取所述最接近的视频帧引用的可独立解码视频帧。
  5. 根据权利要求4所述的终端,其特征在于:
    所述可独立解码视频帧的帧标识包括所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息;
    所述收发单元获取所述最接近的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息获取所述最接近的视频帧引用的可独立解码视频帧。
  6. 根据权利要求5所述的终端,其特征在于:
    所述可独立解码视频帧所在分片为仅包括可独立解码视频帧的单独分片,所述MPD文件中包含所述单独分片的获取地址;
    所述收发单元,用于在所述处理单元根据所述MPD文件确定与所述接入时间点对应的媒体分片之前,进一步用于根据所述单独分片的获取地址获取所述单独分片。
  7. 根据权利要求1-6任一所述的终端,其特征在于:
    所述分片索引中还包含所述可独立解码视频帧和所述非独立解码视频帧的类型;
    所述处理单元,还用于根据所述可独立解码视频帧和所述非独立解码视频 帧的类型确定所述最接近的视频帧为可独立解码视频帧或者为非独立解码视频帧。
  8. 根据权利要求1-7任一所述的终端,其特征在于,所述收发单元,还用于根据所述处理单元确定的所述媒体分片获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息进一步包括:所述可独立解码视频帧为背景帧,所述非独立解码视频帧为背景更新帧或前向预测帧。
  9. 一种网络视频播放的服务器,其特征在于,所述服务器包括:
    接收单元,用于接收终端在确定与接入时间点对应的媒体分片后发送的获取所述媒体分片的分片索引的请求;
    所述发送单元,用于根据所述请求向所述终端返回所述分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
  10. 一种网络视频播放的方法,其特征在于,所述方法包括:
    接收视频的媒体展示描述MPD文件,并获取用户请求的接入时间点;
    根据所述MPD文件确定与所述接入时间点对应的媒体分片;
    获取所述媒体分片的分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息;
    根据所述位置信息,在所述可独立解码视频帧和所述引用可独立解码视频帧的非独立解码视频帧中确定与所述接入时间点最接近的视频帧;
    确定所述最接近的视频帧为非独立解码视频帧,获取所述最接近的视频帧引用的可独立解码视频帧;
    根据所述最接近的视频帧引用的可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
  11. 根据权利要求10所述的方法,其特征在于,所述方法还包括:
    确定所述最接近的视频帧为可独立解码视频帧,从所述最接近的视频帧开始解码并播放视频。
  12. 根据权利要求10或11所述的方法,其特征在于:
    所述分片索引中包含对所述媒体分片的子分片的描述;
    所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息具体为:在所述对所述媒体分片的子分片的描述中包含所述子分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息。
  13. 根据权利要求10-12任一所述的方法,其特征在于:
    所述MPD文件中或所述分片索引中还包括所述非独立解码视频帧所引用的可独立解码视频帧的帧标识;
    所述获取所述最近接的视频帧引用的可独立解码视频帧具体为:根据所述可独立解码视频帧的帧标识获取所述最接近的视频帧引用的可独立解码视频帧。
  14. 根据权利要求13的方法,其特征在于:
    所述可独立解码视频帧的帧标识包括所述可独立解码视频帧所在分片的分片标识、以及所述可独立解码视频帧在所述分片中的位置信息。
  15. 根据权利要求14的方法,其特征在于:
    所述可独立解码视频帧所在分片为仅包括可独立解码视频帧的单独分片,所述MPD文件中包含所述单独分片的获取地址;
    所述根据MPD文件确定与所述接入时间点对应的媒体分片之前,所述方法还包括:根据所述单独分片的获取地址获取所述单独分片。
  16. 根据权利要求10-15任一所述的方法,其特征在于:
    所述分片索引中还包含所述可独立解码视频帧和所述非独立解码视频帧的类型;
    根据所述可独立解码视频帧和所述非独立解码视频帧的类型确定所述最接近的视频帧为可独立解码视频帧或者为非独立解码视频帧。
  17. 根据权利要求10-16任一所述的方法,其特征在于:
    所述可独立解码视频帧为背景帧,所述非独立解码视频帧为背景更新帧或前向预测帧。
  18. 一种网络视频播放的方法,其特征在于,所述方法包括:
    接收终端在确定与接入时间点对应的媒体分片后发送的获取所述媒体分片的分片索引的请求;
    向所述终端返回所述分片索引,所述分片索引包含所述媒体分片中可独立解码视频帧和引用可独立解码视频帧的非独立解码视频帧的位置信息,以便于所述终端根据所述位置信息在所述可独立解码视频帧和所述非独立解码视频帧中确定与所述接入时间点最接近的视频帧后,从所述最接近的视频帧开始解码并播放视频。
PCT/CN2015/077895 2014-05-05 2015-04-30 网络视频播放的方法和装置 WO2015169172A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP15789417.1A EP3142381B1 (en) 2014-05-05 2015-04-30 Network video playing method and device
JP2016566765A JP6314252B2 (ja) 2014-05-05 2015-04-30 ネットワークビデオ再生方法及び装置
US15/343,749 US10476928B2 (en) 2014-05-05 2016-11-04 Network video playback method and apparatus

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410186163.3A CN103957471B (zh) 2014-05-05 2014-05-05 网络视频播放的方法和装置
CN201410186163.3 2014-05-05

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/343,749 Continuation US10476928B2 (en) 2014-05-05 2016-11-04 Network video playback method and apparatus

Publications (1)

Publication Number Publication Date
WO2015169172A1 true WO2015169172A1 (zh) 2015-11-12

Family

ID=51334677

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/077895 WO2015169172A1 (zh) 2014-05-05 2015-04-30 网络视频播放的方法和装置

Country Status (5)

Country Link
US (1) US10476928B2 (zh)
EP (1) EP3142381B1 (zh)
JP (1) JP6314252B2 (zh)
CN (1) CN103957471B (zh)
WO (1) WO2015169172A1 (zh)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103957471B (zh) 2014-05-05 2017-07-14 华为技术有限公司 网络视频播放的方法和装置
CN105847825A (zh) * 2015-01-16 2016-08-10 杭州海康威视数字技术股份有限公司 视频编码码流的编码、索引存储和访问方法及相应装置
KR101916022B1 (ko) * 2015-10-06 2018-11-07 한국전자통신연구원 세그먼트 기반의 방송 콘텐츠 반복 전송 방법 및 장치
WO2018072675A1 (en) * 2016-10-18 2018-04-26 Zhejiang Dahua Technology Co., Ltd. Methods and systems for video processing
CN108632644B (zh) * 2017-03-21 2020-09-11 华为技术有限公司 预览图的展示方法以及设备
CN110351606B (zh) 2018-04-04 2022-12-27 华为技术有限公司 媒体信息处理方法、相关设备及计算机存储介质
CN110876083B (zh) * 2018-08-29 2021-09-21 浙江大学 指定参考图像的方法及装置及处理参考图像请求的方法及装置
CN111436005B (zh) * 2019-01-15 2022-03-08 北京字节跳动网络技术有限公司 用于显示图像的方法和装置
CN110719254A (zh) * 2019-08-29 2020-01-21 北斗航天卫星应用科技集团有限公司 多协议集成的北斗位置服务系统及服务方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110099594A1 (en) * 2009-10-28 2011-04-28 Qualcomm Incorporated Streaming encoded video data
CN102232298A (zh) * 2011-04-07 2011-11-02 华为技术有限公司 媒体内容的传输处理方法、装置与系统
CN102467940A (zh) * 2010-11-08 2012-05-23 安凯(广州)微电子技术有限公司 一种无索引视频文件快进快退方法、装置以及播放系统
CN102761773A (zh) * 2012-07-03 2012-10-31 Ut斯达康通讯有限公司 一种利用i帧改善iptv业务的用户体验的方法
CN103957471A (zh) * 2014-05-05 2014-07-30 华为技术有限公司 网络视频播放的方法和装置

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09182083A (ja) * 1995-12-27 1997-07-11 Matsushita Electric Ind Co Ltd ビデオ画像符号化方法及び復号化方法とその装置
KR100492567B1 (ko) * 2003-05-13 2005-06-03 엘지전자 주식회사 이동통신 시스템의 http 기반 비디오 스트리밍 장치및 방법
US8315307B2 (en) * 2004-04-07 2012-11-20 Qualcomm Incorporated Method and apparatus for frame prediction in hybrid video compression to enable temporal scalability
KR20060088303A (ko) * 2005-02-01 2006-08-04 엘지전자 주식회사 디지털 방송 수신기의 동영상 저장/재생 장치 및 방법
US7646812B2 (en) * 2005-04-01 2010-01-12 Microsoft Corporation Special predictive picture encoding using color key in source content
US9432433B2 (en) * 2006-06-09 2016-08-30 Qualcomm Incorporated Enhanced block-request streaming system using signaling or block creation
KR100765787B1 (ko) * 2006-06-15 2007-10-12 삼성전자주식회사 압축 부호화된 영상 데이터 스트림의 재생 방법 및 그 장치
CN101106637A (zh) * 2006-07-13 2008-01-16 中兴通讯股份有限公司 通过机顶盒实现对外接存储设备的媒体文件播放的方法
US8566860B2 (en) * 2007-11-15 2013-10-22 At&T Intellectual Property I, Lp System and methods for advanced parental control
KR101019634B1 (ko) * 2008-09-04 2011-03-07 에스케이 텔레콤주식회사 미디어 전송 시스템 및 방법
KR101421390B1 (ko) * 2010-06-29 2014-07-18 퀄컴 인코포레이티드 트릭 모드 비디오 표현물에 대한 비디오 샘플의 시그널링
US9485546B2 (en) * 2010-06-29 2016-11-01 Qualcomm Incorporated Signaling video samples for trick mode video representations
CN102123281A (zh) * 2011-01-07 2011-07-13 中兴通讯股份有限公司 视频播放时快进快退处理方法、装置以及移动终端
CN102136948B (zh) * 2011-03-15 2014-04-02 华为技术有限公司 用于统计用户体验的方法、终端设备和系统
US9264717B2 (en) 2011-10-31 2016-02-16 Qualcomm Incorporated Random access with advanced decoded picture buffer (DPB) management in video coding
US9124895B2 (en) * 2011-11-04 2015-09-01 Qualcomm Incorporated Video coding with network abstraction layer units that include multiple encoded picture partitions
EP2615790A1 (en) * 2012-01-12 2013-07-17 Alcatel Lucent Method, system and devices for improved adaptive streaming of media content
JP2013205680A (ja) 2012-03-29 2013-10-07 Panasonic Corp 音声出力装置および音声出力方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110099594A1 (en) * 2009-10-28 2011-04-28 Qualcomm Incorporated Streaming encoded video data
CN102467940A (zh) * 2010-11-08 2012-05-23 安凯(广州)微电子技术有限公司 一种无索引视频文件快进快退方法、装置以及播放系统
CN102232298A (zh) * 2011-04-07 2011-11-02 华为技术有限公司 媒体内容的传输处理方法、装置与系统
CN102761773A (zh) * 2012-07-03 2012-10-31 Ut斯达康通讯有限公司 一种利用i帧改善iptv业务的用户体验的方法
CN103957471A (zh) * 2014-05-05 2014-07-30 华为技术有限公司 网络视频播放的方法和装置

Also Published As

Publication number Publication date
CN103957471B (zh) 2017-07-14
EP3142381A4 (en) 2017-03-29
EP3142381B1 (en) 2020-01-08
JP6314252B2 (ja) 2018-04-18
JP2017519406A (ja) 2017-07-13
US10476928B2 (en) 2019-11-12
CN103957471A (zh) 2014-07-30
EP3142381A1 (en) 2017-03-15
US20170078354A1 (en) 2017-03-16

Similar Documents

Publication Publication Date Title
WO2015169172A1 (zh) 网络视频播放的方法和装置
EP2649792B1 (en) Pre-buffering audio/video stream pairs
US8355433B2 (en) Encoding video streams for adaptive video streaming
US9646141B2 (en) Fast start of streaming digital media playback with deferred license retrieval
US20120246335A1 (en) Method, terminal, and server for implementing fast playout
CN105916058B (zh) 一种流媒体缓冲播放方法、装置及显示设备
WO2017063399A1 (zh) 一种视频播放方法和装置
US10110507B2 (en) Push-based transmission of resources and correlated network quality estimation
US9324375B1 (en) Dynamically adjusting stream quality level
WO2017071604A1 (zh) 一种播放方法、装置及存储介质
TW201123795A (en) System, method and apparatus for dynamic media file streaming
JP2013038766A (ja) 送信装置、送信装置の制御方法、制御プログラム、及び記録媒体
WO2016112639A1 (zh) 流媒体数据传输方法、客户端和服务器
US20210021655A1 (en) System and method for streaming music on mobile devices
WO2011054319A1 (zh) 一种在httpstreaming系统中实现分层请求内容的方法,装置和系统
WO2013185514A1 (zh) 一种播放流媒体的系统和方法
WO2015192683A1 (zh) 一种基于码流自适应技术的内容分发方法、装置及系统
CN110996160B (zh) 视频处理方法、装置、电子设备及计算机可读取存储介质
US20230050596A1 (en) Playback mode determining method and apparatus
US10200433B2 (en) Client device, a method for receiving a streaming media data and a streaming media data transmission system
WO2018014545A1 (zh) 一种码流数据的处理方法及装置
CN106303754A (zh) 一种音频数据播放方法及装置
JP6294527B2 (ja) 送信装置、送信方法、再生装置、及び再生方法
WO2012071932A1 (zh) 一种分段节目快速启动的方法及系统
US20110196981A1 (en) Combined Binary String for Signaling Byte Range of Media Fragments in Adaptive Streaming

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15789417

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2016566765

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2015789417

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2015789417

Country of ref document: EP