WO2019128667A1 - 视频播放方法以及终端、服务器及存储介质 - Google Patents

视频播放方法以及终端、服务器及存储介质 Download PDF

Info

Publication number
WO2019128667A1
WO2019128667A1 PCT/CN2018/119434 CN2018119434W WO2019128667A1 WO 2019128667 A1 WO2019128667 A1 WO 2019128667A1 CN 2018119434 W CN2018119434 W CN 2018119434W WO 2019128667 A1 WO2019128667 A1 WO 2019128667A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
time period
target video
video frame
target
Prior art date
Application number
PCT/CN2018/119434
Other languages
English (en)
French (fr)
Inventor
王奎
陈颖川
白雅贤
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Priority to EP18897218.6A priority Critical patent/EP3734980A4/en
Publication of WO2019128667A1 publication Critical patent/WO2019128667A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/56Provisioning of proxy services
    • H04L67/568Storing data temporarily at an intermediate stage, e.g. caching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234345Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements the reformatting operation being performed only on part of the stream, e.g. a region of the image or a time segment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/23439Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/24Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth, upstream requests
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/24Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth, upstream requests
    • H04N21/2407Monitoring of transmitted content, e.g. distribution time, number of downloads
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/24Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth, upstream requests
    • H04N21/2408Monitoring of the upstream path of the transmission network, e.g. client requests
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25866Management of end-user data
    • H04N21/25891Management of end-user data being end-user preferences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/262Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
    • H04N21/26258Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists for generating a list of items to be played back in a given order, e.g. playlist, or scheduling item distribution according to such list
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/2662Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4331Caching operations, e.g. of an advertisement for later insertion during playback
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/131Protocols for games, networked simulations or virtual reality

Definitions

  • the present application relates to the field of virtual reality technologies, for example, to a video playing method, and a terminal, a server, and a storage medium.
  • VR technology also known as VR technology or artificial environment
  • VR technology is a computer simulation technology that can create and experience virtual worlds, that is, use computer simulation to generate a virtual environment in three-dimensional space. Provides a user's simulation of the senses of sight, hearing, and touch, making the user feel as if they are there.
  • a VR terminal requires a large storage space and a decoding capability when playing a VR video.
  • the VR terminal first detects the user's visual center area; secondly, downloads the corresponding high-definition video slice from the server, decodes and displays it in the visual center area, and simultaneously downloads from the server.
  • the corresponding low definition video is sliced and decoded and displayed in a visual edge area other than the central area of the vision.
  • the VR terminal detects that the user's visual center area changes the corresponding high-definition video segment needs to be downloaded again, and after being downloaded to a certain extent (such as download completion), the decoding is performed and displayed in the changed visual center area. .
  • the corresponding low-definition video segment needs to be re-downloaded, after being downloaded to a certain extent, decoded and displayed in the changed visual edge region.
  • the VR terminal detects the time after the change of the user's visual center (that is, the time corresponding to downloading to a certain degree of video segmentation from the server), the matching video segmentation can be displayed. This causes a certain delay for the VR terminal to play VR video. That is to say, the related video function of the VR terminal is poor.
  • the present application is directed to providing a video playing method, a terminal, a server, and a storage medium, which can reduce the delay existing when the terminal plays a video, and improve the intelligence of playing the video function of the terminal.
  • the embodiment of the present application provides a video playing method, which is applied to a terminal, including: when playing a target video, receiving index information corresponding to the target video sent by the server; when playing the first video frame of the current time period, Determining, according to the index information, a download address of a second video frame that matches the first time period; wherein the first video frame and the second video frame are video frames in the target video, The first time period is the next time period of the current time period; the second video frame is downloaded at the first time according to the download address; wherein the first time is one of the current time periods a preset time; playing the second video frame at a start time of the first time period.
  • An embodiment of the present application provides a terminal, including: a first processor, a first storage medium, a first communication bus, and a player, where the first processor, the first storage medium, and the player pass through a first communication bus connection; the first processor invoking a video playback related program stored in the first storage medium, and performing the following steps: when playing the target video, receiving, by the receiving server, the target video Index information; when playing the first video frame of the current time period, determining, according to the index information, a download address of the second video frame that matches the first time period; wherein the first video frame and the first The second video frame is a video frame in the target video, the first time period is a next time period of the current time period; and the second video frame is downloaded at a first time according to the download address; The first time is a preset time in the current time period; the player is configured to play the second video frame at a start time of the first time period.
  • An embodiment of the present application provides a server, including: a second processor, and a second storage medium storing the second processor executable instructions, where the first storage medium depends on the second communication bus
  • the second processor performs an operation, when the instruction is executed by the second processor, performing the following steps: acquiring focus description information of the target video; encapsulating the focus description information into index information of the target video; Index information to the terminal.
  • the embodiment of the present application provides a computer storage medium, which is applied to a terminal, where the computer storage medium stores one or more video playback related programs, and the one or more video playback related programs may be one or more A processor executes to implement the video playback method.
  • An embodiment of the present application provides a computer storage medium, which is applied to a server, where the computer storage medium stores one or more video playback related programs, and the one or more video playback related programs may be one or more
  • the second processor executes to implement the video playback method.
  • FIG. 1 is a schematic diagram of a spatial division of a panoramic video sphere according to an embodiment of the present application
  • FIG. 2 is a schematic diagram of a spherical video projection of a panoramic video according to an embodiment of the present application
  • FIG. 3 is a schematic diagram of a panoramic video segmentation according to an embodiment of the present application.
  • FIG. 4 is a schematic structural diagram 1 of a video playing system according to an embodiment of the present application.
  • FIG. 5 is a flowchart 1 of a video playing method according to an embodiment of the present application.
  • FIG. 6 is a schematic diagram of a positional relationship of first, second, and third target video segments according to an embodiment of the present disclosure
  • FIG. 7 is a schematic diagram of a video playing time according to an embodiment of the present application.
  • FIG. 8 is a flowchart 2 of a video broadcast method according to an embodiment of the present application.
  • FIG. 9 is a schematic structural diagram 2 of a video playing system according to an embodiment of the present disclosure.
  • FIG. 10 is a schematic structural diagram 3 of a video playing system according to an embodiment of the present disclosure.
  • FIG. 11 is a schematic structural diagram of a terminal according to an embodiment of the present application.
  • FIG. 12 is a schematic structural diagram of a server according to an embodiment of the present application.
  • the video involved in the application may refer to VR video.
  • the VR video is introduced as follows:
  • the VR video is a 360-degree panoramic video, which is a spherical video space, as shown in FIG.
  • the original spherical video space is projected as a rectangle
  • the longitude of the spherical space of the panoramic video is projected to the horizontal direction of the rectangle (0-360)
  • the dimension is projected to the vertical direction ( ⁇ 90), for example, as shown in FIG. .
  • the panoramic video consists of video captured by 32 slots (the number of specific slots is determined by the actual situation), corresponding to 32 perspectives, each of which corresponds to one video segment.
  • the projected rectangular video area is divided into 32 tiles in a grid shape, for example, as shown in FIG. It should be noted that when the user's visual center falls on a certain area, it indicates that the corresponding focus value of the area is the largest.
  • the terminal in the embodiment of the present application may be an electronic device having a video playing function, such as a mobile phone, a tablet computer, a notebook computer, a palmtop computer, a personal digital assistant (PDA), and a portable media player (Portable Media Player). , PMP), video playback devices, mobile devices such as wearable devices, and fixed terminals such as digital television (TV), desktop computers, and the like.
  • a video playing function such as a mobile phone, a tablet computer, a notebook computer, a palmtop computer, a personal digital assistant (PDA), and a portable media player (Portable Media Player).
  • PMP portable media player
  • PMP video playback devices
  • mobile devices such as wearable devices
  • fixed terminals such as digital television (TV), desktop computers, and the like.
  • An embodiment of the present application provides a video playing system, which is implemented by a video playing system.
  • a video playing system is provided in the embodiment of the present application.
  • the system includes a terminal and a server.
  • the server is configured to generate index information, and send the index information to the terminal, and download the corresponding video frame data when receiving the download request sent by the terminal.
  • the terminal is configured to receive the index information, and according to the index information, when playing the video frame of the current time period, download the video frame that matches the next time period of the current time period from the server in advance. Based on this, the video playing system provided by the present application can improve the intelligence of the terminal video playing function.
  • the embodiment of the present application provides a video playing method, which is applied to a terminal. As shown in FIG. 5, the method may include: step S101, step S102, step S103, and step S104.
  • step S101 when the target video is played, the index information corresponding to the target video transmitted by the server is received.
  • the terminal is a terminal having a video played according to a play instruction. Therefore, the application scenario of the embodiment of the present application is a scenario in which a user needs to play a video when watching a video.
  • the terminal when the terminal plays the target video, the terminal first sends a request instruction to the server, where the request instruction includes related information of the target video, such as a target video identifier.
  • the server receives the request instruction of the terminal, the server returns an index information to the terminal.
  • the target video refers to a video that is to be played by the terminal, and is obtained by receiving a play instruction sent by the user and parsing the play instruction.
  • the parsing play command includes a field for characterizing the video information; and searching for a video in the server that matches the video information according to the field of the video information.
  • the index information used to represent the description information of the slice video corresponding to each frame of the video frame in the target video includes: a time segment identifier, a slice identifier (or a slice perspective), an area coordinate, and a focus value obtained by different acquisition methods.
  • the time period identifier may refer to a time period that lasts when the focus value does not change.
  • the slice ID (or slice video view) may refer to the label of the slice video in the entire frame of video. For example, Tile32 in Figure 3.
  • the area coordinate may refer to a fixed point of a whole frame video as the origin, and the coordinate values of the left, right, upper and lower edges of the slice video are represented by p_top, p_bottom, p_left, and p_right.
  • the focus value may refer to a value that describes how much the user is focused on the video slice. The larger the value, the closer the area corresponding to the video segment is to the visual center of the user.
  • the present application provides two ways to obtain the focus values, namely: mode one, statistical historical data analysis, and mode two, according to preset focus description information.
  • the index information is included in a standard HTTP-based HTTP Live Streaming (HLS) protocol, or a Video Streaming technology that transmits dynamic bitrate on the Internet (Dynamic Adaptive Streaming over HTTP, DASH) ) in the relevant attribute field in the protocol.
  • HLS HTTP-based HTTP Live Streaming
  • DASH Video Streaming technology that transmits dynamic bitrate on the Internet
  • the attribute field corresponding to the index information is the EXT-X-VIEW-INF attribute field in the HLS protocol.
  • the attribute field corresponding to the index information is the ViewInfo attribute field in the DASH protocol.
  • index information described above may be expressed in the form shown in Table 1 below.
  • index information in Table 1 above may be expressed as follows:
  • the sharpness level of the first video frame can be set in advance. For example, at the start of playback, the highest level of sharpness, or the next highest level of sharpness, or the lowest level of sharpness, etc., corresponding to the first video frame is downloaded. In order to ensure the smoothness of the target video playing, in an embodiment, the sharpness level of the first video frame is set to the lowest level of sharpness in advance (ie, the method of playing the first frame of the video frame in the related art).
  • the first video frame address corresponding to the lowest level of sharpness may be determined according to the index information (including each video frame in the target video and the download address in different definitions in the index information), Finally, the first video frame is downloaded according to the address.
  • the terminal when the terminal does not receive the index information sent by the server (corresponding to the case where the server does not generate the index information, for example, the target video has just been released, there is no history data), the terminal follows the related video playing method.
  • the target video is played, that is, the first video frame of the current time period is played until the start time of the first time period, and the second video frame is downloaded.
  • the download speed of the second video frame can be increased at the beginning of the first time period.
  • step S102 when the first video frame of the previous time period is played, the download address of the second video frame that matches the first time period is determined according to the index information.
  • the first video frame and the second video frame are video frames in the target video frame; the first time period is the next time period of the current time period.
  • the second video frame of the first time period needs to be downloaded in advance, so as to reduce the delay of playing the video of the terminal, in order to download the first time period in time.
  • the download address of the second video frame needs to be known first.
  • the index information may be a download address of each slice video in each video frame in the target video, where one download address corresponds to one video segment of a definition.
  • the implementation of step S102 may be: searching for index information of the second video frame corresponding to the first time period from the index information. In the index information according to the second video frame, the download address of each video slice in the second video frame is found.
  • the index information includes: a focus value of each video segment in each video frame in the target video, a resolution corresponding to different focus values, and a video slice download address corresponding to different definitions. .
  • the index information when the index information includes a video slice download address corresponding to different definitions, the index information may be expressed by using the following procedure:
  • step S102 can be implemented by the following S1021-S1023.
  • step S1021 based on the index information, the focus value of each video slice in the second video frame is searched for.
  • the index information may be based on the slice identifier of each video segment in the second video frame. Find the focus value corresponding to the same slice identifier as the slice identifier.
  • the index information may be in the form of a table, or may be in the form of a textual expression or the like. The embodiment of the present application is not limited.
  • step S1022 the sharpness of each video slice in the second video frame is determined based on the focus value.
  • the focus value may be divided into three gear positions, for example, a first gear, a second gear, and a third gear, each gear position corresponding to one focus range. Each gear corresponds to a resolution.
  • the focus value can be matched with the example three gear positions. If the focus value falls within the focus value range corresponding to the first gear, the sharpness corresponding to the focus value is the sharpness corresponding to the first gear.
  • step S1023 based on the sharpness of each video slice in the second video, the download address of each video slice corresponding to the sharpness is searched for.
  • the maximum focus value of each video slice focus value may be first determined, and the sharpness of the video slice corresponding to the maximum focus value is determined to be the highest definition.
  • the resolution of the other video segments is determined according to the positional relationship of the video segments corresponding to the maximum focus values of the other video segments. Based on this, S1022 can be replaced with the following steps A to F.
  • step A a first target video slice corresponding to the maximum focus value is determined from each video slice frame according to the maximum focus value in the index information.
  • the target video frame is composed of a plurality of video tiles, and each video tile corresponds to one focus value. Therefore, the largest focus value can be selected from a plurality of focus values.
  • the video segment corresponding to the maximum focus value is determined as the first target video segment.
  • step B the sharpness of the first target video slice is determined to be the highest level of sharpness.
  • the video segment of the first target video segment corresponding to the maximum focus value has been described above. Therefore, the first target video segment is represented as the video segment where most of the user's visual center is located. In an embodiment, the first The sharpness of the target video slice is determined to be the highest level of sharpness.
  • step C a second target video segment within the de-centered neighborhood of the first target video slice is determined.
  • the second target video segment is a center segment of the first target video segment
  • the preset video segment is a video segment within a radius. It should be noted that the “go to heart” description does not include the first target video segment.
  • step D the sharpness of the second target video slice is determined to be the next highest level of sharpness.
  • step E a third target video slice is determined.
  • step F the sharpness of the third target video slice is determined as the lowest level of sharpness.
  • the second video frame is composed of the first target video segment, the second target video segment, and the third target video segment.
  • the positional relationship of the first target video segment, the second target video segment, and the third target video segment is as shown in FIG. 6.
  • the highest level of sharpness, the next highest level of sharpness, and the lowest level of sharpness described above are merely examples. In practical applications, multiple levels of sharpness can be divided. This application is not limited herein.
  • the resolution of the video segment other than the first target video segment in the second video frame may be directly determined as a lower than the highest level of resolution.
  • the clarity of the preset level may be directly determined as a lower than the highest level of resolution.
  • step S103 a download address is obtained according to S102, and the second video frame is downloaded at the first time.
  • the first time is a preset time of the current time period.
  • the terminal Since the second video frame has been downloaded at the first moment, the terminal has acquired the second video frame data when the first time period arrives. Therefore, the embodiment of the present application can effectively reduce the delay problem existing when playing video of the related terminal.
  • the preset time may be an intermediate time corresponding to the current time period, or a 2/3 time or the like. This application is not limited herein.
  • step S104 the second video frame is played at the beginning of the first time period.
  • D1 is the current time period
  • the current playing time is T1.
  • the second video frame is downloaded.
  • T2 is the first moment.
  • the current time is T3, part of the second video frame has been buffered, so when the T3 time arrives, the second video frame can be directly played.
  • the second video frame played at the beginning of the first time period is the best video frame, that is, the video segment where the user's actual visual center is located, and should be the highest-definition video segmentation in the second video frame. .
  • the second video frame downloaded in advance is not obtained according to the actual focus value of the start time of the first time period, the second video frame downloaded in advance is not necessarily the optimal video frame.
  • step A1 at the start time of the first time period, the fourth target video slice corresponding to the actual maximum focus value is detected.
  • the actual maximum focus value refers to the detected focus value of the user's visual center corresponding to the video segment at the beginning time of the first time period. It should be noted that the detection method belongs to the related art and will not be described again.
  • step B1 if the fourth target video slice does not match the first target video slice, the sharpness of the fourth target video slice is determined as the highest level of sharpness.
  • the second video frame is not the optimal video frame. Therefore, it is necessary to adjust the second video frame to the optimal video frame, that is, it is necessary to determine the definition of the fourth target video slice in the second video frame as the highest level of sharpness.
  • step C1 based on the index information, the download address of the fourth target video slice with the highest level of sharpness is determined.
  • step D1 the fourth target video slice is downloaded.
  • step E1 the fourth target video slice is played.
  • the fourth target video segment When the fourth target video segment is downloaded, the fourth target video segment is played.
  • a fourth target video segment may also be played when the fourth target video segment is downloaded to 50%.
  • the terminal since the second time frame corresponding to the first time period has been downloaded at the first moment.
  • the terminal When the start time of the first time period arrives, the terminal has acquired the second video frame, so the terminal does not need to download the second video frame from the server again, that is, the terminal has no delay problem existing when playing the video. Improve the intelligence of the terminal.
  • the following procedure is a program for playing a partial time period of a target video:
  • the embodiment of the present application provides a video playing method, a terminal, a server, and a storage medium.
  • the server acquires focus description information of the target video, encapsulates the focus description information into index information of the target video, and sends the index information to the terminal.
  • the terminal plays the target video
  • the terminal receives the index information corresponding to the target video sent by the server; when playing the first video frame of the current time period, determines, according to the index information, the second video frame that matches the first time period.
  • the terminal can download the second video frame of the first time period in advance during the current time period.
  • the second video frame can be played directly at the beginning of the first time period, and the second video frame does not need to be downloaded at the beginning of the first time period. Therefore, the technical solution provided by the present application can effectively reduce the delay of the video played by the terminal, and further improve the intelligence of the function of playing the video by the terminal.
  • the present application provides a video playing method, which is applied to a server. As shown in FIG. 8, the method may include: step S201, step S202, and step S203.
  • step S201 focus description information of the target video is acquired.
  • the focus description information described above characterizes focus information for each video slice of each video frame in the target video frame (eg, the user's visual center distance from each video slice distance).
  • the foregoing S201 can be obtained by the following manners 1 and 2.
  • Method 1 Obtain preset focus description information of the target video.
  • the preset focus description information is information that is previously set in focus (eg, sharpness, or focus value, etc.) according to user habits. For example, setting a corresponding definition of each video slice of the second video frame, or setting a focus value corresponding to each video slice of the second video frame, or, for each second video frame The download address corresponding to each video segment is set.
  • the foregoing manner 2 can be implemented by the following steps a and b.
  • step a the number of downloads corresponding to each video segment corresponding to the second time in the first preset time period is counted until all the times in the first preset time period are counted.
  • the second moment is any moment of the first preset time period.
  • the foregoing first preset time period refers to a configurable time period of the target video corresponding duration.
  • the duration of the target video is 100 min
  • the first preset time period is 80 min in the target video.
  • the 80 min can also be adjusted to 70 min according to actual needs.
  • the statistical data is historical data of the target video playing, such as playing data of the first to third days of a movie (including download information of each video segment in each video frame).
  • step b the focus value corresponding to each video slice is determined according to the number of downloads.
  • Each video segment in each video frame corresponds to one download number, and the number of downloads reflects the probability that the user visual center corresponding to each video segment falls on each video segment, when most user vision centers fall in it.
  • the corresponding focus value can be determined according to the number of downloads of each video segment.
  • step b when the above step b is implemented, the following steps a1 to d1 can be implemented.
  • step a1 based on the number of downloads, the final video slice whose download ratio reaches the preset download ratio is determined.
  • Each video segment in each video frame corresponds to one download count, so the download ratio of each video segment in a video can be calculated.
  • the download ratio is greater than the preset download ratio, such as greater than 80%
  • the video segment corresponding to the download ratio is downloaded by 80% of the terminals, and the video segment corresponding to the download ratio is called final.
  • Video fragmentation is the video segment corresponding to the download ratio.
  • step b1 the focus value corresponding to the final video slice is set to the maximum value.
  • step a1 it is known from the above step a1 that the final video segment is downloaded by most terminals, for example, by 80% of the terminals, indicating that the final video segment is the video segment corresponding to most user visual centers, therefore, it is required Set the focus value corresponding to the final video slice to the maximum value.
  • step c1 the remaining video segments except the final video segment in each video segment are arranged according to the corresponding number of downloads, and the sorted remaining video segments are obtained.
  • the number of downloads is sequentially arranged from small to large.
  • the number of downloads corresponding to each video segment may be sequentially arranged according to the order of the video segments. It should be noted that the order of the number of downloads is not limited in this application.
  • step d1 the focus values of the sorted remaining video segments are sequentially set.
  • the focus values of the sorted remaining video segments set in order are also arranged in ascending order.
  • the number of downloads corresponding to each video is sequentially arranged according to the order of the video segments, the size relationship of the adjacent two focus values of the remaining video segments after the arrangement is sequentially set, and each video is sequentially arranged.
  • the size relationship of the corresponding two adjacent download times is the same.
  • step S202 the focus description information is encapsulated into index information of the target video.
  • the focus description information obtained by S201 is encapsulated by using the HLS protocol or the DASH protocol.
  • the focus description information is added to the EXT-X-VIEW-INF attribute field of the HLS protocol.
  • the focus description information is added in the ViewInfo attribute field of the HASH protocol.
  • the focus description information may be encapsulated by other encapsulation forms, which is not limited in this application.
  • step S203 the index information is transmitted to the terminal.
  • the index information can be updated according to the latest statistical data.
  • the specific update method is as follows.
  • the index information generated by the server includes the focus description information of each video segment in each video frame in the target video, and the focus description information can represent the situation of the video segment downloaded by the terminal.
  • the resolution information and the like corresponding to each video slice download That is to say, according to the index information, it is possible to predict the situation when the terminal downloads the video slice.
  • the terminal can predict the downloading of the video segment according to the index information, and download the relevant video segment in advance.
  • the server acquires the focus description information of the target video and encapsulates it into index information of the target video, and sends the index information to the terminal when the terminal plays the target video; the terminal receives The index information sent by the server, and when playing the first video frame of the current time period, determining, according to the index information, a download address of the second video frame that matches the first time period, according to the download address, at the first moment The server downloads the second video frame and plays the second video frame at the beginning of the first time period.
  • the terminal when the terminal plays the target video, the index information corresponding to the target video sent by the server is received; and the terminal determines, according to the index information, the first video frame in the current time segment of the target video. a download address of all the lowest level resolution video segments; downloading the first video frame from the server according to the download address; determining the second video frame in the first time period according to the index information when playing the first video frame
  • the focus value of each video slice; the video slice corresponding to the maximum focus value is determined as the video slice of the highest level definition; the video slice corresponding to the focus value except the maximum focus value is determined as the video of the lowest level definition Fragmentation; determining a download address of the highest level definition video segment and the lowest level definition video segment; downloading the second video frame according to the download address; detecting the actual maximum focus value corresponding to the start time of the first time period Whether the video fragments corresponding to the highest resolution level match the video fragments; if they do not match, at the beginning of the first time period, Actual maximum
  • the embodiment of the present application provides a terminal, where the terminal includes:
  • the first processor 11 calls the video playback related program stored by the first storage medium 12, and performs the following steps: when playing the target video, receiving index information corresponding to the target video sent by the server; Determining, according to the index information, a download address of a second video frame that matches a first time period, where the first video frame and the second video frame are The video frame in the target video, the first time period is a next time period of the current time period; and the second video frame is downloaded at a first time according to the download address; wherein the first time The moment is a preset time in the current time period;
  • the player 14 is configured to play the second video frame at a start time of the first time period.
  • the first processor 10 is configured to search for a focus value of each video segment in the second video frame according to the index information; and determine the first according to the focus value. The resolution of each video slice in the two video frames; according to the definition, the download address of each video slice corresponding to the definition is searched for.
  • the first processor 10 is configured to determine, according to a maximum focus value of the focus values of each of the video segments in the second video frame, a first corresponding to the maximum focus value.
  • a target video slice determining a resolution of the first target video slice as a highest level of resolution; determining a second target video slice in a de-centered neighborhood of the first target video slice; The sharpness of the second target video slice is determined to be the next highest level of sharpness; the third target video slice is determined; the sharpness of the third target video slice is determined to be the lowest level of sharpness.
  • the second video frame is composed of the first target video segment, the second target video segment, and the third target video segment.
  • the first processor 10 is further configured to: at a start time of the first time period, detect a fourth target video segment corresponding to an actual maximum focus value; if the fourth target video segment If the slice does not match the first target video segment, determining the resolution of the fourth target video segment as the highest level of resolution; determining, according to the index information, the clarity with the highest level a download address of the fourth target video slice corresponding to the degree; downloading the fourth target video slice.
  • the player 14 is further configured to play the fourth target video segment.
  • the embodiment of the present application provides a server, where the server includes: a second processor 21, and a second storage medium 22 storing executable instructions of the second processor 21, the second The storage medium 22 is operated by the second processor 21 via the second communication bus 23, and when the instruction is executed by the second processor 22, the following steps are performed: acquiring focus description information of the target video; The focus description information is encapsulated into index information of the target video; the index information is sent to the terminal.
  • the second processor 20 is configured to acquire preset focus description information of the target video.
  • the second processor 20 is configured to count a focus value corresponding to the target video in a first preset time period; and use the focus value as the focus description information.
  • the second processor 20 is configured to count, in the first preset time period, the number of downloads of each of the video segments corresponding to the second time corresponding to the highest level of resolution, Up to all the moments in the first preset time period; wherein the second time is any one of the first preset time periods; determining each video score according to the number of downloads The corresponding focus value of the slice.
  • the second processor 20 is configured to determine, according to the number of downloads, a final video segment whose download ratio reaches a preset download ratio; and set a focus value corresponding to the final video segment to Maximum value; for each video segment, the remaining video segments except the final video segment are arranged according to the corresponding number of downloads, and the remaining video segments after sorting are obtained; the remaining videos after sorting are sequentially set.
  • the focus value of the slice is configured to determine, according to the number of downloads, a final video segment whose download ratio reaches a preset download ratio; and set a focus value corresponding to the final video segment to Maximum value; for each video segment, the remaining video segments except the final video segment are arranged according to the corresponding number of downloads, and the remaining video segments after sorting are obtained; the remaining videos after sorting are sequentially set. The focus value of the slice.
  • the above memory may be a volatile memory, such as a random access memory (RAM), or a non-volatile memory, such as a read only memory. (Read-Only Memory, ROM), flash memory, Hard Disk Drive (HDD) or Solid-State Drive (SSD); or a combination of the above types of memory, and provided to the processor Instructions and data.
  • RAM random access memory
  • non-volatile memory such as a read only memory.
  • ROM Read-Only Memory
  • flash memory such as Hard Disk Drive (HDD) or Solid-State Drive (SSD); or a combination of the above types of memory, and provided to the processor Instructions and data.
  • HDD Hard Disk Drive
  • SSD Solid-State Drive
  • the processor may be an Application Specific Integrated Circuit (ASIC), a Digital Signal Processor (DSP), a Digital Signal Processing Device (DSPD), or a Programmable Logic Device (Programmable Logic). Device, PLD), at least one of a Field Programmable Gate Array (FPGA), a Central Processing Unit (CPU), a controller, a microcontroller, and a microprocessor.
  • ASIC Application Specific Integrated Circuit
  • DSP Digital Signal Processor
  • DSPD Digital Signal Processing Device
  • PLD Programmable Logic Device
  • FPGA Field Programmable Gate Array
  • CPU Central Processing Unit
  • controller a controller
  • microcontroller a microcontroller
  • the embodiment of the present application provides a computer storage medium, which is applied to a terminal, where the computer storage medium stores one or more video playback related programs, and the one or more video playback related programs may be one or more A processor executes to implement the video playing method applied to the virtual reality in the terminal in the first embodiment.
  • An embodiment of the present application provides a computer storage medium, which is applied to a server, where the computer storage medium stores one or more video playback related programs, and the one or more video playback related programs may be one or more
  • the second processor executes to implement an embodiment of a video playback method applied to a virtual reality server.
  • embodiments of the present application can be provided as a method, or a computer program product. Accordingly, the application can take the form of a hardware embodiment, a software embodiment, or an embodiment in combination with software and hardware. Moreover, the application can take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage and optical storage, etc.) including computer usable program code.
  • the computer program instructions can also be stored in a computer readable memory that can direct a computer or other programmable data processing device to operate in a particular manner, such that the instructions stored in the computer readable memory produce an article of manufacture comprising the instruction device.
  • the apparatus implements the functions specified in one or more blocks of a flow or a flow and/or block diagram of the flowchart.
  • These computer program instructions can also be loaded onto a computer or other programmable data processing device such that a series of operational steps are performed on a computer or other programmable device to produce computer-implemented processing for execution on a computer or other programmable device.
  • the instructions provide steps for implementing the functions specified in one or more of the flow or in a block or blocks of a flow diagram.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Graphics (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

公开了一种视频播放方法,应用于终端时,包括:在播放目标视频时,接收服务器发送的与目标视频对应的索引信息;当播放当前时间段的第一视频帧时,根据索引信息,确定与第一时间段相匹配的第二视频帧的下载地址;根据下载地址,在第一时刻下载第二视频帧。应用于服务器时,包括:获取目标视频的焦点描述信息;将焦点描述信息,封装为目标视频的索引信息;发送索引信息至终端。

Description

视频播放方法以及终端、服务器及存储介质
本申请要求在2017年12月29日提交中国专利局、申请号为201711482043.8的中国专利申请的优先权,该申请的全部内容通过引用结合在本申请中。
技术领域
本申请涉及虚拟现实技术领域,例如涉及一种视频播放方法以及终端、服务器及存储介质。
背景技术
虚拟现实(Virtual Reality)技术,简称VR技术,也称之为灵境技术或人工环境,是一种可以创建和体验虚拟世界的计算机仿真技术,也就是说利用电脑模拟产生一个三维空间的虚拟环境,提供使用者关于视觉、听觉、触觉的感官的模拟,让使用者如同身临其境一般。
在相关技术中,VR终端在播放VR视频时,需要巨大的存储空间以及解码能力。为了降低VR终端的存储空间以及解码能力,VR终端首先检测使用者的视觉中心区域;其次,从服务器下载对应的高清晰度视频分片,并解码且在视觉中心区域显示,同时,从服务器下载对应的低清晰度视频分片,并解码且在除视觉中心区域外的视觉边缘区域显示。而在VR终端检测到使用者的视觉中心区域发生变化时,需要重新下载对应的高清晰度视频分片,在下载到一定程度(如下载完成)后,解码且在变化后的视觉中心区域显示。同时,需要重新下载对应的低清晰度视频分片,在下载到一定程度后,解码且在变化后的视觉边缘区域显示。
但是,由于VR终端在检测到使用者的视觉中心发生变化后的一段时间(也就是从服务器下载到一定程度的视频分片对应的时间)之后,才可显示出于相匹配的视频分片,这样会导致VR终端播放VR视频时,存在一定的时延,也就是说,相关的VR终端的播放视频功能的智能性差。
发明内容
为解决上述技术问题,本申请期望提供一种视频播放方法以及终端、服务器及存储介质,能够减少终端播放视频时存在的时延,提高终端的播放视频功能的智能性。
本申请实施例提供了一种视频播放方法,应用于终端,包括:在播放目标视频时,接收服务器发送的与所述目标视频对应的索引信息;当播放当前时间段的第一视频帧时,根据所述索引信息,确定与第一时间段相匹配的第二视频帧的下载地址;其中,所述第一视频帧与所述第二视频帧为所述目标视频中的视频帧,所述第一时间段为所述当前时间段的下一时间段;根据所述下载地址,在第一时刻下载所述第二视频帧;其中,所述第一时刻为所述当前时间段内的一预设时刻;在所述第一时间段起始时刻播放所述第二视频帧。
本申请实施例提供了一种终端,包括:第一处理器、第一存储介质、第一通信总线以及播放器,所述第一处理器、所述第一存储介质以及所述播放器通过所述第一通信总线连接;所述第一处理器,调用所述第一存储介质存储的视频播放相关程序,并执行如下步骤:在播放目标视频时,接收服务器发送的与所述目标视频对应的索引信息;当播放当前时间段的第一视频帧时,根据所述索引信息,确定与第一时间段相匹配的第二视频帧的下载地址;其中,所述第一视频帧与所述第二视频帧为所述目标视频中的视频帧,所述第一时间段为所述当前时间段的下一时间段;根据所述下载地址,在第一时刻下载所述第二视频帧;其中,所述第一时刻为所述当前时间段内的一预设时刻;所述播放器,设置为在所述第一时间段起始时刻播放所述第二视频帧。
本申请实施例提供了一种服务器,包括:第二处理器,以及存储有所述第二处理器可执行指令的第二存储介质,所述第一存储介质通过第二通信总线依赖所述第二处理器执行操作,当所述指令被所述第二处理器执行时,执行如下步骤:获取目标视频的焦点描述信息;将所述焦点描述信息,封装为目标视频的索引信息;发送所述索引信息至终端。
本申请实施例提供了一种计算机存储介质,应用于终端中,所述计算机存储介质存储有一个或多个视频播放相关程序,所述一个或者多个视频播放相关程序可被一个或者多个第一处理器执行,以实现所述视频播放方法。
本申请实施例提供了一种计算机存储介质,应用于服务器中,所述计算机存储介质存储有一个或多个视频播放相关程序,所述一个或者多个视频播放相关程序可被一个或者多个第二处理器执行,以实现所述视频播放方法。
附图概述
图1为本申请实施例提供的一种全景视频球面空间划分示意图;
图2为本申请实施例提供的一种全景视频球面投影示意图;
图3为本申请实施例提供的一种全景视频分片示意图;
图4为本申请实施例提供的一种视频播放系统结构示意图一;
图5为本申请实施例提供的一种视频播放方法的流程图一;
图6为本申请实施例提供的一种第一、二以及三目标视频分片的位置关系示意图;
图7为本申请实施例提供的一种视频播放时刻示意图;
图8为本申请实施例提供的一种视频播方法流程图二;
图9为本申请实施例提供的一种视频播放系统结构示意图二;
图10为本申请实施例提供的一种视频播放系统结构示意图三;
图11为本申请实施例提供的一种终端的结构示意图;
图12为本申请实施例提供的一种服务器的结构示意图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述。
需要说明的是,申请中涉及到的视频可以指的是VR视频。如下对VR视频进行介绍:VR视频是360度全景视频,该全景视频为一个球形视频空间,如图1所示。在这里为了描述方便把原始的球形视频空间投影为一个矩形,全景视频球形空间的经度投影至矩形水平方向(0-360),维度投影至垂直方向(±90),例如,如图2所示。
假设全景视频由32个机位拍摄的视频组成(具体机位数量由实际情况决定),对应32个视角,每个视角对应于一个视频分片。为了在不同区域下传输不同清晰度的视频分片,将投影后的矩形视频区域按网格状划分为32片(tile),例如,如图3所示。需要说明的是,当用户的视觉中心落在某个区域时,说明该区域对应的聚焦值最大。
本申请实施例中的终端可以为具有视频播放功能的电子设备等,例如手机、平板电脑、笔记本电脑、掌上电脑、个人数字助理(Personal Digital Assistant,PDA)、便捷式媒体播放器(Portable Media Player,PMP)、视频播放装置、可穿戴设备等移动终端,以及诸如数字电视(Television,TV)、台式计算机等固定终端。
实施例一
本申请实施例提供了一种视频播放方法是基于视频播放系统实现的,本申请实施例提供的一种视频播放系统,如图4所示,该系统包括终端与服务器。其中,服务器设置为生成索引信息,并将该索引信息发送至终端,并在接收到终端发送的下载请求时,下载对应的视频帧数据。终端设置为接收索引信息,并根据索引信息,在播放当前时间段的视频帧时,提前从服务器下载与当前时间段的下一时间段相匹配的视频帧。基于此,本申请提供的视频播放系统可提高终端视频播放功能的智能性。
本申请实施例提供了一种视频播放方法,应用于终端中,如图5所示,该方法可以包括:步骤S101、步骤S102、步骤S103以及步骤S104。
在步骤S101中,在播放目标视频时,接收服务器发送的与目标视频对应的索引信息。
这里,在本申请实施例中,终端是具有根据播放指令播放视频的终端。因此,本申请实施例的应用场景为用户观看视频时,需要播放视频的场景下。
在一种实现方式中,在终端播放目标视频时,终端首先向服务器发送一个请求指令,该请求指令中包含目标视频的相关信息,例如目标视频标识等。当服务器接收到终端的请求指令时,服务器向终端返回一个索引信息。
在一实施例中,上述的目标视频指的是终端即将播放的视频,通过接收用户发送的播放指令,并解析该播放指令得到。在一实施例中,解析播放指令中包括用于表征视频信息的字段;根据视频信息的字段,查找服务器中与该视频信息相匹配的视频。
在一实施例中,上述的索引信息用于表征目标视频中,每一帧视频帧对应的分片视频的描述信息。该索引信息包括:时间段标识、分片标识(或分片视角)、区域坐标,以及不同获取方式获取到的聚焦值。
其中,时间段标识可以指的是聚焦值不发生变化时所持续的时间段。
分片标识(或分片视频视角)可以指的是该分片视频在整帧视频中的标号。例如,图3中的Tile32。
区域坐标可以指的是以一个整帧视频的一固定点为原点,该分片视频的左、右、上、下边缘的坐标值,用p_top、p_bottom,、p_left、以及p_right表示。
聚焦值可以指的是描述用户对该视频分片的聚焦程度的值。该值越大,则表示该视频分片对应的区域越接近用户的视觉中心。本申请提供两种聚焦值的 获取方式,分别为:方式一,统计历史数据分析得到,以及方式二,根据预设焦点描述信息获取。
在一实施例中,该索引信息包含在标准的基于HTTP的流媒体网络传输协议(HTTP Live Streaming,HLS)协议,或在互联网上传送动态码率的Video Streaming技术(Dynamic Adaptive Streaming over HTTP,DASH)协议中的相关属性字段中。在一实施例中,当包含在HLS协议中时,索引信息对应的属性字段为HLS协议中的EXT-X-VIEW-INF属性字段。当包含在DASH协议中时,索引信息添对应的属性字段为DASH协议中的ViewInfo属性字段。
在一实施例中,基于上述内容,上述的索引信息可采用如下表1所示的形式进行表示。
表1
Figure PCTCN2018119434-appb-000001
在一实施例中,上述表1中的索引信息可表示为如下程序:
Figure PCTCN2018119434-appb-000002
Figure PCTCN2018119434-appb-000003
在一实施例中,在目标视频播放的起始时刻,播放目标视频的第一视频帧时,由于该第一视频帧并不存在可提前下载的可能性。因此,可预先设置第一视频帧的清晰度等级。例如,在播放起始时刻,下载最高等级的清晰度,或次高等级的清晰度,或最低等级的清晰度等对应的第一视频帧。为了保证目标视频播放的流畅度,在一实施例中,预先将第一视频帧的清晰度等级设置为最低等级的清晰度(即相关技术中,播放第一帧视频帧的方法)。在目标视频播放的起始时刻,可根据索引信息(索引信息中包括目标视频中每一个视频帧,不同清晰度下的下载地址),确定出最低等级的清晰度对应的第一视频帧地址,最后根据该地址下载第一视频帧。
在一实施例中,当终端未接收到服务器发送的索引信息时(对应于服务器没有生成索引信息的情况,例如,目标视频刚上映,不存在历史数据的情况),终端按照相关的视频播放方法播放目标视频,即在播放当前时间段的第一视频帧,直至第一时间段的起始时刻,才下载第二视频帧。为了解决该种情况下终 端存在时延的问题,可在第一时间段的起始时刻,将第二视频帧的下载速度提高。
在步骤S102中,当播放前时间段的第一视频帧时,根据索引信息,确定与第一时间段相匹配的第二视频帧的下载地址。
其中,第一视频帧与第二视频帧为目标视频帧中的视频帧;第一时间段为当前时间段的下一时间段。
在当前时间段时,终端播放第一视频帧的同时,需要提前下载第一时间段的第二视频帧,这样才能实现减少终端播放视频延时的问题,为了能够及时的下载第一时间段的第二视频帧,需首先获知第二视频帧的下载地址。
在一种示例中,上述的索引信息可以为目标视频中,每一视频帧中的每个分片视频的下载地址,其中,一个下载地址对应一个清晰度的一个视频分片。基于此,步骤S102的实现方式可以是:从索引信息中,查找出第一时间段对应的第二视频帧的索引信息。在根据第二视频帧的索引信息中,查找到第二视频帧中的每个视频分片的下载地址。
在另一种示例中,上述的索引信息包括:目标视频中每一个视频帧中的每个视频分片的聚焦值、不同聚焦值对应的清晰度、以及不同清晰度对应的视频分片下载地址。
在一实施例中,当索引信息包括不同清晰度对应的视频分片下载地址时,索引信息可用如下程序进行表示:
Figure PCTCN2018119434-appb-000004
基于上述内容,上述步骤S102可通过如下S1021-S1023实现。
在步骤S1021中,根据索引信息,查找第二视频帧中的每个视频分片的聚焦值。
由于上述内容已经说明索信息中包含目标视频中每一个视频帧中的每个视频分片的聚焦值,因此,可根据第二视频帧中的每个视频分片的分片标识,在索引信息中查找与该分片标识相同的分片标识对应的聚焦值。这里,索引信息可采用表格的形式,也可采用文字表述的形式等。本申请实施例不作限定。
在步骤S1022中,根据聚焦值,确定第二视频帧中的每个视频分片的清晰度。
在一实施例中,可将聚焦值划分为三个档位,例如,第一档、第二档、以及第三档,每一个档位对应一个聚焦范围。每个档位对应一个清晰度。在执行S1012时,可将聚焦值与示例三个档位进行匹配。若聚焦值落在第一档对应的聚焦值范围里,则该聚焦值对应的清晰度为第一档对应的清晰度。
在步骤S1023中,根据第二视频中的每个视频分片的清晰度,查找与该清晰度对应的每个视频分片的下载地址。
在一种实施例中,可首先确定出每个视频分片聚焦值中的最大聚焦值,将最大聚焦值对应的视频分片的清晰度确定为最高清晰度。再根据其他视频分片与最大聚焦值对应的视频分片的位置关系,确定其他视频分片的清晰度。基于此,S1022可被替换为如下步骤A至步骤F。
在步骤A中,根据索引信息中最大聚焦值,从每个视频分片帧中确定出与最大聚焦值对应的第一目标视频分片。
目标视频帧由多个视频分片组成,而每个视频分片对应一个聚焦值。因此,可从多个聚焦值中选取最大的聚焦值。在一实施例中,从索引信息中,确定出最大聚焦值对应的视频分片,作为第一目标视频分片。
在步骤B中,将第一目标视频分片的清晰度确定为最高等级的清晰度。
上述已经说明第一目标视频分片为最大聚焦值对应的视频分片,因此,表征第一目标视频分片为大部分用户视觉中心所在的视频分片,在一实施例中,需要将第一目标视频分片的清晰度确定为最高等级的清晰度。
在步骤C中,确定出第一目标视频分片的去心邻域内的第二目标视频分片。
在一实施例中,上述的第二目标视频分片为,以第一目标视频分片为圆心,预设个视频分片为半径内的视频分片。需要说明的是,“去心”说明不含第一目标视频分片。
在步骤D中,将第二目标视频分片的清晰度确定为次高等级的清晰度。
在步骤E中,确定第三目标视频分片。
在步骤F中,将第三目标视频分片的清晰度确定为最低等级的清晰度。
其中,第二视频帧由所述第一目标视频分片、第二目标视频分片以及第三目标视频分片组成。上述的第一目标视频分片、第二目标视频分片以及第三目标视频分片的位置关系如图6所示。
在一实施例中,上述的最高等级的清晰度、次高等级的清晰度以及最低等级的清晰度仅仅是一种示例。在实际应用中,可划分多个等级的清晰度。本申请在此并不做限定。
在一种实例中,在执行完上述步骤A后,也可直接将第二视频帧中,除第一目标视频分片之外的视频分片的清晰度确定为一个低于最高等级的清晰度的预设等级的清晰度。
在步骤S103中,根据S102得到下载地址,在第一时刻下载第二视频帧。
其中,第一时刻为当前时间段的一预设时刻。
由于在第一时刻便已经下载了第二视频帧,因此,在第一时间段到达时,终端已经获取到第二视频帧数据。因此,本申请实施例可有效降低相关终端的播放视频时存在的时延问题。
在一实施例中,上述的预设时刻可以为当前时间段对应的中间时刻,或者2/3时刻等。本申请在此并不做限定。
在步骤S104中,在第一时间段的起始时刻播放第二视频帧。
如图7所示,若D1为当前时间段,且当前播放时刻为T1。当当前播放时刻为T2时,下载第二视频帧。其中,T2为第一时刻。当当前时刻为T3时,已经缓存了部分第二视频帧,因此,当T3时刻到达时,可直接播放第二视频帧。
理论上,在第一时间段的起始时刻播放的第二视频帧为最佳视频帧,即用户实际视觉中心所在的视频分片,应该为第二视频帧中最高清晰度的视频分分片。而实际上,由于提前下载的第二视频帧并不是根据第一时间段的起始时刻实际的聚焦值得到,因此,提前下载的第二视频帧并非一定是最佳视频帧。而为了保证在第一时间段的起始时刻播放的第二视频帧为最佳视频帧,需要在第一时间段的起始时刻检测实际最大聚焦值对应的视频分片。
基于上述内容,S104之前,还包括如下步骤A1-E1。
在步骤A1中,在第一时间段的起始时刻,检测实际最大聚焦值对应的第四 目标视频分片。
在一实施例中,上述实际最大聚焦值指的是在第一时间段的起始时刻,检测到的用户视觉中心对应视频分片的聚焦值。需要说明的是,该检测方法属于相关技术,再次不在赘述。
在步骤B1中,若第四目标视频分片与第一目标视频分片不匹配,则将第四目标视频分片的清晰度确定为最高等级的清晰度。
当第四目标视频分片与第一目标视频分片不匹配时,第二视频帧并非最佳视频帧。因此,需要将第二视频帧调整为最佳视频帧,也就是说,需要将确定第二视频帧中第四目标视频分片的清晰度确定为最高等级的清晰度。
在步骤C1中,根据索引信息,确定与最高等级的清晰度的第四目标视频分片的下载地址。
在步骤D1中,下载第四目标视频分片。
在步骤E1中,播放第四目标视频分片。
在第四目标视频分片下载完时,播放第四目标视频分片。当然,也可在第四目标视频分片下载到百分之五十时,播放部分第四目标视频分片。
可以理解的是,由于在第一时刻已经开始下载第一时间段对应的第二视频帧。当第一时间段的起始时刻到达时,终端已经获取到第二视频帧,所以终端无需再从服务器下载第二视频帧,也就是说,终端已经不存在播放视频时存在的时延问题,提高了终端的智能性。
在一实施例中,如下程序为一个目标视频的部分时间段播放时的程序:
Figure PCTCN2018119434-appb-000005
Figure PCTCN2018119434-appb-000006
Figure PCTCN2018119434-appb-000007
Figure PCTCN2018119434-appb-000008
Figure PCTCN2018119434-appb-000009
Figure PCTCN2018119434-appb-000010
Figure PCTCN2018119434-appb-000011
本申请实施例提供了一种视频播放方法、终端、服务器及存储介质,服务器获取目标视频的焦点描述信息;将焦点描述信息,封装为目标视频的索引信息;发送索引信息至终端。终端在在播放目标视频时,接收服务器发送的与目标视频对应的索引信息;当播放当前时间段的第一视频帧时,根据索引信息,确定与第一时间段相匹配的第二视频帧的下载地址;其中,第一视频帧与第二视频帧为目标视频中的视频帧,第一时间段为当前时间段的下一时间段;根据下载地址,在第一时刻下载第二视频帧;其中,第一时刻为当前时间段内的一预设时刻;在第一时间段起始时刻播放第二视频帧。采用上述技术方案,终端可在当前时间段时,提前下载第一时间段的第二视频帧。在第一时间段的起始时刻就可直接播放第二视频帧,并不需要在第一时间段的起始时刻才下载第二视频帧。因此,本申请提供的技术方案可有效的减少终端播放视频存在的时延,进一步提高终端播放视频功能的智能性。
本申请实施提供了一种视频播放方法,应用于服务器,如图8所示,该方法可以包括:步骤S201、步骤S202和步骤S203。
在步骤S201中,获取目标视频的焦点描述信息。
在一实施例中,上述的焦点描述信息表征目标视频帧中的每个视频帧的每个视频分片的聚焦信息(例如,用户的视觉中心距离每个视频分片距离)。
在一实施例中,上述S201可通过如下方式一和方式二获取。
方式一、获取目标视频的预设焦点描述信息。
在一实施例中,上述的预设焦点描述信息为根据用户习惯,预先设置好的与焦点相关(例如清晰度,或聚焦值等)的信息。例如,将第二视频帧的每个视频分片的对应的清晰度进行设置,或者,将第二视频帧的每个视频分片对应 的焦点值进行设置,或者,将第二视频帧的每个视频分片对应的下载地址进行设置。
方式二、统计目标视频在第一预设时间段对应的聚焦值,将聚焦值作为焦点描述信息。
在一实施例中,上述方式二可通过如下步骤a和步骤b实现。
在步骤a中,统计第一预设时间段中,第二时刻对应的每个视频分片对应的下载次数,直至统计完第一预设时间段中的所有时刻。
其中,第二时刻为第一预设时间段的任一时刻。
在一实施例中,上述的第一预设时间段指的是,目标视频对应时长中的一可配置时间段。例如,目标视频的时长为100min,第一预设时间段为目标视频中的80min。当然,也可根据实际需求,将该80min调整为70min。
在一实施例中,上述的统计数据为目标视频播放的历史数据,如某电影第一至三天的播放数据(包括每个视频帧中,每个视频分片的下载信息)。
在步骤b中,根据下载次数,确定每个视频分片对应的聚焦值。
每一视频帧中的每个视频分片对应一个下载次数,下载次数可反映出每个视频分片对应的用户视觉中心落在每个视频分片上的概率,当大多数用户视觉中心落在其中一个视频分片上时,说明该视频分片对应的聚焦值越大。因此,可根据每个视频分片的下载次数,确定出其对应的聚焦值。
在一实施例中,在实现上述步骤b时,可通过如下步骤a1-步骤d1实现。
在步骤a1中,根据下载次数,确定出下载比率达到预设下载比率的最终视频分片。
每一个视频帧中的每个视频分片对应一个下载次数,因此,可以计算出一个视频中每个视频分片的下载比率。当下载比率大于预设下载比率时,如大于百分之八十时,说明该下载比率对应的视频分片被百分之八十的终端下载,将该下载比率对应的视频分片称为最终视频分片。
在步骤b1中,将最终视频分片对应的聚焦值设置为最大值。
从上述步骤a1中得知,最终视频分片被大部分终端下载,例如,被百分之八十的终端下载,说明最终视频分片为大多数用户视觉中心对应的视频分片,因此,需要将最终视频分片对应的聚焦值设置为最大值。
在步骤c1中,对每个视频分片中除最终视频分片外的剩余视频分片,按照各自对应的下载次数进行排列,得到排序后的剩余视频分片。
将下载次数依次从小至大排列,当然也可以根据视频分片的排列顺序,依次排列出每个视频分片对应的下载次数。需要说明的是,本申请对下载次数的排列顺序并不做限定。
在步骤d1中,依序设置排序后的剩余视频分片的聚焦值。
当下载次数依次从小至大排列时,依次设置的排序后的剩余视频分片的聚焦值也是按照从小到大的顺序排列的。当根据视频分片的排列顺序,依次排列出每个视频对应的下载次数时,依次设置的排列后的剩余视频分片的相邻两个聚焦值的大小关系,分别与依次排列出每个视频对应的相邻两个下载次数的大小关系相同。
在步骤S202中,将焦点描述信息,封装为目标视频的索引信息。
将S201得到的焦点描述信息,采用HLS协议,或DASH协议进行封装。在一实施例中,将焦点描述信息添加在HLS协议的EXT-X-VIEW-INF属性字段中。或者,将焦点描述信息添加在HASH协议的ViewInfo属性字段中。当然,也可以采用其他的封装形式对焦点描述信息进行封装,本申请对此并不作限定。
在步骤S203中,发送索引信息至终端。
在一实施例中,通过上述索引信息可根据最新的统计数据进行更新。具体更新方式如上述方式二。
可以理解的是,服务器生成的索引信息中,包含了目标视频中,每个视频帧中的每个视频分片的焦点描述信息,而焦点描述信息可表征终端下载的视频分片的情况。例如,每个视频分片下载时对应的清晰度信息等。也就是说,根据索引信息可以预测出终端在下载视频分片时的情况。
这样,终端便可根据索引信息,预测视频分片的下载情况,并提前下载相关视频分片。
在一种实施例中,如图9所示,服务器获取目标视频的焦点描述信息,并将其封装为目标视频的索引信息,在终端播放目标视频时,将索引信息发送至终端;终端接收到服务器发送的索引信息,并在播放当前时间段的第一视频帧时,根据索引信息,确定与第一时间段相匹配的第二视频帧的下载地址,根据该下载地址,在第一时刻从服务器下载第二视频帧,并在第一时间段起始时刻播放第二视频帧。
在一实施例中,如图10所示,在终端播放目标视频时,接收服务器发送的与目标视频对应的索引信息;终端根据索引信息,确定出目标视频当前时间段 的第一视频帧中的所有的最低等级清晰度的视频分片的下载地址;根据下载地址,从服务器下载第一视频帧;在播放第一视频帧时,根据索引信息确定出第一时间段的第二视频帧中的每个视频分片的聚焦值;将最大聚焦值对应的视频分片确定为最高等级清晰度的视频分片;将最大聚焦值除外的聚焦值对应的视频分片确定为最低等级清晰度的视频分片;确定出最高等级清晰度视频分片与最低等级清晰度视频分片的下载地址;根据下载地址,下载第二视频帧;在第一时间段的起始时刻检测实际最大聚焦值对应的视频分片与最高清晰度等级对应的视频分片是否匹配;若不匹配,在第一时间段的起始时刻,下载实际最大聚焦值对应的视频分片;播放实际最大聚焦值对应的视频分片。
实施例二
如图11所示,本申请实施例提供了一种终端,该终端包括:
第一处理器11、第一存储介质12、第一通信总线13以及播放器14,所述第一处理器11、所述第一存储介质12以及所述播放器14通过所述第一通信总线13连接;
所述第一处理器11,调用所述第一存储介质12存储的视频播放相关程序,并执行如下步骤:在播放目标视频时,接收服务器发送的与所述目标视频对应的索引信息;当播放当前时间段的第一视频帧时,根据所述索引信息,确定与第一时间段相匹配的第二视频帧的下载地址;其中,所述第一视频帧与所述第二视频帧为所述目标视频中的视频帧,所述第一时间段为所述当前时间段的下一时间段;根据所述下载地址,在第一时刻下载所述第二视频帧;其中,所述第一时刻为所述当前时间段内的一预设时刻;
所述播放器14,设置为在所述第一时间段起始时刻播放所述第二视频帧。
在一实施例中,所述第一处理器10,设置为根据所述索引信息,查找所述第二视频帧中的每个视频分片的聚焦值;根据所述聚焦值,确定所述第二视频帧中的每个视频分片的清晰度;根据所述清晰度,查找与所述清晰度对应的每个视频分片的下载地址。
在一实施例中,所述第一处理器10,设置为根据所述第二视频帧中的每个视频分片的聚焦值中的最大聚焦值,确定出与所述最大聚焦值对应的第一目标视频分片;将所述第一目标视频分片的清晰度确定为最高等级的清晰度;确定出所述第一目标视频分片的去心邻域内的第二目标视频分片;将所述第二目标视频分片的清晰度确定为次高等级的清晰度;确定出第三目标视频分片;将所 述第三目标视频分片的清晰度确定为最低等级的清晰度。
其中,所述第二视频帧由所述第一目标视频分片、所述第二目标视频分片以及所述第三目标视频分片组成。
在一实施例中,所述第一处理器10,还设置为在所述第一时间段起始时刻,检测实际最大聚焦值对应的第四目标视频分片;若所述第四目标视频分片与所述第一目标视频分片不匹配,则将所述第四目标视频分片的清晰度确定为所述最高等级的清晰度;根据所述索引信息,确定与所述最高等级的清晰度对应的所述第四目标视频分片的下载地址;下载所述第四目标视频分片。所述播放器14,还设置为播放所述第四目标视频分片。
如图12所示,本申请实施例提供了一种服务器,该服务器包括:第二处理器21,以及存储有所述第二处理器21可执行指令的第二存储介质22,所述第二存储介质22通过第二通信总线23依赖所述第二处理器21执行操作,当所述指令被所述第二处理器22执行时,执行如下步骤:获取目标视频的焦点描述信息;将所述焦点描述信息,封装为目标视频的索引信息;发送所述索引信息至终端。
在一实施例中,所述第二处理器20,设置为获取所述目标视频的预设焦点描述信息。
在一实施例中,所述第二处理器20,设置为统计所述目标视频在第一预设时间段对应的聚焦值;将所述聚焦值作为所述焦点描述信息。
在一实施例中,所述第二处理器20,设置为统计所述第一预设时间段中,第二时刻对应的每个所述视频分片对应所述最高等级清晰度的下载次数,直至统计完所述第一预设时间段中的所有时刻;其中,所述第二时刻为所述第一预设时间段的任一时刻;根据所述下载次数,确定所述每个视频分片对应的聚焦值。
在一实施例中,所述第二处理器20,设置为根据所述下载次数,确定出下载比率达到预设下载比率的最终视频分片;将所述最终视频分片对应的聚焦值设置为最大值;对每个视频分片中除所述最终视频分片外的剩余视频分片,按照各自对应的下载次数进行排列,得到排序后的剩余视频分片;依序设置排序后的剩余视频分片的聚焦值。
在实际应用中,上述的存储器可以是易失性存储器(volatile memory),例如随机存取存储器(Random-Access Memory,RAM);或者非易失性存储器 (non-volatile memory),例如只读存储器(Read-Only Memory,ROM),快闪存储器(flash memory),硬盘(Hard Disk Drive,HDD)或固态硬盘(Solid-State Drive,SSD);或者上述种类的存储器的组合,并向处理器提供指令和数据。
上述处理器可以为特定用途集成电路(Application Specific Integrated Circuit,ASIC)、数字信号处理器(Digital Signal Processor,DSP)、数字信号处理装置(Digital Signal Processing Device,DSPD)、可编程逻辑装置(Programmable Logic Device,PLD)、现场可编程门阵列(Field Programmable Gate Array,FPGA)、中央处理器(Central Processing Unit,CPU)、控制器、微控制器、以及微处理器中的至少一种。可以理解地,对于不同的设备,用于实现上述处理器功能的电子器件还可以为其它,本申请实施例不作具体限定。
本申请实施例提供了一种计算机存储介质,应用于终端中,所述计算机存储介质存储有一个或多个视频播放相关程序,所述一个或者多个视频播放相关程序可被一个或者多个第一处理器执行,以实现实施例一中应用于终端中的虚拟现实的视频播放方法。
本申请实施例提供了一种计算机存储介质,应用于服务器中,所述计算机存储介质存储有一个或多个视频播放相关程序,所述一个或者多个视频播放相关程序可被一个或者多个第二处理器执行,以实现实施例一种应用于服务器的虚拟现实的视频播放方法。
本领域内的技术人员应明白,本申请的实施例可提供为方法、或计算机程序产品。因此,本申请可采用硬件实施例、软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器和光学存储器等)上实施的计算机程序产品的形式。
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和方框中的至少之一、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备 以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。

Claims (13)

  1. 一种视频播放方法,应用于终端,包括:
    在播放目标视频时,接收服务器发送的与所述目标视频对应的索引信息;
    当播放当前时间段的第一视频帧时,根据所述索引信息,确定与第一时间段相匹配的第二视频帧的下载地址;其中,所述第一视频帧与所述第二视频帧为所述目标视频中的视频帧,所述第一时间段为所述当前时间段的下一时间段;
    根据所述下载地址,在第一时刻下载所述第二视频帧;其中,所述第一时刻为所述当前时间段内的一预设时刻;
    在所述第一时间段起始时刻播放所述第二视频帧。
  2. 根据权利要求1所述的方法,其中,所述根据所述索引信息,确定与第一时间段相匹配的第二视频帧的下载地址,包括:
    根据所述索引信息,查找所述第二视频帧中的每个视频分片的聚焦值;
    根据所述聚焦值,确定所述第二视频帧中的每个视频分片的清晰度;
    根据所述清晰度,查找与所述清晰度对应的每个视频分片的下载地址。
  3. 根据权利要求2所述的方法,其中,所述根据所述聚焦值,确定所述第二视频帧中的每个视频分片的清晰度,包括:
    根据所述第二视频帧中的所有视频分片的聚焦值中的最大聚焦值,确定出与所述最大聚焦值对应的第一目标视频分片;
    将所述第一目标视频分片的清晰度确定为最高等级的清晰度;
    确定出所述第一目标视频分片的去心邻域内的第二目标视频分片;
    将所述第二目标视频分片的清晰度确定为次高等级的清晰度;
    确定出第三目标视频分片;
    将所述第三目标视频分片的清晰度确定为最低等级的清晰度;
    其中,所述第二视频帧由所述第一目标视频分片、所述第二目标视频分片以及所述第三目标视频分片组成。
  4. 根据权利要求3所述的方法,在所述在所述第一时间段起始时刻播放第二视频帧之前,还包括:
    在所述第一时间段起始时刻,检测实际最大聚焦值对应的第四目标视频分片;
    若所述第四目标视频分片与所述第一目标视频分片不匹配,则将所述第四目标视频分片的清晰度确定为所述最高等级的清晰度;
    根据所述索引信息,确定与所述最高等级的清晰度对应的所述第四目标视 频分片的下载地址;
    下载所述第四目标视频分片;
    播放所述第四目标视频分片。
  5. 一种视频播放方法,应用于服务器,包括:
    获取目标视频的焦点描述信息;
    将所述焦点描述信息,封装为所述目标视频的索引信息;
    发送所述索引信息至终端。
  6. 根据权利要求5所述的方法,其中,所述获取目标视频的焦点描述信息,包括:
    获取所述目标视频的预设焦点描述信息。
  7. 根据权利要求5所述的方法,其中,所述获取目标视频的焦点描述信息,包括:
    统计所述目标视频在第一预设时间段对应的聚焦值;
    将所述聚焦值作为所述焦点描述信息。
  8. 根据权利要求7所述的方法,其中,所述统计所述目标视频在第一预设时间段对应的所述聚焦值,包括:
    统计所述第一预设时间段中,第一时刻对应的每个视频分片对应的下载次数,直至统计完所述第一预设时间段中的所有时刻;其中,所述第一时刻为所述第一预设时间段的任一时刻;
    根据所述下载次数,确定所述每个视频分片对应的聚焦值。
  9. 根据权利要求8所述的方法,其中,所述根据所述下载次数,确定所述每个视频分片对应的聚焦值,包括:
    根据所述下载次数,确定出下载比率达到预设下载比率的最终视频分片;
    将所述最终视频分片对应的聚焦值设置为最大值;
    对所有视频分片中除所述最终视频分片外的剩余视频分片,按照各自对应的下载次数进行排列,得到排序后的剩余视频分片;
    依序设置排序后的剩余视频分片的聚焦值。
  10. 一种终端,包括:处理器、存储介质、通信总线以及播放器,所述处理器、所述存储介质以及所述播放器通过所述通信总线连接;
    所述处理器,设置为调用所述存储介质存储的视频播放相关程序,并执行如下步骤:
    在播放目标视频时,接收服务器发送的与所述目标视频对应的索引信息;
    当播放当前时间段的第一视频帧时,根据所述索引信息,确定与第一时间段相匹配的第二视频帧的下载地址;其中,所述第一视频帧与所述第二视频帧为所述目标视频中的视频帧,所述第一时间段为所述当前时间段的下一时间段;
    根据所述下载地址,在第一时刻下载所述第二视频帧;其中,所述第一时刻为所述当前时间段内的一预设时刻;
    所述播放器,设置为在所述第一时间段起始时刻播放所述第二视频帧。
  11. 一种服务器,包括:
    处理器,以及存储有所述处理器可执行指令的存储介质,所述存储介质通过通信总线依赖所述第二处理器执行操作,当所述指令被所述处理器执行时,执行如下操作:
    获取目标视频的焦点描述信息;
    将所述焦点描述信息,封装为目标视频的索引信息;
    发送所述索引信息至终端。
  12. 一种计算机存储介质,应用于终端中,所述计算机存储介质存储有至少一个视频播放相关程序,所述至少一个视频播放相关程序可被至少一个第一处理器执行,以实现如权利要求1至4所述的方法。
  13. 一种计算机存储介质,应用于服务器中,所述计算机存储介质存储有至少一个视频播放相关程序,所述至少一个视频播放相关程序可被至少一个第二处理器执行,以实现如权利要求5至9所述的方法。
PCT/CN2018/119434 2017-12-29 2018-12-06 视频播放方法以及终端、服务器及存储介质 WO2019128667A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP18897218.6A EP3734980A4 (en) 2017-12-29 2018-12-06 VIDEO PLAYBACK METHOD AND TERMINAL DEVICE, SERVER AND STORAGE MEDIUM

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201711482043.8A CN109996110B (zh) 2017-12-29 2017-12-29 一种视频播放方法、终端、服务器及存储介质
CN201711482043.8 2017-12-29

Publications (1)

Publication Number Publication Date
WO2019128667A1 true WO2019128667A1 (zh) 2019-07-04

Family

ID=67066463

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/119434 WO2019128667A1 (zh) 2017-12-29 2018-12-06 视频播放方法以及终端、服务器及存储介质

Country Status (3)

Country Link
EP (1) EP3734980A4 (zh)
CN (1) CN109996110B (zh)
WO (1) WO2019128667A1 (zh)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112788362A (zh) * 2020-12-25 2021-05-11 北京小米移动软件有限公司 一种视频播放方法、视频播放装置及存储介质
CN113747212A (zh) * 2021-08-23 2021-12-03 北京奇艺世纪科技有限公司 视频内容下载方法、装置、设备及存储介质
CN114205631A (zh) * 2021-10-28 2022-03-18 浙江大华技术股份有限公司 视频存储、目录生成、迁移方法、装置、设备和介质
CN114584806A (zh) * 2022-03-07 2022-06-03 湖南国科微电子股份有限公司 视频源文件的处理方法、视频播放方法、装置及设备

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113572975B (zh) * 2020-04-29 2023-06-06 华为技术有限公司 视频播放方法、装置及系统、计算机存储介质
CN114513669A (zh) * 2020-11-16 2022-05-17 华为云计算技术有限公司 视频编码及视频播放方法、装置和系统
CN112584255B (zh) * 2020-12-04 2023-05-26 广州虎牙科技有限公司 一种流媒体数据的播放方法、装置、计算机设备和存储介质
CN115243076A (zh) * 2021-04-22 2022-10-25 华为技术有限公司 视频播放方法、装置及系统、计算机可读存储介质
CN114615550B (zh) * 2022-03-17 2023-12-08 北京奇艺世纪科技有限公司 一种视频获取方法及装置

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2360581A (en) * 2000-03-15 2001-09-26 Television Monitoring Services Quantification of brand exposure on screen
EP1215901A1 (en) * 2000-12-07 2002-06-19 e-Seed Telecommunications S.p.A. Method and system of interactive television, possibly stimulated by telephone call
CN106331840A (zh) * 2016-08-31 2017-01-11 青岛海信宽带多媒体技术有限公司 一种基于hls协议的音视频的跳转播放方法及装置
CN106559677A (zh) * 2015-09-30 2017-04-05 华为技术有限公司 终端、缓存服务器及获取视频分片的方法及装置
CN106791956A (zh) * 2016-11-25 2017-05-31 百度在线网络技术(北京)有限公司 网络直播卡顿的处理方法及装置
CN106961630A (zh) * 2017-03-24 2017-07-18 西安理工大学 一种基于dash优化的p2p流媒体视频播放方法
CN107135207A (zh) * 2017-04-19 2017-09-05 中国电子科技集团公司电子科学研究院 一种全景会议直播方法及系统

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103096126B (zh) * 2012-12-28 2015-09-30 中国科学院计算技术研究所 协作式缓存集群中面向视频点播服务的协作式缓存方法及系统
US9509742B2 (en) * 2014-10-29 2016-11-29 DLVR, Inc. Configuring manifest files referencing infrastructure service providers for adaptive streaming video
CN104519372B (zh) * 2014-12-19 2017-11-21 深圳市九洲电器有限公司 一种流媒体播放的切换方法和系统
CN106937180A (zh) * 2015-12-31 2017-07-07 幸福在线(北京)网络技术有限公司 一种虚拟现实视频的播放方法及装置
US10565463B2 (en) * 2016-05-24 2020-02-18 Qualcomm Incorporated Advanced signaling of a most-interested region in an image
KR101945082B1 (ko) * 2016-07-05 2019-02-01 안규태 미디어 컨텐츠 송신 방법, 미디어 컨텐츠 송신 장치, 미디어 컨텐츠 수신 방법, 및 미디어 컨텐츠 수신 장치

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2360581A (en) * 2000-03-15 2001-09-26 Television Monitoring Services Quantification of brand exposure on screen
EP1215901A1 (en) * 2000-12-07 2002-06-19 e-Seed Telecommunications S.p.A. Method and system of interactive television, possibly stimulated by telephone call
CN106559677A (zh) * 2015-09-30 2017-04-05 华为技术有限公司 终端、缓存服务器及获取视频分片的方法及装置
CN106331840A (zh) * 2016-08-31 2017-01-11 青岛海信宽带多媒体技术有限公司 一种基于hls协议的音视频的跳转播放方法及装置
CN106791956A (zh) * 2016-11-25 2017-05-31 百度在线网络技术(北京)有限公司 网络直播卡顿的处理方法及装置
CN106961630A (zh) * 2017-03-24 2017-07-18 西安理工大学 一种基于dash优化的p2p流媒体视频播放方法
CN107135207A (zh) * 2017-04-19 2017-09-05 中国电子科技集团公司电子科学研究院 一种全景会议直播方法及系统

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3734980A4 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112788362A (zh) * 2020-12-25 2021-05-11 北京小米移动软件有限公司 一种视频播放方法、视频播放装置及存储介质
CN112788362B (zh) * 2020-12-25 2023-09-12 北京小米移动软件有限公司 一种视频播放方法、视频播放装置及存储介质
CN113747212A (zh) * 2021-08-23 2021-12-03 北京奇艺世纪科技有限公司 视频内容下载方法、装置、设备及存储介质
CN114205631A (zh) * 2021-10-28 2022-03-18 浙江大华技术股份有限公司 视频存储、目录生成、迁移方法、装置、设备和介质
CN114584806A (zh) * 2022-03-07 2022-06-03 湖南国科微电子股份有限公司 视频源文件的处理方法、视频播放方法、装置及设备

Also Published As

Publication number Publication date
EP3734980A1 (en) 2020-11-04
CN109996110A (zh) 2019-07-09
CN109996110B (zh) 2021-10-22
EP3734980A4 (en) 2021-05-19

Similar Documents

Publication Publication Date Title
WO2019128667A1 (zh) 视频播放方法以及终端、服务器及存储介质
US10313745B2 (en) Adaptive streaming of an immersive video scene
JP6564464B2 (ja) ビデオプログラムのセグメントの検出
US11589110B2 (en) Digital media system
CN107888987B (zh) 一种全景视频播放方法及装置
US20190387214A1 (en) Method for transmitting panoramic videos, terminal and server
WO2018010653A1 (zh) 全景媒体文件推送方法及装置
CN108124202B (zh) 一种使流媒体服务器支持hls协议的方法
US10897637B1 (en) Synchronize and present multiple live content streams
WO2019214371A1 (zh) 图像的显示方法、生成方法、装置、存储介质及电子装置
WO2018103384A1 (zh) 一种360度全景视频的播放方法、装置及系统
JP7182727B2 (ja) セッションのメディア・データを受信する方法、装置及びコンピュータ・プログラム
WO2019192481A1 (zh) 媒体信息处理方法、相关设备及计算机存储介质
CN108810567B (zh) 一种音频与视频视角匹配的方法、客户端和服务器
CN107438203B (zh) 用于建立和接收清单的方法、网络设备及终端
TW201942821A (zh) 資訊處理裝置、資訊處理方法、及程式
CN108574881B (zh) 一种投影类型推荐方法、服务器及客户端
TW201942713A (zh) 資訊處理裝置、資料處理方法、及程式
JP2024517702A (ja) シングルストリームを利用して関心領域の高画質映像を提供する方法、コンピュータ装置、およびコンピュータプログラム
US20220284661A1 (en) Methods for transmitting and rendering a 3d scene, method for generating patches, and corresponding devices and computer programs
US11134236B2 (en) Image processing device and system
US11763493B2 (en) Method and apparatus for tree-based point cloud compression (PCC) media stream using moving picture experts group (MPEG)-dynamic adaptive streaming over HTTP (DASH)
US20240214540A1 (en) Method for transmitting panoramic videos, terminal and server
US10178419B2 (en) Network-based video type identification method, client and server
WO2019148691A1 (zh) 存储介质与启动页面播放方法、装置及终端设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18897218

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2018897218

Country of ref document: EP

Effective date: 20200729