WO2021238454A1 - 直播视频的展示方法、装置、终端和可读存储介质 - Google Patents

直播视频的展示方法、装置、终端和可读存储介质 Download PDF

Info

Publication number
WO2021238454A1
WO2021238454A1 PCT/CN2021/086540 CN2021086540W WO2021238454A1 WO 2021238454 A1 WO2021238454 A1 WO 2021238454A1 CN 2021086540 W CN2021086540 W CN 2021086540W WO 2021238454 A1 WO2021238454 A1 WO 2021238454A1
Authority
WO
WIPO (PCT)
Prior art keywords
angle
live video
video
conversion
panoramic
Prior art date
Application number
PCT/CN2021/086540
Other languages
English (en)
French (fr)
Inventor
岑焕成
Original Assignee
广州华多网络科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 广州华多网络科技有限公司 filed Critical 广州华多网络科技有限公司
Publication of WO2021238454A1 publication Critical patent/WO2021238454A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440218Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by transcoding between formats or standards, e.g. from MPEG-2 to MPEG-4
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video

Definitions

  • This application relates to the technical field of video playback, and in particular to a method, device, terminal, and readable storage medium for displaying live video.
  • the anchor terminal In the process of online video live broadcast, the anchor terminal usually collects video data, encodes the video data and uploads it to the server.
  • the audience terminal obtains and decodes the video data of the current live room from the server to display the anchor video on the audience terminal. In the display window, the audience can watch the anchor video.
  • the video display mode is single, and the human-computer interaction is not smart enough.
  • a method for displaying live video includes:
  • the panoramic live video is a video of different shooting angles using a preset 3D sphere model Data obtained by projection processing;
  • receiving the input video angle conversion instruction includes:
  • the type of touch operation includes at least one of the following: angle rotation operation for the panoramic live video displayed at the first angle, and panoramic live video displayed at the first angle Zooming operation of the screen;
  • the video angle conversion instruction is determined.
  • performing angle conversion on the panoramic live video currently displayed at the first angle to obtain the panoramic live video at the second angle includes:
  • the conversion category includes any one of screen zoom in, screen zoom out, and screen rotation.
  • the conversion range includes zoom in, zoom out, and rotation angle. Any kind
  • the angle conversion is performed on the panoramic live video displayed in the first angle according to the conversion type and the conversion amplitude, to obtain the panoramic live video of the second angle.
  • parsing the video angle conversion command to determine the conversion category and conversion amplitude corresponding to the video angle conversion command includes:
  • the conversion amplitude corresponding to the video angle conversion instruction is determined.
  • the panoramic live video displayed at the first angle includes the synthesized live video frame and interactive interface elements during the live broadcast, and the current panoramic live video displayed at the first angle is angle-converted to obtain the second angle Live panoramic live video of, including:
  • the live video frame and the interactive interface elements are jointly converted into angles to obtain a second-angle panoramic live video.
  • the above method further includes:
  • the scene replacement instruction carries the target scene identifier
  • the target scene identifier determine the scene image corresponding to the scene replacement instruction
  • determining the video angle conversion instruction according to the type of touch operation includes:
  • the method for acquiring the above-mentioned panoramic live video includes:
  • the live video data includes video data of different shooting angles
  • the RGB video data is projected onto the 3D sphere model to obtain the panoramic live video; where the projection matrix includes the coordinate correspondence between the pixel points in the RGB video data and the pixel points on the 3D sphere model.
  • performing format conversion on live video data to obtain RGB video data includes:
  • the YUV video data is formatted to obtain RGB video data; the conversion formula includes a summing relationship and a difference relationship between Y/U/V data values.
  • the above method further includes:
  • the step of format conversion on the live video data is performed.
  • a live video display device which includes:
  • the receiving module is used to receive the input video angle conversion instruction
  • the angle conversion module is used to respond to the video angle conversion instruction to perform angle conversion on the panoramic live video currently displayed at the first angle to obtain the panoramic live video at the second angle; wherein the panoramic live video adopts a preset 3D sphere model Projected video data from different shooting angles;
  • the display module is used to display the panoramic live video from the second angle.
  • a terminal includes a memory and a processor, the memory stores a computer program, and the processor implements the steps of the above live video display method when the computer program is executed.
  • a computer-readable storage medium has a computer program stored thereon, and when the computer program is executed by a processor, the steps of the above-mentioned live video display method are realized.
  • the above-mentioned live video display method, device, terminal and readable storage medium can receive the input video angle conversion instruction; in response to the video angle conversion instruction, perform angle conversion on the panoramic live video currently displayed at the first angle to obtain the second Panoramic live video from different angles; show the live panoramic live video from the second angle.
  • the live video watched by the audience is a panoramic live video, that is, the video can be viewed from multiple angles, and the video display methods are rich; secondly, the audience can change the viewing angle of the panoramic live video at any time according to their needs, which greatly improves the intelligence of human-computer interaction sex.
  • FIG. 1 is a schematic diagram of an application environment of a method for displaying live video in an embodiment
  • FIG. 2 is a schematic flowchart of a method for displaying live video in an embodiment
  • Figure 2a is a schematic diagram of a spherical display of a panoramic live video in an embodiment
  • FIG. 3 is a schematic flowchart of a method for displaying live video in another embodiment
  • Figure 3a is a schematic diagram of a viewer using a mouse to perform touch operations on the current display interface in an embodiment
  • FIG. 4 is a schematic flowchart of a method for displaying live video in another embodiment
  • FIG. 5 is a schematic flowchart of a method for displaying live video in another embodiment
  • FIG. 6 is a schematic flowchart of a method for displaying live video in another embodiment
  • Fig. 6a is a schematic diagram of the effect of projecting RGB video data to a 3D sphere model in an embodiment
  • FIG. 6b is a schematic diagram of an engineering processing flow for obtaining a panoramic live video in an embodiment
  • FIG. 7 is a schematic flowchart of a method for displaying live video in another embodiment
  • Fig. 8 is a structural block diagram of a live video display device in an embodiment
  • Fig. 9 is a structural block diagram of a live video display device in another embodiment.
  • Fig. 10 is a structural block diagram of a live video display device in another embodiment
  • Fig. 11 is a diagram of the internal structure of a terminal in an embodiment.
  • the method for displaying live video can be applied to the application environment as shown in FIG. 1.
  • the host terminal 11 communicates with the server 12, the host can create a live room through the host terminal 11, and send the collected live video stream to the server 12 through the host terminal 11; the audience terminal 13 also communicates with the server 12 The audience can enter the live broadcast room created by the host through the audience terminal 13, and the audience terminal 13 obtains the live video stream corresponding to the live broadcast room from the server 12 for display for the audience to watch and interact.
  • the host can use a panoramic camera or multiple single-view cameras to shoot his own live video from multiple angles.
  • the viewer terminal 13 After the viewer terminal 13 obtains the live video stream from multiple angles, it performs panoramic projection processing to obtain the panoramic video, thereby realizing the audience Watch live video in 360 degrees.
  • the anchor terminal 11 and the audience terminal 13 may include, but are not limited to, personal computers, laptops, smart phones, tablets, and portable wearable devices.
  • the server 12 may be an independent server or a server cluster composed of multiple servers. accomplish.
  • a method for displaying live video is provided.
  • the method is applied to the audience terminal 13 (hereinafter referred to as terminal) in FIG. 1 as an example for description.
  • the viewer drags on the live video area displayed by the terminal to realize the specific process of watching the live video from multiple angles.
  • the method includes the following steps:
  • a viewer uses a terminal to watch a panoramic live video, it is usually viewed from one angle at the current moment, such as watching the video from the front angle of the host; if the viewer wants to change the angle of the video they watch, such as viewing the side angle of the host Video, you can enter the video angle conversion command on the terminal.
  • the viewer can input the above video angle conversion instruction through an input device external to the terminal (such as a mouse), such as clicking the left button of the mouse and sliding left to make the video rotate to the left; if the terminal’s display is a touch screen, the viewer can also use Slide the finger on the touch screen to input the above video angle conversion command, such as sliding the finger to the left to rotate the video to the left; the audience can also input the above video angle conversion command by voice, such as the terminal-facing microphone voice input "rotate left 30 degrees", etc. . Then, the terminal can receive the input video angle conversion instruction in the above-mentioned manner.
  • an input device external to the terminal such as a mouse
  • the viewer can also use Slide the finger on the touch screen to input the above video angle conversion command, such as sliding the finger to the left to rotate the video to the left
  • voice such as the terminal-facing microphone voice input "rotate left 30 degrees", etc.
  • the terminal can receive the input video angle conversion instruction in the above-mentioned manner.
  • S102 In response to the video angle conversion instruction, perform angle conversion on the panoramic live video currently displayed at the first angle to obtain the panoramic live video at the second angle; wherein, the panoramic live video uses a preset 3D sphere model to take different shooting angles The video data is projected.
  • the terminal can respond to the instruction to analyze the conversion operation carried in the video angle conversion instruction, and perform angle conversion on the panoramic live video currently displayed at the first angle according to the conversion operation.
  • the panoramic live video can be regarded as a video in the form of a sphere in a three-dimensional coordinate system, viewed from the center of the sphere in the Y-axis direction at a first angle ⁇ , which is covered by the first angle
  • the sphere part of is the video area that viewers can watch.
  • the first angle is converted into the second angle ⁇ , and the sphere covered by the second angle is taken as the video area to be displayed.
  • the process of angle conversion of the panoramic live video can be understood as: if the panoramic live video needs to be zoomed in or out, it is to change the angle of view on the Y axis; if the panoramic live video needs to be rotated left and right, it is to move around the Y axis. ; If you need to rotate the panoramic live video up and down, move up and down around the Z axis.
  • the terminal parses the video angle conversion instruction, Determine that the corresponding conversion operation is to rotate to the left, and according to the corresponding relationship between the sliding distance and the rotation angle, it can be determined that it needs to be rotated 90 degrees to the left along the Y axis; the panoramic live video of the second angle after the rotation is part r in the figure .
  • the above-mentioned panoramic live video is obtained by using a preset 3D sphere model to perform projection processing on video data of different shooting angles.
  • the host terminal sends the video data of different shooting angles to the server
  • the audience terminal can obtain the video data of the current live room from the server, and then the video data of different shooting angles are projected by the audience terminal to obtain the panoramic live video.
  • the server performs projection processing on the video data of different shooting angles to obtain the panoramic live video, and then the viewer terminal directly obtains the panoramic live broadcast of the current live room from the server video.
  • S103 Display the panoramic live video from the second angle.
  • the terminal After the terminal obtains the panoramic live video of the second angle, it can display the panoramic live video of the second angle in the display area of the display screen, that is, the process of viewing the panoramic live video from different angles is completed.
  • the live video watched by the audience is a panoramic live video, that is, the video can be viewed from multiple angles, and the video display methods are rich.
  • the terminal can receive the video angle conversion instruction input by the viewer, and in response to the video angle conversion instruction, perform angle conversion on the panoramic live video currently displayed at the first angle to obtain and display the panoramic live video at the second angle; that is, the viewer The viewing angle of the panoramic live video can be changed at any time according to their own needs, which greatly improves the intelligence of human-computer interaction.
  • the terminal after the terminal receives the video angle conversion instruction, the terminal itself performs the angle conversion action. After the instruction is not sent to the server for execution, the server sends the converted panoramic video data to the terminal, which greatly reduces The transmission power consumption between the server and the terminal makes the panoramic live video broadcast smoother.
  • the terminal receives an input video angle conversion instruction.
  • the foregoing S101 may include:
  • the type of touch operation includes at least one of the following: angle rotation operation for the panoramic live video displayed at the first angle, and panoramic view for the first angle display Screen zoom operation of live video.
  • the terminal may detect the touch operation on the current display interface, and determine the corresponding type according to the touch operation.
  • the type of the touch operation may include an angle rotation operation for the panoramic live video displayed at the first angle, such as Rotate left, rotate right, rotate up, rotate down, and zoom in and out of the panoramic live video displayed at the first angle, such as zooming in or out.
  • the terminal can detect the type of touch operation of the input device on the current display interface.
  • the input device is a mouse as an example for illustration.
  • the corresponding type can be determined; for example, touch The operation is to click the left mouse button to slide to the left, and the corresponding type is to rotate the panoramic live video to the left, and the touch operation is to click the left mouse button to slide to the right, and the corresponding type is to rotate the panoramic live video to the right, touch The operation is to slide the mouse wheel upward, and its corresponding type is to zoom in the video screen, and the touch operation is to slide the mouse wheel down, and its corresponding type is to zoom out the video screen.
  • Figure 3a For a schematic diagram of the audience using the mouse to perform touch operations on the current display interface, refer to Figure 3a.
  • the terminal can generate a video angle conversion instruction that can be recognized by the terminal according to the above-mentioned touch operation type, and the video angle conversion instruction can also carry the touch change amount corresponding to the touch operation, that is, how far the mouse cursor slides , Or the mouse wheel has rolled a few times, etc.
  • the terminal may also detect whether the received touch operation types are the same within a preset time interval.
  • the time interval can be set to 0.2s.
  • the video angle conversion instruction is determined according to the type of touch operation; if it is not the same (for example, the type of the previous 0.1s is Rotate to the left, and the type after 0.1s is to rotate to the right), the terminal may not respond to the touch operation, that is, discard the touch operation to prevent screen jitter.
  • the terminal detects the type of touch operation on the current display interface to determine the above video angle conversion instruction; by analyzing the touch operation on the current display interface, it is possible to accurately obtain that the viewer wants to switch After responding to this video angle conversion instruction, it can accurately display the panoramic live video of the angle that the viewer wants to watch, which further improves the intelligence of human-computer interaction.
  • the terminal responds to the video angle conversion instruction to perform angle conversion of the panoramic live video currently displayed at the first angle to obtain the panoramic live video at the second angle.
  • the foregoing S102 may include:
  • S301 Analyze the video angle conversion instruction, and determine the conversion category and conversion amplitude corresponding to the video angle conversion instruction; the conversion category includes any one of screen enlargement, screen reduction, and screen rotation, and the conversion amplitude includes enlargement ratio, reduction ratio, and rotation angle Any of them.
  • the terminal can determine the conversion category and conversion amplitude corresponding to the instruction by parsing the above video angle conversion instruction. Still taking the viewer’s input of a video angle conversion instruction through an input device (mouse) as an example, optionally, the terminal can determine the touch type and touch change amount of the input device according to the initial state and termination state of the input device.
  • the specific process is : Record the initial position of the mouse cursor and the initial position of the preset point on the scroll wheel; if the mouse cursor moves, record the end position of the mouse cursor, through the start position and end position, the mouse cursor in the x-axis and y-axis directions can be calculated Displacement difference, the movement direction of the cursor can be used as the touch type, and the displacement difference can be used as the touch change amount; if the mouse wheel rotates, the sliding track and end position of the preset point are recorded. Through the initial position and the end position, you can To determine the track length of the sliding track, the scroll direction of the scroll wheel can be used as the touch type, and the track length can be used as the touch change amount.
  • the terminal determines the conversion type corresponding to the video angle conversion instruction according to the touch type and the preset correspondence between the touch type and the conversion type.
  • the conversion category corresponding to the cursor moving to the left is screen rotation to the left
  • the conversion category corresponding to the cursor moving to the right is screen rotation to the right
  • the conversion category corresponding to scrolling the wheel upward is screen zooming
  • the conversion category corresponding to scrolling the scroll wheel down is The screen is reduced.
  • the conversion amplitude corresponding to the video angle conversion instruction is determined according to the touch change amount and the preset corresponding relationship between the touch change amount and the conversion amplitude.
  • the displacement difference of the mouse cursor moving to the left (or right) is 1cm, and the corresponding conversion amplitude is 30 degrees to the left (or right); the displacement difference of the mouse cursor moving to the left (or right) is 2cm, The corresponding conversion amplitude is 60 degrees to the left (or right); the track length of the scroll wheel up (or down) is half a circle, and the corresponding zoom in (or zoom out) ratio is 0.5 times; the mouse is up (or down) The length of the scrolling track is one circle, and the corresponding enlargement (or reduction) ratio is 1 times and so on.
  • S302 Perform angle conversion on the panoramic live video displayed at the first angle according to the conversion type and the conversion amplitude, to obtain the panoramic live video at the second angle.
  • the terminal can know the direction and amplitude of the angle conversion to be performed on the panoramic live video according to the conversion type and the conversion range.
  • the conversion type is screen rotation
  • the terminal can directly perform angle conversion on the panoramic live video according to the rotation angle, that is, when the conversion amplitude is 30 degrees to the left, the panoramic live video displayed at the first angle is directly rotated around the Y axis Rotate 30 degrees to the left to get the panoramic live video shown at the second angle.
  • the conversion type is screen zoom in or zoom out, the terminal needs to calculate the corresponding second angle according to the first angle and the zoom ratio or zoom ratio.
  • the first angle is a 60-degree viewing angle in the Y-axis direction
  • the terminal determines the conversion category and conversion range corresponding to the video angle conversion instruction by analyzing the video angle conversion instruction, and can accurately determine the video angle that the viewer wants to perform the panoramic live video; Perform angle conversion on the panoramic live video displayed at the first angle according to the conversion category and conversion range, and obtain the panoramic live video at the second angle, which can accurately display the panoramic live video at the angle that the viewer wants to watch, which further improves the human-computer interaction The intelligence.
  • some interactive interface elements can also be included on the video interface, such as the "send a gift” button, the “Like” button, and the “speak” box; in order to further enrich the display methods of the live video, the terminal
  • These interactive interface elements can also be combined with live video frame images to obtain a panoramic live video displayed from the first angle.
  • the interactive interface elements viewed by the audience can present a "three-dimensional” effect, and when the audience rotates or zooms the panorama When the video is broadcast live, the interactive interface elements will also undergo corresponding angle conversions.
  • the above S102 may include: performing angle conversion on the live video frame and the interactive interface element together to obtain the panoramic live video of the second angle.
  • the terminal may store the image of the interactive interface element, and record the characteristic location point of each interactive interface element (for example, the button position in the area of coordinates 0, 0-100, 100), and then compare the image of the interactive interface element with
  • the live video frame images are synthesized to obtain the panoramic live video displayed from the first angle, that is, the image of the interactive interface element and the live video frame image can be regarded as a whole image.
  • the terminal responds to the video angle conversion instruction, the live video frame and the interactive interface element can perform angle conversion together, and the interactive interface element is also rotated/zoomed correspondingly with the rotation/zooming of the live video frame.
  • the display methods of live video can be further enriched, and the viewer's perception of watching live video is greatly improved.
  • the terminal can also receive the scene replacement instruction from the viewer to replace the scene of the live video, such as a snow mountain scene, a forest scene, River scene and so on.
  • the foregoing method may further include:
  • S401 Receive an input scene replacement instruction; the scene replacement instruction carries a target scene identifier.
  • the audience can input the above scene replacement instruction through an input device external to the terminal (such as a mouse), such as clicking the left button of the mouse and clicking the corresponding scene icon; if the terminal’s display is a touch screen, the audience can also touch the touch screen with a finger. Touch the scene icon to input the scene replacement instruction; the audience can also input the above scene replacement instruction by voice, such as voice input "snow mountain scene” by the microphone of the terminal. Then, the terminal can receive the input scene replacement instruction in the above-mentioned manner, and the scene replacement instruction carries the target scene identifier.
  • an input device external to the terminal such as a mouse
  • S402 Determine a scene image corresponding to the scene replacement instruction according to the target scene identifier.
  • the terminal can determine the storage address of the scene image corresponding to the target scene identifier from the database, where the database stores the correspondence between different scene identifiers and the scene image storage addresses; then the terminal obtains The storage address of, obtain the corresponding scene image from this address.
  • the terminal may synthesize the scene image with the live video frame image of the panoramic live video displayed at the second angle, so as to replace the live background of the panoramic live video with the above-mentioned scene image.
  • the terminal may use the host area in the live video frame as the target area, and other areas as the background area, and replace the background area with the aforementioned scene image.
  • the terminal may receive an input scene replacement instruction, determine the scene image according to the target scene identifier carried by the scene replacement instruction, and then replace the live background of the panoramic live video displayed from the second angle with The scene image.
  • the audience can replace the background of the panoramic live video according to their own scene preferences to meet the audience's demand for display methods, and further improve the intelligence of human-computer interaction.
  • the method for acquiring the panoramic live video includes:
  • S501 Acquire live video data sent by a server; the live video data includes video data of different shooting angles.
  • the live video data may be video stream data uploaded to the server by the anchor terminal and obtained from the server by the viewer terminal, which includes video data of different shooting angles, such as four angles of front, back, left, and right shot by the anchor through the anchor terminal Video data.
  • the live video data may also carry a panorama identifier.
  • the video data of different angles shot by the same anchor at the same time have the same panorama identifier, thereby ensuring that the video data at the same time is projected into the same panoramic live video.
  • the above-mentioned live video data is video stream data with an aspect ratio of 2:1; since the host terminal parameters used by each host are different, the uploaded video stream data may not have a size of 2:1, then the viewer terminal After receiving the video stream data, it can also be cropped to a 2:1 size to make the picture more rounded and smooth.
  • S502 Perform format conversion on the live video data to obtain RGB video data.
  • GPU technology may be used to perform data processing on the live video data, and the terminal needs to first convert the live video data into a data format that can be processed by the GPU.
  • the terminal first decodes the live video data to obtain YUV video data; wherein, the YUV planar data is a form of encoded data in which the sampling rate of the chrominance in the horizontal direction is halved, that is, the phase of each line is The two adjacent Y (luminance information) share one U and V (chromaticity information) data, which are mostly used for image data received by the network terminal and/or the playback terminal, which can reduce the bandwidth of transmitting image data. Then, the terminal performs format conversion on the YUV video data according to the preset conversion formula to obtain RGB video data; wherein, the conversion formula includes a summation relationship and a difference relationship between Y/U/V data values.
  • the conversion formula can be It may also include other summation relations and difference relations between Y/U/V data values.
  • the terminal can transmit live video data in an inter-process communication (Inter-Process Communication, IPC) manner to perform processing processes such as decoding and format conversion.
  • IPC Inter-Process Communication
  • the terminal can also use the vertex shader to create three textures to fill Y, U, and V video data respectively.
  • the built-in bilinear filtering algorithm of the graphics card can be used for interpolation.
  • Other image interpolation algorithms can be used to achieve higher image quality.
  • the above interpolation algorithm can be realized by Shader, and then Y, U, and V can be rendered to the texture respectively, and then the YUV video data can be converted into RGB video data by the conversion formula.
  • the terminal can also determine whether the live video data carries a panoramic identifier, whether the terminal itself supports the above format conversion operation, and whether the terminal itself supports panoramic projection processing; if all the above conditions If it is met, the step of S502 is executed; if any of the conditions is not met, the terminal processes the live video data as a normal video, and does not convert it into a panoramic live video. In this way, compatible processing is performed, which greatly improves the error-tolerant rate of processing live video data.
  • S503 Create a 3D sphere model of a preset size, and the 3D sphere model is used for coordinate projection of RGB video data.
  • the terminal needs to map the above-mentioned live video data into a panoramic live video, the panoramic live video can be regarded as a video in the form of a sphere, the terminal also needs to create a 3D sphere model of a preset size for the RGB video data. Coordinate projection.
  • the size of the 3D sphere model may be a 3D sphere with a radius of 1 and a cross-section and a vertical section of 128.
  • the terminal can project the RGB video data on the 3D sphere model based on the projection matrix to obtain the panoramic live video.
  • this embodiment can use the D3D method to perform projection mapping on RGB video data.
  • the D3D method is a method for texture mapping on the surface of a 3D object. By specifying a set of texture coordinates for each vertex of the video image, these The coordinate positions of the vertices on the texture surface (that is, the 3D sphere model) are used to establish the mapping relationship between the video image and the texture surface.
  • Figure 6a For a schematic diagram of the effect of projecting RGB video data to the 3D sphere model, refer to Figure 6a.
  • the vertical observation range of the camera point is the viewing angle fovy (the maximum angle is 120 degrees, the minimum angle is 60 degrees, and the average value is 90 degrees)
  • np is the near clipping plane ( Set to 0.6)
  • N is the distance from the camera point
  • fp is the far clipping plane (set to 100)
  • F is the distance from the camera point;
  • the above projection matrix can be in, (x, y, z) are the coordinates of the pixels in the RGB video data, (x', y', z') are the coordinates of the pixels on the 3D sphere model.
  • the terminal first obtains the live video data sent by the server, and converts the format of the live video data to obtain RGB video data.
  • the GPU technology is used to perform projection processing on the RGB video data, which greatly improves In order to improve the projection efficiency, the audience cannot perceive the stagnation of the video picture.
  • a 3D sphere model of a preset size is created, and the RGB video data is projected onto the 3D sphere model based on the projection matrix to obtain a panoramic live video.
  • the steps of performing projection processing in this embodiment are executed by the audience terminal, and the server does not need to be executed and then sent to the audience terminal, which reduces the communication power consumption between the server and the audience terminal.
  • the server obtains the video data of different shooting angles sent by the host terminal, and performs format conversion on the video data to obtain RGB video data; and based on the projection matrix, the RGB video data is projected onto the 3D sphere model to obtain Panoramic live video.
  • the viewer terminal sends the video angle conversion instruction to the server, and the server performs angle conversion on the panoramic live video of the first angle according to the video angle conversion instruction to obtain the panoramic live broadcast of the second angle video.
  • the server compares the video areas corresponding to the first angle and the second angle to obtain the difference area (that is, the different parts of the video area corresponding to the two angles), and sends the difference area to the audience terminal, which is based on The difference area is intercepted and merged with the panoramic live video originally displayed at the first angle to obtain the panoramic live video displayed at the second angle.
  • the amount of data that the server interacts with the audience terminal can be greatly reduced, the communication power consumption can be reduced, and the amount of calculation of the audience terminal can be reduced at the same time.
  • the live video data is usually video stream data with an aspect ratio of 2:1
  • the host terminal wants the server to send the video data or the server sends the video data to the viewer terminal, it can be compressed, and the viewer terminal receives the video. After the data, zoom in to a 2:1 size. As a result, the bandwidth occupied by the communication between each terminal and the server can also be reduced.
  • the method includes:
  • S602 Determine the video angle conversion instruction according to the type of touch operation
  • S603 Determine the touch type and touch change amount of the input device according to the initial state and the termination state of the input device;
  • S604 Determine the conversion type corresponding to the video angle conversion instruction according to the touch type and the preset correspondence between the touch type and the conversion type.
  • S605 Determine the conversion amplitude corresponding to the video angle conversion instruction according to the touch change amount and the preset correspondence between the touch change amount and the conversion amplitude.
  • S606 Perform angle conversion on the panoramic live video displayed at the first angle according to the conversion category and conversion amplitude, to obtain a panoramic live video at a second angle;
  • S607 Display the panoramic live video from the second angle.
  • a live video display device which includes: a receiving module 21, an angle conversion module 22 and a display module 23.
  • the receiving module 21 is configured to receive an input video angle conversion instruction.
  • the angle conversion module 22 is configured to perform angle conversion on the panoramic live video currently displayed at the first angle in response to the video angle conversion instruction to obtain the panoramic live video at the second angle; wherein the panoramic live video is a preset 3D sphere
  • the model is obtained by projecting the video data of different shooting angles.
  • the display module 23 is used to display the panoramic live video of the second angle.
  • the live video display device provided in this embodiment can execute the foregoing method embodiments, and its implementation principles and technical effects are similar, and will not be repeated here.
  • the receiving module 21 is specifically configured to detect the type of touch operation of the input device on the current display interface; the type of touch operation includes at least one of the following: Angle rotation operation, screen zoom operation for the panoramic live video displayed at the first angle; according to the type of touch operation, the video angle conversion instruction is determined.
  • the angle conversion module 22 is specifically configured to parse the video angle conversion instruction, and determine the conversion category and conversion amplitude corresponding to the video angle conversion instruction; the conversion category includes any one of image enlargement, image reduction, and image rotation , The conversion range includes any one of enlargement ratio, reduction ratio, and rotation angle; according to the conversion category and conversion range, perform angle conversion on the panoramic live video displayed at the first angle to obtain the panoramic live video of the second angle.
  • the angle conversion module 22 is specifically configured to determine the touch type and touch change amount of the input device according to the initial state and the end state of the input device; according to the touch type and the preset touch type and The correspondence between the conversion types is determined, and the conversion type corresponding to the video angle conversion instruction is determined; the conversion amplitude corresponding to the video angle conversion instruction is determined according to the touch change amount and the preset correspondence between the touch change amount and the conversion amplitude.
  • the panoramic live video displayed from the first angle includes the synthesized live video frame and interactive interface elements during the live broadcast.
  • the angle conversion module 22 is specifically configured to perform angle conversion between the live video frame and the interactive interface elements. , Get the panoramic live video of the second angle.
  • the above-mentioned apparatus further includes a determination module 24 and a replacement module 25.
  • the receiving module 21 is also configured to receive an input scene replacement instruction; the scene replacement instruction carries the target scene identifier.
  • the determining module 24 is configured to determine the scene image corresponding to the scene replacement instruction according to the target scene identifier.
  • the replacement module 25 is used to replace the live background of the panoramic live video displayed from the second angle with the scene image.
  • the live video display device provided in this embodiment can execute the foregoing method embodiments, and its implementation principles and technical effects are similar, and will not be repeated here.
  • the receiving module 21 is specifically configured to detect whether the received touch operation types are the same within a preset time interval; if so, determine the video angle conversion instruction according to the touch operation type; if not, It does not respond to touch operations.
  • the foregoing apparatus further includes: an acquisition module 26, a format conversion module 27, a creation module 28, and a projection module 29.
  • the acquisition module 26 is configured to acquire live video data sent by the server; the live video data includes video data of different shooting angles.
  • the format conversion module 27 is used for format conversion of the live video data to obtain RGB video data.
  • the creation module 28 is used to create a 3D sphere model of a preset size, and the 3D sphere model is used for coordinate projection of RGB video data.
  • the projection module 29 is used to project the RGB video data onto the 3D sphere model based on the projection matrix to obtain the panoramic live video; wherein the projection matrix includes the coordinate correspondence between the pixel points in the RGB video data and the pixel points on the 3D sphere model.
  • FIG. 10 is shown on the basis of the embodiment shown in FIG. 8, which is only an example here, and it can also be shown on the basis of the embodiment shown in FIG. 9.
  • the format conversion module 27 is specifically configured to decode live video data to obtain YUV video data; according to a preset conversion formula, format the YUV video data to obtain RGB video data; the conversion formula includes Y /U/V The sum relation and difference relation between the data values.
  • the above-mentioned device further includes a judgment module for judging whether the live video data carries a panoramic identifier, whether the terminal supports format conversion, and whether the terminal supports panoramic projection processing; if the live video data carries a panoramic identifier and the terminal supports the format And the terminal supports panoramic projection processing, the format conversion module 27 is instructed to perform the step of format conversion on the live video data.
  • the live video display device provided in this embodiment can execute the foregoing method embodiments, and its implementation principles and technical effects are similar, and will not be repeated here.
  • the various modules in the above-mentioned live video display device can be implemented in whole or in part by software, hardware, and a combination thereof.
  • the foregoing modules may be embedded in the form of hardware or independent of the processor in the terminal, or may be stored in the memory of the terminal in the form of software, so that the processor can invoke and execute the operations corresponding to the foregoing modules.
  • a terminal is provided, and its internal structure diagram may be as shown in FIG. 11.
  • the terminal includes a processor, a memory, a communication interface, a display screen, and an input device connected through a system bus.
  • the terminal's processor is used to provide computing and control capabilities.
  • the storage of the terminal includes a non-volatile storage medium and internal storage.
  • the non-volatile storage medium stores an operating system and a computer program.
  • the internal memory provides an environment for the operation of the operating system and computer programs in the non-volatile storage medium.
  • the communication interface of the terminal is used for wired or wireless communication with an external terminal.
  • the wireless mode can be implemented through WIFI, operator network, NFC (near field communication) or other technologies.
  • the display screen of the terminal can be a liquid crystal display or an electronic ink display
  • the input device of the terminal can be a touch layer covered on the display, or a button, trackball or touchpad set on the terminal shell, or It is an external keyboard, touchpad or mouse, etc.
  • FIG. 11 is only a block diagram of part of the structure related to the solution of the present application, and does not constitute a limitation on the terminal to which the solution of the present application is applied.
  • the specific terminal may include More or fewer components are shown in the figure, or some components are combined, or have different component arrangements.
  • a terminal including a memory and a processor, a computer program is stored in the memory, and the processor implements the following steps when the processor executes the computer program:
  • the panoramic live video is a video of different shooting angles using a preset 3D sphere model Data obtained by projection processing;
  • the processor further implements the following steps when executing the computer program:
  • the type of touch operation includes at least one of the following: angle rotation operation for the panoramic live video displayed at the first angle, and panoramic live video displayed at the first angle Zooming operation of the screen;
  • the video angle conversion instruction is determined.
  • the processor further implements the following steps when executing the computer program:
  • the conversion category includes any one of screen zoom in, screen zoom out, and screen rotation.
  • the conversion range includes zoom in, zoom out, and rotation angle. Any kind
  • the angle conversion is performed on the panoramic live video displayed in the first angle according to the conversion type and the conversion amplitude, to obtain the panoramic live video of the second angle.
  • the processor further implements the following steps when executing the computer program:
  • the conversion amplitude corresponding to the video angle conversion instruction is determined.
  • the panoramic live video displayed from the first angle includes synthesized live video frames and interactive interface elements during the live broadcast, and the processor further implements the following steps when executing the computer program:
  • the live video frame and the interactive interface elements are jointly converted into angles to obtain a second-angle panoramic live video.
  • the processor further implements the following steps when executing the computer program:
  • the scene replacement instruction carries the target scene identifier
  • the target scene identifier determine the scene image corresponding to the scene replacement instruction
  • the processor further implements the following steps when executing the computer program:
  • the processor further implements the following steps when executing the computer program:
  • the live video data includes video data of different shooting angles
  • the RGB video data is projected onto the 3D sphere model to obtain the panoramic live video; where the projection matrix includes the coordinate correspondence between the pixel points in the RGB video data and the pixel points on the 3D sphere model.
  • the processor further implements the following steps when executing the computer program:
  • the YUV video data is formatted to obtain RGB video data; the conversion formula includes a summing relationship and a difference relationship between Y/U/V data values.
  • the processor further implements the following steps when executing the computer program:
  • the step of format conversion on the live video data is performed.
  • a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, the following steps are implemented:
  • the panoramic live video is a video of different shooting angles using a preset 3D sphere model Data obtained by projection processing;
  • the type of touch operation includes at least one of the following: angle rotation operation for the panoramic live video displayed at the first angle, and panoramic live video displayed at the first angle Zooming operation of the screen;
  • the video angle conversion instruction is determined.
  • the conversion category includes any one of screen zoom in, screen zoom out, and screen rotation.
  • the conversion range includes zoom in, zoom out, and rotation angle. Any kind
  • the angle conversion is performed on the panoramic live video displayed in the first angle according to the conversion type and the conversion amplitude, to obtain the panoramic live video of the second angle.
  • the conversion amplitude corresponding to the video angle conversion instruction is determined.
  • the panoramic live video displayed from the first angle includes synthesized live video frames and interactive interface elements during the live broadcast, and the computer program is executed by the processor to further implement the following steps:
  • the live video frame and the interactive interface elements are jointly converted into angles to obtain a second-angle panoramic live video.
  • the scene replacement instruction carries the target scene identifier
  • the target scene identifier determine the scene image corresponding to the scene replacement instruction
  • the live video data includes video data of different shooting angles
  • the RGB video data is projected onto the 3D sphere model to obtain the panoramic live video; where the projection matrix includes the coordinate correspondence between the pixel points in the RGB video data and the pixel points on the 3D sphere model.
  • the YUV video data is formatted to obtain RGB video data; the conversion formula includes a summing relationship and a difference relationship between Y/U/V data values.
  • the step of format conversion on the live video data is performed.
  • Non-volatile memory may include read-only memory (Read-Only Memory, ROM), magnetic tape, floppy disk, flash memory, or optical storage.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM may be in various forms, such as static random access memory (Static Random Access Memory, SRAM) or dynamic random access memory (Dynamic Random Access Memory, DRAM), etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Controls And Circuits For Display Device (AREA)

Abstract

一种直播视频的展示方法、装置、终端和可读存储介质,该方法包括:接收输入的视频角度转换指令;响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;展示第二角度的全景直播视频。该方法中,首先,观众所观看的直播视频为全景直播视频,即可以从多角度观看视频,视频展示方式丰富;其次,观众可以根据自身需要随时转换全景直播视频的观看角度,大大提高了人机交互的智能性。

Description

直播视频的展示方法、装置、终端和可读存储介质
本申请要求于2020年05月29日提交至中国专利局、申请号为202010476780.2、发明名称为“直播视频的展示方法、装置、终端和可读存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及视频播放技术领域,特别是涉及一种直播视频的展示方法、装置、终端和可读存储介质。
背景技术
在网络视频直播过程中,通常由主播终端采集视频数据,将视频数据编码后上传至服务器,观众终端从服务器中获取当前所在直播间的视频数据并进行解码,以将主播视频展示在观众终端的显示窗口中,供观众观看主播视频。
当前较多的网络视频直播通常是由主播从一个角度,采用终端摄像头(或摄像机)拍摄自身所处的场景视频,那么观众所观看到的主播视频也为该角度的视频画面;当主播自己主动变换拍摄角度时,观众才可观看到其他角度的视频画面。
因此,传统技术的网络视频直播中,视频展示方式单一,人机交互不够智能。
发明内容
基于此,有必要针对传统技术的网络视频直播中,视频展示方式单一,人机交互不够智能的问题,提供一种直播视频的展示方法、装置、终端和可读存储介质。
一种直播视频的展示方法,该方法包括:
接收输入的视频角度转换指令;
响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中,全景直播视频为采用 预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的;
展示第二角度的全景直播视频。
在其中一个实施例中,接收输入的视频角度转换指令,包括:
检测输入设备在当前展示界面上的触控操作的类型;该触控操作的类型包括以下至少一种:针对第一角度展示的全景直播视频的角度旋转操作、针对第一角度展示的全景直播视频的画面缩放操作;
根据触控操作的类型,确定视频角度转换指令。
在其中一个实施例中,响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频,包括:
解析视频角度转换指令,确定该视频角度转换指令对应的转换类别和转换幅度;转换类别包括画面放大、画面缩小和画面旋转中的任一种,转换幅度包括放大比例、缩小比例和旋转角度中的任一种;
根据转换类别和转换幅度对以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频。
在其中一个实施例中,解析视频角度转换指令,确定视频角度转换指令对应的转换类别和转换幅度,包括:
根据输入设备的初始状态和终止状态,确定输入设备的触控类型和触控变化量;
根据触控类型、以及预设的触控类型与转换类别的对应关系,确定视频角度转换指令对应的转换类别;
根据触控变化量、以及预设的触控变化量与转换幅度的对应关系,确定视频角度转换指令对应的转换幅度。
在其中一个实施例中,以第一角度展示的全景直播视频包括合成的直播视频帧以及直播过程中的交互界面元素,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频,包括:
将直播视频帧和交互界面元素共同进行角度转换,得到第二角度的全景直播视频。
在其中一个实施例中,上述方法还包括:
接收输入的场景替换指令;该场景替换指令携带目标场景标识;
根据目标场景标识,确定与场景替换指令对应的场景图像;
将以第二角度展示的全景直播视频的直播背景替换为该场景图像。
在其中一个实施例中,根据触控操作的类型,确定视频角度转换指令,包括:
检测预设的时间间隔内,所接收的触控操作的类型是否相同;
若是,则根据触控操作的类型确定视频角度转换指令;
若否,则不响应触控操作。
在其中一个实施例中,上述全景直播视频的获取方式包括:
获取服务器发送的直播视频数据;该直播视频数据包括不同拍摄角度的视频数据;
对直播视频数据进行格式转换,得到RGB视频数据;
创建预设大小的3D球体模型,该3D球体模型用于RGB视频数据的坐标投影;
基于投影矩阵,将RGB视频数据投影至3D球体模型上,得到全景直播视频;其中,投影矩阵包括RGB视频数据中像素点与3D球体模型上像素点的坐标对应关系。
在其中一个实施例中,对直播视频数据进行格式转换,得到RGB视频数据,包括:
对直播视频数据进行解码,得到YUV视频数据;
根据预设的转换公式,对YUV视频数据进行格式转换,得到RGB视频数据;转换公式包括Y/U/V数据值之间的求和关系式和求差关系式。
在其中一个实施例中,在获取服务器发送的直播视频数据之后,上述方法还包括:
判断直播视频数据是否携带全景标识、终端是否支持格式转换、以及终端是否支持全景投影处理;
若直播视频数据携带全景标识、且终端支持格式转换、且终端支持全景投影处理,则执行对直播视频数据进行格式转换的步骤。
一种直播视频的展示装置,该装置包括:
接收模块,用于接收输入的视频角度转换指令;
角度转换模块,用于响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中, 全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的;
展示模块,用于展示第二角度的全景直播视频。
一种终端,包括存储器和处理器,存储器存储有计算机程序,处理器执行该计算机程序时实现上述直播视频的展示方法的步骤。
一种计算机可读存储介质,其上存储有计算机程序,该计算机程序被处理器执行时实现上述直播视频的展示方法的步骤。
上述直播视频的展示方法、装置、终端和可读存储介质,能够接收输入的视频角度转换指令;响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;展示第二角度的全景直播视频。首先,观众所观看的直播视频为全景直播视频,即可以从多角度观看视频,视频展示方式丰富;其次,观众可以根据自身需要随时转换全景直播视频的观看角度,大大提高了人机交互的智能性。
附图说明
图1为一个实施例中直播视频的展示方法的应用环境示意图;
图2为一个实施例中直播视频的展示方法的流程示意图;
图2a为一个实施例中全景直播视频的球体展示示意图;
图3为另一个实施例中直播视频的展示方法的流程示意图;
图3a为一个实施例中观众使用鼠标在当前展示界面进行触控操作的示意图;
图4为又一个实施例中直播视频的展示方法的流程示意图;
图5为又一个实施例中直播视频的展示方法的流程示意图;
图6为又一个实施例中直播视频的展示方法的流程示意图;
图6a为一个实施例中将RGB视频数据投影至3D球体模型的效果示意图;
图6b为一个实施例中得到全景直播视频的工程处理流程示意图;
图7为又一个实施例中直播视频的展示方法的流程示意图;
图8为一个实施例中直播视频的展示装置的结构框图;
图9为另一个实施例中直播视频的展示装置的结构框图;
图10为又一个实施例中直播视频的展示装置的结构框图;
图11为一个实施例中终端的内部结构图。
附图标记说明:
11:主播终端;  12:服务器;  13:观众终端。
具体实施方式
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。
本申请实施例提供的直播视频的展示方法,可以应用于如图1所示的应用环境中。其中,主播终端11通过与服务器12进行通信连接,主播可以通过主播终端11创建直播间,并将采集的直播视频流通过主播终端11发送至服务器12;观众终端13也通过与服务器12进行通信连接,观众可以通过观众终端13进入主播所创建的直播间,并由观众终端13从服务器12中获取该直播间所对应的直播视频流进行展示,以供观众进行观看及互动。本申请实施例中,主播可以使用全景摄像头或多个单景摄像头从多个角度拍摄自己的直播视频,观众终端13获取到多个角度的直播视频流后进行全景投影处理得到全景视频,实现观众360度观看直播视频。可选地,主播终端11和观众终端13可以包括但不限于个人计算机、笔记本电脑、智能手机、平板电脑和便携式可穿戴设备,服务器12可以用独立的服务器或者是多个服务器组成的服务器集群来实现。
在一个实施例中,如图2所示,提供了一种直播视频的展示方法,以该方法应用于图1中的观众终端13(以下简称为终端)为例进行说明,本实施例涉及的是观众在终端所展示的直播视频区域上进行拖动,以实现多角度观看直播视频的具体过程,该方法包括以下步骤:
S101,接收输入的视频角度转换指令。
具体地,当观众使用终端观看全景直播视频时,在当前时刻通常是从一个角度进行观看的,如观看主播正面角度的视频;若观众想要转换其所观看的视频角度,如观看主播侧面角度的视频,则可以在终端上输入视频角度转换指令。可选地,观众可以通过终端外接的输入设备(如鼠标)输 入上述视频角度转换指令,如点击鼠标的左键向左滑动使视频向左旋转;若终端的显示屏为触摸屏,观众也可以通过手指在触摸屏上滑动输入上述视频角度转换指令,如手指向左滑动使视频向左旋转;观众还可以通过语音输入上述视频角度转换指令,如面向终端的麦克风语音输入“向左旋转30度”等。那么,终端便可以通过上述方式接收到输入的视频角度转换指令。
S102,响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中,全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的。
具体地,终端接收到视频角度转换指令后,便可响应于该指令,解析视频角度转换指令中携带的转换操作,根据该转换操作将当前以第一角度展示的全景直播视频进行角度转换。其中,参见图2a所示,可以将全景直播视频看做一个位于三维坐标系下球体形式的视频,从球体的球心在Y轴方向以第一角度α向外观看,该第一角度所覆盖的球体部分即观众可以观看到的视频区域。当对以第一角度展示的全景直播视频进行角度转换时,即将第一角度转换为第二角度β,将第二角度所覆盖的球体部分作为所要展示的视频区域。对全景直播视频进行角度转换的过程可以理解为:若需将全景直播视频进行放大或缩小,则为改变Y轴上的视角;若需将全景直播视频左右旋转,则为绕着Y轴左右移动;若需将全景直播视频上下旋转,则为绕着Z轴上下移动。
示例性的,继续参见图2a所示,假设观众当前观看的第一角度的全景直播视频为图中的w部分;通过手指在触摸屏上向左滑动3cm,终端对视频角度转换指令进行解析,可以确定对应的转换操作为向左旋转,并根据滑动距离与旋转角度的对应关系,可以确定需要沿着Y轴方向向左旋转90度;旋转之后第二角度的全景直播视频为图中的r部分。
其中,上述全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的。一方面,主播终端将不同拍摄角度的视频数据发送至服务器后,观众终端可以从服务器获取当前直播间的视频数据,然后由观众终端将不同拍摄角度的视频数据进行投影处理,得到全景 直播视频。另一方面,主播终端将不同拍摄角度的视频数据发送至服务器后,服务器将不同拍摄角度的视频数据进行投影处理,得到全景直播视频,然后由观众终端直接从服务器中获取当前直播间的全景直播视频。
S103,展示第二角度的全景直播视频。
具体地,终端得到了第二角度的全景直播视频后,便可以在显示屏的展示区域展示该第二角度的全景直播视频,即完成了从不同角度观看全景直播视频的过程。
本实施例提供的直播视频的展示方法,首先,观众所观看的直播视频为全景直播视频,即可以从多角度观看视频,视频展示方式丰富。其次,终端可以接收观众输入的视频角度转换指令,并响应于该视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频并展示;即观众可以根据自身需要随时转换全景直播视频的观看角度,大大提高了人机交互的智能性。另外,终端接收到视频角度转换指令后,由终端自身执行角度转换的动作,不需要将该指令发送至服务器执行后,再由服务器将转换后的全景视频数据发送给终端,由此大大减少了服务器与终端之间的传输功耗,使得播放的全景直播视频画面更加流畅。
在一个实施例中,如图3所示,涉及的是终端接收输入的视频角度转换指令的具体过程。可选地,上述S101可以包括:
S201,检测输入设备在当前展示界面上的触控操作的类型;该触控操作的类型包括以下至少一种:针对第一角度展示的全景直播视频的角度旋转操作、针对第一角度展示的全景直播视频的画面缩放操作。
具体地,终端可以检测当前展示界面上的触控操作,根据该触控操作确定其对应的类型,该触控操作的类型可以包括针对第一角度展示的全景直播视频的角度旋转操作,如向左旋转、向右旋转、向上旋转、向下旋转,以及针对第一角度展示的全景直播视频的画面缩放操作,如将画面放大或缩小。
可选地,终端可以检测输入设备在当前展示界面上的触控操作的类型,以该输入设备为鼠标为例进行说明,通过检测鼠标的触控操作,可以确定其对应的类型;如触控操作为点击鼠标左键向左滑动、其对应的类型为使 全景直播视频向左旋转,触控操作为点击鼠标左键向右滑动、其对应的类型为使全景直播视频向右旋转,触控操作为向上滑动鼠标滚轮、其对应的类型为放大视频画面,触控操作为向下滑动鼠标滚轮、其对应的类型为缩小视频画面。关于观众使用鼠标在当前展示界面进行触控操作的示意图可以参见图3a所示。
S202,根据触控操作的类型,确定视频角度转换指令。
具体地,终端可以根据上述触控操作的类型,生成终端可以识别的视频角度转换指令,该视频角度转换指令还可以携带触控操作对应的触控变化量,即鼠标光标滑动了多长的距离,或者鼠标滚轮滚动了几圈等。
在一些场景中,观众使用输入设备在当前展示界面上输入触控操作时,有时会出现误操作行为,如点击鼠标左键向左滑动时又突然向右滑动,导致全景直播视频画面抖动的现象。因此,可选地,终端还可以检测预设的时间间隔内,所接收到的触控操作的类型是否相同。可选地,可以设置时间间隔为0.2s,在这0.2s内,若触控操作的类型相同,则根据触控操作的类型确定视频角度转换指令;若不相同(如前0.1s的类型为向左旋转,后0.1s的类型为向右旋转),则终端可以不响应该触控操作,即丢弃此次触控操作,以防止画面抖动现象。
本实施例提供的直播视频的展示方法,终端检测当前展示界面上的触控操作的类型,以确定上述视频角度转换指令;通过分析当前展示界面上的触控操作,可以准确得到观众想要转换的视频角度,那么响应此视频角度转换指令后,便可以准确的展示出观众想要观看角度的全景直播视频,进一步提高了人机交互的智能性。
在一个实施例中,如图4所示,涉及的是终端响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频的具体过程。可选地,上述S102可以包括:
S301,解析视频角度转换指令,确定视频角度转换指令对应的转换类别和转换幅度;该转换类别包括画面放大、画面缩小和画面旋转中的任一种,转换幅度包括放大比例、缩小比例和旋转角度中的任一种。
具体地,终端通过解析上述视频角度转换指令,可以确定该指令对应的转换类别和转换幅度。仍以观众通过输入设备(鼠标)输入视频角度转 换指令为例,可选地,终端可以根据输入设备的初始状态和终止状态,确定该输入设备的触控类型和触控变化量,具体过程为:记录鼠标光标的初始位置以及滚轮上预设点的初始位置;若鼠标光标发生移动,记录鼠标光标的终止位置,通过起始位置和终止位置,可以计算鼠标光标在x轴和y轴方向的位移差,则光标的移动方向可以作为触控类型,位移差可以作为触控变化量;若鼠标滚轮发生转动,则记录该预设点的滑动轨迹和终止位置,通过初始位置和终止位置,可以确定该滑动轨迹的轨迹长度,则滚轮的滚动方向可以作为触控类型,轨迹长度可以作为触控变化量。
然后,终端根据触控类型、以及预设的触控类型与转换类别的对应关系,确定视频角度转换指令对应的转换类别。例如,光标向左移动对应的转换类别为画面向左旋转,光标向右移动对应的转换类别为画面向右旋转,滚轮向上滚动对应的转换类别为画面放大,滚轮向下滚动对应的转换类别为画面缩小。再根据触控变化量、以及预设的触控变化量与转换幅度的对应关系,确定视频角度转换指令对应的转换幅度。例如,鼠标光标向左(或向右)移动的位移差为1cm,对应的转换幅度为向左(或向右)旋转30度;鼠标光标向左(或向右)移动的位移差为2cm,对应的转换幅度为向左(或向右)旋转60度;滚轮向上(或向下)滚动的轨迹长度为半圈,对应的放大(或缩小)比例为0.5倍;鼠标向上(或向下)滚动的轨迹长度为一圈,对应的放大(或缩小)比例为1倍等。
S302,根据转换类别和转换幅度对以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频。
具体地,终端根据转换类别和转换幅度,可以得知要对全景直播视频进行角度转换的方向和幅度。若转换类别为画面旋转,则终端可以直接根据旋转角度对全景直播视频进行角度转换即可,即转换幅度为向左旋转30度时,则直接将第一角度展示的全景直播视频绕Y轴向左旋转30度,得到第二角度展示的全景直播视频。若转换类别为画面放大或画面缩小,则终端需要根据第一角度以及放大比例或缩小比例计算出对应的第二角度,如当第一角度为在Y轴方向的60度视角,需要放大0.5倍时,对应的第二角度为(60+60×0.5)=90度,即最终以90度视角展示全景直播视频。
本实施例提供的直播视频的展示方法,终端通过解析视频角度转换指 令,确定该视频角度转换指令对应的转换类别和转换幅度,可以准确确定出观众想要对全景直播视频进行的视频角度;然后根据转换类别和转换幅度对第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频,便可以准确的展示出观众想要观看角度的全景直播视频,进一步提高了人机交互的智能性。
对于当前观众所观看的直播视频,在视频界面上还可以包括一些交互界面元素,如“送礼物”按钮、“点赞”按钮、“发言”框等;为进一步丰富直播视频的展示方式,终端还可以将这些交互界面元素和直播视频帧图像进行合成,得到以第一角度展示的全景直播视频,观众所观看到的交互界面元素可以呈现出“立体”的效果,并且当观众旋转或缩放全景直播视频时,交互界面元素也会发生相应的角度转换。可选地,上述S102可以包括:将直播视频帧和交互界面元素共同进行角度转换,得到第二角度的全景直播视频。
具体地,终端可以将交互界面元素的图像进行存储,并且记录每个交互界面元素的特征位置点(例如:在坐标0,0-100,100的区域是按钮位置),然后将交互界面元素的图像与直播视频帧图像进行合成,得到以第一角度展示的全景直播视频,即可以将交互界面元素的图像与直播视频帧图像看作一个整体图像。当终端响应于视频角度转换指令后,便可将直播视频帧和交互界面元素共同进行角度转换,交互界面元素随着直播视频帧的旋转/缩放也进行相应的旋转/缩放。由此,可进一步丰富直播视频的展示方式,大大提高了观众观看直播视频的观感性。
另外,当前的直播视频较多的是主播处于室内进行直播,其场景比较单一,因此,终端还可以接收观众的场景替换指令,将直播视频的场景进行替换,例如替换为雪山场景、森林场景、河流场景等。可选地,如图5所示,上述方法还可以包括:
S401,接收输入的场景替换指令;该场景替换指令携带目标场景标识。
具体地,观众可以通过终端外接的输入设备(如鼠标)输入上述场景替换指令,如点击鼠标的左键点击相应的场景图标;若终端的显示屏为触摸屏,观众也可以通过手指在触摸屏上触碰场景图标,以输入场景替换指令;观众还可以通过语音输入上述场景替换指令,如面向终端的麦克风语 音输入“雪山场景”等。那么,终端便可以通过上述方式接收到输入的场景替换指令,且该场景替换指令携带目标场景标识。
S402,根据目标场景标识,确定与场景替换指令对应的场景图像。
具体地,根据上述目标场景标识,终端可以从数据库中确定与该目标场景标识对应的场景图像的存储地址,其中,数据库中存储有不同场景标识与场景图像存储地址的对应关系;然后终端根据获得的存储地址,从该地址中获取对应的场景图像。
S403,将以第二角度展示的全景直播视频的直播背景替换为场景图像。
具体地,终端可以将场景图像与第二角度展示的全景直播视频的直播视频帧图像进行图像合成,以将全景直播视频的直播背景替换为上述场景图像。其中,终端可以将直播视频帧中的主播区域作为目标区域,其他区域作为背景区域,将背景区域替换为上述场景图像。
本实施例提供的直播视频的展示方法,终端可以接收输入的场景替换指令,根据该场景替换指令携带的目标场景标识确定场景图像,进而将以第二角度展示的全景直播视频的直播背景替换为该场景图像。由此观众可根据自己的场景喜好对全景直播视频的背景进行替换,以满足观众对展示方式的需求,进一步提高了人机交互的智能性。
上述实施例介绍了观众在全景直播视频展示过程中的一些交互操作,下面的实施例再对如何获取该全景直播视频进行详细描述。可选地,如图6所示,全景直播视频的获取方式包括:
S501,获取服务器发送的直播视频数据;该直播视频数据包括不同拍摄角度的视频数据。
具体地,该直播视频数据可以为由主播终端上传至服务器,并由观众终端从服务器中获取的视频流数据,其包括不同拍摄角度的视频数据,如主播通过主播终端拍摄的前后左右四个角度的视频数据。其中,该直播视频数据还可以携带全景标识,同一主播在同一时刻拍摄的不同角度的视频数据中全景标识相同,由此可确保将同一时刻的视频数据投影为同一全景直播视频。
可选地,上述直播视频数据为宽高比为2:1的视频流数据;由于每个主播所使用的主播终端参数不同,其上传的视频流数据可能不是2:1的尺 寸,那么观众终端接收到视频流数据后,还可以将其裁剪至2:1的尺寸,使得画面更加圆润平滑。
S502,对直播视频数据进行格式转换,得到RGB视频数据。
具体地,为提高直播视频数据的投影效率,本实施例可以采用GPU技术对直播视频数据进行数据处理,则终端需先将直播视频数据转换为GPU可处理的数据格式。
可选地,终端首先对直播视频数据进行解码,得到YUV视频数据;其中,YUV平面格式(planar)数据为色度在水平方向上的采样率减半的一种编码数据形式,即每行相邻的2个Y(亮度信息)共用一个U、V(色度信息)数据,多用于供网络端接收和/或播放端接收的图像数据,可以降低传输图像数据的频宽。然后终端根据预设的转换公式,对YUV视频数据进行格式转换,得到RGB视频数据;其中,该转换公式包括Y/U/V数据值之间的求和关系式和求差关系式。可选地,该转换公式可以为
Figure PCTCN2021086540-appb-000001
也可以为包含Y/U/V数据值之间的其他求和关系式和求差关系式。其中,终端内部可通过进程间通信(Inter-Process Communication,IPC)方式对直播视频数据进行传输,以进行解码、格式转换等处理过程。
可选地,终端还可以借助顶点着色器先创建三个纹理,分别填充Y、U、V视频数据,对于缺失的U、V采样点,可以利用显卡内置的双线性滤波算法进行插值,也可以采用其他图像插值算法,以达到更高的图像质量。这里可以通过Shader实现上述插值算法,然后分别把Y、U、V渲染到纹理,再通过转换公式将YUV视频数据转换为RGB视频数据。
可选地,在终端获取到服务器发送的直播视频数据后,还可以判断该直播视频数据是否携带全景标识、终端自身是否支持上述格式转换操作、以及终端自身是否支持全景投影处理;若上述条件全都满足,则执行S502的步骤;若其中有任一个条件不满足,则终端将此直播视频数据作为普通 视频进行处理,不转换为全景直播视频。由此进行兼容处理,大大提高处理直播视频数据的容错率。
S503,创建预设大小的3D球体模型,该3D球体模型用于RGB视频数据的坐标投影。
具体地,因终端需将上述直播视频数据映射为全景直播视频,该全景直播视频可看作一球体形式的视频,则终端还需创建一个预设大小的3D球体模型,用于RGB视频数据的坐标投影。其中,该3D球体模型的尺寸可以为半径为1、横向和竖向切面为128的3D球体。
S504,基于投影矩阵,将RGB视频数据投影至3D球体模型上,得到全景直播视频;其中,投影矩阵包括RGB视频数据中像素点与3D球体模型上像素点的坐标对应关系。
具体地,在得到RGB视频数据以及3D球体模型后,终端便可以基于投影矩阵,将RGB视频数据投影至3D球体模型上,得到全景直播视频。其中,本实施例可以采用D3D方法对RGB视频数据进行投影映射,D3D方法为一种用于在3D物体表面进行纹理贴图的方法,通过为视频图像的每个顶点指定一组纹理坐标,表明这些顶点在贴图面(即3D球体模型)中的坐标位置,以建立起视频图像与贴图面的映射关系。关于将RGB视频数据投影至3D球体模型的效果示意图可以参见图6a所示。
可选地,在进行投影过程中,假设3D球体模型中心为摄像点,摄像点的垂直观察范围为视角fovy(最大角度120度、最小角度60度、均值90度),np为近裁剪平面(设置为0.6),N是它到摄像点的距离,fp是远裁剪平面(设置为100),F是它到摄像点距离;则上述投影矩阵可以为
Figure PCTCN2021086540-appb-000002
其中,
Figure PCTCN2021086540-appb-000003
(x,y,z)为RGB视频数据中像素点的坐标,(x’,y’,z’)为3D球体模型上像素点的坐标。关于得到全景直播视频的工程处理流程可以参见图6b所示。
本实施例提供的直播视频的展示方法,终端首先获取服务器发送的直播视频数据,并对该直播视频数据进行格式转换,得到RGB视频数据,以使用GPU技术对RGB视频数据进行投影处理,大大提高了投影效率,使观众感知不到视频画面的停滞感。然后创建预设大小的3D球体模型,基于投影矩阵将RGB视频数据投影至3D球体模型上,得到全景直播视频。另外,本实施例中进行投影处理的步骤由观众终端所执行,不需要服务器执行后发送至观众终端,减少了服务器与观众终端之间的通信功耗。
在另一个实施例中,由服务器获取主播终端发送的不同拍摄角度的视频数据,对该视频数据进行格式转换,得到RGB视频数据;并基于投影矩阵将RGB视频数据投影至3D球体模型上,得到全景直播视频。当观众在观众终端上输入视频角度转换指令后,观众终端将视频角度转换指令发送至服务器,服务器根据该视频角度转换指令将第一角度的全景直播视频进行角度转换,得到第二角度的全景直播视频。然后,服务器对第一角度和第二角度分别所对应的视频区域进行比对,得到差异区域(即两个角度对应的视频区域的不同部分),将该差异区域发送至观众终端,观众终端根据该差异区域对原来以第一角度展示的全景直播视频进行截取融合,得到第二角度展示的全景直播视频。由此,可以大大减少服务器与观众终端进行交互的数据量,减少通信功耗,同时减少了观众终端的计算量。
另外,因直播视频数据通常为宽高比为2:1的视频流数据,由主播终端想服务器发送时视频数据或者服务器向观众终端发送视频数据时,可以将其进行压缩,观众终端接收到视频数据后再进行放大至2:1的尺寸。由此也可减小各终端与服务器之间通信所占用的带宽。
为更好的理解上述直播视频的展示方法的过程,下面以一个整体实施例方式对该方法进行介绍。如图7所示,该方法包括:
S601,检测输入设备在当前展示界面上的触控操作的类型;
S602,根据触控操作的类型,确定所述视频角度转换指令;
S603,根据输入设备的初始状态和终止状态,确定输入设备的触控类型和触控变化量;
S604,根据触控类型、以及预设的触控类型与转换类别的对应关系,确定视频角度转换指令对应的转换类别;
S605,根据触控变化量、以及预设的触控变化量与转换幅度的对应关系,确定视频角度转换指令对应的转换幅度;
S606,根据转换类别和转换幅度对以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;
S607,展示第二角度的全景直播视频。
关于各步骤的实现过程可以参见上述实施例的描述,其实现原理和技术效果类似,在此不再赘述。
应该理解的是,虽然图2-图7的流程图中各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,这些步骤可以以其它的顺序执行。而且,图2-图7中的至少一部分步骤可以包括多个子步骤或者多个阶段,这些子步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些子步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤的子步骤或者阶段的至少一部分轮流或者交替地执行。
在一个实施例中,如图8所示,提供了一种直播视频的展示装置,包括:接收模块21、角度转换模块22和展示模块23。
具体地,接收模块21,用于接收输入的视频角度转换指令。
角度转换模块22,用于响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中,全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的。
展示模块23,用于展示第二角度的全景直播视频。
本实施例提供的直播视频的展示装置,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。
在一个实施例中,接收模块21,具体用于检测输入设备在当前展示界面上的触控操作的类型;该触控操作的类型包括以下至少一种:针对第一角度展示的全景直播视频的角度旋转操作、针对第一角度展示的全景直播视频的画面缩放操作;根据触控操作的类型,确定视频角度转换指令。
在一个实施例中,角度转换模块22,具体用于解析视频角度转换指令, 确定该视频角度转换指令对应的转换类别和转换幅度;转换类别包括画面放大、画面缩小和画面旋转中的任一种,转换幅度包括放大比例、缩小比例和旋转角度中的任一种;根据转换类别和转换幅度对以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频。
在一个实施例中,角度转换模块22,具体用于根据输入设备的初始状态和终止状态,确定输入设备的触控类型和触控变化量;根据触控类型、以及预设的触控类型与转换类别的对应关系,确定视频角度转换指令对应的转换类别;根据触控变化量、以及预设的触控变化量与转换幅度的对应关系,确定视频角度转换指令对应的转换幅度。
在一个实施例中,以第一角度展示的全景直播视频包括合成的直播视频帧以及直播过程中的交互界面元素,角度转换模块22,具体用于将直播视频帧和交互界面元素共同进行角度转换,得到第二角度的全景直播视频。
在一个实施例中,如图9所示,在上述图8所示实施例的基础上,上述装置还包括确定模块24和替换模块25。
具体地,接收模块21,还用于接收输入的场景替换指令;该场景替换指令携带目标场景标识。
确定模块24,用于根据目标场景标识,确定与场景替换指令对应的场景图像。
替换模块25,用于将以第二角度展示的全景直播视频的直播背景替换为该场景图像。
本实施例提供的直播视频的展示装置,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。
在一个实施例中,接收模块21,具体用于检测预设的时间间隔内,所接收的触控操作的类型是否相同;若是,则根据触控操作的类型确定视频角度转换指令;若否,则不响应触控操作。
在一个实施例中,如图10所示,在上述图8或图9所示实施例的基础上,上述装置还包括:获取模块26、格式转换模块27、创建模块28和投影模块29。
具体地,获取模块26,用于获取服务器发送的直播视频数据;该直播视频数据包括不同拍摄角度的视频数据。
格式转换模块27,用于对直播视频数据进行格式转换,得到RGB视频数据。
创建模块28,用于创建预设大小的3D球体模型,该3D球体模型用于RGB视频数据的坐标投影。
投影模块29,用于基于投影矩阵,将RGB视频数据投影至3D球体模型上,得到全景直播视频;其中,投影矩阵包括RGB视频数据中像素点与3D球体模型上像素点的坐标对应关系。
本实施例提供的直播视频的展示装置,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。需要说明的是,图10为在图8所示实施例的基础上进行示出,此处仅为示例,其也可以在图9所示实施例的基础上进行示出。
在一个实施例中,格式转换模块27,具体用于对直播视频数据进行解码,得到YUV视频数据;根据预设的转换公式,对YUV视频数据进行格式转换,得到RGB视频数据;转换公式包括Y/U/V数据值之间的求和关系式和求差关系式。
在一个实施例中,上述装置还包括判断模块,用于判断直播视频数据是否携带全景标识、终端是否支持格式转换、以及终端是否支持全景投影处理;若直播视频数据携带全景标识、且终端支持格式转换、且终端支持全景投影处理,则指示格式转换模块27执行对直播视频数据进行格式转换的步骤。
本实施例提供的直播视频的展示装置,可以执行上述方法实施例,其实现原理和技术效果类似,在此不再赘述。
关于直播视频的展示装置的具体限定可以参见上文中对于直播视频的展示方法的限定,在此不再赘述。上述直播视频的展示装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于终端中的处理器中,也可以以软件形式存储于终端中的存储器中,以便于处理器调用执行以上各个模块对应的操作。
在一个实施例中,提供了一种终端,其内部结构图可以如图11所示。该终端包括通过系统总线连接的处理器、存储器、通信接口、显示屏和输入装置。其中,该终端的处理器用于提供计算和控制能力。该终端的存储 器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统和计算机程序。该内存储器为非易失性存储介质中的操作系统和计算机程序的运行提供环境。该终端的通信接口用于与外部的终端进行有线或无线方式的通信,无线方式可通过WIFI、运营商网络、NFC(近场通信)或其他技术实现。该计算机程序被处理器执行时以实现一种直播视频的展示方法。该终端的显示屏可以是液晶显示屏或者电子墨水显示屏,该终端的输入装置可以是显示屏上覆盖的触摸层,也可以是终端外壳上设置的按键、轨迹球或触控板,还可以是外接的键盘、触控板或鼠标等。
本领域技术人员可以理解,图11中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的终端的限定,具体的终端可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。
在一个实施例中,提供了一种终端,包括存储器和处理器,存储器中存储有计算机程序,该处理器执行计算机程序时实现以下步骤:
接收输入的视频角度转换指令;
响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中,全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的;
展示第二角度的全景直播视频。
本实施例提供的终端,其实现原理和技术效果与上述方法实施例类似,在此不再赘述。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
检测输入设备在当前展示界面上的触控操作的类型;该触控操作的类型包括以下至少一种:针对第一角度展示的全景直播视频的角度旋转操作、针对第一角度展示的全景直播视频的画面缩放操作;
根据触控操作的类型,确定视频角度转换指令。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
解析视频角度转换指令,确定该视频角度转换指令对应的转换类别和转换幅度;转换类别包括画面放大、画面缩小和画面旋转中的任一种,转换幅度包括放大比例、缩小比例和旋转角度中的任一种;
根据转换类别和转换幅度对以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
根据输入设备的初始状态和终止状态,确定输入设备的触控类型和触控变化量;
根据触控类型、以及预设的触控类型与转换类别的对应关系,确定视频角度转换指令对应的转换类别;
根据触控变化量、以及预设的触控变化量与转换幅度的对应关系,确定视频角度转换指令对应的转换幅度。
在一个实施例中,以第一角度展示的全景直播视频包括合成的直播视频帧以及直播过程中的交互界面元素,处理器执行计算机程序时还实现以下步骤:
将直播视频帧和交互界面元素共同进行角度转换,得到第二角度的全景直播视频。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
接收输入的场景替换指令;该场景替换指令携带目标场景标识;
根据目标场景标识,确定与场景替换指令对应的场景图像;
将以第二角度展示的全景直播视频的直播背景替换为该场景图像。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
检测预设的时间间隔内,所接收的触控操作的类型是否相同;
若是,则根据触控操作的类型确定视频角度转换指令;
若否,则不响应触控操作。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
获取服务器发送的直播视频数据;该直播视频数据包括不同拍摄角度的视频数据;
对直播视频数据进行格式转换,得到RGB视频数据;
创建预设大小的3D球体模型,该3D球体模型用于RGB视频数据的坐标投影;
基于投影矩阵,将RGB视频数据投影至3D球体模型上,得到全景直播视频;其中,投影矩阵包括RGB视频数据中像素点与3D球体模型上像 素点的坐标对应关系。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
对直播视频数据进行解码,得到YUV视频数据;
根据预设的转换公式,对YUV视频数据进行格式转换,得到RGB视频数据;转换公式包括Y/U/V数据值之间的求和关系式和求差关系式。
在一个实施例中,处理器执行计算机程序时还实现以下步骤:
判断直播视频数据是否携带全景标识、终端是否支持格式转换、以及终端是否支持全景投影处理;
若直播视频数据携带全景标识、且终端支持格式转换、且终端支持全景投影处理,则执行对直播视频数据进行格式转换的步骤。
在一个实施例中,提供了一种计算机可读存储介质,其上存储有计算机程序,计算机程序被处理器执行时实现以下步骤:
接收输入的视频角度转换指令;
响应于视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中,全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的;
展示第二角度的全景直播视频。
本实施例提供的计算机可读存储介质,其实现原理和技术效果与上述方法实施例类似,在此不再赘述。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
检测输入设备在当前展示界面上的触控操作的类型;该触控操作的类型包括以下至少一种:针对第一角度展示的全景直播视频的角度旋转操作、针对第一角度展示的全景直播视频的画面缩放操作;
根据触控操作的类型,确定视频角度转换指令。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
解析视频角度转换指令,确定该视频角度转换指令对应的转换类别和转换幅度;转换类别包括画面放大、画面缩小和画面旋转中的任一种,转换幅度包括放大比例、缩小比例和旋转角度中的任一种;
根据转换类别和转换幅度对以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
根据输入设备的初始状态和终止状态,确定输入设备的触控类型和触控变化量;
根据触控类型、以及预设的触控类型与转换类别的对应关系,确定视频角度转换指令对应的转换类别;
根据触控变化量、以及预设的触控变化量与转换幅度的对应关系,确定视频角度转换指令对应的转换幅度。
在一个实施例中,以第一角度展示的全景直播视频包括合成的直播视频帧以及直播过程中的交互界面元素,计算机程序被处理器执行时还实现以下步骤:
将直播视频帧和交互界面元素共同进行角度转换,得到第二角度的全景直播视频。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
接收输入的场景替换指令;该场景替换指令携带目标场景标识;
根据目标场景标识,确定与场景替换指令对应的场景图像;
将以第二角度展示的全景直播视频的直播背景替换为该场景图像。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
检测预设的时间间隔内,所接收的触控操作的类型是否相同;
若是,则根据触控操作的类型确定视频角度转换指令;
若否,则不响应触控操作。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
获取服务器发送的直播视频数据;该直播视频数据包括不同拍摄角度的视频数据;
对直播视频数据进行格式转换,得到RGB视频数据;
创建预设大小的3D球体模型,该3D球体模型用于RGB视频数据的坐标投影;
基于投影矩阵,将RGB视频数据投影至3D球体模型上,得到全景直播视频;其中,投影矩阵包括RGB视频数据中像素点与3D球体模型上像素点的坐标对应关系。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
对直播视频数据进行解码,得到YUV视频数据;
根据预设的转换公式,对YUV视频数据进行格式转换,得到RGB视频数据;转换公式包括Y/U/V数据值之间的求和关系式和求差关系式。
在一个实施例中,计算机程序被处理器执行时还实现以下步骤:
判断直播视频数据是否携带全景标识、终端是否支持格式转换、以及终端是否支持全景投影处理;
若直播视频数据携带全景标识、且终端支持格式转换、且终端支持全景投影处理,则执行对直播视频数据进行格式转换的步骤。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一非易失性计算机可读取存储介质中,该计算机程序在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和易失性存储器中的至少一种。非易失性存储器可包括只读存储器(Read-Only Memory,ROM)、磁带、软盘、闪存或光存储器等。易失性存储器可包括随机存取存储器(Random Access Memory,RAM)或外部高速缓冲存储器。作为说明而非局限,RAM可以是多种形式,比如静态随机存取存储器(Static Random Access Memory,SRAM)或动态随机存取存储器(Dynamic Random Access Memory,DRAM)等。
以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。
以上所述实施例仅表达了本申请的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本申请构思的前提下,还可以做出若干变形和改进,这些都属于本申请的保护范围。因此,本申请专利的保护范围应以所附权利要求为准。

Claims (13)

  1. 一种直播视频的展示方法,其特征在于,所述方法包括:
    接收输入的视频角度转换指令;
    响应于所述视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中,所述全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的;
    展示所述第二角度的全景直播视频。
  2. 根据权利要求1所述的方法,其特征在于,所述接收输入的视频角度转换指令,包括:
    检测输入设备在当前展示界面上的触控操作的类型;所述触控操作的类型包括以下至少一种:针对所述第一角度展示的全景直播视频的角度旋转操作、针对所述第一角度展示的全景直播视频的画面缩放操作;
    根据所述触控操作的类型,确定所述视频角度转换指令。
  3. 根据权利要求2所述的方法,其特征在于,所述响应于所述视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频,包括:
    解析所述视频角度转换指令,确定所述视频角度转换指令对应的转换类别和转换幅度;所述转换类别包括画面放大、画面缩小和画面旋转中的任一种,所述转换幅度包括放大比例、缩小比例和旋转角度中的任一种;
    根据所述转换类别和所述转换幅度对所述以第一角度展示的全景直播视频进行角度转换,得到所述第二角度的全景直播视频。
  4. 根据权利要求3所述的方法,其特征在于,所述解析所述视频角度转换指令,确定所述视频角度转换指令对应的转换类别和转换幅度,包括:
    根据所述输入设备的初始状态和终止状态,确定所述输入设备的触控类型和触控变化量;
    根据所述触控类型、以及预设的触控类型与转换类别的对应关系,确定所述视频角度转换指令对应的转换类别;
    根据所述触控变化量、以及预设的触控变化量与转换幅度的对应关系,确定所述视频角度转换指令对应的转换幅度。
  5. 根据权利要求1所述的方法,其特征在于,所述以第一角度展示的全景直播视频包括合成的直播视频帧以及直播过程中的交互界面元素,所述将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频,包括:
    将所述直播视频帧和所述交互界面元素共同进行角度转换,得到所述第二角度的全景直播视频。
  6. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    接收输入的场景替换指令;所述场景替换指令携带目标场景标识;
    根据所述目标场景标识,确定与所述场景替换指令对应的场景图像;
    将以所述第二角度展示的全景直播视频的直播背景替换为所述场景图像。
  7. 根据权利要求2所述的方法,其特征在于,所述根据所述触控操作的类型,确定所述视频角度转换指令,包括:
    检测预设的时间间隔内,所接收的触控操作的类型是否相同;
    若是,则根据所述触控操作的类型确定所述视频角度转换指令;
    若否,则不响应所述触控操作。
  8. 根据权利要求1所述的方法,其特征在于,所述全景直播视频的获取方式包括:
    获取服务器发送的直播视频数据;所述直播视频数据包括不同拍摄角度的视频数据;
    对所述直播视频数据进行格式转换,得到RGB视频数据;
    创建预设大小的3D球体模型,所述3D球体模型用于所述RGB视频数据的坐标投影;
    基于投影矩阵,将所述RGB视频数据投影至所述3D球体模型上,得到所述全景直播视频;其中,所述投影矩阵包括RGB视频数据中像素点与3D球体模型上像素点的坐标对应关系。
  9. 根据权利要求8所述的方法,其特征在于,所述对所述直播视频数据进行格式转换,得到RGB视频数据,包括:
    对所述直播视频数据进行解码,得到YUV视频数据;
    根据预设的转换公式,对所述YUV视频数据进行格式转换,得到所 述RGB视频数据;所述转换公式包括Y/U/V数据值之间的求和关系式和求差关系式。
  10. 根据权利要求8或9所述的方法,其特征在于,在所述获取服务器发送的直播视频数据之后,所述方法还包括:
    判断所述直播视频数据是否携带全景标识、所述终端是否支持所述格式转换、以及所述终端是否支持全景投影处理;
    若所述直播视频数据携带全景标识、且所述终端支持所述格式转换、且所述终端支持全景投影处理,则执行对所述直播视频数据进行格式转换的步骤。
  11. 一种直播视频的展示装置,其特征在于,所述装置包括:
    接收模块,用于接收输入的视频角度转换指令;
    角度转换模块,用于响应于所述视频角度转换指令,将当前以第一角度展示的全景直播视频进行角度转换,得到第二角度的全景直播视频;其中,所述全景直播视频为采用预设的3D球体模型对不同拍摄角度的视频数据进行投影处理得到的;
    展示模块,用于展示所述第二角度的全景直播视频。
  12. 一种终端,包括存储器和处理器,所述存储器存储有计算机程序,其特征在于,所述处理器执行所述计算机程序时实现权利要求1-10中任一项所述方法的步骤。
  13. 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1-10中任一项所述方法的步骤。
PCT/CN2021/086540 2020-05-29 2021-04-12 直播视频的展示方法、装置、终端和可读存储介质 WO2021238454A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010476780.2 2020-05-29
CN202010476780.2A CN111641871A (zh) 2020-05-29 2020-05-29 直播视频的展示方法、装置、终端和可读存储介质

Publications (1)

Publication Number Publication Date
WO2021238454A1 true WO2021238454A1 (zh) 2021-12-02

Family

ID=72332227

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/086540 WO2021238454A1 (zh) 2020-05-29 2021-04-12 直播视频的展示方法、装置、终端和可读存储介质

Country Status (2)

Country Link
CN (1) CN111641871A (zh)
WO (1) WO2021238454A1 (zh)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114222188A (zh) * 2021-12-28 2022-03-22 深圳小湃科技有限公司 基于旋转屏的全屏显示方法、装置、设备及存储介质
CN115052168A (zh) * 2022-05-31 2022-09-13 北京达佳互联信息技术有限公司 一种标签获取方法、装置、服务器及存储介质
WO2023103875A1 (zh) * 2021-12-10 2023-06-15 华为技术有限公司 自由视角视频的视角切换方法、装置及系统
CN116744028A (zh) * 2023-02-21 2023-09-12 深圳市华方信息产业有限公司 一种手术室的示教方法、系统、终端及存储介质

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111641871A (zh) * 2020-05-29 2020-09-08 广州华多网络科技有限公司 直播视频的展示方法、装置、终端和可读存储介质
CN113489935A (zh) * 2020-09-11 2021-10-08 青岛海信电子产业控股股份有限公司 一种视频交互方法及设备
CN112306344B (zh) * 2020-10-19 2023-11-28 武汉中科通达高新技术股份有限公司 一种数据处理方法及移动终端
CN113411684B (zh) * 2021-06-24 2023-05-30 广州酷狗计算机科技有限公司 视频播放方法和装置、存储介质及电子设备
CN116233493A (zh) * 2021-09-29 2023-06-06 腾讯科技(深圳)有限公司 沉浸媒体的数据处理方法、装置、设备及可读存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6559846B1 (en) * 2000-07-07 2003-05-06 Microsoft Corporation System and process for viewing panoramic video
CN105979242A (zh) * 2015-11-23 2016-09-28 乐视网信息技术(北京)股份有限公司 一种视频的播放方法和装置
CN106454401A (zh) * 2016-10-26 2017-02-22 乐视网信息技术(北京)股份有限公司 一种播放视频的方法及装置
CN106598428A (zh) * 2016-11-29 2017-04-26 宇龙计算机通信科技(深圳)有限公司 播放全景视频的方法、系统及终端设备
CN108174265A (zh) * 2016-12-07 2018-06-15 华为技术有限公司 一种360度全景视频的播放方法、装置及系统
CN111641871A (zh) * 2020-05-29 2020-09-08 广州华多网络科技有限公司 直播视频的展示方法、装置、终端和可读存储介质

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6559846B1 (en) * 2000-07-07 2003-05-06 Microsoft Corporation System and process for viewing panoramic video
CN105979242A (zh) * 2015-11-23 2016-09-28 乐视网信息技术(北京)股份有限公司 一种视频的播放方法和装置
CN106454401A (zh) * 2016-10-26 2017-02-22 乐视网信息技术(北京)股份有限公司 一种播放视频的方法及装置
CN106598428A (zh) * 2016-11-29 2017-04-26 宇龙计算机通信科技(深圳)有限公司 播放全景视频的方法、系统及终端设备
CN108174265A (zh) * 2016-12-07 2018-06-15 华为技术有限公司 一种360度全景视频的播放方法、装置及系统
CN111641871A (zh) * 2020-05-29 2020-09-08 广州华多网络科技有限公司 直播视频的展示方法、装置、终端和可读存储介质

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023103875A1 (zh) * 2021-12-10 2023-06-15 华为技术有限公司 自由视角视频的视角切换方法、装置及系统
CN114222188A (zh) * 2021-12-28 2022-03-22 深圳小湃科技有限公司 基于旋转屏的全屏显示方法、装置、设备及存储介质
CN115052168A (zh) * 2022-05-31 2022-09-13 北京达佳互联信息技术有限公司 一种标签获取方法、装置、服务器及存储介质
CN115052168B (zh) * 2022-05-31 2024-03-19 北京达佳互联信息技术有限公司 一种标签获取方法、装置、服务器及存储介质
CN116744028A (zh) * 2023-02-21 2023-09-12 深圳市华方信息产业有限公司 一种手术室的示教方法、系统、终端及存储介质

Also Published As

Publication number Publication date
CN111641871A (zh) 2020-09-08

Similar Documents

Publication Publication Date Title
WO2021238454A1 (zh) 直播视频的展示方法、装置、终端和可读存储介质
US11303881B2 (en) Method and client for playing back panoramic video
US11941748B2 (en) Lightweight view dependent rendering system for mobile devices
US11736801B2 (en) Merging webcam signals from multiple cameras
JP6044328B2 (ja) 画像処理システム、画像処理方法およびプログラム
US10437545B2 (en) Apparatus, system, and method for controlling display, and recording medium
CN112204993B (zh) 使用重叠的被分区的分段的自适应全景视频流式传输
US20120293613A1 (en) System and method for capturing and editing panoramic images
US10235795B2 (en) Methods of compressing a texture image and image data processing system and methods of generating a 360 degree panoramic video thereof
EP3804349B1 (en) Adaptive panoramic video streaming using composite pictures
JP5743016B2 (ja) 画像を生成する装置および方法
US10650592B2 (en) Methods and apparatus for providing rotated spherical viewpoints
US9996311B2 (en) Efficient communication interface for casting interactively controlled visual content
WO2022141781A1 (zh) 一种播放全景视频的方法、系统、存储介质及播放设备
CN113286138A (zh) 一种全景视频显示方法及显示设备
JP2018033107A (ja) 動画の配信装置及び配信方法
JP6394682B2 (ja) 方法および画像処理装置
CN112887695A (zh) 一种全景图共享处理方法、系统以及终端
JP6688367B1 (ja) 動画配信装置、端末、動画配信システム、動画配信方法、及びプログラム
JP7302647B2 (ja) 画像処理システム、画像処理方法およびプログラム
US20240022688A1 (en) Multiuser teleconferencing with spotlight feature
US11720314B1 (en) Electronic devices and corresponding methods for altering content to meet input parameters of a companion device
Borgeat et al. Collaborative visualization and interaction for detailed environment models
CN118678228A (zh) 全景会议控制方法、装置、计算机设备
CN116208725A (zh) 视频处理方法、电子设备和存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21813988

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21813988

Country of ref document: EP

Kind code of ref document: A1