WO2021114710A1 - 直播视频互动方法、装置以及计算机设备 - Google Patents
直播视频互动方法、装置以及计算机设备 Download PDFInfo
- Publication number
- WO2021114710A1 WO2021114710A1 PCT/CN2020/109871 CN2020109871W WO2021114710A1 WO 2021114710 A1 WO2021114710 A1 WO 2021114710A1 CN 2020109871 W CN2020109871 W CN 2020109871W WO 2021114710 A1 WO2021114710 A1 WO 2021114710A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- special effect
- gesture
- gesture action
- face
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 53
- 230000003993 interaction Effects 0.000 title claims abstract description 39
- 230000000694 effects Effects 0.000 claims abstract description 195
- 230000001960 triggered effect Effects 0.000 claims abstract description 7
- 230000009471 action Effects 0.000 claims description 154
- 230000002452 interceptive effect Effects 0.000 claims description 25
- 238000012544 monitoring process Methods 0.000 claims description 14
- 230000002708 enhancing effect Effects 0.000 abstract description 3
- 230000008569 process Effects 0.000 description 11
- 238000010586 diagram Methods 0.000 description 9
- 230000006399 behavior Effects 0.000 description 6
- 230000008859 change Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000003780 insertion Methods 0.000 description 4
- 230000037431 insertion Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 235000001808 Ceanothus spinosus Nutrition 0.000 description 2
- 241001264786 Ceanothus spinosus Species 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000008094 contradictory effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/4223—Cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2187—Live feed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
- H04N21/4316—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47205—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
- H04N21/4788—Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
Definitions
- This application relates to the field of multimedia technology, and in particular to a live video interaction method, device, computer equipment, and computer-readable storage medium.
- the inventor realizes that at present, the display of interactive behaviors such as gift giving, likes, and cll is generally displayed in real-time in the form of graphics and text in the chat public screen area or a fixed area on the live screen.
- the display method is relatively simple and the degree of interaction is limited. .
- This application proposes a live video interaction method, device, computer equipment, and computer readable storage medium, which are used to solve the problem of a single display mode when a user interacts with an anchor during a live broadcast and a limited sense of interaction and participation.
- this application provides a live video interaction method, the method includes:
- the placing the host video and the user picture in the same video play frame for display includes: placing the host video and the user picture in the first part and the second part of the video play frame Seamless splicing playback.
- the displaying the first video special effect in the video playback frame includes: acquiring the first palm position of the first gesture action, and the first person of the live host's face in the host video Face position; the video special effect is extended and displayed with the first palm position as the starting point and the first face position as the ending point.
- the method further includes: monitoring whether the palm of the first gesture action moves; when the palm of the first gesture action moves, acquiring the first hand of the first gesture action after the movement. Second-hand palm position; changing the starting point of the video special effect from the first palm position to the second palm position.
- the method further includes: monitoring whether the live broadcast host’s face moves; when the live broadcast host’s face moves, acquiring a second face position of the live broadcast host’s face after the movement; and uploading the video The end point of the special effect is changed from the first face position to the second face position.
- the displaying the video special effect in the video playback frame includes: acquiring the face image of the live broadcast host in the host video; and corresponding the face image to the first gesture action
- the human face special effects are synthesized and displayed; wherein the human face special effects include at least one of clothing special effects and expression special effects.
- the method further includes: counting the gesture duration of the first gesture action; when the gesture duration is greater than or equal to a preset time threshold, triggering a second video special effect.
- this application also provides a live video interactive device, the device including:
- the acquisition module is used to obtain the host video of the live broadcast host and the user picture captured by the user terminal in real time; the display module is used to place the host video and the user picture in the same video playback frame for display; the identification module is used to obtain And recognize the user's first gesture action in the user screen; the judgment module is used to compare the first gesture action with the second gesture action in the preset gesture special effect correspondence table to determine whether they are consistent Query module, configured to query the video special effect corresponding to the second gesture action in the gesture special effect correspondence table when the first gesture action is consistent with the second gesture action; the display module, It is also used to display the video special effect in the video play frame.
- this application also proposes a computer device, the computer device includes a memory and a processor, the memory stores computer-readable instructions that can run on the processor, and the computer-readable instructions are The following steps are implemented when the processor is executed:
- the present application also provides a computer-readable storage medium, the computer-readable storage medium stores computer-readable instructions, and the computer-readable instructions can be executed by at least one processor to enable The at least one processor performs the following steps:
- the live video interaction method, device, computer equipment, and computer-readable storage medium proposed in this application acquire the host video of the live broadcast host and the user picture captured by the user terminal in real time, and place them in the same video playback frame for display; then, obtain and identify The first gesture action of the user in the user screen is compared with the second gesture action in the preset gesture special effect correspondence table; when the first gesture action is compared with the second gesture action When the two gesture actions are consistent, the video special effect corresponding to the second gesture action is queried in the gesture special effect correspondence table, and displayed in the video playback frame.
- the user and the host screen can be displayed in the same interface, and the interactive behavior between the user and the host can be connected by video special effects, which enriches the interactive display methods and strengthens the sense of interaction and participation.
- FIG. 1 is a scene diagram of an optional application environment of Embodiment 1 of the present application.
- FIG. 2 is a schematic flowchart of an embodiment of a live video interaction method according to the present application
- Fig. 3 is a flowchart of an exemplary embodiment of displaying the first video special effect to the video play frame in step S208 of Fig. 2;
- Figure 4 is an effect diagram of the extended display of video special effects on the live screen
- FIG. 5 is a schematic flowchart of an exemplary embodiment based on FIG. 3;
- FIG. 6 is a schematic flowchart of another exemplary embodiment based on FIG. 3;
- FIG. 7 is a flowchart of another exemplary embodiment of displaying the first video special effect to the video playback frame in step S208 in FIG. 2;
- Figure 8 is an effect diagram of the live screen before face special effects
- Figure 9 is an effect diagram of the live screen after face special effects
- Fig. 10 is a schematic flowchart of an exemplary embodiment based on Fig. 2;
- FIG. 11 is a schematic diagram of an optional hardware architecture of the computer device of the present application.
- FIG. 12 is a schematic diagram of program modules of an embodiment of the live video interactive device of the present application.
- Fig. 1 is a schematic diagram of an application environment of an embodiment of the present application.
- the computer device 1 is connected to the live broadcast main terminal 2 and can receive and obtain the host video data sent by the live broadcast main terminal 2.
- the computer device 1 as a user terminal may be an independent electronic device with a camera function, such as a mobile phone, a portable device, a PC, and so on. It may also be that an independent functional module is attached to the main electronic device with camera function, and then helps the main electronic device to realize the function of live video interaction.
- Fig. 2 is a schematic flowchart of an embodiment of a live video interaction method according to the present application. It can be understood that the flowchart in this method embodiment is not used to limit the order of execution of the steps. The following is an exemplary description with computer equipment as the execution subject.
- the live video interaction method may include steps S200 to S208, where:
- Step S200 Obtain the host video of the live broadcast host and the user picture captured by the user terminal in real time.
- the computer device 1 is connected to the live broadcast master terminal 2 via a network, and then can obtain the anchor video sent by the live broadcast master terminal 2, and the anchor video is the live broadcast host through the live broadcast master terminal 2 Video data for live shooting.
- the computer device 1 is used as a user's terminal device and is provided with a camera unit, which can take a user picture of the user.
- the computer device 1 can obtain the host video of the live broadcast host and the user picture of the user.
- Step S202 Put the anchor video and the user screen in the same video play frame for display.
- the computer device 1 further provides a display interface, and a video playback frame is set on the display interface, and the video playback frame is used to play video data.
- the video playing frame includes a first part and a second part. The first part is used to display the anchor video, and the second part is used to display the user picture, so that the obtained anchor video and user picture are passed through the video playing frame. Play it.
- the computer device 1 seamlessly splices and plays the first part and the second part of the anchor video and the user screen in the video playback frame, such as upper and lower parts or left and right parts. Two parts.
- the computer device 1 may set the first part and the second part to be spliced up and down when the host video is in a horizontal screen, and set the first part and the second part to be spliced left and right when the host video is in a vertical screen.
- the host video and the user screen are played together in the video playback frame in other ways, for example, the host video and the user are scattered in the video playback frame separately, or
- the anchor video and the user picture form a play window of a size that is included and played together, and there is no restriction here.
- the video of the user and the live broadcast host can be displayed in the same frame, so the user can feel close to the live broadcast host.
- Step S204 Acquire and recognize the user's first gesture action in the user screen, and compare the first gesture action with a second gesture action in a preset gesture special effect correspondence table.
- Step S206 When the first gesture action is consistent with the second gesture action, the first video special effect corresponding to the second gesture action is queried in the gesture special effect correspondence table.
- Step S208 Display the first video special effect in the video play frame.
- the computer device 1 displays the user screen and the host video in the video playback frame, it recognizes each frame of the user screen by means of image recognition, so as to obtain all the images in the user screen. Describe the palm image in each frame of the user's screen, and then recognize the gesture action.
- the computer device 1 may pre-store a gesture recognition model capable of recognizing gesture actions, and then obtain images in the video data of the user screen for recognition.
- the computer device recognizes the first gesture action in the user screen, it compares the first gesture action with the second gesture action in the preset gesture special effect correspondence table.
- the gesture special effect correspondence table includes a one-to-one correspondence between gesture actions and video special effects.
- the second gesture action corresponding to the second gesture action is further queried in the gesture special effect correspondence table.
- the first video special effects Finally, the computer device 1 displays the first video special effect in the video playback frame.
- the process of displaying the first video special effect in the video playing frame in step S208 includes steps S300 to S302.
- Step S300 Obtain the first palm position of the first gesture action and the first human face position of the live broadcast host's face in the host video.
- step S302 the video special effect is extended and displayed with the first palm position as the starting point and the first human face position as the ending point.
- the computer device 1 queries the first video special effect corresponding to the second gesture action, it displays the video special effect on the user screen and the host video.
- the computer device 1 first obtains the first palm position of the user performing the first gesture action in the user screen, and then obtains the first face position of the live broadcaster in the anchor video.
- the human body recognition model can identify all the positions. The palm image that performs the first gesture action in the user screen frame of the user screen, and the face image of the live broadcast host in the host video frame of the host video is recognized.
- the computer device 1 marks the first palm position according to the relative position coordinates of the palm image in the user screen frame; and marks the relative position coordinates of the face image in the active video frame as In the first face position, the relative position coordinates indicate the position of the palm image in the user screen frame or the position of the face image in the host video frame.
- the computer device 1 displays the video special effect with the first palm position as the starting point and the first human face position as the ending point for extended display. Please refer to Figure 4, the user gestures to make a "comparison”, and then the video special effect generates a series of red hearts, which extend from the user's screen to the host video. In this way, the computer device 1 displays the video special effects in the form of "interaction" on the user screen and the anchor video, thereby enhancing the sense of interaction in the live video process.
- the process of displaying the first video special effect in the video playback frame in step S208 includes steps S400 to S404 in addition to steps S300 to S302.
- Step S400 monitoring whether the palm of the first gesture action has moved.
- Step S402 when the palm of the first gesture action moves, acquire the second palm position of the palm of the first gesture action after the movement.
- Step S404 Changing the starting point of the video special effect from the first palm position to the second palm position.
- the computer device 1 will continue to monitor the first palm position as the start point and the first face position as the end point for the extended display of the video special effect. Whether the palm of a gesture action moves; when it moves, obtain the second palm position of the palm after the movement, and then change the starting point of the video effect from the first palm position to the second palm position .
- the process of displaying the first video special effect in the video playback frame in step S208 includes steps S300-S302 and steps S500-S504. .
- step S500 it is monitored whether the live broadcast host's face moves.
- Step S502 When the face of the live broadcast master moves, acquire a second face position of the live broadcast master's face after the movement.
- Step S504 Change the end point of the video special effect from the first face position to the second face position.
- the computer device 1 will continue to monitor the live broadcast during the extended display of the video special effect with the first palm position as the starting point and the first face position as the ending point. Whether the main face has moved; when it has moved, obtain the second face position of the moved face, and then change the end point of the video special effect from the first face position to the second face position Face position.
- the start and end points of the video special effects can be adjusted. Therefore, the "targeted" extension of the video special effects in the live video interaction can be improved and strengthened.
- the interactive effect By monitoring the position of the palm of the first gesture and the position of the face of the live broadcaster, the start and end points of the video special effects can be adjusted. Therefore, the "targeted" extension of the video special effects in the live video interaction can be improved and strengthened. The interactive effect.
- the process of displaying the first video special effect in the video playing frame in step S208 includes steps S600 to S602.
- Step S600 Obtain the face image of the live broadcast host in the host video.
- step S602 the face image and the face special effect corresponding to the first gesture action are synthesized and displayed; wherein the face special effect includes at least one of a clothing special effect and an expression special effect.
- the computer device 1 recognizes the first gesture action and determines that it is consistent with the second gesture action in the gesture special effect correspondence table, and therefore triggers the video special effect corresponding to the first gesture action
- the video special effect includes combining and displaying a preset face special effect with the face image of the current live broadcaster in the host video
- the face special effect includes a clothing special effect or an expression special effect.
- Fig. 8 is a normal live broadcast picture
- Fig. 9 is a live broadcast picture after face special effects.
- the user's screen below produces the first gesture of coin insertion, and the first gesture of coin insertion can trigger the face effect of adding the current face image of the live broadcaster to the sunglasses. Therefore, the computer device 1
- the face image of the live broadcaster with face effects will be displayed in the host video above. This is triggered by the user to perform video special effects on the face image of the live broadcaster, so the interactive effect can also be improved.
- the live video playback method includes steps S700-S702 in addition to the above-mentioned steps S200-S208:
- Step S700 Count the gesture duration of the first gesture action.
- Step S702 When the duration of the gesture is greater than or equal to a preset time threshold, trigger a second video special effect.
- the computer device 1 displays the video special effect in the video playback frame, it will continue to monitor the gesture duration of the first gesture action.
- the duration of the gesture is greater than or equal to the preset threshold, the second video special effect is triggered.
- the second video special effect may also be based on further enhancement of the video special effect, such as changes in color, shape, and the like. In this way, the effect of live video interaction is further enhanced.
- the live video interaction method proposed in this embodiment can obtain the host video of the live broadcast host and the user screen shot by the user terminal in real time, and place them in the same video playback frame for display; then, obtain and identify the user screen
- the user’s first gesture action is compared with the second gesture action in the preset gesture special effect correspondence table; when the first gesture action is consistent with the second gesture action
- the video special effect corresponding to the second gesture action is queried in the gesture special effect correspondence table, and displayed in the video playback frame.
- this application also provides a computer device.
- FIG. 11 is a schematic diagram of an optional hardware architecture of the computer device of this application.
- the computer device 1 may include, but is not limited to, a memory 11, a processor 12, and a network interface 13 that can communicate with each other through a system bus.
- the computer device 1 is connected to a network through a network interface 13 (not shown in FIG. 11), and connected to a live broadcast main terminal (not shown in FIG. 11) through the network for data interaction.
- the network may be Intranet, Internet, Global System of Mobile Communication (GSM), Wideband Code Division Multiple Access (WCDMA), 4G network, 5G Network, Bluetooth (Bluetooth), Wi-Fi, call network and other wireless or wired networks.
- GSM Global System of Mobile Communication
- WCDMA Wideband Code Division Multiple Access
- 4G network Fifth Generation
- 5G Network Fifth Generation
- Bluetooth Bluetooth
- Wi-Fi call network and other wireless or wired networks.
- FIG. 11 only shows the computer device 1 with components 11-13, but it should be understood that it is not required to implement all the illustrated components, and more or fewer components may be implemented instead.
- the memory 11 includes at least one type of readable storage medium
- the readable storage medium includes flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), random access memory (RAM) , Static random access memory (SRAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), programmable read only memory (PROM), magnetic memory, magnetic disks, optical disks, etc.
- the memory 11 may be an internal storage unit of the computer device 1, for example, a hard disk or a memory of the computer device 1.
- the memory 11 may also be an external storage device of the computer device 1, such as a plug-in hard disk, a smart media card (SMC), and a secure digital ( Secure Digital, SD card, Flash Card, etc.
- the memory 11 may also include both the internal storage unit of the computer device 1 and its external storage device.
- the memory 11 is generally used to store the operating system and various application software installed in the computer device 1, such as the program code of the baffle application, and the program code of the live video interactive device 200.
- the memory 11 can also be used to temporarily store various types of data that have been output or will be output.
- the processor 12 may be a central processing unit (CPU), a controller, a microcontroller, a microprocessor, or other data processing chips.
- the processor 12 is generally used to control the overall operation of the computer device 1, such as performing data interaction or communication-related control and processing.
- the processor 12 is used to run the program code or process data stored in the memory 11, for example, to run the application program of the live video interactive device 200, which is not limited here.
- the network interface 13 may include a wireless network interface or a wired network interface, and the network interface 13 is usually used to establish a communication connection between the computer device 1 and the live broadcast main terminal.
- a live video interactive device 200 when installed and running in the computer device 1, when the live video interactive device 200 is running, it can acquire the host video of the live host and the user picture captured by the user terminal in real time, and put it Display in the same video playback frame; then obtain and recognize the user's first gesture action in the user screen, and compare the first gesture action with the second gesture action in the preset gesture special effect correspondence table Yes; when the first gesture action is consistent with the second gesture action, the video special effect corresponding to the second gesture action is queried in the gesture special effect correspondence table and displayed in the video playback frame.
- the user and the host screen can be displayed in the same interface, and the interactive behavior between the user and the host can be connected by video special effects, which enriches the interactive display methods and strengthens the sense of interaction and participation.
- FIG. 12 is a program module diagram of an embodiment of the live video interactive device 200 of the present application.
- the live video interaction device 200 includes a series of computer-readable instructions stored on the memory 11, and when the computer-readable instructions are executed by the processor 12, the live video interaction of the embodiment of the present application can be realized.
- the live video interactive device 200 may be divided into one or more modules based on specific operations implemented by various parts of the computer-readable instructions. For example, in FIG. 12, the live video interactive device 200 may be divided into an acquisition module 201, a display module 202, an identification module 203, a judgment module 204, a query module 205, and a monitoring module 206. among them:
- the acquisition module 201 is used to acquire the host video of the live broadcast host and the user picture captured by the user terminal in real time.
- the computer device is connected to the live broadcast master terminal through the network, and then can obtain the anchor video sent by the live broadcast master terminal, where the anchor video is the live broadcast shooting performed by the live broadcast master through the live broadcast master terminal Video data.
- the computer device as the user's terminal device, is provided with a camera unit, which can capture the user's user picture. Therefore, the acquisition module 201 can acquire the host video of the live broadcast host and the user picture of the user.
- the display module 202 is configured to display the anchor video and the user screen in the same video playback frame.
- the computer device further provides a display interface, and a video playback frame is set on the display interface, and the video playback frame is used to play video data.
- the video playback frame includes a first part and a second part. The first part is used to display the anchor video, and the second part is used to display the user screen.
- the display module 202 can pass the obtained anchor video and user screen through all the screens. Said video play box to play.
- the display module 202 seamlessly splices the first part and the second part of the anchor video and the user screen in the video playback frame, such as upper and lower parts or left and right parts. Two parts.
- the computer device 1 may set the first part and the second part to be spliced up and down when the host video is in a horizontal screen, and set the first part and the second part to be spliced left and right when the host video is in a vertical screen.
- the display module 202 can also play the host video and the user screen together in the video playback frame in other ways, for example, the host video and the user are scattered separately In the video playback frame, or the host video and the user screen form a size playback window that is included and played together, there is no limitation here.
- the video of the user and the live broadcast host can be displayed in the same frame, so the user can feel close to the live broadcast host.
- the recognition module 203 is configured to obtain and recognize the user's first gesture action in the user screen.
- the judgment module 204 is configured to compare the first gesture action with the second gesture action in the preset gesture special effect correspondence table to determine whether they are consistent.
- the query module 205 is configured to query the video special effect corresponding to the second gesture action in the gesture special effect correspondence table when the first gesture action is consistent with the second gesture action.
- the display module 202 is also used to display the video special effect in the video play frame.
- the recognition module 203 performs image recognition on each frame of the user screen. Recognition, so as to obtain the palm image in each frame of the image of the user, and then recognize the gesture action.
- the computer device may pre-store a gesture recognition model capable of recognizing gesture actions, and then the recognition module 203 may call the gesture recognition model to recognize images in the video data obtained from the user screen.
- the recognition module 203 recognizes the first gesture action in the user screen
- the judgment module 204 will perform the first gesture action with the second gesture action in the preset gesture special effect correspondence table. Compare and judge whether they are consistent.
- the gesture special effect correspondence table includes a one-to-one correspondence between gesture actions and video special effects.
- the query module 205 further queries the gesture special effect correspondence table.
- the display module 202 displays the first video special effect in the video play frame.
- the process of the display module 202 displaying the first video special effect in the video playback frame includes: obtaining the first palm position of the first gesture action, and the live broadcast in the host video The first face position of the owner's face; the video special effect is extended and displayed with the first palm position as the starting point and the first human face position as the ending point.
- the display module 202 first obtains the first palm position of the user performing the first gesture action in the user screen, and then obtains the first face position of the live broadcaster in the anchor video.
- the human body recognition model can identify all the positions. The palm image that performs the first gesture action in the user screen frame of the user screen, and the face image of the live broadcast host in the host video frame of the host video is recognized.
- the display module 202 marks the first palm position according to the relative position coordinates of the palm image in the user frame; and marks the relative position coordinates of the face image in the active video frame as In the first face position, the relative position coordinates indicate the position of the palm image in the user screen frame or the position of the face image in the host video frame.
- the display module 202 displays the video special effects with the first palm position as the starting point and the first human face position as the ending point for extended display.
- the user gestures to make a "comparison” and then the video special effect generates a series of red hearts, which extend from the user's screen to the host video.
- the computer device 1 displays the video special effects in the form of "interaction" on the user screen and the anchor video, thereby enhancing the sense of interaction in the live video process.
- the monitoring module 206 is configured to monitor whether the palm of the first gesture action moves when the display module 202 displays the first video special effect in the video playback frame.
- the display module 202 is further configured to obtain the second palm position of the palm of the first gesture motion after the movement when the palm of the first gesture motion moves, and then set the starting point of the video special effect from the The first palm position is changed to the second palm position.
- the monitoring module 206 is further configured to monitor whether the face of the live broadcast host moves during the process of the display module 202 displaying the first video special effect to the video playback frame.
- the display module 202 is further configured to obtain the second face position of the live broadcast host’s face after the movement when the live broadcast host’s face moves, and to obtain the second palm position of the palm of the first gesture action after the move. , And then change the end point of the video special effect from the first face position to the second face position.
- the start and end points of the video special effects can be adjusted. Therefore, the "targeted" extension of the video special effects in the live video interaction can be improved and strengthened.
- the interactive effect By monitoring the position of the palm of the first gesture and the position of the face of the live broadcaster, the start and end points of the video special effects can be adjusted. Therefore, the "targeted" extension of the video special effects in the live video interaction can be improved and strengthened. The interactive effect.
- the display module 202 is further configured to obtain the face image of the live broadcast host in the host video, and then synthesize the face image with the face special effect corresponding to the first gesture action And display; wherein, the face special effects include at least one of clothing special effects and expression special effects.
- the computer device pre-setting the video special effects includes: synthesizing and displaying the preset face special effects with the face image of the current live broadcaster in the host video, and the face special effects include clothing special effects Or emoji effects.
- Fig. 8 is a normal live broadcast picture
- Fig. 9 is a live broadcast picture after face special effects.
- the user screen below produces the first gesture of coin insertion, and the first gesture of coin insertion can trigger the addition of the current face image of the live broadcaster to the face special effects of sunglasses. Therefore, the display module 202 will display the face image of the live broadcaster with face effects in the host video above. This is triggered by the user to perform video special effects on the face image of the live broadcaster, so the interactive effect can also be improved.
- the monitoring module 206 is also used to count the gesture duration of the first gesture action.
- the display module 202 is further configured to trigger a second video special effect when the duration of the gesture is greater than or equal to a preset time threshold.
- the monitoring module 206 continues to monitor the gesture duration of the first gesture action.
- the display module 202 triggers the display of the second video special effect.
- the second video special effect may also be based on further enhancement of the video special effect, such as changes in color, shape, and the like. In this way, the effect of live video interaction is further enhanced.
- the computer device can obtain the host video of the live broadcast host and the user picture captured by the user terminal in real time, and place them in the same video playback frame for display; then obtain and recognize the user's first gesture in the user picture Action, compare the first gesture action with the second gesture action in the preset gesture special effect correspondence table; when the first gesture action is consistent with the second gesture action, in the gesture
- the special effect correspondence table queries the video special effect corresponding to the second gesture action, and displays it in the video play frame.
- the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to make a terminal device (which can be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present application, including:
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Marketing (AREA)
- Image Analysis (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
本申请公开了一种直播视频互动方法,该方法包括:获取直播主的主播视频和用户终端实时拍摄的用户画面,投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,查询出所述第二手势动作对应的视频特效,显示到所述视频播放框。本申请还提供一种直播视频互动装置、计算机设备以及计算机可读存储介质。本申请能够将主播视频和用户画面进行同框显示,根据用户画面中的手势动作触发视频特效,显示到包括用户画面和主播视频的视频播放框,从而加强了直播时主播与用户之间的互动。
Description
本申请要求于2019年12月09日提交中国专利局、申请号为201911251115.7、发明名称为“直播视频互动方法、装置以及计算机设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
本申请涉及多媒体技术领域,尤其涉及一种直播视频互动方法、装置、计算机设备及计算机可读存储介质。
随着互联网技术的快速发展,视频直播由于具备直观、快速、交互性强的特点已经成为一种非常广泛的娱乐方式,受到了大众的喜爱。在视频直播的过程中,用户与直播主的常见的互动行为包括送礼,点赞,打cll;丰富的礼物展示形式更能有效促进优质直播内容的产出。
发明人意识到,目前,送礼,点赞,打cll等互动行为的展示一般是在聊天公屏区域或直播画面上的固定区域以图文形式实时展示,其展示的方式较为单一,互动程度有限。
发明内容
本申请提出一种直播视频互动方法、装置、计算机设备及计算机可读存储介质,用于解决直播过程中用户跟主播进行互动时展示的方式较为单一,互动参与感有限的问题。
首先,为实现上述目的,本申请提供一种直播视频互动方法,所述方法包括:
获取直播主的主播视频和用户终端实时拍摄的用户画面;将所述主播视频和所述用户画面投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效;将所述第一视频特效显示到所述视频播放框。
在一个例子中,所述将所述主播视频和所述用户画面投放在同一个视频播放框显示包括:将所述主播视频和所述用户画面在所述视频播放框中第一部分和第二部分无缝拼接播放。
在一个例子中,所述将所述第一视频特效显示到所述视频播放框包括:获取所述第一手势动作的第一手掌位置,以及所述主播视频中的直播主人脸的第一人脸位置;将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。
在一个例子中,所述方法还包括:监测所述第一手势动作的手掌是否发生移动;当所述第一手势动作的手掌发生移动时,获取移动后所述第一手势动作的手掌的第二手掌位置;将所述视频特效的起点从所述第一手掌位置变更为所述第二手掌位置。
在一个例子中,所述方法还包括:监测所述直播主人脸是否发生移动;当所述直播主人脸发生移动时,获取移动后所述直播主人脸的第二人脸位置;将所述视频特效的终点从所述第一人脸位置变更为所述第二人脸位置。
在一个例子中,所述将所述视频特效显示到所述视频播放框包括:获取所述主播视频中的直播主的人脸图像;将所述人脸图像与所述第一手势动作对应的人脸特效进行合成并显示;其中,所述人脸特效包括服饰特效和表情特效中的至少一种。
在一个例子中,所述方法还包括:统计所述第一手势动作的手势持续时间;当所述手势持续时间大于或等于预设的时间阈值时,触发第二视频特效。
此外,为实现上述目的,本申请还提供一种直播视频互动装置,所述装置包括:
获取模块,用于获取直播主的主播视频和用户终端实时拍摄的用户画面;显示模块,用于将所述主播视频和所述用户画面投放在同一个视频播放框显示;识别模块,用于获取并识别出所述用户画面中的用户的第一手势动作;判断模块,用于将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对,判断是否一致;查询模块,用于当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的视频特效;所述显示模块,还用于将所述视频特效显示到所述视频播放框。
进一步地,本申请还提出一种计算机设备,所述计算机设备包括存储器、处理器,所述存储器上存储有可在所述处理器上运行的计算机可读指令,所述计算机可读指令被所述处理器执行时实现以下步骤:
获取直播主的主播视频和用户终端实时拍摄的用户画面;将所述主播视频和所述用户画面投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效;将所述第一视频特效显示到所述视频播放框。
进一步地,为实现上述目的,本申请还提供一种计算机可读存储介质,所述计算机可读存储介质存储有计算机可读指令,所述计算机可读指令可被至少一个处理器执行,以使 所述至少一个处理器执行以下步骤:
获取直播主的主播视频和用户终端实时拍摄的用户画面;将所述主播视频和所述用户画面投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效;将所述第一视频特效显示到所述视频播放框。
本申请所提出的直播视频互动方法、装置、计算机设备及计算机可读存储介质,获取直播主的主播视频和用户终端实时拍摄的用户画面,投放在同一个视频播放框显示;然后获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的视频特效,显示到所述视频播放框。通过以上方式,能够在同一界面内展示用户与主播画面,利用视频特效连接用户与主播之间的互动行为,丰富了互动展示的方式,加强了互动参与感。
图1是本申请实施例一可选的应用环境的情景图;
图2是本申请直播视频互动方法一实施例的流程示意图;
图3是图2步骤S208中的将所述第一视频特效显示到所述视频播放框的一示例性实施例的流程图;
图4是是视频特效在直播画面延伸显示的效果图;
图5是基于图3的一示例性实施例的流程示意图;
图6是基于图3的另一示例性实施例的流程示意图;
图7是图2步骤S208中的将所述第一视频特效显示到所述视频播放框的另一示例性实施例的流程图;
图8是人脸特效之前的直播画面效果图;
图9是人脸特效之后的直播画面效果图;
图10是基于图2的一示例性实施例的流程示意图;
图11是本申请计算机设备一可选的硬件架构的示意图;
图12是本申请直播视频互动装置一实施例的程序模块示意图。
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
需要说明的是,在本申请中涉及“第一”、“第二”等的描述仅用于描述目的,而不能理解为指示或暗示其相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括至少一个该特征。另外,各个实施例之间的技术方案可以相互结合,但是必须是以本领域普通技术人员能够实现为基础,当技术方案的结合出现相互矛盾或无法实现时应当认为这种技术方案的结合不存在,也不在本申请要求的保护范围之内。
图1是本申请一实施例的应用环境示意图。参阅图1所示,所述计算机设备1与直播主终端2连接,能够接收并获取到由所述直播主终端2发送过来的主播视频数据。在本实施例中,所述计算机设备1作为用户终端,可以是具有摄像功能的独立电子设备,比如手机、便携设备,PC机等。也可以是一个独立的功能模块附加在具备有摄像功能的主电子设备上,然后帮所述主电子设备实现直播视频互动的功能。
图2是本申请直播视频互动方法一实施例的流程示意图。可以理解,本方法实施例中的流程图不用于对执行步骤的顺序进行限定。下面以计算机设备为执行主体进行示例性描述。
如图2所示,所述直播视频互动方法可以包括步骤S200~S208,其中:
步骤S200,获取直播主的主播视频和用户终端实时拍摄的用户画面。
具体地,所述计算机设备1通过网络与所述直播主终端2连接,然后则可以获取由所述直播主终端2发送的主播视频,所述主播视频是所述直播主通过所述直播主终端2进行直播拍摄的视频数据。当然,所述计算机设备1作为用户的终端设备,设置有摄像单元,能够拍摄用户的用户画面。也就是说,所述计算机设备1能够获取到直播主的主播视频和用户的用户画面。
步骤S202,将所述主播视频和所述用户画面投放在同一个视频播放框显示。
具体地,所述计算机设备1还提供一个显示界面,所述显示界面上设置一个视频播放框,所述视频播放框用于播放视频数据。本实施例中,视频播放框包括第一部分与第二部分,第一部分用于展示主播视频,第二部分用于展示用户画面,从而实现将获取到的主播视频和用户画面通过所述视频播放框进行播放。
需要注意的是,在本实施例中,所述计算机设备1将所述主播视频和所述用户画面在 所述视频播放框中第一部分和第二部分无缝拼接播放,比如上下两部分或者左右两部分。如,计算机设备1可以在主播视频为横屏时,设置第一部分与第二部分为上下拼接,在主播视频为竖屏时,设置第一部分与第二部分为左右拼接。当然,在其他实施例中,所述主播视频与所述用户画面在所述视频播放框中以其他方式共同播放,比如,所述主播视频与所述用户分开散落在所述视频播放框,或者所述主播视频和所述用户画面形成大小播放窗包含在一起播放,这里不做限制。在本实施例中,将主播视频和用户画面进行拼接播放,可以将用户和直播主的视频同框显示,因此可以给用户一种与直播主近距离的感觉。
步骤S204,获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对。
步骤S206,当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效。
步骤S208,将所述第一视频特效显示到所述视频播放框。
具体地,所述计算机设备1将所述用户画面和所述主播视频显示到所述视频播放框之后,则通过图像识别的方式对所述用户画面中的每一帧图像进行识别,从而获取所述用户画面中的每一帧图像中的手掌图像,然后识别手势动作。在本实施例中,所述计算机设备1可以预先存储一个能够识别手势动作的手势识别模型,然后获取所述用户画面的视频数据中的图像进行识别。当所述计算机设备识别出所述用户画面中的第一手势动作时,则会将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对。其中,所述手势特效对应表包括手势动作与视频特效的一一对应关系。
当所述计算机设备1识别出的第一手势动作与所述手势特效对应表中的第二手势动作一致时,则进一步在所述手势特效对应表中查询出所述第二手势动作对应的第一视频特效。最后,所述计算机设备1将所述第一视频特效显示到所述视频播放框。
如图3所示,在一示例性的实施例中,步骤S208中的将所述第一视频特效显示到所述视频播放框的过程,包括步骤S300~S302。
步骤S300,获取所述第一手势动作的第一手掌位置,以及所述主播视频中的直播主人脸的第一人脸位置。
步骤S302,将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。
具体地,所述计算机设备1在查询出所述第二手势动作对应的第一视频特效之后,则将所述视频特效显示到所述用户画面和所述主播视频。所述计算机设备1先获取所述用户画面中的用户执行所述第一手势动作的第一手掌位置,然后再获取所述主播视频中直播主 的第一人脸位置。在本实施例中,由于所述用户画面和所述主播视频在所述视频播放框中属于相对的固定位置关系(即短时间内位置固定不变),因此,可以通过人体识别模型识别出所述用户画面的用户画面帧中的执行所述第一手势动作的手掌图像,以及识别出所述主播视频的主播视频帧中的直播主的人脸图像。
接着,所述计算机设备1根据所述手掌图像在所述用户画面帧的相对位置坐标标记为所述第一手掌位置;以及根据所述人脸图像在所述主动视频帧的相对位置坐标标记为所述第一人脸位置,其中,相对位置坐标表示手掌图像在所述用户画面帧中的位置或者所述人脸图像在所述主播视频帧中的位置。最后,所述计算机设备1将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。请参阅图4所示,用户手势比划出“比心”,然后视频特效产生一连串红心,从用户画面延伸显示到主播视频。通过这种方式,所述计算机设备1将视频特效以“互动”形式显示在所述用户画面和所述主播视频中,从而提升了直播视频过程中的互动感。
如图5所示,在一示例性的实施例中,步骤S208中的将所述第一视频特效显示到所述视频播放框的过程,除了包括步骤S300~S302,还包括步骤S400~S404。
步骤S400,监测所述第一手势动作的手掌是否发生移动。
步骤S402,当所述第一手势动作的手掌发生移动时,获取移动后所述第一手势动作的手掌的第二手掌位置。
步骤S404,将所述视频特效的起点从所述第一手掌位置变更为所述第二手掌位置。
在本实施例中,所述计算机设备1在将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示的过程中,还会持续监测所述第一手势动作的手掌是否发生移动;当发生移动时,则获取移动后所述手掌的第二手掌位置,然后将所述视频特效的起点由所述第一手掌位置变更为所述第二手掌位置。
如图6所示,在另一示例性的实施例中,步骤S208中的将所述第一视频特效显示到所述视频播放框的过程,除了包括步骤S300~S302,还包括步骤S500~S504。
步骤S500,监测所述直播主人脸是否发生移动。
步骤S502,当所述直播主人脸发生移动时,获取移动后所述直播主人脸的第二人脸位置。
步骤S504,将所述视频特效的终点从所述第一人脸位置变更为所述第二人脸位置。
在本实施例中,所述计算机设备1在将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示的过程中,还会持续监测所述直播主的人脸是否发生移动;当发生移动时,则获取移动后所述人脸的第二人脸位置,然后将所述视频特效的终 点由所述第一人脸位置变更为所述第二人脸位置。
通过监测所述第一手势的手掌的位置以及所述直播主的人脸的位置从而调整所述视频特效的起点和终点,因此,可以提升直播视频互动中视频特效的“针对性”延伸,加强了互动效果。
如图7所示,在一示例性的实施例中,步骤S208中的将所述第一视频特效显示到所述视频播放框的过程,包括步骤S600~S602。
步骤S600,获取所述主播视频中的直播主的人脸图像。
步骤S602,将所述人脸图像与所述第一手势动作对应的人脸特效进行合成并显示;其中,所述人脸特效包括服饰特效和表情特效中的至少一种。
在本实施例中,所述计算机设备1识别出所述第一手势动作并判断为与所述手势特效对应表中的第二手势动作一致,因此触发所述第一手势动作对应的视频特效,其中,所述视频特效包括,将预设的人脸特效与所述主播视频中当前的直播主的人脸图像进合成并显示,人脸特效包括服饰特效或表情特效。例如,图8为正常的直播画面,图9是人脸特效之后的直播画面。如图9所示,下方的用户画面产生了投币的第一手势,而投币的第一手势能够触发将直播主当前的人脸图像加上墨镜的人脸特效,因此,所述计算机设备1则会在上方的主播视频中显示加上人脸特效的直播主的人脸图像。这种由用户触发,对直播主的人脸图像进行视频特效,因此,也能够很好提升互动效果。
如图10所示,在一示例性的实施例中,所述直播视频播放方法除了包括以上所述的步骤S200~S208,还包括步骤S700~S702:
步骤S700,统计所述第一手势动作的手势持续时间。
步骤S702,当所述手势持续时间大于或等于预设的时间阈值时,触发第二视频特效。
具体地,所述计算机设备1在将所述视频特效显示到所述视频播放框之后,还会继续监控所述第一手势动作的手势持续时间。当所述手势持续时间大于或等于预设阈值时,则触发第二视频特效。当然,所述第二视频特效也可以是基于所述视频特效的进一步加强,比如颜色、形状等变化。通过这种方式进一步加强直播视频互动的效果。
从上文可知,本实施例所提出的直播视频互动方法能够获取直播主的主播视频和用户终端实时拍摄的用户画面,投放在同一个视频播放框显示;然后获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的视频特效,显示到所述视频播放框。通过以上方式,能够在同一界面内展示用户与主播画面,利用视频特效连接用户与主播之间的互动行为,丰富 了互动展示的方式,加强了互动参与感。
此外,本申请还提供一种计算机设备,参阅图11所示,是本申请计算机设备一可选的硬件架构的示意图。
本实施例中,所述计算机设备1可包括,但不仅限于,可通过系统总线相互通信连接存储器11、处理器12、网络接口13。所述计算机设备1通过网络接口13连接网络(图11未标出),通过网络连接到直播主终端等(图11未标出)进行数据交互。所述网络可以是企业内部网(Intranet)、互联网(Internet)、全球移动通讯系统(Global System of Mobile communication,GSM)、宽带码分多址(Wideband Code Division Multiple Access,WCDMA)、4G网络、5G网络、蓝牙(Bluetooth)、Wi-Fi、通话网络等无线或有线网络。
需要指出的是,图11仅示出了具有组件11-13的计算机设备1,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。
其中,所述存储器11至少包括一种类型的可读存储介质,所述可读存储介质包括闪存、硬盘、多媒体卡、卡型存储器(例如,SD或DX存储器等)、随机访问存储器(RAM)、静态随机访问存储器(SRAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、可编程只读存储器(PROM)、磁性存储器、磁盘、光盘等。在一些实施例中,所述存储器11可以是所述计算机设备1的内部存储单元,例如该计算机设备1的硬盘或内存。在另一些实施例中,所述存储器11也可以是所述计算机设备1的外部存储设备,例如该计算机设备1配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。当然,所述存储器11还可以既包括所述计算机设备1的内部存储单元也包括其外部存储设备。本实施例中,所述存储器11通常用于存储安装于所述计算机设备1的操作系统和各类应用软件,例如挡板应用的程序代码,以及直播视频互动装置200的程序代码等。此外,所述存储器11还可以用于暂时地存储已经输出或者将要输出的各类数据。
所述处理器12在一些实施例中可以是中央处理器(Central Processing Unit,CPU)、控制器、微控制器、微处理器、或其他数据处理芯片。该处理器12通常用于控制所述计算机设备1的总体操作,例如执行数据交互或者通信相关的控制和处理等。本实施例中,所述处理器12用于运行所述存储器11中存储的程序代码或者处理数据,例如运行所述直播视频互动装置200的应用程序,这里不做限制。
所述网络接口13可包括无线网络接口或有线网络接口,该网络接口13通常用于在所述计算机设备1与直播主终端之间建立通信连接。
本实施例中,所述计算机设备1内安装并运行有直播视频互动装置200时,当所述直 播视频互动装置200运行时,能够获取直播主的主播视频和用户终端实时拍摄的用户画面,投放在同一个视频播放框显示;然后获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的视频特效,显示到所述视频播放框。通过以上方式,能够在同一界面内展示用户与主播画面,利用视频特效连接用户与主播之间的互动行为,丰富了互动展示的方式,加强了互动参与感。
至此,己经详细介绍了本申请计算机设备的硬件结构和功能。下面,将基于上述计算机设备,提出本申请的各个实施例。
参阅图12所示,是本申请直播视频互动装置200一实施例的程序模块图。
本实施例中,所述直播视频互动装置200包括一系列的存储于存储器11上的计算机可读指令,当该计算机可读指令被处理器12执行时,可以实现本申请实施例的直播视频互动功能。在一些实施例中,基于该计算机可读指令各部分所实现的特定的操作,直播视频互动装置200可以被划分为一个或多个模块。例如,在图12中,所述直播视频互动装置200可以被分割成获取模块201、显示模块202、识别模块203、判断模块204、查询模块205和监测模块206。其中:
所述获取模块201,用于获取直播主的主播视频和用户终端实时拍摄的用户画面。
具体地,所述计算机设备通过网络与所述直播主终端连接,然后则可以获取由所述直播主终端发送的主播视频,所述主播视频是所述直播主通过所述直播主终端进行直播拍摄的视频数据。当然,所述计算机设备作为用户的终端设备,设置有摄像单元,能够拍摄用户的用户画面。因此,所述获取模块201能够获取到直播主的主播视频和用户的用户画面。
所述显示模块202,用于将所述主播视频和所述用户画面投放在同一个视频播放框显示。
具体地,所述计算机设备还提供一个显示界面,所述显示界面上设置一个视频播放框,所述视频播放框用于播放视频数据。本实施例中,视频播放框包括第一部分与第二部分,第一部分用于展示主播视频,第二部分用于展示用户画面,所述显示模块202能够将获取到的主播视频和用户画面通过所述视频播放框进行播放。
需要注意的是,在本实施例中,所述显示模块202将所述主播视频和所述用户画面在所述视频播放框中第一部分和第二部分无缝拼接播放,比如上下两部分或者左右两部分。如,计算机设备1可以在主播视频为横屏时,设置第一部分与第二部分为上下拼接,在主播视频为竖屏时,设置第一部分与第二部分为左右拼接。当然,在其他实施例中,所述显 示模块202也可以将所述主播视频与所述用户画面在所述视频播放框中以其他方式共同播放,比如,所述主播视频与所述用户分开散落在所述视频播放框,或者所述主播视频和所述用户画面形成大小播放窗包含在一起播放,这里不做限制。在本实施例中,将主播视频和用户画面进行拼接播放,可以将用户和直播主的视频同框显示,因此可以给用户一种与直播主近距离的感觉。
所述识别模块203,用于获取并识别出所述用户画面中的用户的第一手势动作。
所述判断模块204,用于将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对,判断是否一致。
所述查询模块205,用于当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的视频特效。
所述显示模块202,还用于将所述视频特效显示到所述视频播放框。
具体地,所述显示模块202将所述用户画面和所述主播视频显示到所述视频播放框之后,所述识别模块203则通过图像识别的方式对所述用户画面中的每一帧图像进行识别,从而获取所述用户画面中的每一帧图像中的手掌图像,然后识别手势动作。在本实施例中,所述计算机设备可以预先存储一个能够识别手势动作的手势识别模型,然后所述识别模块203可以调用该手势识别模型对获取所述用户画面的视频数据中的图像进行识别。当所述识别模块203识别出所述用户画面中的第一手势动作时,所述判断模块204则会将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对,判断是否一致。其中,所述手势特效对应表包括手势动作与视频特效的一一对应关系。
当所述判断模块204判断出所述第一手势动作与所述手势特效对应表中的第二手势动作一致时,所述查询模块205则进一步在所述手势特效对应表中查询出所述第二手势动作对应的第一视频特效。最后,所述显示模块202将所述第一视频特效显示到所述视频播放框。
在本实施例中,所述显示模块202将所述第一视频特效显示到所述视频播放框的过程包括:获取所述第一手势动作的第一手掌位置,以及所述主播视频中的直播主人脸的第一人脸位置;将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。
具体地,所述显示模块202先获取所述用户画面中的用户执行所述第一手势动作的第一手掌位置,然后再获取所述主播视频中直播主的第一人脸位置。在本实施例中,由于所述用户画面和所述主播视频在所述视频播放框中属于相对的固定位置关系(即短时间内位置固定不变),因此,可以通过人体识别模型识别出所述用户画面的用户画面帧中的执行所 述第一手势动作的手掌图像,以及识别出所述主播视频的主播视频帧中的直播主的人脸图像。
接着,所述显示模块202根据所述手掌图像在所述用户画面帧的相对位置坐标标记为所述第一手掌位置;以及根据所述人脸图像在所述主动视频帧的相对位置坐标标记为所述第一人脸位置,其中,相对位置坐标表示手掌图像在所述用户画面帧中的位置或者所述人脸图像在所述主播视频帧中的位置。最后,所述显示模块202将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。请参阅图4所示,用户手势比划出“比心”,然后视频特效产生一连串红心,从用户画面延伸显示到主播视频。通过这种方式,所述计算机设备1将视频特效以“互动”形式显示在所述用户画面和所述主播视频中,从而提升了直播视频过程中的互动感。
所述监测模块206,用于在所述显示模块202将所述第一视频特效显示到所述视频播放框的过程中监测所述第一手势动作的手掌是否发生移动。
所述显示模块202,还用于当所述第一手势动作的手掌发生移动时,获取移动后所述第一手势动作的手掌的第二手掌位置,然后将所述视频特效的起点从所述第一手掌位置变更为所述第二手掌位置。
所述监测模块206,还用于在所述显示模块202将所述第一视频特效显示到所述视频播放框的过程中监测所述直播主人脸是否发生移动。
所述显示模块202,还用于当所述直播主人脸发生移动时,获取移动后所述直播主人脸的第二人脸位置,获取移动后所述第一手势动作的手掌的第二手掌位置,然后将所述视频特效的终点从所述第一人脸位置变更为所述第二人脸位置。
通过监测所述第一手势的手掌的位置以及所述直播主的人脸的位置从而调整所述视频特效的起点和终点,因此,可以提升直播视频互动中视频特效的“针对性”延伸,加强了互动效果。
在另一实施例中,所述显示模块202还用于获取所述主播视频中的直播主的人脸图像,然后将所述人脸图像与所述第一手势动作对应的人脸特效进行合成并显示;其中,所述人脸特效包括服饰特效和表情特效中的至少一种。
在本实施例中,所述计算机设备预先设置所述视频特效包括:将预设的人脸特效与所述主播视频中当前的直播主的人脸图像进合成并显示,人脸特效包括服饰特效或表情特效。例如,图8为正常的直播画面,图9是人脸特效之后的直播画面。如图9所示,下方的用户画面产生了投币的第一手势,而投币的第一手势能够触发将直播主当前的人脸图像加上墨镜的人脸特效,因此,所述显示模块202则会在上方的主播视频中显示加上人脸特效的 直播主的人脸图像。这种由用户触发,对直播主的人脸图像进行视频特效,因此,也能够很好提升互动效果。
当然,在另一实施例中,所述监测模块206还用于统计所述第一手势动作的手势持续时间。
所述显示模块202,还用于当所述手势持续时间大于或等于预设的时间阈值时,触发第二视频特效。
具体地,所述显示模块202在将所述视频特效显示到所述视频播放框之后,所述监测模块206则会继续监控所述第一手势动作的手势持续时间。当所述监测模块206统计出所述手势持续时间大于或等于预设阈值时,那么,所述显示模块202则触发显示第二视频特效。当然,所述第二视频特效也可以是基于所述视频特效的进一步加强,比如颜色、形状等变化。通过这种方式进一步加强直播视频互动的效果。
从上文可知,所述计算机设备能够获取直播主的主播视频和用户终端实时拍摄的用户画面,投放在同一个视频播放框显示;然后获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的视频特效,显示到所述视频播放框。通过以上方式,能够在同一界面内展示用户与主播画面,利用视频特效连接用户与主播之间的互动行为,丰富了互动展示的方式,加强了互动参与感。
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本申请各个实施例所述的方法,包括:
获取直播主的主播视频和用户终端实时拍摄的用户画面;将所述主播视频和所述用户画面投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效;将所述第一视频特效显示到所述视频播放框。
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说 明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。
Claims (20)
- 一种直播视频互动方法,所述方法包括步骤:获取直播主的主播视频和用户终端实时拍摄的用户画面;将所述主播视频和所述用户画面投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效;将所述第一视频特效显示到所述视频播放框。
- 如权利要求1所述的直播视频互动方法,所述将所述主播视频和所述用户画面投放在同一个视频播放框显示包括:将所述主播视频和所述用户画面在所述视频播放框中第一部分和第二部分无缝拼接播放。
- 如权利要求1所述的直播视频互动方法,所述将所述第一视频特效显示到所述视频播放框包括:获取所述第一手势动作的第一手掌位置,以及所述主播视频中的直播主人脸的第一人脸位置;将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。
- 如权利要求3所述的直播视频互动方法,所述方法还包括:监测所述第一手势动作的手掌是否发生移动;当所述第一手势动作的手掌发生移动时,获取移动后所述第一手势动作的手掌的第二手掌位置;将所述视频特效的起点从所述第一手掌位置变更为所述第二手掌位置。
- 如权利要求3所述的直播视频互动方法,所述方法还包括:监测所述直播主人脸是否发生移动;当所述直播主人脸发生移动时,获取移动后所述直播主人脸的第二人脸位置;将所述视频特效的终点从所述第一人脸位置变更为所述第二人脸位置。
- 如权利要求1所述的直播视频互动方法,所述将所述视频特效显示到所述视频播放框包括:获取所述主播视频中的直播主的人脸图像;将所述人脸图像与所述第一手势动作对应的人脸特效进行合成并显示;其中,所述人 脸特效包括服饰特效和表情特效中的至少一种。
- 如权利要求1所述的直播视频互动方法,所述方法还包括:统计所述第一手势动作的手势持续时间;当所述手势持续时间大于或等于预设的时间阈值时,触发第二视频特效。
- 一种直播视频互动装置,所述装置包括:获取模块,用于获取直播主的主播视频和用户终端实时拍摄的用户画面;显示模块,用于将所述主播视频和所述用户画面投放在同一个视频播放框显示;识别模块,用于获取并识别出所述用户画面中的用户的第一手势动作;判断模块,用于将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对,判断是否一致;查询模块,用于当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的视频特效;所述显示模块,还用于将所述视频特效显示到所述视频播放框。
- 一种计算机设备,所述计算机设备包括存储器、处理器,所述存储器上存储有可在所述处理器上运行的计算机可读指令,所述计算机可读指令被所述处理器执行时实现以下步骤:获取直播主的主播视频和用户终端实时拍摄的用户画面;将所述主播视频和所述用户画面投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效;将所述第一视频特效显示到所述视频播放框。
- 如权利要求9所述的计算机设备,所述将所述主播视频和所述用户画面投放在同一个视频播放框显示包括:将所述主播视频和所述用户画面在所述视频播放框中第一部分和第二部分无缝拼接播放。
- 如权利要求9所述的计算机设备,所述将所述第一视频特效显示到所述视频播放框包括:获取所述第一手势动作的第一手掌位置,以及所述主播视频中的直播主人脸的第一人脸位置;将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。
- 如权利要求11所述的计算机设备,所述计算机可读指令被所述处理器执行时还实现以下步骤:监测所述第一手势动作的手掌是否发生移动;当所述第一手势动作的手掌发生移动时,获取移动后所述第一手势动作的手掌的第二手掌位置;将所述视频特效的起点从所述第一手掌位置变更为所述第二手掌位置。
- 如权利要求11所述的计算机设备,所述计算机可读指令被所述处理器执行时还实现以下步骤:监测所述直播主人脸是否发生移动;当所述直播主人脸发生移动时,获取移动后所述直播主人脸的第二人脸位置;将所述视频特效的终点从所述第一人脸位置变更为所述第二人脸位置。
- 如权利要求9所述的计算机设备,所述将所述视频特效显示到所述视频播放框包括:获取所述主播视频中的直播主的人脸图像;将所述人脸图像与所述第一手势动作对应的人脸特效进行合成并显示;其中,所述人脸特效包括服饰特效和表情特效中的至少一种。
- 如权利要求9所述的计算机设备,所述计算机可读指令被所述处理器执行时还实现以下步骤:统计所述第一手势动作的手势持续时间;当所述手势持续时间大于或等于预设的时间阈值时,触发第二视频特效。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机可读指令,所述计算机可读指令可被至少一个处理器执行,以使所述至少一个处理器执行以下步骤:获取直播主的主播视频和用户终端实时拍摄的用户画面;将所述主播视频和所述用户画面投放在同一个视频播放框显示;获取并识别出所述用户画面中的用户的第一手势动作,将所述第一手势动作与预设的手势特效对应表中的第二手势动作进行比对;当所述第一手势动作与所述第二手势动作一致时,在所述手势特效对应表查询出所述第二手势动作对应的第一视频特效;将所述第一视频特效显示到所述视频播放框。
- 如权利要求16所述的计算机可读存储介质,所述将所述主播视频和所述用户画面 投放在同一个视频播放框显示包括:将所述主播视频和所述用户画面在所述视频播放框中第一部分和第二部分无缝拼接播放。
- 如权利要求16所述的计算机可读存储介质,所述将所述第一视频特效显示到所述视频播放框包括:获取所述第一手势动作的第一手掌位置,以及所述主播视频中的直播主人脸的第一人脸位置;将所述视频特效以所述第一手掌位置为起点、所述第一人脸位置为终点进行延伸显示。
- 如权利要求16所述的计算机可读存储介质,所述将所述视频特效显示到所述视频播放框包括:获取所述主播视频中的直播主的人脸图像;将所述人脸图像与所述第一手势动作对应的人脸特效进行合成并显示;其中,所述人脸特效包括服饰特效和表情特效中的至少一种。
- 如权利要求18所述的计算机可读存储介质,所述计算机可读指令被所述处理器执行,以使所述至少一个处理器还实现以下步骤:监测所述第一手势动作的手掌是否发生移动;当所述第一手势动作的手掌发生移动时,获取移动后所述第一手势动作的手掌的第二手掌位置;将所述视频特效的起点从所述第一手掌位置变更为所述第二手掌位置;或者,监测所述直播主人脸是否发生移动;当所述直播主人脸发生移动时,获取移动后所述直播主人脸的第二人脸位置;将所述视频特效的终点从所述第一人脸位置变更为所述第二人脸位置;或者,统计所述第一手势动作的手势持续时间;当所述手势持续时间大于或等于预设的时间阈值时,触发第二视频特效。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/783,267 US11778263B2 (en) | 2019-12-09 | 2020-08-18 | Live streaming video interaction method and apparatus, and computer device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911251115.7A CN113038149A (zh) | 2019-12-09 | 2019-12-09 | 直播视频互动方法、装置以及计算机设备 |
CN201911251115.7 | 2019-12-09 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021114710A1 true WO2021114710A1 (zh) | 2021-06-17 |
Family
ID=76329464
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/109871 WO2021114710A1 (zh) | 2019-12-09 | 2020-08-18 | 直播视频互动方法、装置以及计算机设备 |
Country Status (3)
Country | Link |
---|---|
US (1) | US11778263B2 (zh) |
CN (1) | CN113038149A (zh) |
WO (1) | WO2021114710A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113453032A (zh) * | 2021-06-28 | 2021-09-28 | 广州虎牙科技有限公司 | 手势互动方法、装置、系统、服务器和存储介质 |
CN114327059A (zh) * | 2021-12-24 | 2022-04-12 | 北京百度网讯科技有限公司 | 手势处理方法、装置、设备以及存储介质 |
WO2023040970A1 (zh) * | 2021-09-15 | 2023-03-23 | 北京字跳网络技术有限公司 | 交互方法、装置、电子设备及存储介质 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021059770A1 (ja) * | 2019-09-24 | 2021-04-01 | ソニー株式会社 | 情報処理装置、情報処理システム、および情報処理方法、並びにプログラム |
CN111669623B (zh) * | 2020-06-28 | 2023-10-13 | 腾讯科技(深圳)有限公司 | 视频特效的处理方法、装置以及电子设备 |
CN114051151B (zh) * | 2021-11-23 | 2023-11-28 | 广州博冠信息科技有限公司 | 直播互动方法、装置、存储介质与电子设备 |
CN114979746B (zh) * | 2022-05-13 | 2024-03-12 | 北京字跳网络技术有限公司 | 一种视频处理方法、装置、设备及存储介质 |
CN117911238A (zh) * | 2024-01-29 | 2024-04-19 | 江苏臻星美悦健康科技有限公司 | 人脸识别处理的特效策略分化处理系统 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013027893A1 (ko) * | 2011-08-22 | 2013-02-28 | Kang Jun-Kyu | 통신단말장치의 감정 컨텐츠 서비스 장치 및 방법, 이를 위한 감정 인지 장치 및 방법, 이를 이용한 감정 컨텐츠를 생성하고 정합하는 장치 및 방법 |
KR20140140916A (ko) * | 2013-05-30 | 2014-12-10 | 삼성전자주식회사 | 영상 통화 시 관심 대상을 표시하기 위한 장치 및 그 방법 |
CN108600680A (zh) * | 2018-04-11 | 2018-09-28 | 南京粤讯电子科技有限公司 | 视频处理方法、终端及计算机可读存储介质 |
CN109922352A (zh) * | 2019-02-26 | 2019-06-21 | 李钢江 | 一种数据处理方法、装置、电子设备及可读存储介质 |
CN110163045A (zh) * | 2018-06-07 | 2019-08-23 | 腾讯科技(深圳)有限公司 | 一种手势动作的识别方法、装置以及设备 |
CN110442240A (zh) * | 2019-08-09 | 2019-11-12 | 杭州学两手网络科技有限公司 | 一种基于手势识别的教学交互系统 |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9792716B2 (en) * | 2014-06-13 | 2017-10-17 | Arcsoft Inc. | Enhancing video chatting |
US10592098B2 (en) * | 2016-05-18 | 2020-03-17 | Apple Inc. | Devices, methods, and graphical user interfaces for messaging |
EP3482344B1 (en) * | 2016-07-07 | 2022-08-31 | Harman International Industries, Incorporated | Portable personalization |
US10674187B2 (en) * | 2016-07-26 | 2020-06-02 | Facebook, Inc. | Systems and methods for shared broadcasting |
US10809808B2 (en) * | 2016-10-14 | 2020-10-20 | Intel Corporation | Gesture-controlled virtual reality systems and methods of controlling the same |
CN107765855A (zh) * | 2017-10-25 | 2018-03-06 | 电子科技大学 | 一种基于手势识别控制机器人运动的方法和系统 |
CN109963187B (zh) * | 2017-12-14 | 2021-08-31 | 腾讯科技(深圳)有限公司 | 一种动画实现方法和装置 |
CN108462883B (zh) * | 2018-01-08 | 2019-10-18 | 平安科技(深圳)有限公司 | 一种直播互动方法、装置、终端设备及存储介质 |
US11395089B2 (en) * | 2018-05-08 | 2022-07-19 | Google Llc | Mixing audio based on a pose of a user |
CN109344755B (zh) * | 2018-09-21 | 2024-02-13 | 广州市百果园信息技术有限公司 | 视频动作的识别方法、装置、设备及存储介质 |
WO2020117823A1 (en) * | 2018-12-03 | 2020-06-11 | Smule, Inc. | Augmented reality filters for captured audiovisual performances |
CN111629156A (zh) * | 2019-02-28 | 2020-09-04 | 北京字节跳动网络技术有限公司 | 图像特效的触发方法、装置和硬件装置 |
US10873697B1 (en) * | 2019-03-29 | 2020-12-22 | Facebook, Inc. | Identifying regions of interest in captured video data objects by detecting movement within higher resolution frames of the regions |
-
2019
- 2019-12-09 CN CN201911251115.7A patent/CN113038149A/zh active Pending
-
2020
- 2020-08-18 WO PCT/CN2020/109871 patent/WO2021114710A1/zh active Application Filing
- 2020-08-18 US US17/783,267 patent/US11778263B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013027893A1 (ko) * | 2011-08-22 | 2013-02-28 | Kang Jun-Kyu | 통신단말장치의 감정 컨텐츠 서비스 장치 및 방법, 이를 위한 감정 인지 장치 및 방법, 이를 이용한 감정 컨텐츠를 생성하고 정합하는 장치 및 방법 |
KR20140140916A (ko) * | 2013-05-30 | 2014-12-10 | 삼성전자주식회사 | 영상 통화 시 관심 대상을 표시하기 위한 장치 및 그 방법 |
CN108600680A (zh) * | 2018-04-11 | 2018-09-28 | 南京粤讯电子科技有限公司 | 视频处理方法、终端及计算机可读存储介质 |
CN110163045A (zh) * | 2018-06-07 | 2019-08-23 | 腾讯科技(深圳)有限公司 | 一种手势动作的识别方法、装置以及设备 |
CN109922352A (zh) * | 2019-02-26 | 2019-06-21 | 李钢江 | 一种数据处理方法、装置、电子设备及可读存储介质 |
CN110442240A (zh) * | 2019-08-09 | 2019-11-12 | 杭州学两手网络科技有限公司 | 一种基于手势识别的教学交互系统 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113453032A (zh) * | 2021-06-28 | 2021-09-28 | 广州虎牙科技有限公司 | 手势互动方法、装置、系统、服务器和存储介质 |
WO2023040970A1 (zh) * | 2021-09-15 | 2023-03-23 | 北京字跳网络技术有限公司 | 交互方法、装置、电子设备及存储介质 |
CN114327059A (zh) * | 2021-12-24 | 2022-04-12 | 北京百度网讯科技有限公司 | 手势处理方法、装置、设备以及存储介质 |
CN114327059B (zh) * | 2021-12-24 | 2024-08-09 | 北京百度网讯科技有限公司 | 手势处理方法、装置、设备以及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
US11778263B2 (en) | 2023-10-03 |
US20230012089A1 (en) | 2023-01-12 |
CN113038149A (zh) | 2021-06-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021114710A1 (zh) | 直播视频互动方法、装置以及计算机设备 | |
US20180041796A1 (en) | Method and device for displaying information on video image | |
WO2019101038A1 (zh) | 弹幕内容控制方法、计算机设备和存储介质 | |
CN108024079B (zh) | 录屏方法、装置、终端及存储介质 | |
CN108958610A (zh) | 基于人脸的特效生成方法、装置和电子设备 | |
CN107911736B (zh) | 直播互动方法及系统 | |
US9524587B2 (en) | Adapting content to augmented reality virtual objects | |
US9817235B2 (en) | Method and apparatus for prompting based on smart glasses | |
WO2021114709A1 (zh) | 直播视频互动方法、装置以及计算机设备 | |
WO2021159792A1 (zh) | 虚拟物品交互方法、装置、计算机设备及存储介质 | |
WO2023051185A1 (zh) | 图像处理方法、装置、电子设备及存储介质 | |
CN109582122B (zh) | 增强现实信息提供方法、装置及电子设备 | |
WO2018000619A1 (zh) | 一种数据展示方法、装置、电子设备与虚拟现实设备 | |
JP7231638B2 (ja) | 映像に基づく情報取得方法及び装置 | |
US20210077911A1 (en) | Method of determining exciting moments in a game video and method of playing a game video | |
CN114450969B (zh) | 视频截屏方法、终端及计算机可读存储介质 | |
US12034996B2 (en) | Video playing method, apparatus and device, storage medium, and program product | |
US20230316529A1 (en) | Image processing method and apparatus, device and storage medium | |
WO2020052062A1 (zh) | 检测方法和装置 | |
CN110839175A (zh) | 一种基于智能电视的交互方法、存储介质以及智能电视 | |
CN109302636B (zh) | 提供数据对象全景图信息的方法及装置 | |
WO2019119643A1 (zh) | 移动直播的互动终端、方法及计算机可读存储介质 | |
WO2024164983A1 (zh) | 一种特效生成方法、装置、计算机设备及存储介质 | |
US20240163392A1 (en) | Image special effect processing method and apparatus, and electronic device and computer readable storage medium | |
WO2019015411A1 (zh) | 录屏方法、装置及电子设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20899714 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20899714 Country of ref document: EP Kind code of ref document: A1 |