CN110519544B - Video call method and electronic equipment - Google Patents

Video call method and electronic equipment Download PDF

Info

Publication number
CN110519544B
CN110519544B CN201910817845.2A CN201910817845A CN110519544B CN 110519544 B CN110519544 B CN 110519544B CN 201910817845 A CN201910817845 A CN 201910817845A CN 110519544 B CN110519544 B CN 110519544B
Authority
CN
China
Prior art keywords
video
video object
electronic device
tag
call
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910817845.2A
Other languages
Chinese (zh)
Other versions
CN110519544A (en
Inventor
韩桂敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN201910817845.2A priority Critical patent/CN110519544B/en
Publication of CN110519544A publication Critical patent/CN110519544A/en
Application granted granted Critical
Publication of CN110519544B publication Critical patent/CN110519544B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72454User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/275Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone

Abstract

The embodiment of the invention discloses a video call method and electronic equipment, wherein the video call method comprises the following steps: receiving a first input of a user to N tags in a video call picture, wherein each tag indicates a video object in the video call picture; in response to the first input, sending target data to a second electronic device to cause the second electronic device to process each call video frame of the first electronic device based on the first tag control data set. The video call method disclosed by the embodiment of the invention can effectively hide one or more video objects in the video call picture.

Description

Video call method and electronic equipment
Technical Field
The embodiment of the invention relates to the technical field of communication, in particular to a video call method and electronic equipment.
Background
With the rapid development of the mobile internet, the video call function of the electronic device greatly facilitates the communication between people. Although the existing electronic equipment has a video call function, the function is not perfect enough in some special scenes, and the video call experience of a user is influenced. For example: when the user A and the user B carry out video call, the opposite side B is inconvenient to see certain video objects in the video picture due to the environment of the user A and the like, and the video call can be switched to the voice call only in a compromise mode, so that the communication experience of the two sides is influenced.
Disclosure of Invention
The embodiment of the invention provides a video call method, which aims to solve the problem that a certain video object in a video picture cannot be hidden during video call in the prior art.
In order to solve the technical problem, the invention is realized as follows:
in a first aspect, an embodiment of the present invention provides a video call method, which is applied to a first electronic device, where the method includes: receiving a first input of a user to N tags in a video call picture, wherein each tag indicates a video object in the video call picture; in response to the first input, sending target data to a second electronic device to cause the second electronic device to process each call video frame of the first electronic device based on the first tag control data set, the target data comprising a first tag control data set; wherein the first tag control data set comprises N sets, each set comprising a first element, a second element and a third element, the first element indicating a tag ID of one tag, the second element indicating a display state of one video object, and the third element indicating a processing manner of one video object.
In a second aspect, an embodiment of the present invention provides a video call method, which is applied to a second electronic device, where the method includes: receiving a first label control data set sent by first electronic equipment; processing each call video frame of the first electronic device based on the first tag control dataset; displaying each processed call video frame of the first electronic device; wherein the first tag control data set comprises N sets, each set comprising a first element indicating a tag ID of each tag, a second element indicating a display state of a video object, and a third element indicating a processing manner of the video object.
In a third aspect, an embodiment of the present invention provides an electronic device, which is used as a first electronic device, where the electronic device includes: the device comprises a first input receiving module, a first output receiving module and a second input receiving module, wherein the first input of a user to N labels in a video call picture is received, and each label indicates a video object in the video call picture; a first sending module, configured to send, in response to the first input, target data to a second terminal, so that the second terminal processes each call video frame of the first electronic device based on the first tag control data set, where the target data includes the first tag control data set; wherein the first tag control data set comprises N sets, each set comprising a first element, a second element and a third element, the first element indicating a tag ID of one tag, the second element indicating a display state of one video object, and the third element indicating a processing manner of one video object.
Preferably, the electronic device further includes: the first acquisition module is used for acquiring a call video image of a frame of video call picture before the first input receiving module receives first input of a user to N labels in the video call picture; the first identification module is used for identifying N video objects in the call video image; and the first marking module is used for respectively marking the N video objects through N labels.
Preferably, the first input receiving module includes: the first sub-module is used for receiving a first sub-input of a user to the N labels in the video call picture, and the first sub-input is used for selecting the N labels; the second sub-module is used for receiving a second sub-input of the user to the N tags, wherein the second sub-input is used for setting the display state of the corresponding video object of each tag; and the third receiving submodule is used for receiving a third sub-input of the user to the N tags, and the third sub-input is used for setting a processing mode of a corresponding video object of each tag.
Preferably, the first transmitting module includes: the first determining submodule is used for determining the display state and the processing mode of each video object based on the first input; the first generation submodule is used for generating N sets based on the N labels and the display state and the processing mode of each video object; a second generation submodule, configured to generate a first tag control data set based on the N sets; and the first sending submodule is used for sending the first label control data set as target data to second electronic equipment.
Preferably, the first generation submodule is specifically configured to: and generating an ith set based on the ith label and the display state and the processing mode of the ith video object indicated by the ith label.
Preferably, the display state comprises a first state and a second state, the first state indicating that the video object is displayed and the second state indicating that the video object is hidden; when the display state is the first state, the processing mode includes: rendering the video object and then displaying the rendered video object, or replacing the video object with a 3D model corresponding to the video object to display the rendered video object; when the display state is the second state, the processing mode includes: and filling the area where the video object is located by adopting the background image or filling the area where the video object is located by adopting a preset pattern.
Preferably, the electronic device further includes: a second obtaining module, configured to, after the first generating sub-module, obtain 3D modeling data of a video object corresponding to an ith tag when the processing manner of the video object corresponding to the ith tag is that a 3D model corresponding to the video object is used to replace the video object for display; an element generation module to generate a fourth element based on the 3D modeling data; an element adding module, configured to add the fourth element to the ith set.
Preferably, the electronic device further includes: a mark display module, configured to mark, by M tags, M video objects that are not included in the first tag control data set when M video objects appear in a video call screen after the first sending module sends target data to the second electronic device in response to the first input, and display the M tags in a preset manner; a data set generating module, configured to generate a second tag control data set when a second input to the tags of the M video objects is received within a preset time period, where the second tag data set includes a set corresponding to the tags of the M video objects; a second sending module, configured to send the second tag control data set to the second electronic device; wherein M is a positive integer.
Preferably, the electronic device further includes: a third input receiving module, configured to receive, after the first sending module sends the target data to the second electronic device in response to the first input, a third input of the user to at least one of the N tags in a video call with the second electronic device; an adjusting module, configured to adjust at least one of a display state and a processing manner of the video object corresponding to the at least one tag in response to the third input, and update the first tag control data set; and the third sending module is configured to send the updated first tag control data set to the second electronic device, so that the second electronic device processes each call video frame of the first electronic device according to the updated first tag control data set.
In a fourth aspect, an embodiment of the present invention provides an electronic device, which serves as a second electronic device, where the electronic device includes: the first data set receiving module is used for receiving a first label control data set sent by first electronic equipment; the video processing module is used for processing each call video frame of the first electronic equipment based on the first label control data set; the display module is used for displaying each processed call video frame of the first electronic equipment; wherein the first tag control data set comprises N sets, each set comprising a first element indicating a tag ID of each tag, a second element indicating a display state of a video object, and a third element indicating a processing manner of the video object.
Preferably, the video processing module includes: the video object determining submodule is used for respectively determining a video object corresponding to the kth label contained in each call video frame of the first electronic equipment; the processing mode determining submodule is used for determining the display state and the processing mode of the video object corresponding to the kth label based on the second element and the third element corresponding to the ith label; the processing submodule is used for processing the video object corresponding to the kth label in each call video frame based on the display state and the processing mode of the video object corresponding to the kth label; wherein k is a positive integer and is less than or equal to N.
Preferably, the display state comprises a first state and a second state, the first state indicating that the video object is displayed and the second state indicating that the video object is hidden; the processing submodule comprises: the first processing unit is used for rendering and displaying a first video object with a first display state, or replacing the video object with a 3D model corresponding to the video object for displaying; and the second processing unit is used for filling the area where the video object is located with a background image or filling the area where the video object is located with a preset pattern in the video frame for a second video object of which the display state is a second state.
Preferably, the ith set of the N sets includes a fourth element, which is 3D modeling data of the video object corresponding to the ith tag in the ith set; the video processing module includes: a 3D data acquisition sub-module for acquiring the 3D modeling data in an ith set of the N sets; and the replacement display submodule is used for replacing the video object corresponding to the ith label in the ith set in each call video frame of the first electronic equipment through the 3D modeling data to display.
Preferably, the electronic device further includes: the second data set receiving module is used for receiving a second tag control data set sent by the first electronic device after the first data set receiving module receives the first tag control data set sent by the first electronic device; the element extraction module is used for extracting the display states indicated by the M labels in the second label control data set and the second element of the set where each label is located and the processing mode indicated by the third element; and the image processing module is used for carrying out image processing on the M video objects corresponding to the M labels based on the display state indicated by the second element of the set where each label is located and the processing mode indicated by the third element.
Preferably, the electronic device further includes: a third data set receiving module, configured to receive, after the first data set receiving module receives the first tag control data set sent by the first electronic device, an updated first tag control data set sent by the first electronic device; a replacement module for replacing the stored first tag control data set with the updated first tag control data set; and the video frame processing module is used for processing each call video frame sent by the first electronic equipment based on the updated first tag control data set.
In a fifth aspect, an embodiment of the present invention provides an electronic device, which includes a processor, a memory, and a computer program stored on the memory and executable on the processor, and when the computer program is executed by the processor, the electronic device implements the steps of any one of the video call methods described in the embodiments of the present invention.
In a fourth aspect, an embodiment of the present invention provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the computer program implements the steps of any one of the video call methods described in the embodiments of the present invention.
In the embodiment of the invention, a first input of a user to N labels in a video call picture is received; and responding to the first input, sending the target data to the second electronic equipment so that the second electronic equipment processes each call video frame of the first electronic equipment based on the first label control data set, and effectively hiding one or more video objects in the video call picture.
Drawings
Fig. 1 is a flowchart of a video call method according to an embodiment of the present invention;
fig. 2 is a second flowchart of a video call method according to an embodiment of the present invention;
FIG. 3 is a schematic view of a call video image;
FIG. 4 is a comparison graph before and after processing of a video call picture;
FIG. 5 is a comparison graph before and after processing of a video call picture;
fig. 6 is a third flowchart of a video call method according to an embodiment of the present invention;
fig. 7 is a fourth flowchart of a video call method according to an embodiment of the present invention;
FIG. 8 is a block diagram of an electronic device of an embodiment of the invention;
FIG. 9 is a block diagram of an electronic device of an embodiment of the invention;
fig. 10 is a schematic diagram of a hardware structure of an electronic device according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1, a flowchart of a video call method according to an embodiment of the present invention is shown.
In the embodiment of the present invention, a video call method is described from a first electronic device side, and the video call method in the embodiment of the present invention includes the following steps:
step 101: first input of a user to the N labels in the video call picture is received.
The video call method provided by the embodiment of the invention is applied to a scene of video call between first equipment and second equipment. Wherein each tag indicates a video object in a video call screen. After the first device and the second device establish video call connection, labels corresponding to all video objects in a video call picture can be displayed in the first device, and a user of the first device inputs and sets second elements and third elements corresponding to all the labels. Wherein the second element indicates a display state of one video object and the third element indicates a processing manner of one video object.
Specifically, a display state pull-down option and a video object processing mode pull-down option may be set under each tag, and the user selects a display state and a video object processing method corresponding to the tag from the pull-down options set under the tag. And N is an integer greater than or equal to 1, and when N is greater than 1, the first electronic equipment user can carry out input setting on the second elements and the third elements corresponding to the labels one by one.
Step 102: in response to the first input, sending the target data to the second electronic device to cause the second electronic device to process each call video frame of the first electronic device based on the first tag control data set.
The object data includes a first tag control data set, wherein the first tag control data set includes N sets, each set including a first element indicating a tag ID of one tag, a second element indicating a display state of one video object, and a third element indicating a processing manner of one video object.
Display states may include, but are not limited to: a first state indicating that the video object is displayed and a second state indicating that the video object is hidden. In a specific implementation process, if only individual video objects in the video call picture need to be displayed on the second device, the display state of the tag corresponding to the video object that does not need to be displayed may be set to the second state, and the tag corresponding to the video object that needs to be displayed does not need to be set.
If the display states corresponding to the video objects are different, the corresponding processing modes are also different. For example: when the display state is the first state, the corresponding processing manner may include, but is not limited to: and rendering the video object and then displaying the rendered video object, or replacing the video object with a 3D model corresponding to the video object for displaying. When the display state is the second state, the corresponding processing manner may include, but is not limited to: and filling the area where the video object is located by adopting the background image or filling the area where the video object is located by adopting a preset pattern.
And the second electronic equipment processes each call video frame of the first electronic equipment according to the first label control data set and then displays the processed call video frames on the second electronic equipment. It should be noted that the first tag control data set may be sent only when the first device sends the first call video frame to the second device, or the first tag control data set may be sent to the second device once every preset time interval during the video call.
According to the video call method provided by the embodiment of the invention, first input of a user to N labels in a video call picture is received; and responding to the first input, sending the target data to the second electronic equipment so that the second electronic equipment processes each call video frame of the first electronic equipment based on the first label control data set, and effectively hiding one or more video objects in the call video picture.
Referring to fig. 2, a second flowchart of a video call method according to an embodiment of the present invention is shown.
In the embodiment of the present invention, the video call method is still described from the first electronic device side, and the video call method in the embodiment of the present invention includes the following steps:
step 201: and acquiring a call video image of a frame of video call picture.
The video call method provided by the embodiment of the invention is applied to a scene of video call between first electronic equipment and second electronic equipment. When a video call picture is displayed on first electronic equipment in the initial stage of video call, the first electronic equipment does not send a call video frame to second electronic equipment, and a user of the second electronic equipment cannot see the video call picture of the first electronic equipment. The first electronic equipment acquires a frame of call video image and sets elements corresponding to the labels in the frame of call video image.
Step 202: n video objects in the call video image are identified.
The call video image comprises a plurality of video objects, the first electronic equipment identifies multiple subjects of the frame of video call image, each video object is a subject, and each subject corresponds to a label. And respectively generating a label for the N video objects in the video call image after the identification is finished and displaying the label.
Step 203: by means of the N tags, N video objects are respectively marked.
Fig. 3 is a schematic view of a call video image, where each video object corresponds to one tag after multi-topic identification, which is a tag one 301 and a tag two 302. The first electronic equipment user can select one or more labels and set the elements corresponding to the selected labels respectively.
Step 204: first input of a user to the N labels in the video call picture is received.
Each tag indicates a video object in a video call screen. The first input to the tag may include, but is not limited to: and setting a second element and a third element corresponding to the label. The second element indicates a display state of one video object and the third element indicates a processing manner of one video object.
One way to preferably receive a first input from a user to N tags in a video call screen is to:
receiving a first sub-input of a user to N labels in a video call picture, wherein the first sub-input is used for selecting the N labels; receiving a second sub-input of the user to the N labels, wherein the second sub-input is used for setting the display state of the corresponding video object of each label; and receiving a third sub-input of the N labels from the user, wherein the third sub-input is used for setting a processing mode of a corresponding video object of each label.
The display state comprises a first state and a second state, the first state indicates that the video object is displayed, and the second state indicates that the video object is hidden; when the display state is the first state, the processing mode comprises the following steps: rendering the video object and then displaying the rendered video object, or replacing the video object with a 3D model corresponding to the video object to display the rendered video object; when the display state is the second state, the processing mode comprises the following steps: and filling the area where the video object is located by adopting the background image or filling the area where the video object is located by adopting a preset pattern.
One exemplary way to set the second element and the third element corresponding to the tag is to select the tag to display a second element pull-down menu, where an "x" and an "check mark" option are set in the pull-down menu, where clicking "x" is to hide a video object corresponding to the tag in a video call screen displayed on the second device, and clicking "check mark" is to display a video object corresponding to the tag in the video call screen displayed on the second device. And displaying a third element drop-down menu after the second element is detected to be set, wherein the third element drop-down menu comprises a processing mode option corresponding to the set second element, and the first equipment user can finish the setting of a specific processing mode by selecting the corresponding option.
In a specific implementation process, when the input operation is performed on the tag, only the tag corresponding to the video object which needs to be kept displayed can be set, and also only the tag corresponding to the video object which needs to be hidden and displayed can be set, and a specific setting mode can be selected by a first electronic device user according to actual requirements. Referring to fig. 4 showing a comparison before and after processing of a video call screen, fig. 4 specifically includes fig. 4a and fig. 4b, where fig. 4a is a schematic diagram before processing of the video call screen on the first electronic device side, fig. 4b is a schematic diagram after processing of the video call screen on the second electronic device side, and if the second element corresponding to the second tab 302 shown in fig. 4a is set to "x" as shown in fig. 4, the video object corresponding to the second tab 302 is not displayed when displaying the video call screen in the second electronic device. Referring to fig. 5 showing a comparison before and after video call screen processing, fig. 5 specifically includes fig. 5a and 5b, where fig. 5a is a schematic diagram before video call screen processing on the first electronic device side, fig. 5b is a schematic diagram after video call screen processing on the second electronic device side, and if the second element corresponding to the label two 302 shown in fig. 5a is set to "v" as shown in fig. 5, only the video object corresponding to the label two 302 is displayed when video call screen display is performed in the second electronic device.
Step 205: based on the first input, a display state and a processing manner of each video object are determined.
One video object corresponds to one label, the display state of the video object is a second element in a set corresponding to the label, and the processing mode is a third element in the set corresponding to the label.
Step 206: and generating N sets based on the N labels and the display state and the processing mode of each video object.
Specifically, the ith set is generated based on the ith label and the display state and processing mode of the ith video object indicated by the ith label.
Each set includes a first element indicating a tag ID of a tag, a second element indicating a display state of a video object, and a third element indicating a processing manner of a video object.
Under the condition that the processing mode of the video object corresponding to the ith label is that the 3D model corresponding to the video object is adopted to replace the video object for displaying, after N sets are generated based on the N labels and the display state and the processing mode of each video object, the first electronic equipment acquires 3D modeling data of the video object corresponding to the ith label; generating a fourth element based on the 3D modeling data; add the fourth element to the ith set.
After the fourth element is added to the set corresponding to the tag, when the second electronic device processes the video object corresponding to the tag according to each element in the set, the 3D image is generated according to the 3D modeling data, and the generated 3D image is adopted to replace the video object for display.
Step 207: based on the N sets, a first set of tag control data is generated.
The first tag control data set contains a set corresponding to the selected N tags.
A first tag control data set UserA { [ tag 1, display status, processing mode ], [ tag 2, display status, processing mode ], [ tag 3, display status, processing mode ], … … [ tag N, display status, processing mode ] }. Correspondingly, if a certain tag corresponds to 3D modeling data, the 3D modeling data is added as a fourth element to the tag set in the first tag control data set.
Step 208: and sending the first label control data set as target data to the second electronic equipment.
The first electronic equipment sends the first label control data set to appointed second electronic equipment, and simultaneously can send the first frame call video frame to the second electronic equipment, the second electronic equipment processes and displays the first frame call video frame according to the first label control data set, and at the moment, smooth video call starts. In the subsequent video call process, the second electronic device processes each call video frame of the first electronic device based on the first tag control data set. The specific processing manner may refer to the description in the related embodiment of the second electronic device, which is not described in detail in the embodiment of the present invention.
In the video call process of the first electronic device and the second electronic device, a new video object may appear in a call video picture of the first electronic device, and under the condition that M video objects which are not contained in the first tag control data set, namely the new video object, appear in the video call picture, M video objects are marked through the M tags, and the M tags are displayed according to a preset mode; generating a second tag control data set under the condition that a second input to the tags of the M video objects is received within a preset time length, wherein the second tag data set comprises a set corresponding to the tags of the M video objects; sending the second tag control data set to the second electronic device; wherein M is a positive integer.
The second electronic equipment processes the newly added video object according to the second label control data set, and the first electronic equipment user can flexibly control whether to display the newly added video object on the second electronic equipment by setting the set corresponding to the label of the newly added video object, and if the newly added video object is displayed in what way, the interestingness of video call can be improved.
In addition to the above, in the video call process between the first electronic device and the second electronic device, the first electronic device user may also need to adjust an element in a set corresponding to a certain tag of the N tags, and for this case, the first electronic device user may perform a third input on a tag to be adjusted of the N tags. The first electronic equipment receives a third input of the user to at least one label in the N labels in the process of video call with the second electronic equipment; responding to a third input, adjusting at least one of a display state and a processing mode of the video object corresponding to at least one label, and updating the first label control data set; and sending the updated first tag control data set to the second electronic equipment so that the second electronic equipment processes each call video frame of the first electronic equipment according to the updated first tag control data set.
The mode is convenient for the first electronic equipment user to flexibly adjust the display mode of the video object in the video call process, and the video call experience and the interestingness can be improved.
According to the video call method provided by the embodiment of the invention, first input of a user to N labels in a video call picture is received; and responding to the first input, sending the target data to the second electronic equipment so that the second electronic equipment processes each call video frame of the first electronic equipment based on the first label control data set, and effectively hiding one or more video objects in the video call picture. In addition, the video call method provided by the embodiment of the invention can also adjust the display state and the processing mode of the newly added video object and the video object set when the video call starts in the video call process, and can improve the interestingness of the video call.
Referring to fig. 6, a third flowchart of a video call method according to an embodiment of the present invention is shown.
In the embodiment of the present invention, a video call method is described from a second electronic device side, and the video call method in the embodiment of the present invention includes the following steps:
step 501: a first tag control data set transmitted by a first electronic device is received.
When the first electronic device and the second electronic device carry out video call, a user of the first electronic device can set the display state, the processing mode and the like of one or more video objects in a video call picture, so that the purpose of controlling the display effect of the video objects is achieved. One video object corresponds to one tag, one tag corresponds to one set, and each set comprises elements for controlling the display effect of the video object. Specifically, the first tag control data set includes N sets, each set including a first element indicating a tag ID of each tag, a second element indicating a display state of the video object, and a third element indicating a processing manner of the video object.
For the generation process of the first tag control data set, reference may be made to the foregoing related embodiment of the first electronic device side, which is not described in detail in this embodiment of the present invention.
Step 502: each call video frame of the first electronic device is processed based on the first tag control data set.
The first tag control data set comprises N sets, each set corresponds to one video object, and the second electronic equipment processes the corresponding video object according to each set.
For example: and when the second electronic device processes the call video frame, determining the video object corresponding to the ith label in the call video frame, and rendering and displaying the video object.
Step 503: and displaying each call video frame of the first electronic equipment after the processing.
The video object displayed in each call video of the first electronic device after processing and the display effect of the displayed video object are consistent with the requirements of the first electronic device user. For example: some video objects in the call video frame can be hidden and displayed, or the displayed video objects can be personalized and the like.
According to the video call method provided by the embodiment of the invention, first electronic equipment receives first input of a user on N labels in a video call picture; and responding to the first input, sending the target data to the second electronic equipment, and displaying each video object by the second electronic equipment according to the first label control data set, so that one or more video objects in the video call picture can be effectively hidden.
Referring to fig. 7, a fourth flowchart of a video call method according to an embodiment of the present invention is shown.
In the embodiment of the present invention, the video call method is still described from the second electronic device side, and the video call method in the embodiment of the present invention includes the following steps:
step 601: a first tag control data set transmitted by a first electronic device is received.
The first tag control data set comprises N sets, each set comprises a first element, a second element and a third element, the first element indicates the tag ID of each tag, the second element indicates the display state of the video object, and the third element indicates the processing mode of the video object.
Step 602: and respectively determining a video object corresponding to the kth label contained in each call video frame of the first electronic equipment.
The first tag control data set comprises N sets, each set corresponds to one tag, each tag corresponds to one video object, the second electronic equipment respectively processes all call video frames, and the processing modes of all call video frames are the same. Specifically, when the second electronic device processes a single call video frame, each to-be-processed video object is determined from the call video frame according to the tag corresponding to each set.
Step 603: and determining the display state and the processing mode of the video object corresponding to the kth label based on the second element and the third element corresponding to the kth label.
The display state includes a first state indicating that the video object is displayed and a second state indicating that the video object is hidden.
The processing mode comprises the following steps: rendering and displaying the video object, replacing the video object with a 3D model corresponding to the video object for displaying, filling the area where the video object is located with a background image, or filling the area where the video object is located with a preset pattern.
Step 604: and processing the video object corresponding to the kth label in each call video frame based on the display state and the processing mode of the video object corresponding to the kth label.
Wherein k is a positive integer and is less than or equal to N.
A method for processing a video object corresponding to a kth tag in each call video frame, preferably based on a display state and a processing method of the video object corresponding to the kth tag, includes:
for a first video object with a display state in a first state, rendering the first video object and then displaying the first video object, or replacing the video object with a 3D model corresponding to the video object to display the first video object; and for the second video object with the display state being the second state, filling the area where the video object is located with a background image in the video frame, or filling the area where the video object is located with a preset pattern.
Among other things, the rendering of the video object may include, but is not limited to: adding and replacing dressing, clothes and the like. The preset pattern may be a mosaic, a single tone image, or the like. The specific style of the preset pattern can be set by those skilled in the art according to actual requirements, which is not specifically limited in the embodiment of the present invention.
The first tag control data set includes N sets, and each set may include three elements, namely a first element, a second element, and a third element, or may include four elements, namely a first element, a second element, a third element, and a fourth element. When the ith set in the N sets comprises a fourth element, the fourth element is the 3D modeling data of the video object corresponding to the ith label in the ith set; in this case, based on the first tag control data set, when each video frame of the first electronic device is processed, 3D modeling data in the ith set of the N sets is acquired; and replacing the video object corresponding to the ith label in the ith set in each video frame of the first electronic equipment by the 3D modeling data for displaying.
Step 605: and displaying each call video frame of the first electronic equipment after the processing.
In a specific implementation process, the first electronic device may generate a first tag control data set according to a first input of a user when a video call starts, and may also generate a second tag control data set according to a second input of the user when a video object is newly added to a video call picture at the first electronic device during the video call and when the user sets a display of the newly added video object, and send the second tag control data set to the second device, which displays video objects corresponding to N sets included in the second device according to the first tag data set, and displays video objects corresponding to M sets included in the second device according to the second tag data set.
The process of the second device performing display processing on the video objects corresponding to the M sets included in the second tag data set according to the second tag data set includes: receiving a second label control data set sent by the first electronic equipment; extracting M labels in the second label control data set and the display state indicated by the second element and the processing mode indicated by the third element of the set where each label is located; and performing image processing on the M video objects corresponding to the M labels based on the display state indicated by the second element and the processing mode indicated by the third element of the set in which each label is positioned.
For example: in the initial stage of video call, the video call picture of the first device contains A, B and C three video objects, and after a period of time of video call, video objects D and E appear in the video call picture of the first device, in this case, the first electronic device can display a tag corresponding to a newly appearing video object D, E in the video call picture, the user can set the display state and display mode of the video object D, E by performing a second input on the tag, generate a second tag control data set according to the setting, send the second tag control data set to the second device, and the second device processes and displays the corresponding video object in the call video frame according to the received second tag control data set.
According to the mode of receiving the second label control data set sent by the first electronic equipment after receiving the first label control data set and processing each call video frame sent by the first electronic equipment according to the second label control data set and the first label control data set, a user can conveniently and flexibly control the display effect of each video object in the call video frame.
In addition to the above display setting of the newly added video object in the video call process, the element in the set corresponding to a certain tag of the N tags in the first tag control data set may be adjusted, and for this situation, the user of the first electronic device may perform a third input on the tag to be adjusted of the N tags, and the first electronic device updates the first tag control data set according to the third input and sends the updated first tag control data set to the second electronic device. The second electronic equipment receives the updated first label control data set sent by the first electronic equipment; replacing the stored first tag control data set with the updated first tag control data set; and processing each call video frame sent by the first electronic equipment based on the updated first tag control data set.
According to the method for processing the call video frame according to the updated first tag control data set after receiving the first tag control data set, the adjustment requirement of the first electronic equipment user on the video object display effect can be synchronized with the real video object display effect of the second electronic equipment, and the experience of video call can be improved.
According to the video call method provided by the embodiment of the invention, each call video frame of the first electronic equipment is processed based on the first label control data set by receiving the first label control data set sent by the first electronic equipment, so that one or more video objects in a video call picture can be effectively hidden, or interesting display is carried out on one or more video objects. In addition, the video call method provided by the embodiment of the invention can also adjust the display state and the processing mode of the newly added video object or the video object set at the beginning of the call video according to the second tag control data set sent by the first electron or the updated first tag control data set in the video call process, so that the video call experience of the user can be improved.
Referring to fig. 8, a block diagram of an electronic device of an embodiment of the invention is shown. The electronic device can realize the details of the video call method in the first two method embodiments and achieve the same effect.
The electronic device of the embodiment of the present invention is a first electronic device, wherein the electronic device includes: a first input receiving module 701, configured to receive a first input of a user to N tags in a video call screen, where each tag indicates a video object in the video call screen; a first sending module 702, configured to send, in response to the first input, target data to a second electronic device, so that the second electronic device processes each call video frame of the first electronic device based on the first tag control data set, where the target data includes the first tag control data set; wherein the first tag control data set comprises N sets, each set comprising a first element, a second element and a third element, the first element indicating a tag ID of one tag, the second element indicating a display state of one video object, and the third element indicating a processing manner of one video object.
Preferably, the electronic device further includes: the first acquisition module is used for acquiring a call video image of a frame of video call picture before the first input receiving module receives first input of a user to N labels in the video call picture; the first identification module is used for identifying N video objects in the call video image; and the first marking module is used for respectively marking the N video objects through N labels.
Preferably, the first input receiving module includes: the first sub-module is used for receiving a first sub-input of a user to the N labels in the video call picture, and the first sub-input is used for selecting the N labels; the second sub-module is used for receiving a second sub-input of the user to the N tags, wherein the second sub-input is used for setting the display state of the corresponding video object of each tag; and the third receiving submodule is used for receiving a third sub-input of the user to the N tags, and the third sub-input is used for setting a processing mode of a corresponding video object of each tag.
Preferably, the first transmitting module includes: the first determining submodule is used for determining the display state and the processing mode of each video object based on the first input; the first generation submodule is used for generating N sets based on the N labels and the display state and the processing mode of each video object; a second generation submodule, configured to generate a first tag control data set based on the N sets; and the first sending submodule is used for sending the first label control data set as target data to second electronic equipment.
Preferably, the first generation submodule is specifically configured to: and generating an ith set based on the ith label and the display state and the processing mode of the ith video object indicated by the ith label.
Preferably, the display state comprises a first state and a second state, the first state indicating that the video object is displayed and the second state indicating that the video object is hidden; when the display state is the first state, the processing mode includes: rendering the video object and then displaying the rendered video object, or replacing the video object with a 3D model corresponding to the video object to display the rendered video object; when the display state is the second state, the processing mode includes: and filling the area where the video object is located by adopting the background image or filling the area where the video object is located by adopting a preset pattern.
Preferably, the electronic device further includes: a second obtaining module, configured to, after the first generating sub-module, obtain 3D modeling data of a video object corresponding to an ith tag when the processing manner of the video object corresponding to the ith tag is that a 3D model corresponding to the video object is used to replace the video object for display; an element generation module to generate a fourth element based on the 3D modeling data; an element adding module, configured to add the fourth element to the ith set.
Preferably, the electronic device further includes: a mark display module, configured to mark, by M tags, M video objects that are not included in the first tag control data set when M video objects appear in a video call screen after the first sending module sends target data to the second electronic device in response to the first input, and display the M tags in a preset manner; a data set generating module, configured to generate a second tag control data set when a second input to the tags of the M video objects is received within a preset time period, where the second tag data set includes a set corresponding to the tags of the M video objects; a second sending module, configured to send the second tag control data set to the second electronic device; wherein M is a positive integer.
Preferably, the electronic device further includes: a third input receiving module, configured to receive, after the first sending module sends the target data to the second electronic device in response to the first input, a third input of the user to at least one of the N tags in a video call with the second electronic device; an adjusting module, configured to adjust at least one of a display state and a processing manner of the video object corresponding to the at least one tag in response to the third input, and update the first tag control data set; and the third sending module is configured to send the updated first tag control data set to the second electronic device, so that the second electronic device processes each call video frame of the first electronic device according to the updated first tag control data set.
The electronic device provided in the embodiment of the present invention can implement each process implemented by the first electronic device in the method embodiments of fig. 1 to fig. 5, and details are not repeated here to avoid repetition.
According to the electronic equipment provided by the embodiment of the invention, first input of a user to N labels in a video call picture is received; and responding to the first input, sending the target data to the second electronic equipment so that the second electronic equipment processes each call video frame of the first electronic equipment based on the first label control data set, and effectively hiding one or more video objects in the video call picture or displaying the one or more video objects with interest.
Referring to fig. 9, a block diagram of an electronic device of an embodiment of the invention is shown. The electronic device can realize the details of the video call method in the last two method embodiments and achieve the same effect.
The electronic device of the embodiment of the present invention is a second electronic device, wherein the electronic device includes: a first data set receiving module 801, configured to receive a first tag control data set sent by a first electronic device; a video processing module 802, configured to process each call video frame of the first electronic device based on the first tag control data set; a display module 803, configured to display each processed call video frame of the first electronic device; wherein the first tag control data set comprises N sets, each set comprising a first element indicating a tag ID of each tag, a second element indicating a display state of a video object, and a third element indicating a processing manner of the video object.
Preferably, the video processing module includes: the video object determining submodule is used for respectively determining a video object corresponding to the kth label contained in each call video frame of the first electronic equipment; the processing mode determining submodule is used for determining the display state and the processing mode of the video object corresponding to the kth label based on the second element and the third element corresponding to the kth label; the processing submodule is used for processing the video object corresponding to the kth label in each call video frame based on the display state and the processing mode of the video object corresponding to the kth label; wherein k is a positive integer and is less than or equal to N.
Preferably, the display state comprises a first state and a second state, the first state indicating that the video object is displayed and the second state indicating that the video object is hidden; the processing submodule comprises: the first processing unit is used for rendering and displaying a first video object with a first display state, or replacing the video object with a 3D model corresponding to the video object for displaying; and the second processing unit is used for filling the area where the video object is located with a background image or filling the area where the video object is located with a preset pattern in the video frame for a second video object of which the display state is a second state.
Preferably, the ith set of the N sets includes a fourth element, which is 3D modeling data of the video object corresponding to the ith tag in the ith set; the video processing module includes: a 3D data acquisition sub-module for acquiring the 3D modeling data in an ith set of the N sets; and the replacement display submodule is used for replacing the video object corresponding to the ith label in the ith set in each call video frame of the first electronic equipment through the 3D modeling data to display.
Preferably, the electronic device further includes: the second data set receiving module is used for receiving a second tag control data set sent by the first electronic device after the first data set receiving module receives the first tag control data set sent by the first electronic device; the element extraction module is used for extracting the display states indicated by the M labels in the second label control data set and the second element of the set where each label is located and the processing mode indicated by the third element; and the image processing module is used for carrying out image processing on the M video objects corresponding to the M labels based on the display state indicated by the second element of the set where each label is located and the processing mode indicated by the third element.
Preferably, the electronic device further includes: a third data set receiving module, configured to receive, after the first data set receiving module receives the first tag control data set sent by the first electronic device, an updated first tag control data set sent by the first electronic device; a replacement module for replacing the stored first tag control data set with the updated first tag control data set; and the video frame processing module is used for processing each call video frame sent by the first electronic equipment based on the updated first tag control data set.
The electronic device provided in the embodiment of the present invention can implement each process implemented by the second electronic device in the method embodiments of fig. 6 to fig. 7, and details are not repeated here to avoid repetition.
According to the electronic device provided by the embodiment of the invention, each call video frame of the first electronic device is processed based on the first tag control data set by receiving the first tag control data set sent by the first electronic device, so that one or more video objects in a video call picture can be effectively hidden, or interesting display is performed on one or more video objects.
Referring to fig. 10, a hardware structure diagram of an electronic device according to an embodiment of the present invention is shown.
Fig. 10 is a schematic diagram of a hardware structure of an electronic device for implementing various embodiments of the present invention, where the electronic device 900 includes, but is not limited to: a radio frequency unit 901, a network module 902, an audio output unit 903, an input unit 904, a sensor 905, a display unit 906, a user input unit 907, an interface unit 908, a memory 909, a processor 910, and a power supply 911. Those skilled in the art will appreciate that the electronic device configuration shown in fig. 10 does not constitute a limitation of the electronic device, and that the electronic device may include more or fewer components than shown, or some components may be combined, or a different arrangement of components. In the embodiment of the present invention, the electronic device includes, but is not limited to, a mobile phone, a tablet computer, a notebook computer, a palm computer, a vehicle-mounted terminal, a wearable device, a pedometer, and the like.
A processor 910 configured to receive a first input from a user of N tags in a video call screen, wherein each tag indicates a video object in the video call screen; in response to the first input, sending target data to a second electronic device to cause the second electronic device to process each call video frame of the first electronic device based on the first tag control data set, the target data comprising a first tag control data set; wherein the first tag control data set comprises N sets, each set comprising a first element, a second element and a third element, the first element indicating a tag ID of one tag, the second element indicating a display state of one video object, and the third element indicating a processing manner of one video object.
Alternatively, the processor 910 may be further configured to receive a first tag control data set transmitted by the first electronic device; processing each call video frame of the first electronic device based on the first tag control dataset; displaying each processed call video frame of the first electronic device; wherein the first tag control data set comprises N sets, each set comprising a first element indicating a tag ID of each tag, a second element indicating a display state of a video object, and a third element indicating a processing manner of the video object.
According to the electronic equipment provided by the embodiment of the invention, first input of a user to N labels in a video call picture is received; and responding to the first input, sending the target data to the second electronic equipment so that the second electronic equipment processes each call video frame of the first electronic equipment based on the first label control data set, and effectively hiding one or more video objects in the video call picture or displaying the one or more video objects with interest.
It should be understood that, in the embodiment of the present invention, the radio frequency unit 901 may be used for receiving and sending signals during a message transmission and reception process or a call process, and specifically, after receiving downlink data from a base station, the downlink data is processed by the processor 910; in addition, the uplink data is transmitted to the base station. Generally, the radio frequency unit 901 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 901 can also communicate with a network and other devices through a wireless communication system.
The electronic device provides wireless broadband internet access to the user via the network module 902, such as assisting the user in sending and receiving e-mails, browsing web pages, and accessing streaming media.
The audio output unit 903 may convert audio data received by the radio frequency unit 901 or the network module 902 or stored in the memory 909 into an audio signal and output as sound. Also, the audio output unit 903 may provide audio output related to a specific function performed by the electronic device 900 (e.g., a call signal reception sound, a message reception sound, etc.). The audio output unit 903 includes a speaker, a buzzer, a receiver, and the like.
The input unit 904 is used to receive audio or video signals. The input Unit 904 may include a Graphics Processing Unit (GPU) 9041 and a microphone 9042, and the Graphics processor 9041 processes image data of a still picture or video obtained by an image capturing device (such as a camera) in a video capture mode or an image capture mode. The processed image frames may be displayed on the display unit 906. The image frames processed by the graphic processor 9041 may be stored in the memory 909 (or other storage medium) or transmitted via the radio frequency unit 901 or the network module 902. The microphone 9042 can receive sounds and can process such sounds into audio data. The processed audio data may be converted into a format output transmittable to a mobile communication base station via the radio frequency unit 901 in case of the phone call mode.
The electronic device 900 also includes at least one sensor 905, such as light sensors, motion sensors, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor may adjust the brightness of the display panel 9061 according to the brightness of ambient light, and the proximity sensor may turn off the display panel 9061 and/or the backlight when the electronic device 900 is moved to the ear. As one type of motion sensor, an accelerometer sensor can detect the magnitude of acceleration in each direction (generally three axes), detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of an electronic device (such as horizontal and vertical screen switching, related games, magnetometer posture calibration), and vibration identification related functions (such as pedometer, tapping); the sensors 905 may also include a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, an infrared sensor, etc., which are not described in detail herein.
The display unit 906 is used to display information input by the user or information provided to the user. The Display unit 906 may include a Display panel 9061, where the Display panel 9061 may be configured in a form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like, where the Display panel 9061 is a screen of an electronic device, the screen of the electronic device includes N regions, and one photosensitive sensor is correspondingly disposed under each region.
The user input unit 907 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the electronic device. Specifically, the user input unit 907 includes a touch panel 9071 and other input devices 9072. The touch panel 9071, also referred to as a touch screen, may collect touch operations by a user on or near the touch panel 9071 (e.g., operations by a user on or near the touch panel 9071 using a finger, a stylus, or any other suitable object or accessory). The touch panel 9071 may include two parts, a touch detection device and a touch controller. The touch detection device detects the touch direction of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, sends the touch point coordinates to the processor 910, receives a command from the processor 910, and executes the command. In addition, the touch panel 9071 may be implemented by using various types such as a resistive type, a capacitive type, an infrared ray, and a surface acoustic wave. The user input unit 907 may include other input devices 9072 in addition to the touch panel 9071. Specifically, the other input devices 9072 may include, but are not limited to, a physical keyboard, function keys (such as a volume control key, a switch key, and the like), a track ball, a mouse, and a joystick, which are not described herein again.
Further, the touch panel 9071 may be overlaid on the display panel 9061, and when the touch panel 9071 detects a touch operation on or near the touch panel 9071, the touch panel is transmitted to the processor 910 to determine the type of the touch event, and then the processor 910 provides a corresponding visual output on the display panel 9061 according to the type of the touch event. Although in fig. 10, the touch panel 9071 and the display panel 9061 are two independent components to implement the input and output functions of the electronic device, in some embodiments, the touch panel 9071 and the display panel 9061 may be integrated to implement the input and output functions of the electronic device, which is not limited herein.
The interface unit 908 is an interface for connecting an external device to the electronic apparatus 900. For example, the external device may include a wired or wireless headset port, an external power supply (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 908 may be used to receive input from external devices (e.g., data information, power, etc.) and transmit the received input to one or more elements within the electronic device 900 or may be used to transmit data between the electronic device 900 and external devices.
The memory 909 may be used to store software programs as well as various data. The memory 909 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the cellular phone, and the like. Further, the memory 909 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 910 is a control center of the electronic device, connects various parts of the entire electronic device using various interfaces and lines, and performs various functions of the electronic device and processes data by running or executing software programs and/or modules stored in the memory 909 and calling data stored in the memory 909, thereby performing overall monitoring of the electronic device. Processor 910 may include one or more processing units; preferably, the processor 910 may integrate an application processor, which mainly handles operating systems, user interfaces, application programs, etc., and a modem processor, which mainly handles wireless communications. It is to be appreciated that the modem processor described above may not be integrated into processor 910.
The electronic device 900 may further include a power supply 911 (e.g., a battery) for supplying power to various components, and preferably, the power supply 911 may be logically connected to the processor 910 through a power management system, so as to manage charging, discharging, and power consumption management functions through the power management system.
In addition, the electronic device 900 includes some functional modules that are not shown, and thus are not described in detail herein.
Preferably, an embodiment of the present invention further provides an electronic device, which includes a processor 910, a memory 909, and a computer program that is stored in the memory 909 and can be run on the processor 910, and when the computer program is executed by the processor 910, the processes of the video call method embodiment are implemented, and the same technical effect can be achieved, and in order to avoid repetition, details are not described here again.
The embodiment of the present invention further provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the computer program implements each process of the video call method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here. The computer-readable storage medium may be a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present invention may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present invention.
While the present invention has been described with reference to the embodiments shown in the drawings, the present invention is not limited to the embodiments, which are illustrative and not restrictive, and it will be apparent to those skilled in the art that various changes and modifications can be made therein without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (19)

1. A video call method is applied to a first electronic device, and is characterized by comprising the following steps:
receiving a first input of a user to N tags in a video call picture, wherein each tag indicates a video object in the video call picture;
in response to the first input, sending target data to a second electronic device to cause the second electronic device to process each call video frame of the first electronic device based on a first tag control data set, the target data comprising the first tag control data set;
wherein the first tag control data set comprises N sets, each set comprising a first element, a second element and a third element, the first element indicating a tag ID of one tag, the second element indicating a display state of one video object, and the third element indicating a processing manner of one video object.
2. The method of claim 1, wherein prior to receiving the first user input of the N tags in the video call screen, further comprising:
acquiring a call video image of a frame of video call picture;
identifying N video objects in the call video image;
and marking the N video objects respectively through N labels.
3. The method of claim 1, wherein receiving a first user input of N tags in a video call screen comprises:
receiving a first sub-input of a user to N labels in the video call picture, wherein the first sub-input is used for selecting the N labels;
receiving a second sub-input of the user to the N tags, wherein the second sub-input is used for setting the display state of the corresponding video object of each tag;
and receiving a third sub-input of the N labels from the user, wherein the third sub-input is used for setting a processing mode of a corresponding video object of each label.
4. The method of claim 3, wherein sending target data to a second electronic device in response to the first input comprises:
determining a display state and a processing mode of each video object based on the first input;
generating N sets based on the N labels and the display state and the processing mode of each video object;
generating a first set of tag control data based on the N sets;
and sending the first label control data set as target data to a second electronic device.
5. The method of claim 4, wherein generating N sets based on the N tags and the display status and processing manner of each video object comprises:
and generating an ith set based on the ith label and the display state and the processing mode of the ith video object indicated by the ith label.
6. The method of claim 5, wherein the display state comprises a first state and a second state, wherein the first state indicates displaying the video object and the second state indicates hiding the video object;
when the display state is the first state, the processing mode includes: rendering the video object and then displaying the rendered video object, or replacing the video object with a 3D model corresponding to the video object to display the rendered video object;
when the display state is the second state, the processing mode includes: and filling the area where the video object is located by adopting the background image or filling the area where the video object is located by adopting a preset pattern.
7. The method according to claim 6, wherein when the processing manner of the video object corresponding to the ith tag is to replace the video object with a 3D model corresponding to the video object for display, after the generating N sets based on the N tags and the display state and the processing manner of each video object, the method further comprises:
acquiring 3D modeling data of the video object corresponding to the ith label;
generating a fourth element based on the 3D modeling data;
adding the fourth element to the ith set.
8. The method of claim 1, wherein after said transmitting target data to a second electronic device in response to said first input, the method further comprises:
under the condition that M video objects which are not contained in the first label control data set appear in a video call picture, marking the M video objects through the M labels, and displaying the M labels according to a preset mode;
generating a second tag control data set under the condition that a second input to the tags of the M video objects is received within a preset time length, wherein the second tag data set comprises a set corresponding to the tags of the M video objects;
sending the second tag control data set to the second electronic device;
wherein M is a positive integer.
9. The method of claim 1, wherein after said transmitting target data to a second electronic device in response to said first input, the method further comprises:
receiving a third input of a user to at least one of the N tags in the process of video call with the second electronic device;
responding to the third input, adjusting at least one of a display state and a processing mode of the video object corresponding to the at least one label, and updating the first label control data set;
and sending the updated first tag control data set to the second electronic equipment, so that the second electronic equipment processes each call video frame of the first electronic equipment according to the updated first tag control data set.
10. A video call method is applied to a second electronic device, and is characterized by comprising the following steps:
receiving a first label control data set sent by first electronic equipment;
processing each video frame of the first electronic device based on the first tag control dataset;
displaying each processed call video frame of the first electronic device;
wherein the first tag control data set comprises N sets, each set comprising a first element indicating a tag ID of each tag, a second element indicating a display state of a video object, and a third element indicating a processing manner of the video object.
11. The method of claim 10, wherein processing each call video frame of the first electronic device based on the first tag control dataset comprises:
respectively determining a video object corresponding to a kth label contained in each call video frame of the first electronic equipment;
determining the display state and the processing mode of the video object corresponding to the kth label based on the second element and the third element corresponding to the kth label;
processing the video object corresponding to the kth label in each call video frame based on the display state and the processing mode of the video object corresponding to the kth label;
wherein k is a positive integer and is less than or equal to N.
12. The method of claim 11, wherein the display state comprises a first state and a second state, wherein the first state indicates displaying the video object and the second state indicates hiding the video object;
the processing the video object corresponding to the kth tag in each call video frame based on the display state and the processing mode of the video object corresponding to the kth tag includes:
for a first video object with a display state being a first state, rendering the first video object and then displaying the first video object, or replacing the video object with a 3D model corresponding to the video object for displaying;
and for a second video object with a display state of a second state, filling the area of the video object with a background image in the video frame, or filling the area of the video object with a preset pattern.
13. The method of claim 10, wherein an ith set of the N sets includes a fourth element that is 3D modeling data for a video object corresponding to an ith tag of the ith set; processing each call video frame of the first electronic device based on the first tag control dataset comprises:
obtaining the 3D modeling data in an ith set of the N sets;
and replacing a video object corresponding to the ith label in the ith set in each call video frame of the first electronic equipment by the 3D modeling data for displaying.
14. The method of claim 10, wherein after receiving the first tag control data set transmitted by the first electronic device, the method further comprises:
receiving a second tag control data set sent by the first electronic device;
extracting the display states indicated by the M labels in the second label control data set and the second element of the set where each label is located and the processing mode indicated by the third element;
and performing image processing on the M video objects corresponding to the M labels based on the display state indicated by the second element and the processing mode indicated by the third element of the set in which each label is positioned.
15. The method of claim 10, wherein after receiving the first tag control data set transmitted by the first electronic device, the method further comprises:
receiving an updated first tag control data set sent by the first electronic device;
replacing the stored first tag control data set with the updated first tag control data set;
and processing each call video frame sent by the first electronic equipment based on the updated first tag control data set.
16. An electronic device, the electronic device being a first electronic device, the electronic device comprising:
the device comprises a first input receiving module, a first output receiving module and a second input receiving module, wherein the first input of a user to N labels in a video call picture is received, and each label indicates a video object in the video call picture;
a first sending module, configured to send, in response to the first input, target data to a second electronic device, so that the second electronic device processes each call video frame of the first electronic device based on a first tag control data set, where the target data includes the first tag control data set;
wherein the first tag control data set comprises N sets, each set comprising a first element, a second element and a third element, the first element indicating a tag ID of one tag, the second element indicating a display state of one video object, and the third element indicating a processing manner of one video object.
17. An electronic device, the electronic device being a second electronic device, the electronic device comprising:
the first data set receiving module is used for receiving a first label control data set sent by first electronic equipment;
the video processing module is used for processing each call video frame of the first electronic equipment based on the first label control data set;
the display module is used for displaying each processed call video frame of the first electronic equipment;
wherein the first tag control data set comprises N sets, each set comprising a first element indicating a tag ID of each tag, a second element indicating a display state of a video object, and a third element indicating a processing manner of the video object.
18. A terminal device comprising a processor, a memory and a computer program stored on the memory and executable on the processor, the computer program, when executed by the processor, implementing the steps of the video telephony method of any one of claims 1 to 15.
19. A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, carries out the steps of the video call method according to any one of claims 1 to 15.
CN201910817845.2A 2019-08-30 2019-08-30 Video call method and electronic equipment Active CN110519544B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910817845.2A CN110519544B (en) 2019-08-30 2019-08-30 Video call method and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910817845.2A CN110519544B (en) 2019-08-30 2019-08-30 Video call method and electronic equipment

Publications (2)

Publication Number Publication Date
CN110519544A CN110519544A (en) 2019-11-29
CN110519544B true CN110519544B (en) 2021-03-23

Family

ID=68628633

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910817845.2A Active CN110519544B (en) 2019-08-30 2019-08-30 Video call method and electronic equipment

Country Status (1)

Country Link
CN (1) CN110519544B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111601063B (en) * 2020-04-29 2021-12-14 维沃移动通信有限公司 Video processing method and electronic equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101626482A (en) * 2008-07-11 2010-01-13 华为技术有限公司 Method, device and system for implementing video conference
CN105472308A (en) * 2015-12-14 2016-04-06 湖北工业大学 Multi-view naked eye 3D video conference system

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7227567B1 (en) * 2004-09-14 2007-06-05 Avaya Technology Corp. Customizable background for video communications
GB0810205D0 (en) * 2008-06-04 2008-07-09 Advanced Risc Mach Ltd Graphics processing systems
US9124762B2 (en) * 2012-12-20 2015-09-01 Microsoft Technology Licensing, Llc Privacy camera
CN106358069A (en) * 2016-10-31 2017-01-25 维沃移动通信有限公司 Video data processing method and mobile terminal
CN109672822A (en) * 2018-12-29 2019-04-23 努比亚技术有限公司 A kind of method for processing video frequency of mobile terminal, mobile terminal and storage medium

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101626482A (en) * 2008-07-11 2010-01-13 华为技术有限公司 Method, device and system for implementing video conference
CN105472308A (en) * 2015-12-14 2016-04-06 湖北工业大学 Multi-view naked eye 3D video conference system

Also Published As

Publication number Publication date
CN110519544A (en) 2019-11-29

Similar Documents

Publication Publication Date Title
CN111049979B (en) Application sharing method, electronic equipment and computer readable storage medium
CN109525874B (en) Screen capturing method and terminal equipment
CN109240577B (en) Screen capturing method and terminal
CN110096326B (en) Screen capturing method, terminal equipment and computer readable storage medium
CN110109593B (en) Screen capturing method and terminal equipment
CN111010608B (en) Video playing method and electronic equipment
CN111026484A (en) Application sharing method, first electronic device and computer-readable storage medium
CN109343788B (en) Operation control method of mobile terminal and mobile terminal
CN109618218B (en) Video processing method and mobile terminal
CN108174109B (en) Photographing method and mobile terminal
CN110933494A (en) Picture sharing method and electronic equipment
CN108196781B (en) Interface display method and mobile terminal
CN111125800B (en) Icon display method and electronic equipment
CN110413363B (en) Screenshot method and terminal equipment
CN109639981B (en) Image shooting method and mobile terminal
CN109166164B (en) Expression picture generation method and terminal
CN111061404A (en) Control method and first electronic device
CN107729100B (en) Interface display control method and mobile terminal
CN111061407B (en) Video program operation control method, electronic device, and storage medium
CN111261128B (en) Screen brightness adjusting method and electronic equipment
CN110213437B (en) Editing method and mobile terminal
CN110519544B (en) Video call method and electronic equipment
CN109062483B (en) Image processing method and terminal equipment
CN111443968A (en) Screenshot method and electronic equipment
CN110851219A (en) Information processing method and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant