WO2021103610A1 - 展示对象的控制方法与装置、电子设备及存储介质 - Google Patents

展示对象的控制方法与装置、电子设备及存储介质 Download PDF

Info

Publication number
WO2021103610A1
WO2021103610A1 PCT/CN2020/104483 CN2020104483W WO2021103610A1 WO 2021103610 A1 WO2021103610 A1 WO 2021103610A1 CN 2020104483 W CN2020104483 W CN 2020104483W WO 2021103610 A1 WO2021103610 A1 WO 2021103610A1
Authority
WO
WIPO (PCT)
Prior art keywords
display
target object
target
screen
displayed
Prior art date
Application number
PCT/CN2020/104483
Other languages
English (en)
French (fr)
Chinese (zh)
Inventor
孙贺然
Original Assignee
北京市商汤科技开发有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京市商汤科技开发有限公司 filed Critical 北京市商汤科技开发有限公司
Priority to KR1020217015205A priority Critical patent/KR20210075188A/ko
Priority to JP2021527860A priority patent/JP2022515317A/ja
Publication of WO2021103610A1 publication Critical patent/WO2021103610A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content

Definitions

  • the present disclosure relates to the field of computer technology, and in particular to a method and device for controlling display objects, electronic equipment and storage media.
  • the present disclosure provides at least a solution for controlling the display object.
  • the present disclosure provides a method for controlling a display object, including: displaying the acquired video picture on a display screen; in a case where it is detected that the video picture includes a target object, displaying on the display screen A display object associated with the target object; based on a cumulative detection result of the target object in the video screen, controlling the display object in the display screen to change a display state.
  • the above provides an automatic detection scheme for the target object, which can detect whether the target object is included in the video image based on the video image displayed on the display screen, and display the target object when the target object is detected in the video image.
  • the associated display object controls the display object to change the display state.
  • the target object changes state information (such as gestures, smiles, etc.)
  • the cumulative detection result can also be changed accordingly, and the display object can be controlled to change the display state, thereby enriching the display effect and enhancing the interaction with the target object.
  • the present disclosure provides a display object control device, including: a first display module for displaying the acquired video picture on a display screen; a second display module for displaying the video picture in the If the target object is included, the display object associated with the target object is displayed on the display screen; the control module is configured to control the display screen based on the cumulative detection result of the target object in the video screen The display object in changes the display state.
  • the present disclosure provides an electronic device including: a processor; and a memory, wherein the memory stores machine-readable instructions executable by the processor, and the machine-readable instructions are executed by the processor At the time, the processor is made to implement the control method of the display object described in the first aspect.
  • the present disclosure provides a computer-readable storage medium having a computer program stored on the computer-readable storage medium, wherein when the computer program is run by a processor, the processor realizes the above-mentioned first aspect The control method of the display object.
  • FIG. 1 shows a schematic flowchart of a method for controlling a display object provided by an embodiment of the present disclosure
  • FIG. 2 shows a schematic diagram of an interface of a first display area provided by an embodiment of the present disclosure
  • FIG. 3 shows a schematic diagram of a special effect display provided by an embodiment of the present disclosure
  • FIG. 4 shows a schematic diagram of a second display object display provided by an embodiment of the present disclosure
  • FIG. 5 shows a schematic diagram of a method for determining a movement route of a target object provided by an embodiment of the present disclosure
  • FIG. 6 shows a schematic diagram of the distribution of various display areas in a display screen provided by an embodiment of the present disclosure
  • FIG. 7 shows a schematic diagram of the distribution of various display areas in another display screen provided by an embodiment of the present disclosure.
  • FIG. 8 shows a schematic structural diagram of a display object control device provided by an embodiment of the present disclosure
  • FIG. 9 shows a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • first, second, third, etc. may be used in the present disclosure to describe various information, these information should not be limited by these terms. These terms are only used to distinguish the same type of information from each other.
  • first information may also be referred to as second information, and similarly, the second information may also be referred to as first information.
  • word “if” as used herein can be interpreted as "when” or “when” or “in response to”.
  • the present disclosure can be applied to electronic equipment with data processing capabilities.
  • the electronic equipment can be equipped with a display device for displaying data processing results, or an external display device for displaying data processing results.
  • the connection method is not limited to wired connection, and/ Or wireless connection.
  • the electronic device may also be connected to at least one camera device, and the connection mode is not limited to wired connection and/or wireless connection.
  • the electronic equipment and the camera connected to it can be deployed in the same location area. After the camera takes the video, it is transmitted to the electronic equipment for processing. After the processing is completed, the electronic equipment can be externally connected through the display device configured by the electronic equipment or the electronic equipment.
  • Display device (such as a display screen) for display.
  • the electronic device may be, for example, a mobile phone, a tablet computer, a smart TV, a computer, and the like.
  • the target object may be all or part of the users appearing in the video screen, or users appearing in the video screen and conforming to a certain set attribute, or a specific object set in advance. Different target objects can be determined according to different application scenarios.
  • the cumulative detection result is the cumulative value of the detection results of a certain attribute of the target object or a certain posture of the target object at different times.
  • Specific attributes or postures can be set according to different application scenarios, and the detection results are not limited to the above two cases.
  • a method for controlling a display object includes steps S101 to S103.
  • the video picture displayed on the display screen may be obtained by real-time shooting by a camera device, which may be a camera built into the electronic device or an external camera device. If the camera device is an external device of the electronic device, the camera device and the display screen can be deployed in the same location area, for example, can be deployed in the same exhibition or in the same room. Considering that the video screen shot by the camera device is displayed on the display screen, in order to make it easier for users to watch their posture (such as facial posture and/or body posture), the camera device can be placed on the same vertical plane as the display screen. . Alternatively, the position where the user can capture the complete posture of the user when the user views the display screen can be used as the deployment position of the camera device.
  • a camera device which may be a camera built into the electronic device or an external camera device. If the camera device is an external device of the electronic device, the camera device and the display screen can be deployed in the same location area, for example, can be deployed in the same exhibition or in the same room.
  • Target object There can be multiple display objects associated with a target object, and different target objects can be associated with different display objects. The specifics will be described in detail in the following embodiments, and the description will not be expanded here.
  • the target object it is possible to detect whether the target object is included in the video picture based on the video picture displayed on the display screen, and when the target object is detected in the video picture, display the display object associated with the target object, and based on the comparison
  • the cumulative detection results of the target object in the video screen and control the display object to change the display state.
  • the target object changes state information (such as gestures, smiles, etc.)
  • the cumulative detection result can also be changed accordingly, and the display object can be controlled to change the display state, thereby enriching the display effect and enhancing the interaction with the target object.
  • the target object in step 102 may be all objects that appear in the image acquisition range, or may be part of the object that can obtain a complete face in the image acquisition range, or may be within the image acquisition range that meets the preset attribute characteristics Object.
  • the preset attribute characteristics can be, for example, gender, hair length, clothes color, and so on. If the target object is an object that meets the preset attribute characteristics, after acquiring the video image, the attribute characteristics of each object in the video image can be identified first, and then the object that meets the preset attribute characteristics is determined as the target object.
  • the target object may also be a specific object (there may be one or more), where the specific object may be, for example, a person who is scheduled to participate in an exhibition.
  • the biological characteristics of the specific object may be stored in advance, for example, facial characteristics, gait characteristics, pedestrian re-identification (Person Re-identification, REID) characteristics, and so on.
  • Detecting whether the target object is included in the video frame may include: detecting the biological characteristics of each object in the video frame, matching the detected biological characteristics with the biological characteristics of the specific object stored in advance, and determining the successfully matched object as the target object.
  • displaying the display object associated with the target object may include one or more of the following six situations.
  • Case 1 In the case where the target object is detected in the video frame, the first display object in the initial display state can be displayed in the first display area of the display screen, where the initial display state is used to prompt the target object to make a target gesture.
  • Target gestures include but are not limited to postures and facial expressions, such as hand gestures, smiling expressions, etc.
  • the acquired video images can also be displayed in the first display area of the display screen.
  • the key points of each target object in the video images can be detected first, and then based on the key points
  • the first display object in the initial display state is displayed in the first display area.
  • FIG. 2 is a schematic diagram of the interface of the first display area.
  • the first display area displays a video picture taken by the camera.
  • the video picture includes multiple target objects, and the head of each target object Special effects of flowers are displayed in this position.
  • the displayed flower is the first display object in the initial display state.
  • the first display area may display the classroom images captured by the camera in real time, and each student in the captured classroom images is a target object.
  • each student's head position there can be a palm special effect, and the palm special effect is the first display object.
  • the palm special effect in the initial display state can be a palm with a dashed border, with a prompt "raise hand to ask questions" around the palm to remind students to raise their hands for interaction.
  • the displayed prompt words can also be voice-broadcasted to give voice prompts to the students.
  • controlling the display object in the display screen to change the display state may include: detecting the target posture made by the target object in the video frame, and making the cumulative target posture based on the target object The duration or cumulative number of times controls the change of the display state of the first display object in the first display area.
  • the controlling the first display object in the first display area to change the display state based on the cumulative duration or cumulative number of times the target object makes the target gesture may include: when the cumulative duration or cumulative number of times has not reached In the case of the set value, the first display object is controlled to change the display state to the intermediate display state; wherein, the intermediate display state changes with the cumulative value of the cumulative duration or cumulative times to indicate the cumulative value.
  • the color of the flower can be changed according to the smile duration of the target user, or Light up the square in the progress bar below the flower. For example, if a square in the progress bar means a smile for 2 seconds, when the target object smiles for 10 seconds, light up 5 squares in the progress bar under the flower from bottom to top.
  • the color of the squares can be changed according to the number of illuminated squares. For example, when only 5 squares are lit, the color of the squares can be blue; as the number of lighted squares increases, the color of the squares can be gradually changed; when all the squares are lighted, the squares can be red.
  • the number of hands raised by the student ie the target object
  • the displayed palm special effects can be changed according to the number of hands raised by the student. If the student raises their hands 4 times, they can light up the palm special effect and add a " ⁇ 4" sign behind the palm special effect. If the student raises his hand for the first time, he can show only one palm special effect at the top of his head, and add the “ ⁇ 1” sign behind the displayed palm special effect.
  • controlling the first display object in the first display area to change the display state based on the cumulative duration or cumulative number of the target gestures made by the target object may include: When the number of times reaches the set value, the first display object is controlled to change the display state to the target display state to indicate that the target posture made by the target object meets the set condition.
  • the first display object when it is detected that the number of times the student raises their hands reaches the set number of times, the first display object may include not only the palm special effect, but also the star special effect; the target display state of the first display object may be, for example, the star special effect
  • the five stars of will light up in turn to indicate that the number of times the student has raised their hands has reached the set number of times; at the same time, a sign indicating the number of times the hand has been raised is added after the palm special effect.
  • the target object can be prompted to make a target posture, thereby increasing the interaction with the target object, and by controlling the first display object to change the display state, it becomes the first display object Provides more ways to display.
  • Case 2 In the case where the target object is detected in the video frame, the attribute feature of the target object in the video frame can be detected, and the second display object corresponding to the attribute feature can be displayed in the first display area of the display screen.
  • the second display object includes at least one of the following information: the identity of the target object, age value, smile value, charm value, viewing duration, duration of different emotions, duration of attention.
  • the type of the second display object corresponding to the attribute feature may be different, which can be specifically set according to actual conditions.
  • the target object may be a student in the classroom
  • the corresponding second display object may include at least one of the following information: duration of class, duration of attention , Positive emotion duration, negative emotion duration.
  • positive emotions can be happiness, excitement, excitement, etc.
  • negative emotions can be fear, sadness, etc.
  • class duration can be the length of time between when the electronic device starts to acquire the video image captured by the camera and the current moment
  • the attention time period may be the time period during which the camera device can collect the complete facial features of the student.
  • the length of attention can be determined in the following manner: by detecting the head position of the student in each frame of the video picture, the offset angle between the head position of the student and the preset position (for example, a blackboard) is determined; When the offset angle is within the preset angle range, it is determined that the student is paying attention to the blackboard at the moment; the attention duration is determined by counting the number of video frames in which the student is paying attention to the blackboard.
  • the preset position for example, a blackboard
  • the positive emotion duration and negative emotion duration of the students can be detected by the following method: input the captured video images into the pre-trained emotion recognition model to predict the corresponding value of each student in the video image. Emotions; determine whether the emotion is a positive emotion or a negative emotion based on the predicted emotions; count the duration of the video images where the student’s emotions belong to the positive emotion and the video images where the student’s emotions belong to the negative emotion as of the current moment The duration of the positive emotion and the duration of the negative emotion corresponding to each student are determined based on the duration of the video images in which each emotion appears.
  • the second display object may include one or more of identity, age, charm value, viewing time, expression, and smile value.
  • the attribute characteristics of the target object may be the biological characteristics of the target object (such as facial characteristics, gait characteristics, REID characteristics, etc.).
  • the biological characteristics of the target object in the video frame can be extracted first, and then the extracted biological characteristics can be matched with the biological characteristics of the known identities stored in the database, and the identity of the biological characteristics in the database that has been successfully matched
  • the identification is determined as the identification identification of the target object; if the matching is unsuccessful, an identification identification is assigned to the target object, and the identification identification assigned to the target object and its corresponding biological characteristics are stored in the database.
  • the video image can be input into a pre-trained neural network model to predict the age of the corresponding target object in the video image.
  • the detection method of the target object's charm value, expression, smile value and other attribute characteristics can be similar to the detection method of the target object's age, which will not be repeated here, but it needs to be explained that the neural network model used to predict different attribute characteristics Different, the training process of different neural network models is also different.
  • the attribute characteristics of the target object include multiple characteristics of age, charm value, expression, and smile value
  • the labels of the supervised data need to be multiple feature labels.
  • the method for determining the viewing time of the target object may be the same as the method for determining the student's attention time in the classroom scene, which will not be repeated here.
  • the attribute characteristics of the target object may also change. For example, if the target object changes facial expression, the expression, smile value, charm value, etc. of the detected target object may change.
  • the second display object in the case of detecting that the attribute characteristic of the target object changes, it is also possible to control the second display object in the first display area to change the display state according to the change of the attribute characteristic.
  • the second display object may be as shown in FIG. 4, and a display frame may be displayed around each target object.
  • the display frame may display at least one of the identity, age, charm value, expression, viewing time, and smile value of the target object, and the display position of the display frame may be a fixed position around the target object.
  • the key points of the body of the target object can be detected, and then the display position can be determined according to the key points of the body, and then the display frame corresponding to the target object can be displayed.
  • Case 3 When it is detected that the target object is included in the video picture, the sign-in information of the target object may be displayed in the second display area of the display screen.
  • the target object may be a predetermined attendee. For example, if the application scenario is a meeting, the target object is a predetermined participant.
  • the biological characteristics of the target object can be input in advance. After the video frame is acquired, the biological characteristics of each object in the video frame are detected, the detected biological characteristics are compared with the biological characteristics input in advance, and the objects whose comparison results are successful are determined as the target objects.
  • the check-in information of the target object may be displayed in the second display area, and the check-in information may include, for example, a photo of the target object and/or the identity information of the target object. After detecting that the target object is included in the video picture, the display state of the sign-in information of the target object in the second display area can be changed.
  • the sign-in information may be displayed in the form of information cards in the second display area, and each information card displays a photo of the target object and/or identity information of the target object.
  • the background color of the information card can be changed when the display status of the sign-in information is changed, or the color of the information card can be gray before the display status of the sign-in information is changed, and the display status of the sign-in information can be changed by lighting the information card, etc.
  • the display in the second display area may also be marked The sign-in information of the target object.
  • the display object in the first display area is a flower special effect
  • the color of the flower special effect is controlled to change, that is, the display state of the display object is the target display state.
  • a mark may be added to the check-in information of the corresponding target object in the second display area, for example, a “VIP” mark may be added to the check-in information of the target object in the second display area.
  • the photos included in the check-in information may be pictures stored in the database in advance.
  • the intercepted photos can be taken randomly, or taken from a frame of video image with the highest definition.
  • the position in the second display area is limited, and the number of target objects is multiple.
  • the check-in information of the target object may be displayed scrolling, or, after a period of time after adding a mark to the check-in information of the target object, the check-in information of the target object may not be displayed in the second display area.
  • the automatic sign-in of the target object can be realized, the sign-in steps can be simplified, and the sign-in efficiency can be improved; by marking the sign-in information of the target object, the display method of the sign-in information can be enriched, and the target object whose display state is the target display state can be distinguished , And display status are not the target audience of the target display status.
  • Case 4 In the case of detecting that the target object is included in the video screen, the description information of the target object may be displayed in the third display area of the display screen.
  • the description information of the target object may be the attribute information of the target object and the face image of the target object intercepted from the video screen.
  • the attribute information of the target object can be different.
  • the attribute information of the target object can include student ID, gender, age, number of classes, etc.; while in public places such as exhibitions and shopping malls, the target object’s attribute information
  • the attribute information may include identification, gender, number of visits, length of stay, etc.
  • the description information of the target object that can be displayed in the third display area is limited.
  • the third display area can only display the description information of N target objects.
  • the third display area can display the description information of all target objects;
  • the number of target objects included in the screen is greater than N, for example, the video screen includes N+1 target objects, the description information of the target object displayed first can be deleted, and the latest N+1th target object can be displayed
  • N is a positive integer.
  • the third display area may also only display the description information of the first M target objects appearing in the video screen, and the description information of the target objects appearing after the Mth position in the video screen is not displayed. Show that M is a positive integer, and M and N can be equal or unequal. Or, display the description information of the first M target objects that meet the characteristics of a specific attribute.
  • the description information of the target object displayed in the third display area may also be marked.
  • the flower special effect and the prompt "smile becomes a VIP" can be displayed in the corresponding position.
  • you can change the color of the flower special effect that is, the display state of the display object in the first display area becomes the target display state
  • the description information of the target object can be displayed, and after the display state of the display object is changed to the target display state, a mark is added to the description information, which enriches the display method of the description information and increases the interaction with the user.
  • Case 5 In the case where it is detected that the target object is included in the video screen, the business content can be displayed in the fourth display area of the display screen.
  • the business content displayed in the fourth display area may be, for example, advertisements, promotional videos, or venue introduction information, etc.
  • the form of the business content is not limited to pictures, text, audio, video, etc.
  • controlling the display state of the display object in the display screen may include: in the case of detecting that the posture of the target object in the video screen is switched from the first posture to the second posture, controlling the change The business content displayed in the fourth display area is switched from the first business content to the second business content.
  • the current posture of the target object is the first posture
  • the preset posture is the second posture
  • the business content currently displayed in the fourth display area is the first business content
  • the business content other than the first business content is the second Business content.
  • An indicator may be displayed in the fourth display area for instructing the target object to make a second gesture to switch the business content.
  • the indicator may be "please wave your hand to switch the display content".
  • the current posture of the target object is the first posture.
  • the business content displayed in the fourth display area can be controlled to switch. In this way, the interaction with the target object is increased, and the switching method of business content is enriched.
  • Case 6 In a case where it is detected that the video frame includes at least one target object, the feature distribution information of the at least one target object is displayed in the fifth display area of the display screen.
  • the feature distribution information may include, but is not limited to: age distribution, gender distribution, attendance distribution, popular route distribution, regional population density distribution, etc.
  • the electronic device when the feature distribution information includes the distribution of popular routes, the electronic device can be connected to multiple camera devices (the connection method can be wired connection or wireless connection); the movement route of each target object can be determined first, and then the moving route of each target object can be determined. Routes filter out popular routes and display them.
  • the method for determining the movement route of the target object may include steps S501 to S504.
  • S501 Acquire collected data of the first camera device, where the collected data includes the collected face image and identification information of the first camera device.
  • S502 Acquire location description information corresponding to the identification information of the first camera device.
  • S503 Determine the movement data of the target object identified by the face image based on the position description information of the first camera device.
  • the first camera device may be a camera device that currently photographs the target object, and the collected data of the first camera device also includes the collection time of the face image;
  • the historical movement data may be updated based on the position description information of the first camera device to obtain the movement data; the setting before the collection time is detected If the historical movement data of the target user is not obtained within the time period, the position description information of the first camera device is used as the movement data.
  • the historical movement data includes position description information of at least one second camera device, where the second camera device is a camera device that collects a face image of the target object within a set time period before the collection time.
  • the controlling the display object in the display screen to change the display state based on the cumulative detection result of the target object in the video screen may include: based on the cumulative detection of at least one target object As a result, the control updates the feature distribution information displayed in the fifth display area.
  • the displayed feature distribution information may show that the ratio of male to female is 5:2; or the gender distribution is displayed
  • the display can be performed in the form of a pie chart, and the display mode is not limited in the embodiment of the present disclosure.
  • the distribution of each display area in the display screen can be as shown in FIG. 6
  • the first display area and the fourth display area may be the same area, that is, the acquired video picture, the first display object, and the display service content are displayed in different display modes.
  • the distribution of each display area in the display screen can also be exemplified as shown in FIG. 7 .
  • the writing order of the steps does not mean a strict execution order but constitutes any limitation on the implementation process.
  • the specific execution order of each step should be based on its function and possibility.
  • the inner logic is determined.
  • the embodiment of the present disclosure also provides a display object control device.
  • the structure diagram of the display object control device provided by the embodiment of the present disclosure includes a first display module 801, The second display module 802 and the control module 803 are specifically:
  • the first display module 801 is used to display the acquired video picture on the display screen
  • the second display module 802 is configured to display the display object associated with the target object on the display screen when it is detected that the target object is included in the video frame;
  • the control module 803 is configured to control the display object in the display screen to change the display state based on the cumulative detection result of the target object in the video frame.
  • the second display module 802 may be used to display the initial display in the first display area of the display screen when the target object is detected in the video screen.
  • the first display object in the state; wherein the initial display state is used to prompt the target object to make a target gesture.
  • control module 803 may be used to: detect the target gesture made by the target object in the video frame; or based on the accumulated time or length of the target gesture made by the target object The cumulative number of times controls the change of the display state of the first display object in the first display area.
  • control module 803 may be used to control the first display object to change the display state to an intermediate display state when the accumulated duration or the accumulated number of times has not reached a set value; Wherein, the intermediate display state changes with the cumulative value of the cumulative duration or the cumulative number of times to indicate the cumulative value.
  • control module 803 may be configured to control the first display object to change the display state to the first target display state when the cumulative duration or cumulative number of times reaches a set value. , To indicate that the target posture made by the target object has met the set conditions.
  • the second display module 802 may be used to: detect the attribute characteristics of the target object in the video frame; display the attributes and characteristics of the target object in the first display area of the display screen. The corresponding second display object.
  • control module 803 may be configured to: in the case of detecting that the target attribute characteristics of the target object change, control the second display object in the first display area to follow The change of the attribute characteristics changes the display state.
  • the second display object includes at least one of the following information: the identity of the target object, age value, smile value, charm value, viewing time, duration of different emotions, attention duration.
  • the second display module 802 may be used to display the target object in the second display area of the display screen in the case of detecting that the video frame includes the target object.
  • the sign-in information of the target object may be used to display the target object in the second display area of the display screen in the case of detecting that the video frame includes the target object.
  • control module 803 may be further configured to: control the display object in the first display area of the display screen to change the display state, and detect that the changed display state is the target display In the case of the state, mark the sign-in information of the target object displayed in the second display area.
  • the second display module 802 may be used to display the target object in the third display area of the display screen in the case where it is detected that the video frame includes the target object. Descriptive information of the target object.
  • control module 803 may be further configured to: control the display object in the first display area of the display screen to change the display state, and detect that the changed display state is the target display In the case of the state, marking the description information of the target object displayed in the third display area is performed.
  • the second display module 803 may be configured to display business content in the fourth display area of the display screen in the case of detecting that the target object is included in the video screen .
  • control module 803 may be configured to: in the case of detecting that the posture of the target object in the video frame is switched from the first posture to the second posture, control the fourth posture to be The business content displayed in the display area is switched from the first business content to the second business content.
  • the second display module 802 may be configured to display in the fifth display area of the display screen when it is detected that the video frame includes at least one of the target objects At least one feature distribution information of the target object.
  • control module 803 may be configured to: based on a cumulative detection result of at least one of the target objects, control to update the feature distribution information displayed in the fifth display area.
  • the functions or templates contained in the device provided in the embodiments of the present disclosure can be used to execute the methods described in the above method embodiments.
  • the functions or templates contained in the device provided in the embodiments of the present disclosure can be used to execute the methods described in the above method embodiments.
  • the embodiments of the present disclosure also provide an electronic device.
  • the electronic device may include: a processor 901, a memory 902, and a bus 903.
  • the memory 902 includes a memory 9021 and an external memory 9022.
  • the memory 9021 here is also called internal memory, which is used to temporarily store the calculation data in the processor 901 and the data exchanged with the external memory 9022 such as a hard disk.
  • the processor 901 exchanges data with the external memory 9022 through the memory 9021.
  • the processor 901 and the memory 902 communicate through the bus 903.
  • the memory 902 stores machine-readable instructions executable by the processor 901.
  • the processor 901 can implement any of the above-mentioned methods for displaying movement state information.
  • the embodiments of the present disclosure also provide a computer-readable storage medium with a computer program stored on the computer-readable storage medium, wherein, when the computer program is executed by a processor, the processor can implement any of the above Display the control method of the object.
  • the computer program product of the method for controlling the display object provided by the embodiment of the present disclosure includes a computer-readable storage medium storing program code.
  • the program code includes instructions that can be used to execute the display object described in the above method embodiment. For the steps of the control method, refer to the foregoing method embodiment for details, which will not be repeated here.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the embodiments of the present disclosure.
  • the functional units in the various embodiments of the present disclosure may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the function is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a non-volatile computer readable storage medium executable by a processor.
  • the computer software product is stored in a storage medium and includes a number of instructions to enable a computer device (which may be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the methods described in the various embodiments of the present disclosure.
  • the aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disks or optical disks and other media that can store program codes. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
PCT/CN2020/104483 2019-11-28 2020-07-24 展示对象的控制方法与装置、电子设备及存储介质 WO2021103610A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
KR1020217015205A KR20210075188A (ko) 2019-11-28 2020-07-24 전시 대상 제어 방법, 장치, 전자 디바이스 및 기록 매체
JP2021527860A JP2022515317A (ja) 2019-11-28 2020-07-24 展示対象制御方法、装置、電子デバイス、及び記録媒体

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911190043.X 2019-11-28
CN201911190043.XA CN110968239B (zh) 2019-11-28 2019-11-28 一种展示对象的控制方法、装置、设备及存储介质

Publications (1)

Publication Number Publication Date
WO2021103610A1 true WO2021103610A1 (zh) 2021-06-03

Family

ID=70031963

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/104483 WO2021103610A1 (zh) 2019-11-28 2020-07-24 展示对象的控制方法与装置、电子设备及存储介质

Country Status (5)

Country Link
JP (1) JP2022515317A (ja)
KR (1) KR20210075188A (ja)
CN (1) CN110968239B (ja)
TW (1) TWI758837B (ja)
WO (1) WO2021103610A1 (ja)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110968239B (zh) * 2019-11-28 2022-04-05 北京市商汤科技开发有限公司 一种展示对象的控制方法、装置、设备及存储介质
CN111539339A (zh) * 2020-04-26 2020-08-14 北京市商汤科技开发有限公司 数据处理方法及装置、电子设备和存储介质
CN111625101B (zh) * 2020-06-03 2024-05-17 上海商汤智能科技有限公司 一种展示控制方法及装置
CN111857476B (zh) * 2020-07-17 2022-07-08 维沃移动通信有限公司 显示方法、装置及电子设备

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104246661A (zh) * 2012-04-16 2014-12-24 高通股份有限公司 使用手势与装置交互
US20170018023A1 (en) * 2015-07-16 2017-01-19 Countr, Inc. System and computer method for tracking online actions
CN107066983A (zh) * 2017-04-20 2017-08-18 腾讯科技(上海)有限公司 一种身份验证方法及装置
CN108053700A (zh) * 2018-01-02 2018-05-18 北京建筑大学 一种人工智能教学辅助系统
CN110968239A (zh) * 2019-11-28 2020-04-07 北京市商汤科技开发有限公司 一种展示对象的控制方法、装置、设备及存储介质

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5219184B2 (ja) * 2007-04-24 2013-06-26 任天堂株式会社 トレーニングプログラム、トレーニング装置、トレーニングシステムおよびトレーニング方法
JP5601045B2 (ja) * 2010-06-24 2014-10-08 ソニー株式会社 ジェスチャ認識装置、ジェスチャ認識方法およびプログラム
US8542879B1 (en) * 2012-06-26 2013-09-24 Google Inc. Facial recognition
JP6254785B2 (ja) * 2012-07-24 2017-12-27 サイトセンシング株式会社 視聴率調査システム、並びに表情情報生成装置及び表情情報生成プログラム
US11226686B2 (en) * 2014-01-20 2022-01-18 Lenovo (Singapore) Pte. Ltd. Interactive user gesture inputs
WO2017137948A1 (en) * 2016-02-10 2017-08-17 Vats Nitin Producing realistic body movement using body images
JP6516702B2 (ja) * 2016-05-24 2019-05-22 リズム時計工業株式会社 人数集計システム、人数集計方法、及び人数集計結果の閲覧方法
JP2018085597A (ja) * 2016-11-22 2018-05-31 パナソニックIpマネジメント株式会社 人物行動監視装置および人物行動監視システム
US10607035B2 (en) * 2017-08-31 2020-03-31 Yeo Messaging Ltd. Method of displaying content on a screen of an electronic processing device
CN110121117A (zh) * 2018-02-06 2019-08-13 优酷网络技术(北京)有限公司 视频结构化信息展示方法及装置
CN208141466U (zh) * 2018-05-17 2018-11-23 塔米智能科技(北京)有限公司 一种基于机器人的签到装置及系统

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104246661A (zh) * 2012-04-16 2014-12-24 高通股份有限公司 使用手势与装置交互
US20170018023A1 (en) * 2015-07-16 2017-01-19 Countr, Inc. System and computer method for tracking online actions
CN107066983A (zh) * 2017-04-20 2017-08-18 腾讯科技(上海)有限公司 一种身份验证方法及装置
CN108053700A (zh) * 2018-01-02 2018-05-18 北京建筑大学 一种人工智能教学辅助系统
CN110968239A (zh) * 2019-11-28 2020-04-07 北京市商汤科技开发有限公司 一种展示对象的控制方法、装置、设备及存储介质

Also Published As

Publication number Publication date
TWI758837B (zh) 2022-03-21
KR20210075188A (ko) 2021-06-22
TW202121250A (zh) 2021-06-01
CN110968239A (zh) 2020-04-07
CN110968239B (zh) 2022-04-05
JP2022515317A (ja) 2022-02-18

Similar Documents

Publication Publication Date Title
WO2021103610A1 (zh) 展示对象的控制方法与装置、电子设备及存储介质
US9898647B2 (en) Systems and methods for detecting, identifying and tracking objects and events over time
US11736756B2 (en) Producing realistic body movement using body images
WO2021135197A1 (zh) 状态识别方法、装置、电子设备及存储介质
US8508571B2 (en) Teleconference system
CN103733196A (zh) 用于实现现实世界用户经历的可搜索历史的方法及设备
CN112383830A (zh) 视频封面确定方法及装置、存储介质
CN110674664A (zh) 视觉注意力的识别方法及系统、存储介质、处理器
US20180158102A1 (en) Advertising display system using smart film screen
CN111640202A (zh) 一种ar场景特效生成的方法及装置
US10602091B2 (en) Method and system for providing video call service
JP2013157984A (ja) Ui提供方法およびそれを適用した映像受信装置
CN110868538A (zh) 推荐拍摄姿态的方法和电子设备
KR102178396B1 (ko) 증강현실 기반의 이미지출력물 제작 방법 및 장치
CN111464859B (zh) 一种在线视频展示的方法、装置、计算机设备及存储介质
US20210075754A1 (en) Method for sharing a photograph
US11409788B2 (en) Method for clustering at least two timestamped photographs
JP2018060375A (ja) 情報処理システム、情報処理装置およびプログラム
US9269159B2 (en) Systems and methods for tracking object association over time
JP2023014402A (ja) 情報処理装置、情報提示システム、情報処理方法、及び情報処理プログラム
US20210072869A1 (en) Method for retrieving at least two captured photographs
CN109688318B (zh) 一种录制教学视频的方法及装置
CN111626521A (zh) 一种游览路线生成的方法及装置
US20190180042A1 (en) Image display device, image display control device, and image display control method
Nasution et al. Person identification by face recognition on portable device for teaching-aid system: Preliminary report

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2021527860

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 20217015205

Country of ref document: KR

Kind code of ref document: A

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20894159

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20894159

Country of ref document: EP

Kind code of ref document: A1