WO2021036624A1 - 交互方法、装置、设备以及存储介质 - Google Patents
交互方法、装置、设备以及存储介质 Download PDFInfo
- Publication number
- WO2021036624A1 WO2021036624A1 PCT/CN2020/104466 CN2020104466W WO2021036624A1 WO 2021036624 A1 WO2021036624 A1 WO 2021036624A1 CN 2020104466 W CN2020104466 W CN 2020104466W WO 2021036624 A1 WO2021036624 A1 WO 2021036624A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- objects
- user
- information
- interactive
- image
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/002—Specific input/output arrangements not covered by G06F3/01 - G06F3/16
- G06F3/005—Input arrangements through a video camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/751—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
Definitions
- the present disclosure relates to the field of computer vision technology, and in particular to an interaction method, device, equipment, and storage medium.
- the way of human-computer interaction is mostly: the user inputs based on keys, touch, and voice, and the device responds by presenting images and text on the display screen.
- the device responds by presenting images and text on the display screen.
- most virtual characters are improved on the basis of voice assistants, and only the voice of the device is output, and the interaction between the user and the virtual characters is still on the surface.
- the embodiments of the present disclosure provide an interaction solution.
- an interaction method includes: acquiring an image of the periphery of a display device collected by a camera, the display device displaying interactive objects through a transparent display screen; and performing operations on one or more objects involved in the image. Detection; in response to detecting that at least two objects are involved in the image, a target object is selected from the at least two objects based on the detected feature information of the at least two objects; based on the detection of the target object As a result, the interactive object displayed on the transparent display screen of the display device is driven to respond to the target object.
- the characteristic information includes object posture information and/or object attribute information.
- the selecting the target object from the at least two objects according to the detected feature information of the at least two objects includes: according to the object posture information of each of the at least two objects The degree of posture matching with the set posture feature, or the target object is selected from the at least two objects according to the degree of match between the object attribute information of each of the at least two objects and the attribute of the set attribute feature .
- a suitable object By selecting a target object from multiple objects according to feature information such as object posture information and object attribute information of each object, a suitable object can be selected as the target object for interaction, thereby improving interaction efficiency and service experience.
- the selecting the target object from the at least two objects according to the detected feature information of the at least two objects includes: according to the object posture information of each of the at least two objects , Selecting one or more first objects that meet the characteristics of the set posture; when there are at least two first objects, driving the interactive objects to guide the at least two first objects to output setting information, respectively, And the target object is determined according to the order in which the detected first objects output the setting information respectively.
- the target object with high willingness to cooperate can be selected from the objects that meet the characteristics of the set posture, which can improve the interaction efficiency and service experience.
- the selecting the target object from the at least two objects according to the detected feature information of the at least two objects includes: according to the object posture information of each of the at least two objects , Select one or more first objects that meet the characteristics of the set posture; if there are at least two first objects, determine the at least two first objects according to the respective object attribute information of the at least two first objects The respective interaction response priority of each first object, and the target object is determined according to the interaction response priority.
- the method further includes: after selecting a target object from the at least two objects, driving the interactive object to output confirmation information to the target object.
- the object By outputting confirmation information to the target object, the object can be made clear that the object is currently in an interactive state, and the interaction efficiency is improved.
- the method further includes: responding to that the object is not detected from the image at the current moment, and the object is not detected and not tracked from the image within a set time period before the current moment Object, determining that the object to be interacted with of the interactive object is empty, and making the display device enter the waiting object state.
- the method further includes: in response to no object being detected from the image at the current moment, and the object is detected or tracked from the image within a set time period before the current moment, The to-be-interacted object of the interactive object is determined to be the object that interacted most recently.
- the display state of the interactive object is more in line with actual interaction requirements, More targeted.
- the display device displays the reflection of the interaction object through the transparent display screen, or the display device displays the reflection of the interaction object on the bottom plate.
- the displayed interactive objects can be made more three-dimensional and vivid.
- the interactive object includes a virtual character with a three-dimensional effect.
- an interactive device in a second aspect, includes: an image acquisition unit for acquiring images around a display device collected by a camera, the display device displaying interactive objects through a transparent display screen; One or more objects involved in the image are detected; an object selection unit is configured to respond to the detection unit detecting that at least two objects are involved in the image, according to the detected feature information of the at least two objects , Select a target object from the at least two objects; a driving unit, configured to drive the interactive object displayed on the transparent display screen of the display device to perform the target object on the target object based on the detection result of the target object Response.
- the characteristic information includes object posture information and/or object attribute information.
- the object selection unit is specifically configured to: according to the degree of matching of the object posture information of each of the at least two objects with the posture of the set posture feature, or according to the The object attribute information of each object matches the degree of the attribute of the set attribute feature, and the target object is selected from the at least two objects.
- the object selection unit is specifically configured to: select one or more first objects that meet the set posture characteristics according to the object posture information of each of the at least two objects; When there are at least two objects, the driving unit is caused to drive the interactive objects to guide the at least two first objects to output setting information, and to output the settings according to the detected first objects. The order of the information determines the target object.
- the object selection unit is specifically configured to: select one or more first objects that meet the set posture characteristics according to the object posture information of each of the at least two objects; In the case that there are at least two objects, the respective interaction response priorities of the at least two first objects are determined according to the respective object attribute information of the at least two first objects, and the interaction response priorities are determined according to the interaction response priorities.
- the target object is specifically configured to: select one or more first objects that meet the set posture characteristics according to the object posture information of each of the at least two objects; In the case that there are at least two objects, the respective interaction response priorities of the at least two first objects are determined according to the respective object attribute information of the at least two first objects, and the interaction response priorities are determined according to the interaction response priorities.
- the target object is specifically configured to: select one or more first objects that meet the set posture characteristics according to the object posture information of each of the at least two objects; In the case that there are at least two objects, the respective interaction response priorities of the at least two first objects are determined according to the respective object attribute information of the at least
- the device further includes a confirmation unit configured to: in response to the object selection unit selecting a target object from the at least two objects, cause the driving unit to drive the interactive object Output confirmation information to the target object.
- the device further includes a waiting state unit configured to respond to the detection unit not detecting an object from the image at the current moment, and at a set time before the current moment In the segment, no object is detected and no object is tracked from the image, it is determined that the to-be-interacted object of the interactive object is empty, and the display device enters the waiting object state.
- a waiting state unit configured to respond to the detection unit not detecting an object from the image at the current moment, and at a set time before the current moment In the segment, no object is detected and no object is tracked from the image, it is determined that the to-be-interacted object of the interactive object is empty, and the display device enters the waiting object state.
- the device further includes an end state unit configured to respond to the detection unit not detecting an object from the image at the current moment, and at a set time before the current moment An object is detected or tracked from the image in the segment, and the object to be interacted with the interactive object is determined to be the object that interacted most recently.
- the display device also displays the reflection of the interaction object through the transparent display screen, or the display device also displays the reflection of the interaction object on the bottom plate.
- the interactive object includes a virtual character with a three-dimensional effect.
- an interactive device in a third aspect, includes a processor; a memory for storing instructions executable by the processor, and when the instructions are executed, the processor is prompted to implement any implementation provided in the present disclosure The interactive method described in the method.
- a computer-readable storage medium having a computer program stored thereon, and when the computer program is executed by a processor, the processor is prompted to implement the interaction method according to any of the embodiments provided in the present disclosure .
- Fig. 1 shows a flowchart of an interaction method according to at least one embodiment of the present disclosure
- Fig. 2 shows a schematic diagram of displaying interactive objects according to at least one embodiment of the present disclosure
- Fig. 3 shows a schematic structural diagram of an interactive device according to at least one embodiment of the present disclosure
- Fig. 4 shows a schematic structural diagram of an interactive device according to at least one embodiment of the present disclosure.
- FIG. 1 shows a flowchart of an interaction method according to at least one embodiment of the present disclosure. As shown in FIG. 1, the method includes steps 101 to 104.
- step 101 an image of the periphery of a display device collected by a camera is acquired, and the display device displays interactive objects through a transparent display screen.
- the periphery of the display device includes any direction within the setting range of the display device, for example, it may include one or more of the front direction, the side direction, the rear direction, and the upper direction of the display device.
- the camera used to collect images can be set on the display device or used as an external device, independent of the display device. And the image collected by the camera can also be displayed on the transparent display screen in the display device.
- the number of the cameras can be multiple.
- the image collected by the camera may be a frame in the video stream, or may be an image obtained in real time.
- step 102 one or more users involved in the image are detected.
- the one or more users in the image described herein refer to one or more objects involved in the detection process of the image.
- object and “user” can be used interchangeably, and for convenience of presentation, they are collectively referred to as "user”.
- the detection results are obtained, such as whether there are users around the display device and how many users are there, and information about the detected users can also be obtained, for example, from image recognition technology
- the detection result may also include other information.
- step 103 in response to detecting that at least two users are involved in the image, a target user is selected from the at least two users according to the detected characteristic information of the at least two users.
- users can be selected according to corresponding feature information.
- step 104 based on the detection result of the target user, the interactive object displayed on the transparent display screen of the display device is driven to respond to the target user.
- the interactive object In response to the detection results of different target users, the interactive object will be driven to respond correspondingly to the different target users.
- the interactive object displayed on the transparent display screen of the display device is driven to respond to the target user , Can choose the target user suitable for the current scene to interact in the multi-user scene, which improves the interaction efficiency and service experience.
- the interactive objects displayed on the transparent display screen of the display device include virtual characters with a three-dimensional effect.
- the interaction process can be made more natural and the user's interaction experience can be improved.
- the interactive objects are not limited to virtual characters with three-dimensional effects, but may also be virtual animals, virtual items, cartoon characters, and other virtual images capable of realizing interactive functions.
- the three-dimensional effect of the interactive object displayed on the transparent display screen can be realized by the following method.
- Whether the human eye sees an object in three dimensions is usually determined by the shape of the object itself and the light and shadow effects of the object.
- the light and shadow effects are, for example, high light and dark light in different areas of the object, and the projection of the light on the ground after the object is irradiated (that is, the reflection).
- the reflection of the interactive object is also displayed on the transparent display screen, so that the human eye can observe the stereoscopic Picture.
- a bottom plate is provided under the transparent display screen, and the transparent display is perpendicular or inclined to the bottom plate. While the transparent display screen displays the stereoscopic video or image of the interactive object, the reflection of the interactive object is displayed on the bottom plate, so that the human eye can observe the stereoscopic image.
- the display device further includes a box body, and the front side of the box body is set to be transparent, for example, the transparent setting is realized by materials such as glass or plastic.
- the transparent setting is realized by materials such as glass or plastic.
- one or more light sources are also provided in the box to provide light to the transparent display screen to form a reflection.
- the three-dimensional video or image of the interactive object is displayed on the transparent display screen, and the reflection of the interactive object is formed on the transparent display screen or the bottom plate to achieve the three-dimensional effect, so that the displayed interactive object It is more three-dimensional and vivid, and enhances the user's interactive experience.
- the characteristic information includes user posture information and/or user attribute information
- the target user can be selected from at least two detected users according to the user posture information and/or user attribute information.
- the user gesture information refers to characteristic information obtained by performing image recognition in an image, such as user actions, gestures, and so on.
- User attribute information refers to the characteristic information about the user, including the user's identity (for example, whether it is a VIP user), service record, time of arrival at the current location, and so on.
- the attribute feature information may be obtained from user history records stored on the display device or the cloud, and the user history records may be related to the user's face and/or human body characteristics by retrieving on the display device or the cloud. Information is obtained from matching records.
- the target user may be selected from the at least two users according to the degree of match between the user posture information of each of the at least two users and the posture of the set posture feature.
- the user with the highest degree of posture matching among the matching results of the at least two users can be determined by matching the user posture information of the at least two users with the hand-raising action. For the target user.
- the target user may be selected from the at least two users according to the degree of matching between the user attribute information of each of the at least two users and the attributes of the set attribute characteristics.
- the attribute matching degree in the matching results of the at least two users may be matched by matching the user attribute information of the at least two users with the set attribute characteristics. The highest user is determined as the interactive object.
- a user suitable for the current application scenario can be selected as the target user for interaction , So as to improve the interaction efficiency and service experience.
- the target user can be selected from the at least two users in the following manner:
- the first user who meets the set posture feature is selected.
- conforming to the set posture feature means that the posture matching degree of the user posture information and the set posture feature is greater than a set value, for example, greater than 80%.
- the posture feature is set as a hand-raising action.
- the first user whose posture information matches the posture of the hand-raising action higher than 80% (the user is considered to have performed the hand-raising action). That is, all users who have performed the gesture of raising their hands are selected.
- the target user can be further determined by the following method: driving the interactive object to guide the at least two first users to output setting information respectively, and according to the detected first user The order in which each user outputs the setting information determines the target user.
- the setting information output by the first user may be one or more of actions, expressions, and voices.
- at least two first users are guided to perform a jumping action, and the first user who performs the jumping action first is determined as the target user.
- a target user with high willingness to cooperate can be selected from users who meet the characteristics of the set posture, which can improve the interaction efficiency and service experience.
- the target user can be further determined by the following method:
- the interaction response priority is determined according to the user attribute information of each first user, and the first user with the highest priority is determined Determined as the target user.
- the user attribute information used as the basis for selection can be comprehensively judged in combination with the user's current needs and actual scenes. For example, in the scenario of queuing to buy tickets, the time of arrival at the current location can be used as the basis of user attribute information to determine the interaction priority.
- the user who arrives first has the highest interactive response priority and can be determined as the target user; in other service locations, the target user can also be determined based on other user attribute information, for example, the interaction priority is determined based on the user's points in the location , So that the user with the highest score has the highest interactive response priority.
- each user may be further guided to output setting information. If the number of first users who output the setting information is still more than one, the user with the highest interactive response priority can be determined as the target user.
- the target user is selected from multiple detected users in combination with user attribute information, user posture information, and application scenarios, and different interactive response priorities can be set to provide corresponding services to the target user. Selecting a suitable user as the target user for interaction improves the interaction efficiency and service experience.
- the user can be notified that the user is selected by outputting confirmation information to the user.
- the interactive object may be driven to point to the user with a finger, or the interactive object may be driven to highlight the user in the camera preview screen, or the confirmation information may be output by other means.
- the user by outputting confirmation information to the target user, the user can be made sure that the user is currently in an interactive state, and the interaction efficiency is improved.
- the interaction object After a certain user is selected as the target user for interaction, the interaction object only responds or preferentially responds to the instructions of the target user until the target user leaves the shooting range of the camera.
- the user When the user is not detected in the image around the device, it means that there is no user around the display device, that is, the device is not currently in a state of interacting with the user.
- This state includes that there is no user interacting with the device in the set time period before the current time, that is, waiting for the user state; it also includes the user interacting with the user in the set time period before the current time, and the device Is in the user away state.
- the interactive object should be driven to react differently. For example, for the waiting user state, the interactive object can be driven to respond to welcome the user in combination with the current environment; while for the user leaving state, the interactive object can be driven to respond to the user who interacted most recently to end the service.
- no user in response to the user not being detected from the image at the current moment, and within a set time period before the current moment, for example, within 5 seconds, no user is detected from the image and no user is detected from the image.
- the user is tracked, the user to be interacted with the interactive object is determined to be empty, and the interactive object on the display device is driven to enter the waiting user state.
- the interaction object in response to the user being not detected from the image at the current moment, and the user is detected or tracked from the image within a set period of time before the current moment, it is determined that the interaction object is The user to be interacted is the user who interacted most recently.
- the interactive object when there is no user interacting with the interactive object, by determining that the device is currently waiting for the user or the user leaving the state, and driving the interactive object to make different responses, the interactive object is displayed The status is more in line with the interaction needs and more targeted.
- the detection result may also include the current service status of the device.
- the current service status may also include discovering the user status and so on.
- the current service state of the device may also include other states, and is not limited to the above.
- a human face and/or a human body is detected from the image around the device, it means that there is a user around the display device, and the state at the moment when the user is detected may be determined as the user-discovered state.
- the user history information stored in the display device can also be obtained, and/or the user history information stored in the cloud can be obtained to determine whether the user is a regular user , Or whether it is a VIP customer.
- the user history information may also include the user's name, gender, age, service record, remarks, and so on.
- the user history information may include information input by the user, or may include information recorded by the display device and/or cloud.
- the user history information matching the user may be found based on the detected feature information of the user's face and/or human body.
- the interactive object When the display device is in a user discovery state, the interactive object can be driven to respond according to the current service state of the display device, user attribute information obtained from the image, and user history information obtained through search.
- the user history information When a user is detected for the first time, the user history information may be empty, that is, the interaction object is driven according to the current service state, the user attribute information, and the environment information.
- the user’s face and/or human body can be recognized through the image first to obtain basic user attribute information about the user.
- the user is a female and is at the age of Between 20 and 30 years old; then, according to the user’s face and/or body feature information, search on the display device and/or the cloud to find user history information that matches the feature information, for example, the user Name, service record, etc.
- the interactive object is driven to make a targeted welcoming action to the female user, and to show the female user the services that can be provided for the female user.
- the order of providing services can be adjusted, so that the user can find the service items of interest more quickly.
- feature information of the at least two users can be obtained first, and the feature information can include at least one of user posture information and user attribute information, and The feature information corresponds to user history information, where the user posture information can be obtained by recognizing the user's actions in the image.
- the target user among the at least two users is determined according to the obtained characteristic information of the at least two users.
- the characteristic information of each user can be comprehensively evaluated in combination with the actual scene to determine the target user to be interacted with.
- the interactive object displayed on the display device can be driven to respond to the target user.
- the user when the user is found, after driving the interactive object to respond, by tracking the user detected in the image surrounding the display device, for example, the facial expression of the user can be tracked, and/or, Tracking the user's actions, etc., and judging whether to make the display device enter the service activation state by judging whether the user has actively interacted expressions and/or actions.
- specific trigger information may be set, such as common facial expressions and/or actions for greetings between people, such as blinking, nodding, waving, raising hands, slaps, and so on.
- the specified trigger information set here may be referred to as the first trigger information.
- the display device enters the service activation state, and the interactive object is driven to display the provided service, for example, it can be displayed in language or The text information displayed on the screen is displayed.
- the current common somatosensory interaction requires the user to raise his hand for a period of time to activate the service. After selecting the service, the user needs to keep his hand still for several seconds to complete the activation.
- the interactive method provided by the embodiments of the present disclosure does not require the user to raise the hand for a period of time to activate the service, nor does it need to keep the hand position different to complete the selection.
- the service can be automatically activated and the device is in the service activation state. , Avoiding users raising their hands and waiting for a period of time, improving user experience.
- specific trigger information can be set, such as a specific gesture action, and/or a specific voice command.
- the specified trigger information set here may be referred to as second trigger information.
- the second trigger information In the case of detecting the second trigger information output by the user, it is determined that the display device enters the in-service state, and the interactive object is driven to provide a service matching the second trigger information.
- the corresponding service is executed through the second trigger information output by the user.
- the services that can be provided to users include: the first service option, the second service option, the third service option, etc., and the corresponding second trigger information can be configured for the first service option.
- the voice "one” can be set. "Is the second trigger information corresponding to the first service option, and the voice "two” is set as the second trigger information corresponding to the second service option, and so on.
- the display device is caused to enter the service option corresponding to the second trigger information, and the interactive object is driven to provide the service according to the content set by the service option.
- the first granularity (coarse-grained) identification method is to make the device enter the service activation state and drive the interactive object to display the provided service when the first trigger information output by the user is detected;
- the recognition method is to make the device enter the in-service state in the case of detecting the second trigger information output by the user, and drive the interactive object to provide the corresponding service.
- the user does not need to enter keys, touches, or voice input, and only stand around the display device.
- the interactive objects displayed in the display device can make targeted welcoming actions and follow the user’s instructions.
- the needs or interests of the users show the service items that can be provided, and enhance the user experience.
- the environment information of the display device may be acquired, and the interactive object displayed on the display device can be driven to respond according to the detection result and the environment information.
- the environmental information of the display device may be acquired through the geographic location of the display device and/or the application scenario of the display device.
- the environmental information may be, for example, the geographic location of the display device, an Internet Protocol (IP) address, or the weather, date, etc. of the area where the display device is located.
- IP Internet Protocol
- the interactive object may be driven to respond according to the current service state and environment information of the display device.
- the environmental information includes time, location, and weather conditions, which can drive the interactive objects displayed on the display device to make welcome actions and gestures, or make some interesting actions, and output
- the voice "It is XX time, X month X day, X year X, XX weather, welcome to XX shopping mall in XX city, I am very happy to serve you".
- the current time, location, and weather conditions are also added, which not only provides more information, but also makes the response of interactive objects more in line with interaction needs and more targeted.
- the interactive objects displayed in the display device are driven to respond, so that the response of the interactive objects is more in line with the interaction requirements, and the user The interaction with the interactive objects is more real and vivid, thereby enhancing the user experience.
- a matching predetermined response label may be obtained according to the detection result and the environmental information; then, the interactive object is driven to make a corresponding response according to the response label. This application is not limited to this.
- the response tag may correspond to the driving text of one or more of the action, expression, gesture, and language of the interactive object. For different detection results and environmental information, corresponding driving text can be obtained according to the determined response label, so that the interactive object can be driven to output one or more of corresponding actions, expressions, and languages.
- the corresponding response label may be: the action is a welcome action, and the voice is "Welcome to Shanghai”.
- the corresponding response label can be: the action is welcome action, and the voice is " Good morning, Ms. Zhang, welcome, and I am very happy to serve you.”
- the interactive object By configuring corresponding response labels for the combination of different detection results and different environmental information, and using the response labels to drive the interactive object to output one or more of the corresponding actions, expressions, and languages, the interactive object can be driven according to Different states and different scenarios of the device make different responses, so as to make the responses of the interactive objects more diversified.
- the response tag may be input to a pre-trained neural network, and the driving text corresponding to the response tag may be output, so as to drive the interactive object to output one of corresponding actions, expressions, and language. Or multiple.
- the neural network may be trained by a sample response label set, wherein the sample response label is annotated with corresponding driving text. After being trained, the neural network can output corresponding driving text for the output response label, so as to drive the interactive object to output one or more of corresponding actions, expressions, and languages. Compared with searching the corresponding driving text directly on the display device or the cloud, using a pre-trained neural network, driving text can also be generated for response labels that are not preset with driving text to drive the interactive object to respond appropriately.
- the driving text can be manually configured for the corresponding response label.
- the corresponding driving text is automatically called to drive the interactive object to respond, so that the actions and expressions of the interactive object are more natural.
- the display device in response to the display device being in a user discovery state, obtain the position information of the user relative to the interactive object in the display device according to the position of the user in the image; and The position information adjusts the orientation of the interactive object so that the interactive object faces the user.
- the image of the interactive object is captured by a virtual camera.
- the virtual camera is a virtual software camera applied to 3D software and used to collect images, and the interactive object is displayed on the screen through the 3D image collected by the virtual camera. Therefore, the user's perspective can be understood as the perspective of the virtual camera in the 3D software, which will cause a problem that the interactive objects cannot achieve eye contact between users.
- the line of sight of the interactive object is also kept aligned with the virtual camera. Since the interactive object faces the user during the interaction process, and the line of sight remains aligned with the virtual camera, the user will have the illusion that the interactive object is looking at himself, which can improve the comfort of interaction between the user and the interactive object.
- FIG. 3 shows a schematic structural diagram of an interaction device according to at least one embodiment of the present disclosure.
- the device may include: an image acquisition unit 301, a detection unit 302, a user selection unit 303, and a driving unit 304.
- the image acquisition unit 301 is configured to acquire images around the display device collected by the camera, and the display device displays interactive objects through a transparent display;
- the detection unit 302 is configured to perform operations on one or more users involved in the image. Detection;
- user selection unit 303 in response to the detection unit 302 detects that the image involves at least two users, according to the detected feature information of the at least two users, from the at least two users Select the target user;
- the driving unit 304 is configured to drive the interactive object displayed on the transparent display screen of the display device to respond to the target user based on the detection result of the target user.
- the one or more users in the image described herein refer to one or more objects involved in the detection process of the image. In the following, "object” and "user” can be used interchangeably, and for convenience of presentation, they are collectively referred to as "user”.
- the characteristic information includes user posture information and/or user attribute information.
- the user selection unit 303 is specifically configured to: according to the degree of match between the user posture information of each of the at least two users and the posture of the set posture feature, or according to the at least two users The user attribute information of each user in each user matches the degree of the attribute of the set attribute feature, and the target user is selected from the at least two users.
- the user selection unit 303 is specifically configured to: select one or more first users that meet the set gesture characteristics according to the user gesture information of each of the at least two users; In the case that there are at least two first users, the driving unit 304 is caused to drive the interactive object to guide the at least two first users to output setting information; According to the sequence of the setting information, the target user is determined.
- the user selection unit 303 is specifically configured to: select one or more first users that meet the set gesture characteristics according to the user gesture information of each of the at least two users; In the case that there are at least two first users, determine the respective interaction response priorities of the at least two first users according to the respective user attribute information of the at least two first users; and according to the interaction response The priority determines the target user.
- the device further includes a confirmation unit configured to: in response to the user selection unit 303 selecting a target user from the at least two users, causing the driving unit to drive the The interactive object outputs confirmation information to the target user.
- the device further includes a waiting state unit for: responding to the detection unit 302 not detecting the user from the image at the current moment, and setting the device before the current moment No user is detected from the image and no user is tracked within a certain period of time, the user to be interacted with the interactive object is determined to be empty, and the display device enters a state of waiting for the user.
- a waiting state unit for: responding to the detection unit 302 not detecting the user from the image at the current moment, and setting the device before the current moment No user is detected from the image and no user is tracked within a certain period of time, the user to be interacted with the interactive object is determined to be empty, and the display device enters a state of waiting for the user.
- the device further includes an end state unit for: responding to the detection unit 302 not detecting a user from the image at the current moment, and setting the device before the current moment.
- a user is detected or tracked from the image within a predetermined period of time, and the user to be interacted with the interactive object is determined to be the user who interacted most recently.
- the display device displays the reflection of the interaction object through the transparent display screen, or the display device displays the reflection of the interaction object on the bottom plate.
- the interactive object includes a virtual character with a three-dimensional effect.
- At least one embodiment of the present disclosure also provides an interactive device.
- the device includes a memory 401 and a processor 402.
- the memory 401 is used to store instructions executable by the processor, and when the instructions are executed, the processor 402 is prompted to implement the interaction method described in any embodiment of the present disclosure.
- At least one embodiment of the present disclosure also provides a computer-readable storage medium on which a computer program is stored.
- the computer program When the computer program is executed by a processor, the processor realizes the interaction described in any embodiment of the present disclosure. method.
- one or more embodiments of the present disclosure may be provided as a method, a system, or a computer program product. Therefore, one or more embodiments of the present disclosure may adopt the form of a complete hardware embodiment, a complete software embodiment, or an embodiment combining software and hardware. Moreover, one or more embodiments of the present disclosure may adopt computer programs implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program codes. The form of the product.
- computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
- Embodiments of the subject matter in the present disclosure can be implemented as one or more computer programs, that is, one or more of computer program instructions encoded on a tangible non-transitory program carrier to be executed by a data processing device or to control the operation of the data processing device Modules.
- the program instructions may be encoded on artificially generated propagated signals, such as machine-generated electrical, optical or electromagnetic signals, which are generated to encode information and transmit it to a suitable receiver device for data transmission.
- the processing device executes.
- the computer storage medium may be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of them.
- the processing and logic flow in the present disclosure can be executed by one or more programmable computers executing one or more computer programs to perform corresponding functions by operating according to input data and generating output.
- the processing and logic flow can also be executed by a dedicated logic circuit, such as FPGA (Field Programmable Gate Array) or ASIC (Application Specific Integrated Circuit), and the device can also be implemented as a dedicated logic circuit.
- FPGA Field Programmable Gate Array
- ASIC Application Specific Integrated Circuit
- Computers suitable for executing computer programs include, for example, general-purpose and/or special-purpose microprocessors, or any other type of central processing unit.
- the central processing unit will receive instructions and data from a read-only memory and/or a random access memory.
- the basic components of a computer include a central processing unit for implementing or executing instructions and one or more memory devices for storing instructions and data.
- the computer will also include one or more mass storage devices for storing data, such as magnetic disks, magneto-optical disks, or optical disks, or the computer will be operatively coupled to this mass storage device to receive data from or send data to it. It transmits data, or both.
- the computer does not have to have such equipment.
- the computer can be embedded in another device, such as a mobile phone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a global positioning system (GPS) receiver, or, for example, a universal serial bus (USB ) Flash drives are portable storage devices, just to name a few.
- PDA personal digital assistant
- GPS global positioning system
- USB universal serial bus
- Computer-readable media suitable for storing computer program instructions and data include all forms of non-volatile memory, media, and memory devices, including, for example, semiconductor memory devices (such as EPROM, EEPROM, and flash memory devices), magnetic disks (such as internal hard disks or Removable disks), magneto-optical disks, CD ROM and DVD-ROM disks.
- semiconductor memory devices such as EPROM, EEPROM, and flash memory devices
- magnetic disks such as internal hard disks or Removable disks
- magneto-optical disks CD ROM and DVD-ROM disks.
- the processor and the memory can be supplemented by or incorporated into a dedicated logic circuit.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Databases & Information Systems (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Evolutionary Computation (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- User Interface Of Digital Computer (AREA)
- Transition And Organic Metals Composition Catalysts For Addition Polymerization (AREA)
- Indexing, Searching, Synchronizing, And The Amount Of Synchronization Travel Of Record Carriers (AREA)
- Holo Graphy (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
Claims (22)
- 一种交互方法,所述方法包括:获取摄像头采集的显示设备周边的图像,所述显示设备通过透明显示屏显示交互对象;对所述图像中涉及的一个或多个对象进行检测;响应于检测到所述图像中涉及至少两个对象,根据检测到的所述至少两个对象的特征信息,从所述至少两个对象中选择目标对象;基于对所述目标对象的检测结果,驱动所述显示设备的透明显示屏上显示的所述交互对象对所述目标对象进行回应。
- 根据权利要求1所述的方法,其中,所述特征信息包括对象姿态信息和/或对象属性信息。
- 根据权利要求2所述的方法,其中,所述根据检测到的所述至少两个对象的特征信息,从所述至少两个对象中选择目标对象,包括:根据所述至少两个对象中每个对象的对象姿态信息与设定姿态特征的姿态匹配程度,或,根据所述至少两个对象中每个对象的对象属性信息与设定属性特征的属性匹配程度,从所述至少两个对象中选择所述目标对象。
- 根据权利要求2所述的方法,其中,所述根据检测到的所述至少两个对象的特征信息,从所述至少两个对象中选择目标对象,包括:根据所述至少两个对象中每个对象的对象姿态信息,选取符合设定姿态特征的一个或多个第一对象;在所述第一对象有至少两个的情况下,驱动所述交互对象引导所述至少两个第一对象各自输出设定信息,并根据检测到的所述第一对象各自输出所述设定信息的顺序,确定所述目标对象。
- 根据权利要求2所述的方法,其中,所述根据检测到的所述至少两个对象的特征信息,从所述至少两个对象中选择目标对象,包括:根据所述至少两个对象中每个对象的对象姿态信息,选取符合设定姿态特征的一个或多个第一对象;在所述第一对象有至少两个的情况下,根据所述至少两个第一对象各自的对象属性信息,确定所述至少两个第一对象各自的交互响应优先级,并根据所述交互响应优先级确定所述目标对象。
- 根据权利要求1至5任一项所述的方法,所述方法还包括:在从所述至少两个对象中选择目标对象后,驱动所述交互对象对所述目标对象输出确认信息。
- 根据权利要求1至6任一项所述的方法,所述方法还包括:响应于在当前时刻从所述图像中未检测到对象,且在当前时刻之前的设定时间段内从所述图像中未检测到对象且未追踪到对象,确定所述交互对象的待交互对象为空,并使所述显示设备进入等待对象状态。
- 根据权利要求1至6任一项所述的方法,所述方法还包括:响应于在当前时刻从所述图像中未检测到对象,且在当前时刻之前的设定时间段内从所述图像中检测到对象或追踪到对象,确定所述交互对象的待交互对象为最近一次进行交互的对象。
- 根据权利要求1至8任一项所述的方法,其中,所述显示设备通过所述透明显示屏显示所述交互对象的倒影,或者,所述显示设备在底板上显示所述交互对象的倒影。
- 根据权利要求1至9任一项所述的方法,其中,所述交互对象包括具有立体效果的虚拟人物。
- 一种交互装置,所述装置包括:图像获取单元,用于获取摄像头采集的显示设备周边的图像,所述显示设备通过透明显示屏显示交互对象;检测单元,用于对所述图像中涉及的一个或多个对象进行检测;对象选择单元,用于响应于所述检测单元检测到所述图像中涉及至少两个对象,根据检测到的所述至少两个对象的特征信息,从所述至少两个对象中选择目标对象;驱动单元,用于基于对所述目标对象的检测结果,驱动所述显示设备的透明显示屏上显示的所述交互对象对所述目标对象进行回应。
- 根据权利要求11所述的装置,其中,所述特征信息包括对象姿态信息和/或对象属性信息。
- 根据权利要求12所述的装置,其中,所述对象选择单元用于:根据所述至少两个对象中每个对象的对象姿态信息与设定姿态特征的姿态匹配程度,或,根据所述至少两个对象中每个对象的对象属性信息与设定属性特征的属性匹配程度,从所述至少两个对象中选择所述目标对象。
- 根据权利要求12所述的装置,其中,所述对象选择单元用于:根据所述至少两个对象中每个对象的对象姿态信息,选取符合设定姿态特征的一个或多个第一对象;在所述第一对象有至少两个的情况下,使所述驱动单元驱动所述交互对象引导所述至少两个第一对象各自输出设定信息,并根据检测到的所述第一对象各自输出所述设定信息的顺序,确定所述目标对象。
- 根据权利要求12所述的装置,其中,所述对象选择单元用于:根据所述至少两个对象中每个对象的对象姿态信息,选取符合设定姿态特征的一个或多个第一对象;在所述第一对象有至少两个的情况下,根据所述至少连个第一对象各自的对象属性信息,确定所述至少两个第一对象各自的交互响应优先级,并根据所述交互响应优先级确定所述目标对象。
- 根据权利要求11至15任一项所述的装置,其中,所述装置还包括确认单元,所述确认单元用于:响应于所述对象选择单元从所述至少两个对象中选择了目标对象,使所述驱动单元驱动所述交互对象对所述目标对象输出确认信息。
- 根据权利要求11至16任一项所述的装置,其特征在于,所述装置还包括等待状态单元,所述等待状态单元用于:响应于所述检测单元在当前时刻从所述图像中未检测到对象,且在当前时刻之前的设定时间段内从所述图像中未检测到对象且未追踪到对象,确定所述交互对象的待交互对象为空,并使所述显示设备进入等待对象状态。
- 根据权利要求11至16任一项所述的装置,其中,所述装置还包括结束状态单元,所述结束状态单元用于:响应于所述检测单元在当前时刻从所述图像中未检测到对象,且在当前时刻之前的设定时间段内从所述图像中检测到对象或追踪到对象,确定所述交互对象的待交互对象为最近一次进行交互的对象。
- 根据权利要求11至18任一项所述的装置,其中,所述显示设备通过所述透明显示屏显示所述交互对象的倒影,或者,所述显示设备在底板上显示所述交互对象的倒影。
- 根据权利要求11至19任一项所述的装置,其中,所述交互对象包括具有立体效果的虚拟人物。
- 一种交互设备,所述设备包括:处理器;以及用于存储可由所述处理器执行的指令的存储器,其中,所述指令在被执行时,促使所述处理器实现根据权利要求1至10任一项所述的交互方法。
- 一种计算机可读存储介质,其上存储有计算机程序,其中,所述计算机程序被处理器执行时,使所述处理器实现根据权利要求1至10任一项所述的交互方法。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021556968A JP7224488B2 (ja) | 2019-08-28 | 2020-07-24 | インタラクティブ方法、装置、デバイス、及び記憶媒体 |
KR1020217031185A KR102707660B1 (ko) | 2019-08-28 | 2020-07-24 | 인터렉티브 방법, 장치, 디바이스 및 기록 매체 |
US17/681,026 US20220179609A1 (en) | 2019-08-28 | 2022-02-25 | Interaction method, apparatus and device and storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910803899.3 | 2019-08-28 | ||
CN201910803899.3A CN110716634A (zh) | 2019-08-28 | 2019-08-28 | 交互方法、装置、设备以及显示设备 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/681,026 Continuation US20220179609A1 (en) | 2019-08-28 | 2022-02-25 | Interaction method, apparatus and device and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021036624A1 true WO2021036624A1 (zh) | 2021-03-04 |
Family
ID=69209574
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/104466 WO2021036624A1 (zh) | 2019-08-28 | 2020-07-24 | 交互方法、装置、设备以及存储介质 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20220179609A1 (zh) |
JP (1) | JP7224488B2 (zh) |
CN (1) | CN110716634A (zh) |
TW (1) | TWI775134B (zh) |
WO (1) | WO2021036624A1 (zh) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110716634A (zh) * | 2019-08-28 | 2020-01-21 | 北京市商汤科技开发有限公司 | 交互方法、装置、设备以及显示设备 |
CN110716641B (zh) * | 2019-08-28 | 2021-07-23 | 北京市商汤科技开发有限公司 | 交互方法、装置、设备以及存储介质 |
CN111443801B (zh) * | 2020-03-25 | 2023-10-13 | 北京百度网讯科技有限公司 | 人机交互方法、装置、设备及存储介质 |
CN111459452B (zh) * | 2020-03-31 | 2023-07-18 | 北京市商汤科技开发有限公司 | 交互对象的驱动方法、装置、设备以及存储介质 |
CN111627097B (zh) * | 2020-06-01 | 2023-12-01 | 上海商汤智能科技有限公司 | 一种虚拟景物的展示方法及装置 |
CN111640197A (zh) * | 2020-06-09 | 2020-09-08 | 上海商汤智能科技有限公司 | 一种增强现实ar特效控制方法、装置及设备 |
CN116528046A (zh) * | 2020-11-09 | 2023-08-01 | 华为技术有限公司 | 目标用户追焦拍摄方法、电子设备及存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102221886A (zh) * | 2010-06-11 | 2011-10-19 | 微软公司 | 通过化身与用户界面交互 |
EP2919094A1 (en) * | 2014-03-10 | 2015-09-16 | BAE Systems PLC | Interactive information display |
CN106325517A (zh) * | 2016-08-29 | 2017-01-11 | 袁超 | 一种基于虚拟现实的目标对象触发方法、系统和穿戴设备 |
CN107728782A (zh) * | 2017-09-21 | 2018-02-23 | 广州数娱信息科技有限公司 | 交互方法及交互系统、服务器 |
CN106203364B (zh) * | 2016-07-14 | 2019-05-24 | 广州帕克西软件开发有限公司 | 一种3d眼镜互动试戴系统及方法 |
CN110716634A (zh) * | 2019-08-28 | 2020-01-21 | 北京市商汤科技开发有限公司 | 交互方法、装置、设备以及显示设备 |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6720949B1 (en) * | 1997-08-22 | 2004-04-13 | Timothy R. Pryor | Man machine interfaces and applications |
JP2005189426A (ja) | 2003-12-25 | 2005-07-14 | Nippon Telegr & Teleph Corp <Ntt> | 情報表示装置および情報入出力装置 |
US8555207B2 (en) | 2008-02-27 | 2013-10-08 | Qualcomm Incorporated | Enhanced input using recognized gestures |
JP6322927B2 (ja) | 2013-08-14 | 2018-05-16 | 富士通株式会社 | インタラクション装置、インタラクションプログラムおよびインタラクション方法 |
TW201614423A (en) * | 2014-10-03 | 2016-04-16 | Univ Southern Taiwan Sci & Tec | Operation system for somatosensory device |
CN104978029B (zh) * | 2015-06-30 | 2018-11-23 | 北京嘿哈科技有限公司 | 一种屏幕操控方法及装置 |
KR20170029320A (ko) * | 2015-09-07 | 2017-03-15 | 엘지전자 주식회사 | 이동 단말기 및 그 제어방법 |
WO2017086108A1 (ja) | 2015-11-16 | 2017-05-26 | 大日本印刷株式会社 | 情報提示装置、情報提示方法、プログラム、情報処理装置及び案内ロボット制御システム |
JP6768597B2 (ja) * | 2017-06-08 | 2020-10-14 | 株式会社日立製作所 | 対話システム、対話システムの制御方法、及び装置 |
CN107728780B (zh) * | 2017-09-18 | 2021-04-27 | 北京光年无限科技有限公司 | 一种基于虚拟机器人的人机交互方法及装置 |
CN108153425A (zh) * | 2018-01-25 | 2018-06-12 | 余方 | 一种基于全息投影的互动娱乐系统和方法 |
CN108780361A (zh) * | 2018-02-05 | 2018-11-09 | 深圳前海达闼云端智能科技有限公司 | 人机交互方法、装置、机器人及计算机可读存储介质 |
CN108415561A (zh) * | 2018-02-11 | 2018-08-17 | 北京光年无限科技有限公司 | 基于虚拟人的手势交互方法及系统 |
CN108470205A (zh) * | 2018-02-11 | 2018-08-31 | 北京光年无限科技有限公司 | 基于虚拟人的头部交互方法及系统 |
CN108363492B (zh) * | 2018-03-09 | 2021-06-25 | 南京阿凡达机器人科技有限公司 | 一种人机交互方法及交互机器人 |
CN108682202A (zh) * | 2018-04-27 | 2018-10-19 | 伍伟权 | 一种文科用全息投影教学设备 |
CN109522790A (zh) * | 2018-10-08 | 2019-03-26 | 百度在线网络技术(北京)有限公司 | 人体属性识别方法、装置、存储介质及电子设备 |
CN109739350A (zh) * | 2018-12-24 | 2019-05-10 | 武汉西山艺创文化有限公司 | 基于透明液晶显示屏的ai智能助理设备及其交互方法 |
CN110119197A (zh) * | 2019-01-08 | 2019-08-13 | 佛山市磁眼科技有限公司 | 一种全息互动系统 |
-
2019
- 2019-08-28 CN CN201910803899.3A patent/CN110716634A/zh active Pending
-
2020
- 2020-07-24 WO PCT/CN2020/104466 patent/WO2021036624A1/zh active Application Filing
- 2020-07-24 JP JP2021556968A patent/JP7224488B2/ja active Active
- 2020-08-25 TW TW109128905A patent/TWI775134B/zh active
-
2022
- 2022-02-25 US US17/681,026 patent/US20220179609A1/en not_active Abandoned
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102221886A (zh) * | 2010-06-11 | 2011-10-19 | 微软公司 | 通过化身与用户界面交互 |
EP2919094A1 (en) * | 2014-03-10 | 2015-09-16 | BAE Systems PLC | Interactive information display |
CN106203364B (zh) * | 2016-07-14 | 2019-05-24 | 广州帕克西软件开发有限公司 | 一种3d眼镜互动试戴系统及方法 |
CN106325517A (zh) * | 2016-08-29 | 2017-01-11 | 袁超 | 一种基于虚拟现实的目标对象触发方法、系统和穿戴设备 |
CN107728782A (zh) * | 2017-09-21 | 2018-02-23 | 广州数娱信息科技有限公司 | 交互方法及交互系统、服务器 |
CN110716634A (zh) * | 2019-08-28 | 2020-01-21 | 北京市商汤科技开发有限公司 | 交互方法、装置、设备以及显示设备 |
Also Published As
Publication number | Publication date |
---|---|
CN110716634A (zh) | 2020-01-21 |
TWI775134B (zh) | 2022-08-21 |
TW202109246A (zh) | 2021-03-01 |
KR20210131415A (ko) | 2021-11-02 |
JP2022526772A (ja) | 2022-05-26 |
JP7224488B2 (ja) | 2023-02-17 |
US20220179609A1 (en) | 2022-06-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021036624A1 (zh) | 交互方法、装置、设备以及存储介质 | |
WO2021036622A1 (zh) | 交互方法、装置、设备以及存储介质 | |
JP7483798B2 (ja) | ワードフロー注釈 | |
US10817760B2 (en) | Associating semantic identifiers with objects | |
CN109635621B (zh) | 用于第一人称视角中基于深度学习识别手势的系统和方法 | |
EP2877254B1 (en) | Method and apparatus for controlling augmented reality | |
US9280972B2 (en) | Speech to text conversion | |
EP2912659B1 (en) | Augmenting speech recognition with depth imaging | |
KR101832693B1 (ko) | 직관적 컴퓨팅 방법들 및 시스템들 | |
JP2019197499A (ja) | プログラム、記録媒体、拡張現実感提示装置及び拡張現実感提示方法 | |
CN109815409A (zh) | 一种信息的推送方法、装置、穿戴设备及存储介质 | |
JP2022531055A (ja) | インタラクティブ対象の駆動方法、装置、デバイス、及び記録媒体 | |
KR102707660B1 (ko) | 인터렉티브 방법, 장치, 디바이스 및 기록 매체 | |
KR20220111716A (ko) | 디바이스 로컬리제이션을 위한 디바이스 및 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20859149 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021556968 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20217031185 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20859149 Country of ref document: EP Kind code of ref document: A1 |