CN107957781B - Information display method and device - Google Patents

Information display method and device Download PDF

Info

Publication number
CN107957781B
CN107957781B CN201711326613.4A CN201711326613A CN107957781B CN 107957781 B CN107957781 B CN 107957781B CN 201711326613 A CN201711326613 A CN 201711326613A CN 107957781 B CN107957781 B CN 107957781B
Authority
CN
China
Prior art keywords
input device
cursor
environment
display
environment picture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711326613.4A
Other languages
Chinese (zh)
Other versions
CN107957781A (en
Inventor
南迪尔
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Xiaomi Mobile Software Co Ltd
Original Assignee
Beijing Xiaomi Mobile Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Xiaomi Mobile Software Co Ltd filed Critical Beijing Xiaomi Mobile Software Co Ltd
Priority to CN201711326613.4A priority Critical patent/CN107957781B/en
Publication of CN107957781A publication Critical patent/CN107957781A/en
Application granted granted Critical
Publication of CN107957781B publication Critical patent/CN107957781B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04812Interaction techniques based on cursor appearance or behaviour, e.g. being affected by the presence of displayed objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality

Abstract

The invention discloses an information display method and device, belonging to the field of virtual reality, wherein the method comprises the following steps: detecting whether an environment picture currently displayed by display equipment contains an input equipment cursor corresponding to the input equipment or not, wherein the environment picture currently displayed by the display equipment is a part of a three-dimensional virtual environment generated by a VR host; when the environment picture does not contain the input equipment cursor, determining the relative position of the input equipment cursor and the environment picture; and displaying the indication information in the environment picture according to the relative position. In the embodiment of the disclosure, the indication information for indicating the relative position of the cursor of the input device is automatically displayed, so that the user can purposefully adjust the orientation of the input device, and the efficiency of moving the cursor of the input device into the current environment picture is further improved.

Description

Information display method and device
Technical Field
The embodiment of the disclosure relates to the field of Virtual Reality (VR), and in particular, to an information display method and apparatus.
Background
VR systems, as an immersive interactive system, typically include a VR host, a display device, and an input device. The VR host is used for rendering and generating a three-dimensional virtual environment, the display device is used for displaying an environment picture of the three-dimensional virtual environment for a user, and the input device is used for controlling a virtual object in the three-dimensional virtual environment.
Because the environment picture displayed by the display device is limited, when the input device cursor corresponding to the input device in the three-dimensional virtual environment is outside the environment picture currently displayed by the display device, the user needs to manually adjust the orientation of the input device, so as to move the input device cursor into the current environment picture.
Disclosure of Invention
In order to solve the problem that a user needs to try many times to move an input device cursor to a current environment picture under the condition that the user cannot know the position of the input device cursor outside the current environment picture, and the moving efficiency is low, the embodiment of the disclosure provides an information display method and an information display device, and the technical scheme is as follows:
in a first aspect, an information display method is provided, and the method includes:
detecting whether an environment picture currently displayed by display equipment contains an input equipment cursor corresponding to the input equipment or not, wherein the environment picture currently displayed by the display equipment is a part of a three-dimensional virtual environment generated by a VR host;
when the environment picture does not contain the input equipment cursor, determining the relative position of the input equipment cursor and the environment picture;
and controlling the display device to display the indication information in the environment picture according to the relative position.
Optionally, detecting whether an environment picture currently displayed by the display device includes an input device cursor corresponding to the input device includes:
acquiring a coordinate range of an environment picture in a three-dimensional virtual environment;
acquiring a first coordinate of a cursor of an input device in a three-dimensional virtual environment;
and if no intersection exists between the first coordinate and the coordinate range, determining that the environment picture does not contain the input equipment cursor.
Optionally, determining the relative position of the input device cursor and the environment picture includes:
acquiring a first coordinate of a cursor of an input device in a three-dimensional virtual environment;
acquiring a second coordinate of the midpoint of the environment picture in the three-dimensional virtual environment;
and determining the relative position of the cursor of the input device and the environment picture according to the first coordinate and the second coordinate.
Optionally, before determining the relative position of the input device cursor and the environment picture, the method further includes:
acquiring sensor data sent by input equipment, wherein the sensor data is acquired by a motion sensor in the input equipment, and the motion sensor is an angular velocity sensor and/or an acceleration sensor;
calculating a vibration frequency of the input device from the sensor data;
determining a relative position of an input device cursor and an environment picture, comprising:
and when the environment picture does not contain the input device cursor and the vibration frequency of the input device is greater than the frequency threshold value, determining the relative position of the input device cursor and the environment picture.
Optionally, before determining the relative position of the input device cursor and the environment picture, the method further includes:
acquiring the continuous use duration of the input equipment;
determining a relative position of an input device cursor and an environment picture, comprising:
and when the environment picture does not contain the input equipment cursor and the continuous use time of the input equipment is greater than the time length threshold value, determining the relative position of the input equipment cursor and the environment picture.
Optionally, controlling the display device to display the indication information in the environment picture according to the relative position includes:
generating a directional arrow according to the relative position, wherein the direction of the directional arrow points to an input device cursor from the center of the environment picture;
and controlling the display device to display a directional arrow at a preset position in the environment picture.
Optionally, the method further includes:
and when the display duration of the indication information is greater than the threshold value, or when the input equipment cursor is detected to move to the environment picture, controlling the display equipment to stop displaying the indication information.
In a second aspect, there is provided an information display apparatus comprising:
the detection module is configured to detect whether an environment picture currently displayed by the display device contains an input device cursor corresponding to the input device, wherein the environment picture currently displayed by the display device is a part of a three-dimensional virtual environment generated by the VR host;
a determining module configured to determine a relative position of the input device cursor and the environment screen when the input device cursor is not included in the environment screen;
and the display module is configured to control the display device to display the indication information in the environment picture according to the relative position.
Optionally, the detection module includes:
the first obtaining sub-module is configured to obtain a coordinate range of the environment picture in the three-dimensional virtual environment;
the second acquisition submodule is configured to acquire a first coordinate of a cursor of the input device in the three-dimensional virtual environment;
and the first determining submodule is configured to determine that the environment picture does not contain the input device cursor if no intersection exists between the first coordinate and the coordinate range.
Optionally, the determining module includes:
a third obtaining submodule configured to obtain a first coordinate of a cursor of the input device in the three-dimensional virtual environment;
a fourth obtaining sub-module configured to obtain a second coordinate of the midpoint of the environment picture in the three-dimensional virtual environment;
and the second determining submodule is configured to determine the relative position of the input device cursor and the environment picture according to the first coordinate and the second coordinate.
Optionally, the apparatus further includes:
the device comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is configured to acquire sensor data sent by the input device, the sensor data is acquired by a motion sensor in the input device, and the motion sensor is an angular velocity sensor and/or an acceleration sensor;
a calculation module configured to calculate a vibration frequency of the input device from the sensor data;
the determining module is further configured to determine a relative position of the input device cursor and the environment picture when the environment picture does not contain the input device cursor and the vibration frequency of the input device is greater than the frequency threshold.
Optionally, the apparatus further includes:
the second acquisition module is configured to acquire the continuous use time length of the input device;
and the determining module is further configured to determine the relative position of the input device cursor and the environment picture when the environment picture does not contain the input device cursor and the continuous use time of the input device is greater than the time threshold.
Optionally, the display module includes:
the generation submodule is configured to generate a directional arrow according to the relative position, and the direction of the directional arrow is that the center of the environment picture points to the cursor of the input device;
a control sub-module configured to control the display device to display a directional arrow at a predetermined position within the environment screen.
Optionally, the apparatus further includes:
and the control module is configured to control the display device to stop displaying the indication information when the display duration of the indication information is greater than a threshold value or the input device cursor is detected to move to the environment screen.
In a third aspect, there is provided an information display apparatus comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to:
detecting whether an environment picture currently displayed by display equipment contains an input equipment cursor corresponding to the input equipment or not, wherein the environment picture currently displayed by the display equipment is a part of a three-dimensional virtual environment generated by a VR host;
when the environment picture does not contain the input equipment cursor, determining the relative position of the input equipment cursor and the environment picture;
and controlling the display device to display the indication information in the environment picture according to the relative position.
In a fourth aspect, there is provided a computer readable medium having stored thereon program instructions which, when executed by a processor, implement the information display method as described in the first aspect above.
The technical scheme provided by the embodiment of the disclosure has the following beneficial effects:
when the VR host detects that the cursor of the input device is outside the environment picture currently displayed by the display device, the relative position of the cursor of the input device and the environment picture in the three-dimensional virtual environment is determined, and the display device is controlled to display corresponding indication information in the environment picture according to the relative position, so that a user can quickly move the cursor of the input device to the environment picture according to the indication information; compared with the prior art that the user can only adjust the position of the input device blindly, in the embodiment of the disclosure, the user can purposefully adjust the orientation of the input device by automatically displaying the indication information for indicating the relative position of the cursor of the input device, so that the efficiency of moving the cursor of the input device to the current environment picture is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic diagram of a VR system provided by an exemplary embodiment of the present disclosure;
FIG. 2 is a flow chart of an information display method provided by an exemplary embodiment of the present disclosure;
FIG. 3 is a flow chart of an information display method provided by another exemplary embodiment of the present disclosure;
FIG. 4 is a schematic diagram of a spatial rectangular coordinate system provided by an exemplary embodiment of the present disclosure;
FIG. 5 is a schematic illustration of an interface display provided by an exemplary embodiment of the present disclosure;
FIG. 6 is a flow chart of an information display method provided by another exemplary embodiment of the present disclosure;
FIG. 7 is a flow chart of an information display method provided by another exemplary embodiment of the present disclosure;
fig. 8 is a structural diagram of an information display apparatus provided in an exemplary embodiment of the present disclosure;
fig. 9 is a block diagram of a VR host provided in an example embodiment of the present disclosure.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, embodiments of the present invention will be described in detail with reference to the accompanying drawings.
Please refer to fig. 1, which illustrates a schematic structural diagram of a VR system according to an embodiment of the present disclosure. The VR system includes: VR host 120, display device 140, and input device 160.
The VR host 120 is used for modeling a three-dimensional virtual environment, generating a three-dimensional display screen corresponding to the three-dimensional virtual environment, generating a virtual object in the three-dimensional virtual environment, and the like. Certainly, the VR host 120 may also model a two-dimensional virtual environment, generate a two-dimensional display frame corresponding to the two-dimensional virtual environment, and generate a virtual object in the two-dimensional virtual environment; alternatively, the VR host 120 may model a three-dimensional virtual environment, generate a two-dimensional display screen corresponding to the three-dimensional virtual environment according to the viewing position of the user, generate a two-dimensional projection screen of a virtual object in the three-dimensional virtual environment, and the like, which is not limited in this embodiment.
VR host 120 receives an input signal from input device 160 and displays an input device cursor, which may be an arrow, cross, or virtual hand, icon in a three-dimensional virtual environment based on the input signal. VR host 120 is typically implemented by a processor, memory, image VR host, etc. electronics disposed on a circuit board. Optionally, the VR host 120 further includes an image capturing device for capturing the head movement of the user and changing the display in the display device 140 according to the head movement of the user. The information display method provided by the various embodiments of the present disclosure is used in the VR host 120.
The display device 140 is a display for wearing on the head of the user to display images. The display device 140 generally includes a wearing portion including temples and an elastic band for wearing the display device 140 on the head of a user, and a display portion including a left-eye display screen and a right-eye display screen. Optionally, the display device 140 may display different images on the left-eye display screen and the right-eye display screen, so as to simulate a three-dimensional virtual environment for the user; or directly display an environment screen of the three-dimensional virtual environment generated by the VR host 120. In this embodiment, an example is described in which the display device 140 directly displays an environment screen of the three-dimensional virtual environment generated by the VR host 120, where the environment screen displayed on the display device 140 is only a partial screen in the three-dimensional virtual environment within the field of view because the display range of the display device 140 is limited.
Optionally, a motion sensor is disposed on the display device 140 for capturing head movements of the user, so that the VR host 120 changes the environment picture displayed in the display device 140 according to the head movements of the user.
The display device 140 is electrically connected to the VR host 120 through a flexible circuit board or a hardware interface or a data line or a wireless network.
Input device 160 is an input peripheral for controlling virtual objects in a three-dimensional virtual environment. The device can be at least one of a somatosensory glove, a somatosensory handle, a remote controller, a treadmill, a mouse, a keyboard and a human eye focusing device. Input device 160 typically includes physical keys for enabling activation and/or deactivation of the input device, for enabling detection of whether a user is holding the input device, for invoking a menu bar, etc., and the embodiments are not further listed herein.
Optionally, part or all of the physical keys may be implemented as virtual keys implemented by a touch screen, which is not limited in this embodiment.
Optionally, a motion sensor is disposed on the input device 160, and is configured to acquire a motion state of the input device 160 and send the motion state to the VR host 120 in the form of sensor data, so that the VR host 120 adjusts a position of a cursor of the input device according to the sensor data. The motion sensor may be any one of an acceleration sensor and an angular velocity sensor, and the number of each type of motion sensor may be one or more, which is not limited in this embodiment.
The input device 160 is connected to the VR host 120 via cable, Bluetooth, or Wi-Fi (Wireless-Fidelity).
It should be noted that VR host 120 may be integrated within display device 140 or integrated in another device different from display device 140. In this embodiment, the VR host 120 is integrated in another device different from the display device 140. The other device may be a desktop computer or a server, which is not limited in this embodiment.
Fig. 2 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. This embodiment is exemplified by applying the method to the VR host 120 shown in fig. 1, and the information display method includes the following steps:
in step 201, it is detected whether an environment screen currently displayed by the display device includes an input device cursor corresponding to the input device, where the environment screen currently displayed by the display device is a part of the three-dimensional virtual environment generated by the VR host.
Because the display range of the display device is limited (usually, the display range of the display device facing a predetermined direction and with a predetermined size is a display range simulating the visual range of human eyes), in the normal use process, when the spatial position of the input device is changed and leaves the display range, the environment picture currently displayed by the display device does not contain a cursor corresponding to the input device; or, because the VR host integrates data collected by the inertial sensor in the input device to determine the position of the input device, as the continuous operating time of the input device increases, errors generated by integral calculation are accumulated, which causes an angle deviation between the cursor of the input device in the three-dimensional virtual environment and the actual position of the input device, and causes the cursor of the input device that should be originally displayed in the environment picture to be displayed outside the environment picture.
To identify whether the input device cursor is outside of the currently displayed environmental frame, in one possible implementation, the VR host detects whether the input device cursor is in the environmental frame based on a range of coordinates of the currently displayed environmental frame of the display device and coordinates of the input device cursor.
In step 202, when the input device cursor is not included in the environment screen, the relative position of the input device cursor to the environment screen is determined.
After determining that the environment picture does not contain the input device cursor, the VR host further determines the relative position of the input device cursor and the environment picture in order to indicate the position of the input device cursor outside the environment picture.
Optionally, the VR host determines the relative position of the cursor of the input device and the environment picture according to the coordinate of the cursor of the input device in the three-dimensional virtual environment and the coordinate of the environment picture (a certain reference point) in the three-dimensional virtual environment.
In step 203, the display device is controlled to display the indication information in the environment screen according to the relative position.
In order to facilitate the user to adjust the position of the input device with a target, the input device cursor is moved to the currently displayed environment picture, and after the relative position of the input device cursor and the environment picture is determined, the VR host controls the display device to display corresponding indication information according to the relative position.
The indication information is used for indicating the relative direction of the input device cursor and the environment picture, and the indication information can be in the form of characters, icons or animations.
In summary, in the embodiment of the present disclosure, when the VR host detects that the cursor of the input device is outside the environment picture currently displayed by the display device, the VR host determines the relative position of the cursor of the input device and the environment picture in the three-dimensional virtual environment, and controls the display device to display corresponding indication information in the environment picture according to the relative position, so that the user can quickly move the cursor of the input device to the environment picture according to the indication information; compared with the prior art that the user can only adjust the position of the input device blindly, in the embodiment of the disclosure, the user can purposefully adjust the orientation of the input device by automatically displaying the indication information for indicating the relative position of the cursor of the input device, so that the efficiency of moving the cursor of the input device to the current environment picture is improved.
Fig. 3 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. This embodiment is exemplified by applying the method to the VR host 120 shown in fig. 1, and the information display method includes the following steps:
in step 301, a coordinate range of an environment picture currently displayed by a display device in the three-dimensional virtual environment is obtained, where the environment picture currently displayed by the display device is a part of the three-dimensional virtual environment generated by the VR host.
In a possible implementation manner, a spatial rectangular coordinate system based on a three-dimensional virtual environment is pre-stored in the VR host, and when it is detected that an environment picture is displayed in the display device, a coordinate range of the environment picture in the three-dimensional virtual environment can be obtained, that is, a set of coordinates of all pixel points in the environment picture in the spatial rectangular coordinate system, where when a posture of the display device changes, a coordinate range of the currently displayed environment picture also changes correspondingly.
Schematically, as shown in fig. 4, the x-coordinate range of the environment screen 41 is x e [70,150], the y-coordinate range is y e [70,110], and the z-coordinate range is z e [20,140], and further, all the coordinates satisfying the above coordinate ranges are determined as the coordinate ranges of the environment screen 41 in the three-dimensional virtual environment.
In step 302, first coordinates of an input device cursor in a three-dimensional virtual environment are obtained.
After the input device is connected to the VR host, the VR host determines an initial coordinate of a cursor of the input device in the three-dimensional virtual environment. Furthermore, in the movement process of the input device, the VR host acquires sensor data acquired by a movement sensor in the input device, and determines the movement direction and the posture of the input device according to the sensor data, so that actions in the real environment are mapped to the three-dimensional virtual environment. Further, the VR host can acquire a first coordinate of the input device cursor in the three-dimensional virtual environment.
Illustratively, as shown in FIG. 4, the VR host determines that the input device cursor B has a first coordinate of (-30,15, 20).
In step 303, if there is no intersection between the first coordinate and the coordinate range, it is determined that the environment screen does not include the input device cursor.
After a first coordinate corresponding to the input equipment and a coordinate range of the environment picture are determined, the VR host continues to judge whether an intersection exists between the first coordinate and the coordinate range, and when the intersection does not exist between the first coordinate and the coordinate range of the environment picture, the VR host determines that the environment picture does not contain an input equipment cursor; when the first coordinate and the coordinate range of the environment picture have intersection, the VR host determines that the environment picture contains an input device cursor.
In connection with the example in step 301 and step 302, as shown in fig. 4, since the first coordinate of the input device cursor B is not within the coordinate range of the environment screen 41, the VR host determines that there is no intersection between the first coordinate and the coordinate range, and further determines that the environment screen 41 does not include the input device cursor.
In step 304, when the input device cursor is not included in the environment picture, a second coordinate of the midpoint of the environment picture in the three-dimensional virtual environment is acquired.
When the input device cursor is determined to be out of the currently displayed environment picture, the VR host further determines the relative position between the input device cursor and the environment picture according to the coordinates of a certain reference point in the environment picture. Optionally, the reference point is a central point of the environment picture.
In one possible implementation, after determining the coordinate range of the environmental picture, the VR host determines a median of the x, y, and z coordinate ranges to be the x, y, and z coordinates of a midpoint of the environmental picture.
In connection with the example in step 303, as shown in fig. 4, the VR host determines that the coordinate of the point a in the environment picture is the median 110 of [70,150], the coordinate of the point a in the environment picture is the median 90 of [70,110], and the coordinate of the point a in the environment picture is the median 80 of [20,140 ].
In step 305, a relative position of the input device cursor to the environmental screen is determined based on the first coordinate and the second coordinate.
After a first coordinate corresponding to the input device cursor and a second coordinate corresponding to the middle point of the environment picture are obtained, the VR host determines the relative position of the input device cursor and the environment picture according to the two coordinates.
Optionally, the VR host calculates a direction vector according to the first coordinate and the second coordinate, where the direction vector is the second coordinate — the first coordinate, and further, the VR host determines the direction of the direction vector as the relative direction between the input device cursor and the environment frame, and determines the size of the direction vector as the distance between the input device cursor and the middle point of the environment frame.
Schematically, as shown in FIG. 4, the coordinates of the point A in the environment frameIs (110,90,80), the input device cursor B has coordinates of (-30,15,20), and further, the VR host calculates a direction vector
Figure BDA0001505785270000091
And to direct the vector
Figure BDA0001505785270000092
Is determined as the relative direction of the cursor of the input device and the environment picture, and the direction vector
Figure BDA0001505785270000101
Is determined as the distance of the input device cursor from the midpoint of the environmental picture.
In step 306, a directional arrow is generated according to the relative position, and the direction of the directional arrow is from the center of the environment screen to the input device cursor.
And after the VR host determines the relative position of the input device cursor and the environment picture, generating a directional arrow pointing to the input device cursor from the center of the environment picture according to the relative position.
In step 307, the display device is controlled to display a directional arrow at a predetermined position within the environment screen.
And the VR host determines a preset position displayed in the environment picture according to the direction pointed by the directional arrow, and then controls the display equipment to display the directional arrow at the preset position in the environment picture. Optionally, the predetermined position is a left edge, a right edge, an upper edge, a lower edge, or the like of the environment picture.
For example, when the VR host detects that the directional arrow points to the right side of the environment picture, the VR host controls the display device to display the directional arrow at the right edge of the screen picture; and when the VR host detects that the directional arrow points to the left side of the environment picture, controlling the display device to display the directional arrow at the left edge of the screen picture.
In connection with the example in step 304, schematically, as shown in fig. 5, a directional arrow 52 is displayed at the left edge of the environment screen 51.
After the directional arrow 52 is displayed in the environment picture 51, in order to move the input device cursor to the current environment picture, the user moves the input device in the reverse direction along the direction indicated by the directional arrow 52, and in real time, the VR host receives sensor data acquired by the motion sensor in the input device, and then moves the input device cursor to the environment picture 51 according to the position change of the input device in the moving process.
In step 308, when the display duration of the indication information is greater than the threshold value, or when it is detected that the cursor of the input device moves to the environment screen, the display device is controlled to stop displaying the indication information.
In order to avoid that the long-time display of the prompt information shields the picture and influences the user to carry out immersive interaction, the VR host acquires the display duration of the indication information, and when the display duration is greater than a threshold value, the display equipment is controlled to stop displaying the indication information. In this embodiment, the indication information is a directional arrow. The threshold is 5s, and optionally, the threshold can be set by the user in the setting interface.
In a possible implementation manner, after the prompt information is displayed in the environment picture, according to the adjustment condition of the orientation of the input device, the VR host detects whether the cursor of the input device moves into the environment picture in real time, and controls the display device to stop displaying the indication information after the VR host detects that the cursor of the input device moves into the environment picture.
In this embodiment, when the input device cursor is not within the display range, the VR host determines the relative position of the input device cursor and the environment picture according to the coordinates of the input device cursor and the environment picture in the first spatial rectangular coordinate system; and then, the VR host controls the display device to display a directional arrow pointing to the cursor of the input device from the center of the environment picture at a preset position of the environment picture according to the relative position. The user can purposefully change the position of the input device according to the directional arrow, so that the cursor of the input device is displayed in the environment picture. The efficiency of displaying the input device cursor within the environmental picture is improved.
In this embodiment, when the display duration of the indication information reaches the threshold, or when it is detected that the cursor of the input device moves to the environment screen, the VR host automatically stops displaying the indication information, so that the influence of the long-time display of the indication information on the immersive experience of the user is avoided.
In general, since the input device cursor is not displayed in the environment screen, that is, the user may shake the input device tentatively when the user cannot see the input device cursor, in the following embodiments, the example of triggering the display of the instruction information using the shake of the input device as a trigger condition is described.
In a possible implementation manner, as shown in fig. 6, step 309 to step 311 are further included before step 304.
In step 309, sensor data sent by the input device is obtained, the sensor data is collected by a motion sensor in the input device, and the motion sensor is an angular velocity sensor and/or an acceleration sensor.
Optionally, when the motion sensor in the input device is an angular velocity sensor, the sensor data sent by the input device to the VR host is angular velocity data, where the angular velocity data includes an angular velocity direction and an angular velocity magnitude; accordingly, the VR host obtains angular velocity data sent by the input device.
Optionally, when a motion sensor in the input device is an acceleration sensor, sensor data sent by the input device to the VR host is acceleration data, where the acceleration data includes an acceleration direction and an acceleration magnitude; correspondingly, the VR host acquires acceleration data sent by the input device.
In step 310, the vibration frequency of the input device is calculated from the sensor data.
Wherein, the calculation formula of the vibration frequency is as follows: vibration frequency is the vibration times divided by the vibration time length.
In a possible implementation manner, when the input device vibrates, the angular speed direction of the input device changes, so that when a motion sensor in the input device is an angular speed sensor, the VR host determines the vibration frequency of the input device according to the change frequency of the angular speed direction after acquiring the angular speed direction and the angular speed sent by the angular speed sensor, and further calculates the vibration frequency according to the vibration frequency and the vibration duration;
similarly, when the motion sensor in the input device is an acceleration sensor, the VR host determines the vibration frequency of the input device according to the acceleration direction change condition indicated by the acceleration data, and further calculates the vibration frequency of the input device.
In step 311, it is determined whether the vibration frequency exceeds a frequency threshold.
Because when the user seeks the cursor, can subconsciously rock input device, consequently set up the vibration frequency threshold value in advance in the VR host computer, after the VR host computer obtains input device's vibration frequency through the calculation, judge whether vibration frequency exceeds the frequency threshold value, wherein, the frequency threshold value can be 4 times/second.
Further, when the environment screen does not include the input device cursor and the vibration frequency of the input device is greater than the frequency threshold, the VR host continues to perform the steps 304 to 305 of determining the relative position of the input device cursor and the environment screen.
In this embodiment, when the VR host detects that the environment screen does not include the input device cursor and the motion state of the input device matches the motion state of the user when searching for the cursor, the indication information is displayed, and the trigger condition for displaying the indication information matches the usage habit of the user, thereby improving the efficiency of the user moving the input device cursor into the environment screen.
During continuous operation of the input device, the input device cursor may angularly deviate from the actual position of the input device. Unlike the case where the input device cursor is outside the environment screen due to normal operation, when the input device cursor is outside the environment screen due to the angle deviation, the user cannot determine the specific position of the input device cursor, so that the display device needs to prompt the position of the input device cursor, as shown in fig. 7, before the step 304, steps 312 and 313 are further included:
in step 312, the duration of continuous use of the input device is obtained.
After the input device works continuously for a long time, the environment picture currently displayed by the display device does not contain the input device cursor due to the existence of the angle deviation.
And when the environment picture currently displayed by the display device does not contain the input device cursor, the VR host further acquires the continuous use time of the input device.
In step 313, it is determined whether the continuous use time period exceeds a time period threshold.
In a possible implementation manner, a duration threshold is stored in the VR host, and after the duration of continuous use of the input device is obtained, the VR host determines whether the duration of continuous use exceeds the duration threshold. Wherein the duration threshold may be set to 5 min.
Further, when the environment frame does not include the input device cursor and the continuous use duration of the input device is greater than the duration threshold, the VR host continues to perform the step of determining the relative position of the input device cursor and the environment frame in steps 304 to 305.
In this embodiment, when the VR host acquires that the duration of continuous use of the input device exceeds the duration threshold and the environment screen does not contain the input device cursor, the indication information is displayed, so that the problem that the user cannot determine the specific position of the input device cursor due to the angular deviation between the input device cursor and the actual position of the input device is solved.
The following are embodiments of the disclosed apparatus that may be used to perform embodiments of the disclosed methods. For details not disclosed in the embodiments of the apparatus of the present disclosure, refer to the embodiments of the method of the present disclosure.
Referring to fig. 8, there is shown a block diagram of an information display device provided in an embodiment of the present disclosure. The apparatus may be implemented as all or a portion of the terminal in software, hardware, or a combination of both. The device includes: detection module 810, determination module 820, and display module 830.
The detection module 810 is configured to detect whether an environment picture currently displayed by the display device includes an input device cursor corresponding to the input device, where the environment picture currently displayed by the display device is a part of a three-dimensional virtual environment generated by the VR host;
a determining module 820 configured to determine a relative position of the input device cursor and the environment screen when the input device cursor is not included in the environment screen;
and a display module 830 configured to control the display device to display the indication information in the environment screen according to the relative position.
Optionally, the detecting module 810 includes:
the first obtaining sub-module is configured to obtain a coordinate range of the environment picture in the three-dimensional virtual environment;
the second acquisition submodule is configured to acquire a first coordinate of a cursor of the input device in the three-dimensional virtual environment;
and the first determining submodule is configured to determine that the environment picture does not contain the input device cursor if no intersection exists between the first coordinate and the coordinate range.
Optionally, the determining module 820 includes:
a third obtaining submodule configured to obtain a first coordinate of a cursor of the input device in the three-dimensional virtual environment;
a fourth obtaining sub-module configured to obtain a second coordinate of the midpoint of the environment picture in the three-dimensional virtual environment;
and the second determining submodule is configured to determine the relative position of the input device cursor and the environment picture according to the first coordinate and the second coordinate.
Optionally, the apparatus further includes:
the device comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is configured to acquire sensor data sent by the input device, the sensor data is acquired by a motion sensor in the input device, and the motion sensor is an angular velocity sensor and/or an acceleration sensor;
a calculation module configured to calculate a vibration frequency of the input device from the sensor data;
the determining module is further configured to determine a relative position of the input device cursor and the environment picture when the environment picture does not contain the input device cursor and the vibration frequency of the input device is greater than the frequency threshold.
Optionally, the apparatus further includes:
the second acquisition module is configured to acquire the continuous use time length of the input device;
the determining module is further configured to determine a relative position of the input device cursor and the environment picture when the environment picture does not contain the input device cursor and the continuous use duration of the input device is greater than the duration threshold.
Optionally, the display module 830 includes:
the generation submodule is configured to generate a directional arrow according to the relative position, and the direction of the directional arrow is that the center of the environment picture points to the cursor of the input device;
a control sub-module configured to control the display device to display a directional arrow at a predetermined position within the environment screen.
Optionally, the apparatus further includes:
and the control module is configured to control the display device to stop displaying the indication information when the display duration of the indication information is greater than a threshold value or the input device cursor is detected to move to the environment screen.
In summary, in the embodiment of the present disclosure, when the VR host detects that the cursor of the input device is outside the environment picture currently displayed by the display device, the VR host determines the relative position of the cursor of the input device and the environment picture in the three-dimensional virtual environment, and controls the display device to display corresponding indication information in the environment picture according to the relative position, so that the user can quickly move the cursor of the input device to the environment picture according to the indication information; compared with the prior art that the user can only adjust the position of the input device blindly, in the embodiment of the disclosure, the user can purposefully adjust the orientation of the input device by automatically displaying the indication information for indicating the relative position of the cursor of the input device, so that the efficiency of moving the cursor of the input device to the current environment picture is improved.
In this embodiment, when the input device cursor is not within the display range, the VR host determines the relative position of the input device cursor and the environment picture according to the coordinates of the input device cursor and the environment picture in the first spatial rectangular coordinate system; and then, the VR host controls the display device to display a directional arrow pointing to the cursor of the input device from the center of the environment picture at a preset position of the environment picture according to the relative position. The user can purposefully change the position of the input device according to the directional arrow, so that the cursor of the input device is displayed in the environment picture. The efficiency of displaying the input device cursor within the environmental picture is improved.
In this embodiment, when the display duration of the indication information reaches the threshold, or when it is detected that the cursor of the input device moves to the environment screen, the VR host automatically stops displaying the indication information, so that the influence of the long-time display of the indication information on the immersive experience of the user is avoided.
In this embodiment, when the VR host detects that the environment screen does not include the input device cursor and the motion state of the input device matches the motion state of the user when searching for the cursor, the indication information is displayed, and the trigger condition for displaying the indication information matches the usage habit of the user, thereby improving the efficiency of the user moving the input device cursor into the environment screen.
In this embodiment, when the VR host acquires that the duration of continuous use of the input device exceeds the duration threshold and the environment screen does not contain the input device cursor, the indication information is displayed, so that the problem that the user cannot determine the specific position of the input device cursor due to the angular deviation between the input device cursor and the actual position of the input device is solved.
Fig. 9 is a block diagram illustrating a VR host 900 according to an example embodiment. Referring to fig. 9, VR host 900 may include one or more of the following components: a processing component 902, a memory 904, a power component 906, and a communication component 916.
The processing component 902 generally controls the overall operation of the VR host 900. Processing component 902 may include one or more processors 920 to execute instructions to perform all or a portion of the steps of the methods described above. Further, processing component 902 can include one or more modules that facilitate interaction between processing component 902 and other components.
The memory 904 is configured to store various types of data to support operations at the VR host 900. Examples of such data include instructions for any application or method operating on VR host 900. The memory 904 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
The power supply component 906 provides power to the various components of the VR host 900. The power components 906 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the VR host 900.
The communication component 916 is configured to facilitate wired or wireless communication between the VR host 900 and the access devices and display devices. The VR host 900 may access a wireless network based on a communication standard, such as WiFi, 2G, or 3G, or a combination thereof. In an exemplary embodiment, the communication component 916 receives a broadcast signal or broadcast associated information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 916 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, a non-transitory computer-readable storage medium including instructions, such as the memory 904 including instructions, executable by the processor 920 of the VR host 900 to perform the information display method described above is also provided. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (16)

1. An information display method is applied to a VR host of a Virtual Reality (VR) system, wherein the VR host is respectively connected with an input device and a display device, and the method comprises the following steps:
detecting whether an environment picture currently displayed by the display device contains an input device cursor corresponding to the input device, wherein the environment picture currently displayed by the display device is a part of a three-dimensional virtual environment generated by the VR host;
when the input device cursor is not contained in the environment picture, determining the relative position of the input device cursor and the environment picture;
and controlling the display equipment to display indication information in the environment picture according to the relative position, wherein the indication information is used for indicating the relative direction of the input equipment cursor and the environment picture.
2. The method according to claim 1, wherein the detecting whether an input device cursor corresponding to the input device is included in the environment picture currently displayed by the display device comprises:
acquiring a coordinate range of the environment picture in the three-dimensional virtual environment;
acquiring a first coordinate of the input device cursor in the three-dimensional virtual environment;
and if no intersection exists between the first coordinate and the coordinate range, determining that the environment picture does not contain the input equipment cursor.
3. The method of claim 1, wherein the determining the relative position of the input device cursor to the environmental screen comprises:
acquiring a first coordinate of the input device cursor in the three-dimensional virtual environment;
acquiring a second coordinate of the midpoint of the environment picture in the three-dimensional virtual environment;
and determining the relative position of the input equipment cursor and the environment picture according to the first coordinate and the second coordinate.
4. The method of any of claims 1-3, wherein prior to determining the relative position of the input device cursor to the environmental screen, the method further comprises:
acquiring sensor data sent by the input equipment, wherein the sensor data is acquired by a motion sensor in the input equipment, and the motion sensor is an angular velocity sensor and/or an acceleration sensor;
calculating a vibration frequency of the input device from the sensor data;
the determining the relative position of the input device cursor and the environment picture comprises:
and when the environment picture does not contain the input equipment cursor and the vibration frequency of the input equipment is greater than a frequency threshold value, determining the relative position of the input equipment cursor and the environment picture.
5. The method of any of claims 1-3, wherein prior to determining the relative position of the input device cursor to the environmental screen, the method further comprises:
acquiring the continuous use duration of the input equipment;
the determining the relative position of the input device cursor and the environment picture comprises:
and when the environment picture does not contain the input equipment cursor and the continuous use time of the input equipment is longer than a time threshold, determining the relative position of the input equipment cursor and the environment picture.
6. The method according to any one of claims 1 to 3, wherein the controlling the display device to display indication information in the environment picture according to the relative position comprises:
generating a directional arrow according to the relative position, wherein the direction of the directional arrow is from the center of the environment picture to the input device cursor;
and controlling the display device to display the directional arrow at a preset position in the environment picture.
7. The method of any of claims 1 to 3, further comprising:
and when the display duration of the indication information is greater than a threshold value, or when the input equipment cursor is detected to move to the environment picture, controlling the display equipment to stop displaying the indication information.
8. The utility model provides an information display device which characterized in that, is applied to in the VR host computer of virtual reality VR system, the VR host computer links to each other with input device and display device respectively, the device includes:
a detection module configured to detect whether an environment picture currently displayed by the display device includes an input device cursor corresponding to the input device, where the environment picture currently displayed by the display device is a part of a three-dimensional virtual environment generated by the VR host;
a determination module configured to determine a relative position of the input device cursor to the environment screen when the input device cursor is not included in the environment screen;
and the display module is configured to control the display device to display indication information in the environment picture according to the relative position, wherein the indication information is used for indicating the relative direction of the input device cursor and the environment picture.
9. The apparatus of claim 8, wherein the detection module comprises:
a first obtaining sub-module configured to obtain a coordinate range of the environment picture in the three-dimensional virtual environment;
a second obtaining submodule configured to obtain a first coordinate of the input device cursor in the three-dimensional virtual environment;
a first determining sub-module configured to determine that the input device cursor is not included in the environment picture if there is no intersection between the first coordinate and the coordinate range.
10. The apparatus of claim 8, wherein the determining module comprises:
a third obtaining submodule configured to obtain a first coordinate of the input device cursor in the three-dimensional virtual environment;
a fourth obtaining sub-module configured to obtain a second coordinate of the midpoint of the environment picture in the three-dimensional virtual environment;
a second determining submodule configured to determine a relative position of the input device cursor and the environment screen according to the first coordinate and the second coordinate.
11. The apparatus of any one of claims 8 to 10, further comprising:
a first acquisition module configured to acquire sensor data sent by the input device, wherein the sensor data is acquired by a motion sensor in the input device, and the motion sensor is an angular velocity sensor and/or an acceleration sensor;
a calculation module configured to calculate a vibration frequency of the input device from the sensor data;
the determining module is further configured to determine a relative position of the input device cursor and the environment screen when the input device cursor is not included in the environment screen and the vibration frequency of the input device is greater than a frequency threshold.
12. The apparatus of any one of claims 8 to 10, further comprising:
a second obtaining module configured to obtain a continuous use duration of the input device;
the determining module is further configured to determine a relative position of the input device cursor and the environment picture when the environment picture does not include the input device cursor and the continuous use duration of the input device is greater than a duration threshold.
13. The apparatus of any one of claims 8 to 10, wherein the display module comprises:
a generation submodule configured to generate a directional arrow in a direction from the center of the environment screen to the input device cursor according to the relative position;
a control sub-module configured to control the display device to display the directional arrow at a predetermined position within the environment screen.
14. The apparatus of any one of claims 8 to 10, further comprising:
the control module is configured to control the display device to stop displaying the indication information when the display duration of the indication information is greater than a threshold value or the input device cursor is detected to move to the environment screen.
15. An information display apparatus, characterized in that the apparatus comprises:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to:
detecting whether an environment picture currently displayed by display equipment contains an input equipment cursor corresponding to the input equipment or not, wherein the environment picture currently displayed by the display equipment is a part of a three-dimensional virtual environment generated by a Virtual Reality (VR) host;
when the input device cursor is not contained in the environment picture, determining the relative position of the input device cursor and the environment picture;
and controlling the display equipment to display indication information in the environment picture according to the relative position, wherein the indication information is used for indicating the relative direction of the input equipment cursor and the environment picture.
16. A computer-readable medium, having stored thereon program instructions which, when executed by a processor, implement the information display method of any one of claims 1 to 7.
CN201711326613.4A 2017-12-13 2017-12-13 Information display method and device Active CN107957781B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711326613.4A CN107957781B (en) 2017-12-13 2017-12-13 Information display method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711326613.4A CN107957781B (en) 2017-12-13 2017-12-13 Information display method and device

Publications (2)

Publication Number Publication Date
CN107957781A CN107957781A (en) 2018-04-24
CN107957781B true CN107957781B (en) 2021-02-09

Family

ID=61957808

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711326613.4A Active CN107957781B (en) 2017-12-13 2017-12-13 Information display method and device

Country Status (1)

Country Link
CN (1) CN107957781B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108619721B (en) * 2018-04-27 2020-08-11 腾讯科技(深圳)有限公司 Distance information display method and device in virtual scene and computer equipment
CN114138120B (en) * 2021-12-09 2023-02-14 北京冰镇山竹科技有限公司 Virtual visual angle tracking method and system applied to virtual reality

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9335912B2 (en) * 2007-09-07 2016-05-10 Apple Inc. GUI applications for use with 3D remote controller
JP5531616B2 (en) * 2007-12-07 2014-06-25 ソニー株式会社 Control device, input device, control system, control method, and handheld device
US8806381B2 (en) * 2009-11-18 2014-08-12 International Business Machines Corporation Method and system to improve GUI use efficiency
US9665250B2 (en) * 2011-02-07 2017-05-30 Blackberry Limited Portable electronic device and method of controlling same
KR20150025114A (en) * 2013-08-28 2015-03-10 엘지전자 주식회사 Apparatus and Method for Portable Device displaying Augmented Reality image
US9389706B2 (en) * 2014-11-19 2016-07-12 Screenovate Technologies Ltd. Method and system for mouse control over multiple screens
US20160239110A1 (en) * 2015-02-13 2016-08-18 I/O Interconnect Inc. Method for automatically transferring control right of mouse between personal computer and handheld computer
CN105159522B (en) * 2015-07-22 2018-03-13 深圳多新哆技术有限责任公司 A kind of method of virtual reality display device response peripheral apparatus operation
CN107085489A (en) * 2017-03-21 2017-08-22 联想(北京)有限公司 A kind of control method and electronic equipment

Also Published As

Publication number Publication date
CN107957781A (en) 2018-04-24

Similar Documents

Publication Publication Date Title
EP2996017B1 (en) Method, apparatus and computer program for displaying an image of a physical keyboard on a head mountable display
US8933882B2 (en) User centric interface for interaction with visual display that recognizes user intentions
US10540022B2 (en) Interactive input controls in a simulated three-dimensional (3D) environment
US20180253152A1 (en) Gesture-controlled augmented reality experience using a mobile communications device
CN108038726B (en) Article display method and device
KR20220008281A (en) Systems and methods for generating dynamic obstacle collision warnings for head mounted displays
US10488918B2 (en) Analysis of user interface interactions within a virtual reality environment
RU2751130C1 (en) Method for coordinate alignment of coordinate systems used by computer-generated reality apparatus and tactile sensation transmission apparatus
US20160334884A1 (en) Remote Sensitivity Adjustment in an Interactive Display System
JP6399692B2 (en) Head mounted display, image display method and program
US10296096B2 (en) Operation recognition device and operation recognition method
US10978019B2 (en) Head mounted display system switchable between a first-person perspective mode and a third-person perspective mode, related method and related non-transitory computer readable storage medium
US20190227693A1 (en) Virtual space positioning method and apparatus
KR20150040580A (en) virtual multi-touch interaction apparatus and method
CN107957781B (en) Information display method and device
CN110688002B (en) Virtual content adjusting method, device, terminal equipment and storage medium
US20130316816A1 (en) Display control system, display control method, display control device, and computer-readable storage medium
US10073609B2 (en) Information-processing device, storage medium, information-processing method and information-processing system for controlling movement of a display area
US20240054975A1 (en) Head mounted display and setting method
WO2021004413A1 (en) Handheld input device and blanking control method and apparatus for indication icon of handheld input device
US20200327867A1 (en) Head mounted display system capable of displaying a virtual scene and a map of a real environment in a picture-in-picture mode, related method and related non-transitory computer readable storage medium
JP7287172B2 (en) Display control device, display control method, and program
US11604517B2 (en) Information processing device, information processing method for a gesture control user interface
KR20230124363A (en) Electronic apparatus and method for controlling thereof
CN113010057B (en) Pose control method, electronic device and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant