WO2020140905A1 - Virtual content interaction system and method - Google Patents

Virtual content interaction system and method Download PDF

Info

Publication number
WO2020140905A1
WO2020140905A1 PCT/CN2019/130646 CN2019130646W WO2020140905A1 WO 2020140905 A1 WO2020140905 A1 WO 2020140905A1 CN 2019130646 W CN2019130646 W CN 2019130646W WO 2020140905 A1 WO2020140905 A1 WO 2020140905A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
terminal device
virtual content
head
screen
Prior art date
Application number
PCT/CN2019/130646
Other languages
French (fr)
Chinese (zh)
Inventor
胡永涛
戴景文
贺杰
Original Assignee
广东虚拟现实科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN201910005562.8A external-priority patent/CN111399630B/en
Priority claimed from CN201910005848.6A external-priority patent/CN111399631B/en
Priority claimed from CN201910295517.0A external-priority patent/CN111818326B/en
Application filed by 广东虚拟现实科技有限公司 filed Critical 广东虚拟现实科技有限公司
Publication of WO2020140905A1 publication Critical patent/WO2020140905A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer

Definitions

  • This application relates to the field of augmented reality, in particular to a virtual content interaction system and method.
  • Augmented reality technology uses computer graphics technology and visualization technology to construct virtual content that does not exist in the real environment, and accurately integrates the virtual content into the real environment through image recognition and positioning technology, and integrates the virtual content and the real environment with the help of display devices. And show the user a real sensory experience.
  • the primary technical problem to be solved by augmented reality technology is how to accurately integrate the virtual content into the real world, that is, to make the virtual content appear in the correct position of the real scene at the correct angle posture, so as to generate a strong visual reality. Therefore, the interaction with virtual content is an important research direction of augmented reality technology.
  • the embodiments of the present application propose a virtual content interaction system and method.
  • an embodiment of the present application provides a virtual content interaction system, including a terminal device and an external head-mounted display device, the terminal device establishing a communication connection with the head-mounted display device, wherein the terminal device For acquiring first relative spatial position information of the head-mounted display device and the terminal device, determining a display position of virtual content according to the first relative spatial position information, and rendering the virtual content according to the display position , And obtain the display data of the virtual content, and transmit the display data to the head-mounted display device; the head-mounted display device is used to receive the display data sent by the terminal device, and according to the The display data displays the virtual content.
  • an embodiment of the present application provides a virtual content interaction method, which is applied to a terminal device, and the terminal device is in communication connection with an external head-mounted display device.
  • the method includes: acquiring the head-mounted display device and The first relative spatial position of the terminal device; determining the display position of the virtual content according to the first relative spatial position; rendering the virtual content according to the display position, and obtaining display data of the virtual content; and
  • the display data is transmitted to the head-mounted display device, and the display data is used to instruct the head-mounted display device to display the virtual content.
  • an embodiment of the present application provides a virtual content interaction method, which is applied to a head-mounted display device, and the head-mounted display device establishes a communication connection with a terminal device, including: collecting an image containing a marker through a camera.
  • the marker is provided on the terminal device; the marker image is transmitted to the terminal device, and the image is used to instruct the terminal device to recognize the marker contained in the image and obtain the A first relative spatial position of the head-mounted display device and the terminal device; and receiving display data sent by the terminal device, and displaying virtual content according to the display data, the display data is determined by the terminal device according to the first
  • the virtual content is obtained by rendering the virtual content at a relative spatial position.
  • an embodiment of the present application provides a virtual content display method, which is applied to a first terminal, and the first terminal is in communication connection with a second terminal.
  • the method includes: acquiring the first terminal and the first terminal First relative spatial position information between the two terminals; obtaining display content data from the second terminal, wherein the display content data includes at least the data of the display content currently displayed by the second terminal; and according to the first A relative spatial position information and the display content data to display virtual content, wherein the virtual content includes the display content displayed by the second terminal and the extended content corresponding to the display content.
  • an embodiment of the present application provides an image processing method, which is applied to a terminal device, and the terminal device is in communication connection with a head-mounted display device.
  • the method includes: acquiring the terminal device and the head-mounted display device Relative spatial position relationship between them; according to the relative spatial position relationship, acquiring a projection area of the virtual content displayed by the head-mounted display device on the screen of the terminal device, the projection area being displayed by the head-mounted display The projection area of the virtual content on the screen observed by the device; obtaining the image content corresponding to the projection area in the screen content according to the screen content to be displayed on the screen; and performing on the image content Specify the process and display the screen content containing the image content after the specified process, and the difference in tone between the first tone of the specified image content and the second tone of the virtual content is greater than the first threshold.
  • an embodiment of the present application provides a terminal device, including: one or more processors; a memory; wherein the memory stores one or more application programs, and the one or more application programs are configured to be configured by the one Or executed by multiple processors, and the one or more programs are configured to execute the method provided above.
  • an embodiment of the present application provides a computer-readable storage medium.
  • the computer-readable storage medium stores a program code, and the program code can be called by a processor to execute the method provided above.
  • FIG. 1 is a schematic diagram of an application scenario according to an embodiment of the present application.
  • FIG. 2 is a schematic diagram of an augmented reality system according to an embodiment of this application.
  • FIG. 3 is a schematic diagram of another application scenario according to an embodiment of the present application.
  • FIG. 4 is a flowchart of a virtual content interaction method in an embodiment.
  • FIG. 5 is a schematic diagram showing virtual content in an embodiment.
  • FIG. 6 is a schematic diagram showing extended content on a screen of a terminal device in an embodiment.
  • FIG. 7 is a schematic diagram showing extended content on a screen of a terminal device in an embodiment.
  • FIG. 8 is a flowchart of a virtual content interaction method in an embodiment.
  • FIG. 9 is a flowchart of determining the display position of virtual content in one embodiment.
  • FIG. 10 is a schematic diagram showing virtual content in one embodiment.
  • FIG. 11 is a schematic diagram showing that the virtual content displayed in one embodiment overlaps with the screen content of the terminal device.
  • FIG. 12 is a schematic diagram showing virtual content in one embodiment.
  • 13A-13B are schematic diagrams of updating the displayed virtual content according to the change of the screen content of the terminal device in one embodiment.
  • 14A-14C are schematic diagrams of the sliding of the control area of the terminal device in an embodiment.
  • 15 is a schematic diagram of controlling virtual content according to a finger sliding trajectory in one embodiment.
  • FIG. 16 is a schematic diagram of controlling virtual content according to a finger sliding trajectory in another embodiment.
  • 17A-17B are schematic diagrams of controlling virtual content according to a finger sliding trajectory in yet another embodiment.
  • FIG. 18 is a schematic diagram of controlling virtual content according to a sliding track of a finger in still another embodiment.
  • FIG. 19 is a flowchart of a virtual content interaction method in yet another embodiment.
  • Figure 20 is a block diagram of an augmented reality system in one embodiment.
  • 21 is a flowchart of an image processing method in an embodiment.
  • 22 is a schematic diagram of a projection area of virtual content on a screen of a terminal device in an embodiment.
  • 23 is a schematic diagram showing virtual content and processed screen content in an embodiment.
  • 24 is a flowchart of an image processing method in another embodiment.
  • FIG. 25 is a flowchart of acquiring a projection area of virtual content on a screen of a terminal device in an embodiment.
  • 26 is a schematic diagram of a projection area of virtual content on a screen of a terminal device in another embodiment.
  • FIG. 27 is a flowchart of acquiring the first projection area of the left-eye image on the screen in one embodiment.
  • FIG. 28 is a schematic diagram showing virtual content and processed screen content in another embodiment.
  • FIG. 29 is a flowchart of an image processing method in still another embodiment.
  • FIG. 30 is a block diagram of a head-mounted display device in one embodiment.
  • the augmented reality system 10 provided by the embodiment of the present application includes: a terminal device 100 and a head-mounted display device 200 connected to the terminal device 100.
  • the terminal device 100 may be held and controlled by a user, and may be an electronic device capable of running application programs, such as a mobile phone, a smart watch, a tablet computer, an e-reader, and a notebook computer.
  • application programs such as a mobile phone, a smart watch, a tablet computer, an e-reader, and a notebook computer.
  • the head-mounted display device 200 may be an external head-mounted display device, that is, the head-mounted display device 200 includes a display module, a communication module, a camera, etc. for display.
  • the processor, memory, etc. of the terminal device 100 control the displayed virtual content.
  • the display module may include a display screen (or a projection device) and a display lens to display virtual content.
  • the head-mounted display device 200 can transmit information and instructions with the terminal device 100.
  • the transmitted information may include virtual content displayed by the head-mounted display device 200.
  • the terminal device 100 and the head-mounted display device 200 can be connected via wireless communication methods such as Bluetooth, WiFi, ZigBee, etc., and can also be connected by wired communication through an interface such as a USB, as shown in FIG. Wired communication connections such as terminals.
  • the marker 101 is provided on the terminal device 100.
  • the marker 101 can be any graphic or object with identifiable features, which can be collected and tracked by the head-mounted display device 200.
  • the marker 101 can be a pattern with a topology structure, which refers to the marker The connected relationship between the sub-markers and the feature points.
  • the camera can collect an image containing the marker 101 in real time and transmit the image to the terminal device 100, and the terminal device 100 can obtain the marker according to the image The relative position, orientation, and other spatial position information between the object 101 and the head-mounted display device 200 to obtain relative spatial position information between the terminal device 100 and the head-mounted display device 200.
  • the terminal device 100 may render a corresponding virtual object based on the spatial position information between the head-mounted display device 200 and transmit the display screen of the virtual object to the head-mounted display device 200, and the head-mounted display device 200 may display the virtual object through the display module Object.
  • the head-mounted display device 200 can also track the shape of the terminal device 100 to determine the relative spatial position relationship between the terminal device 100 and the head-mounted display device 200.
  • the terminal device 100 is wirelessly connected to the head-mounted display device 200, and the user can scan the marker 101 on the terminal device 100 through the head-mounted display device 200 worn to watch a virtual space scene containing multiple virtual planets Superimposed on the real space, where the virtual space scene corresponds to the space scene displayed by the terminal device 100.
  • the terminal device 100 and the head-mounted display device 200 are connected by wire communication, and the user scans the marker 101 on the terminal device 100 through the head-mounted display device 200 worn, and can view the superimposed display of the medical human model including the virtual reality in reality On the surface of the terminal device 100 in the space.
  • an embodiment of the present application provides a virtual content interaction method, which is applied to a terminal device.
  • the terminal device is in communication connection with an external head-mounted display device.
  • the method includes the following steps.
  • Step 410 Obtain the first relative spatial position of the head-mounted display device and the terminal device.
  • the first relative spatial position may include relative position and posture information between the head-mounted display device and the terminal device, and the posture information may be the orientation and rotation angle of the terminal device relative to the head-mounted display device.
  • the terminal device includes an inertial measurement unit (IMU), which may first acquire the measurement data of the IMU, and determine the first relative spatial position between the terminal device and the head-mounted display device based on the measurement data.
  • IMU inertial measurement unit
  • a light spot may also be provided on the terminal device, and the head-mounted display device collects an image containing the light spot through the camera and sends the image to the terminal device.
  • the terminal device may identify the light spot in the image to determine the first relative spatial position of the head-mounted display device and the terminal device.
  • the light spot set on the terminal device may be a visible light spot or an infrared light spot.
  • an infrared camera may be provided on the head-mounted display device to collect an image of the infrared light spot.
  • the light spot set on the terminal device may be one, or may be a light spot sequence composed of multiple light spots.
  • the light spot may be arranged on the casing of the terminal device, for example, may be arranged around the screen.
  • the light point can also be set on the protective cover of the terminal device.
  • the protective cover containing the light point can be put on, so that the terminal device can be tracked.
  • different light spots can be set around the screen of the terminal device, such as different numbers of light spots or different color spots around the terminal device, so that the terminal device can determine The relative spatial position of the head-mounted display device.
  • Step 420 Determine the display position of the virtual content according to the first relative spatial position.
  • the terminal device can convert the first relative spatial position from the real space to the position coordinates in the virtual space, and according to the positional relationship between the virtual content to be displayed and the terminal device in the virtual space and the position coordinates, refer to the head-mounted display device as a reference Calculate the spatial position of the virtual content to be displayed in the virtual space relative to the head-mounted display device, that is, obtain the display position of the virtual content in the virtual space.
  • the display position refers to the three-dimensional space coordinates of the virtual content in the virtual space with the head-mounted display device as the origin (which can also be regarded as the origin of the human eye).
  • Step 430 Render the virtual content according to the display position and obtain the display data of the virtual content.
  • the terminal device may acquire the data of the virtual content to be displayed, construct the virtual content according to the data of the virtual content, and render the virtual content according to the display position.
  • the data of the virtual content to be displayed may include model data of the virtual content, which is used to construct a three-dimensional model of the virtual content.
  • the model data may include data such as the color, vertex coordinates, and outline of the model.
  • the data of the virtual content may also be downloaded from the server by the terminal device, or may be obtained from other terminals for the terminal device.
  • Step 440 Transmit the display data to the head-mounted display device to instruct the head-mounted display device to display virtual content.
  • the terminal device may obtain display data of the rendered virtual content, and the display data may include RGB values of respective pixels in the display screen and corresponding pixel coordinates, and the like.
  • the terminal device can transmit the display data to the head-mounted display device, and the head-mounted display device can generate a display screen according to the display data, and project the display screen onto the optical lens to display the virtual content.
  • the user By wearing the optical lens of the display device, the user can see that the virtual content is superimposed and displayed in the real world, realizing the effect of augmented reality.
  • the terminal device obtains the data of the virtual content for rendering.
  • the data of the virtual content may be the content data displayed on the current screen of the terminal device, or it may be the extended content data corresponding to the content displayed on the current screen, or it may be the data on the current screen.
  • the complete content may include the content displayed on the current screen and the extended content corresponding to the displayed content.
  • the content displayed on the current screen may be a part of the complete content, and the extended content is a part of the complete content except the displayed content.
  • the content currently displayed by the terminal device 100 is a part of the China map
  • the terminal device 100 transmits the data of the part of the China map
  • the virtual content 300 a displayed by the head-mounted display device 200 is the part of the China map.
  • the terminal device 100 transmits data of a complete China map
  • the virtual content 300b displayed by the head-mounted display device 200 is a complete China map.
  • the content currently displayed by the terminal device 100 is a game map
  • the terminal device 100 transmits the data of the extended content (game character) corresponding to the game map to the head-mounted display device, and the game character 300c that the user can see through the head-mounted display device Superimposed on the real space to solve the problem that the display content is limited by the screen of the mobile terminal.
  • buttons and other interface elements can be displayed on the screen of the terminal device, and the user can enter different modes by selecting and clicking different buttons, for example, entering or exiting the augmented reality mode; in the augmented reality mode, Select different display modes, for example, display the same content as the screen of the terminal device, display extended content corresponding to the screen of the terminal device, or display preset content irrelevant to the screen content of the terminal device, etc.
  • the terminal device may transmit the display data corresponding to the selected display mode to the head-mounted display device, so that the head-mounted display device displays virtual content according to the display data.
  • the terminal device may only serve as the processor and memory of the head-mounted display device to determine the virtual content displayed by the head-mounted display device, and the virtual content is not related to the content currently displayed on the screen of the terminal device.
  • the data of the virtual content may be generated in real time according to the actual environment, or may be stored in advance.
  • the terminal device may not display any content.
  • a virtual content interaction method provided by an embodiment of the present application is applied to a terminal device, and the terminal device is in communication connection with an external head-mounted display device, including the following steps.
  • Step 810 Acquire a first relative spatial position of the head-mounted display device and the terminal device.
  • the terminal device is provided with markers.
  • the markers may be provided on the casing of the terminal device, or may be displayed on the screen of the terminal device in the form of an image.
  • the marker may also be an external marker.
  • the terminal device can be inserted through the USB or the earphone hole, etc., so as to track the terminal device.
  • Step 820 Determine the display position of the virtual content according to the first relative spatial position.
  • determining the display position of the virtual content according to the first relative spatial position includes steps 822 and 824.
  • Step 822 Acquire the second relative spatial position of the virtual content and the terminal device.
  • the second relative spatial position may include the position and posture information of the virtual content in the virtual space relative to the terminal device, and may also be understood as the relative spatial position relationship between the superimposed position of the virtual content in the real space and the terminal device.
  • the superimposed position of the virtual content in the real space can be superimposed on the terminal device, such as on the screen area of the terminal device, as shown in FIG. 12, the user can see that the virtual medical mannequin is superimposed on the screen area of the terminal device through the head-mounted display device ;
  • Virtual content can also be located on the edge or around the terminal device, as shown in Figure 6, China map 300b is superimposed on the upper right of the terminal device 100;
  • the virtual content can also be in a different plane from the terminal device, such as the plane where the terminal device is located and the virtual content The plane is vertical, and the superimposition position of the virtual content is not limited.
  • the second relative spatial position may be stored in the terminal device in advance, or may be set according to the content displayed on the current screen of the terminal device.
  • the content displayed on the current screen of the terminal device is different, and the second relative spatial position may be different.
  • the current screen of the terminal device 100 displays a part of the China map, and the virtual content 300b including the complete China map is in the superimposed position of the terminal device in the real space.
  • the second relative spatial position may also be set according to the inclusion relationship between the content displayed on the current screen of the terminal device and the virtual content.
  • the second relative spatial position of the virtual content and the terminal device may be determined according to the position of the content displayed on the screen in the virtual content, so that the displayed content included in the virtual content It overlaps with the content displayed on the terminal device screen.
  • the content currently displayed by the terminal device 100 is the surrounding map 110 of Hebei province, and the virtual content may be the map 300b of China.
  • the map 300b of China and the mobile terminal can be set
  • the second relative spatial position enables the user to see the surrounding map of Hebei Republic in the China map 300b through the head-mounted display device, which overlaps with the surrounding map 110 of Hebei Republic currently displayed on the terminal device.
  • Step 824 Determine the display position of the virtual content according to the first relative spatial position and the second relative spatial position.
  • the terminal device may determine the spatial coordinates of the terminal device in the virtual space according to the first relative spatial position, and based on the spatial coordinates and the second relative spatial position, use the head-mounted display device as a reference to determine the space of the virtual content relative to the head-mounted display device Position to get the display position of the virtual content.
  • the terminal device may determine whether the display position of the virtual content overlaps with the terminal device.
  • the display position of the virtual content may be a superimposition position of the virtual content in the real space, or a space position of the virtual content in the virtual space with the head-mounted display device as the origin.
  • the corresponding display area of the content displayed on the current screen of the terminal device in the virtual content can be determined, and the display area is the overlapping area between the display position of the virtual content and the terminal device.
  • the content of the display area is the same as the content displayed on the current screen of the terminal device.
  • the terminal device may perform a specified process on the display area, may adjust the content color of the display area to the specified color, and may also adjust the content transparency of the display area to the specified transparency.
  • the brightness value of each color component of the specified color is lower than the first threshold, and the specified transparency is lower than the second threshold.
  • the first threshold is the maximum brightness value of each color component when the virtual content cannot be normally superimposed and displayed.
  • the first threshold may be set to 13 brightness, which is 95% black, or may be set to 0 brightness, which is black.
  • the second threshold is the maximum transparency when the virtual content cannot be observed normally.
  • the second threshold may be set to 1, which is 90% transparent, or may be set to 0, which is 100% transparent.
  • the content displayed on the current screen of the terminal device is different, and the display position corresponding to the virtual content is different.
  • the corresponding relationship can be stored in the terminal device.
  • the display position of the virtual content can be updated according to the changed content.
  • the terminal device may receive the user's sliding operation on the screen, and change the content displayed on the screen according to the sliding operation.
  • the terminal device may re-determine the corresponding display area of the changed content displayed on the screen in the virtual content. Based on the newly determined display area, the updated display position of the virtual content is calculated, and the virtual content is re-rendered based on the updated display position, so that the user can observe that the position of the virtual content changes as the content displayed on the terminal device changes.
  • FIGS. 13A and 13B the user slides to the right on the screen of the terminal device 100, and the content displayed by the terminal device 100 changes from the surrounding map 110 of Hebei to the surrounding map 120 of Ningxia. The user can see the China map 300 through the head-mounted display device 200 Has moved, and the map 300 of China is accurately connected with the content displayed on the mobile terminal.
  • Step 830 Render the virtual content according to the display position and obtain the display data of the virtual content.
  • Step 840 Transmit the display data to the head-mounted display device, and the display data is used to instruct the head-mounted display device to display the virtual content.
  • Step 850 When receiving the control operation, generate a control instruction according to the control operation.
  • the terminal device may receive the user's control operation, generate a control instruction according to the control operation, and control the virtual content displayed by the head-mounted display device.
  • the terminal device includes a manipulation area, which can receive a user's control operation in the manipulation area.
  • the manipulation area may include at least one of a touch screen and keys, where the user's control operations include, but are not limited to, single-finger sliding, clicking, pressing, multi-finger coordinated sliding, etc., acting on the manipulation area of the terminal device.
  • the control instructions generated according to the control operation may include a movement instruction, a zoom-in instruction, a zoom-out instruction, a rotation instruction, a selection instruction, etc., so as to realize the display effect of controlling the movement, zooming, rotation, and selection of virtual content.
  • the terminal device may generate a control instruction according to one or more of the detected number of fingers when the control operation is performed, the gesture action when the control operation is performed, and the finger sliding trajectory when the control operation is performed.
  • the number of fingers when performing control operations in the manipulation area can be detected in real time, and different control instructions can be generated according to different numbers of fingers.
  • FIG. 14A when it is detected that the user performs a single-finger slide control operation in the manipulation area of the terminal device, a control instruction for moving virtual content is generated.
  • the control instruction is to control the head-mounted display device to move the currently displayed virtual map to the right relative to the user’s perspective.
  • FIG. 14A when it is detected that the user performs a single-finger slide control operation in the manipulation area of the terminal device, a control instruction for moving virtual content is generated.
  • the control instruction is to control the head-mounted display device to move the currently displayed virtual map to the right relative to the user’s perspective.
  • FIG. 14A when it is detected that the user performs a single
  • a control instruction for reducing the virtual content is generated.
  • the control instruction is to control the head-mounted display device to reduce the currently displayed virtual map relative to the user's viewing angle.
  • the finger sliding trajectory of the control operation may be detected in real time, and different control instructions may be generated according to different finger sliding trajectories.
  • FIG. 15 when the user performs a sliding operation on the touch screen area of the terminal device 100 with a finger to the right, a control instruction to rotate the virtual content 300 (virtual medical human body) to the right is generated.
  • control operation may also be a gesture control of the user.
  • the head-mounted display device can collect the user's gesture image through the camera and recognize the gesture in the gesture image.
  • the recognized gesture is a preset gesture
  • a control instruction is generated.
  • the preset gestures can be ascending, descending, waving left and right gestures.
  • the terminal device may also be connected to the controller, and may receive user control operations in the control area of the controller.
  • the user's control operations include, but are not limited to, single-finger slide, click, press, multi-finger cooperative slide acting on the control area of the controller.
  • the same control operation may correspond to different control instructions.
  • the terminal device may generate a control instruction corresponding to the control operation according to the virtual content and the received control operation.
  • Step 860 Control the displayed virtual content according to the control instruction.
  • the terminal device may adjust the display state of the virtual content according to the generated control instruction, and transmit the adjusted virtual content to the head-mounted display device, so that the head-mounted display device displays the adjusted virtual content.
  • the user controls the virtual content by operating the terminal device, and can also view the display effect of the virtual content while operating on the terminal device, thereby realizing the interaction between the terminal device and the head-mounted display device.
  • the content displayed on the screen of the terminal device may also be controlled according to the control instruction.
  • the content displayed by the terminal device is a map of Beijing, Shanxi province and other cities.
  • the virtual content displayed by the head-mounted display device is a map of China.
  • the control command is an enlarged display command, as shown in Figure 17B, the user displays through the head-mounted display The enlarged map of China seen by the device, and at the same time, the maps of Beijing, Shanxi province and other cities displayed on the terminal device are enlarged and displayed to the map of Beijing.
  • controlling virtual content may include selecting specified virtual content, moving virtual content, zooming in on virtual content, and so on. As shown in FIG. 18, when the area selected by the user on the virtual content 300 is a map of Beijing, the user can see that the map of Beijing is displayed on the terminal device 100 simultaneously.
  • the terminal device in addition to being a processor and a memory of the head-mounted display device, can also provide power to the head-mounted display device, keep the head-mounted display device light, and reduce the manufacturing cost of the head-mounted display device.
  • the virtual content interaction method provided in the above embodiments may also be independently implemented in a head-mounted display device, and the head-mounted display device may have a processor and a memory to generate and display virtual content.
  • the content currently displayed by the terminal device and the virtual content displayed by the head-mounted display device are accurately connected to improve the display effect; and the displayed virtual content can be controlled to improve the terminal device and the head-mounted device. Show interactivity between devices.
  • an embodiment of the present application provides a virtual content interaction method, which is applied to a head-mounted display device.
  • the head-mounted display device and a terminal device communicate and connect, including the following steps.
  • Step 1910 Acquire an image containing a marker, which is set on the terminal device.
  • Step 1920 Transmit the collected image to the terminal device.
  • the image is used to instruct the terminal device to identify the marker in the image, and obtain the first relative spatial position of the head-mounted display device and the terminal device.
  • Step 1930 Receive the display data sent by the terminal device, and display the virtual content according to the display data.
  • the display data is rendered by the terminal device according to the first relative spatial position.
  • an augmented reality system provided by an embodiment of the present application includes a terminal device 11 and a head-mounted display device 12 communicatively connected to the terminal device 11.
  • the terminal device 11 is used to obtain first relative spatial position information of the head-mounted display device 12 and the terminal device 11, determine the display position of the virtual content according to the first relative spatial position information, render the virtual content according to the display position, and obtain the virtual content
  • the display data is transmitted to the head-mounted display device 12, and the display data is used to instruct the head-mounted display device 12 to display virtual content.
  • the head-mounted display device 12 is used to receive the display data sent by the terminal device 11 and display the virtual content according to the display data.
  • an embodiment of the present application provides an image processing method, which can be applied to a terminal device.
  • the terminal device and a head-mounted display device are communicatively connected.
  • the method includes the following steps.
  • Step 2110 Obtain the relative spatial position relationship between the terminal device and the head-mounted display device.
  • Step 2120 Acquire the projection area of the virtual content displayed on the head-mounted display device on the screen of the terminal device according to the relative spatial position relationship, and the projection area is the projection area of the virtual content observed on the screen by the human eye through the head-mounted display device.
  • the projection area can be understood as, when the shape of the virtual content is projected on the plane where the screen of the terminal device is based on the human eye observing the virtual content as a reference point, between the projection area obtained on the plane and the screen area of the terminal device Coincident area.
  • the projection area may be represented by the coordinates of the screen coordinate system of the terminal device, and the screen coordinate system uses one corner of the screen, for example, the lower left corner, as the origin.
  • the projection area of the virtual content on the screen of the terminal device may be obtained according to the virtual content and the spatial position of the terminal device in the virtual space.
  • the virtual space may include a virtual camera, which is used to simulate the human eye of the user, and the position of the virtual camera in the virtual space may be regarded as the position of the head-mounted display device in the virtual space.
  • the spatial position of the virtual content and the terminal device in the virtual space may be the position of the virtual content and the terminal device in a preset coordinate system of the virtual space, and the preset coordinate system may include a world coordinate system, The spatial coordinate system with the virtual camera as the origin, etc.
  • the terminal device can obtain the extension line of the connection between the virtual camera and each vertex in the virtual content according to the virtual camera, the virtual content, and the spatial position of the terminal device in the preset coordinate system, and calculate each extension line and the terminal in the virtual space separately
  • the intersection point of the screen of the device determines the projection point of each vertex of the virtual content on the screen of the terminal device, and obtains the spatial coordinates of each projection point in the preset coordinate system.
  • the spatial coordinates of the projection point can be converted into the two-dimensional coordinates of the screen coordinate system of the terminal device to obtain the virtual content on the screen of the terminal device On the projection area.
  • the terminal device 220 is a tablet computer, with the human eye 221 as the origin (also can be regarded as the origin of the head-mounted display device), the extension line of the human eye 221 and the virtual content 223, and the tablet screen
  • the area formed by the intersection point on the screen is the projection area 225.
  • the specific method for obtaining the projection area may not be limited in this application.
  • the outline of the projection area of the virtual content on the screen of the terminal device may be obtained based only on the outline vertex of the virtual content and the spatial position of the terminal device in the virtual space. To get the projection area.
  • the head-mounted display device may transmit the display data of the virtual content to the terminal device, the display data may include the spatial location of the virtual content in the virtual space, and the terminal device may use the above-mentioned methods according to the display data of the virtual content Get the projection area.
  • the terminal device may also obtain the projection area from the head-mounted display device. The head-mounted display device may obtain the projection area in the above manner, and then transmit the data of the projection area to the terminal device.
  • Step 2130 Acquire image content corresponding to the projection area in the screen content according to the screen content to be displayed on the screen.
  • the image 227 is the image content corresponding to the projection area in the screen content.
  • the terminal device can obtain the two-dimensional coordinates of the projection area on the screen of the terminal device, and can convert the two-dimensional coordinates to the coordinates of the image coordinate system of the screen content according to the conversion parameters of the screen coordinate system and the image coordinate system of the screen content to obtain the projection
  • the area corresponds to the image area in the screen content, and the image content in the image area is acquired.
  • Step 2140 Perform specified processing on the image content and display the screen content containing the specified processed image content.
  • the difference between the first hue of the specified processed image content and the second hue of the virtual content is greater than the first threshold .
  • the terminal device can specify the image content corresponding to the projection area in the screen content and display the processed screen content, which can reduce the interference with the virtual content displayed by the head-mounted display device.
  • the difference between the first color tone of the designated processed image content and the second color tone of the virtual content is greater than a first threshold, which can be set reasonably according to the user's viewing comfort.
  • a first threshold which can be set reasonably according to the user's viewing comfort.
  • the virtual content overlay display can be readable and recognizable on the designated image content after processing.
  • the image content can highlight the virtual content and reduce the interference of the content displayed by the terminal device with the virtual content.
  • Hue can refer to the lightness or darkness of an image.
  • the hue includes color and transparency
  • the specified processing may be to set the color of the image content or to set the transparency of the image content. For example, adjust the color of the image content to a solid color (such as black, gray, etc.), adjust the transparency of the image content to 50% transparency, and overlay a solid color picture on the image content.
  • the terminal device 230 establishes a wireless communication connection with the head-mounted display device, the head-mounted display device recognizes the marker 231 on the terminal device 230 and displays the virtual animal 233, and the user can see the virtual animal 233 through the worn head-mounted display device Superimposed and displayed on the terminal device 230 in the real space, and the color of the image content 235 (stone, grass, etc.) corresponding to the projection area of the virtual content in the screen content displayed on the terminal device 230 is adjusted to gray, which can highlight the virtual animal 233, Reduce the interference of image content 235 to virtual animals.
  • the image processing method of the above embodiment reduces the interference of the screen content displayed by the terminal device on the virtual content when the virtual content is displayed by the head-mounted display device, highlights the virtual content displayed by the head-mounted display device, and improves the realism of the virtual content in augmented reality And display effects; and only process the screen content corresponding to the projection area of the virtual content on the screen, while highlighting the virtual content, to ensure the normal display of other content on the screen.
  • FIG. 24 another embodiment of the present application provides an image processing method, which is applied to a terminal device.
  • the terminal device and a head-mounted display device are communicatively connected, including the following steps.
  • Step 2410 Obtain the relative spatial position relationship between the terminal device and the head-mounted display device.
  • Step 2420 Obtain a projection area of the virtual content displayed on the head-mounted display device on the screen of the terminal device according to the relative spatial position relationship, and the projection area is a projection area of the virtual content observed on the screen by the human eye through the head-mounted display device.
  • the above-mentioned obtaining the projection area of the virtual content displayed on the head-mounted display device on the screen of the terminal device according to the relative spatial position relationship may include steps 2422-2428.
  • Step 2422 Obtain a left-eye image and a right-eye image for displaying virtual content in the head-mounted display device.
  • the display data of the virtual content may include a left-eye image and a right-eye image for forming three-dimensional virtual content in the human eye through the reflection of the optical element.
  • the left-eye image and the right-eye image have parallax.
  • the left-eye image is used to project to the user's left eye through the optical lens during display
  • the right-eye image is used to project to the user's right eye through the optical lens during display.
  • the fusion of the image and the right-eye image through the user's brain can form a stereoscopic image, so that the user can see the display effect of the stereoscopic image.
  • the head-mounted display device can obtain rendering coordinates of the virtual content, and render the virtual content based on the rendering coordinates.
  • the rendering coordinates may be the spatial coordinates of each point of the virtual content in the virtual space with the virtual camera as the origin.
  • the virtual camera includes a left virtual camera and a right virtual camera.
  • the left virtual camera is used to simulate the left eye of the human eye
  • the right virtual camera is used to simulate the right eye of the human eye.
  • the left rendering coordinates of the virtual content in the first spatial coordinate system with the left virtual camera as the origin and the right rendering coordinates in the second spatial coordinate system with the right virtual camera as the origin can be obtained.
  • the virtual content can be rendered according to the left rendering coordinates to obtain the left-eye image of the virtual content, and the virtual content can be rendered according to the right rendering coordinates to obtain the right-eye image of the virtual content.
  • the head-mounted display device can transmit the left-eye image and the right-eye image of the virtual content to the terminal device.
  • Step 2424 Acquire the first projection area of the left-eye image on the screen of the terminal device according to the relationship between the left-eye image and the relative spatial position.
  • the terminal device may acquire the first projection area of the left-eye image on the screen according to the relative spatial position relationship between the left-eye image and the head-mounted display device.
  • the first projection area may be a plane on which the shape of the virtual content in the left-eye image is projected onto the screen of the terminal device, and the projection area on the plane coincides with the screen area of the terminal device.
  • the projection area may partially overlap the screen area of the terminal device, or may be entirely contained in the screen area. As shown in FIG.
  • the terminal device 260 is a tablet computer, and the left-eye display image of the virtual content 261 can be incident on the left eye 262 of the user after being reflected by the optical lens of the head-mounted display device, and the left-eye display image corresponds to the screen of the tablet computer 'S first projection area 264.
  • the terminal device may acquire the virtual content in the left virtual camera and the left-eye image according to the spatial position of the left virtual camera, the left-eye image, and the terminal device in the same spatial coordinate system (such as the world coordinate system) in the virtual space.
  • the intersection of the extension line of each vertex and the screen of the terminal device, the intersection is the projection point of the vertex on the screen.
  • the space coordinates of the projection point of the left-eye image on the screen of the virtual space can be converted into two-dimensional coordinates of the screen coordinate system of the terminal device according to the conversion parameters of the space coordinate system of the virtual space and the screen coordinate system of the terminal device The first projection area of the left-eye image on the screen.
  • the first projection area of the left-eye image on the screen of the terminal device may be acquired based on the spatial position of the outline area of the virtual content in the left-eye image on the virtual space to simplify the calculation steps and optimize the processing process of the terminal device.
  • the acquiring the first projection area of the left-eye image on the screen of the terminal device according to the left-eye image and the relative spatial position relationship may include steps 2702-2708.
  • Step 2702 Acquire the first outline coordinate set of virtual content in the virtual space in the left-eye image.
  • the terminal device may acquire the first outline coordinate set of the virtual content of the left-eye image in the virtual space to obtain the first projection area according to the first outline coordinate set.
  • the first set of outline coordinates may be the coordinates of each vertex of the outline area of the virtual content in the left-eye image in the space coordinate system.
  • the space coordinates of the outline area of the virtual content in the first space coordinate system can be directly used as the first outline coordinate set, or the conversion parameters of the first space coordinate system and the world coordinate system can be used to display the content of the virtual content in the left-eye display image.
  • the spatial coordinates of the contour area in the first spatial coordinate system are converted into the spatial coordinates in the world coordinate system to obtain the first contour coordinate set.
  • Step 2704 Acquire the screen coordinate set of the screen in the virtual space according to the relative spatial position relationship.
  • the terminal device may acquire the spatial position of the screen of the terminal device in the real space according to the relative position relationship, and convert the spatial position into spatial coordinates in the virtual space to obtain the screen coordinate set of the screen in the virtual space.
  • Step 2706 Establish a first connection between the left virtual camera in the virtual space and each point in the first outline coordinate set, and obtain the coordinates of the point on each first connection in the screen coordinate set to obtain the first A set of coordinates.
  • the terminal device may establish a first connection between the left virtual camera in the virtual space and each point in the first outline coordinate set, and obtain the coordinates of the point on each first connection established in the screen coordinate set to obtain the first A set of intersection coordinates.
  • the coordinates of the left virtual camera, the first outline coordinate set, the screen coordinate set, and the first intersection coordinate set are coordinates under the same spatial coordinate system.
  • the terminal device may determine whether the coordinates of the point on the first connection line exist in the screen coordinate set. If there is a point on the first line, it can be determined that the first line crosses the screen, otherwise, it is determined that the first line does not cross the screen.
  • the first intersection coordinate set can be obtained according to the coordinates of the point on the first line in the screen coordinate set.
  • the first intersection coordinate set includes coordinates of projection points of each vertex of the outline area of the virtual content in the left-eye image on the screen.
  • Step 2708 Acquire the first projection area of the left-eye image on the screen according to the first intersection coordinate set.
  • the terminal device may convert the first intersection coordinate set into two-dimensional coordinates of the terminal device's screen coordinate system according to the conversion parameters of the space coordinate system in the virtual space and the screen coordinate system of the terminal device to obtain the first projection area.
  • Step 2426 Acquire the second projection area of the right-eye display image on the screen according to the relationship between the right-eye image and the relative spatial position.
  • the terminal device may refer to the above corresponding step of acquiring the first projection area of the left-eye image on the screen of the terminal device according to the left-eye image and the relative spatial position relationship, and acquire the second projection area of the right-eye image on the screen.
  • the right-eye image of the virtual content 261 can be incident on the right eye 263 of the user after being reflected by the optical lens of the head-mounted display device, and the right-eye image corresponds to the second projection area 265 on the screen of the tablet computer.
  • the terminal device may acquire the second outline coordinate set of the virtual content in the right-eye image in the virtual space, obtain the screen coordinate set of the screen in the virtual space according to the relative spatial position relationship, and respectively establish a right virtual camera in the virtual space
  • the second connection line with each point in the second outline coordinate set and obtain the coordinates of the points on each established second connection line in the screen coordinate set to obtain the second intersection coordinate set, and then according to the second coordinate set, Acquire the second projection area of the right-eye image on the screen.
  • Step 2428 Acquire the synthesis area of the first projection area and the second projection area, and use the synthesis area as the projection area of the virtual content on the screen.
  • the combined area of the first projection area and the second projection area refers to the area formed by the coordinate set obtained by combining the first intersection coordinate set corresponding to the first projection area and the second intersection coordinate set corresponding to the second projection area, That is, the union area of the first projection area and the second projection area.
  • the above process of obtaining the projection area of the virtual content on the screen may also be performed on the head-mounted display device, and then transmit the data of the projection area to the terminal device to reduce the calculation amount of the terminal device and optimize The processing of terminal equipment.
  • Step 2430 Acquire image content corresponding to the projection area in the screen content according to the screen content to be displayed on the screen.
  • Step 2440 Perform specified processing on the image content and display the screen content containing the specified processed image content.
  • the difference between the first hue of the specified processed image content and the second hue of the virtual content is greater than the first threshold .
  • the above-mentioned process of specifying the image content includes any one or more of the following: overlaying the content of the specified color on the image content; adjusting the color of the image content to the specified color; and transparency of the image content The value is adjusted to the specified transparency value.
  • the color difference between the specified color and the color of the virtual content is greater than the second threshold, and the difference between the specified transparency value and the transparency value of the virtual content is greater than the third threshold.
  • the overlay content may be a picture of a specified color, or a newly-created layer of a specified color, which is not limited herein, and only needs to cover the image content, and the presentation color may be the specified color.
  • the specified color can be a solid color, such as black, gray, etc.
  • the second threshold can be set reasonably according to the user's viewing comfort. For example, when the color of the virtual content is blue, the specified color can be gray.
  • the specified transparency value can be used to reduce the display effect of the image content.
  • the smaller the transparency value the higher the transparency. If the specified transparency value can be set to 5, which is 50% transparent, it can also be set to 0, which is 100% transparent.
  • the third threshold can be set reasonably according to the user's observation comfort, which is not limited here. For example, the third threshold is set to 5, when the transparency value of the virtual content is 10, that is, 0% transparency, the specified transparency value may be 1, that is, 90% transparency.
  • the transparency value of the image content may also be set to be less than a preset threshold, for example, the preset threshold may be 1 or 2, and so on.
  • the head-mounted display device may only include a display module, a communication module, and a camera for display, and control of the displayed virtual content may be performed by a processor, memory, etc. of the terminal device.
  • the terminal device includes a manipulation area, and when the manipulation area detects a manipulation operation, a control instruction may be generated according to the manipulation operation, the virtual content displayed in the head-mounted display device may be adjusted according to the control instruction, and the adjusted virtual content The corresponding display data is sent to the head-mounted display device.
  • the terminal device may reacquire the projection area based on the adjusted virtual content, and specify the image content corresponding to the newly acquired projection area in the screen content.
  • the user views the virtual animal 233 superimposed on the screen of the terminal device through the worn head-mounted display device, can move the virtual animal 233 to the right, and re-determines the projection area of the virtual animal 233 on the screen,
  • the color of the image content 235 corresponding to the projection area in the screen content is adjusted to gray, and the virtual animal is highlighted, so that the user can clearly observe the virtual animal 233 all the time, reducing the interference of the screen content on the virtual animal.
  • an embodiment of the present application provides an image processing method, which is applied to a head-mounted display device, and the head-mounted display device is in communication connection with a terminal device.
  • the method includes the following steps.
  • Step 2910 Display the virtual content.
  • Step 2920 Obtain the relative spatial position relationship between the terminal device and the head-mounted display device.
  • Step 2930 Acquire the projection area of the virtual content on the screen of the terminal device according to the relative spatial position relationship.
  • the projection area is the projection area of the virtual content on the screen observed by the human eye through the head-mounted display device.
  • Step 2940 Send the data of the projection area to the terminal device.
  • the data of the projection area is used to instruct the terminal device to specify the image content corresponding to the projection area in the screen content to be displayed, and display the content containing the image content after the specified processing.
  • the screen content specifies the difference in tone between the first hue of the processed image content and the second hue of the virtual content is greater than the first threshold.
  • the image processing method provided by the above embodiment reduces the interference of the screen content displayed by the terminal device on the virtual content when the virtual content is displayed on the head-mounted display device, thereby highlighting the virtual content and improving the realism and display effect of the virtual content in augmented reality . Moreover, only the screen content corresponding to the projection area of the virtual content on the screen is processed, while the virtual content is highlighted, the normal display of other screen content is ensured, which is convenient for the user to interact with the mobile terminal.
  • a terminal device may be an electronic device capable of running an application program, such as a smartphone or a tablet computer.
  • the terminal device may include one or more of the following components: a processor and a memory, where the memory stores one or more application programs, the one or more application programs may be configured to be executed by one or more processors, one or Multiple programs are configured to perform the method as described in the foregoing method embodiments.
  • the processor may include one or more processing cores.
  • the processor uses various interfaces and lines to connect the various parts of the entire terminal device, executes each instruction of the terminal device by running or executing instructions, programs, code sets or instruction sets stored in the memory, and calling data stored in the memory kindss of functions and processing data.
  • the processor may be implemented in at least one hardware form of digital signal processing, field programmable gate array, and programmable logic array.
  • the processor may integrate one or a combination of a central processor, an image processor, and a modem. Among them, the CPU mainly handles the operating system, user interface and application programs, etc.; the GPU is used for rendering and rendering of the displayed content; and the modem is used for processing wireless communication. It can be understood that the above-mentioned modem may not be integrated into the processor 120, and may be implemented by a communication chip alone.
  • the memory may include random access memory or read-only memory.
  • the memory can be used to store instructions, programs, codes, code sets or instruction sets.
  • the memory 130 may include a storage program area and a storage data area, where the storage program area may store instructions for implementing an operating system and instructions for implementing at least one function (such as a touch function, a sound playback function, an image playback function, etc.) , Instructions for implementing the following method embodiments.
  • the storage data area can also store data created by the terminal device in use, and the like.
  • a head-mounted display device which includes a camera for collecting images of real objects and collecting scene images of a target scene.
  • the camera may be an infrared camera or a visible light camera, and the specific type is not limited.
  • the first terminal 100 may further include one or more components as follows: a display module, an optical module, a communication module, and a power supply.
  • the display module may include a display control unit for receiving a display image of the virtual content rendered by the processor, displaying and projecting the display image onto the optical module, so that the user can view the virtual content through the optical module .
  • the display module may be a display screen or a projection device, etc., for displaying images.
  • the optical module may use an off-axis optical system or a waveguide optical system. After the display image displayed by the display module passes through the optical module, it can be projected to the user's eyes. The user can see the display image projected by the display module through the optical module.
  • the user can also observe the real environment through the optical module and feel the visual effect of the superimposed virtual content and the real environment.
  • the communication module may be a module such as Bluetooth, WiFi, or ZigBee.
  • the terminal device may communicate with the interactive device through the communication module to exchange information and instructions.
  • the power supply can supply power to the entire terminal equipment to ensure the normal operation of various components of the terminal equipment.
  • the head-mounted display device may further include a processor and a memory, where the memory stores one or more application programs, the one or more application programs may be configured to be executed by one or more processors, one Or more programs are configured to perform the method as described in the foregoing method embodiments.
  • a computer-readable storage medium stores program code, and the program code may be called by a processor to execute the method described in the foregoing method embodiments.
  • the computer-readable storage medium may be an electronic memory such as flash memory, EEPROM, EPROM, hard disk, or ROM.
  • the computer-readable storage medium includes a non-volatile computer-readable medium.
  • the computer-readable storage medium 800 has a storage space for the program code 810 that performs any of the method steps described above. These program codes can be read from or written into one or more computer program products.
  • the program code 810 may be compressed in an appropriate form, for example.

Abstract

A virtual content interaction system, comprising a terminal device and an external head-mounted display device. The terminal device establishes a communicational connection with the head-mounted display device; the terminal device is configured to obtain first relative space position information of the head-mounted display device and the terminal device, determine a display position of virtual content according to the first relative space position information, render the virtual content according to the display position, obtain display data of the virtual content, and transmit the display data to the head-mounted display device; the head-mounted display device is configured to receive the display data sent by the terminal device and display the virtual content according to the display data.

Description

虚拟内容交互系统及方法Virtual content interaction system and method 技术领域Technical field
本申请涉及增强现实领域,尤其涉及一种虚拟内容交互系统及方法。This application relates to the field of augmented reality, in particular to a virtual content interaction system and method.
背景技术Background technique
随着科技的发展,机器智能化及信息智能化日益普及,通过机器视觉或者虚拟视觉等图像采集装置来识别用户影像以实现人机交互的技术越来越重要。增强现实技术借助计算机图形技术和可视化技术构建现实环境中不存在的虚拟内容,并通过图像识别定位技术将虚拟内容准确地融合到真实环境中,借助显示设备将虚拟内容与真实环境融为一体,并显示给使用者真实的感观体验。增强现实技术要解决的首要技术难题是如何将虚拟内容准确地融合到真实世界中,也就是要使虚拟内容以正确的角度姿态出现在真实场景的正确位置上,以产生强烈的视觉真实感。因此,与虚拟内容的交互是增强现实技术的重要研究方向。With the development of science and technology, machine intelligence and information intelligence are becoming more and more popular. The technology of identifying user images through machine vision or virtual vision and other image acquisition devices to realize human-computer interaction is becoming more and more important. Augmented reality technology uses computer graphics technology and visualization technology to construct virtual content that does not exist in the real environment, and accurately integrates the virtual content into the real environment through image recognition and positioning technology, and integrates the virtual content and the real environment with the help of display devices. And show the user a real sensory experience. The primary technical problem to be solved by augmented reality technology is how to accurately integrate the virtual content into the real world, that is, to make the virtual content appear in the correct position of the real scene at the correct angle posture, so as to generate a strong visual reality. Therefore, the interaction with virtual content is an important research direction of augmented reality technology.
发明内容Summary of the invention
本申请实施例提出了一种虚拟内容交互系统及方法。The embodiments of the present application propose a virtual content interaction system and method.
第一方面,本申请实施例提供了一种虚拟内容交互系统,包括终端设备以及外接式的头戴显示设备,所述终端设备与所述头戴显示设备建立通信连接,其中,所述终端设备,用于获取所述头戴显示设备与所述终端设备的第一相对空间位置信息,根据所述第一相对空间位置信息,确定虚拟内容的显示位置,根据所述显示位置渲染所述虚拟内容,并获取所述虚拟内容的显示数据,以及将所述显示数据传输至所述头戴显示设备;所述头戴显示设备,用于接收所述终端设备发送的所述显示数据,并根据所述显示数据显示所述虚拟内容。In a first aspect, an embodiment of the present application provides a virtual content interaction system, including a terminal device and an external head-mounted display device, the terminal device establishing a communication connection with the head-mounted display device, wherein the terminal device For acquiring first relative spatial position information of the head-mounted display device and the terminal device, determining a display position of virtual content according to the first relative spatial position information, and rendering the virtual content according to the display position , And obtain the display data of the virtual content, and transmit the display data to the head-mounted display device; the head-mounted display device is used to receive the display data sent by the terminal device, and according to the The display data displays the virtual content.
第二方面,本申请实施例提供了一种虚拟内容交互方法,应用于终端设备,所述终端设备与外接式的头戴显示设备通信连接,所述方法包括:获取所述头戴显示设备与所述终端设备的第一相对空间位置;根据所述第一相对空间位置,确定虚拟内容的显示位置;根据所述显示位置渲染所述虚拟内容,并获取所述虚拟内容的显示数据;及将所述显示数据传输至所述头戴显示设备,所述显示数据用于指示所述头戴显示设备显示所述虚拟内容。In a second aspect, an embodiment of the present application provides a virtual content interaction method, which is applied to a terminal device, and the terminal device is in communication connection with an external head-mounted display device. The method includes: acquiring the head-mounted display device and The first relative spatial position of the terminal device; determining the display position of the virtual content according to the first relative spatial position; rendering the virtual content according to the display position, and obtaining display data of the virtual content; and The display data is transmitted to the head-mounted display device, and the display data is used to instruct the head-mounted display device to display the virtual content.
第三方面,本申请实施例提供了一种虚拟内容交互方法,应用于头戴显示设备,所述头戴显示设备与终端设备建立通信连接,包括:通过相机采集包含有标记物的图像,所述标记物为设置于所述终端设备上;将所述标记物图像传输至所述终端设备,所述图像用于指示所述终端设备对所述图像中包含标记物进行识别,并获取所述头戴显示设备与所述终端设备的第一相对空间位置;及接收所述终端设备发送的显示数据,并根据所述显示数据显示虚拟内容,所述显示数据由所述终端设备根据所述第一相对空间位置渲染所述虚拟内容得到。In a third aspect, an embodiment of the present application provides a virtual content interaction method, which is applied to a head-mounted display device, and the head-mounted display device establishes a communication connection with a terminal device, including: collecting an image containing a marker through a camera. The marker is provided on the terminal device; the marker image is transmitted to the terminal device, and the image is used to instruct the terminal device to recognize the marker contained in the image and obtain the A first relative spatial position of the head-mounted display device and the terminal device; and receiving display data sent by the terminal device, and displaying virtual content according to the display data, the display data is determined by the terminal device according to the first The virtual content is obtained by rendering the virtual content at a relative spatial position.
第四方面,本申请实施例提供了一种虚拟内容显示方法,应用于第一终端,所述第一终端与第二终端通信连接,所述方法包括:获取所述第一终端与所述第二终端之间的第一相对空间位置信息;从所述第二终端获取显示内容数据,其中,所述显示内容数据至少包括所述第二终端当前显示的显示内容的数据;及根据所述第一相对空间位置信息以及所述显示内容数据,对虚拟内容进行显示,其中,所述虚拟内容包括所述第二终端显示的显示内容以及与所述显示内容对应的扩展内容。According to a fourth aspect, an embodiment of the present application provides a virtual content display method, which is applied to a first terminal, and the first terminal is in communication connection with a second terminal. The method includes: acquiring the first terminal and the first terminal First relative spatial position information between the two terminals; obtaining display content data from the second terminal, wherein the display content data includes at least the data of the display content currently displayed by the second terminal; and according to the first A relative spatial position information and the display content data to display virtual content, wherein the virtual content includes the display content displayed by the second terminal and the extended content corresponding to the display content.
第五方面,本申请实施例提供了一种图像处理方法,应用于终端设备,所述终端设备与头戴显示设备通信连接,所述方法包括:获取所述终端设备与所述头戴显示设备 之间的相对空间位置关系;根据所述相对空间位置关系,获取所述头戴显示设备显示的虚拟内容在所述终端设备的屏幕上的投影区域,所述投影区域为通过所述头戴显示设备观察到的所述虚拟内容在所述屏幕上的投射区域;根据所述屏幕待显示的屏幕内容,获取所述屏幕内容中与所述投影区域对应的图像内容;及对所述图像内容进行指定处理,并显示包含指定处理后的图像内容的屏幕内容,所述指定处理后的图像内容的第一色调与所述虚拟内容的第二色调之间的色调差值大于第一阈值。According to a fifth aspect, an embodiment of the present application provides an image processing method, which is applied to a terminal device, and the terminal device is in communication connection with a head-mounted display device. The method includes: acquiring the terminal device and the head-mounted display device Relative spatial position relationship between them; according to the relative spatial position relationship, acquiring a projection area of the virtual content displayed by the head-mounted display device on the screen of the terminal device, the projection area being displayed by the head-mounted display The projection area of the virtual content on the screen observed by the device; obtaining the image content corresponding to the projection area in the screen content according to the screen content to be displayed on the screen; and performing on the image content Specify the process and display the screen content containing the image content after the specified process, and the difference in tone between the first tone of the specified image content and the second tone of the virtual content is greater than the first threshold.
第六方面,本申请实施例提供了一种终端设备,包括:一个或多个处理器;存储器;其中存储器存储有一个或多个应用程序,一个或多个应用程序被配置为由所述一个或多个处理器执行,所述一个或多个程序配置用于执行上述提供的方法。In a sixth aspect, an embodiment of the present application provides a terminal device, including: one or more processors; a memory; wherein the memory stores one or more application programs, and the one or more application programs are configured to be configured by the one Or executed by multiple processors, and the one or more programs are configured to execute the method provided above.
第七方面,本申请实施例提供了一种计算机可读取存储介质,所述计算机可读取存储介质中存储有程序代码,所述程序代码可被处理器调用执行上述提供的方法。According to a seventh aspect, an embodiment of the present application provides a computer-readable storage medium. The computer-readable storage medium stores a program code, and the program code can be called by a processor to execute the method provided above.
附图说明BRIEF DESCRIPTION
图1为本申请实施例的一种应用场景示意图。FIG. 1 is a schematic diagram of an application scenario according to an embodiment of the present application.
图2为本申请实施例的增强现实系统示意图。FIG. 2 is a schematic diagram of an augmented reality system according to an embodiment of this application.
图3为本申请实施例的另一种应用场景示意图。FIG. 3 is a schematic diagram of another application scenario according to an embodiment of the present application.
图4为一个实施例中虚拟内容交互方法流程图。FIG. 4 is a flowchart of a virtual content interaction method in an embodiment.
图5为一个实施例中显示虚拟内容的示意图。FIG. 5 is a schematic diagram showing virtual content in an embodiment.
图6为一个实施例中显示终端设备屏幕上的扩展内容的示意图。FIG. 6 is a schematic diagram showing extended content on a screen of a terminal device in an embodiment.
图7为一个实施例中显示终端设备屏幕上的扩展内容的示意图。FIG. 7 is a schematic diagram showing extended content on a screen of a terminal device in an embodiment.
图8为一个实施例中虚拟内容交互方法流程图。FIG. 8 is a flowchart of a virtual content interaction method in an embodiment.
图9为一个实施例中确定虚拟内容的显示位置的流程图。FIG. 9 is a flowchart of determining the display position of virtual content in one embodiment.
图10为一个实施例中显示虚拟内容的示意图。FIG. 10 is a schematic diagram showing virtual content in one embodiment.
图11为一个实施例中显示的虚拟内容与终端设备的屏幕内容存在重叠的示意图。FIG. 11 is a schematic diagram showing that the virtual content displayed in one embodiment overlaps with the screen content of the terminal device.
图12为一个实施例中显示虚拟内容的示意图。FIG. 12 is a schematic diagram showing virtual content in one embodiment.
图13A-13B为一个实施例中根据终端设备的屏幕内容变化更新显示的虚拟内容示意图。13A-13B are schematic diagrams of updating the displayed virtual content according to the change of the screen content of the terminal device in one embodiment.
图14A-14C为一个实施例中终端设备的操控区滑动的示意图。14A-14C are schematic diagrams of the sliding of the control area of the terminal device in an embodiment.
图15为一个实施例中根据手指滑动轨迹控制虚拟内容的示意图。15 is a schematic diagram of controlling virtual content according to a finger sliding trajectory in one embodiment.
图16为另一个实施例中根据手指滑动轨迹控制虚拟内容的示意图。FIG. 16 is a schematic diagram of controlling virtual content according to a finger sliding trajectory in another embodiment.
图17A-17B为又一个实施例中根据手指滑动轨迹控制虚拟内容的示意图。17A-17B are schematic diagrams of controlling virtual content according to a finger sliding trajectory in yet another embodiment.
图18为再一个实施例中根据手指滑动轨迹控制虚拟内容的示意图。FIG. 18 is a schematic diagram of controlling virtual content according to a sliding track of a finger in still another embodiment.
图19为又一个实施例中虚拟内容交互方法流程图。FIG. 19 is a flowchart of a virtual content interaction method in yet another embodiment.
图20为一个实施例中增强现实系统的框图。Figure 20 is a block diagram of an augmented reality system in one embodiment.
图21为一个实施例中图像处理方法的流程图。21 is a flowchart of an image processing method in an embodiment.
图22为一个实施例中虚拟内容在终端设备的屏幕上的投影区域的示意图。22 is a schematic diagram of a projection area of virtual content on a screen of a terminal device in an embodiment.
图23为一个实施例中显示虚拟内容及处理后的屏幕内容的示意图。23 is a schematic diagram showing virtual content and processed screen content in an embodiment.
图24为另一个实施例中图像处理方法的流程图。24 is a flowchart of an image processing method in another embodiment.
图25为一个实施例中获取虚拟内容在终端设备的屏幕上的投影区域的流程图。FIG. 25 is a flowchart of acquiring a projection area of virtual content on a screen of a terminal device in an embodiment.
图26为另一个实施例中虚拟内容在终端设备的屏幕上的投影区域的示意图。26 is a schematic diagram of a projection area of virtual content on a screen of a terminal device in another embodiment.
图27为一个实施例中获取左眼图像在屏幕上的第一投影区域的流程图。FIG. 27 is a flowchart of acquiring the first projection area of the left-eye image on the screen in one embodiment.
图28为另一个实施例中显示虚拟内容及处理后的屏幕内容的示意图。FIG. 28 is a schematic diagram showing virtual content and processed screen content in another embodiment.
图29为又一个实施例中图像处理方法的流程图。FIG. 29 is a flowchart of an image processing method in still another embodiment.
图30为一个实施例中头戴显示设备的框图。30 is a block diagram of a head-mounted display device in one embodiment.
具体实施方式detailed description
请参图1-2,本申请实施例提供的增强现实系统10包括:终端设备100、与终端设备100连接的头戴显示设备200。Referring to FIGS. 1-2, the augmented reality system 10 provided by the embodiment of the present application includes: a terminal device 100 and a head-mounted display device 200 connected to the terminal device 100.
在一实施例中,终端设备100可被用户手持并进行操控,可为手机、智能手表、平板电脑、电子阅读器、笔记本电脑等能够运行应用程序的电子设备。In an embodiment, the terminal device 100 may be held and controlled by a user, and may be an electronic device capable of running application programs, such as a mobile phone, a smart watch, a tablet computer, an e-reader, and a notebook computer.
在一些实施例中,头戴显示设备200可为外接式头戴显示设备,即头戴显示设备200包含用于显示的显示模组、通信模块和相机等,依靠与头戴显示设备200连接的终端设备100的处理器、存储器等对显示的虚拟内容进行控制。其中,显示模组可包括显示屏(或投影装置)和显示镜片,以对虚拟内容进行显示。In some embodiments, the head-mounted display device 200 may be an external head-mounted display device, that is, the head-mounted display device 200 includes a display module, a communication module, a camera, etc. for display. The processor, memory, etc. of the terminal device 100 control the displayed virtual content. The display module may include a display screen (or a projection device) and a display lens to display virtual content.
其中,头戴显示设备200可与终端设备100进行信息及指令的传输。其中,传输的信息可包括头戴显示设备200显示的虚拟内容。终端设备100与头戴显示设备200可通过蓝牙、WiFi、ZigBee等无线通信方式连接,也可通过USB等接口进行有线通信连接,如图2,头戴显示设备200通过USB接口与平板电脑、手机终端等有线通信连接。Among them, the head-mounted display device 200 can transmit information and instructions with the terminal device 100. Among them, the transmitted information may include virtual content displayed by the head-mounted display device 200. The terminal device 100 and the head-mounted display device 200 can be connected via wireless communication methods such as Bluetooth, WiFi, ZigBee, etc., and can also be connected by wired communication through an interface such as a USB, as shown in FIG. Wired communication connections such as terminals.
在一些实施例中,终端设备100上设置有标记物101。其中,标记物101可为任意具有可识别特征标记的图形或物体,能被头戴显示设备200采集追踪即可,例如,标记物101可为具有拓扑结构的图案,拓扑结构是指标记物中的子标记物和特征点等之间连通关系。当标记物101处于头戴显示设备200的相机的视觉范围内时,该相机可实时采集包含该标记物101的图像,并将该图像传输给终端设备100,终端设备100可根据该图像获取标记物101与头戴显示设备200之间的相对位置、朝向等空间位置信息,以获得终端设备100与头戴显示设备200之间的相对空间位置信息。终端设备100可基于与头戴显示设备200之间的空间位置信息渲染相应的虚拟对象,并将虚拟对象的显示画面传输给头戴显示设备200,头戴显示设备200可通过显示模组显示虚拟对象。In some embodiments, the marker 101 is provided on the terminal device 100. The marker 101 can be any graphic or object with identifiable features, which can be collected and tracked by the head-mounted display device 200. For example, the marker 101 can be a pattern with a topology structure, which refers to the marker The connected relationship between the sub-markers and the feature points. When the marker 101 is within the visual range of the camera of the head-mounted display device 200, the camera can collect an image containing the marker 101 in real time and transmit the image to the terminal device 100, and the terminal device 100 can obtain the marker according to the image The relative position, orientation, and other spatial position information between the object 101 and the head-mounted display device 200 to obtain relative spatial position information between the terminal device 100 and the head-mounted display device 200. The terminal device 100 may render a corresponding virtual object based on the spatial position information between the head-mounted display device 200 and transmit the display screen of the virtual object to the head-mounted display device 200, and the head-mounted display device 200 may display the virtual object through the display module Object.
在一些实施方式中,头戴显示设备200还可追踪终端设备100的形状,确定终端设备100与头戴显示设备200之间的相对空间位置关系。In some embodiments, the head-mounted display device 200 can also track the shape of the terminal device 100 to determine the relative spatial position relationship between the terminal device 100 and the head-mounted display device 200.
如图1,终端设备100与头戴显示设备200无线通信连接,用户可通过佩戴的头戴显示设备200,扫描终端设备100上的标记物101,观看到包含有多个虚拟星球的虚拟太空场景叠加在现实空间,其中,虚拟太空场景与终端设备100显示的太空场景对应。As shown in FIG. 1, the terminal device 100 is wirelessly connected to the head-mounted display device 200, and the user can scan the marker 101 on the terminal device 100 through the head-mounted display device 200 worn to watch a virtual space scene containing multiple virtual planets Superimposed on the real space, where the virtual space scene corresponds to the space scene displayed by the terminal device 100.
如图3,终端设备100与头戴显示设备200有线通信连接,用户通过佩戴的头戴显示设备200,扫描终端设备100上的标记物101,可观看到包含虚拟的医学人体模型叠加显示在现实空间中的终端设备100的表面上。As shown in FIG. 3, the terminal device 100 and the head-mounted display device 200 are connected by wire communication, and the user scans the marker 101 on the terminal device 100 through the head-mounted display device 200 worn, and can view the superimposed display of the medical human model including the virtual reality in reality On the surface of the terminal device 100 in the space.
请参图4,本申请实施例提供了一种虚拟内容交互方法,应用于终端设备,终端设备与外接式的头戴显示设备通信连接,该方法包括如下步骤。Referring to FIG. 4, an embodiment of the present application provides a virtual content interaction method, which is applied to a terminal device. The terminal device is in communication connection with an external head-mounted display device. The method includes the following steps.
步骤410:获取头戴显示设备与终端设备的第一相对空间位置。Step 410: Obtain the first relative spatial position of the head-mounted display device and the terminal device.
第一相对空间位置可包括头戴显示设备与终端设备之间的相对位置及姿态信息,姿态信息可为终端设备相对头戴显示设备的朝向及旋转角度。The first relative spatial position may include relative position and posture information between the head-mounted display device and the terminal device, and the posture information may be the orientation and rotation angle of the terminal device relative to the head-mounted display device.
在一些实施方式中,终端设备包括惯性测量单元(IMU),可先获取IMU的测量数据,并根据该测量数据,确定终端设备与头戴显示设备之间的第一相对空间位置。In some embodiments, the terminal device includes an inertial measurement unit (IMU), which may first acquire the measurement data of the IMU, and determine the first relative spatial position between the terminal device and the head-mounted display device based on the measurement data.
在一些实施方式中,终端设备上还可设置有光点,头戴显示设备通过相机采集包含光点的图像,并将该图像发送给终端设备。终端设备可对该图像中的光点进行识别,确定头戴显示设备与终端设备的第一相对空间位置。终端设备上设置的光点可为可见光点,也可为红外光点等。当光点为红外光点时,头戴显示设备上可设置有红外相机,用于采集该红外光点的图像。终端设备上设置的光点可为一个,也可为由多个光点组成的光点序列。In some embodiments, a light spot may also be provided on the terminal device, and the head-mounted display device collects an image containing the light spot through the camera and sends the image to the terminal device. The terminal device may identify the light spot in the image to determine the first relative spatial position of the head-mounted display device and the terminal device. The light spot set on the terminal device may be a visible light spot or an infrared light spot. When the light spot is an infrared light spot, an infrared camera may be provided on the head-mounted display device to collect an image of the infrared light spot. The light spot set on the terminal device may be one, or may be a light spot sequence composed of multiple light spots.
在一个实施例中,光点可设置在终端设备的外壳上,例如可设置在屏幕的周围。光点也可设置在终端设备的保护套上,终端设备进行使用时,可套上包含光点的保护套,从而可实现对终端设备的追踪。例如,可在终端设备的屏幕四周分别设置不同的光点,如在四周设置不同数量的光点,或不同颜色的光点,以使终端设备根据光点图像中各个光点的分布,确定与头戴显示设备的相对空间位置。In one embodiment, the light spot may be arranged on the casing of the terminal device, for example, may be arranged around the screen. The light point can also be set on the protective cover of the terminal device. When the terminal device is used, the protective cover containing the light point can be put on, so that the terminal device can be tracked. For example, different light spots can be set around the screen of the terminal device, such as different numbers of light spots or different color spots around the terminal device, so that the terminal device can determine The relative spatial position of the head-mounted display device.
步骤420:根据第一相对空间位置,确定虚拟内容的显示位置。Step 420: Determine the display position of the virtual content according to the first relative spatial position.
终端设备可将该第一相对空间位置从现实空间转换为虚拟空间中的位置坐标,并根据需要显示的虚拟内容与终端设备在虚拟空间的位置关系及该位置坐标,以头戴显示设备作为参照,计算虚拟空间中需要显示的虚拟内容相对头戴显示设备的空间位置,即得到虚拟内容在虚拟空间中的显示位置。该显示位置指的是虚拟内容在虚拟空间中以头戴显示设备为原点(也可看作是以人眼为原点)的三维空间坐标。The terminal device can convert the first relative spatial position from the real space to the position coordinates in the virtual space, and according to the positional relationship between the virtual content to be displayed and the terminal device in the virtual space and the position coordinates, refer to the head-mounted display device as a reference Calculate the spatial position of the virtual content to be displayed in the virtual space relative to the head-mounted display device, that is, obtain the display position of the virtual content in the virtual space. The display position refers to the three-dimensional space coordinates of the virtual content in the virtual space with the head-mounted display device as the origin (which can also be regarded as the origin of the human eye).
步骤430:根据显示位置渲染虚拟内容,并获取虚拟内容的显示数据。Step 430: Render the virtual content according to the display position and obtain the display data of the virtual content.
终端设备可获取待显示的虚拟内容的数据,根据虚拟内容的数据构建虚拟内容,并根据显示位置渲染虚拟内容。其中,待显示的虚拟内容的数据,可包括虚拟内容的模型数据,用于构建虚拟内容三维模型。例如,模型数据可包括模型的颜色、顶点坐标、轮廓等数据。The terminal device may acquire the data of the virtual content to be displayed, construct the virtual content according to the data of the virtual content, and render the virtual content according to the display position. The data of the virtual content to be displayed may include model data of the virtual content, which is used to construct a three-dimensional model of the virtual content. For example, the model data may include data such as the color, vertex coordinates, and outline of the model.
在一些实施方式中,虚拟内容的数据也可为终端设备从服务器处下载得到,还可为终端设备从其他终端处获取得到。In some embodiments, the data of the virtual content may also be downloaded from the server by the terminal device, or may be obtained from other terminals for the terminal device.
步骤440:将显示数据传输至头戴显示设备,以指示头戴显示设备显示虚拟内容。Step 440: Transmit the display data to the head-mounted display device to instruct the head-mounted display device to display virtual content.
终端设备可获取渲染后的虚拟内容的显示数据,该显示数据可包括显示画面中各个像素点的RGB值及对应的像素点坐标等。终端设备可将显示数据传输至头戴显示设备,头戴显示设备可根据该显示数据生成显示画面,并将显示画面投射到光学镜片上,显示虚拟内容。用户通过头戴显示设备的光学镜片,可看到虚拟内容叠加显示在真实世界中,实现增强现实的效果。The terminal device may obtain display data of the rendered virtual content, and the display data may include RGB values of respective pixels in the display screen and corresponding pixel coordinates, and the like. The terminal device can transmit the display data to the head-mounted display device, and the head-mounted display device can generate a display screen according to the display data, and project the display screen onto the optical lens to display the virtual content. By wearing the optical lens of the display device, the user can see that the virtual content is superimposed and displayed in the real world, realizing the effect of augmented reality.
终端设备获取虚拟内容的数据进行渲染,该虚拟内容的数据可为终端设备当前屏幕上显示的内容数据,也可为与当前屏幕上显示的内容对应的扩展内容数据,还可为与当前屏幕上的内容对应的完整内容数据。完整内容可包括当前屏幕上显示的内容及与显示的内容对应的扩展内容,当前屏幕上显示的内容可为完整内容中的一部分,该扩展内容为完整内容中除显示的内容外的部分。The terminal device obtains the data of the virtual content for rendering. The data of the virtual content may be the content data displayed on the current screen of the terminal device, or it may be the extended content data corresponding to the content displayed on the current screen, or it may be the data on the current screen. The complete content data corresponding to the content of. The complete content may include the content displayed on the current screen and the extended content corresponding to the displayed content. The content displayed on the current screen may be a part of the complete content, and the extended content is a part of the complete content except the displayed content.
如图5,终端设备100当前显示的内容为部分中国地图,终端设备100传输该部分中国地图的数据,头戴显示设备200显示的虚拟内容300a为该部分中国地图。如图6,终端设备100传输完整的中国地图的数据,头戴显示设备200显示的虚拟内容300b为完整的中国地图。如图7,终端设备100当前显示的内容为游戏地图,终端设备100向头戴显示设备传输游戏地图对应的扩展内容(游戏人物)的数据,用户可通过头戴显示设备看到的游戏人物300c叠加在现实空间,解决显示内容被移动终端的屏幕所限制的问题。As shown in FIG. 5, the content currently displayed by the terminal device 100 is a part of the China map, the terminal device 100 transmits the data of the part of the China map, and the virtual content 300 a displayed by the head-mounted display device 200 is the part of the China map. As shown in FIG. 6, the terminal device 100 transmits data of a complete China map, and the virtual content 300b displayed by the head-mounted display device 200 is a complete China map. As shown in FIG. 7, the content currently displayed by the terminal device 100 is a game map, and the terminal device 100 transmits the data of the extended content (game character) corresponding to the game map to the head-mounted display device, and the game character 300c that the user can see through the head-mounted display device Superimposed on the real space to solve the problem that the display content is limited by the screen of the mobile terminal.
在一些实施例中,终端设备的屏幕上可显示不同的按钮等界面元素,用户可通过选择点击不同的按钮进入不同的模式,例如,进入或退出增强现实模式;在增强现实模式中,又可选择不同的显示模式,例如,显示与终端设备的屏幕相同的内容、显示与终端设备的屏幕对应的扩展内容,或显示与终端设备的屏幕内容无关的预设内容等。终端设备可将选择的显示模式对应的显示数据传输至头戴显示设备,以使头戴显示设备根据该显示数据显示虚拟内容。In some embodiments, different buttons and other interface elements can be displayed on the screen of the terminal device, and the user can enter different modes by selecting and clicking different buttons, for example, entering or exiting the augmented reality mode; in the augmented reality mode, Select different display modes, for example, display the same content as the screen of the terminal device, display extended content corresponding to the screen of the terminal device, or display preset content irrelevant to the screen content of the terminal device, etc. The terminal device may transmit the display data corresponding to the selected display mode to the head-mounted display device, so that the head-mounted display device displays virtual content according to the display data.
在一个实施例中,终端设备可仅作为头戴显示设备的处理器和存储器,确定头戴显示设备显示的虚拟内容,该虚拟内容与终端设备当前屏幕显示的内容无关。虚拟内容的数据可根据现实环境实时生成,也可预先存储,头戴显示设备显示虚拟内容时,终端设备可不显示任何内容。In one embodiment, the terminal device may only serve as the processor and memory of the head-mounted display device to determine the virtual content displayed by the head-mounted display device, and the virtual content is not related to the content currently displayed on the screen of the terminal device. The data of the virtual content may be generated in real time according to the actual environment, or may be stored in advance. When the virtual content is displayed on the head-mounted display device, the terminal device may not display any content.
请参图8,本申请一实施例提供的一种虚拟内容交互方法,应用于终端设备,终端设备与外接式的头戴显示设备通信连接,包括如下步骤。Referring to FIG. 8, a virtual content interaction method provided by an embodiment of the present application is applied to a terminal device, and the terminal device is in communication connection with an external head-mounted display device, including the following steps.
步骤810:获取头戴显示设备与终端设备的第一相对空间位置。Step 810: Acquire a first relative spatial position of the head-mounted display device and the terminal device.
在一些实施方式中,终端设备上设置有标记物,标记物可设置在终端设备的外壳上,也可采用图像的形式显示在终端设备的屏幕上,标记物还可为外接式标记物,在使用时可通过USB或耳机孔等插入终端设备,以实现对终端设备的追踪。In some embodiments, the terminal device is provided with markers. The markers may be provided on the casing of the terminal device, or may be displayed on the screen of the terminal device in the form of an image. The marker may also be an external marker. When in use, the terminal device can be inserted through the USB or the earphone hole, etc., so as to track the terminal device.
步骤820:根据第一相对空间位置,确定虚拟内容的显示位置。Step 820: Determine the display position of the virtual content according to the first relative spatial position.
在一些实施方式中,如图9,上述根据第一相对空间位置,确定虚拟内容的显示位置,包括步骤822、824。In some embodiments, as shown in FIG. 9, determining the display position of the virtual content according to the first relative spatial position includes steps 822 and 824.
步骤822:获取虚拟内容与终端设备的第二相对空间位置。Step 822: Acquire the second relative spatial position of the virtual content and the terminal device.
第二相对空间位置可包括虚拟空间中虚拟内容相对终端设备的位置及姿态信息,也可理解为虚拟内容在现实空间的叠加位置与终端设备之间的相对空间位置关系。The second relative spatial position may include the position and posture information of the virtual content in the virtual space relative to the terminal device, and may also be understood as the relative spatial position relationship between the superimposed position of the virtual content in the real space and the terminal device.
虚拟内容在现实空间中的叠加位置,可重叠于终端设备,如在终端设备的屏幕区域上,如图12,用户通过头戴显示设备可看到虚拟医学人体模型重叠于终端设备的屏幕区域上;虚拟内容也可处于终端设备的边缘或周围,如图6,中国地图300b叠加在终端设备100的右上方;虚拟内容还可与终端设备处于不同平面,如终端设备所在的平面与虚拟内容所在平面垂直,虚拟内容的叠加位置不作限定。The superimposed position of the virtual content in the real space can be superimposed on the terminal device, such as on the screen area of the terminal device, as shown in FIG. 12, the user can see that the virtual medical mannequin is superimposed on the screen area of the terminal device through the head-mounted display device ; Virtual content can also be located on the edge or around the terminal device, as shown in Figure 6, China map 300b is superimposed on the upper right of the terminal device 100; the virtual content can also be in a different plane from the terminal device, such as the plane where the terminal device is located and the virtual content The plane is vertical, and the superimposition position of the virtual content is not limited.
在一些实施方式中,第二相对空间位置可预先存储于终端设备,也可根据终端设备当前屏幕显示的内容设定。终端设备当前屏幕显示的内容不同,第二相对空间位置可不同,如图6,终端设备100当前屏幕显示部分中国地图,包含完整的中国地图的虚拟内容300b在现实空间的叠加位置处于终端设备的右上方;如图10,终端设备100当前屏幕显示一张照片,包含多张照片的虚拟内容300d在现实空间的叠加位置处于终端设备100的正上方。In some embodiments, the second relative spatial position may be stored in the terminal device in advance, or may be set according to the content displayed on the current screen of the terminal device. The content displayed on the current screen of the terminal device is different, and the second relative spatial position may be different. As shown in FIG. 6, the current screen of the terminal device 100 displays a part of the China map, and the virtual content 300b including the complete China map is in the superimposed position of the terminal device in the real space. Top right; as shown in FIG. 10, the current screen of the terminal device 100 displays a photo, and the virtual content 300d containing multiple photos is located directly above the terminal device 100 in the superimposed position of the real space.
在一些实施方式中,第二相对空间位置还可根据终端设备当前屏幕显示的内容与虚拟内容的包含关系设定。当虚拟内容包括终端设备屏幕显示的内容时,可根据该屏幕显示的内容在虚拟内容中的位置,确定虚拟内容与终端设备的第二相对空间位置,以使虚拟内容中包含的该显示的内容与终端设备屏幕上显示的内容重叠。In some embodiments, the second relative spatial position may also be set according to the inclusion relationship between the content displayed on the current screen of the terminal device and the virtual content. When the virtual content includes the content displayed on the screen of the terminal device, the second relative spatial position of the virtual content and the terminal device may be determined according to the position of the content displayed on the screen in the virtual content, so that the displayed content included in the virtual content It overlaps with the content displayed on the terminal device screen.
如图11,终端设备100当前显示的内容为河北省周边地图110,虚拟内容可为中国地图300b,可根据河北省周边地图110在中国地图300b中的位置,设定中国地图300b与手机终端的第二相对空间位置,使用户可通过头戴显示设备看到中国地图300b中的河北省周边地图重叠于终端设备当前显示的河北省周边地图110。As shown in FIG. 11, the content currently displayed by the terminal device 100 is the surrounding map 110 of Hebei Province, and the virtual content may be the map 300b of China. According to the location of the surrounding map 110 of Hebei Province in the map 300b of China, the map 300b of China and the mobile terminal can be set The second relative spatial position enables the user to see the surrounding map of Hebei Province in the China map 300b through the head-mounted display device, which overlaps with the surrounding map 110 of Hebei Province currently displayed on the terminal device.
步骤824:根据第一相对空间位置以及第二相对空间位置,确定虚拟内容的显示位置。Step 824: Determine the display position of the virtual content according to the first relative spatial position and the second relative spatial position.
终端设备可根据第一相对空间位置确定终端设备在虚拟空间中的空间坐标,并根据该空间坐标及第二相对空间位置,以头戴显示设备作为参照,确定虚拟内容相对头戴显示设备的空间位置,得到虚拟内容的显示位置。The terminal device may determine the spatial coordinates of the terminal device in the virtual space according to the first relative spatial position, and based on the spatial coordinates and the second relative spatial position, use the head-mounted display device as a reference to determine the space of the virtual content relative to the head-mounted display device Position to get the display position of the virtual content.
在一些实施方式中,终端设备可判断虚拟内容的显示位置与终端设备是否存在重叠。虚拟内容的显示位置可为虚拟内容在现实空间中的叠加位置,也可为虚拟内容在虚拟空间中以头戴显示设备为原点的空间位置。当虚拟内容的显示位置与终端设备存在重叠时,可确定终端设备当前屏幕显示的内容在虚拟内容中对应的显示区域,显示区域即为虚拟内容的显示位置与终端设备之间的重叠区域。显示区域的内容与终端设备当前屏幕显示的内容相同。In some embodiments, the terminal device may determine whether the display position of the virtual content overlaps with the terminal device. The display position of the virtual content may be a superimposition position of the virtual content in the real space, or a space position of the virtual content in the virtual space with the head-mounted display device as the origin. When the display position of the virtual content overlaps with the terminal device, the corresponding display area of the content displayed on the current screen of the terminal device in the virtual content can be determined, and the display area is the overlapping area between the display position of the virtual content and the terminal device. The content of the display area is the same as the content displayed on the current screen of the terminal device.
在一些实施方式中,终端设备可对显示区域进行指定处理,可将显示区域的内容颜色调整为指定颜色,也可将显示区域的内容透明度调整为指定透明度。其中,该指定颜色的各颜色分量的亮度值低于第一阈值,该指定透明度低于第二阈值。In some embodiments, the terminal device may perform a specified process on the display area, may adjust the content color of the display area to the specified color, and may also adjust the content transparency of the display area to the specified transparency. Wherein, the brightness value of each color component of the specified color is lower than the first threshold, and the specified transparency is lower than the second threshold.
第一阈值为虚拟内容无法正常叠加显示时的各颜色分量的最大亮度值。例如,第一阈值可设定为13亮度,即95%黑,也可设定为0亮度,即黑色。第二阈值为虚拟内容无法被正常观察到时的透明度最大值。例如,第二阈值可设定为1,即90%透明,也可设定为0,即100%透明。对第一显示区域进行指定显示处理后,用户无法观察到第一显示区域中的虚拟内容,使用户可看到终端设备当前显示的内容与头戴显示设备显示的虚拟内容准确地衔接在一起。The first threshold is the maximum brightness value of each color component when the virtual content cannot be normally superimposed and displayed. For example, the first threshold may be set to 13 brightness, which is 95% black, or may be set to 0 brightness, which is black. The second threshold is the maximum transparency when the virtual content cannot be observed normally. For example, the second threshold may be set to 1, which is 90% transparent, or may be set to 0, which is 100% transparent. After performing the designated display processing on the first display area, the user cannot observe the virtual content in the first display area, so that the user can see that the content currently displayed on the terminal device and the virtual content displayed on the head-mounted display device are accurately connected together.
终端设备当前屏幕显示的内容与虚拟内容的显示位置存在对应关系,屏幕显示的内容不同,对应虚拟内容不同的显示位置。该对应关系可存储于终端设备中。当终端设备的屏幕显示的内容发生变化时,可根据变化后的内容,更新虚拟内容的显示位置。There is a corresponding relationship between the content displayed on the current screen of the terminal device and the display position of the virtual content, the content displayed on the screen is different, and the display position corresponding to the virtual content is different. The corresponding relationship can be stored in the terminal device. When the content displayed on the screen of the terminal device changes, the display position of the virtual content can be updated according to the changed content.
作为一种实施方式,终端设备可接收用户在屏幕上的滑动操作,并根据滑动操作改变 屏幕显示的内容。终端设备可重新确定屏幕上显示的变化后的内容在虚拟内容中对应的显示区域。并根据重新确定的显示区域,计算虚拟内容更新后的显示位置,并基于更新后的显示位置重新渲染虚拟内容,使用户可观察到虚拟内容的位置随终端设备显示的内容的变动而变动。如图13A和图13B,用户在终端设备100的屏幕上向右滑动,终端设备100显示的内容从河北周边地图110变为宁夏周边地图120,用户通过头戴显示设备200可看到中国地图300的显示位置发生了移动,且中国地图300与手机终端显示的内容准确衔接在一起。As an embodiment, the terminal device may receive the user's sliding operation on the screen, and change the content displayed on the screen according to the sliding operation. The terminal device may re-determine the corresponding display area of the changed content displayed on the screen in the virtual content. Based on the newly determined display area, the updated display position of the virtual content is calculated, and the virtual content is re-rendered based on the updated display position, so that the user can observe that the position of the virtual content changes as the content displayed on the terminal device changes. As shown in FIGS. 13A and 13B, the user slides to the right on the screen of the terminal device 100, and the content displayed by the terminal device 100 changes from the surrounding map 110 of Hebei to the surrounding map 120 of Ningxia. The user can see the China map 300 through the head-mounted display device 200 Has moved, and the map 300 of China is accurately connected with the content displayed on the mobile terminal.
步骤830,根据显示位置渲染虚拟内容,并获取虚拟内容的显示数据。Step 830: Render the virtual content according to the display position and obtain the display data of the virtual content.
步骤840:将显示数据传输至头戴显示设备,显示数据用于指示头戴显示设备显示虚拟内容。Step 840: Transmit the display data to the head-mounted display device, and the display data is used to instruct the head-mounted display device to display the virtual content.
步骤850:在接收到控制操作时,根据控制操作生成控制指令。Step 850: When receiving the control operation, generate a control instruction according to the control operation.
终端设备可接收用户的控制操作,根据控制操作生成控制指令,控制头戴显示设备显示的虚拟内容。作为一种实施方式,终端设备包括操控区,可接收用户在操控区的控制操作。操控区可以包括触摸屏、按键中的至少一种,其中,用户的控制操作包括但不限于作用于终端设备的操控区的单指滑动、点击、按压、多指配合滑动等。根据控制操作生成的控制指令可包括移动指令、放大指令、缩小指令、旋转指令、选取指令等,以实现控制虚拟内容的移动、缩放、旋转、选中的显示效果。The terminal device may receive the user's control operation, generate a control instruction according to the control operation, and control the virtual content displayed by the head-mounted display device. As an implementation manner, the terminal device includes a manipulation area, which can receive a user's control operation in the manipulation area. The manipulation area may include at least one of a touch screen and keys, where the user's control operations include, but are not limited to, single-finger sliding, clicking, pressing, multi-finger coordinated sliding, etc., acting on the manipulation area of the terminal device. The control instructions generated according to the control operation may include a movement instruction, a zoom-in instruction, a zoom-out instruction, a rotation instruction, a selection instruction, etc., so as to realize the display effect of controlling the movement, zooming, rotation, and selection of virtual content.
在一些实施例中,终端设备可根据检测到的控制操作执行时的手指数量、控制操作执行时的手势动作、控制操作执行时的手指滑动轨迹中的一种或多种,生成控制指令。作为一种实施方式,可实时检测在操控区执行控制操作时的手指数量,并根据不同的手指数量,生成不同的控制指令。如图14A,在检测到用户在终端设备的操控区进行单指滑动的控制操作时,生成移动虚拟内容的控制指令。如图13A-13B,该控制指令为控制头戴显示设备将当前显示的虚拟地图相对于用户的视角向右移动。如图14B,在检测到用户在终端设备的操控区进行双指的距离相对收缩合并的控制操作时,生成缩小虚拟内容的控制指令。如图6和图16,该控制指令为控制头戴显示设备将当前显示的虚拟地图相对于用户的视角缩小。In some embodiments, the terminal device may generate a control instruction according to one or more of the detected number of fingers when the control operation is performed, the gesture action when the control operation is performed, and the finger sliding trajectory when the control operation is performed. As an embodiment, the number of fingers when performing control operations in the manipulation area can be detected in real time, and different control instructions can be generated according to different numbers of fingers. As shown in FIG. 14A, when it is detected that the user performs a single-finger slide control operation in the manipulation area of the terminal device, a control instruction for moving virtual content is generated. As shown in FIGS. 13A-13B, the control instruction is to control the head-mounted display device to move the currently displayed virtual map to the right relative to the user’s perspective. As shown in FIG. 14B, when it is detected that the user performs a control operation of relatively narrowing and merging the distance between two fingers in the manipulation area of the terminal device, a control instruction for reducing the virtual content is generated. As shown in FIGS. 6 and 16, the control instruction is to control the head-mounted display device to reduce the currently displayed virtual map relative to the user's viewing angle.
作为一种实施方式,可实时检测控制操作的手指滑动轨迹,并根据不同的手指滑动轨迹,生成不同的控制指令。如图15,用户通过手指在终端设备100的触控屏幕区域进行向右滑动操作时,生成向右旋转虚拟内容300(虚拟医学人体)的控制指令。As an embodiment, the finger sliding trajectory of the control operation may be detected in real time, and different control instructions may be generated according to different finger sliding trajectories. As shown in FIG. 15, when the user performs a sliding operation on the touch screen area of the terminal device 100 with a finger to the right, a control instruction to rotate the virtual content 300 (virtual medical human body) to the right is generated.
在一些实施例中,控制操作还可为用户的手势控制。头戴显示设备可通过相机采集到用户的手势图像,并识别手势图像中的手势,当识别到的手势为预设手势时,生成控制指令。预设手势可为上升、落下、左右挥手等手势。In some embodiments, the control operation may also be a gesture control of the user. The head-mounted display device can collect the user's gesture image through the camera and recognize the gesture in the gesture image. When the recognized gesture is a preset gesture, a control instruction is generated. The preset gestures can be ascending, descending, waving left and right gestures.
在一些实施例中,终端设备还可与控制器连接,可接收用户在控制器的操控区的控制操作。用户的控制操作包括但不限于作用于控制器的操控区的单指滑动、点击、按压、多指配合滑动等。在一些实施方式中,对于不同的虚拟内容,同一个控制操作可对应不同的控制指令。终端设备可根据虚拟内容及接收的控制操作,生成与该控制操作对应的控制指令。In some embodiments, the terminal device may also be connected to the controller, and may receive user control operations in the control area of the controller. The user's control operations include, but are not limited to, single-finger slide, click, press, multi-finger cooperative slide acting on the control area of the controller. In some embodiments, for different virtual content, the same control operation may correspond to different control instructions. The terminal device may generate a control instruction corresponding to the control operation according to the virtual content and the received control operation.
步骤860:根据控制指令,控制显示的虚拟内容。Step 860: Control the displayed virtual content according to the control instruction.
在一些实施方式中,终端设备可根据生成的控制指令,调整虚拟内容的显示状态,并将调整后的虚拟内容传输至头戴显示设备,以使头戴显示设备显示调整后的虚拟内容。用户通过操作终端设备控制虚拟内容,在终端设备上进行操作的同时,也能观看到虚拟内容的显示效果,实现了终端设备与头戴显示设备的交互。In some embodiments, the terminal device may adjust the display state of the virtual content according to the generated control instruction, and transmit the adjusted virtual content to the head-mounted display device, so that the head-mounted display device displays the adjusted virtual content. The user controls the virtual content by operating the terminal device, and can also view the display effect of the virtual content while operating on the terminal device, thereby realizing the interaction between the terminal device and the head-mounted display device.
在一些实施方式中,若虚拟内容与终端设备的当前显示的显示内容对应,还可根据控制指令,控制终端设备屏幕显示的内容。如图17A,终端设备显示的内容为北京市、山西省等城市的地图,头戴显示设备显示的虚拟内容为中国地图,若控制指令为放大显示指令时,如图17B,用户通过头戴显示设备看到的放大后的中国地图,同时,终端设备显示的 北京市、山西省等城市的地图放大显示至北京市的地图。In some embodiments, if the virtual content corresponds to the currently displayed display content of the terminal device, the content displayed on the screen of the terminal device may also be controlled according to the control instruction. As shown in Figure 17A, the content displayed by the terminal device is a map of Beijing, Shanxi Province and other cities. The virtual content displayed by the head-mounted display device is a map of China. If the control command is an enlarged display command, as shown in Figure 17B, the user displays through the head-mounted display The enlarged map of China seen by the device, and at the same time, the maps of Beijing, Shanxi Province and other cities displayed on the terminal device are enlarged and displayed to the map of Beijing.
在一些实施方式中,控制虚拟内容可包括选取指定虚拟内容、移动虚拟内容、放大虚拟内容等。如图18,用户在虚拟内容300上选中的区域为北京市地图时,可看到终端设备100上同步显示北京市地图。In some embodiments, controlling virtual content may include selecting specified virtual content, moving virtual content, zooming in on virtual content, and so on. As shown in FIG. 18, when the area selected by the user on the virtual content 300 is a map of Beijing, the user can see that the map of Beijing is displayed on the terminal device 100 simultaneously.
在一些实施例中,终端设备除了作为头戴显示设备的处理器及存储器外,还可向头戴显示设备提供电源,保持头戴显示设备的轻便,并降低头戴显示设备的制作成本。In some embodiments, in addition to being a processor and a memory of the head-mounted display device, the terminal device can also provide power to the head-mounted display device, keep the head-mounted display device light, and reduce the manufacturing cost of the head-mounted display device.
在一些实施例中,上述实施例提供的虚拟内容交互方法也可在头戴显示设备中独立实施,头戴显示设备中可具有处理器及存储器,以生成并显示虚拟内容。In some embodiments, the virtual content interaction method provided in the above embodiments may also be independently implemented in a head-mounted display device, and the head-mounted display device may have a processor and a memory to generate and display virtual content.
上述实施例提供的虚拟内容交互方法,终端设备当前显示的内容与头戴显示设备显示的虚拟内容准确地衔接在一起,提高显示效果;且可控制显示的虚拟内容,提高了终端设备与头戴显示设备之间的交互性。In the virtual content interaction method provided in the above embodiment, the content currently displayed by the terminal device and the virtual content displayed by the head-mounted display device are accurately connected to improve the display effect; and the displayed virtual content can be controlled to improve the terminal device and the head-mounted device. Show interactivity between devices.
请参图19,本申请实施例提供了一种虚拟内容交互方法,应用于头戴显示设备,头戴显示设备与终端设备通信连接,包括如下步骤。Referring to FIG. 19, an embodiment of the present application provides a virtual content interaction method, which is applied to a head-mounted display device. The head-mounted display device and a terminal device communicate and connect, including the following steps.
步骤1910:采集包含有标记物的图像,标记物为设置于终端设备上。Step 1910: Acquire an image containing a marker, which is set on the terminal device.
步骤1920:将采集的图像传输至终端设备,该图像用于指示终端设备对图像中标记物进行识别,并获取头戴显示设备与终端设备的第一相对空间位置。Step 1920: Transmit the collected image to the terminal device. The image is used to instruct the terminal device to identify the marker in the image, and obtain the first relative spatial position of the head-mounted display device and the terminal device.
步骤1930:接收终端设备发送的显示数据,并根据显示数据显示虚拟内容,显示数据由终端设备根据第一相对空间位置渲染得到。Step 1930: Receive the display data sent by the terminal device, and display the virtual content according to the display data. The display data is rendered by the terminal device according to the first relative spatial position.
请参图20,本申请实施例提供的一种增强现实系统,包括终端设备11及与终端设备11通信连接的头戴显示设备12。终端设备11用于获取头戴显示设备12与终端设备11的第一相对空间位置信息,根据该第一相对空间位置信息确定虚拟内容的显示位置,根据该显示位置渲染虚拟内容,并获取虚拟内容的显示数据,及将显示数据传输至头戴显示设备12,显示数据用于指示头戴显示设备12显示虚拟内容。头戴显示设备12用于接收终端设备11发送的显示数据,并根据显示数据显示虚拟内容。Referring to FIG. 20, an augmented reality system provided by an embodiment of the present application includes a terminal device 11 and a head-mounted display device 12 communicatively connected to the terminal device 11. The terminal device 11 is used to obtain first relative spatial position information of the head-mounted display device 12 and the terminal device 11, determine the display position of the virtual content according to the first relative spatial position information, render the virtual content according to the display position, and obtain the virtual content The display data is transmitted to the head-mounted display device 12, and the display data is used to instruct the head-mounted display device 12 to display virtual content. The head-mounted display device 12 is used to receive the display data sent by the terminal device 11 and display the virtual content according to the display data.
请参图21,本申请实施例提供了一种图像处理方法,可应用于终端设备,终端设备与头戴显示设备通信连接,该方法包括如下步骤。Referring to FIG. 21, an embodiment of the present application provides an image processing method, which can be applied to a terminal device. The terminal device and a head-mounted display device are communicatively connected. The method includes the following steps.
步骤2110:获取终端设备与头戴显示设备之间的相对空间位置关系。Step 2110: Obtain the relative spatial position relationship between the terminal device and the head-mounted display device.
步骤2120:根据相对空间位置关系,获取头戴显示设备显示的虚拟内容在终端设备的屏幕上的投影区域,投影区域为人眼通过头戴显示设备观察到的虚拟内容在屏幕上的投射区域。Step 2120: Acquire the projection area of the virtual content displayed on the head-mounted display device on the screen of the terminal device according to the relative spatial position relationship, and the projection area is the projection area of the virtual content observed on the screen by the human eye through the head-mounted display device.
投影区域可理解为,以观察虚拟内容的人眼为基准点,将虚拟内容的形状投射到终端设备的屏幕所在的平面上时,在该平面上得到的投射区域与终端设备的屏幕区域之间的重合区域。在一些实施方式中,投影区域可用终端设备的屏幕坐标系的坐标进行表示,该屏幕坐标系以该屏幕的一个角,例如左下角,为原点。The projection area can be understood as, when the shape of the virtual content is projected on the plane where the screen of the terminal device is based on the human eye observing the virtual content as a reference point, between the projection area obtained on the plane and the screen area of the terminal device Coincident area. In some embodiments, the projection area may be represented by the coordinates of the screen coordinate system of the terminal device, and the screen coordinate system uses one corner of the screen, for example, the lower left corner, as the origin.
在一些实施方式中,可根据虚拟内容和终端设备在虚拟空间中的空间位置,获取虚拟内容在终端设备的屏幕上的投影区域。其中,虚拟空间中可包括虚拟相机,该虚拟相机用于模拟用户的人眼,虚拟相机在虚拟空间中的位置可看作头戴显示设备在虚拟空间中的位置。In some embodiments, the projection area of the virtual content on the screen of the terminal device may be obtained according to the virtual content and the spatial position of the terminal device in the virtual space. The virtual space may include a virtual camera, which is used to simulate the human eye of the user, and the position of the virtual camera in the virtual space may be regarded as the position of the head-mounted display device in the virtual space.
在一些实施例中,虚拟内容和终端设备在虚拟空间中的空间位置,可为虚拟内容和终端设备在虚拟空间的预设坐标系中的位置,该预设坐标系可包括世界坐标系、以虚拟相机为原点的空间坐标系等。终端设备可根据虚拟相机、虚拟内容和终端设备在预设坐标系中的空间位置,获取虚拟相机与虚拟内容中各个顶点的连线的延长线,并分别计算各条延长线与虚拟空间中终端设备的屏幕的交点,确定虚拟内容的各个顶点在终端设备的屏幕上的投影点,并得到各个投影点在该预设坐标系中的空间坐标。可根据虚拟空间的预设坐标系与真实环境中终端设备的屏幕坐标系的转化参数,将投影点的空间坐标转化为终端设备的屏幕坐标系的二维坐标,得到虚拟内容在终端设备的屏幕上的投影区域。In some embodiments, the spatial position of the virtual content and the terminal device in the virtual space may be the position of the virtual content and the terminal device in a preset coordinate system of the virtual space, and the preset coordinate system may include a world coordinate system, The spatial coordinate system with the virtual camera as the origin, etc. The terminal device can obtain the extension line of the connection between the virtual camera and each vertex in the virtual content according to the virtual camera, the virtual content, and the spatial position of the terminal device in the preset coordinate system, and calculate each extension line and the terminal in the virtual space separately The intersection point of the screen of the device determines the projection point of each vertex of the virtual content on the screen of the terminal device, and obtains the spatial coordinates of each projection point in the preset coordinate system. According to the conversion parameters of the preset coordinate system of the virtual space and the screen coordinate system of the terminal device in the real environment, the spatial coordinates of the projection point can be converted into the two-dimensional coordinates of the screen coordinate system of the terminal device to obtain the virtual content on the screen of the terminal device On the projection area.
如图22,终端设备220为平板电脑,以人眼221为原点(也可看作是以头戴显示设备为原点),人眼221与虚拟内容223的连线的延长线与平板电脑的屏幕的交点在屏幕上形成的区域为投影区域225。As shown in FIG. 22, the terminal device 220 is a tablet computer, with the human eye 221 as the origin (also can be regarded as the origin of the head-mounted display device), the extension line of the human eye 221 and the virtual content 223, and the tablet screen The area formed by the intersection point on the screen is the projection area 225.
具体的投影区域的获取方式在本申请中可不限定,例如,可仅根据虚拟内容的轮廓顶点和终端设备在虚拟空间中的空间位置,获取虚拟内容在终端设备的屏幕上的投影区域的轮廓,以得到投影区域。The specific method for obtaining the projection area may not be limited in this application. For example, the outline of the projection area of the virtual content on the screen of the terminal device may be obtained based only on the outline vertex of the virtual content and the spatial position of the terminal device in the virtual space. To get the projection area.
在一些实施例中,头戴显示设备可将虚拟内容的显示数据传输至终端设备,该显示数据可包括虚拟内容在虚拟空间中的空间位置,终端设备可根据虚拟内容的显示数据,通过上述方式获取到投影区域。在一些实施例中,终端设备也可从头戴显示设备处获取投影区域,头戴显示设备可通过上述方式得到投影区域,再将该投影区域的数据传输至终端设备。In some embodiments, the head-mounted display device may transmit the display data of the virtual content to the terminal device, the display data may include the spatial location of the virtual content in the virtual space, and the terminal device may use the above-mentioned methods according to the display data of the virtual content Get the projection area. In some embodiments, the terminal device may also obtain the projection area from the head-mounted display device. The head-mounted display device may obtain the projection area in the above manner, and then transmit the data of the projection area to the terminal device.
步骤2130:根据屏幕待显示的屏幕内容,获取屏幕内容中与投影区域对应的图像内容。Step 2130: Acquire image content corresponding to the projection area in the screen content according to the screen content to be displayed on the screen.
如图22,图像227为屏幕内容中与投影区域对应的图像内容。终端设备可获取投影区域在终端设备屏幕上的二维坐标,可根据屏幕坐标系与屏幕内容的图像坐标系的转换参数,将该二维坐标转换为屏幕内容的图像坐标系的坐标,得到投影区域在屏幕内容中对应的图像区域,并获取处于该图像区域的图像内容。As shown in FIG. 22, the image 227 is the image content corresponding to the projection area in the screen content. The terminal device can obtain the two-dimensional coordinates of the projection area on the screen of the terminal device, and can convert the two-dimensional coordinates to the coordinates of the image coordinate system of the screen content according to the conversion parameters of the screen coordinate system and the image coordinate system of the screen content to obtain the projection The area corresponds to the image area in the screen content, and the image content in the image area is acquired.
步骤2140:对图像内容进行指定处理,并显示包含指定处理后的图像内容的屏幕内容,指定处理后的图像内容的第一色调与虚拟内容的第二色调之间的色调差值大于第一阈值。Step 2140: Perform specified processing on the image content and display the screen content containing the specified processed image content. The difference between the first hue of the specified processed image content and the second hue of the virtual content is greater than the first threshold .
终端设备可对屏幕内容中与投影区域对应的图像内容进行指定处理,并显示处理后的屏幕内容,处理后的屏幕内容可减少对头戴显示设备显示的虚拟内容的干扰。The terminal device can specify the image content corresponding to the projection area in the screen content and display the processed screen content, which can reduce the interference with the virtual content displayed by the head-mounted display device.
在一些实施例中,指定处理后的图像内容的第一色调与虚拟内容的第二色调之间的色调差值大于第一阈值,该第一阈值可根据用户的观察舒适度进行合理设定。指定处理后的图像内容的第一色调与叠加显示的虚拟内容的第二色调之间具有反差,虚拟内容叠加显示在指定处理后的图像内容上可具有可读性和识别性,该处理后的图像内容可突出虚拟内容,减少终端设备显示的内容对虚拟内容的干扰。In some embodiments, the difference between the first color tone of the designated processed image content and the second color tone of the virtual content is greater than a first threshold, which can be set reasonably according to the user's viewing comfort. There is a contrast between the first hue of the designated image content and the second hue of the virtual content superimposed. The virtual content overlay display can be readable and recognizable on the designated image content after processing. The image content can highlight the virtual content and reduce the interference of the content displayed by the terminal device with the virtual content.
色调可指的是图像的明暗程度。在一些实施例中,色调包括颜色和透明度,指定处理可为设整图像内容的颜色,也可为设整图像内容的透明度。例如,将图像内容的颜色调为纯色(如黑色、灰色等)、将图像内容的透明度调整为50%透明、在图像内容上覆盖纯色图片等。Hue can refer to the lightness or darkness of an image. In some embodiments, the hue includes color and transparency, and the specified processing may be to set the color of the image content or to set the transparency of the image content. For example, adjust the color of the image content to a solid color (such as black, gray, etc.), adjust the transparency of the image content to 50% transparency, and overlay a solid color picture on the image content.
如图23,终端设备230与头戴显示设备建立无线通信连接,头戴显示设备识别终端设备230上的标记物231并显示虚拟动物233,用户通过佩戴的头戴显示设备可看到虚拟动物233叠加显示于现实空间的终端设备230上,且终端设备230显示的屏幕内容中与虚拟内容的投影区域对应的图像内容235(石头、草地等)的颜色被调整为灰色,可突出虚拟动物233,减少图像内容235对虚拟动物的干扰。As shown in FIG. 23, the terminal device 230 establishes a wireless communication connection with the head-mounted display device, the head-mounted display device recognizes the marker 231 on the terminal device 230 and displays the virtual animal 233, and the user can see the virtual animal 233 through the worn head-mounted display device Superimposed and displayed on the terminal device 230 in the real space, and the color of the image content 235 (stone, grass, etc.) corresponding to the projection area of the virtual content in the screen content displayed on the terminal device 230 is adjusted to gray, which can highlight the virtual animal 233, Reduce the interference of image content 235 to virtual animals.
上述实施例的图像处理方法,在头戴显示设备显示虚拟内容时,减少终端设备显示的屏幕内容对虚拟内容的干扰,突出头戴显示设备显示的虚拟内容,提高增强现实中虚拟内容的真实感及显示效果;且仅对虚拟内容在屏幕上的投影区域对应的屏幕内容进行处理,在突出显示虚拟内容的同时,保证屏幕其他内容的正常显示。The image processing method of the above embodiment reduces the interference of the screen content displayed by the terminal device on the virtual content when the virtual content is displayed by the head-mounted display device, highlights the virtual content displayed by the head-mounted display device, and improves the realism of the virtual content in augmented reality And display effects; and only process the screen content corresponding to the projection area of the virtual content on the screen, while highlighting the virtual content, to ensure the normal display of other content on the screen.
请参图24,本申请另一实施例提供一种图像处理方法,应用于终端设备,终端设备与头戴显示设备通信连接,包括如下步骤。Referring to FIG. 24, another embodiment of the present application provides an image processing method, which is applied to a terminal device. The terminal device and a head-mounted display device are communicatively connected, including the following steps.
步骤2410:获取终端设备与头戴显示设备之间的相对空间位置关系。Step 2410: Obtain the relative spatial position relationship between the terminal device and the head-mounted display device.
步骤2420:根据相对空间位置关系,获取头戴显示设备显示的虚拟内容在终端设备的屏幕上的投影区域,投影区域为人眼通过头戴显示设备观察到的虚拟内容在屏幕上的投射区域。Step 2420: Obtain a projection area of the virtual content displayed on the head-mounted display device on the screen of the terminal device according to the relative spatial position relationship, and the projection area is a projection area of the virtual content observed on the screen by the human eye through the head-mounted display device.
在一个实施例中,请参图25,上述根据相对空间位置关系,获取头戴显示设备显示的虚拟内容在终端设备的屏幕上的投影区域,可包括步骤2422~2428。In one embodiment, referring to FIG. 25, the above-mentioned obtaining the projection area of the virtual content displayed on the head-mounted display device on the screen of the terminal device according to the relative spatial position relationship may include steps 2422-2428.
步骤2422:获取头戴显示设备中用于显示虚拟内容的左眼图像及右眼图像。Step 2422: Obtain a left-eye image and a right-eye image for displaying virtual content in the head-mounted display device.
虚拟内容的显示数据可包括左眼图像及右眼图像,以用于通过光学元件的反射在人眼形成三维的虚拟内容。左眼图像与右眼图像具有视差,左眼图像用于在显示时经过光学镜片投射到用户的左眼,右眼图像用于在显示时经过光学镜片投射到用户的右眼,具有视差左眼图像及右眼图像经过用户大脑的融合后可形成立体图像,使用户看到立体图像的显示效果。The display data of the virtual content may include a left-eye image and a right-eye image for forming three-dimensional virtual content in the human eye through the reflection of the optical element. The left-eye image and the right-eye image have parallax. The left-eye image is used to project to the user's left eye through the optical lens during display, and the right-eye image is used to project to the user's right eye through the optical lens during display. The fusion of the image and the right-eye image through the user's brain can form a stereoscopic image, so that the user can see the display effect of the stereoscopic image.
头戴显示设备可获取虚拟内容的渲染坐标,并基于该渲染坐标渲染虚拟内容。该渲染坐标可为虚拟内容的各个点在虚拟空间中以虚拟相机为原点的空间坐标。虚拟相机包括左虚拟相机及右虚拟相机。其中,左虚拟相机用于模拟人眼左眼,右虚拟相机用于模拟人眼右眼。可获取虚拟内容在以左虚拟相机为原点的第一空间坐标系中的左渲染坐标及在以右虚拟相机为原点的第二空间坐标系中的右渲染坐标。可根据左渲染坐标渲染虚拟内容,得到虚拟内容的左眼图像,根据右渲染坐标渲染虚拟内容,得到虚拟内容的右眼图像。头戴显示设备可将虚拟内容的左眼图像以及右眼图像传输至终端设备。The head-mounted display device can obtain rendering coordinates of the virtual content, and render the virtual content based on the rendering coordinates. The rendering coordinates may be the spatial coordinates of each point of the virtual content in the virtual space with the virtual camera as the origin. The virtual camera includes a left virtual camera and a right virtual camera. The left virtual camera is used to simulate the left eye of the human eye, and the right virtual camera is used to simulate the right eye of the human eye. The left rendering coordinates of the virtual content in the first spatial coordinate system with the left virtual camera as the origin and the right rendering coordinates in the second spatial coordinate system with the right virtual camera as the origin can be obtained. The virtual content can be rendered according to the left rendering coordinates to obtain the left-eye image of the virtual content, and the virtual content can be rendered according to the right rendering coordinates to obtain the right-eye image of the virtual content. The head-mounted display device can transmit the left-eye image and the right-eye image of the virtual content to the terminal device.
步骤2424:根据左眼图像及相对空间位置关系,获取左眼图像在终端设备的屏幕上的第一投影区域。Step 2424: Acquire the first projection area of the left-eye image on the screen of the terminal device according to the relationship between the left-eye image and the relative spatial position.
终端设备在可根据左眼图像及与头戴显示设备之间相对空间位置关系,获取左眼图像在屏幕上的第一投影区域。其中,第一投影区域可为将左眼图像中的虚拟内容的形状投射到终端设备的屏幕所在的平面,在该平面上的投射区域与终端设备的屏幕区域之间的重合区域。该投射区域可部分与终端设备的屏幕区域重合,也可全部包含于屏幕区域中。如图26,终端设备260为平板电脑,虚拟内容261的左眼显示图像在经过头戴显示设备的光学镜片反射后,可入射至用户左眼262,左眼显示图像在平板电脑的屏幕上对应的第一投影区域264。The terminal device may acquire the first projection area of the left-eye image on the screen according to the relative spatial position relationship between the left-eye image and the head-mounted display device. The first projection area may be a plane on which the shape of the virtual content in the left-eye image is projected onto the screen of the terminal device, and the projection area on the plane coincides with the screen area of the terminal device. The projection area may partially overlap the screen area of the terminal device, or may be entirely contained in the screen area. As shown in FIG. 26, the terminal device 260 is a tablet computer, and the left-eye display image of the virtual content 261 can be incident on the left eye 262 of the user after being reflected by the optical lens of the head-mounted display device, and the left-eye display image corresponds to the screen of the tablet computer 'S first projection area 264.
在一些实施方式中,终端设备可根据左虚拟相机、左眼图像以及终端设备在虚拟空间中同一个空间坐标系(如世界坐标系)的空间位置,获取左虚拟相机与左眼图像中虚拟内容的各个顶点的连线的延长线与终端设备的屏幕的交点,该交点即为顶点在屏幕上的投影点。可根据虚拟空间的该空间坐标系与终端设备的屏幕坐标系的转化参数,将左眼图像在虚拟空间的屏幕上的投影点的空间坐标转化为终端设备的屏幕坐标系的二维坐标,得到左眼图像在屏幕上的第一投影区域。In some embodiments, the terminal device may acquire the virtual content in the left virtual camera and the left-eye image according to the spatial position of the left virtual camera, the left-eye image, and the terminal device in the same spatial coordinate system (such as the world coordinate system) in the virtual space. The intersection of the extension line of each vertex and the screen of the terminal device, the intersection is the projection point of the vertex on the screen. The space coordinates of the projection point of the left-eye image on the screen of the virtual space can be converted into two-dimensional coordinates of the screen coordinate system of the terminal device according to the conversion parameters of the space coordinate system of the virtual space and the screen coordinate system of the terminal device The first projection area of the left-eye image on the screen.
进一步地,可仅根据左眼图像中虚拟内容的轮廓区域在虚拟空间的空间位置,获取左眼图像在终端设备的屏幕上的第一投影区域,以简化计算步骤,优化终端设备的处理过程。Further, the first projection area of the left-eye image on the screen of the terminal device may be acquired based on the spatial position of the outline area of the virtual content in the left-eye image on the virtual space to simplify the calculation steps and optimize the processing process of the terminal device.
在一些实施例中,请参图27,上述根据左眼图像以及相对空间位置关系,获取左眼图像在终端设备的屏幕上的第一投影区域,可包括步骤2702~2708。In some embodiments, please refer to FIG. 27, the acquiring the first projection area of the left-eye image on the screen of the terminal device according to the left-eye image and the relative spatial position relationship may include steps 2702-2708.
步骤2702:获取左眼图像中虚拟内容在虚拟空间的第一轮廓坐标集合。Step 2702: Acquire the first outline coordinate set of virtual content in the virtual space in the left-eye image.
在一些实施例中,终端设备可获取左眼图像的虚拟内容在虚拟空间的第一轮廓坐标集合,以根据第一轮廓坐标集合获取第一投影区域。其中,第一轮廓坐标集合可为左眼图像中虚拟内容的轮廓区域的各个顶点在空间坐标系中的坐标。可直接将虚拟内容的轮廓区域在第一空间坐标系中的空间坐标作为第一轮廓坐标集合,也可根据第一空间坐标系与世界坐标系的转换参数,将左眼显示图像中虚拟内容的轮廓区域在第一空间坐标系中的空间坐标转化为在世界坐标系中的空间坐标,得到第一轮廓坐标集合。In some embodiments, the terminal device may acquire the first outline coordinate set of the virtual content of the left-eye image in the virtual space to obtain the first projection area according to the first outline coordinate set. The first set of outline coordinates may be the coordinates of each vertex of the outline area of the virtual content in the left-eye image in the space coordinate system. The space coordinates of the outline area of the virtual content in the first space coordinate system can be directly used as the first outline coordinate set, or the conversion parameters of the first space coordinate system and the world coordinate system can be used to display the content of the virtual content in the left-eye display image. The spatial coordinates of the contour area in the first spatial coordinate system are converted into the spatial coordinates in the world coordinate system to obtain the first contour coordinate set.
步骤2704:根据相对空间位置关系获取屏幕在虚拟空间的屏幕坐标集合。Step 2704: Acquire the screen coordinate set of the screen in the virtual space according to the relative spatial position relationship.
终端设备可根据相对位置关系获取终端设备的屏幕在现实空间中的空间位置,将该空间位置转换为虚拟空间中的空间坐标,得到屏幕在虚拟空间的屏幕坐标集合。The terminal device may acquire the spatial position of the screen of the terminal device in the real space according to the relative position relationship, and convert the spatial position into spatial coordinates in the virtual space to obtain the screen coordinate set of the screen in the virtual space.
步骤2706:分别建立虚拟空间中的左虚拟相机与第一轮廓坐标集合中各个点的第一连线,并获取屏幕坐标集合中处于建立的各条第一连线上的点的坐标,得到第一坐标集合。Step 2706: Establish a first connection between the left virtual camera in the virtual space and each point in the first outline coordinate set, and obtain the coordinates of the point on each first connection in the screen coordinate set to obtain the first A set of coordinates.
终端设备可分别建立虚拟空间中的左虚拟相机与第一轮廓坐标集合中各个点的第一连线,并获取屏幕坐标集合中处于建立的各条第一连线上的点的坐标,得到第一交点坐标集合。其中,左虚拟相机的坐标、第一轮廓坐标集合、屏幕坐标集合及第一交点坐标集合为 同一空间坐标系下的坐标。The terminal device may establish a first connection between the left virtual camera in the virtual space and each point in the first outline coordinate set, and obtain the coordinates of the point on each first connection established in the screen coordinate set to obtain the first A set of intersection coordinates. Among them, the coordinates of the left virtual camera, the first outline coordinate set, the screen coordinate set, and the first intersection coordinate set are coordinates under the same spatial coordinate system.
终端设备可判断屏幕坐标集合中,是否存在处于第一连线上的点的坐标。如果存在处于第一连线上的点,可确定该第一连线与屏幕交叉,反之,则确定该第一连线不与屏幕交叉。可根据屏幕坐标集合中处于第一连线上的点的坐标,得到第一交点坐标集合。第一交点坐标集合包括左眼图像中虚拟内容的轮廓区域的各个顶点在屏幕上的投影点的坐标。The terminal device may determine whether the coordinates of the point on the first connection line exist in the screen coordinate set. If there is a point on the first line, it can be determined that the first line crosses the screen, otherwise, it is determined that the first line does not cross the screen. The first intersection coordinate set can be obtained according to the coordinates of the point on the first line in the screen coordinate set. The first intersection coordinate set includes coordinates of projection points of each vertex of the outline area of the virtual content in the left-eye image on the screen.
步骤2708:根据第一交点坐标集合,获取左眼图像在屏幕上的第一投影区域。Step 2708: Acquire the first projection area of the left-eye image on the screen according to the first intersection coordinate set.
终端设备可根据虚拟空间中的空间坐标系与终端设备的屏幕坐标系的转化参数,将第一交点坐标集合转化为终端设备的屏幕坐标系的二维坐标,得到第一投影区域。The terminal device may convert the first intersection coordinate set into two-dimensional coordinates of the terminal device's screen coordinate system according to the conversion parameters of the space coordinate system in the virtual space and the screen coordinate system of the terminal device to obtain the first projection area.
步骤2426,根据右眼图像及相对空间位置关系,获取右眼显示图像在屏幕上的第二投影区域。Step 2426: Acquire the second projection area of the right-eye display image on the screen according to the relationship between the right-eye image and the relative spatial position.
终端设备可参照上述根据左眼图像以及相对空间位置关系,获取左眼图像在终端设备的屏幕上的第一投影区域的对应步骤,获取右眼图像在屏幕上的第二投影区域。如图26,虚拟内容261的右眼图像在经过头戴显示设备的光学镜片反射后,可入射至用户右眼263,右眼图像在平板电脑的屏幕上对应的第二投影区域265。The terminal device may refer to the above corresponding step of acquiring the first projection area of the left-eye image on the screen of the terminal device according to the left-eye image and the relative spatial position relationship, and acquire the second projection area of the right-eye image on the screen. As shown in FIG. 26, the right-eye image of the virtual content 261 can be incident on the right eye 263 of the user after being reflected by the optical lens of the head-mounted display device, and the right-eye image corresponds to the second projection area 265 on the screen of the tablet computer.
在一个实施例中,终端设备可获取右眼图像中虚拟内容在虚拟空间的第二轮廓坐标集合,根据相对空间位置关系获取屏幕在虚拟空间的屏幕坐标集合,分别建立虚拟空间中的右虚拟相机与第二轮廓坐标集合中各个点的第二连线,并获取屏幕坐标集合中处于建立的各条第二连线上的点的坐标,得到第二交点坐标集合,再根据第二坐标集合,获取右眼图像在屏幕上的第二投影区域。In one embodiment, the terminal device may acquire the second outline coordinate set of the virtual content in the right-eye image in the virtual space, obtain the screen coordinate set of the screen in the virtual space according to the relative spatial position relationship, and respectively establish a right virtual camera in the virtual space The second connection line with each point in the second outline coordinate set, and obtain the coordinates of the points on each established second connection line in the screen coordinate set to obtain the second intersection coordinate set, and then according to the second coordinate set, Acquire the second projection area of the right-eye image on the screen.
步骤2428:获取第一投影区域以及第二投影区域的合成区域,并将合成区域作为虚拟内容在屏幕上的投影区域。Step 2428: Acquire the synthesis area of the first projection area and the second projection area, and use the synthesis area as the projection area of the virtual content on the screen.
第一投影区域及第二投影区域的合成区域指的是将第一投影区域对应的第一交点坐标集合与第二投影区域对应的第二交点坐标集合合并后得到的坐标集合所形成的区域,即第一投影区域及第二投影区域的并集区域。在一些实施例中,上述获取虚拟内容在屏幕上的投影区域的处理过程也可在头戴显示设备中进行,再将该投影区域的数据传输至终端设备,以减少终端设备的计算量,优化终端设备的处理过程。The combined area of the first projection area and the second projection area refers to the area formed by the coordinate set obtained by combining the first intersection coordinate set corresponding to the first projection area and the second intersection coordinate set corresponding to the second projection area, That is, the union area of the first projection area and the second projection area. In some embodiments, the above process of obtaining the projection area of the virtual content on the screen may also be performed on the head-mounted display device, and then transmit the data of the projection area to the terminal device to reduce the calculation amount of the terminal device and optimize The processing of terminal equipment.
步骤2430:根据屏幕待显示的屏幕内容,获取屏幕内容中与投影区域对应的图像内容。Step 2430: Acquire image content corresponding to the projection area in the screen content according to the screen content to be displayed on the screen.
步骤2440:对图像内容进行指定处理,并显示包含指定处理后的图像内容的屏幕内容,指定处理后的图像内容的第一色调与虚拟内容的第二色调之间的色调差值大于第一阈值。Step 2440: Perform specified processing on the image content and display the screen content containing the specified processed image content. The difference between the first hue of the specified processed image content and the second hue of the virtual content is greater than the first threshold .
在一些实施例中,上述对图像内容进行指定处理,包括以下任意一种或多种:将指定颜色的覆盖内容覆盖于图像内容上;将图像内容的颜色调整为指定颜色;将图像内容的透明度值调整为指定透明度值。其中,指定颜色与虚拟内容的颜色之间的色差大于第二阈值,指定透明度值与虚拟内容的透明度值之间的差值大于第三阈值。In some embodiments, the above-mentioned process of specifying the image content includes any one or more of the following: overlaying the content of the specified color on the image content; adjusting the color of the image content to the specified color; and transparency of the image content The value is adjusted to the specified transparency value. The color difference between the specified color and the color of the virtual content is greater than the second threshold, and the difference between the specified transparency value and the transparency value of the virtual content is greater than the third threshold.
其中,覆盖内容可为指定颜色的图片,也可为新建的指定颜色的图层,在此不作限定,仅需覆盖图像内容,且呈现颜色为指定颜色即可。Among them, the overlay content may be a picture of a specified color, or a newly-created layer of a specified color, which is not limited herein, and only needs to cover the image content, and the presentation color may be the specified color.
指定颜色可为纯色,如黑色、灰色等,第二阈值可根据用户的观察舒适度进行合理设定,例如,当虚拟内容的颜色为蓝色时,指定颜色可为灰色。The specified color can be a solid color, such as black, gray, etc. The second threshold can be set reasonably according to the user's viewing comfort. For example, when the color of the virtual content is blue, the specified color can be gray.
指定透明度值可用于减弱图像内容的显示效果,透明度值越小,透明程度越高。如指定透明度值可设定为5,即50%透明,也可设定为0,即100%透明。第三阈值可根据用户的观察舒适度进行合理设定,在此不作限定。例如,第三阈值设定为5,当虚拟内容的透明度值为10,即0%透明时,指定透明度值可为1,即90%透明。The specified transparency value can be used to reduce the display effect of the image content. The smaller the transparency value, the higher the transparency. If the specified transparency value can be set to 5, which is 50% transparent, it can also be set to 0, which is 100% transparent. The third threshold can be set reasonably according to the user's observation comfort, which is not limited here. For example, the third threshold is set to 5, when the transparency value of the virtual content is 10, that is, 0% transparency, the specified transparency value may be 1, that is, 90% transparency.
作为一种实施方式,也可设定图像内容的透明度值小于预设阈值,例如,预设阈值可为1或2等。As an implementation manner, the transparency value of the image content may also be set to be less than a preset threshold, for example, the preset threshold may be 1 or 2, and so on.
作为一种实施方式,头戴显示设备可仅包含用于显示的显示模组、通信模块和相机,对显示的虚拟内容的控制可以是依靠终端设备的处理器、存储器等进行。As an embodiment, the head-mounted display device may only include a display module, a communication module, and a camera for display, and control of the displayed virtual content may be performed by a processor, memory, etc. of the terminal device.
在一些实施例中,终端设备包括操控区域,当操作区域检测到操控操作时,可根据操 控操作生成控制指令,根据控制指令调整头戴显示设备中显示的虚拟内容,并将调整后的虚拟内容对应的显示数据发送给头戴显示设备。终端设备可基于调整后的虚拟内容重新获取投影区域,并对屏幕内容中与重新获取的投影区域对应的图像内容进行指定处理。In some embodiments, the terminal device includes a manipulation area, and when the manipulation area detects a manipulation operation, a control instruction may be generated according to the manipulation operation, the virtual content displayed in the head-mounted display device may be adjusted according to the control instruction, and the adjusted virtual content The corresponding display data is sent to the head-mounted display device. The terminal device may reacquire the projection area based on the adjusted virtual content, and specify the image content corresponding to the newly acquired projection area in the screen content.
如图23和图28,用户通过佩戴的头戴显示设备观看到叠加于终端设备的屏幕上的虚拟动物233,可向右移动虚拟动物233,并重新确定虚拟动物233在屏幕上的投影区域,并将屏幕内容中与该投影区域对应的图像内容235的颜色调整为灰色,突出显示虚拟动物,使用户一直能够清楚地观察到虚拟动物233,减少了屏幕内容对虚拟动物的干扰。As shown in FIGS. 23 and 28, the user views the virtual animal 233 superimposed on the screen of the terminal device through the worn head-mounted display device, can move the virtual animal 233 to the right, and re-determines the projection area of the virtual animal 233 on the screen, The color of the image content 235 corresponding to the projection area in the screen content is adjusted to gray, and the virtual animal is highlighted, so that the user can clearly observe the virtual animal 233 all the time, reducing the interference of the screen content on the virtual animal.
请参图29,本申请实施例提供一种图像处理方法,应用于头戴显示设备,头戴显示设备与终端设备通信连接,该方法包括如下步骤。Referring to FIG. 29, an embodiment of the present application provides an image processing method, which is applied to a head-mounted display device, and the head-mounted display device is in communication connection with a terminal device. The method includes the following steps.
步骤2910:将虚拟内容进行显示。Step 2910: Display the virtual content.
步骤2920:获取终端设备与头戴显示设备之间的相对空间位置关系。Step 2920: Obtain the relative spatial position relationship between the terminal device and the head-mounted display device.
步骤2930:根据相对空间位置关系,获取虚拟内容在终端设备的屏幕上的投影区域,投影区域为人眼通过头戴显示设备观察到的虚拟内容在屏幕上的投射区域。Step 2930: Acquire the projection area of the virtual content on the screen of the terminal device according to the relative spatial position relationship. The projection area is the projection area of the virtual content on the screen observed by the human eye through the head-mounted display device.
步骤2940:将投影区域的数据发送至终端设备,投影区域的数据用于指示终端设备对待显示的屏幕内容中,与投影区域对应的图像内容进行指定处理,并显示包含指定处理后的图像内容的屏幕内容,指定处理后的图像内容的第一色调与虚拟内容的第二色调之间的色调差值大于第一阈值。Step 2940: Send the data of the projection area to the terminal device. The data of the projection area is used to instruct the terminal device to specify the image content corresponding to the projection area in the screen content to be displayed, and display the content containing the image content after the specified processing. The screen content specifies the difference in tone between the first hue of the processed image content and the second hue of the virtual content is greater than the first threshold.
上述实施例提供的图像处理方法,在头戴显示设备显示虚拟内容时,减少终端设备显示的屏幕内容对虚拟内容的干扰,从而突出显示虚拟内容,提高增强现实中虚拟内容的真实感及显示效果。而且,仅对虚拟内容在屏幕上的投影区域对应的屏幕内容进行处理,在突出显示虚拟内容的同时,保证其他屏幕内容的正常显示,方便用户与移动终端进行交互。The image processing method provided by the above embodiment reduces the interference of the screen content displayed by the terminal device on the virtual content when the virtual content is displayed on the head-mounted display device, thereby highlighting the virtual content and improving the realism and display effect of the virtual content in augmented reality . Moreover, only the screen content corresponding to the projection area of the virtual content on the screen is processed, while the virtual content is highlighted, the normal display of other screen content is ensured, which is convenient for the user to interact with the mobile terminal.
在一个实施例中,提供的一种终端设备,该终端设备可为智能手机、平板电脑等能够运行应用程序的电子设备。该终端设备可包括一个或多个如下部件:处理器及存储器,其中存储器存储有一个或多个应用程序,该一个或多个应用程序可被配置为由一个或多个处理器执行,一个或多个程序配置用于执行如前述方法实施例所描述的方法。In one embodiment, a terminal device is provided, and the terminal device may be an electronic device capable of running an application program, such as a smartphone or a tablet computer. The terminal device may include one or more of the following components: a processor and a memory, where the memory stores one or more application programs, the one or more application programs may be configured to be executed by one or more processors, one or Multiple programs are configured to perform the method as described in the foregoing method embodiments.
处理器可包括一个或者多个处理核。处理器利用各种接口和线路连接整个终端设备内的各个部分,通过运行或执行存储在存储器内的指令、程序、代码集或指令集,以及调用存储在存储器内的数据,执行终端设备的各种功能和处理数据。可选地,处理器可采用数字信号处理、现场可编程门阵列、可编程逻辑阵列中的至少一种硬件形式来实现。处理器可集成中央处理器、图像处理器和调制解调器等中的一种或几种的组合。其中,CPU主要处理操作系统、用户界面和应用程序等;GPU用于负责显示内容的渲染和绘制;调制解调器用于处理无线通信。可以理解的是,上述调制解调器也可以不集成到处理器120中,单独通过一块通信芯片进行实现。The processor may include one or more processing cores. The processor uses various interfaces and lines to connect the various parts of the entire terminal device, executes each instruction of the terminal device by running or executing instructions, programs, code sets or instruction sets stored in the memory, and calling data stored in the memory Kinds of functions and processing data. Alternatively, the processor may be implemented in at least one hardware form of digital signal processing, field programmable gate array, and programmable logic array. The processor may integrate one or a combination of a central processor, an image processor, and a modem. Among them, the CPU mainly handles the operating system, user interface and application programs, etc.; the GPU is used for rendering and rendering of the displayed content; and the modem is used for processing wireless communication. It can be understood that the above-mentioned modem may not be integrated into the processor 120, and may be implemented by a communication chip alone.
存储器可包括随机存储器,也可包括只读存储器。存储器可用于存储指令、程序、代码、代码集或指令集。存储器130可包括存储程序区和存储数据区,其中,存储程序区可存储用于实现操作系统的指令、用于实现至少一个功能的指令(比如触控功能、声音播放功能、图像播放功能等)、用于实现下述各个方法实施例的指令等。存储数据区还可以存储终端设备在使用中所创建的数据等。The memory may include random access memory or read-only memory. The memory can be used to store instructions, programs, codes, code sets or instruction sets. The memory 130 may include a storage program area and a storage data area, where the storage program area may store instructions for implementing an operating system and instructions for implementing at least one function (such as a touch function, a sound playback function, an image playback function, etc.) , Instructions for implementing the following method embodiments. The storage data area can also store data created by the terminal device in use, and the like.
如图30,提供一种头戴显示设备,包括相机,用于采集现实物体的图像及采集目标场景的场景图像。相机可为红外相机,也可为可见光相机,具体类型并不限定。As shown in FIG. 30, a head-mounted display device is provided, which includes a camera for collecting images of real objects and collecting scene images of a target scene. The camera may be an infrared camera or a visible light camera, and the specific type is not limited.
在一个实施例中,第一终端100还可包括如下一个或多个部件:显示模组、光学模组、通信模块以及电源。显示模组可包括显示控制单元,显示控制单元用于接收处理器渲染后的虚拟内容的显示图像,将该显示图像显示并投射至光学模组上,使用户能够通过光学模组观看到虚拟内容。其中,显示模组可为显示屏或投射装置等,用于显示图像。光学模组可采用离轴光学系统或波导光学系统,显示模组显示的显示图像经光学模组后,能够被投射至用户的眼睛。用户通过光学模组可看到显示模组投射的显示图像。在一些实施方式中, 用户还能够透过光学模组观察到现实环境,感受虚拟内容与现实环境叠加后的视觉效果。通信模块可是蓝牙、WiFi、ZigBee等模块,终端设备可通过通信模块与交互设备通信连接,以进行信息以及指令的交互。电源可为整个终端设备进行供电,保证终端设备各个部件的正常运行。In one embodiment, the first terminal 100 may further include one or more components as follows: a display module, an optical module, a communication module, and a power supply. The display module may include a display control unit for receiving a display image of the virtual content rendered by the processor, displaying and projecting the display image onto the optical module, so that the user can view the virtual content through the optical module . Among them, the display module may be a display screen or a projection device, etc., for displaying images. The optical module may use an off-axis optical system or a waveguide optical system. After the display image displayed by the display module passes through the optical module, it can be projected to the user's eyes. The user can see the display image projected by the display module through the optical module. In some embodiments, the user can also observe the real environment through the optical module and feel the visual effect of the superimposed virtual content and the real environment. The communication module may be a module such as Bluetooth, WiFi, or ZigBee. The terminal device may communicate with the interactive device through the communication module to exchange information and instructions. The power supply can supply power to the entire terminal equipment to ensure the normal operation of various components of the terminal equipment.
在一些实施方式中,头戴显示设备还可包括处理器及存储器,其中存储器存储有一个或多个应用程序,该一个或多个应用程序可被配置为由一个或多个处理器执行,一个或多个程序配置用于执行如前述方法实施例所描述的方法。In some embodiments, the head-mounted display device may further include a processor and a memory, where the memory stores one or more application programs, the one or more application programs may be configured to be executed by one or more processors, one Or more programs are configured to perform the method as described in the foregoing method embodiments.
在一个实施例中,提供一种计算机可读存储介质,该计算机可读存储介质中存储有程序代码,程序代码可被处理器调用执行上述方法实施例中所描述的方法。In one embodiment, a computer-readable storage medium is provided. The computer-readable storage medium stores program code, and the program code may be called by a processor to execute the method described in the foregoing method embodiments.
计算机可读存储介质可为诸如闪存、EEPROM、EPROM、硬盘或者ROM之类的电子存储器。可选地,计算机可读存储介质包括非易失性计算机可读介质。计算机可读存储介质800具有执行上述方法中的任何方法步骤的程序代码810的存储空间。这些程序代码可以从一个或者多个计算机程序产品中读出或者写入到这一个或者多个计算机程序产品中。程序代码810可以例如以适当形式进行压缩。The computer-readable storage medium may be an electronic memory such as flash memory, EEPROM, EPROM, hard disk, or ROM. Optionally, the computer-readable storage medium includes a non-volatile computer-readable medium. The computer-readable storage medium 800 has a storage space for the program code 810 that performs any of the method steps described above. These program codes can be read from or written into one or more computer program products. The program code 810 may be compressed in an appropriate form, for example.

Claims (22)

  1. 一种虚拟内容交互系统,其特征在于,包括终端设备及头戴显示设备,所述终端设备与所述头戴显示设备建立通信连接,其中,A virtual content interaction system is characterized by comprising a terminal device and a head-mounted display device, the terminal device establishing a communication connection with the head-mounted display device, wherein,
    所述终端设备,用于获取所述头戴显示设备与所述终端设备的第一相对空间位置信息,根据所述第一相对空间位置信息,确定虚拟内容的显示位置,根据所述显示位置渲染所述虚拟内容,并获取所述虚拟内容的显示数据,及将所述显示数据传输至所述头戴显示设备;The terminal device is configured to acquire first relative spatial position information of the head-mounted display device and the terminal device, determine the display position of the virtual content according to the first relative spatial position information, and render according to the display position The virtual content, and obtaining display data of the virtual content, and transmitting the display data to the head-mounted display device;
    所述头戴显示设备,用于接收所述终端设备发送的所述显示数据,并根据所述显示数据显示所述虚拟内容。The head-mounted display device is configured to receive the display data sent by the terminal device and display the virtual content according to the display data.
  2. 根据权利要求1所述的系统,其特征在于,所述终端设备设置有标记物;The system according to claim 1, wherein the terminal device is provided with a marker;
    所述头戴显示设备还用于通过相机采集包含所述标记物的图像,并将所述图像发送给所述终端设备;The head-mounted display device is also used to collect an image containing the marker through a camera and send the image to the terminal device;
    所述终端设备还用于接收所述头戴显示设备发送的所述图像,并识别所述图像中包含的标记物,获取所述头戴显示设备与所述终端设备之间的第一相对空间位置。The terminal device is further configured to receive the image sent by the head-mounted display device, and identify markers contained in the image, to obtain a first relative space between the head-mounted display device and the terminal device position.
  3. 根据权利要求1所述的系统,其特征在于,所述终端设备还用于接收控制操作,根据所述控制操作生成控制指令,并根据所述控制指令调整所述虚拟内容,将调整后的虚拟内容的显示数据传输至所述头戴显示设备;The system according to claim 1, wherein the terminal device is further configured to receive a control operation, generate a control instruction according to the control operation, and adjust the virtual content according to the control instruction, and adjust the adjusted virtual The display data of the content is transmitted to the head-mounted display device;
    所述头戴显示设备还用于接收所述调整后的虚拟内容的显示数据,并根据所述调整后的虚拟内容的显示数据更新显示的虚拟内容。The head-mounted display device is further configured to receive the display data of the adjusted virtual content, and update the displayed virtual content according to the adjusted display data of the virtual content.
  4. 根据权利要求3所述的系统,其特征在于,所述控制操作包括终端设备的操控区检测到的操控操作、头戴显示设备或终端设备采集到的用户手势及与所述终端设备连接的控制器接收到控制操作中的至少一种。The system according to claim 3, wherein the control operation includes a manipulation operation detected by a manipulation area of a terminal device, a user gesture collected by a head-mounted display device or a terminal device, and a control connected to the terminal device The controller receives at least one of the control operations.
  5. 根据权利要求1所述的系统,其特征在于,所述终端设备还用于获取屏幕显示的屏幕内容及与所述屏幕内容对应的扩展内容,根据所述屏幕内容及扩展内容获取虚拟内容的显示数据;The system according to claim 1, wherein the terminal device is further used to obtain screen content displayed on the screen and extended content corresponding to the screen content, and obtain virtual content display according to the screen content and the extended content data;
    所述头戴显示设备还用于根据所述显示数据显示包括所述屏幕内容及扩展内容的虚拟内容。The head-mounted display device is also used to display virtual content including the screen content and the extended content according to the display data.
  6. 根据权利要求5所述的系统,其特征在于,所述终端设备还用于获取所述虚拟内容与所述终端设备之间的第二相对空间位置,并根据所述第一相对空间位置以及所述第二相对空间位置,确定所述虚拟内容的显示位置。The system according to claim 5, wherein the terminal device is further used to obtain a second relative spatial position between the virtual content and the terminal device, and according to the first relative spatial position and The second relative spatial position determines the display position of the virtual content.
  7. 根据权利要求6所述的系统,其特征在于,所述终端设备还用于当所述虚拟内容的显示位置与所述终端设备的屏幕存在重叠时,确定所述屏幕的重叠部分的屏幕内容,并确定所述重叠部分的屏幕内容在所述虚拟内容中对应的显示区域,对所述显示区域进行预设处理,所述虚拟内容在所述头戴显示设备进行显示时,进行预设处理后的显示区域的内容处于不可见状态。The system according to claim 6, wherein the terminal device is further configured to determine the screen content of the overlapping portion of the screen when the display position of the virtual content overlaps with the screen of the terminal device, And determine the corresponding display area of the screen content of the overlapping portion in the virtual content, and perform preset processing on the display area, after the virtual content is displayed on the head-mounted display device, after the preset processing The content of the display area is invisible.
  8. 根据权利要求1所述的系统,其特征在于,所述终端设备还用于根据所述第一相对空间位置获取所述头戴显示设备显示的虚拟内容在所述终端设备的屏幕上的投影区域,所述投影区域为通过所述头戴显示设备观察到的所述虚拟内容在所述屏幕上的投射区域,获取待显示的屏幕内容中与所述投影区域对应的图像内容,并对所述图像内容进行指定处理,通过所述屏幕显示包含指定处理后的图像内容的屏幕内容,所述指定处理后的图像内容的第一色调与所述虚拟内容的第二色调之间的色调差值大于第一阈值。The system according to claim 1, wherein the terminal device is further configured to obtain a projection area of the virtual content displayed by the head-mounted display device on the screen of the terminal device according to the first relative spatial position , The projection area is the projection area of the virtual content observed on the screen through the head-mounted display device, obtains the image content corresponding to the projection area in the screen content to be displayed, and The image content is subjected to designation processing, and the screen content containing the image content after the designation processing is displayed on the screen, and the difference between the first color tone of the image content after the designation processing and the second color tone of the virtual content is greater than The first threshold.
  9. 根据权利要求8所述的系统,其特征在于,所述终端设备还用于获取所述头戴显示设备中用于显示虚拟内容的左眼图像以及右眼图像,根据所述左眼图像以及所述第 一相对空间位置,获取所述左眼图像在所述终端设备的屏幕上的第一投影区域,根据所述右眼图像以及所述第一相对空间位置,获取所述右眼图像在所述屏幕上的第二投影区域,确定所述虚拟内容在所述屏幕上的投影区域,所述投影区域为所述第一投影区域及所述第二投影区域的合成区域。The system according to claim 8, wherein the terminal device is further configured to acquire a left-eye image and a right-eye image for displaying virtual content in the head-mounted display device, and according to the left-eye image and the Acquiring the first relative spatial position, acquiring a first projection area of the left-eye image on the screen of the terminal device, and acquiring the right-eye image at all positions according to the right-eye image and the first relative spatial position A second projection area on the screen to determine a projection area of the virtual content on the screen, the projection area being a composite area of the first projection area and the second projection area.
  10. 根据权利要求8所述的系统,其特征在于,所述指定处理包括以下处理中的至少一种:The system according to claim 8, wherein the designated processing includes at least one of the following processing:
    将指定颜色的覆盖内容覆盖于所述图像内容上;Overlay the overlay content of the specified color on the image content;
    将所述图像内容的颜色调整为指定颜色;及Adjust the color of the image content to the specified color; and
    将所述图像内容的透明度值调整为指定透明度值;Adjusting the transparency value of the image content to a specified transparency value;
    其中,所述指定颜色与所述虚拟内容的颜色之间的色差大于第二阈值,所述指定透明度值与所述虚拟内容的透明度值之间的差值大于第三阈值。Wherein, the color difference between the specified color and the color of the virtual content is greater than a second threshold, and the difference between the specified transparency value and the transparency value of the virtual content is greater than a third threshold.
  11. 一种虚拟内容交互方法,其特征在于,应用于终端设备,所述终端设备与外接式的头戴显示设备通信连接,所述方法包括:A virtual content interaction method is characterized in that it is applied to a terminal device that is in communication connection with an external head-mounted display device. The method includes:
    获取所述头戴显示设备与所述终端设备的第一相对空间位置;Acquiring a first relative spatial position of the head-mounted display device and the terminal device;
    根据所述第一相对空间位置,确定虚拟内容的显示位置;Determine the display position of the virtual content according to the first relative spatial position;
    根据所述显示位置渲染所述虚拟内容,并获取所述虚拟内容的显示数据;及Rendering the virtual content according to the display position, and obtaining display data of the virtual content; and
    将所述显示数据传输至所述头戴显示设备,所述显示数据用于指示所述头戴显示设备显示所述虚拟内容。Transmitting the display data to the head-mounted display device, where the display data is used to instruct the head-mounted display device to display the virtual content.
  12. 根据权利要求11所述的方法,其特征在于,所述根据所述第一相对空间位置,确定虚拟内容的显示位置,包括:The method according to claim 11, wherein the determining the display position of the virtual content according to the first relative spatial position comprises:
    获取虚拟内容与所述终端设备的第二相对空间位置;及Acquiring the second relative spatial position of the virtual content and the terminal device; and
    根据所述第一相对空间位置以及所述第二相对空间位置,确定所述虚拟内容的显示位置。The display position of the virtual content is determined according to the first relative spatial position and the second relative spatial position.
  13. 根据权利要求11所述的方法,其特征在于,在所述将所述虚拟内容传输至所述头戴显示设备之后,所述方法还包括:The method according to claim 11, wherein after the transmitting the virtual content to the head-mounted display device, the method further comprises:
    接收控制操作,根据所述控制操作生成控制指令;及Receiving a control operation and generating a control instruction according to the control operation; and
    根据所述控制指令,控制所述虚拟内容。According to the control instruction, the virtual content is controlled.
  14. 根据权利要求13所述的方法,其特征在于,所述控制操作包括终端设备的操控区检测到的操控操作、头戴显示设备或终端设备采集到的用户手势及与所述终端设备连接的控制器接收到控制操作中的至少一种。The method according to claim 13, wherein the control operation includes a manipulation operation detected by a manipulation area of a terminal device, a user gesture collected by a head-mounted display device or a terminal device, and a control connected to the terminal device The controller receives at least one of the control operations.
  15. 根据权利要求13所述的方法,其特征在于,所述根据所述控制操作生成控制指令,包括:The method according to claim 13, wherein the generating a control instruction according to the control operation includes:
    根据检测到的所述控制操作执行时的手指数量、所述控制操作执行时的手势动作、所述控制操作执行时的手指滑动轨迹中的至少一种,生成控制指令。A control instruction is generated according to at least one of the detected number of fingers when the control operation is performed, the gesture action when the control operation is performed, and the finger sliding trajectory when the control operation is performed.
  16. 根据权利要求11所述的方法,其特征在于,所述终端设备包括标记物,所述获取所述头戴显示设备与所述终端设备之间的第一相对空间位置,包括:The method according to claim 11, wherein the terminal device includes a marker, and the acquiring the first relative spatial position between the head-mounted display device and the terminal device includes:
    接收所述头戴显示设备发送的包含标记物的图像,所述图像为所述头戴显示设备通过相机采集得到;及Receiving an image containing a marker sent by the head-mounted display device, the image being acquired by the head-mounted display device through a camera; and
    识别所述图像中包含的标记物,并获取所述头戴显示设备与所述终端设备之间的第一相对空间位置。Identify the marker contained in the image, and obtain the first relative spatial position between the head-mounted display device and the terminal device.
  17. 根据权利要求11所述的方法,其特征在于,所述方法还包括:The method according to claim 11, wherein the method further comprises:
    根据所述第一相对空间位置获取所述头戴显示设备显示的虚拟内容在所述终端设备的屏幕上的投影区域,所述投影区域为通过所述头戴显示设备观察到的所述虚拟内容在所述屏幕上的投射区域;Acquiring a projection area of the virtual content displayed by the head-mounted display device on the screen of the terminal device according to the first relative spatial position, the projection area being the virtual content observed by the head-mounted display device The projection area on the screen;
    获取待显示的屏幕内容中与所述投影区域对应的图像内容,并对所述图像内容进行指定处理;及Acquiring image content corresponding to the projection area in the screen content to be displayed, and performing specified processing on the image content; and
    通过所述屏幕显示包含指定处理后的图像内容的屏幕内容,所述指定处理后的图像内容的第一色调与所述虚拟内容的第二色调之间的色调差值大于第一阈值。The screen content containing the designated processed image content is displayed through the screen, and the difference in tone between the first tone of the specified processed image content and the second tone of the virtual content is greater than the first threshold.
  18. 一种虚拟内容交互方法,其特征在于,应用于头戴显示设备,所述头戴显示设备与终端设备建立通信连接,包括:A virtual content interaction method is characterized in that it is applied to a head-mounted display device, and the head-mounted display device establishes a communication connection with a terminal device, including:
    通过相机采集包含有标记物的图像,所述标记物为设置于所述终端设备上;Collect an image containing a marker through the camera, the marker is provided on the terminal device;
    将所述标记物图像传输至所述终端设备,所述图像用于指示所述终端设备对所述图像中包含标记物进行识别,并获取所述头戴显示设备与所述终端设备的第一相对空间位置;及Transmitting the marker image to the terminal device, where the image is used to instruct the terminal device to recognize the marker contained in the image, and obtain the first of the head-mounted display device and the terminal device Relative spatial position; and
    接收所述终端设备发送的显示数据,并根据所述显示数据显示虚拟内容,所述显示数据由所述终端设备根据所述第一相对空间位置渲染所述虚拟内容得到。Receiving display data sent by the terminal device, and displaying virtual content according to the display data, the display data being obtained by the terminal device rendering the virtual content according to the first relative spatial position.
  19. 根据权利要求18所述的方法,其特征在于,在所述通过相机采集包含有标记物的图像之后,所述方法还包括:The method according to claim 18, wherein after the image containing the marker is collected by the camera, the method further comprises:
    识别所述图像中包含的标记物,获取所述头戴显示设备与所述终端设备的第一相对空间位置;Identify the markers contained in the image, and obtain the first relative spatial position of the head-mounted display device and the terminal device;
    从所述终端设备接收所述终端设备的屏幕显示的屏幕内容,以及与所述屏幕内容对应的扩展内容;及Receiving screen content displayed on the screen of the terminal device from the terminal device, and extended content corresponding to the screen content; and
    根据所述屏幕内容及扩展内容生成虚拟内容,并根据所述第一相对空间位置渲染包含所述屏幕内容及扩展内容的虚拟内容。Generate virtual content according to the screen content and extended content, and render virtual content including the screen content and extended content according to the first relative spatial position.
  20. 根据权利要求19所述的方法,其特征在于,所述方法还包括:The method of claim 19, further comprising:
    当所述虚拟内容的显示位置与所述终端设备的屏幕存在重叠时,确定所述屏幕的重叠部分的屏幕内容;When the display position of the virtual content overlaps with the screen of the terminal device, determine the screen content of the overlapping portion of the screen;
    确定所述重叠部分的屏幕内容在所述虚拟内容中对应的显示区域;及Determine the corresponding display area of the screen content of the overlapping portion in the virtual content; and
    对所述显示区域进行预设处理,所述虚拟内容在所述头戴显示设备进行显示时,进行预设处理后的显示区域的内容处于不可见状态。Perform preset processing on the display area, and when the virtual content is displayed on the head-mounted display device, the content of the display area after the preset processing is in an invisible state.
  21. 一种虚拟内容显示方法,其特征在于,应用于第一终端,所述第一终端与第二终端通信连接,所述方法包括:A virtual content display method is characterized in that it is applied to a first terminal, and the first terminal is in communication connection with a second terminal. The method includes:
    获取所述第一终端与所述第二终端之间的第一相对空间位置信息;Acquiring first relative spatial position information between the first terminal and the second terminal;
    从所述第二终端获取显示内容数据,其中,所述显示内容数据至少包括所述第二终端当前显示的显示内容的数据;及Acquiring display content data from the second terminal, wherein the display content data includes at least data of display content currently displayed by the second terminal; and
    根据所述第一相对空间位置信息以及所述显示内容数据,对虚拟内容进行显示,其中,所述虚拟内容包括所述第二终端显示的显示内容以及与所述显示内容对应的扩展内容。The virtual content is displayed according to the first relative spatial position information and the display content data, where the virtual content includes the display content displayed by the second terminal and the extended content corresponding to the display content.
  22. 一种图像处理方法,其特征在于,应用于终端设备,所述终端设备与头戴显示设备通信连接,所述方法包括:An image processing method, characterized in that it is applied to a terminal device that is in communication connection with a head-mounted display device, the method includes:
    获取所述终端设备与所述头戴显示设备之间的相对空间位置关系;Acquiring the relative spatial position relationship between the terminal device and the head-mounted display device;
    根据所述相对空间位置关系,获取所述头戴显示设备显示的虚拟内容在所述终端设备的屏幕上的投影区域,所述投影区域为通过所述头戴显示设备观察到的所述虚拟内容在所述屏幕上的投射区域;Acquiring a projection area of the virtual content displayed on the head-mounted display device on the screen of the terminal device according to the relative spatial position relationship, the projection area being the virtual content observed through the head-mounted display device The projection area on the screen;
    根据所述屏幕待显示的屏幕内容,获取所述屏幕内容中与所述投影区域对应的图像内容;及Acquiring image content corresponding to the projection area in the screen content according to the screen content to be displayed on the screen; and
    对所述图像内容进行指定处理,并显示包含指定处理后的图像内容的屏幕内容,所述指定处理后的图像内容的第一色调与所述虚拟内容的第二色调之间的色调差值大于第一阈值。Performing a designated process on the image content, and displaying a screen content containing the designated processed image content, the difference in tone between the first hue of the designated processed image content and the second hue of the virtual content being greater than The first threshold.
PCT/CN2019/130646 2019-01-03 2019-12-31 Virtual content interaction system and method WO2020140905A1 (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
CN201910005848.6 2019-01-03
CN201910005562.8A CN111399630B (en) 2019-01-03 2019-01-03 Virtual content interaction method and device, terminal equipment and storage medium
CN201910005562.8 2019-01-03
CN201910005848.6A CN111399631B (en) 2019-01-03 2019-01-03 Virtual content display method and device, terminal equipment and storage medium
CN201910295517.0A CN111818326B (en) 2019-04-12 2019-04-12 Image processing method, device, system, terminal device and storage medium
CN201910295517.0 2019-04-12

Publications (1)

Publication Number Publication Date
WO2020140905A1 true WO2020140905A1 (en) 2020-07-09

Family

ID=71407159

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/130646 WO2020140905A1 (en) 2019-01-03 2019-12-31 Virtual content interaction system and method

Country Status (1)

Country Link
WO (1) WO2020140905A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114167986A (en) * 2021-12-03 2022-03-11 杭州灵伴科技有限公司 Head-mounted display device assembly including handle and virtual image control method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130293468A1 (en) * 2012-05-04 2013-11-07 Kathryn Stone Perez Collaboration environment using see through displays
CN107852488A (en) * 2015-05-22 2018-03-27 三星电子株式会社 System and method for showing virtual image by HMD device
CN108401463A (en) * 2017-08-11 2018-08-14 深圳前海达闼云端智能科技有限公司 Virtual display device, intelligent interaction method and cloud server

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130293468A1 (en) * 2012-05-04 2013-11-07 Kathryn Stone Perez Collaboration environment using see through displays
CN107852488A (en) * 2015-05-22 2018-03-27 三星电子株式会社 System and method for showing virtual image by HMD device
CN108401463A (en) * 2017-08-11 2018-08-14 深圳前海达闼云端智能科技有限公司 Virtual display device, intelligent interaction method and cloud server

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114167986A (en) * 2021-12-03 2022-03-11 杭州灵伴科技有限公司 Head-mounted display device assembly including handle and virtual image control method

Similar Documents

Publication Publication Date Title
US10095458B2 (en) Information processing apparatus, information processing method, non-transitory computer-readable storage medium, and system
CN110456907A (en) Control method, device, terminal device and the storage medium of virtual screen
JP4679661B1 (en) Information presenting apparatus, information presenting method, and program
JP5966510B2 (en) Information processing system
WO2016185845A1 (en) Interface control system, interface control device, interface control method and program
US9740282B1 (en) Gaze direction tracking
CN110163942B (en) Image data processing method and device
JPWO2014141504A1 (en) 3D user interface device and 3D operation processing method
EP3262505B1 (en) Interactive system control apparatus and method
WO2020078443A1 (en) Method and system for displaying virtual content based on augmented reality and terminal device
TW202025719A (en) Method, apparatus and electronic device for image processing and storage medium thereof
US20180203706A1 (en) Transitioning Between 2D and Stereoscopic 3D Webpage Presentation
WO2014128751A1 (en) Head mount display apparatus, head mount display program, and head mount display method
US10257500B2 (en) Stereoscopic 3D webpage overlay
JP2017120556A (en) Head-mounted display for operation, control method of head-mounted display for operation, and program for head-mounted display for operation
WO2022005715A1 (en) Augmented reality eyewear with 3d costumes
JP2022183213A (en) Head-mounted display
CN111813214B (en) Virtual content processing method and device, terminal equipment and storage medium
CN111818326B (en) Image processing method, device, system, terminal device and storage medium
WO2020140905A1 (en) Virtual content interaction system and method
CN111913560A (en) Virtual content display method, device, system, terminal equipment and storage medium
US20220301264A1 (en) Devices, methods, and graphical user interfaces for maps
CN111913564B (en) Virtual content control method, device, system, terminal equipment and storage medium
CN111399630B (en) Virtual content interaction method and device, terminal equipment and storage medium
CN111651031B (en) Virtual content display method and device, terminal equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19907914

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19907914

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 19907914

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 21.01.2022)

122 Ep: pct application non-entry in european phase

Ref document number: 19907914

Country of ref document: EP

Kind code of ref document: A1