WO2021203902A1 - Virtual image realization method and apparatus, and storage medium and terminal device - Google Patents

Virtual image realization method and apparatus, and storage medium and terminal device Download PDF

Info

Publication number
WO2021203902A1
WO2021203902A1 PCT/CN2021/080040 CN2021080040W WO2021203902A1 WO 2021203902 A1 WO2021203902 A1 WO 2021203902A1 CN 2021080040 W CN2021080040 W CN 2021080040W WO 2021203902 A1 WO2021203902 A1 WO 2021203902A1
Authority
WO
WIPO (PCT)
Prior art keywords
virtual image
augmented reality
reality scene
target object
terminal
Prior art date
Application number
PCT/CN2021/080040
Other languages
French (fr)
Chinese (zh)
Inventor
季春林
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Publication of WO2021203902A1 publication Critical patent/WO2021203902A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]

Definitions

  • the present disclosure relates to the technical field of virtual reality and augmented reality, and in particular to virtual image realization methods, virtual image realization devices, computer-readable storage media and terminal equipment.
  • the user can see the superimposition effect of the virtual image projected onto the real scene, thereby bringing the user an immersive visual experience.
  • the present disclosure provides a method for realizing a virtual image, a device for realizing a virtual image, a computer-readable storage medium, and terminal equipment.
  • a virtual image realization method which is applied to a first terminal, the method includes: acquiring edit information for a target object input by a user; and generating a virtual image of the target object according to the edit information. Image; upload the virtual image to the augmented reality scene where the target object is located, the augmented reality scene is stored in the cloud, so that the second terminal displays the virtual image by acquiring the augmented reality scene from the cloud .
  • a virtual image realization method applied to a second terminal comprising: acquiring an augmented reality scene from the cloud, the augmented reality scene including at least a virtual image of a target object, the The virtual image is uploaded to the augmented reality scene by the first terminal; relocation is performed based on the augmented reality scene, and when the relocation is successful, the virtual image or prompt information of the virtual image is displayed.
  • a virtual image realization device which is provided in a first terminal, the device includes a processor, and the processor is configured to execute the following program modules stored in a memory: an acquisition module for acquiring a user Input editing information for the target object; a generating module for generating a virtual image of the target object according to the editing information; an uploading module for uploading the virtual image to the augmented reality scene where the target object is located , The augmented reality scene is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image.
  • a virtual image realization device which is provided in a second terminal, the device includes a processor, and the processor is configured to execute the following program modules stored in a memory: an acquisition module for downloading from the cloud Acquire an augmented reality scene, the augmented reality scene includes at least a virtual image of a target object, the virtual image is uploaded to the augmented reality scene by the first terminal; the relocation module is configured to perform relocation based on the augmented reality scene Positioning, when repositioning is successful, display the virtual image or prompt information of the virtual image.
  • a computer-readable storage medium having a computer program stored thereon, and the computer program, when executed by a processor, implements the virtual image realization method described in the first or second aspect and the same Possible implementation.
  • a terminal device including: a processor; a memory for storing executable instructions of the processor; and a display; wherein the processor is configured to execute the executable Instructions to execute the virtual image realization method described in the first or second aspect and possible implementations thereof.
  • FIG. 1 shows a schematic diagram of the architecture of an AR system in this exemplary embodiment
  • FIG. 2 shows a schematic structural diagram of a mobile terminal in this exemplary embodiment
  • FIG. 3 shows a flowchart of a method for implementing virtual images in this exemplary embodiment
  • Fig. 4 shows a flowchart of another virtual image realization method in this exemplary embodiment
  • FIG. 5 shows a flowchart of a relocation method in this exemplary embodiment
  • Fig. 6 shows a flowchart of a method for modifying a virtual image in this exemplary embodiment
  • FIG. 7 shows a schematic structural diagram of a virtual image realization device provided in a first terminal in this exemplary embodiment
  • FIG. 8 shows a schematic structural diagram of another virtual image realization device provided in the first terminal in this exemplary embodiment
  • FIG. 9 shows a schematic structural diagram of a virtual image realization device provided in a second terminal in this exemplary embodiment
  • FIG. 10 shows a schematic structural diagram of another virtual image realization device provided in the second terminal in this exemplary embodiment.
  • the AR virtual images that the user can view are generally set in advance by the program, for example, the user walks to a certain place to trigger the appearance of a specific virtual character.
  • This kind of AR experience process is immutable, it is difficult to arouse the user's interest in multiple experiences, and its practicability is low.
  • exemplary embodiments of the present disclosure provide a virtual image realization method and a virtual image realization device.
  • FIG. 1 shows a schematic diagram of the architecture of an AR system suitable for implementing an exemplary embodiment of the present disclosure.
  • the AR system 100 may include: a first terminal 110, a second terminal 120, and a cloud 130.
  • the first terminal 110 and the second terminal 120 are used by users, and may be various electronic devices with display functions, including but not limited to smart phones, tablet computers, portable computers, smart wearable devices (such as AR glasses), and so on. Both the first terminal 110 and the second terminal 120 can run AR program clients.
  • the cloud 130 is a collection of hardware devices on the server side where AR programs are deployed, which can include a series of physical architectures such as proxy servers, application servers, databases, etc. This disclosure does not make a special distinction on this, and abstracts them as the cloud 130 to provide AR cloud. service.
  • the first terminal 110 and the second terminal 120 may be connected to the server 130 through a network to perform data exchange.
  • terminal devices in FIG. 1 is only illustrative, and any number of terminal devices can be provided according to implementation needs, for example, a third terminal, a fourth terminal, etc. can be provided.
  • the exemplary embodiment of the present disclosure also provides a terminal device, which may be the above-mentioned first terminal 110 or the second terminal 120, which may be realized in various forms, for example, may include smart Mobile phones, tablet computers, portable computers, smart wearable devices (such as AR glasses), etc.
  • a terminal device which may be the above-mentioned first terminal 110 or the second terminal 120, which may be realized in various forms, for example, may include smart Mobile phones, tablet computers, portable computers, smart wearable devices (such as AR glasses), etc.
  • the mobile terminal 200 in FIG. 2 as an example, the structure of the terminal device will be exemplarily described below.
  • the mobile terminal 200 may include more or fewer components than shown, or combine certain components, or split certain components, or arrange different components.
  • the illustrated components can be implemented in hardware, software, or a combination of software and hardware.
  • the interface connection relationship between the components is only shown schematically, and does not constitute a structural limitation of the mobile terminal 200.
  • the mobile terminal 200 may also adopt a different interface connection manner
  • the mobile terminal 200 may specifically include: a processor 210, an internal memory 221, an external memory interface 222, a Universal Serial Bus (USB) interface 230, a charging management module 240, a power management module 241, Battery 242, antenna 1, antenna 2, mobile communication module 250, wireless communication module 260, audio module 270, speaker 271, receiver 272, microphone 273, earphone interface 274, sensor module 280, display 290, camera module 291, indicator 292, a motor 293, a button 294, a Subscriber Identification Module (SIM) card interface 295, and so on.
  • the sensor module 280 may include a depth sensor 2801, a pressure sensor 2802, a gyroscope sensor 2803, an air pressure sensor 2804, and the like.
  • the processor 210 may include one or more processing units.
  • the processor 210 may include an application processor (AP), a modem processor, a graphics processing unit (GPU), and an image signal processor. (Image Signal Processor, ISP), controller, video codec, digital signal processor (Digital Signal Processor, DSP), baseband processor and/or Neural-Network Processing Unit (NPU), etc.
  • AP application processor
  • modem processor GPU
  • image signal processor ISP
  • controller Video codec
  • DSP Digital Signal Processor
  • NPU Neural-Network Processing Unit
  • the different processing units may be independent devices or integrated in one or more processors.
  • the controller can generate operation control signals according to the instruction operation code and timing signals to complete the control of reading instructions and executing instructions.
  • a memory may also be provided in the processor 210 for storing instructions and data.
  • the memory can store instructions for implementing six modular functions: detection instructions, connection instructions, information management instructions, analysis instructions, data transmission instructions, and notification instructions, and the processor 210 controls the execution.
  • the memory in the processor 210 is a cache memory.
  • the memory can store instructions or data that have just been used or recycled by the processor 210. If the processor 210 needs to use the instruction or data again, it can be directly called from the memory. Repeated accesses are avoided, the waiting time of the processor 210 is reduced, and the efficiency of the system is improved.
  • the processor 210 may include one or more interfaces.
  • Interfaces can include integrated circuit (Inter-Integrated Circuit, I2C) interface, integrated circuit built-in audio (Inter-Integrated Circuit Sound, I2S) interface, pulse code modulation (Pulse Code Modulation, PCM) interface, universal asynchronous transceiver (Universal Asynchronous) Receiver/Transmitter, UART) interface, Mobile Industry Processor Interface (MIPI), General-Purpose Input/Output (GPIO) interface, Subscriber Identity Module (SIM) interface and/ Or Universal Serial Bus (Universal Serial Bus, USB) interface, etc. Connections are formed with other components of the mobile terminal 200 through different interfaces.
  • I2C Inter-Integrated Circuit
  • I2S integrated circuit built-in audio
  • PCM pulse code modulation
  • PCM Pulse Code Modulation
  • UART Universal Asynchronous transceiver
  • MIPI Mobile Industry Processor Interface
  • GPIO General-Purpose Input/Output
  • SIM Subscriber Identity Module
  • USB Universal Serial Bus
  • the USB interface 230 is an interface that complies with the USB standard specification, and specifically may be a MiniUSB interface, a MicroUSB interface, a USBTypeC interface, and so on.
  • the USB interface 230 can be used to connect a charger to charge the mobile terminal 200, can also connect to earphones to play audio through the earphones, and can also be used to connect the mobile terminal 200 to other touch terminals, such as computers, peripherals, and the like.
  • the charging management module 240 is used to receive charging input from the charger.
  • the charger can be a wireless charger or a wired charger.
  • the charging management module 240 may receive the charging input of the wired charger through the USB interface 230.
  • the charging management module 240 may receive the wireless charging input through the wireless charging coil of the mobile terminal 200. While the charging management module 240 charges the battery 242, it can also supply power to the touch terminal through the power management module 241.
  • the power management module 241 is used to connect the battery 242, the charging management module 240, and the processor 210.
  • the power management module 241 receives the input of the battery 242 and/or the charge management module 240, and supplies power to the processor 210, internal memory 221, display 290, camera module 291, and wireless communication module 260. It can also be used to monitor battery capacity. The number of cycles, battery health status (leakage, impedance) and other parameters.
  • the wireless communication function of the mobile terminal 200 can be implemented by the antenna 1, the antenna 2, the mobile communication module 250, the wireless communication module 260, the modem processor, and the baseband processor.
  • the antenna 1 and the antenna 2 are used to transmit and receive electromagnetic wave signals.
  • Each antenna in the mobile terminal 200 can be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
  • Antenna 1 can be multiplexed as a diversity antenna of a wireless local area network.
  • the antenna can be used in combination with a tuning switch.
  • the mobile communication module 250 may provide a wireless communication solution including 2G/3G/4G/5G and the like applied to the mobile terminal 200.
  • the mobile communication module 250 may include at least one filter, a switch, a power amplifier, a low noise amplifier (LNA), and so on.
  • the mobile communication module 250 can receive electromagnetic waves by the antenna 1, and perform processing such as filtering and amplifying the received electromagnetic waves, and transmit them to the modem processor for demodulation.
  • the mobile communication module 250 can also amplify the signal modulated by the modem processor, and convert it into electromagnetic wave radiation via the antenna 1.
  • the wireless communication module 260 can provide applications on the mobile terminal 200, including wireless local area networks (Wireless Local Area Networks, WLAN) (such as Wireless Fidelity (Wi-Fi) networks), Bluetooth (Bluetooth, BT), and global navigation satellites. System (Global Navigation Satellite System, GNSS), Frequency Modulation (FM), Near Field Communication (NFC), Infrared Technology (Infrared, IR) and other wireless communication solutions.
  • the wireless communication module 260 may be one or more devices integrating at least one communication processing module.
  • the wireless communication module 260 receives electromagnetic waves via the antenna 2, frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 210.
  • the wireless communication module 260 may also receive the signal to be sent from the processor 210, perform frequency modulation, amplify, and convert it into electromagnetic waves and radiate it through the antenna 2.
  • the antenna 1 of the mobile terminal 200 is coupled with the mobile communication module 250, and the antenna 2 is coupled with the wireless communication module 260, so that the mobile terminal 200 can communicate with the network and other devices through wireless communication technology.
  • the wireless communication technology may include Global System for Mobile Communications (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (CDMA), and broadband Code Division Multiple Access (Wideband Code Division Multiple Access, WCDMA), Time Division Code Division Multiple Access (Time Division-Synchronous Code Division Multiple Access, TD-SCDMA), Long Term Evolution (LTE), New Radio (New Radio, NR), BT, GNSS, WLAN, NFC, FM, and/or IR technology, etc.
  • GSM Global System for Mobile Communications
  • GPRS General Packet Radio Service
  • CDMA Code Division Multiple Access
  • WCDMA Wideband Code Division Multiple Access
  • Time Division Code Division Multiple Access Time Division-Synchronous Code Division Multiple Access
  • TD-SCDMA Time Division Code Division Multiple Access
  • LTE Long Term Evolution
  • the GNSS may include Global Positioning System (GPS), Global Navigation Satellite System (GLONASS), Beidou Navigation Satellite System (BDS), Quasi-Zenith Satellite System (Quasi -Zenith Satellite System, QZSS) and/or Satellite Based Augmentation Systems (SBAS).
  • GPS Global Positioning System
  • GLONASS Global Navigation Satellite System
  • BDS Beidou Navigation Satellite System
  • QZSS Quasi-Zenith Satellite System
  • SBAS Satellite Based Augmentation Systems
  • the mobile terminal 200 implements a display function through a GPU, a display 290, an application processor, and the like.
  • the GPU is a microprocessor that connects the display 290 and the application processor.
  • the GPU is used to perform mathematical and geometric calculations and is used for graphics rendering.
  • the processor 210 may include one or more GPUs that execute program instructions to generate or change display information.
  • the mobile terminal 200 may include one or more displays 290 for displaying images, videos, and the like.
  • the display 290 includes a display panel.
  • the display panel can be a Liquid Crystal Display (LCD), Organic Light-Emitting Diode (OLED), Active-Matrix Organic Light-Emitting Diode, or Active-Matrix Organic Light-Emitting Diode (Active-Matrix Organic Light-Emitting Diode).
  • Emitting Diode AMOLED
  • Flexible Light Emitting Diode Flexible Light Emitting Diode (Flexlight-Emitting Diode, FLED), Miniled, MicroLed, Micro-oLed, Quantum Dot Light Emitting Diodes (QLED), etc.
  • the mobile terminal 200 can implement a shooting function through an ISP, a camera module 291, a video codec, a GPU, a display 290, and an application processor.
  • the ISP is used to process the data fed back from the camera module 291. For example, when taking a picture, the shutter is opened, the light is transmitted to the photosensitive element of the camera through the lens, the light signal is converted into an electrical signal, and the photosensitive element of the camera transfers the electrical signal to the ISP for processing and is converted into an image visible to the naked eye.
  • the ISP may be provided in the camera module 291.
  • the mobile terminal 200 may include one or more camera modules 291 for capturing still images or videos.
  • the object generates an optical image through the lens and is projected to the photosensitive element.
  • the photosensitive element converts the optical signal into an electrical signal, and then transfers the electrical signal to the ISP to convert it into a digital image signal.
  • ISP outputs digital image signals to DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
  • Video codecs are used to compress or decompress digital video.
  • the mobile terminal 200 may support one or more video codecs. In this way, the mobile terminal 200 can play or record videos in multiple encoding formats, such as: Moving Picture Experts Group (MPEG) 1, MPEG2, MPEG3, MPEG4, etc.
  • MPEG Moving Picture Experts Group
  • MPEG2 MPEG2, MPEG3, MPEG4, etc.
  • the external memory interface 222 may be used to connect an external memory card, such as a Micro SD card, so as to expand the storage capacity of the mobile terminal 200.
  • the external memory card communicates with the processor 210 through the external memory interface 222 to realize the data storage function. For example, save music, video and other files in an external memory card.
  • the internal memory 221 may be used to store computer executable program code, where the executable program code includes instructions.
  • the internal memory 221 may include a storage program area and a storage data area.
  • the storage program area can store an operating system, an application program (such as a sound playback function, an image playback function, etc.) required by at least one function, and the like.
  • the data storage area can store data (such as audio data, phone book, etc.) created during the use of the mobile terminal 200.
  • the internal memory 221 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash storage (Universal Flash Storage, UFS), and the like.
  • the processor 210 executes various functional applications and data processing of the mobile terminal 200 by running instructions stored in the internal memory 221 and/or instructions stored in a memory provided in the processor.
  • the mobile terminal 200 can implement audio functions through an audio module 270, a speaker 271, a receiver 272, a microphone 273, a headset interface 274, an application processor, and the like. For example, music playback, recording, etc.
  • the audio module 270 is used to convert digital audio information into an analog audio signal for output, and is also used to convert an analog audio input into a digital audio signal.
  • the audio module 270 can also be used to encode and decode audio signals.
  • the audio module 270 may be provided in the processor 210, or part of the functional modules of the audio module 270 may be provided in the processor 210.
  • the speaker 271 also called a “speaker” is used to convert audio electrical signals into sound signals.
  • the mobile terminal 200 can listen to music through the speaker 271, or listen to a hands-free call.
  • the receiver 272 also called “earpiece” is used to convert audio electrical signals into sound signals.
  • the mobile terminal 200 answers a call or voice message, it can receive the voice by bringing the receiver 272 close to the human ear.
  • the microphone 273, also called “microphone” or “microphone”, is used to convert sound signals into electrical signals.
  • the user can approach the microphone 273 through the mouth to make a sound, and input the sound signal to the microphone 273.
  • the mobile terminal 200 may be provided with multiple microphones 273 to implement functions such as noise reduction, sound source identification, and directional recording.
  • the earphone interface 274 is used to connect wired earphones.
  • the earphone interface 274 can be a USB interface 230, or a 3.5mm open mobile terminal platform (Open Mobile Terminal Platform, OMTP) standard interface, and the US Cellular Telecommunications Industry Association (Cellular Telecommunications Industry Association of the USA, CTIA) standard interface .
  • OMTP Open Mobile Terminal Platform
  • CTIA Cellular Telecommunications Industry Association
  • the depth sensor 2801 is used to obtain depth information of the scene.
  • the depth sensor may be provided in the camera module 291.
  • the pressure sensor 2802 is used to sense the pressure signal, and can convert the pressure signal into an electrical signal for implementing functions such as pressure touch.
  • the gyroscope sensor 2803 can be used to determine the motion posture of the mobile terminal 200, and can be used to shoot scenes such as anti-shake, navigation, and somatosensory games.
  • the air pressure sensor 2804 is used to measure air pressure, and can assist positioning and navigation by calculating the altitude.
  • sensors with other functions can be set in the sensor module 280, such as magnetic sensors, acceleration sensors, distance sensors, proximity light sensors, fingerprint sensors, temperature sensors, touch sensors, ambient light sensors, bone conduction sensors, etc. Wait.
  • the indicator 292 can be an indicator light, which can be used to indicate the charging status, power change, or to indicate messages, missed calls, notifications, and so on.
  • the motor 293 can generate vibration prompts, such as vibration prompts for incoming calls, alarm clocks, received messages, etc., and can also be used for touch vibration feedback and the like.
  • the button 294 includes a power-on button, a volume button, and so on.
  • the button 294 may be a mechanical button. It can also be a touch button.
  • the mobile terminal 200 may receive key input, and generate key signal input related to user settings and function control of the mobile terminal 200.
  • the mobile terminal 200 may support one or more SIM card interfaces 295 for connecting to the SIM card, so that the mobile terminal 200 can interact with the network through the SIM card to implement functions such as call and data communication.
  • the virtual image realization method can be realized in the form of an AR program.
  • FIG. 3 shows a schematic flow of the virtual image realization method executed by the first terminal 110.
  • the first terminal 110 may implement the following steps S310 to S330 through a client running an AR program:
  • Step S310 Obtain editing information for the target object input by the user.
  • the target object can be any object in the real world environment, such as a refrigerator, a pool, etc. in a room.
  • the user can input editing information in any form and content for the target object, which can include any one or more of text, pictures, hand-drawn drawings, animation effects, video, and audio.
  • the user can input a paragraph of text, set the corresponding font, color, etc., and edit the frame of the note paper style for it, as well as the animation effect of the text display, such as gradually appearing in a faded manner, flying in from a certain direction, etc.; Or the user can manually draw the picture and add a certain background texture to it; or the user can edit a virtual icon and add a piece of audio, the virtual icon can trigger the audio to be played, and so on.
  • the target object can be determined by scanning with a camera. Specifically, before step S310, the following steps may be performed:
  • the editing interface is triggered to be displayed.
  • the real scene refers to the real world environment where the target object is located, such as a real room, coffee shop, shopping mall, or street.
  • the user wants to edit a virtual image of an object, he can turn on the camera of the first terminal, aim at the object, and the camera scans the collected image.
  • the editing interface is triggered, or the message "Is it right?" Edit virtual image", the user selects "Yes” and the editing interface pops up.
  • the editing interface is used to input editing information, and can include editing columns for various information, such as text columns, picture columns, and animation effect columns.
  • the client can guide the user to move multiple angles, for example, the scanning interface displays "please aim at the target from another angle", “please aim at the back of the target” "And other guidance information, so as to collect the full picture of the target object.
  • the user can click to select the target object on the client.
  • the client can download the scene information from the server in advance and open the display. The user clicks on an object in the scene, determines it as the target object, and enters the editing interface.
  • step S320 a virtual image of the target object is generated according to the above-mentioned editing information.
  • a virtual image can be generated correspondingly.
  • the user's editing information is the main content of the virtual image.
  • the client can directly use the editing information as the virtual image, or based on this, add some virtual visual effects, or combine the virtual image with the target object to a certain extent .
  • the user edits a virtual note with the refrigerator in the room as the target object, including the size, shape, and color of the note, and enters text in the note.
  • the client can automatically generate a virtual note for the virtual note.
  • the user can also edit the display position of the virtual image to include it in the above-mentioned editing information. For example, the user clicks on a certain position on the target object to edit and generate a virtual note at that position, or the user edits the virtual note and moves it to a certain position, that is, the display position of the virtual image.
  • step S330 the virtual image is uploaded to the augmented reality scene where the target object is located, and the augmented reality scene is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image.
  • Augmented reality scene refers to a data collection established for the real scene where the target object is located, which can include map data of the real scene, cloud anchor point data, 3D point cloud data of each object, and added virtual image data (such as user It can be seen that all AR data related to the scene can be stored in the corresponding augmented reality scene. When new data is added or data is updated, it can be synchronized to the augmented reality scene. Thus, after the client generates a virtual image and uploads it to the cloud, it can be directly stored in the corresponding augmented reality scene.
  • the user scans the real scene for the first time, that is, the cloud does not record any AR data about the real scene.
  • the map data can be a three-dimensional point cloud model reflecting the characteristics of a real scene, including information such as terrain, texture, and obstacles.
  • the map data generated by the first terminal can adopt the SLAM (Simultaneous Localization And Mapping, simultaneous positioning and mapping) method.
  • the map data of the scene usually only includes a small part of the scene, or is different from the actual scene; after that, every time a frame of image is collected, the map data can be updated and optimized according to the image, and the map points that are not in the map can be added. Or modify the position of the existing map point, etc.; until the entire scene is traversed, the complete map data is obtained.
  • SLAM collects scene images
  • a certain number of key frame images are usually selected from continuous frame images.
  • a collection of key frame images is usually packaged into the map data for subsequent use.
  • the map data can then be uploaded to the cloud, for example, it can be uploaded together with the virtual image of the target object, or it can be uploaded before the virtual image, so that an augmented reality scene corresponding to the real scene can be established on the cloud. Therefore, it is easy to determine that the target object belongs to the established augmented reality scene.
  • user A uses his own kitchen as a real scene, scans the kitchen to create map data, scans the refrigerator in the kitchen, and edits virtual images.
  • the client uploads map data and virtual images to the cloud to create a corresponding augmented reality scene.
  • the scene can be named "User A's Kitchen” and the virtual image is added to the augmented reality scene.
  • an augmented reality scene includes key frame images of some or all objects. For example, when the kitchen is scanned for the first time, key frame images are recorded for more characteristic objects such as refrigerators and sinks in the kitchen.
  • the client can match the image of the target object with the key frame image.
  • the target object is determined to be a recorded object, and the editing interface can be triggered at this time.
  • the client does not enter a certain augmented reality scene, and the user directly scans the target object and edits related information, when uploading a virtual image, you can also select at least one key frame image from the collected images of the target object, and Upload to the cloud.
  • the cloud can match the key frame image with the key frame image in the stored augmented reality scene to determine which augmented reality scene the target object belongs to.
  • the first terminal is uploading virtual images and target objects
  • GPS Global Positioning System, Global Positioning System
  • other positioning data can also be sent, so that the cloud can narrow the matching range based on the positioning data.
  • the user Before each virtual image is generated, the user is required to scan the target object and the real scene, and send the key frame image, map data, and virtual image to the cloud.
  • the cloud finds the corresponding augmented reality scene based on the map data, and if found, it will be directly retrieved Augmented reality scene, if not found, create a new augmented reality scene.
  • the virtual image edited on the first terminal is added to the augmented reality scene
  • a second terminal obtains the information of the augmented reality scene
  • the virtual image can be obtained synchronously, so that the virtual image can be displayed on the second terminal.
  • user A edits and generates a virtual note of the refrigerator and uploads it to the cloud; after user B downloads the augmented reality scene, the virtual note can be displayed while watching the refrigerator.
  • the prompt information of the virtual image can also be generated according to the position of the target object in the augmented reality scene, and when the virtual image is uploaded, the prompt information is uploaded to the augmented reality scene together.
  • the prompt information is used to prompt the orientation of the virtual image, so that when the user is in different positions in the augmented reality scene, the virtual image can be found according to the prompt information.
  • the prompt information can be generated at the same time as the virtual image, or can be generated later than the virtual image.
  • the client first determines the position of the target object in the augmented reality scene, which is usually the position of the virtual image (most of the virtual images are displayed on the target object), and then can use this position as a reference to convert the augmented reality scene Divided into multiple areas, a guiding virtual indicator arrow is generated for each area to point to the position of the virtual image.
  • the client first determines the position of the target object in the augmented reality scene, which is usually the position of the virtual image (most of the virtual images are displayed on the target object), and then can generate a virtual icon at this position. It is used to prompt the virtual image, and the virtual icon can be displayed when the user is far away from the position of the virtual image.
  • FIG. 4 shows a schematic flow of the virtual image realization method executed by the above-mentioned second terminal 120.
  • the second terminal 120 may implement the following steps S410 and S420 through a client running an AR program:
  • Step S410 Obtain an augmented reality scene from the cloud.
  • the augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal.
  • step S420 relocation is performed based on the above-mentioned augmented reality scene, and when the relocation is successful, the above-mentioned virtual image or prompt information of the virtual image is displayed.
  • the cloud can send an update notification to the second terminal.
  • the new virtual image is obtained; or the first terminal and the second terminal are connected to the cloud at the same time, and the augmented reality is turned on
  • the cloud can be synchronized to the second terminal.
  • corresponding permissions may be set for acquiring the augmented reality scene. For example, when user A creates an augmented reality scene through the first terminal, set a password, and other users need to enter the correct password to download the augmented reality scene from the cloud; or when user A creates an augmented reality scene through the first terminal, set an access whitelist , Only users on the whitelist can download the augmented reality scene from the cloud, and so on.
  • the relocation process is generally as follows: the second terminal starts the camera, and user B holds (or wears) the second terminal and moves in the real scene; the second terminal collects images of the real scene and matches it with the map data in the augmented reality scene.
  • the matching is successful, it is determined that the relocation is successful, and the virtual image or its prompt message can be displayed at this time.
  • the virtual image is displayed or the prompt information is specifically related to the location of the user.
  • a virtual image is displayed, otherwise, a prompt message is displayed to guide the user to move to another location where the virtual image can be seen.
  • the virtual image is generally displayed on the target object. If the first terminal edits the display position of the virtual image, the second terminal can also obtain the position when acquiring the augmented reality scene and display the virtual image at that position. .
  • step S420 may include the following steps S501 to S504:
  • Step S501 Perform relocation based on the augmented reality scene, and when the relocation is successful, determine the current pose of the second terminal in the augmented reality scene;
  • Step S502 Determine the current view area of the second terminal according to the foregoing current pose
  • Step S503 when the display position of the virtual image is within the current view area, display the virtual image
  • Step S504 When the display position of the virtual image is outside the current view area, display the prompt information of the virtual image.
  • the relocation process is also the process of aligning the camera coordinate system of the second terminal with the world coordinate system of the augmented reality scene.
  • the current pose of the second terminal in the world coordinate system of the augmented reality scene can be determined;
  • the current pose and the field of view of the second terminal (for example, 120 degrees) can determine the current field of view area to simulate the range of the area that can be captured by the camera of the second terminal, or the user is in the current pose and can see It can then determine whether the user can see the virtual image theoretically, that is, when the display position of the virtual image is within the current field of view, it is judged that the user can see the virtual image, then the virtual image is displayed, and the display position of the virtual image is When the current field of view is outside, it is determined that the user cannot see the virtual image, and a prompt message is displayed.
  • the relocation may also be performed based on the target object, so that when the relocation is successful, it can be directly determined that the user can see the virtual image, and therefore the virtual image is displayed.
  • the key frame image of the target object can be obtained from the augmented reality scene, and then the collected current image can be matched with the key frame image.
  • the second terminal is just aligned with the target object, the captured current image and key frame The image can be matched successfully, and then the virtual image is displayed.
  • the prompt information of the virtual image when the prompt information of the virtual image is displayed, if the user is far away from the target object and the target object is no longer in the field of view, the user can follow the direction of the virtual image at the edge of the current interface.
  • Display prompt information such as virtual guide arrows, etc.; if the user is far away from the target object, but the target object is within the field of view, it can display the prompt information that displays the virtual image in the form of a virtual icon, such as displaying a virtual icon above the target object , And use an arrow to point to the target object to indicate that the location has a virtual image.
  • the user may be allowed to modify the virtual image, as shown in FIG. 6, which is implemented through the following steps S601 and S602:
  • step S601 the user's modification information for the virtual image is obtained, and a new virtual image is obtained according to the modification information.
  • Step S602 upload the above-mentioned new virtual image to the augmented reality scene to update the virtual image stored on the cloud.
  • the user can modify any one or more aspects of the virtual image, such as modifying text content, replacing pictures, modifying animation effects, moving positions, and so on.
  • the second terminal modifies and obtains the new virtual image, it is uploaded to the cloud, and the cloud updates the original virtual image, which can overwrite the original virtual image, and can also save the new virtual image and the original virtual image at the same time.
  • other terminals can obtain new virtual images when downloading augmented reality scenes from the cloud.
  • any user who enters the augmented reality scene can modify the virtual image in it; or the user sets a password when editing and generating the virtual image, and other users can modify it after entering the correct password, and so on.
  • the first terminal obtains the editing information for the target object input by the user, generates a virtual image of the target object according to the editing information, and uploads it to the augmented reality scene stored in the cloud;
  • the second terminal can display the virtual image or its prompt information through relocation.
  • it provides an interactive way that allows users to freely edit virtual content, and different virtual elements can be added based on the same real scene, thereby improving the diversified experience of augmented reality.
  • by editing and displaying virtual images it provides a channel for information transmission and sharing, which is highly practical and interesting.
  • the data related to the augmented reality scene is stored in the cloud, thereby saving the storage space of the terminal and improving the operation efficiency, and the data is stored and read in units of the augmented reality scene, ensuring the privacy of information.
  • Exemplary embodiments of the present disclosure also provide a virtual image realization device, which may be installed in the above-mentioned first terminal 110.
  • the virtual image realization device 700 may include a processor 710 and a memory 720.
  • the memory 720 stores the following program modules:
  • the obtaining module 721 is used to obtain edit information for the target object input by the user;
  • the generating module 722 is configured to generate a virtual image of the target object according to the above-mentioned editing information
  • the upload module 723 is configured to upload the above-mentioned virtual image to an augmented reality scene, the augmented reality scene is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image;
  • the processor 710 is configured to execute the foregoing program modules.
  • the acquisition module 721 is further configured to trigger the display of an editing interface when a target object is scanned from a real scene, and the editing interface is used to input editing information.
  • the upload module 723 is also used to select at least one key frame image from the collected images of the target object when uploading the virtual image, and upload it to the augmented reality scene.
  • the generating module 722 is further configured to generate prompt information of the virtual image according to the position of the target object in the augmented reality scene, and the prompt information is used to prompt the orientation of the virtual image.
  • the upload module 730 is also used to upload the prompt information to the augmented reality scene.
  • the generating module 722 is configured to generate a virtual indicating arrow pointing from the area to the virtual image for each area in the augmented reality scene.
  • the above-mentioned editing information may include any one or more of text, pictures, hand-drawn drawings, animation effects, video, and audio; in addition, the above-mentioned editing information may also include the display position of the virtual image. .
  • the generating module 722 is also used to scan the actual scene where the target object is located to generate map data.
  • the upload module 723 is also used to upload map data to the cloud to establish an augmented reality scene corresponding to the real scene.
  • Exemplary embodiments of the present disclosure also provide another virtual image realization device, which can be set in the above-mentioned first terminal 110.
  • the virtual image realization device 800 may include:
  • the obtaining module 810 is used to obtain edit information for the target object input by the user;
  • the generating module 820 is configured to generate a virtual image of the target object according to the above-mentioned editing information
  • the upload module 830 is configured to upload the above-mentioned virtual image to an augmented reality scene, which is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image.
  • the acquisition module 810 is further configured to trigger the display of an editing interface when a target object is scanned from a real scene, and the editing interface is used to input editing information.
  • the upload module 830 is further configured to select at least one key frame image from the collected images of the target object when uploading the virtual image, and upload it to the augmented reality scene.
  • the generating module 820 is further configured to generate prompt information of the virtual image according to the position of the target object in the augmented reality scene, and the prompt information is used to prompt the orientation of the virtual image.
  • the upload module 730 is also used to upload the prompt information to the augmented reality scene.
  • the generating module 820 is configured to generate a virtual indicating arrow pointing from the area to the virtual image for each area in the augmented reality scene.
  • the above-mentioned editing information may include any one or more of text, pictures, hand-drawn drawings, animation effects, video, and audio; in addition, the above-mentioned editing information may also include the display position of the virtual image. .
  • the generating module 820 is also used to scan the actual scene where the target object is located to generate map data.
  • the upload module 830 is also used to upload map data to the cloud to establish an augmented reality scene corresponding to the real scene.
  • Exemplary embodiments of the present disclosure also provide a virtual image realization device, which may be installed in the above-mentioned second terminal 120.
  • the virtual image realization device 900 may include a processor 910 and a memory 920.
  • the memory 920 stores the following program modules:
  • the acquiring module 921 is configured to acquire an augmented reality scene from the cloud, the augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal;
  • the relocation module 922 is used to perform relocation based on the augmented reality scene, and when the relocation is successful, display a virtual image or a prompt message of the virtual image;
  • the processor 910 is configured to execute the foregoing program modules.
  • the relocation module 922 is configured to:
  • the relocation module 922 is configured to:
  • the relocation module 922 is configured to:
  • the relocation module 922 is configured to:
  • the virtual image realization apparatus 900 may further include:
  • the modification module is used to obtain the user's modification information for the virtual image, obtain a new virtual image according to the modification information, and upload the new virtual image to the augmented reality scene to update the virtual image stored on the cloud.
  • Exemplary embodiments of the present disclosure also provide another virtual image realization device, which can be set in the above-mentioned second terminal 120.
  • the virtual image realization device 1000 may include:
  • the acquiring module 1010 is configured to acquire an augmented reality scene from the cloud, the augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal;
  • the relocation module 1020 is used to perform relocation based on the augmented reality scene, and when the relocation is successful, display a virtual image or a prompt message of the virtual image.
  • the relocation module 1020 is configured to:
  • the relocation module 1020 is configured to:
  • the relocation module 1020 is configured to:
  • the relocation module 1020 is configured to:
  • the virtual image realization apparatus 1000 may further include:
  • the modification module is used to obtain the user's modification information for the virtual image, obtain a new virtual image according to the modification information, and upload the new virtual image to the augmented reality scene to update the virtual image stored on the cloud.
  • Exemplary embodiments of the present disclosure also provide a computer-readable storage medium, which can be implemented in the form of a program product, which includes program code.
  • the program product runs on a terminal device, the program code is used to make the terminal device Perform the steps according to various exemplary embodiments of the present disclosure described in the above-mentioned "Exemplary Method" section of this specification.
  • the program product can adopt a portable compact disk read-only memory (CD-ROM) and include program code, and can run on a terminal device, such as a personal computer.
  • CD-ROM portable compact disk read-only memory
  • the program product of the present disclosure is not limited thereto.
  • the readable storage medium can be any tangible medium that contains or stores a program, and the program can be used by or in combination with an instruction execution system, device, or device.
  • the program product can adopt any combination of one or more readable media.
  • the readable medium may be a readable signal medium or a readable storage medium.
  • the readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or a combination of any of the above. More specific examples (non-exhaustive list) of readable storage media include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Type programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • the computer-readable signal medium may include a data signal propagated in baseband or as a part of a carrier wave, and readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • the readable signal medium may also be any readable medium other than a readable storage medium, and the readable medium may send, propagate, or transmit a program for use by or in combination with the instruction execution system, apparatus, or device.
  • the program code contained on the readable medium can be transmitted by any suitable medium, including but not limited to wireless, wired, optical cable, RF, etc., or any suitable combination of the foregoing.
  • the program code for performing the operations of the present disclosure can be written in any combination of one or more programming languages.
  • the programming languages include object-oriented programming languages—such as Java, C++, etc., as well as conventional procedural programming. Language-such as "C" language or similar programming language.
  • the program code can be executed entirely on the user's computing device, partly on the user's device, executed as an independent software package, partly on the user's computing device and partly executed on the remote computing device, or entirely on the remote computing device or server Executed on.
  • the remote computing device can be connected to a user computing device through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computing device (for example, using Internet service providers). Business to connect via the Internet).
  • LAN local area network
  • WAN wide area network
  • Internet service providers for example, using Internet service providers
  • modules or units of the device for action execution are mentioned in the above detailed description, this division is not mandatory.
  • the features and functions of two or more modules or units described above may be embodied in one module or unit.
  • the features and functions of a module or unit described above can be further divided into multiple modules or units to be embodied.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Architecture (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

A virtual image realization method, a virtual image realization apparatus, a storage medium and a terminal device. The virtual image realization method is applied to a first terminal and comprises: acquiring editing information input by a user for a target object (S310); generating a virtual image of the target object according to the editing information (S320); and uploading the virtual image to an augmented reality scene where the target object is located, wherein the augmented reality scene is stored in a cloud end, so that a second terminal acquires the augmented reality scene from the cloud end to display the virtual image (S330). The diversified experience of augmented reality is improved, and a channel for realizing information transfer and sharing is provided. (FIG. 3)

Description

虚拟影像实现方法、装置、存储介质与终端设备Virtual image realization method, device, storage medium and terminal equipment
本申请要求于2020年04月08日提交的,申请号为202010269991.9,名称为“虚拟影像实现方法、装置、存储介质与终端设备”的中国专利申请的优先权,该中国专利申请的全部内容通过引用结合在本文中。This application claims the priority of a Chinese patent application filed on April 8, 2020 with the application number 202010269991.9 and the title "Virtual image realization method, device, storage medium and terminal equipment". The entire content of the Chinese patent application is approved Reference is incorporated in this article.
技术领域Technical field
本公开涉及虚拟现实与增强现实技术领域,尤其涉及虚拟影像实现方法、虚拟影像实现装置、计算机可读存储介质与终端设备。The present disclosure relates to the technical field of virtual reality and augmented reality, and in particular to virtual image realization methods, virtual image realization devices, computer-readable storage media and terminal equipment.
背景技术Background technique
在AR(Augmented Reality,增强现实)场景或AR游戏中,用户能够看到虚拟影像投射至现实场景的叠加效果,由此带给用户沉浸式的视觉体验。In an AR (Augmented Reality) scene or AR game, the user can see the superimposition effect of the virtual image projected onto the real scene, thereby bringing the user an immersive visual experience.
发明内容Summary of the invention
本公开提供虚拟影像实现方法、虚拟影像实现装置、计算机可读存储介质与终端设备。The present disclosure provides a method for realizing a virtual image, a device for realizing a virtual image, a computer-readable storage medium, and terminal equipment.
根据本公开的第一方面,提供一种虚拟影像实现方法,应用于第一终端,所述方法包括:获取用户输入的针对目标物体的编辑信息;根据所述编辑信息生成所述目标物体的虚拟影像;将所述虚拟影像上传至所述目标物体所在的增强现实场景中,所述增强现实场景存储于云端,使第二终端通过从所述云端获取所述增强现实场景而显示所述虚拟影像。According to a first aspect of the present disclosure, there is provided a virtual image realization method, which is applied to a first terminal, the method includes: acquiring edit information for a target object input by a user; and generating a virtual image of the target object according to the edit information. Image; upload the virtual image to the augmented reality scene where the target object is located, the augmented reality scene is stored in the cloud, so that the second terminal displays the virtual image by acquiring the augmented reality scene from the cloud .
根据本公开的第二方面,提供一种虚拟影像实现方法,应用于第二终端,所述方法包括:从云端获取增强现实场景,所述增强现实场景至少包括一目标物体的虚拟影像,所述虚拟影像由第一终端上传至所述增强现实场景中;基于所述增强现实场景进行重定位,当重定位成功时,显示所述虚拟影像或者所述虚拟影像的提示信息。According to a second aspect of the present disclosure, there is provided a virtual image realization method applied to a second terminal, the method comprising: acquiring an augmented reality scene from the cloud, the augmented reality scene including at least a virtual image of a target object, the The virtual image is uploaded to the augmented reality scene by the first terminal; relocation is performed based on the augmented reality scene, and when the relocation is successful, the virtual image or prompt information of the virtual image is displayed.
根据本公开的第三方面,提供一种虚拟影像实现装置,设置于第一终端,所述装置包括处理器,所述处理器用于执行存储器中存储的以下程序模块:获取模块,用于获取用户输入的针对目标物体的编辑信息;生成模块,用于根据所述编辑信息生成所述目标物体的虚拟影像;上传模块,用于将所述虚拟影像上传至所述目标物体所在的增强现实场景中,所述增强现实场景存储于云端,使第二终端通过从所述云端获取所述增强现实场景而显示所述虚拟影像。According to a third aspect of the present disclosure, there is provided a virtual image realization device, which is provided in a first terminal, the device includes a processor, and the processor is configured to execute the following program modules stored in a memory: an acquisition module for acquiring a user Input editing information for the target object; a generating module for generating a virtual image of the target object according to the editing information; an uploading module for uploading the virtual image to the augmented reality scene where the target object is located , The augmented reality scene is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image.
根据本公开的第四方面,提供一种虚拟影像实现装置,设置于第二终端,所述装置包括处理器,所述处理器用于执行存储器中存储的以下程序模块:获取模块,用于从云端获取增强现实场景,所述增强现实场景至少包括一目标物体的虚拟影像,所述虚拟影像由第一终端上传至所述增强现实场景中;重定位模块,用于基于所述增强现实场景进行重定位,当重定位成功时,显示所述虚拟影像或者所述虚拟影像的提示信息。According to a fourth aspect of the present disclosure, there is provided a virtual image realization device, which is provided in a second terminal, the device includes a processor, and the processor is configured to execute the following program modules stored in a memory: an acquisition module for downloading from the cloud Acquire an augmented reality scene, the augmented reality scene includes at least a virtual image of a target object, the virtual image is uploaded to the augmented reality scene by the first terminal; the relocation module is configured to perform relocation based on the augmented reality scene Positioning, when repositioning is successful, display the virtual image or prompt information of the virtual image.
根据本公开的第五方面,提供一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现上述第一或第二方面所述的虚拟影像实现方法及其可能的实施方式。According to a fifth aspect of the present disclosure, there is provided a computer-readable storage medium having a computer program stored thereon, and the computer program, when executed by a processor, implements the virtual image realization method described in the first or second aspect and the same Possible implementation.
根据本公开的第六方面,提供一种终端设备,包括:处理器;存储器,用于存储所述处理器的可执行指令;以及显示器;其中,所述处理器配置为经由执行所述可执行指令来执行上述第一或第二方面所述的虚拟影像实现方法及其可能的实施方式。According to a sixth aspect of the present disclosure, there is provided a terminal device, including: a processor; a memory for storing executable instructions of the processor; and a display; wherein the processor is configured to execute the executable Instructions to execute the virtual image realization method described in the first or second aspect and possible implementations thereof.
附图说明Description of the drawings
图1示出本示例性实施方式中一种AR系统的架构示意图;FIG. 1 shows a schematic diagram of the architecture of an AR system in this exemplary embodiment;
图2示出本示例性实施方式中一种移动终端的结构示意图;FIG. 2 shows a schematic structural diagram of a mobile terminal in this exemplary embodiment;
图3示出本示例性实施方式中一种虚拟影像实现方法的流程图;Fig. 3 shows a flowchart of a method for implementing virtual images in this exemplary embodiment;
图4示出本示例性实施方式中另一种虚拟影像实现方法的流程图;Fig. 4 shows a flowchart of another virtual image realization method in this exemplary embodiment;
图5示出本示例性实施方式中一种重定位方法的流程图;Fig. 5 shows a flowchart of a relocation method in this exemplary embodiment;
图6示出本示例性实施方式中一种修改虚拟影像方法的流程图;Fig. 6 shows a flowchart of a method for modifying a virtual image in this exemplary embodiment;
图7示出本示例性实施方式中一种设置于第一终端的虚拟影像实现装置的结构示意图;FIG. 7 shows a schematic structural diagram of a virtual image realization device provided in a first terminal in this exemplary embodiment;
图8示出本示例性实施方式中另一种设置于第一终端的虚拟影像实现装置的结构示意图;FIG. 8 shows a schematic structural diagram of another virtual image realization device provided in the first terminal in this exemplary embodiment;
图9示出本示例性实施方式中一种设置于第二终端的虚拟影像实现装置的结构示意图;FIG. 9 shows a schematic structural diagram of a virtual image realization device provided in a second terminal in this exemplary embodiment;
图10示出本示例性实施方式中另一种设置于第二终端的虚拟影像实现装置的结构示意图。FIG. 10 shows a schematic structural diagram of another virtual image realization device provided in the second terminal in this exemplary embodiment.
具体实施方式Detailed ways
现在将参考附图更全面地描述示例实施方式。然而,示例实施方式能够以多种形式实施,且不应被理解为限于在此阐述的范例;相反,提供这些实施方式使得本公开将更加全面和完整,并将示例实施方式的构思全面地传达给本领域的技术人员。所描述的特征、结构或特性可以以任何合适的方式结合在一个或更多实施方式中。在下面的描述中,提供许多具体细节从而给出对本公开的实施方式的充分理解。然而,本领域技术人员将意识到,可以实践本公开的技术方案而省略所述特定细节中的一个或更多,或者可以采用其它的方法、组元、装置、步骤等。在其它情况下,不详细示出或描述公知技术方案以避免喧宾夺主而使得本公开的各方面变得模糊。Example embodiments will now be described more fully with reference to the accompanying drawings. However, the example embodiments can be implemented in various forms, and should not be construed as being limited to the examples set forth herein; on the contrary, these embodiments are provided so that the present disclosure will be more comprehensive and complete, and the concept of the example embodiments will be fully conveyed To those skilled in the art. The described features, structures or characteristics can be combined in one or more embodiments in any suitable way. In the following description, many specific details are provided to give a sufficient understanding of the embodiments of the present disclosure. However, those skilled in the art will realize that the technical solutions of the present disclosure can be practiced without one or more of the specific details, or other methods, components, devices, steps, etc. can be used. In other cases, the well-known technical solutions are not shown or described in detail in order to avoid overwhelming the crowd and obscure all aspects of the present disclosure.
此外,附图仅为本公开的示意性图解,并非一定是按比例绘制。图中相同的附图标记表示相同或类似的部分,因而将省略对它们的重复描述。附图中所示的一些方框图是功能实体,不一定必须与物理或逻辑上独立的实体相对应。可以采用软件形式来实现这些功能实体,或在一个或多个硬件模块或集成电路中实现这些功能实体,或在不同网络和/或处理器装置和/或微控制器装置中实现这些功能实体。In addition, the drawings are only schematic illustrations of the present disclosure, and are not necessarily drawn to scale. The same reference numerals in the figures denote the same or similar parts, and thus their repeated description will be omitted. Some of the block diagrams shown in the drawings are functional entities and do not necessarily correspond to physically or logically independent entities. These functional entities may be implemented in the form of software, or implemented in one or more hardware modules or integrated circuits, or implemented in different networks and/or processor devices and/or microcontroller devices.
相关技术中,用户所能观看到的AR虚拟影像一般是程序事先设定的,例如用户走到某个地点,触发特定的虚拟人物出现。这样的AR体验过程一成不变,难以引起用户多次体验的兴趣,且实用性较低。In related technologies, the AR virtual images that the user can view are generally set in advance by the program, for example, the user walks to a certain place to trigger the appearance of a specific virtual character. This kind of AR experience process is immutable, it is difficult to arouse the user's interest in multiple experiences, and its practicability is low.
鉴于上述问题,本公开的示例性实施方式提供虚拟影像实现方法和虚拟影像实现装置。In view of the foregoing problems, exemplary embodiments of the present disclosure provide a virtual image realization method and a virtual image realization device.
图1示出了适于实现本公开示例性实施方式的AR系统的架构示意图。FIG. 1 shows a schematic diagram of the architecture of an AR system suitable for implementing an exemplary embodiment of the present disclosure.
如图1所示,AR系统100可以包括:第一终端110、第二终端120和云端130。其中,第一终端110、第二终端120由用户使用,可以是具有显示功能的各种电子设备,包括但不限于智能手机、平板电脑、便携式计算机、智能可穿戴设备(如AR眼镜)等。第一终端110、第二终端120均可以运行AR程序的客户端。云端130是部署AR程序的服务端的硬件设备集合,可以包括代理服务器、应用服务器、数据库等一系列实体架构,本公开对此不做特别区分,将其统一抽象为云端130,用于提供AR云服务。第一终端110、第二终端120可以通过网络与服务器130连接,进行数据交互。As shown in FIG. 1, the AR system 100 may include: a first terminal 110, a second terminal 120, and a cloud 130. Among them, the first terminal 110 and the second terminal 120 are used by users, and may be various electronic devices with display functions, including but not limited to smart phones, tablet computers, portable computers, smart wearable devices (such as AR glasses), and so on. Both the first terminal 110 and the second terminal 120 can run AR program clients. The cloud 130 is a collection of hardware devices on the server side where AR programs are deployed, which can include a series of physical architectures such as proxy servers, application servers, databases, etc. This disclosure does not make a special distinction on this, and abstracts them as the cloud 130 to provide AR cloud. service. The first terminal 110 and the second terminal 120 may be connected to the server 130 through a network to perform data exchange.
应当理解的是,图1中终端设备的数量仅仅是示意性的,根据实现需要,可以具有任意数目的终端设备,比如可以设置第三终端、第四终端等等。It should be understood that the number of terminal devices in FIG. 1 is only illustrative, and any number of terminal devices can be provided according to implementation needs, for example, a third terminal, a fourth terminal, etc. can be provided.
为了实现上述虚拟影像实现方法和实现装置,本公开的示例性实施方式还提供一种终端设备,可以是上述第一终端110或第二终端120,其可以以各种形式实现,例如可以包括智能手机、平板电脑、便携式计算机、智能可穿戴设备(如AR眼镜)等。下面以图2中的移动终端200为例,对终端设备的构造进行示例性说明。在另一些实施方式中,移动终端200可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件、软件或软件和硬件的组合实现。各部件间的接口连接关系只是示意性示出,并不构成对移动终端200的结构限定。在另一些实施方式中,移动终端200也可以采用与图2不同的接口连接方式,或多种接口连接方式的组合。In order to realize the above virtual image realization method and realization device, the exemplary embodiment of the present disclosure also provides a terminal device, which may be the above-mentioned first terminal 110 or the second terminal 120, which may be realized in various forms, for example, may include smart Mobile phones, tablet computers, portable computers, smart wearable devices (such as AR glasses), etc. Taking the mobile terminal 200 in FIG. 2 as an example, the structure of the terminal device will be exemplarily described below. In other embodiments, the mobile terminal 200 may include more or fewer components than shown, or combine certain components, or split certain components, or arrange different components. The illustrated components can be implemented in hardware, software, or a combination of software and hardware. The interface connection relationship between the components is only shown schematically, and does not constitute a structural limitation of the mobile terminal 200. In other embodiments, the mobile terminal 200 may also adopt a different interface connection manner from that shown in FIG. 2 or a combination of multiple interface connection manners.
如图2所示,移动终端200具体可以包括:处理器210、内部存储器221、外部存储器接口222、通用串行总线(Universal Serial Bus,USB)接口230、充电管理模块240、电源管理模块241、电池242、天线1、天线2、移动通信模块250、无线通信模块260、音频模块270、扬声器271、受话器272、麦克风273、耳机接口274、传感器模块280、显示器290、摄像模组291、指示器292、马达293、按键294以及用户标识模块(Subscriber Identification Module,SIM)卡接口295等。其中传感器模块280可以包括深度传感器2801、压力传感器2802、陀螺仪传感器2803、气压传感器2804等。As shown in FIG. 2, the mobile terminal 200 may specifically include: a processor 210, an internal memory 221, an external memory interface 222, a Universal Serial Bus (USB) interface 230, a charging management module 240, a power management module 241, Battery 242, antenna 1, antenna 2, mobile communication module 250, wireless communication module 260, audio module 270, speaker 271, receiver 272, microphone 273, earphone interface 274, sensor module 280, display 290, camera module 291, indicator 292, a motor 293, a button 294, a Subscriber Identification Module (SIM) card interface 295, and so on. The sensor module 280 may include a depth sensor 2801, a pressure sensor 2802, a gyroscope sensor 2803, an air pressure sensor 2804, and the like.
处理器210可以包括一个或多个处理单元,例如:处理器210可以包括应用处理器(Application Processor,AP)、调制解调处理器、图形处理器(Graphics Processing Unit,GPU)、图像信号处理器(Image Signal Processor,ISP)、控制器、视频编解码器、数字信号处理器(Digital Signal Processor,DSP)、基带处理器和/或神经网络处理器(Neural-Network Processing Unit,NPU)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。The processor 210 may include one or more processing units. For example, the processor 210 may include an application processor (AP), a modem processor, a graphics processing unit (GPU), and an image signal processor. (Image Signal Processor, ISP), controller, video codec, digital signal processor (Digital Signal Processor, DSP), baseband processor and/or Neural-Network Processing Unit (NPU), etc. Among them, the different processing units may be independent devices or integrated in one or more processors.
控制器可以根据指令操作码和时序信号,产生操作控制信号,完成读取指令和执行指令的控制。The controller can generate operation control signals according to the instruction operation code and timing signals to complete the control of reading instructions and executing instructions.
处理器210中还可以设置存储器,用于存储指令和数据。存储器可以存储用于实现六个模块化功能的指令:检测指令、连接指令、信息管理指令、分析指令、数据传输指令和通知指令,并由处理器210来控制执行。在一些实施方式中,处理器210中的存储器为高速缓冲存储器。该存储器可以保存处理器210刚用过或循环使用的指令或数据。如果处理器210需要再次使用该指令或数据,可从所述存储器中直接调用。避免了重复存取,减少了处理器210的等待时间,因而提高了系统的效率。A memory may also be provided in the processor 210 for storing instructions and data. The memory can store instructions for implementing six modular functions: detection instructions, connection instructions, information management instructions, analysis instructions, data transmission instructions, and notification instructions, and the processor 210 controls the execution. In some embodiments, the memory in the processor 210 is a cache memory. The memory can store instructions or data that have just been used or recycled by the processor 210. If the processor 210 needs to use the instruction or data again, it can be directly called from the memory. Repeated accesses are avoided, the waiting time of the processor 210 is reduced, and the efficiency of the system is improved.
在一些实施方式中,处理器210可以包括一个或多个接口。接口可以包括集成电路(Inter-Integrated Circuit,I2C)接口、集成电路内置音频(Inter-Integrated Circuit Sound,I2S)接口、脉冲编码调制(Pulse Code Modulation,PCM)接口、通用异步收发传输器(Universal Asynchronous Receiver/Transmitter,UART)接口、移动产业处理器接口(Mobile Industry Processor Interface,MIPI)、通用输入输出(General-Purpose Input/Output,GPIO)接口、用户标识模块(Subscriber Identity Module,SIM)接口和/或通用串行总线(Universal Serial Bus,USB)接口等。通过不同的接口和移动终端200的其他部件形成连接。In some embodiments, the processor 210 may include one or more interfaces. Interfaces can include integrated circuit (Inter-Integrated Circuit, I2C) interface, integrated circuit built-in audio (Inter-Integrated Circuit Sound, I2S) interface, pulse code modulation (Pulse Code Modulation, PCM) interface, universal asynchronous transceiver (Universal Asynchronous) Receiver/Transmitter, UART) interface, Mobile Industry Processor Interface (MIPI), General-Purpose Input/Output (GPIO) interface, Subscriber Identity Module (SIM) interface and/ Or Universal Serial Bus (Universal Serial Bus, USB) interface, etc. Connections are formed with other components of the mobile terminal 200 through different interfaces.
USB接口230是符合USB标准规范的接口,具体可以是MiniUSB接口,MicroUSB接口,USBTypeC接口等。USB接口230可以用于连接充电器为移动终端200充电,也可以连接耳机,通过耳机播放音频,还可以用于移动终端200连接其他触控终端,例如连接电脑、外围设备等。The USB interface 230 is an interface that complies with the USB standard specification, and specifically may be a MiniUSB interface, a MicroUSB interface, a USBTypeC interface, and so on. The USB interface 230 can be used to connect a charger to charge the mobile terminal 200, can also connect to earphones to play audio through the earphones, and can also be used to connect the mobile terminal 200 to other touch terminals, such as computers, peripherals, and the like.
充电管理模块240用于从充电器接收充电输入。其中,充电器可以是无线充电器,也可以是有线充电器。在一些有线充电的实施方式中,充电管理模块240可以通过USB接口230接收有线充电器的充电输入。在一些无线充电的实施方式中,充电管理模块240可以通过移动终端200的无线充电线圈接收无线充电输入。充电管理模块240为电池242充电的同时,还可以通过电源管理模块241为触控终端供电。The charging management module 240 is used to receive charging input from the charger. Among them, the charger can be a wireless charger or a wired charger. In some wired charging implementations, the charging management module 240 may receive the charging input of the wired charger through the USB interface 230. In some implementations of wireless charging, the charging management module 240 may receive the wireless charging input through the wireless charging coil of the mobile terminal 200. While the charging management module 240 charges the battery 242, it can also supply power to the touch terminal through the power management module 241.
电源管理模块241用于连接电池242、充电管理模块240与处理器210。电源管理模块241接收电池242和/或充电管理模块240的输入,为处理器210、内部存储器221、显示器290、摄像模组291和无线通信模块260等供电,还可以用于监测电池容量,电池循环次数,电池健康状态(漏电,阻抗)等参数。The power management module 241 is used to connect the battery 242, the charging management module 240, and the processor 210. The power management module 241 receives the input of the battery 242 and/or the charge management module 240, and supplies power to the processor 210, internal memory 221, display 290, camera module 291, and wireless communication module 260. It can also be used to monitor battery capacity. The number of cycles, battery health status (leakage, impedance) and other parameters.
移动终端200的无线通信功能可以通过天线1、天线2、移动通信模块250、无线通信模块260、调制解调处理器以及基带处理器等实现。The wireless communication function of the mobile terminal 200 can be implemented by the antenna 1, the antenna 2, the mobile communication module 250, the wireless communication module 260, the modem processor, and the baseband processor.
天线1和天线2用于发射和接收电磁波信号。移动终端200中的每个天线可用于覆盖单个或多个通信频带。不同的天线还可以复用,以提高天线的利用率。例如:可以将天线1复用为无线局域网的分集天线。在另外一些实施方式中,天线可以和调谐开关结合使用。The antenna 1 and the antenna 2 are used to transmit and receive electromagnetic wave signals. Each antenna in the mobile terminal 200 can be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization. For example: Antenna 1 can be multiplexed as a diversity antenna of a wireless local area network. In other embodiments, the antenna can be used in combination with a tuning switch.
移动通信模块250可以提供应用在移动终端200上的包括2G/3G/4G/5G等无线通信的解决方案。移动通信模块250可以包括至少一个滤波器,开关,功率放大器,低噪声放大 器(Low Noise Amplifier,LNA)等。移动通信模块250可以由天线1接收电磁波,并对接收的电磁波进行滤波、放大等处理,传送至调制解调处理器进行解调。移动通信模块250还可以对经调制解调处理器调制后的信号放大,经天线1转为电磁波辐射出去。The mobile communication module 250 may provide a wireless communication solution including 2G/3G/4G/5G and the like applied to the mobile terminal 200. The mobile communication module 250 may include at least one filter, a switch, a power amplifier, a low noise amplifier (LNA), and so on. The mobile communication module 250 can receive electromagnetic waves by the antenna 1, and perform processing such as filtering and amplifying the received electromagnetic waves, and transmit them to the modem processor for demodulation. The mobile communication module 250 can also amplify the signal modulated by the modem processor, and convert it into electromagnetic wave radiation via the antenna 1.
无线通信模块260可以提供应用在移动终端200上的包括无线局域网(Wireless Local Area Networks,WLAN)(如无线保真(Wireless Fidelity,Wi-Fi)网络)、蓝牙(Bluetooth,BT)、全球导航卫星系统(Global Navigation Satellite System,GNSS)、调频(Frequency Modulation,FM)、近距离无线通信技术(Near Field Communication,NFC)、红外技术(Infrared,IR)等无线通信的解决方案。无线通信模块260可以是集成至少一个通信处理模块的一个或多个器件。无线通信模块260经由天线2接收电磁波,将电磁波信号调频以及滤波处理,将处理后的信号发送到处理器210。无线通信模块260还可以从处理器210接收待发送的信号,对其进行调频,放大,经天线2转为电磁波辐射出去。The wireless communication module 260 can provide applications on the mobile terminal 200, including wireless local area networks (Wireless Local Area Networks, WLAN) (such as Wireless Fidelity (Wi-Fi) networks), Bluetooth (Bluetooth, BT), and global navigation satellites. System (Global Navigation Satellite System, GNSS), Frequency Modulation (FM), Near Field Communication (NFC), Infrared Technology (Infrared, IR) and other wireless communication solutions. The wireless communication module 260 may be one or more devices integrating at least one communication processing module. The wireless communication module 260 receives electromagnetic waves via the antenna 2, frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 210. The wireless communication module 260 may also receive the signal to be sent from the processor 210, perform frequency modulation, amplify, and convert it into electromagnetic waves and radiate it through the antenna 2.
在一些实施方式中,移动终端200的天线1和移动通信模块250耦合,天线2和无线通信模块260耦合,使得移动终端200可以通过无线通信技术与网络以及其他设备通信。所述无线通信技术可以包括全球移动通讯系统(Global System for Mobile communications,GSM),通用分组无线服务(General Packet Radio Service,GPRS),码分多址接入(Code Division Multiple Access,CDMA),宽带码分多址(Wideband Code Division Multiple Access,WCDMA),时分码分多址(Time Division-Synchronous Code Division Multiple Access,TD-SCDMA),长期演进(Long Term Evolution,LTE),新空口(New Radio,NR),BT,GNSS,WLAN,NFC,FM,和/或IR技术等。所述GNSS可以包括全球卫星定位系统(Global Positioning System,GPS),全球导航卫星系统(Global Navigation Satellite System,GLONASS),北斗卫星导航系统(Beidou Navigation Satellite System,BDS),准天顶卫星系统(Quasi-Zenith Satellite System,QZSS)和/或星基增强系统(Satellite Based Augmentation Systems,SBAS)。In some embodiments, the antenna 1 of the mobile terminal 200 is coupled with the mobile communication module 250, and the antenna 2 is coupled with the wireless communication module 260, so that the mobile terminal 200 can communicate with the network and other devices through wireless communication technology. The wireless communication technology may include Global System for Mobile Communications (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (CDMA), and broadband Code Division Multiple Access (Wideband Code Division Multiple Access, WCDMA), Time Division Code Division Multiple Access (Time Division-Synchronous Code Division Multiple Access, TD-SCDMA), Long Term Evolution (LTE), New Radio (New Radio, NR), BT, GNSS, WLAN, NFC, FM, and/or IR technology, etc. The GNSS may include Global Positioning System (GPS), Global Navigation Satellite System (GLONASS), Beidou Navigation Satellite System (BDS), Quasi-Zenith Satellite System (Quasi -Zenith Satellite System, QZSS) and/or Satellite Based Augmentation Systems (SBAS).
移动终端200通过GPU、显示器290及应用处理器等实现显示功能。GPU为的微处理器,连接显示器290和应用处理器。GPU用于执行数学和几何计算,用于图形渲染。处理器210可包括一个或多个GPU,其执行程序指令以生成或改变显示信息。The mobile terminal 200 implements a display function through a GPU, a display 290, an application processor, and the like. The GPU is a microprocessor that connects the display 290 and the application processor. The GPU is used to perform mathematical and geometric calculations and is used for graphics rendering. The processor 210 may include one or more GPUs that execute program instructions to generate or change display information.
移动终端200可以包括一个或多个显示器290,用于显示图像,视频等。显示器290包括显示面板。显示面板可以采用液晶显示屏(Liquid Crystal Display,LCD),有机发光二极管(Organic Light-Emitting Diode,OLED),有源矩阵有机发光二极体或主动矩阵有机发光二极体(Active-Matrix Organic Light Emitting Diode,AMOLED),柔性发光二极管(Flexlight-Emitting Diode,FLED),Miniled,MicroLed,Micro-oLed,量子点发光二极管(Quantum dot Light Emitting Diodes,QLED)等。The mobile terminal 200 may include one or more displays 290 for displaying images, videos, and the like. The display 290 includes a display panel. The display panel can be a Liquid Crystal Display (LCD), Organic Light-Emitting Diode (OLED), Active-Matrix Organic Light-Emitting Diode, or Active-Matrix Organic Light-Emitting Diode (Active-Matrix Organic Light-Emitting Diode). Emitting Diode, AMOLED), Flexible Light Emitting Diode (Flexlight-Emitting Diode, FLED), Miniled, MicroLed, Micro-oLed, Quantum Dot Light Emitting Diodes (QLED), etc.
移动终端200可以通过ISP、摄像模组291、视频编解码器、GPU、显示器290及应用处理器等实现拍摄功能。The mobile terminal 200 can implement a shooting function through an ISP, a camera module 291, a video codec, a GPU, a display 290, and an application processor.
ISP用于处理摄像模组291反馈的数据。例如,拍照时,打开快门,光线通过镜头被 传递到摄像头感光元件上,光信号转换为电信号,摄像头感光元件将所述电信号传递给ISP处理,转化为肉眼可见的图像。在一些实施方式中,ISP可以设置在摄像模组291中。The ISP is used to process the data fed back from the camera module 291. For example, when taking a picture, the shutter is opened, the light is transmitted to the photosensitive element of the camera through the lens, the light signal is converted into an electrical signal, and the photosensitive element of the camera transfers the electrical signal to the ISP for processing and is converted into an image visible to the naked eye. In some embodiments, the ISP may be provided in the camera module 291.
移动终端200可以包括一个或多个摄像模组291,用于捕获静态图像或视频。物体通过镜头生成光学图像投射到感光元件。感光元件把光信号转换成电信号,之后将电信号传递给ISP转换成数字图像信号。ISP将数字图像信号输出到DSP加工处理。DSP将数字图像信号转换成标准的RGB,YUV等格式的图像信号。The mobile terminal 200 may include one or more camera modules 291 for capturing still images or videos. The object generates an optical image through the lens and is projected to the photosensitive element. The photosensitive element converts the optical signal into an electrical signal, and then transfers the electrical signal to the ISP to convert it into a digital image signal. ISP outputs digital image signals to DSP for processing. DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
视频编解码器用于对数字视频压缩或解压缩。移动终端200可以支持一种或多种视频编解码器。这样,移动终端200可以播放或录制多种编码格式的视频,例如:动态图像专家组(Moving Picture Experts Group,MPEG)1,MPEG2,MPEG3,MPEG4等。Video codecs are used to compress or decompress digital video. The mobile terminal 200 may support one or more video codecs. In this way, the mobile terminal 200 can play or record videos in multiple encoding formats, such as: Moving Picture Experts Group (MPEG) 1, MPEG2, MPEG3, MPEG4, etc.
外部存储器接口222可以用于连接外部存储卡,例如Micro SD卡,实现扩展移动终端200的存储能力。外部存储卡通过外部存储器接口222与处理器210通信,实现数据存储功能。例如将音乐,视频等文件保存在外部存储卡中。The external memory interface 222 may be used to connect an external memory card, such as a Micro SD card, so as to expand the storage capacity of the mobile terminal 200. The external memory card communicates with the processor 210 through the external memory interface 222 to realize the data storage function. For example, save music, video and other files in an external memory card.
内部存储器221可以用于存储计算机可执行程序代码,所述可执行程序代码包括指令。内部存储器221可以包括存储程序区和存储数据区。其中,存储程序区可存储操作系统,至少一个功能所需的应用程序(比如声音播放功能,图像播放功能等)等。存储数据区可存储移动终端200使用过程中所创建的数据(比如音频数据,电话本等)等。此外,内部存储器221可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件,闪存器件,通用闪存存储器(Universal Flash Storage,UFS)等。处理器210通过运行存储在内部存储器221的指令和/或存储在设置于处理器中的存储器的指令,执行移动终端200的各种功能应用以及数据处理。The internal memory 221 may be used to store computer executable program code, where the executable program code includes instructions. The internal memory 221 may include a storage program area and a storage data area. Among them, the storage program area can store an operating system, an application program (such as a sound playback function, an image playback function, etc.) required by at least one function, and the like. The data storage area can store data (such as audio data, phone book, etc.) created during the use of the mobile terminal 200. In addition, the internal memory 221 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash storage (Universal Flash Storage, UFS), and the like. The processor 210 executes various functional applications and data processing of the mobile terminal 200 by running instructions stored in the internal memory 221 and/or instructions stored in a memory provided in the processor.
移动终端200可以通过音频模块270、扬声器271、受话器272、麦克风273、耳机接口274及应用处理器等实现音频功能。例如音乐播放、录音等。The mobile terminal 200 can implement audio functions through an audio module 270, a speaker 271, a receiver 272, a microphone 273, a headset interface 274, an application processor, and the like. For example, music playback, recording, etc.
音频模块270用于将数字音频信息转换成模拟音频信号输出,也用于将模拟音频输入转换为数字音频信号。音频模块270还可以用于对音频信号编码和解码。在一些实施方式中,音频模块270可以设置于处理器210中,或将音频模块270的部分功能模块设置于处理器210中。The audio module 270 is used to convert digital audio information into an analog audio signal for output, and is also used to convert an analog audio input into a digital audio signal. The audio module 270 can also be used to encode and decode audio signals. In some embodiments, the audio module 270 may be provided in the processor 210, or part of the functional modules of the audio module 270 may be provided in the processor 210.
扬声器271,也称“喇叭”,用于将音频电信号转换为声音信号。移动终端200可以通过扬声器271收听音乐,或收听免提通话。The speaker 271, also called a "speaker", is used to convert audio electrical signals into sound signals. The mobile terminal 200 can listen to music through the speaker 271, or listen to a hands-free call.
受话器272,也称“听筒”,用于将音频电信号转换成声音信号。当移动终端200接听电话或语音信息时,可以通过将受话器272靠近人耳接听语音。The receiver 272, also called "earpiece", is used to convert audio electrical signals into sound signals. When the mobile terminal 200 answers a call or voice message, it can receive the voice by bringing the receiver 272 close to the human ear.
麦克风273,也称“话筒”,“传声器”,用于将声音信号转换为电信号。当拨打电话或发送语音信息时,用户可以通过人嘴靠近麦克风273发声,将声音信号输入到麦克风273。移动终端200可以设置多个麦克风273,以实现降噪、识别声音来源、定向录音等功能。The microphone 273, also called "microphone" or "microphone", is used to convert sound signals into electrical signals. When making a call or sending a voice message, the user can approach the microphone 273 through the mouth to make a sound, and input the sound signal to the microphone 273. The mobile terminal 200 may be provided with multiple microphones 273 to implement functions such as noise reduction, sound source identification, and directional recording.
耳机接口274用于连接有线耳机。耳机接口274可以是USB接口230,也可以是3.5mm的开放移动触控终端平台(Open Mobile Terminal Platform,OMTP)标准接口,美国蜂窝 电信工业协会(Cellular Telecommunications Industry Association of the USA,CTIA)标准接口。The earphone interface 274 is used to connect wired earphones. The earphone interface 274 can be a USB interface 230, or a 3.5mm open mobile terminal platform (Open Mobile Terminal Platform, OMTP) standard interface, and the US Cellular Telecommunications Industry Association (Cellular Telecommunications Industry Association of the USA, CTIA) standard interface .
深度传感器2801用于获取景物的深度信息。在一些实施方式中,深度传感器可以设置于摄像模组291。压力传感器2802用于感受压力信号,可以将压力信号转换成电信号,用于实现压力触控等功能。陀螺仪传感器2803可以用于确定移动终端200的运动姿态,可用于拍摄防抖、导航、体感游戏等场景。气压传感器2804用于测量气压,可通过计算海拔高度,辅助定位和导航。The depth sensor 2801 is used to obtain depth information of the scene. In some embodiments, the depth sensor may be provided in the camera module 291. The pressure sensor 2802 is used to sense the pressure signal, and can convert the pressure signal into an electrical signal for implementing functions such as pressure touch. The gyroscope sensor 2803 can be used to determine the motion posture of the mobile terminal 200, and can be used to shoot scenes such as anti-shake, navigation, and somatosensory games. The air pressure sensor 2804 is used to measure air pressure, and can assist positioning and navigation by calculating the altitude.
此外,根据实际需要,还可以在传感器模块280中设置其他功能的传感器,例如磁传感器、加速度传感器、距离传感器、接近光传感器、指纹传感器、温度传感器、触摸传感器、环境光传感器、骨传导传感器等等。In addition, according to actual needs, sensors with other functions can be set in the sensor module 280, such as magnetic sensors, acceleration sensors, distance sensors, proximity light sensors, fingerprint sensors, temperature sensors, touch sensors, ambient light sensors, bone conduction sensors, etc. Wait.
指示器292可以是指示灯,可以用于指示充电状态,电量变化,也可以用于指示消息,未接来电,通知等。The indicator 292 can be an indicator light, which can be used to indicate the charging status, power change, or to indicate messages, missed calls, notifications, and so on.
马达293可以产生振动提示,例如来电、闹钟、接收信息等的振动提示,也可以用于触摸振动反馈等。The motor 293 can generate vibration prompts, such as vibration prompts for incoming calls, alarm clocks, received messages, etc., and can also be used for touch vibration feedback and the like.
按键294包括开机键,音量键等。按键294可以是机械按键。也可以是触摸式按键。移动终端200可以接收按键输入,产生与移动终端200的用户设置以及功能控制有关的键信号输入。The button 294 includes a power-on button, a volume button, and so on. The button 294 may be a mechanical button. It can also be a touch button. The mobile terminal 200 may receive key input, and generate key signal input related to user settings and function control of the mobile terminal 200.
移动终端200可以支持一个或多个SIM卡接口295,用于连接SIM卡,使移动终端200通过SIM卡和网络交互,实现通话以及数据通信等功能。The mobile terminal 200 may support one or more SIM card interfaces 295 for connecting to the SIM card, so that the mobile terminal 200 can interact with the network through the SIM card to implement functions such as call and data communication.
下面对本公开示例性实施方式的虚拟影像实现方法和虚拟影像实现装置进行具体说明。The virtual image realization method and virtual image realization device of the exemplary embodiment of the present disclosure will be described in detail below.
虚拟影像实现方法可以以AR程序的方式实现。图3示出由上述第一终端110执行的虚拟影像实现方法的示意性流程,第一终端110可以通过运行AR程序的客户端,实现以下步骤S310至S330:The virtual image realization method can be realized in the form of an AR program. FIG. 3 shows a schematic flow of the virtual image realization method executed by the first terminal 110. The first terminal 110 may implement the following steps S310 to S330 through a client running an AR program:
步骤S310,获取用户输入的针对目标物体的编辑信息。Step S310: Obtain editing information for the target object input by the user.
其中,目标物体可以是真实世界环境中的任意物体,如房间中的冰箱、水池等。用户可以针对该目标物体,输入任何形式、任何内容的编辑信息,其可以包括文本、图片、手绘图画、动画效果、视频、音频中的任意一种或多种。例如,用户可以输入一段文字,设置相应的字体、颜色等,并为其编辑便签纸样式的图框,以及文字显示的动画效果,如以淡化的方式逐渐出现、从某个方向飞入等;或者用户可以手动绘制图画,并为其添加一定的背景纹理等;或者用户可以编辑一个虚拟图标,并添加一段音频,该虚拟图标可以触发播放该音频,等等。Among them, the target object can be any object in the real world environment, such as a refrigerator, a pool, etc. in a room. The user can input editing information in any form and content for the target object, which can include any one or more of text, pictures, hand-drawn drawings, animation effects, video, and audio. For example, the user can input a paragraph of text, set the corresponding font, color, etc., and edit the frame of the note paper style for it, as well as the animation effect of the text display, such as gradually appearing in a faded manner, flying in from a certain direction, etc.; Or the user can manually draw the picture and add a certain background texture to it; or the user can edit a virtual icon and add a piece of audio, the virtual icon can trigger the audio to be played, and so on.
在一种实施方式中,可以通过摄像头扫描的方式确定目标物体。具体来说,在步骤S310前,可以先执行以下步骤:In an embodiment, the target object can be determined by scanning with a camera. Specifically, before step S310, the following steps may be performed:
当从现实场景中扫描到目标物体时,触发显示编辑界面。When the target object is scanned from the real scene, the editing interface is triggered to be displayed.
其中,现实场景是指目标物体所在的真实世界环境,如现实的房间、咖啡店、商场或者街道等。当用户想要对某个物体编辑虚拟影像时,可以打开第一终端的摄像头,对准该物体,摄像头扫描所采集的图像,当识别到该物体时,触发弹出编辑界面,或者弹出消息“是否编辑虚拟影像”,用户选择“是”则弹出编辑界面。编辑界面用于输入编辑信息,可以包括各种信息的编辑栏,如文本栏、图片栏、动画效果栏等。Among them, the real scene refers to the real world environment where the target object is located, such as a real room, coffee shop, shopping mall, or street. When the user wants to edit a virtual image of an object, he can turn on the camera of the first terminal, aim at the object, and the camera scans the collected image. When the object is recognized, the editing interface is triggered, or the message "Is it right?" Edit virtual image", the user selects "Yes" and the editing interface pops up. The editing interface is used to input editing information, and can include editing columns for various information, such as text columns, picture columns, and animation effect columns.
进一步的,为了准确识别到目标物体,当用户打开摄像头扫描时,客户端可以引导用户移动多个角度,例如在扫描界面上显示“请换一个角度对准目标”、“请对准目标的背面”等引导信息,从而采集到目标物体的全貌。Further, in order to accurately identify the target object, when the user turns on the camera to scan, the client can guide the user to move multiple angles, for example, the scanning interface displays "please aim at the target from another angle", "please aim at the back of the target" "And other guidance information, so as to collect the full picture of the target object.
在另一种实施方式中,用户可以在客户端上点击选取目标物体。具体来说,客户端可以事先从服务器下载场景信息并打开显示,用户在场景中点击某一物体,将其确定为目标物体,并进入编辑界面。In another embodiment, the user can click to select the target object on the client. Specifically, the client can download the scene information from the server in advance and open the display. The user clicks on an object in the scene, determines it as the target object, and enters the editing interface.
步骤S320,根据上述编辑信息生成目标物体的虚拟影像。In step S320, a virtual image of the target object is generated according to the above-mentioned editing information.
在编辑完成后,可以对应生成虚拟影像。一般的,用户的编辑信息即虚拟影像的主要内容,客户端可以直接将编辑信息作为虚拟影像,也可以以此为基础,添加一些虚拟的视觉效果,或者将虚拟影像与目标物体进行一定的融合。举例来说,用户以在房间内的冰箱为目标物体,编辑了一张虚拟便签,包括便签的大小、形状、颜色等,并在便签中输入了文本,客户端可以自动为虚拟便签生成虚拟的光影,或者设置一定的透明度,以放置在冰箱的一个平面上,或者对虚拟标签做一定的变形,使其更具立体感,或者对虚拟标签增加动态效果,如随风抖动等,或者对其中的文本生成对应的一段语音等等。After the editing is completed, a virtual image can be generated correspondingly. Generally, the user's editing information is the main content of the virtual image. The client can directly use the editing information as the virtual image, or based on this, add some virtual visual effects, or combine the virtual image with the target object to a certain extent . For example, the user edits a virtual note with the refrigerator in the room as the target object, including the size, shape, and color of the note, and enters text in the note. The client can automatically generate a virtual note for the virtual note. Light and shadow, or set a certain degree of transparency to be placed on a plane of the refrigerator, or deform the virtual label to make it more three-dimensional, or add dynamic effects to the virtual label, such as shaking with the wind, or to it The text generates a corresponding piece of speech and so on.
在一种可选的实施方式中,用户还可以编辑虚拟影像的显示位置,使其包括于上述编辑信息中。例如用户点击目标物体上的某一位置,在该位置上编辑生成虚拟便签,或者用户在编辑虚拟便签后,将其移动到某一位置上,即虚拟影像的显示位置。In an optional implementation manner, the user can also edit the display position of the virtual image to include it in the above-mentioned editing information. For example, the user clicks on a certain position on the target object to edit and generate a virtual note at that position, or the user edits the virtual note and moves it to a certain position, that is, the display position of the virtual image.
步骤S330,将虚拟影像上传至目标物体所在的增强现实场景中,增强现实场景存储于云端,使第二终端通过从云端获取增强现实场景而显示虚拟影像。In step S330, the virtual image is uploaded to the augmented reality scene where the target object is located, and the augmented reality scene is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image.
本示例性实施方式中,云端对于AR数据的存储是以增强现实场景为单位实现的。增强现实场景是指对目标物体所在的现实场景建立的数据集合,可以包括现实场景的地图数据、云锚点数据、各个物体的三维点云数据、已添加的虚拟影像数据(如用户对其他物体编辑的虚拟影像)等,可见,凡是与该场景相关的AR数据,都可以存储在对应的增强现实场景中。当增加新的数据,或者数据发生更新时,都可以同步到增强现实场景中。由此,当客户端生成虚拟影像并上传云端后,可以直接存储到对应的增强现实场景中。In this exemplary embodiment, the storage of AR data in the cloud is implemented in units of augmented reality scenes. Augmented reality scene refers to a data collection established for the real scene where the target object is located, which can include map data of the real scene, cloud anchor point data, 3D point cloud data of each object, and added virtual image data (such as user It can be seen that all AR data related to the scene can be stored in the corresponding augmented reality scene. When new data is added or data is updated, it can be synchronized to the augmented reality scene. Thus, after the client generates a virtual image and uploads it to the cloud, it can be directly stored in the corresponding augmented reality scene.
关于如何确定目标物体所在的增强现实场景,下面提供几个实施方式。Regarding how to determine the augmented reality scene where the target object is located, several implementation manners are provided below.
(1)用户首次扫描现实场景,即云端未记录关于现实场景的任何AR数据。在这种情况下,除了扫描目标物体外,还需要扫描目标物体所在的现实场景,生成地图数据。地图数据可以是反映现实场景特征的三维点云模型,包括地形、纹理、障碍物等信息。第一终端生成地图数据可以采用SLAM(Simultaneous Localization And Mapping,同时定位与建 图)方法。具体来说,启动摄像头后,用户A手持(或佩戴)第一终端,在现实场景内移动;第一终端采集场景图像,并同时记录位姿信息;在获得一定帧数的图像后,可以初始化场景的地图数据,通常只包括场景的一小部分,或者与实际场景有差别;此后,每采集一帧的图像,都可以根据该图像对地图数据进行更新优化,增加地图中没有的地图点,或者修正已有地图点的位置等;直到遍历整个场景,得到完整的地图数据。在SLAM采集场景图像时,通常从连续帧图像中选取一定数量的关键帧图像,这是为了减少建模过程中的信息冗余而选取的具有代表性的帧,通常可以每间隔一定的帧数选取一帧为关键帧,或者在图像内容变化较多时提取关键帧图像。在创建地图数据时,通常将关键帧图像的集合打包到地图数据中,以便于后续使用。(1) The user scans the real scene for the first time, that is, the cloud does not record any AR data about the real scene. In this case, in addition to scanning the target object, it is also necessary to scan the actual scene where the target object is located to generate map data. The map data can be a three-dimensional point cloud model reflecting the characteristics of a real scene, including information such as terrain, texture, and obstacles. The map data generated by the first terminal can adopt the SLAM (Simultaneous Localization And Mapping, simultaneous positioning and mapping) method. Specifically, after starting the camera, user A holds (or wears) the first terminal and moves in the real scene; the first terminal collects scene images and records pose information at the same time; after obtaining a certain number of frames of images, it can be initialized The map data of the scene usually only includes a small part of the scene, or is different from the actual scene; after that, every time a frame of image is collected, the map data can be updated and optimized according to the image, and the map points that are not in the map can be added. Or modify the position of the existing map point, etc.; until the entire scene is traversed, the complete map data is obtained. When SLAM collects scene images, a certain number of key frame images are usually selected from continuous frame images. This is a representative frame selected in order to reduce information redundancy in the modeling process, usually a certain number of frames per interval Choose a frame as the key frame, or extract the key frame image when the image content changes a lot. When creating map data, a collection of key frame images is usually packaged into the map data for subsequent use.
之后可以将地图数据上传至云端,例如可以和目标物体的虚拟影像一起上传,也可以先于虚拟影像上传,这样可以在云端上建立现实场景对应的增强现实场景。由此,很容易确定目标物体属于所建立的增强现实场景。The map data can then be uploaded to the cloud, for example, it can be uploaded together with the virtual image of the target object, or it can be uploaded before the virtual image, so that an augmented reality scene corresponding to the real scene can be established on the cloud. Therefore, it is easy to determine that the target object belongs to the established augmented reality scene.
举例来说,用户A以自家厨房为现实场景,扫描厨房建立地图数据,同时扫描厨房中的冰箱,编辑虚拟影像。客户端将地图数据和虚拟影像上传到云端,建立对应的增强现实场景,如可以将场景命名为“用户A的厨房”,同时将虚拟影像添加到该增强现实场景。For example, user A uses his own kitchen as a real scene, scans the kitchen to create map data, scans the refrigerator in the kitchen, and edits virtual images. The client uploads map data and virtual images to the cloud to create a corresponding augmented reality scene. For example, the scene can be named "User A's Kitchen" and the virtual image is added to the augmented reality scene.
(2)如果之前已经建立过增强现实场景,用户在扫描目标物体前,可以先在客户端上进入增强现实场景。例如可以输入场景编号或场景名,发送请求到云端,在云端反馈允许后,客户端下载增强现实场景的信息。一般的,增强现实场景中包括一些或者全部物体的关键帧图像,例如第一次扫描厨房时,对厨房中的冰箱、水池等特征性较强的物体记录关键帧图像。当用户扫描目标物体时,客户端可以将目标物体的图像与关键帧图像进行匹配,当匹配成功时,确定目标物体为已记录的某个物体,此时可以触发显示编辑界面。(2) If an augmented reality scene has been created before, the user can enter the augmented reality scene on the client before scanning the target object. For example, you can enter the scene number or scene name, send a request to the cloud, and after the cloud feedback allows, the client downloads the information of the augmented reality scene. Generally, an augmented reality scene includes key frame images of some or all objects. For example, when the kitchen is scanned for the first time, key frame images are recorded for more characteristic objects such as refrigerators and sinks in the kitchen. When the user scans the target object, the client can match the image of the target object with the key frame image. When the matching is successful, the target object is determined to be a recorded object, and the editing interface can be triggered at this time.
(3)如果客户端未进入某个增强现实场景,而用户直接扫描目标物体并编辑相关信息,在上传虚拟影像时,还可以从所采集的目标物体的图像中选取至少一关键帧图像,一同上传至云端。云端可以将该关键帧图像与已存储的增强现实场景中的关键帧图像进行匹配,以确定目标物体属于哪个增强现实场景。考虑到匹配计算的数据量较大,且容易出现误匹配的情况(如冰箱的外观都差不多,用户A家的冰箱可能匹配到用户B家的厨房),第一终端在上传虚拟影像和目标物体的关键帧图像时,还可以发送GPS(Global Positioning System,全球定位系统)等定位数据,使云端根据定位数据缩小匹配范围。(3) If the client does not enter a certain augmented reality scene, and the user directly scans the target object and edits related information, when uploading a virtual image, you can also select at least one key frame image from the collected images of the target object, and Upload to the cloud. The cloud can match the key frame image with the key frame image in the stored augmented reality scene to determine which augmented reality scene the target object belongs to. Considering that the amount of data for matching calculation is large and mismatching is easy to occur (for example, the appearance of the refrigerator is similar, the refrigerator of user A may match the kitchen of user B), the first terminal is uploading virtual images and target objects When the key frame image is selected, GPS (Global Positioning System, Global Positioning System) and other positioning data can also be sent, so that the cloud can narrow the matching range based on the positioning data.
(4)每次生成虚拟影像前,要求用户扫描目标物体和现实场景,将关键帧图像、地图数据、虚拟影像发送至云端,云端根据地图数据查找对应的增强现实场景,如果找到则直接调取增强现实场景,如果未找到则新建一个增强现实场景。(4) Before each virtual image is generated, the user is required to scan the target object and the real scene, and send the key frame image, map data, and virtual image to the cloud. The cloud finds the corresponding augmented reality scene based on the map data, and if found, it will be directly retrieved Augmented reality scene, if not found, create a new augmented reality scene.
实际应用中,可以根据情况采用或结合上述不同的方式,本公开对此不做限定。In actual applications, the above-mentioned different methods can be adopted or combined according to the situation, which is not limited in the present disclosure.
在将第一终端上编辑的虚拟影像添加到增强现实场景后,如果有第二终端获取增强现实场景的信息,可以同步获取到该虚拟影像,从而可以在第二终端上显示该虚拟影像。例如用户A编辑生成了冰箱的虚拟便签,上传到云端;用户B下载增强现实场景后,在观看 冰箱时,可以显示该虚拟便签。After the virtual image edited on the first terminal is added to the augmented reality scene, if a second terminal obtains the information of the augmented reality scene, the virtual image can be obtained synchronously, so that the virtual image can be displayed on the second terminal. For example, user A edits and generates a virtual note of the refrigerator and uploads it to the cloud; after user B downloads the augmented reality scene, the virtual note can be displayed while watching the refrigerator.
在一种可选的实施方式中,还可以根据目标物体在增强现实场景中的位置,生成虚拟影像的提示信息,并在上传虚拟影像时,将提示信息一同上传至增强现实场景中。该提示信息用于提示虚拟影像的方位,以便于用户处于增强现实场景中的不同位置时,都可以按照提示信息找到虚拟影像。该提示信息可以与虚拟影像同时生成,也可以晚于虚拟影像生成。下面对如何生成提示信息,提供两个具体实施方式:In an optional implementation manner, the prompt information of the virtual image can also be generated according to the position of the target object in the augmented reality scene, and when the virtual image is uploaded, the prompt information is uploaded to the augmented reality scene together. The prompt information is used to prompt the orientation of the virtual image, so that when the user is in different positions in the augmented reality scene, the virtual image can be found according to the prompt information. The prompt information can be generated at the same time as the virtual image, or can be generated later than the virtual image. The following provides two specific implementation methods on how to generate prompt information:
在一种实施方式中,客户端首先确定目标物体在增强现实场景中的位置,通常也是虚拟影像的位置(虚拟影像大多显示在目标物体上),然后可以以该位置为基准,将增强现实场景划分为多个区域,针对每个区域生成一条引导性的虚拟指示箭头,以指向虚拟影像的位置。In one embodiment, the client first determines the position of the target object in the augmented reality scene, which is usually the position of the virtual image (most of the virtual images are displayed on the target object), and then can use this position as a reference to convert the augmented reality scene Divided into multiple areas, a guiding virtual indicator arrow is generated for each area to point to the position of the virtual image.
在另一种实施方式中,客户端首先确定目标物体在增强现实场景中的位置,通常也是虚拟影像的位置(虚拟影像大多显示在目标物体上),然后可以在该位置上生成一个虚拟图标,用于提示该虚拟影像,当用户距离虚拟影像的位置较远时,可以显示该虚拟图标。In another embodiment, the client first determines the position of the target object in the augmented reality scene, which is usually the position of the virtual image (most of the virtual images are displayed on the target object), and then can generate a virtual icon at this position. It is used to prompt the virtual image, and the virtual icon can be displayed when the user is far away from the position of the virtual image.
图4示出由上述第二终端120执行的虚拟影像实现方法的示意性流程,第二终端120可以通过运行AR程序的客户端,实现以下步骤S410和S420:FIG. 4 shows a schematic flow of the virtual image realization method executed by the above-mentioned second terminal 120. The second terminal 120 may implement the following steps S410 and S420 through a client running an AR program:
步骤S410,从云端获取增强现实场景,该增强现实场景至少包括一目标物体的虚拟影像,该虚拟影像由第一终端上传至该增强现实场景中。Step S410: Obtain an augmented reality scene from the cloud. The augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal.
步骤S420,基于上述增强现实场景进行重定位,当重定位成功时,显示上述虚拟影像或者虚拟影像的提示信息。In step S420, relocation is performed based on the above-mentioned augmented reality scene, and when the relocation is successful, the above-mentioned virtual image or prompt information of the virtual image is displayed.
举例来说:假设用户A使用第一终端,编辑生成了虚拟影像上传到云端,之后用户B使用第二终端下载增强现实场景,则获得该虚拟影像;或者第二终端曾经下载过该增强现实场景,当第一终端上传虚拟影像后,云端可以向第二终端发送更新通知,用户B选择更新后,获取到新增的虚拟影像;或者第一终端和第二终端同时连接云端,并打开增强现实场景,第一终端生成并发送虚拟影像后,云端可以同步到第二终端。For example: suppose that user A uses the first terminal to edit and generate a virtual image and upload it to the cloud, and then user B uses the second terminal to download an augmented reality scene, then the virtual image is obtained; or the second terminal has downloaded the augmented reality scene After the first terminal uploads the virtual image, the cloud can send an update notification to the second terminal. After user B chooses to update, the new virtual image is obtained; or the first terminal and the second terminal are connected to the cloud at the same time, and the augmented reality is turned on In the scenario, after the first terminal generates and sends the virtual image, the cloud can be synchronized to the second terminal.
在一种可选的实施方式中,可以为获取增强现实场景设置相应的权限。例如,用户A通过第一终端创建增强现实场景时,设置密码,其他用户需要输入正确的密码才能从云端下载该增强现实场景;或者用户A通过第一终端创建增强现实场景时,设置访问白名单,只有在白名单上的用户才能从云端下载该增强现实场景,等等。In an optional implementation manner, corresponding permissions may be set for acquiring the augmented reality scene. For example, when user A creates an augmented reality scene through the first terminal, set a password, and other users need to enter the correct password to download the augmented reality scene from the cloud; or when user A creates an augmented reality scene through the first terminal, set an access whitelist , Only users on the whitelist can download the augmented reality scene from the cloud, and so on.
在获取增强现实场景后,需要进行重定位,以确定第二终端所处的现实场景即增强现实场景。重定位过程一般如下:第二终端启动摄像头,用户B手持(或佩戴)第二终端,在现实场景内移动;第二终端采集现实场景的图像,与增强现实场景中的地图数据进行匹配,当匹配成功时,确定重定位成功,此时可以显示虚拟影像或其提示信息。具体显示虚拟影像还是提示信息,与用户所处的位置有关。一般的,用户距离目标物体较近,且目标物体在视野范围以内时,显示虚拟影像,否则显示提示信息,以引导用户移动到可看到虚拟影像的其他位置。After acquiring the augmented reality scene, relocation is required to determine the reality scene where the second terminal is located, that is, the augmented reality scene. The relocation process is generally as follows: the second terminal starts the camera, and user B holds (or wears) the second terminal and moves in the real scene; the second terminal collects images of the real scene and matches it with the map data in the augmented reality scene. When the matching is successful, it is determined that the relocation is successful, and the virtual image or its prompt message can be displayed at this time. Whether the virtual image is displayed or the prompt information is specifically related to the location of the user. Generally, when the user is close to the target object and the target object is within the field of view, a virtual image is displayed, otherwise, a prompt message is displayed to guide the user to move to another location where the virtual image can be seen.
需要说明的是,虚拟影像一般显示于目标物体上,如果第一终端编辑了虚拟影像的显示位置,第二终端在获取增强现实场景时,也可以得到该位置,并将虚拟影像显示于该位置。It should be noted that the virtual image is generally displayed on the target object. If the first terminal edits the display position of the virtual image, the second terminal can also obtain the position when acquiring the augmented reality scene and display the virtual image at that position. .
在一种可选的实施方式中,参考图5所示,步骤S420可以包括以下步骤S501至S504:In an optional implementation manner, referring to FIG. 5, step S420 may include the following steps S501 to S504:
步骤S501,基于增强现实场景进行重定位,当重定位成功时,确定第二终端在增强现实场景中的当前位姿;Step S501: Perform relocation based on the augmented reality scene, and when the relocation is successful, determine the current pose of the second terminal in the augmented reality scene;
步骤S502,根据上述当前位姿确定第二终端的当前视野区域;Step S502: Determine the current view area of the second terminal according to the foregoing current pose;
步骤S503,当虚拟影像的显示位置处于当前视野区域内时,显示虚拟影像;Step S503, when the display position of the virtual image is within the current view area, display the virtual image;
步骤S504,当虚拟影像的显示位置处于当前视野区域外时,显示虚拟影像的提示信息。Step S504: When the display position of the virtual image is outside the current view area, display the prompt information of the virtual image.
重定位的过程也是将第二终端的相机坐标系与增强现实场景的世界坐标系进行对准的过程,通过重定位可以确定第二终端在增强现实场景的世界坐标系中的当前位姿;根据当前位姿,以及第二终端的视场角(例如可以是120度),可以确定当前视野区域,以模拟第二终端的相机可以拍摄到的区域范围,或者用户处于当前位姿下,可以看到的视野范围;进而可以确定用户理论上是否能够看到虚拟影像,即虚拟影像的显示位置处于当前视野区域内时,判断用户能够看到虚拟影像,则显示虚拟影像,虚拟影像的显示位置处于当前视野区域外时,判断用户不能看到虚拟影像,显示提示信息。The relocation process is also the process of aligning the camera coordinate system of the second terminal with the world coordinate system of the augmented reality scene. Through relocation, the current pose of the second terminal in the world coordinate system of the augmented reality scene can be determined; The current pose and the field of view of the second terminal (for example, 120 degrees) can determine the current field of view area to simulate the range of the area that can be captured by the camera of the second terminal, or the user is in the current pose and can see It can then determine whether the user can see the virtual image theoretically, that is, when the display position of the virtual image is within the current field of view, it is judged that the user can see the virtual image, then the virtual image is displayed, and the display position of the virtual image is When the current field of view is outside, it is determined that the user cannot see the virtual image, and a prompt message is displayed.
在一种可选的实施方式中,也可以基于目标物体进行重定位,这样在重定位成功时,可以直接确定用户能够看到虚拟影像,因此显示虚拟影像。具体来说,可以从增强现实场景中获取目标物体的关键帧图像,然后将采集的当前图像与关键帧图像进行匹配,当第二终端刚好对准目标物体时,所拍摄的当前图像与关键帧图像能够匹配成功,进而显示虚拟影像。In an optional implementation manner, the relocation may also be performed based on the target object, so that when the relocation is successful, it can be directly determined that the user can see the virtual image, and therefore the virtual image is displayed. Specifically, the key frame image of the target object can be obtained from the augmented reality scene, and then the collected current image can be matched with the key frame image. When the second terminal is just aligned with the target object, the captured current image and key frame The image can be matched successfully, and then the virtual image is displayed.
在一种可选的实施方式中,在显示虚拟影像的提示信息时,如果用户距离目标物体较远,且目标物体不再视野范围内,则可以沿虚拟影像的方向,在当前界面的边缘处显示提示信息,如虚拟引导箭头等;如果用户距离目标物体较远,但目标物体处于视野范围内,则可以显示以虚拟图标的形式显示虚拟影像的提示信息,例如在目标物体的上方显示虚拟图标,并用一个箭头指向目标物体,以表示该位置具有虚拟影像。In an optional implementation manner, when the prompt information of the virtual image is displayed, if the user is far away from the target object and the target object is no longer in the field of view, the user can follow the direction of the virtual image at the edge of the current interface. Display prompt information, such as virtual guide arrows, etc.; if the user is far away from the target object, but the target object is within the field of view, it can display the prompt information that displays the virtual image in the form of a virtual icon, such as displaying a virtual icon above the target object , And use an arrow to point to the target object to indicate that the location has a virtual image.
在一种可选的实施方式中,可以允许用户修改虚拟影像,参考图6所示,通过以下步骤S601和S602实现:In an optional implementation manner, the user may be allowed to modify the virtual image, as shown in FIG. 6, which is implemented through the following steps S601 and S602:
步骤S601,获取用户针对虚拟影像的修改信息,根据修改信息得到新的虚拟影像。In step S601, the user's modification information for the virtual image is obtained, and a new virtual image is obtained according to the modification information.
步骤S602,将上述新的虚拟影像上传至增强现实场景中,以更新云端上存储的虚拟影像。Step S602, upload the above-mentioned new virtual image to the augmented reality scene to update the virtual image stored on the cloud.
其中,用户可以针对虚拟影像的任意一方面或多方面进行修改,如修改文字内容,替换图片,修改动画效果,移动位置等。第二终端修改得到新的虚拟影像后,上传至云端,云端更新原虚拟影像,可以覆盖原虚拟影像,也可以同时保存新的虚拟影像和原虚拟影像。此后,其他终端在从云端下载增强现实场景时,可以得到新的虚拟影像。Among them, the user can modify any one or more aspects of the virtual image, such as modifying text content, replacing pictures, modifying animation effects, moving positions, and so on. After the second terminal modifies and obtains the new virtual image, it is uploaded to the cloud, and the cloud updates the original virtual image, which can overwrite the original virtual image, and can also save the new virtual image and the original virtual image at the same time. After that, other terminals can obtain new virtual images when downloading augmented reality scenes from the cloud.
此外,图6所示的方法过程也可以由第一终端或其他任意终端执行。例如,凡是进入增强现实场景的用户,均可以修改其中的虚拟影像;或者用户在编辑生成虚拟影像时设定密码,其他用户输入正确的密码后才能修改,等等。In addition, the method process shown in FIG. 6 may also be executed by the first terminal or any other terminal. For example, any user who enters the augmented reality scene can modify the virtual image in it; or the user sets a password when editing and generating the virtual image, and other users can modify it after entering the correct password, and so on.
综上所述,本示例性实施方式中,由第一终端获取用户输入的针对目标物体的编辑信息,根据编辑信息生成目标物体的虚拟影像,并上传至存储于云端的增强现实场景中;第二终端在获取该增强现实场景后,通过重定位可以显示该虚拟影像或其提示信息。一方面,提供了一种允许用户自由编辑虚拟内容的交互方式,基于同一现实场景可以增加不同的虚拟元素,从而提高增强现实的多样化体验。另一方面,通过编辑与显示虚拟影像的方式,提供了一种实现信息传递与共享的渠道,具有较高的实用性与趣味性。再一方面,增强现实场景的相关数据存储于云端,从而节省终端的存储空间,提高运行效率,并且数据是以增强现实场景为单位存储与读取的,保证了信息的隐私性。In summary, in this exemplary embodiment, the first terminal obtains the editing information for the target object input by the user, generates a virtual image of the target object according to the editing information, and uploads it to the augmented reality scene stored in the cloud; After acquiring the augmented reality scene, the second terminal can display the virtual image or its prompt information through relocation. On the one hand, it provides an interactive way that allows users to freely edit virtual content, and different virtual elements can be added based on the same real scene, thereby improving the diversified experience of augmented reality. On the other hand, by editing and displaying virtual images, it provides a channel for information transmission and sharing, which is highly practical and interesting. On the other hand, the data related to the augmented reality scene is stored in the cloud, thereby saving the storage space of the terminal and improving the operation efficiency, and the data is stored and read in units of the augmented reality scene, ensuring the privacy of information.
本公开示例性实施方式还提供一种虚拟影像实现装置,可以设置于上述第一终端110。如图7所示,该虚拟影像实现装置700可以包括处理器710与存储器720。其中,存储器720存储有以下程序模块:Exemplary embodiments of the present disclosure also provide a virtual image realization device, which may be installed in the above-mentioned first terminal 110. As shown in FIG. 7, the virtual image realization device 700 may include a processor 710 and a memory 720. Among them, the memory 720 stores the following program modules:
获取模块721,用于获取用户输入的针对目标物体的编辑信息;The obtaining module 721 is used to obtain edit information for the target object input by the user;
生成模块722,用于根据上述编辑信息生成目标物体的虚拟影像;The generating module 722 is configured to generate a virtual image of the target object according to the above-mentioned editing information;
上传模块723,用于将上述虚拟影像上传至增强现实场景中,该增强现实场景存储于云端,使第二终端通过从云端获取增强现实场景而显示虚拟影像;The upload module 723 is configured to upload the above-mentioned virtual image to an augmented reality scene, the augmented reality scene is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image;
处理器710用于执行上述程序模块。The processor 710 is configured to execute the foregoing program modules.
在一种可选的实施方式中,获取模块721,还用于当从现实场景中扫描到目标物体时,触发显示编辑界面,编辑界面用于输入编辑信息。In an optional implementation manner, the acquisition module 721 is further configured to trigger the display of an editing interface when a target object is scanned from a real scene, and the editing interface is used to input editing information.
在一种可选的实施方式中,上传模块723,还用于在上传虚拟影像时,从所采集的目标物体的图像中选取至少一关键帧图像,上传至增强现实场景中。In an optional implementation manner, the upload module 723 is also used to select at least one key frame image from the collected images of the target object when uploading the virtual image, and upload it to the augmented reality scene.
在一种可选的实施方式中,生成模块722,还用于根据目标物体在增强现实场景中的位置,生成虚拟影像的提示信息,提示信息用于提示虚拟影像的方位。上传模块730,还用于将提示信息上传至增强现实场景中。In an optional implementation manner, the generating module 722 is further configured to generate prompt information of the virtual image according to the position of the target object in the augmented reality scene, and the prompt information is used to prompt the orientation of the virtual image. The upload module 730 is also used to upload the prompt information to the augmented reality scene.
在一种可选的实施方式中,生成模块722,被配置为针对增强现实场景中的每个区域,生成由该区域指向虚拟影像的虚拟指示箭头。In an optional implementation manner, the generating module 722 is configured to generate a virtual indicating arrow pointing from the area to the virtual image for each area in the augmented reality scene.
在一种可选的实施方式中,上述编辑信息可以包括文本、图片、手绘图画、动画效果、视频、音频中的任意一种或多种;此外,上述编辑信息还可以包括虚拟影像的显示位置。In an optional implementation manner, the above-mentioned editing information may include any one or more of text, pictures, hand-drawn drawings, animation effects, video, and audio; in addition, the above-mentioned editing information may also include the display position of the virtual image. .
在一种可选的实施方式中,生成模块722,还用于扫描目标物体所在的现实场景,生成地图数据。上传模块723,还用于将地图数据上传至云端,以建立现实场景对应的增强现实场景。In an optional implementation manner, the generating module 722 is also used to scan the actual scene where the target object is located to generate map data. The upload module 723 is also used to upload map data to the cloud to establish an augmented reality scene corresponding to the real scene.
本公开示例性实施方式还提供另一种虚拟影像实现装置,可以设置于上述第一终端110。如图8所示,该虚拟影像实现装置800可以包括:Exemplary embodiments of the present disclosure also provide another virtual image realization device, which can be set in the above-mentioned first terminal 110. As shown in FIG. 8, the virtual image realization device 800 may include:
获取模块810,用于获取用户输入的针对目标物体的编辑信息;The obtaining module 810 is used to obtain edit information for the target object input by the user;
生成模块820,用于根据上述编辑信息生成目标物体的虚拟影像;The generating module 820 is configured to generate a virtual image of the target object according to the above-mentioned editing information;
上传模块830,用于将上述虚拟影像上传至增强现实场景中,该增强现实场景存储于云端,使第二终端通过从云端获取增强现实场景而显示虚拟影像。The upload module 830 is configured to upload the above-mentioned virtual image to an augmented reality scene, which is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image.
在一种可选的实施方式中,获取模块810,还用于当从现实场景中扫描到目标物体时,触发显示编辑界面,编辑界面用于输入编辑信息。In an optional implementation manner, the acquisition module 810 is further configured to trigger the display of an editing interface when a target object is scanned from a real scene, and the editing interface is used to input editing information.
在一种可选的实施方式中,上传模块830,还用于在上传虚拟影像时,从所采集的目标物体的图像中选取至少一关键帧图像,上传至增强现实场景中。In an optional implementation manner, the upload module 830 is further configured to select at least one key frame image from the collected images of the target object when uploading the virtual image, and upload it to the augmented reality scene.
在一种可选的实施方式中,生成模块820,还用于根据目标物体在增强现实场景中的位置,生成虚拟影像的提示信息,提示信息用于提示虚拟影像的方位。上传模块730,还用于将提示信息上传至增强现实场景中。In an optional implementation manner, the generating module 820 is further configured to generate prompt information of the virtual image according to the position of the target object in the augmented reality scene, and the prompt information is used to prompt the orientation of the virtual image. The upload module 730 is also used to upload the prompt information to the augmented reality scene.
在一种可选的实施方式中,生成模块820,被配置为针对增强现实场景中的每个区域,生成由该区域指向虚拟影像的虚拟指示箭头。In an optional implementation manner, the generating module 820 is configured to generate a virtual indicating arrow pointing from the area to the virtual image for each area in the augmented reality scene.
在一种可选的实施方式中,上述编辑信息可以包括文本、图片、手绘图画、动画效果、视频、音频中的任意一种或多种;此外,上述编辑信息还可以包括虚拟影像的显示位置。In an optional implementation manner, the above-mentioned editing information may include any one or more of text, pictures, hand-drawn drawings, animation effects, video, and audio; in addition, the above-mentioned editing information may also include the display position of the virtual image. .
在一种可选的实施方式中,生成模块820,还用于扫描目标物体所在的现实场景,生成地图数据。上传模块830,还用于将地图数据上传至云端,以建立现实场景对应的增强现实场景。In an optional implementation manner, the generating module 820 is also used to scan the actual scene where the target object is located to generate map data. The upload module 830 is also used to upload map data to the cloud to establish an augmented reality scene corresponding to the real scene.
本公开示例性实施方式还提供一种虚拟影像实现装置,可以设置于上述第二终端120。如图9所示,该虚拟影像实现装置900可以包括处理器910与存储器920。其中,存储器920存储有以下程序模块:Exemplary embodiments of the present disclosure also provide a virtual image realization device, which may be installed in the above-mentioned second terminal 120. As shown in FIG. 9, the virtual image realization device 900 may include a processor 910 and a memory 920. Among them, the memory 920 stores the following program modules:
获取模块921,用于从云端获取增强现实场景,该增强现实场景至少包括一目标物体的虚拟影像,该虚拟影像由第一终端上传至增强现实场景中;The acquiring module 921 is configured to acquire an augmented reality scene from the cloud, the augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal;
重定位模块922,用于基于增强现实场景进行重定位,当重定位成功时,显示虚拟影像或者虚拟影像的提示信息;The relocation module 922 is used to perform relocation based on the augmented reality scene, and when the relocation is successful, display a virtual image or a prompt message of the virtual image;
处理器910用于执行上述程序模块。The processor 910 is configured to execute the foregoing program modules.
在一种可选的实施方式中,重定位模块922,被配置为:In an optional implementation manner, the relocation module 922 is configured to:
从增强现实场景中获取目标物体的关键帧图像;Obtain the key frame image of the target object from the augmented reality scene;
将采集的当前图像与关键帧图像进行匹配,当匹配成功时,显示虚拟影像。Match the collected current image with the key frame image, and display a virtual image when the matching is successful.
在一种可选的实施方式中,重定位模块922,被配置为:In an optional implementation manner, the relocation module 922 is configured to:
基于增强现实场景进行重定位,当重定位成功时,确定第二终端在增强现实场景中的当前位姿;Perform relocation based on the augmented reality scene, and when the relocation is successful, determine the current pose of the second terminal in the augmented reality scene;
根据当前位姿确定第二终端的当前视野区域;Determine the current field of view area of the second terminal according to the current pose;
当虚拟影像的显示位置处于当前视野区域内时,显示虚拟影像;When the display position of the virtual image is within the current view area, display the virtual image;
当虚拟影像的显示位置处于当前视野区域外时,显示虚拟影像的提示信息。When the display position of the virtual image is outside the current field of view, prompt information of the virtual image is displayed.
在一种可选的实施方式中,重定位模块922,被配置为:In an optional implementation manner, the relocation module 922 is configured to:
将采集的当前图像与增强现实场景的地图数据进行匹配,根据匹配结果确定重定位结果。Match the collected current image with the map data of the augmented reality scene, and determine the relocation result according to the matching result.
在一种可选的实施方式中,重定位模块922,被配置为:In an optional implementation manner, the relocation module 922 is configured to:
从增强现实场景中获取虚拟影像的显示位置;或者将目标物体的位置确定为虚拟影像的显示位置。Obtain the display position of the virtual image from the augmented reality scene; or determine the position of the target object as the display position of the virtual image.
在一种可选的实施方式中,虚拟影像实现装置900还可以包括:In an optional implementation manner, the virtual image realization apparatus 900 may further include:
修改模块,用于获取用户针对虚拟影像的修改信息,根据修改信息得到新的虚拟影像,以及将新的虚拟影像上传至增强现实场景中,以更新云端上存储的虚拟影像。The modification module is used to obtain the user's modification information for the virtual image, obtain a new virtual image according to the modification information, and upload the new virtual image to the augmented reality scene to update the virtual image stored on the cloud.
本公开示例性实施方式还提供另一种虚拟影像实现装置,可以设置于上述第二终端120。如图10所示,该虚拟影像实现装置1000可以包括:Exemplary embodiments of the present disclosure also provide another virtual image realization device, which can be set in the above-mentioned second terminal 120. As shown in FIG. 10, the virtual image realization device 1000 may include:
获取模块1010,用于从云端获取增强现实场景,该增强现实场景至少包括一目标物体的虚拟影像,该虚拟影像由第一终端上传至增强现实场景中;The acquiring module 1010 is configured to acquire an augmented reality scene from the cloud, the augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal;
重定位模块1020,用于基于增强现实场景进行重定位,当重定位成功时,显示虚拟影像或者虚拟影像的提示信息。The relocation module 1020 is used to perform relocation based on the augmented reality scene, and when the relocation is successful, display a virtual image or a prompt message of the virtual image.
在一种可选的实施方式中,重定位模块1020,被配置为:In an optional implementation manner, the relocation module 1020 is configured to:
从增强现实场景中获取目标物体的关键帧图像;Obtain the key frame image of the target object from the augmented reality scene;
将采集的当前图像与关键帧图像进行匹配,当匹配成功时,显示虚拟影像。Match the collected current image with the key frame image, and display a virtual image when the matching is successful.
在一种可选的实施方式中,重定位模块1020,被配置为:In an optional implementation manner, the relocation module 1020 is configured to:
基于增强现实场景进行重定位,当重定位成功时,确定第二终端在增强现实场景中的当前位姿;Perform relocation based on the augmented reality scene, and when the relocation is successful, determine the current pose of the second terminal in the augmented reality scene;
根据当前位姿确定第二终端的当前视野区域;Determine the current field of view area of the second terminal according to the current pose;
当虚拟影像的显示位置处于当前视野区域内时,显示虚拟影像;When the display position of the virtual image is within the current view area, display the virtual image;
当虚拟影像的显示位置处于当前视野区域外时,显示虚拟影像的提示信息。When the display position of the virtual image is outside the current field of view, prompt information of the virtual image is displayed.
在一种可选的实施方式中,重定位模块1020,被配置为:In an optional implementation manner, the relocation module 1020 is configured to:
将采集的当前图像与增强现实场景的地图数据进行匹配,根据匹配结果确定重定位结果。Match the collected current image with the map data of the augmented reality scene, and determine the relocation result according to the matching result.
在一种可选的实施方式中,重定位模块1020,被配置为:In an optional implementation manner, the relocation module 1020 is configured to:
从增强现实场景中获取虚拟影像的显示位置;或者将目标物体的位置确定为虚拟影像的显示位置。Obtain the display position of the virtual image from the augmented reality scene; or determine the position of the target object as the display position of the virtual image.
在一种可选的实施方式中,虚拟影像实现装置1000还可以包括:In an optional implementation manner, the virtual image realization apparatus 1000 may further include:
修改模块,用于获取用户针对虚拟影像的修改信息,根据修改信息得到新的虚拟影像,以及将新的虚拟影像上传至增强现实场景中,以更新云端上存储的虚拟影像。The modification module is used to obtain the user's modification information for the virtual image, obtain a new virtual image according to the modification information, and upload the new virtual image to the augmented reality scene to update the virtual image stored on the cloud.
上述虚拟影像实现装置700、虚拟影像实现装置800、虚拟影像实现装置900与虚拟影像实现装置1000中各部分的具体细节在方法部分实施方式中已经详细说明,未披露的 细节内容可以参见方法部分的实施方式内容,因而不再赘述。The specific details of each part of the above-mentioned virtual image realization device 700, virtual image realization device 800, virtual image realization device 900 and virtual image realization device 1000 have been described in detail in the method part implementation. For undisclosed details, please refer to the method part. The content of the implementation mode will not be repeated here.
本公开的示例性实施方式还提供了一种计算机可读存储介质,可以实现为一种程序产品的形式,其包括程序代码,当程序产品在终端设备上运行时,程序代码用于使终端设备执行本说明书上述“示例性方法”部分中描述的根据本公开各种示例性实施方式的步骤。该程序产品可以采用便携式紧凑盘只读存储器(CD-ROM)并包括程序代码,并可以在终端设备,例如个人电脑上运行。然而,本公开的程序产品不限于此,在本文件中,可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。Exemplary embodiments of the present disclosure also provide a computer-readable storage medium, which can be implemented in the form of a program product, which includes program code. When the program product runs on a terminal device, the program code is used to make the terminal device Perform the steps according to various exemplary embodiments of the present disclosure described in the above-mentioned "Exemplary Method" section of this specification. The program product can adopt a portable compact disk read-only memory (CD-ROM) and include program code, and can run on a terminal device, such as a personal computer. However, the program product of the present disclosure is not limited thereto. In this document, the readable storage medium can be any tangible medium that contains or stores a program, and the program can be used by or in combination with an instruction execution system, device, or device.
程序产品可以采用一个或多个可读介质的任意组合。可读介质可以是可读信号介质或者可读存储介质。可读存储介质例如可以为但不限于电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。The program product can adopt any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or a combination of any of the above. More specific examples (non-exhaustive list) of readable storage media include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Type programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了可读程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。可读信号介质还可以是可读存储介质以外的任何可读介质,该可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。The computer-readable signal medium may include a data signal propagated in baseband or as a part of a carrier wave, and readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing. The readable signal medium may also be any readable medium other than a readable storage medium, and the readable medium may send, propagate, or transmit a program for use by or in combination with the instruction execution system, apparatus, or device.
可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于无线、有线、光缆、RF等等,或者上述的任意合适的组合。The program code contained on the readable medium can be transmitted by any suitable medium, including but not limited to wireless, wired, optical cable, RF, etc., or any suitable combination of the foregoing.
可以以一种或多种程序设计语言的任意组合来编写用于执行本公开操作的程序代码,程序设计语言包括面向对象的程序设计语言—诸如Java、C++等,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算设备上执行、部分地在用户设备上执行、作为一个独立的软件包执行、部分在用户计算设备上部分在远程计算设备上执行、或者完全在远程计算设备或服务器上执行。在涉及远程计算设备的情形中,远程计算设备可以通过任意种类的网络,包括局域网(LAN)或广域网(WAN),连接到用户计算设备,或者,可以连接到外部计算设备(例如利用因特网服务提供商来通过因特网连接)。The program code for performing the operations of the present disclosure can be written in any combination of one or more programming languages. The programming languages include object-oriented programming languages—such as Java, C++, etc., as well as conventional procedural programming. Language-such as "C" language or similar programming language. The program code can be executed entirely on the user's computing device, partly on the user's device, executed as an independent software package, partly on the user's computing device and partly executed on the remote computing device, or entirely on the remote computing device or server Executed on. In the case of a remote computing device, the remote computing device can be connected to a user computing device through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computing device (for example, using Internet service providers). Business to connect via the Internet).
应当注意,尽管在上文详细描述中提及了用于动作执行的设备的若干模块或者单元,但是这种划分并非强制性的。实际上,根据本公开的示例性实施方式,上文描述的两个或更多模块或者单元的特征和功能可以在一个模块或者单元中具体化。反之,上文描述的一个模块或者单元的特征和功能可以进一步划分为由多个模块或者单元来具体化。It should be noted that although several modules or units of the device for action execution are mentioned in the above detailed description, this division is not mandatory. In fact, according to the exemplary embodiments of the present disclosure, the features and functions of two or more modules or units described above may be embodied in one module or unit. Conversely, the features and functions of a module or unit described above can be further divided into multiple modules or units to be embodied.
所属技术领域的技术人员能够理解,本公开的各个方面可以实现为系统、方法或程序产品。因此,本公开的各个方面可以具体实现为以下形式,即:完全的硬件实施方式、完 全的软件实施方式(包括固件、微代码等),或硬件和软件方面结合的实施方式,这里可以统称为“电路”、“模块”或“系统”。本领域技术人员在考虑说明书及实践这里公开的发明后,将容易想到本公开的其他实施方式。本申请旨在涵盖本公开的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本公开的一般性原理并包括本公开未公开的本技术领域中的公知常识或惯用技术手段。说明书和实施方式仅被视为示例性的,本公开的真正范围和精神由权利要求指出。Those skilled in the art can understand that various aspects of the present disclosure can be implemented as a system, a method, or a program product. Therefore, various aspects of the present disclosure can be specifically implemented in the following forms, namely: complete hardware implementation, complete software implementation (including firmware, microcode, etc.), or a combination of hardware and software implementations, which may be collectively referred to herein as "Circuit", "Module" or "System". Those skilled in the art will easily think of other embodiments of the present disclosure after considering the specification and practicing the invention disclosed herein. This application is intended to cover any variations, uses, or adaptive changes of the present disclosure. These variations, uses, or adaptive changes follow the general principles of the present disclosure and include common knowledge or conventional technical means in the technical field that are not disclosed in the present disclosure. . The description and the embodiments are only regarded as exemplary, and the true scope and spirit of the present disclosure are pointed out by the claims.
应当理解的是,本公开并不局限于上面已经描述并在附图中示出的精确结构,并且可以在不脱离其范围进行各种修改和改变。本公开的范围仅由所附的权利要求来限定。It should be understood that the present disclosure is not limited to the precise structure that has been described above and shown in the drawings, and various modifications and changes can be made without departing from its scope. The scope of the present disclosure is limited only by the appended claims.

Claims (20)

  1. 一种虚拟影像实现方法,应用于第一终端,其特征在于,所述方法包括:A virtual image realization method applied to a first terminal, characterized in that the method includes:
    获取用户输入的针对目标物体的编辑信息;Obtain the editing information for the target object input by the user;
    根据所述编辑信息,生成所述目标物体的虚拟影像;Generating a virtual image of the target object according to the editing information;
    将所述虚拟影像上传至所述目标物体所在的增强现实场景中,所述增强现实场景存储于云端,使第二终端通过从所述云端获取所述增强现实场景而显示所述虚拟影像。The virtual image is uploaded to the augmented reality scene where the target object is located, and the augmented reality scene is stored in the cloud, so that the second terminal obtains the augmented reality scene from the cloud to display the virtual image.
  2. 根据权利要求1所述的方法,其特征在于,在获取所述编辑信息前,所述方法还包括:The method according to claim 1, characterized in that, before obtaining the editing information, the method further comprises:
    当从现实场景中扫描到所述目标物体时,触发显示编辑界面,所述编辑界面用于输入编辑信息。When the target object is scanned from the real scene, the display of the editing interface is triggered, and the editing interface is used to input editing information.
  3. 根据权利要求2所述的方法,其特征在于,在上传所述虚拟影像时,所述方法还包括:The method according to claim 2, wherein when uploading the virtual image, the method further comprises:
    从所采集的所述目标物体的图像中选取至少一关键帧图像,上传至所述增强现实场景。At least one key frame image is selected from the collected images of the target object and uploaded to the augmented reality scene.
  4. 根据权利要求1所述的方法,其特征在于,所述方法还包括:The method according to claim 1, wherein the method further comprises:
    根据所述目标物体在所述增强现实场景中的位置,生成所述虚拟影像的提示信息,所述提示信息用于提示所述虚拟影像的方位;Generating prompt information of the virtual image according to the position of the target object in the augmented reality scene, where the prompt information is used to prompt the orientation of the virtual image;
    将所述提示信息上传至所述增强现实场景中。Upload the prompt information to the augmented reality scene.
  5. 根据权利要求4所述的方法,其特征在于,所述根据所述目标物体在所述增强现实场景中的位置,生成所述虚拟影像的提示信息,包括:The method according to claim 4, wherein the generating the prompt information of the virtual image according to the position of the target object in the augmented reality scene comprises:
    针对所述增强现实场景中的每个区域,生成由所述区域指向所述虚拟影像的虚拟指示箭头。For each area in the augmented reality scene, a virtual indicator arrow pointing from the area to the virtual image is generated.
  6. 根据权利要求1所述的方法,其特征在于,所述编辑信息包括文本、图片、手绘图画、动画效果、视频、音频中的任意一种或多种,所述编辑信息还包括所述虚拟影像的显示位置。The method according to claim 1, wherein the editing information includes any one or more of text, pictures, hand-drawn drawings, animation effects, video, and audio, and the editing information further includes the virtual image Display position.
  7. 根据权利要求1所述的方法,其特征在于,所述方法还包括:The method according to claim 1, wherein the method further comprises:
    扫描所述目标物体所在的现实场景,生成地图数据;Scanning the real scene where the target object is located to generate map data;
    将所述地图数据上传至所述云端,以建立所述现实场景对应的增强现实场景。Upload the map data to the cloud to establish an augmented reality scene corresponding to the real scene.
  8. 一种虚拟影像实现方法,应用于第二终端,其特征在于,所述方法包括:A virtual image realization method applied to a second terminal, characterized in that the method includes:
    从云端获取增强现实场景,所述增强现实场景至少包括一目标物体的虚拟影像,所述虚拟影像由第一终端上传至所述增强现实场景中;Acquiring an augmented reality scene from the cloud, where the augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal;
    基于所述增强现实场景进行重定位,当重定位成功时,显示所述虚拟影像或者所述虚拟影像的提示信息。Relocation is performed based on the augmented reality scene, and when the relocation is successful, the virtual image or prompt information of the virtual image is displayed.
  9. 根据权利要求8所述的方法,其特征在于,所述基于所述增强现实场景进行重定位,当重定位成功时,显示所述虚拟影像或者所述虚拟影像的提示信息,包括:The method according to claim 8, wherein the performing relocation based on the augmented reality scene, and when the relocation is successful, displaying the virtual image or the prompt information of the virtual image comprises:
    从所述增强现实场景中获取所述目标物体的关键帧图像;Acquiring a key frame image of the target object from the augmented reality scene;
    将采集的当前图像与所述关键帧图像进行匹配,当匹配成功时,显示所述虚拟影像。Match the collected current image with the key frame image, and when the matching is successful, display the virtual image.
  10. 根据权利要求8所述的方法,其特征在于,所述基于所述增强现实场景进行重定位,当重定位成功时,显示所述虚拟影像或者所述虚拟影像的提示信息,包括:The method according to claim 8, wherein the performing relocation based on the augmented reality scene, and when the relocation is successful, displaying the virtual image or the prompt information of the virtual image comprises:
    基于所述增强现实场景进行重定位,当重定位成功时,确定所述第二终端在所述增强现实场景中的当前位姿;Perform relocation based on the augmented reality scene, and when the relocation is successful, determine the current pose of the second terminal in the augmented reality scene;
    根据所述当前位姿确定所述第二终端的当前视野区域;Determining the current view area of the second terminal according to the current pose;
    当所述虚拟影像的显示位置处于所述当前视野区域内时,显示所述虚拟影像;Displaying the virtual image when the display position of the virtual image is within the current view area;
    当所述虚拟影像的显示位置处于所述当前视野区域外时,显示所述虚拟影像的提示信息。When the display position of the virtual image is outside the current view area, display prompt information of the virtual image.
  11. 根据权利要求10所述的方法,其特征在于,所述基于所述增强现实场景进行重定位,包括:The method according to claim 10, wherein the relocation based on the augmented reality scene comprises:
    将采集的当前图像与所述增强现实场景的地图数据进行匹配,根据匹配结果确定重定位结果。Match the collected current image with the map data of the augmented reality scene, and determine the relocation result according to the matching result.
  12. 根据权利要求8所述的方法,其特征在于,所述方法还包括:The method according to claim 8, wherein the method further comprises:
    从所述增强现实场景中获取所述虚拟影像的显示位置;或者Acquiring the display position of the virtual image from the augmented reality scene; or
    将所述目标物体的位置确定为所述虚拟影像的显示位置。The position of the target object is determined as the display position of the virtual image.
  13. 根据权利要求8所述的方法,其特征在于,所述方法还包括:The method according to claim 8, wherein the method further comprises:
    获取用户针对所述虚拟影像的修改信息,根据所述修改信息得到新的虚拟影像;Acquiring user modification information for the virtual image, and obtaining a new virtual image according to the modification information;
    将所述新的虚拟影像上传至所述增强现实场景中,以更新所述云端上存储的所述虚拟影像。Uploading the new virtual image to the augmented reality scene to update the virtual image stored on the cloud.
  14. 一种虚拟影像实现装置,设置于第一终端,其特征在于,所述装置包括处理器,所述处理器用于执行存储器中存储的以下程序模块:A virtual image realization device is provided in a first terminal, and is characterized in that the device includes a processor, and the processor is configured to execute the following program modules stored in a memory:
    获取模块,用于获取用户输入的针对目标物体的编辑信息;The obtaining module is used to obtain the editing information for the target object input by the user;
    生成模块,用于根据所述编辑信息生成所述目标物体的虚拟影像;A generating module, configured to generate a virtual image of the target object according to the editing information;
    上传模块,用于将所述虚拟影像上传至所述目标物体所在的增强现实场景中,所述增强现实场景存储于云端,使第二终端通过从所述云端获取所述增强现实场景而显示所述虚拟影像。The upload module is configured to upload the virtual image to the augmented reality scene where the target object is located, and the augmented reality scene is stored in the cloud, so that the second terminal displays the augmented reality scene by acquiring the augmented reality scene from the cloud. The virtual image.
  15. 根据权利要求14所述的装置,其特征在于,所述生成模块,还用于根据所述目标物体在所述增强现实场景中的位置,生成所述虚拟影像的提示信息,所述提示信息用于提示所述虚拟影像的方位;The device according to claim 14, wherein the generating module is further configured to generate prompt information of the virtual image according to the position of the target object in the augmented reality scene, and the prompt information is used To prompt the orientation of the virtual image;
    所述上传模块,还用于将所述提示信息上传至所述增强现实场景中。The upload module is also used to upload the prompt information to the augmented reality scene.
  16. 一种虚拟影像实现装置,设置于第二终端,其特征在于,所述装置包括处理器,所述处理器用于执行存储器中存储的以下程序模块:A device for realizing a virtual image, which is provided in a second terminal, is characterized in that the device includes a processor, and the processor is configured to execute the following program modules stored in a memory:
    获取模块,用于从云端获取增强现实场景,所述增强现实场景至少包括一目标物体的虚拟影像,所述虚拟影像由第一终端上传至所述增强现实场景中;An acquisition module, configured to acquire an augmented reality scene from the cloud, the augmented reality scene includes at least a virtual image of a target object, and the virtual image is uploaded to the augmented reality scene by the first terminal;
    重定位模块,用于基于所述增强现实场景进行重定位,当重定位成功时,显示所述虚拟影像或者所述虚拟影像的提示信息。The relocation module is configured to perform relocation based on the augmented reality scene, and when the relocation is successful, display the virtual image or prompt information of the virtual image.
  17. 根据权利要求16所述的装置,其特征在于,所述重定位模块,被配置为:The device according to claim 16, wherein the relocation module is configured to:
    从所述增强现实场景中获取所述目标物体的关键帧图像;Acquiring a key frame image of the target object from the augmented reality scene;
    将采集的当前图像与所述关键帧图像进行匹配,当匹配成功时,显示所述虚拟影像。Match the collected current image with the key frame image, and when the matching is successful, display the virtual image.
  18. 根据权利要求16所述的装置,其特征在于,所述重定位模块,被配置为:The device according to claim 16, wherein the relocation module is configured to:
    基于所述增强现实场景进行重定位,当重定位成功时,确定所述第二终端在所述增强现实场景中的当前位姿;Perform relocation based on the augmented reality scene, and when the relocation is successful, determine the current pose of the second terminal in the augmented reality scene;
    根据所述当前位姿确定所述第二终端的当前视野区域;Determining the current view area of the second terminal according to the current pose;
    当所述虚拟影像的显示位置处于所述当前视野区域内时,显示所述虚拟影像;Displaying the virtual image when the display position of the virtual image is within the current view area;
    当所述虚拟影像的显示位置处于所述当前视野区域外时,显示所述虚拟影像的提示信息。When the display position of the virtual image is outside the current view area, display prompt information of the virtual image.
  19. 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现权利要求1至13任一项所述的方法。A computer-readable storage medium having a computer program stored thereon, wherein the computer program implements the method according to any one of claims 1 to 13 when the computer program is executed by a processor.
  20. 一种终端设备,其特征在于,包括:A terminal device, characterized in that it comprises:
    处理器;processor;
    存储器,用于存储所述处理器的可执行指令;以及A memory for storing executable instructions of the processor; and
    显示器;monitor;
    其中,所述处理器配置为经由执行所述可执行指令来执行权利要求1至13任一项所述的方法。Wherein, the processor is configured to execute the method according to any one of claims 1 to 13 by executing the executable instructions.
PCT/CN2021/080040 2020-04-08 2021-03-10 Virtual image realization method and apparatus, and storage medium and terminal device WO2021203902A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010269991.9 2020-04-08
CN202010269991.9A CN111476911B (en) 2020-04-08 2020-04-08 Virtual image realization method, device, storage medium and terminal equipment

Publications (1)

Publication Number Publication Date
WO2021203902A1 true WO2021203902A1 (en) 2021-10-14

Family

ID=71750040

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/080040 WO2021203902A1 (en) 2020-04-08 2021-03-10 Virtual image realization method and apparatus, and storage medium and terminal device

Country Status (2)

Country Link
CN (1) CN111476911B (en)
WO (1) WO2021203902A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114363161A (en) * 2022-01-11 2022-04-15 中国工商银行股份有限公司 Abnormal equipment positioning method, device, equipment and medium
CN115049804A (en) * 2022-05-16 2022-09-13 北京微视威信息科技有限公司 Editing method, device, equipment and medium for virtual scene
WO2023071381A1 (en) * 2021-10-29 2023-05-04 深圳市慧鲤科技有限公司 Virtual-real fusion method and apparatus, and device, storage medium and program product

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111476911B (en) * 2020-04-08 2023-07-25 Oppo广东移动通信有限公司 Virtual image realization method, device, storage medium and terminal equipment
CN112070906A (en) * 2020-08-31 2020-12-11 北京市商汤科技开发有限公司 Augmented reality system and augmented reality data generation method and device
CN112070907A (en) * 2020-08-31 2020-12-11 北京市商汤科技开发有限公司 Augmented reality system and augmented reality data generation method and device
CN112053450B (en) * 2020-09-10 2024-07-16 脸萌有限公司 Text display method and device, electronic equipment and storage medium
CN112330819B (en) * 2020-11-04 2024-02-06 腾讯科技(深圳)有限公司 Interaction method and device based on virtual article and storage medium
CN112699331A (en) * 2020-12-31 2021-04-23 深圳市慧鲤科技有限公司 Message information display method and device, electronic equipment and storage medium
CN113724398A (en) * 2021-09-01 2021-11-30 北京百度网讯科技有限公司 Augmented reality method, apparatus, device and storage medium
CN113989470A (en) * 2021-11-15 2022-01-28 北京有竹居网络技术有限公司 Picture display method and device, storage medium and electronic equipment
CN114092675A (en) * 2021-11-22 2022-02-25 北京百度网讯科技有限公司 Image display method, image display device, electronic apparatus, and storage medium
CN114332417B (en) * 2021-12-13 2023-07-14 亮风台(上海)信息科技有限公司 Method, equipment, storage medium and program product for interaction of multiple scenes
CN114283263A (en) * 2021-12-27 2022-04-05 深圳市慧鲤科技有限公司 Virtual display method, device and related equipment and storage medium
CN114419292A (en) * 2022-01-21 2022-04-29 北京字跳网络技术有限公司 Image processing method, device, equipment and storage medium
CN115174993B (en) * 2022-08-09 2024-02-13 北京字跳网络技术有限公司 Method, apparatus, device and storage medium for video production
CN118524247A (en) * 2023-02-13 2024-08-20 华为云计算技术有限公司 Synchronous display method, electronic equipment, server and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110055135A1 (en) * 2009-08-26 2011-03-03 International Business Machines Corporation Deferred Teleportation or Relocation in Virtual Worlds
CN103853913A (en) * 2012-12-03 2014-06-11 三星电子株式会社 Method for operating augmented reality contents and device and system for supporting the same
CN108510597A (en) * 2018-03-09 2018-09-07 北京小米移动软件有限公司 Edit methods, device and the non-transitorycomputer readable storage medium of virtual scene
CN110716645A (en) * 2019-10-15 2020-01-21 北京市商汤科技开发有限公司 Augmented reality data presentation method and device, electronic equipment and storage medium
CN111476911A (en) * 2020-04-08 2020-07-31 Oppo广东移动通信有限公司 Virtual image implementation method and device, storage medium and terminal equipment

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160133230A1 (en) * 2014-11-11 2016-05-12 Bent Image Lab, Llc Real-time shared augmented reality experience
CN106033333A (en) * 2015-03-10 2016-10-19 沈阳中云普华科技有限公司 A visual augmented reality scene making system and method
CN106155311A (en) * 2016-06-28 2016-11-23 努比亚技术有限公司 AR helmet, AR interactive system and the exchange method of AR scene
CN106683201A (en) * 2016-12-23 2017-05-17 深圳市豆娱科技有限公司 Scene editing method and device based on three-dimensional virtual reality
CN109117034A (en) * 2017-06-23 2019-01-01 卢俊谚 The editing device and system of augmented reality are integrated on line
CN107808416A (en) * 2017-09-29 2018-03-16 北京金山安全软件有限公司 Image processing method, image display method, image processing device, image display device, electronic equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110055135A1 (en) * 2009-08-26 2011-03-03 International Business Machines Corporation Deferred Teleportation or Relocation in Virtual Worlds
CN103853913A (en) * 2012-12-03 2014-06-11 三星电子株式会社 Method for operating augmented reality contents and device and system for supporting the same
CN108510597A (en) * 2018-03-09 2018-09-07 北京小米移动软件有限公司 Edit methods, device and the non-transitorycomputer readable storage medium of virtual scene
CN110716645A (en) * 2019-10-15 2020-01-21 北京市商汤科技开发有限公司 Augmented reality data presentation method and device, electronic equipment and storage medium
CN111476911A (en) * 2020-04-08 2020-07-31 Oppo广东移动通信有限公司 Virtual image implementation method and device, storage medium and terminal equipment

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023071381A1 (en) * 2021-10-29 2023-05-04 深圳市慧鲤科技有限公司 Virtual-real fusion method and apparatus, and device, storage medium and program product
CN114363161A (en) * 2022-01-11 2022-04-15 中国工商银行股份有限公司 Abnormal equipment positioning method, device, equipment and medium
CN114363161B (en) * 2022-01-11 2024-03-22 中国工商银行股份有限公司 Abnormal equipment positioning method, device, equipment and medium
CN115049804A (en) * 2022-05-16 2022-09-13 北京微视威信息科技有限公司 Editing method, device, equipment and medium for virtual scene
CN115049804B (en) * 2022-05-16 2023-08-15 北京微视威信息科技有限公司 Editing method, device, equipment and medium for virtual scene

Also Published As

Publication number Publication date
CN111476911A (en) 2020-07-31
CN111476911B (en) 2023-07-25

Similar Documents

Publication Publication Date Title
WO2021203902A1 (en) Virtual image realization method and apparatus, and storage medium and terminal device
WO2020244495A1 (en) Screen projection display method and electronic device
WO2021213120A1 (en) Screen projection method and apparatus, and electronic device
WO2020244492A1 (en) Screen projection display method and electronic device
EP4080464A1 (en) Augmented reality processing method, apparatus and system, and storage medium, and electronic device
WO2020062267A1 (en) Information prompt method and electronic device
WO2020186988A1 (en) Information display method and device, terminal, and storage medium
WO2021184952A1 (en) Augmented reality processing method and apparatus, storage medium, and electronic device
WO2019114514A1 (en) Method and apparatus for displaying pitch information in live broadcast room, and storage medium
WO2019128592A1 (en) Method and apparatus for live broadcasting
CN112118477B (en) Virtual gift display method, device, equipment and storage medium
JP7173670B2 (en) VOICE CONTROL COMMAND GENERATION METHOD AND TERMINAL
US11682148B2 (en) Method for displaying advertisement picture, method for uploading advertisement picture, and apparatus
WO2021213031A1 (en) Image synthesis method and related apparatus
WO2021197020A1 (en) Audio processing method and apparatus, readable medium, and electronic device
WO2022148319A1 (en) Video switching method and apparatus, storage medium, and device
CN114185503B (en) Multi-screen interaction system, method, device and medium
CN114697742A (en) Video recording method and electronic equipment
WO2023284591A1 (en) Video capture method and apparatus, electronic device, and storage medium
WO2022127670A1 (en) Call method and system, and related device
CN112269554B (en) Display system and display method
CN114449090B (en) Data sharing method, device and system, electronic equipment, storage medium and chip
CN113747056A (en) Photographing method and device and electronic equipment
CN113542575A (en) Device pose adjusting method, image shooting method and electronic device
WO2020051916A1 (en) Method for transmitting information and electronic device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21785259

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21785259

Country of ref document: EP

Kind code of ref document: A1