WO2015154627A1 - 一种虚拟现实组件系统 - Google Patents

一种虚拟现实组件系统 Download PDF

Info

Publication number
WO2015154627A1
WO2015154627A1 PCT/CN2015/075213 CN2015075213W WO2015154627A1 WO 2015154627 A1 WO2015154627 A1 WO 2015154627A1 CN 2015075213 W CN2015075213 W CN 2015075213W WO 2015154627 A1 WO2015154627 A1 WO 2015154627A1
Authority
WO
WIPO (PCT)
Prior art keywords
sensor
control unit
controller
display control
user
Prior art date
Application number
PCT/CN2015/075213
Other languages
English (en)
French (fr)
Inventor
覃政
Original Assignee
北京蚁视科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京蚁视科技有限公司 filed Critical 北京蚁视科技有限公司
Publication of WO2015154627A1 publication Critical patent/WO2015154627A1/zh

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/25Output arrangements for video game devices
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/428Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/80Special adaptations for executing a specific game genre or game mode
    • A63F13/837Shooting of targets

Definitions

  • the present invention relates to the field of virtual reality, and in particular, to a system for implementing a virtual reality scene.
  • Virtual Reality is a high-tech technology that has emerged in recent years.
  • Virtual reality is the use of computer simulation to generate a virtual world in a three-dimensional space, providing users with simulations of visual, auditory, tactile and other senses, allowing users to observe things in the three-dimensional space in a timely and unrestricted manner.
  • Virtual reality technology covers real-time 3D computer graphics technology, wide-angle (wide-view) stereo display technology, tracking sensor technology for observers' heads, eyes and hands, as well as tactile/force feedback, stereo, network transmission, voice input and output.
  • a user typically uses a controller and various sensors to track various movements of the head and the body by wearing a head-mounted display, thereby allowing the user to experience an immersive experience.
  • the existing product adopts a scheme in which the display portion and the controller portion are detachable, that is, detachable under the operation of the user.
  • the board is separately mounted in the control box, and then the controller is connected to the head mounted display by a cable or the like.
  • this will bring great inconvenience to the user's use.
  • control box Since the control box needs to be placed separately, for example, on a table or placed on the ground, it is not convenient for the user to hold it, so that the scope of the user's activities and its limitations are hindered and affected. In addition, the connection between the control box and the head mounted display also hinders the user's movement, causing inconvenience and reducing the user experience.
  • first-person perspective shooting game many games are first-person perspective, such as a first-person perspective shooting game (FPS).
  • FPS first-person perspective shooting game
  • the existing FPS is not realistic enough in the case of using virtual reality technology to simulate the first person perspective.
  • the existing game control controls the front, rear, left and right movement of the character through the button or the joystick, which is far from true in the environment.
  • the virtual reality device uses a computer vision method to track users, thereby implementing motion tracking. This solution system is large and complex, and is not easy to promote for civilian use.
  • It is an object of the present invention to provide a virtual reality system comprising a near-eye display, a display control unit, and a controller, the near-eye display being separable from the display control unit, wherein the near-eye display comprises a first sensor, a three-axis attitude and/or a three-axis position for sensing the near-eye display;
  • the display control unit includes a first connector, a communication module, and a power supply module, wherein the first connector is for connecting the controller
  • the communication module is configured to receive and transmit data/control signals in a wired or wireless manner and transmit to the near-eye display and receive signals transmitted from the near-eye display, the power supply module being configured to be the display control unit Powering;
  • the controller includes a second connector for connecting the display control unit, and a second sensor for sensing a triaxial attitude of the controller and / or three-axis position.
  • the near-eye display and the display control unit are connected by wires and transmit data and/or power supply signals.
  • the display control unit includes an interface module for connecting to other display terminals or peripherals.
  • the first sensor and the second sensor are selected from the group consisting of at least one or a combination of an acceleration sensor, an angular velocity sensor, and a magnetic induction sensor.
  • the acceleration sensor is configured to sense whether the linear acceleration or the linear velocity of the controller in the inertial system reaches a certain threshold, and triggers the control signal.
  • the angular velocity sensor is configured to sense whether the angular acceleration or angular velocity of the controller in the inertial frame reaches a certain threshold, and then triggers a control signal.
  • an event composed of a plurality of instructions is triggered by recording a parameter sensed by the sensor at the trigger threshold time as an initial parameter.
  • the first sensor or the second sensor senses that the user generates a displacement or tilt with respect to a first direction in its own coordinate axis, and after the displacement or tilt exceeds a certain threshold, the sensor triggers an instruction to The motion control signal in the first direction is continuously emitted.
  • the control signal is triggered to stop the previous continuous movement in the first direction or to trigger the displacement control signal in the second direction.
  • the display control unit and the controller are detachable combination kits, and the combination thereof is a snap type or a magnetic type.
  • the combination kit is in the shape of a hand-held gun body, the display control unit is a barrel portion, and the controller is a handle portion.
  • a trip mechanism is further included for disengaging the display control unit from the controller under a pull of the user's one-handed finger toward the first direction.
  • the tripping mechanism is operative to trigger operation of the controller when the user is biased toward a second direction that is different from the first direction.
  • the tripping mechanism triggers an operation of the controller to turn on a throwing action of the sensing user.
  • the communication module of the wireless mode is selected from at least one of WiFi, Bluetooth, infrared, ultrasonic, Wireless USB, RFID, and NFC.
  • the power supply module is configured to supply power to the display control unit and/or the controller.
  • the power supply module sets a battery slot for accommodating the battery to pass through the body of the display control unit, so that when the user puts a new battery on the side of the body of the display control unit, the old one is simultaneously The battery is pushed out of the fuselage.
  • the display control unit further comprises at least one manipulation button.
  • the virtual reality system according to the present invention allows the user to comfortably and naturally manipulate the display
  • the display control unit section avoids the restriction of the external cable to the user's head.
  • the one-hand operation of the escape mechanism eliminates the need for the user's hands to operate, improving convenience.
  • the virtual reality system according to the present invention simulates the user operation of the throw by using the real swaying action, and controls the movement of the avatar by sensing the real movement of the user's body, thereby improving the user experience.
  • FIG. 1 is a schematic block diagram showing a virtual reality system according to the present invention.
  • FIG. 2 is a connection form of a display control unit package controller in accordance with one embodiment of the present invention.
  • FIG. 3 shows a product structure diagram of a near-eye display in accordance with an embodiment of the present invention.
  • FIG. 4 illustrates the process of triggering an instruction when a sensor in a controller senses a user's sway or throw action and reaches a certain threshold in accordance with one embodiment of the present invention.
  • FIG. 5 illustrates a process of triggering an instruction when a sensor in a near-eye display or controller senses a user's displacement action and reaches a certain threshold, in accordance with one embodiment of the present invention.
  • the virtual reality system 100 of the present invention includes a near-eye display 101, a display control unit 102, and a controller 103, wherein the near-eye display 101 is used to provide display on the user's head.
  • Function, near-eye display 101 and display control unit 102 are separate, and data and/or power supply signals can be connected and transmitted via connection 104.
  • the control unit 102 is a manipulation part held by the user for manipulating in the experience of the virtual reality, such as manipulating the movement of the training or the character in the game, and the like.
  • the display control unit 102 receives a control signal from the controller 103 to generate a display control signal for display by the near-eye display 101.
  • display control unit 102 and controller 103 may be provided as a detachable combination kit.
  • the combination kit can be configured, for example, in the shape of a hand-held gun body, the display control unit 102 is a barrel portion, and the controller 103 is a handle portion.
  • the barrel portion and the shank portion are combined by the connecting members 102a and 103a in the direction of the double arrow as shown in FIG.
  • the combination of the connectors 102a and 103a may be, for example, a snap type or a magnetic type.
  • a form of a male or female snap as shown in FIG.
  • combination kit can also be provided in other ways that are convenient for the user to hold, such as the joystick, handle, steering wheel.
  • the near-eye display 101 needs to be worn on the user's head for a long time in use, and therefore needs to be lightweight to ensure comfortable wearing.
  • the near-eye display 101 and the display control unit are preferably connected by a wired connection 104.
  • the connection 104 can not only transmit data and signals, but also transmit power signals through the power supply module 102e (described below) disposed in the display control unit 102 to avoid the increased weight of the near-eye display 101 requiring the configuration of additional power supply units.
  • the controller 103 includes a trip mechanism 103b for controlling tripping of the display control unit 102 and the controller 103.
  • the trip mechanism 103b may be provided in the shape of a trigger. It can be used for one-hand control to control the trip of the display control unit 102 and the controller 103 during use, which greatly improves the convenience of use.
  • the user can disengage the display control unit 102 from the controller 103 by pulling the trip mechanism 103b in a direction d1 as shown in FIG. 2 with a one-handed finger.
  • the tripping mechanism 103b can also be set to move in other directions to trigger an additional control function, for example, the user in a direction d2 different from the direction d1 as shown in FIG.
  • the trip mechanism 103b is pulled to trigger other operations of the controller. These operations will be described in detail below.
  • the display control unit 102 includes a connector 102a, an interface module 102b, a central processing unit 102c, a communication module 102d, and a power supply module 102e.
  • the interface module 102b includes at least one video/audio display interface for connecting to other display terminals.
  • the interface module 102b may be selected, for example, from an audio/video display output interface commonly used in the art, and is selected from the group consisting of an HDMI interface, a micro-HDMI interface, a VGA interface, and the like.
  • the display control unit 102 itself may transmit the video signal originally transmitted to the near-eye display 101 to an external display device, such as a television, projection, etc., for providing a video output signal thereto.
  • the interface module 102b can also be a USB interface for connecting other peripherals such as a charger, an external sensor (such as an infrared or laser ranging sensor, etc.).
  • the central processing unit 102e is configured to control the operation of each module of the display control unit 102, and control the communication module 102d to receive and transmit data/control signals from the host in a wired or wireless manner, and transmit the signals to the near-eye display 101 for display. Signals transmitted from the near-eye display 101 (eg, sensor-sensed signals) may be received and processed and sent back to the host for signal interaction.
  • the host is, for example, a computer, game machine or the like that can run software or games for human-computer interaction in accordance with the virtual reality system 100 of the present invention.
  • the communication module 102d preferably performs signal transmission in a wireless manner for the sake of user manipulation flexibility.
  • the wireless communication module 102d may be selected from at least one of WiFi, Bluetooth, infrared, ultrasonic, Wireless USB, RFID, and NFC. It will be understood by those skilled in the art that connection and data signal transmission can also be performed in a wired manner.
  • the power supply module 102e is used to power the display control unit 102.
  • the near-eye display 101 can also be powered.
  • the controller 103 can also be powered.
  • the power supply module 102e can adopt various methods such as a rechargeable battery and a disposable battery. It is preferable to use a replaceable battery power supply mode, so that the user can change the battery to extend the use time.
  • the battery slot of the power supply module 102e accommodating the battery may be disposed to pass through the body of the display control unit 102, so that when the user replaces the battery, a new battery may be placed on the side of the body of the display control unit 102, and the old battery top is Out of the fuselage, convenient for users to operate with one hand.
  • the display control unit 102 can also include at least one manipulation button 102f.
  • the manipulation of the virtual reality scene is performed when the manipulation button on the controller 103 is inconvenient to use or when the display control unit 102 is used alone.
  • the manipulation button 102f may include up, down, left, and right buttons for controlling the direction keys, and buttons for selecting, starting, switching, and the like.
  • the controller 103 includes a connector 103a, a tripping mechanism 103b, and a sensor 103c.
  • the sensor 103c is for sensing a motion signal when the user's hand-held controller 103 is moving.
  • the motion is one or more of throwing, shaking, swaying, throwing, moving, rotating, and the like.
  • the sensor 103c can be used to sense a three-axis (x, y, z) attitude and/or a three-axis (x, y, z) position of the controller 103.
  • the sensor 103c may be selected, for example, from at least one or a combination of an acceleration sensor (gravity sensor), an angular velocity sensor (gyro), and a magnetic induction sensor. The sensing operation of the sensor will be described in detail below.
  • the controller 103 may further include a communication module 103d and a manipulation button 103e.
  • the communication module 102d is configured to receive and transmit data/control signals in a wired or wireless manner with a host or display control unit.
  • the communication module 102d preferably performs signal transmission in a wireless manner.
  • the wireless communication module 102d may be selected from at least one of WiFi, Bluetooth, infrared, ultrasonic, Wireless USB, RFID, and NFC. It will be understood by those skilled in the art that connection and data signal transmission can also be performed in a wired manner.
  • the manipulation button 102f may include, for example, upper, lower, left, and right buttons for controlling the direction keys, and buttons for selecting, starting, switching, and the like. More preferably, the controller 103 can be equipped with a separate power supply unit (not shown) to power the controller 103 when it is independent of the display control unit 102.
  • the near-eye display 101 includes a display screen 101a and a sensor 101b.
  • the sensor 101b is for sensing a motion signal when the user wears the near-eye display 101 to move.
  • the sensor 101b can be used, for example, to sense a three-axis attitude and a three-axis position of the near-eye display 101.
  • the sensor 101b may be selected, for example, from at least one of an acceleration sensor, a gyroscope, a gravity sensor, and an angular velocity sensor. The sensing operation of the sensor will be described in detail below.
  • FIG. 3 shows a product structure diagram of a near-eye display 300 in accordance with an embodiment of the present invention.
  • the near-eye display 101 may be provided as a goggle type, including a display housing 301, a display screen 302 embedded in the display housing 301 on the side close to the human face and corresponding to the left and right eyes of the person, and a display
  • the housing 301 is coupled and extends rearwardly to secure the annular headband 303 of the display housing about the user's head.
  • the display housing 301 is similar in appearance to a goggle, and is recessed inwardly toward a side of the face by a predetermined depth.
  • the recess is provided with a display screen 302.
  • the recess is surrounded by a forehead support portion 304 supported on the upper portion of the head and supported on the lower portion of the head.
  • the cheek support portion 305, the nose pad 306 supported in the middle of the face, and the head band 303 supported on the left and right side portions of the head are surrounded to form a close fit with the human face.
  • the outside visible light is isolated and cannot enter.
  • the controller 103 may include a sensor for sensing a motion signal or a human body posture signal when the user's hand-held controller is moving.
  • the motion and posture are one or more of throwing, shaking, swaying, throwing, moving, turning, tilting, and the like.
  • the sensor may be an acceleration sensor (gravity sensor) that may output a signal indicating an acceleration motion of the controller due to being thrown, shaken, swayed, etc. by the user.
  • the controller 103 can monitor the output signal from the acceleration sensor and trigger an instruction when a threshold of the output signal is detected. For example, the threshold is reached when the user pushes the throw with a certain amount of force or the controller 103 reaches a certain acceleration.
  • the sensor may also be an angular velocity sensor (gyro) that outputs a signal indicative of a change in angular velocity produced by the controller as a result of being thrown, shaken, swayed, etc. by the user.
  • the senor may further comprise a magnetic induction sensor for sensing a change in the magnetic field generated by the motion of the user to perform an offset correction on the acceleration sensor or the angular velocity sensor to make the result of the sensing more accurate.
  • a magnetic induction sensor for sensing a change in the magnetic field generated by the motion of the user to perform an offset correction on the acceleration sensor or the angular velocity sensor to make the result of the sensing more accurate.
  • the sensor can trigger an instruction when the user action reaches a certain threshold, and the instruction can be an instruction to execute the next step in the process including a series of steps.
  • the instruction can be, for example, controlling a virtual grenade in a virtual reality environment game to start to leave the user's hand and move by itself.
  • the self-moving trajectory can be obtained by recording the parabolic motion simulation calculation by recording the sensing parameters such as the user's acceleration and the throwing direction sensed by the sensor as the initial parameters at the trigger threshold time.
  • the senor may also be an angular velocity sensor for sensing angular velocity, such as a gyroscope.
  • the angular velocity sensor 401 senses that the user's throw controller 400 is After the angular acceleration or angular velocity in the inertial system reaches a certain threshold, the command is triggered.
  • the sensor employed by the controller 103 according to the present invention is a nine-axis spatially-degree-of-freedom motion sensor integrated with an acceleration sensor, a gyroscope, and a magnetic induction sensor.
  • the controller 103 can also cooperate with an external sensor for further accurate correction and measurement.
  • an external sensor for further accurate correction and measurement.
  • an external range finder infrared, laser
  • the controller 103 can also cooperate with an external sensor for further accurate correction and measurement.
  • an external range finder infrared, laser
  • the controller 103 is connected through the interface of the display control unit 102 to measure the distance between the human body and the fixed marker, thereby more accurately sensing the change of the movement or posture of the human body.
  • the instructions may be predefined by the user.
  • the threshold size of the command can also be pre-defined by the user.
  • the threshold value can be used to reflect the sensitivity of the trigger sensor. Too low a threshold may cause the user to inadvertently move the controller and erroneously trigger the operation in the corresponding virtual reality environment. .
  • the controller 400 includes a resilient strap for the user to tie the controller 400 to the wrist through the lanyard to prevent damage caused by the user throwing the controller 400 out of the way during the throwing process.
  • a particular change in speed or acceleration is produced, and the change in speed or acceleration can also be used as a trigger control signal for the sensor to trigger subsequent commands or events.
  • the triggering of the instructions in the virtual reality environment can also be triggered by the tripping mechanism 103b shown in FIG. 2.
  • the triggering of the tripping mechanism 103b is similar to the physical button triggering, and can be implemented by various keystrokes, such as long press, short press, tap, and the like. Different signal functions can be triggered by setting the different durations of the press. For example, when the user is in the virtual reality environment of the battle game, at this time the user holds the controller 400 as a grenade in the virtual game, and the user presses and holds in a certain direction (for example, the direction d2 shown in FIG. 2). The tripping mechanism 103b continues for a specific time, for example, 3-5 seconds. This duration can also be set by the user. After reaching or exceeding the predetermined length of time by the tripping mechanism 103b, the user releases the tripping mechanism 103b, triggering the command to achieve the throwing of the grenade.
  • the triggering of the command in the virtual reality environment can also be triggered by the tripping action of the display control unit 102 and the controller 103 by the operation of the tripping mechanism 103b shown in FIG.
  • the connection portion of the display control unit 102 and the controller 103 may be provided with at least one electrical contact, and when the user pulls the tripping mechanism 103b in one direction during use, the control unit 102 can be controlled and controlled by one hand.
  • the trip of the device 103 at which time the display control unit 102 is disconnected from at least one of the connection portions of the controller 103, triggers an instruction.
  • the control controls the trip of the display control unit 102 and the controller 103, resulting in the display control unit 102 and the control.
  • Device At least one of the electrical contacts of the connecting portion of 103 is disconnected.
  • a trigger signal is generated to turn on the user's throwing action, for example, in the "thundering" mode in the virtual reality game environment, the user can then trigger the motion speed by using the sensor described above, or by long pressing the tripping authority 103b. Subsequent operations.
  • a sensor may also be included in the near-eye display for sensing a motion signal when a user's head wears a tilting or directional tilting motion of the near-eye display.
  • the sensor may be, for example, an acceleration (gravity) sensor that may output a signal indicative of motion resulting from a movement of the user in position, thereby triggering further instructions.
  • the display control unit can monitor a signal sensed by the sensor from the near-eye display and trigger an instruction when a threshold of the output signal is detected. For example, when the user shifts or tilts in a direction from his or her own coordinate axis, and when the displacement or tilt angle of the direction exceeds a certain threshold, the command is triggered.
  • the position tracking of the user of the virtual reality system of the present invention can be realized, so that the reality and the virtual synchronization.
  • the sensing of the above-mentioned user from a displacement or tilt angle in a certain direction of the own coordinate axis can also be realized by a sensor in the display control unit.
  • Fig. 5 schematically shows the process of this position movement tracking.
  • the sensor senses whether the user has shifted to a certain extent in any direction, and when the displacement reaches a certain threshold, an instruction may be triggered, which may be performed during a process including a series of steps. The instruction for the next step. For example, when the user is in the process of walking the virtual character in the virtual reality environment of the game, the user wears the near-eye display 501 at this time, and the user moves in any direction as shown in FIG. 5, when the displacement is sensed in the near-eye display 501.
  • the sensor senses that the user produces a displacement relative to a direction in its own coordinate axis, and when the displacement in a certain direction exceeds a certain threshold, the sensor triggers an instruction to control the movement of the character in the virtual scene to continue in this direction.
  • the control signal that is, begins to walk in the direction in which the user moves.
  • the second control signal may be triggered to stop the previous edge
  • the continuous movement in one direction or the displacement control signal in the second direction may also be triggered after the displacement in the second direction exceeds another threshold.
  • the sensor in the sensor display control unit 502 in the near-eye display 501 can also pass the sensing.
  • the user generates an instruction to trigger an inclination relative to a direction in one of its own axes.
  • a certain threshold for example, a certain angle
  • the sensor triggers an instruction to control the movement control signal in the direction of the character in the virtual scene to start moving in the direction of the user's movement.
  • the second control signal may be triggered to stop the previous edge Continuous displacement movement in one direction.
  • the displacement control signal to the second direction may also be triggered after the tilt in the second direction exceeds another threshold.
  • the angle between the previous direction and the second direction is greater than 90 degrees.
  • the next step of the trigger can also be used as a starting signal for a series of subsequent operations. For example, after the trigger starts moving in a certain direction, it can continuously move to the direction itself to reach the predetermined target.
  • the trajectory, the initial speed, and the like of the self-moving can be obtained by recording the sensing parameters such as the moving amplitude, the moving acceleration, and the linear velocity sensed by the sensor at the triggering threshold time as the initial parameters, and performing the calculation of the moving motion simulation.
  • the instructions may be predefined by the user.
  • the threshold size of the command can also be pre-defined by the user.
  • the threshold value can be used to reflect the sensitivity of the trigger sensor. Too low a threshold may cause the user to inadvertently move the near-eye display and erroneously trigger the operation in the corresponding virtual reality environment. .
  • the senor for sensing the displacement of the user disposed in the near-eye display may also be disposed in the controller or the display control unit to sense the displacement of the user when the user connects the external display through the display control unit. motion.
  • the virtual reality system according to the present invention allows the user to comfortably and naturally manipulate the control unit portion of the display, avoiding the restriction of the external cable to the user's head.
  • the one-hand operation of the escape mechanism eliminates the need for the user's hands to operate, improving convenience.
  • the virtual reality system according to the present invention simulates the user operation of the throw by using the real swaying action, and controls the movement of the avatar by sensing the real movement of the user's body, thereby improving the user experience.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

一种虚拟现实系统(100),包括近眼显示器(101)、显示器控制单元(102),以及控制器(103),所述近眼显示器(101)与所述显示器控制单元(102)是可分离的,其中,所述近眼显示器(101)包括第一传感器(101b),用于感测所述近眼显示器(101)的三轴姿态和/或三轴位置;所述显示器控制单元(102)包括第一连接件(102a)、通信模块(102d)和供电模块(102e),其中所述第一连接件(102a)用于连接所述控制器(103),所述通信模块(102d)用于以有线或无线方式接收并发送数据/控制信号,并传输至所述近眼显示器(101)以及接收来自所述近眼显示器(101)发送的信号,所述供电模块(102e)用于为所述显示器控制单元(102)供电;所述控制器(103)包括第二连接件(103a)和第二传感器(103c),其中所述第二连接件(103a)用于连接所述显示器控制单元(102),所述第二传感器(103c)用于感测所述控制器(103)的三轴姿态和/或三轴位置。

Description

一种虚拟现实组件系统 技术领域
本发明涉及虚拟现实领域,特别涉及一种实现虚拟现实场景的系统。
背景技术
虚拟现实(Virtual Reality,简称VR)是近年来出现的高科技技术。虚拟现实是利用电脑模拟产生一个三维空间的虚拟世界,提供使用者关于视觉、听觉、触觉等感官的模拟,让使用者如同身历其境一般,可以及时、没有限制地观察三度空间内的事物。虚拟现实技术涵盖了包括实时三维计算机图形技术,广角(宽视野)立体显示技术,对观察者头、眼和手的跟踪传感技术,以及触觉/力觉反馈、立体声、网络传输、语音输入输出等多种技术。已经广泛应用在医学、娱乐、军事航天、训练等多个领域中。
在虚拟现实应用的场景中,通常使用者通过佩戴头戴式显示器,利用控制器以及各种传感器来跟踪头部和身体的各种运动,从而使使用者体验到身临其境的感受。为了减轻头戴式显示器的重量,现有的产品采用将显示器部分和控制器部分可分离的方案,即在用户的操作下可分离。例如,将电路板单独装在控制盒中,然后通过电缆线等有线方式将控制器和头戴式显示器相连。但这样会为用户的使用带来极大的操控上的不便性。由于控制盒需要单独安放,例如放在桌子上,或者放在地上,因此无法方便地让用户手持,使得用户的活动范围及其有限,受到阻碍和影响。另外,因此控制盒和头戴式显示器之间的连线也会阻碍用户的运动,造成不便,降低用户体验。
另外,在虚拟现实技术应用的游戏场景中,很多是以第一人称视角的游戏,例如第一人称视角的射击游戏(FPS)。但现有的FPS在利用虚拟现实技术模拟第一人称视角的情形中,体验都不够真实。例如,当在 游戏中需要玩家进行物品抛掷时,只能通过按动控制器上的按键来模拟,并不能真实的模拟抛掷的用户体验。当玩家控制角色进行移动时,现有的游戏控制都是通过按键或摇杆来控制角色的前后左右移动,这和真是环境中的移动方式相差甚远,不够真实。而且,虚拟现实设备中采用计算机视觉方法跟踪用户,从而实现运动跟踪,这个方案系统庞大复杂,不易于民用推广。
因此,需要一种体验更好的虚拟现实系统来解决上述问题。
发明内容
本发明的目的在于提供一种虚拟现实系统,包括近眼显示器、显示器控制单元,以及控制器,所述近眼显示器与所述显示器控制单元是可分离的,其中,所述近眼显示器包括第一传感器,用于感测所述近眼显示器的三轴姿态和/或三轴位置;所述显示器控制单元包括第一连接件、通信模块和供电模块,其中所述第一连接件用于连接所述控制器,所述通信模块用于以有线或无线方式接收并发送数据/控制信号,并传输至所述近眼显示器以及接收来自所述近眼显示器发送的信号,所述供电模块用于为所述显示器控制单元供电;所述控制器包括第二连接件和第二传感器,其中所述第二连接件用于连接所述显示器控制单元,所述第二传感器用于感测所述控制器的三轴姿态和/或三轴位置。
优选地,所述近眼显示器与所述显示器控制单元通过连线连接并传输数据和/或供电信号。
优选地,所述显示器控制单元包括接口模块用于连接到其他显示终端或外设。
优选地,所述第一传感器和第二传感器选自加速度传感器、角速度传感器和磁感应传感器中的至少一种或多种的组合。
优选地,所述加速度传感器用于感测所述控制器在惯性系中的线加速度或线速度是否达到一定阈值,是则触发控制信号。
优选地,所述角速度传感器用于感测所述控制器在惯性系中的角加速度或角速度是否达到一定阈值,是则触发控制信号。
优选地,当所述传感器感测到的运动达到一定阈值后,通过记录触发阈值时刻由传感器感测到的参数作为初始参量,触发由多个指令构成的事件。
优选地,当所述第一传感器或第二传感器感测到用户产生相对于自身坐标轴中第一方向的位移或倾斜,且当所述位移或倾斜超过一定阈值之后,所述传感器触发指令以持续发出所第一述方向上的移动控制信号。
优选地,当所述第一传感器或第二传感器感测到用户产生了相对于所述第一方向不同的第二方向的移动或倾斜时,且当该第二方向上的位移或倾斜超过一定阈值之后,触发控制信号以停止之前沿所述第一方向的持续移动,或者触发向该第二方向上的位移控制信号。
优选地,所述显示器控制单元和控制器为可分离式组合套件,其组合方式为卡扣式或磁吸式。
优选地,所述组合套件为手持枪体形状,所述显示器控制单元为枪管部分,所述控制器为枪柄部分。
优选地,还包括脱扣机关,用于在用户的单手手指朝向第一方向的扳动下使所述显示器控制单元与所述控制器脱开。
优选地,所述脱扣机关用于在用户朝向与所述第一方向不同的第二方向的扳动下触发所述控制器的操作。
优选地,所述脱扣机关触发所述控制器的操作为开启感测用户的抛掷动作。
优选地,所述无线方式的所述通信模块选自WiFi、蓝牙、红外线、超声波、Wireless USB、RFID、NFC中的至少一种。
优选地,所述供电模块用于为所述显示器控制单元和/或所述控制器进行供电。
优选地,所述供电模块将容纳电池的电池槽设置为穿通所述显示器控制单元的机身,以使用户在所述显示器控制单元的机身一侧放入新的电池时,同时将旧的电池顶出机身。
优选地,所述显示器控制单元还包括至少一个操控按钮。
根据本发明的虚拟现实系统可以让用户舒适而自然地操控显示器的 显示器控制单元部分,避免了外界线缆对用户头部的限制。单手操作脱出机构,无需用户的双手操作,提高了便利性。根据本发明的虚拟现实系统采用真实的甩动动作来模拟抛掷的用户操作,通过感测用户身体的真实移动来控制虚拟人物的运动,提高了用户的使用体验。
应当理解,前述大体的描述和后续详尽的描述均为示例性说明和解释,并不应当用作对本发明所要求保护内容的限制。
附图说明
参考随附的附图,本发明更多的目的、功能和优点将通过本发明实施方式的如下描述得以阐明,其中:
图1是根据本发明的虚拟现实系统的概要结构示意图。
图2是根据本发明一个实施例的显示器控制单元包裹控制器的连接形式。
图3示出了根据本发明一个具体实施例的近眼显示器的产品结构图。
图4示出了根据本发明一个实施例当控制器中的传感器感测用户的甩动或抛掷动作并达到一定阈值时触发指令的过程。
图5示出了根据本发明一个实施例当近眼显示器或控制器中的传感器感测用户发生位移动作并达到一定阈值时触发指令的过程。
具体实施方式
通过参考示范性实施例,本发明的目的和功能以及用于实现这些目的和功能的方法将得以阐明。然而,本发明并不受限于以下所公开的示范性实施例;可以通过不同形式来对其加以实现。说明书的实质仅仅是帮助相关领域技术人员综合理解本发明的具体细节。
在下文中,将参考附图描述本发明的实施例。在附图中,相同的附图标记代表相同或类似的部件,或者相同或类似的步骤。
图1是根据本发明的虚拟现实系统100的结构示意图。如图1所示,本发明的虚拟现实系统100包括近眼显示器101、显示器控制单元102,以及控制器103,其中,近眼显示器101用于佩戴在使用者头部提供显示 功能,近眼显示器101与显示器控制单元102是分体的,之间可以通过连线104连接并传输数据和/或供电信号。控制单元102为用户手持的操控部分,用于在虚拟现实的体验中进行操控,如操控训练或游戏中人物的动作等等。显示器控制单元102接收来自控制器103的控制信号,生成供近眼显示器101显示的显示控制信号。
优选地,显示器控制单元102和控制器103可以设置为可分离式组合套件。如图1所示,根据本发明的一个实施例,所述组合套件例如可以设置为手持枪体形状,显示器控制单元102为枪管部分,控制器103为枪柄部分。枪管部分和枪柄部分通过连接件102a和103a沿如图1所示的双向箭头方向进行组合。连接件102a和103a的组合方式例如可以为卡扣式或磁吸式。例如可以采用如图1所示的阴阳卡扣的形式,也可以采用包覆式卡扣形式,例如图2所示的由显示器控制单元102包裹控制器103的连接形式。可以理解的是,所述组合套件还可以设置为其他便于用户手持的方式,例如操纵杆、手柄、方向盘的方式。
近眼显示器101需要在使用时较长时间佩戴在使用者的头部,因此需要重量轻以保证佩戴的舒适性。根据本发明的一个实施例,近眼显示器101与显示器控制单元之间优选通过有线方式的连线104连接。连线104不仅可以传输数据和信号,还可以通过设置在显示器控制单元102中的供电模块102e(将在下文描述)传输供电信号,以免除近眼显示器101需要配置额外的供电单元所增加的重量。
如图1所示,控制器103包括脱扣机关103b用于控制显示器控制单元102与控制器103的脱扣。在组合套件设置为手持枪体形状的实施例中,脱扣机关103b可以设置为扳机的形状。用于在使用时可以单手操控以控制控制显示器控制单元102与控制器103的脱扣,大大提高了使用的便捷性。如图2所示,用户可以通过单手手指朝向如图2所示的方向d1扳动脱扣机关103b从而使所述显示器控制单元102与所述控制器103脱开。
更优选地,作为控制器103的控制按键之一,脱扣机关103b还可以设置在其他方向运动从而触发额外的控制功能,例如,用户沿如图2所示的与方向d1不同的方向d2的扳动脱扣机关103b,从而触发所述控制器的其他的操作。这些操作将在下文详细说明。
如图1所示,显示器控制单元102包括连接件102a、接口模块102b、中央处理单元102c、通信模块102d和供电模块102e。接口模块102b包括至少一个视/音频显示接口,用于连接到其他显示终端。接口模块102b例如可以选自本领域常用的视/音频显示输出接口,选自HDMI接口、micro-HDMI接口、VGA接口等。当连接到其他视频显示装置时,显示器控制单元102本身可以将原本传输至近眼显示器101的视频信号传输至外接的显示设备,例如电视、投影等,为其提供视频输出信号。接口模块102b还可以是USB接口用于连接其他外设,如充电器、外接传感器(如红外或激光测距传感器等)。
中央处理单元102e用于控制显示器控制单元102各个模块的工作,并控制通信模块102d用于从主机以有线或无线方式接收并发送数据/控制信号,并将信号传输至近眼显示器101进行显示,还可以接收来自近眼显示器101发送的信号(例如传感器感测的信号)并处理后发送回主机进行信号交互。所述主机例如为电脑、游戏机等可运行供根据本发明的虚拟现实系统100实现人机交互的软件或游戏。出于使用者操控灵活性的考虑,通信模块102d优选以无线方式进行信号传送。所述无线方式的通信模块102d可以选自WiFi、蓝牙、红外线、超声波、Wireless USB、RFID、NFC中的至少一种。本领域技术人员可以理解的是,也可以采用有线方式进行连接和数据信号传送。
供电模块102e用于为显示器控制单元102供电。优选地,在显示器控制单元102和近眼显示器101通过有线方式连接的情况下,还可以为近眼显示器101供电。更优选地,也可以为控制器103进行供电。供电模块102e可采用充电电池、一次性电池等多种方式。优选采用可替换式电池供电方式,以方便使用者可以通过更改电池延长使用时间。供电模块102e容纳电池的电池槽可以设置为穿通显示器控制单元102的机身,这样用户在更换电池时,可以在显示器控制单元102的机身一侧放入新的电池,同时将旧的电池顶出机身,方便用户单手操作。
优选地,显示器控制单元102还可以包括至少一个操控按钮102f。在控制器103上的操控按钮不方便使用时或是单独使用显示器控制单元102时进行虚拟现实场景的操控。操控按钮102f可以包括控制方向键的上、下、左、右按键以及选择、开始、切换等按键。
控制器103包括连接件103a、脱扣机关103b、传感器103c。传感器103c用于感测当用户手持控制器103发生运动时的运动信号。所述运动例如抛掷、晃动、甩动、抛接、移动、转动等一种或多种。传感器103c可以用于感测所述控制器103的三轴(x,y,z)姿态和/或三轴(x,y,z)位置。传感器103c例如可以选自加速度传感器(重力传感器)、角速度传感器(陀螺仪)、磁感应传感器中的至少一种或多种的组合。传感器的感测操作将在下文详细说明。
优选地,控制器103还可以包括通信模块103d和操控按键103e。通信模块102d用于与主机或显示器控制单元以有线或无线方式接收并发送数据/控制信号。通信模块102d优选以无线方式进行信号传送。所述无线方式的通信模块102d可以选自WiFi、蓝牙、红外线、超声波、Wireless USB、RFID、NFC中的至少一种。本领域技术人员可以理解的是,也可以采用有线方式进行连接和数据信号传送。操控按钮102f例如可以包括控制方向键的上、下、左、右按键以及选择、开始、切换等按键。更优选地,控制器103可以配备独立的供电单元(图中未示出)以便在控制器103独立于显示器控制单元102工作时为其供电。
如图1所示,近眼显示器101包括显示屏101a和传感器101b。传感器101b用于感测当用户佩戴近眼显示器101发生运动时的运动信号。传感器101b例如可以用于感测所述近眼显示器101的三轴姿态和三轴位置。传感器101b例如可以选自加速度传感器、陀螺仪、重力传感器、角速度传感器中的至少一种。传感器的感测操作将在下文详细说明。
图3示出了根据本发明一个具体实施例的近眼显示器300的产品结构图。如图3所示,近眼显示器101可以设置为护目镜式,包括显示器壳体301、内嵌于该显示器壳体301的靠近人脸一侧且与人左右眼对应的显示屏302、以及与显示器壳体301连接并向后延伸以围绕使用者头部固定该显示器壳体的环状头带303。
所述显示器壳体301外观类似护目镜,其靠近人脸一侧向内凹陷一预定深度,凹陷部布置有显示屏302,凹陷部四周由支撑于头上部的前额支撑部304、支撑于头下部的脸颊支撑部305、支撑于人脸中间的鼻托306、以及支撑于头左右侧部的头带303包围,形成与人脸的紧密贴合,如此 外界可见光被隔离而不能进来。通过调节头带303,使用者在佩戴上该护目镜式显示器时,可以不受外界可见光影响且清楚地看到屏幕上的游戏画面。
根据本发明的实施例,控制器103可以包括传感器用于感测当用户手持控制器发生运动时的运动信号或人体姿态信号。所述运动和姿态例如抛掷、晃动、甩动、抛接、移动、转动、倾斜等一种或多种。例如,传感器可以是加速度传感器(重力传感器),可以输出指示控制器由于被用户抛掷、晃动、甩动等而产生的加速运动的信号。控制器103可以监测来自该加速传感器的输出信号,并且当检测到输出信号的阈值时触发指令。例如,当用户以一定的力度甩动抛掷或控制器103到达一定加速度时会达到该阈值。传感器也可以是角速度传感器(陀螺仪),输出指示控制器由于被用户抛掷、晃动、甩动等而产生的角速度变化的信号。
更优选地,传感器还可以包括磁感应传感器,用于感测用户运动产生的磁场变化从而对加速度传感器或角速度传感器进行偏移矫正,以使得感测的结果更为精确。
如图4所示,在一些实施例中,传感器感测用户动作达到一定阈值时可以触发指令,这种指令可以是执行包含有一系列步骤的过程中的下一步骤的指令。例如,当用户处于战斗游戏的虚拟现实环境中时,此时用户握持住控制器400作为虚拟的游戏中的手雷,用户沿如图4所示方向甩动或抛掷控制器400,当加速度传感器401感测到用户抛掷控制器400在惯性系中的线加速度或线速度达到一定阈值后,触发指令,该指令例如可以是控制虚拟现实环境游戏中的虚拟手雷开始脱离开用户的手并自行运动以达到预定的目标。自行运动的轨迹可以通过记录触发阈值时刻由传感器感测到的用户的加速度、抛掷方向等感测参数作为初始参量,进行抛物线运动模拟的计算从而得到。
可选地,传感器还可以是用来感测角速度的角速度传感器,例如陀螺仪,当用户沿如图4所示方向甩动或抛掷控制器400,角速度传感器401感测到用户抛掷控制器400在惯性系中的角加速度或角速度达到一定阈值后,触发指令。
优选地,根据本发明的控制器103采用的传感器为集成加速度传感器、陀螺仪和磁感应传感器一体的九轴空间自由度运动传感器。
更优选地,控制器103还可以配合外接传感器进行进一步精确的矫正和测量。例如通过显示器控制单元102的接口连接外接的测距仪(红外、激光)来测量人体与固定标志物之间的距离,从而更为精确的感测人体的运动或姿态的变化。
可以理解的是,在一些实施例中,该指令可以由用户预先定义。指令的阈值大小也可以由用户自行预先定义,阈值的大小可以用来反应触发传感器的灵敏度,过低的阈值有可能造成用户在不经意间移动控制器而错误地触发相应的虚拟现实环境中的操作。
优选地,所述控制器400包括有弹性的系带,用于使用户将控制器400通过系带系在手腕上,可以防止抛掷过程中用户脱手误将控制器400丢出去造成的损坏。根据一个实施例,由于系带在用户抛掷时受到拉拽,而产生一个特定的速度或加速度的变化,也可以将该速度或加速度的变化作为传感器的触发控制信号从而触发后续的指令或事件。
根据本发明的又一实施例,在虚拟现实环境中指令的触发还可以通过图2中所示的脱扣机关103b来实现触发。脱扣机关103b的触发类似于物理按键式触发,可以通过各种按键方式来实现,例如长按、短按、点按等。通过设置按动不同的时长,可以触发不同的信号功能。例如,当用户处于战斗游戏的虚拟现实环境中时,此时用户握持住控制器400作为虚拟的游戏中的手雷,用户通过向某一方向(例如图2中所示的方向d2)按住脱扣机关103b持续特定的时间,例如3-5秒。该时长也可以由用户自行设定。当按照脱扣机关103b达到或超过预定的时长后,用户释放脱扣机关103b,触发指令实现手雷的抛掷。
根据本发明的又一实施例,在虚拟现实环境中指令的触发还可以通过图2中所示的脱扣机关103b的操作造成显示器控制单元102与控制器103的脱扣动作来触发。可以将显示器控制单元102与控制器103的连接部分设置至少一个电触点,当用户在使用时通过向一个方向扳动脱扣机关103b时,可以单手操控以控制控制显示器控制单元102与控制器103的脱扣,此时显示器控制单元102与控制器103的连接部分中的至少一个电触点断开,触发指令。例如,当用户处于战斗游戏的虚拟现实环境中时,此时用户通过向一个方向扳动脱扣机关103b时,控制控制显示器控制单元102与控制器103的脱扣,导致显示器控制单元102与控制器 103的连接部分中的至少一个电触点断开。此时产生触发信号开启感测用户的抛掷动作,例如在虚拟现实游戏环境中的“扔手雷”模式,用户可以接着通过上面描述的利用传感器检测运动速度、或通过长按脱扣机关103b来触发后续的操作。
根据本发明的实施例,近眼显示器中也可以包括传感器用于感测当用户的头部佩戴所述近眼显示器发生平移或带有方向性的倾斜运动时的运动信号。所述传感器例如可以是加速度(重力)传感器,可以输出指示由于用户发生位置上的移动而产生的运动的信号,从而触发进一步的指令。显示器控制单元可以监测来自近眼显示器的该传感器所感测而输出的信号,并且当检测到输出信号的阈值时触发指令。例如,当用户偏离自身坐标轴中某一方向发生位移或倾斜,且当该方向的位移或倾斜角度超过一定阈值之后,会触发指令。这样可以实现本发明的虚拟现实系统对使用者的位置跟踪,使得现实和虚拟同步。可替换地,也可以通过显示器控制单元中的传感器实现上述用户偏离自身坐标轴中某一方向上位移或倾斜角度的感测。
图5示意性地示出了这一位置移动跟踪的过程。如图5所示,在一些实施例中,传感器感测用户是否向任意方向发生了一定幅度的位移,当位移达到一定阈值时可以触发指令,这种指令可以是执行包含有一系列步骤的过程中的下一步骤的指令。例如,当用户处于游戏的虚拟现实环境中的虚拟人物行走的过程中时,此时用户佩戴近眼显示器501,用户沿如图5所示任一方向发生移动,当近眼显示器501中感测位移的传感器感测到用户产生相对于自身坐标轴中某一方向的位移,且当某一方向的位移超过一定阈值之后,传感器就会触发指令以控制在虚拟场景中的人物持续发出这个方向上的移动控制信号,即开始向用户移动的方向行走。当传感器感测到用户接着产生了相对于前一位移方向不同的第二位移方向的移动时,且当该第二方向上的位移超过一定阈值之后,可以触发第二控制信号,以停止之前沿某一方向的持续运动,或者当该第二方向上的位移超过另一阈值之后,也可以触发向该第二方向上的位移控制信号。
根据又一实施例,当用户沿如图5所示任一方向发生移动,近眼显示器501中的传感器显示器控制单元502中的传感器也可以通过感测用 户产生相对于自身坐标轴中某一方向的倾斜而触发指令。且当某一方向的倾斜角度超过一定阈值(例如一定的角度)之后,传感器就会触发指令以控制在虚拟场景中的人物持续发出这个方向上的移动控制信号,即开始向用户移动的方向行走。当传感器感测到用户接着产生了相对于前一方向不同的第二方向的倾斜时,且当该第二方向上的倾斜超过一定角度达到阈值之后,可以触发第二控制信号,以停止之前沿某一方向的持续位移运动。或者当该第二方向上的倾斜超过另一阈值之后,也可以触发向该第二方向上的位移控制信号。优选地,该前一方向与第二方向之间的夹角大于90度。
当传感器感测到的用户位移发生变化并达到特定的阈值时,触发的下一步操作也可以作为之后一系列操作的起始信号。例如,可以在触发向某一个方向开始移动后,持续地向该方向一直自行移动以达到预定的目标。自行移动的轨迹、初始速度等可以通过记录触发阈值时刻由传感器感测到的用户的移动幅度、移动加速度、线速度等感测参数作为初始参量,进行移动运动模拟的计算从而得到。
可以理解的是,在一些实施例中,该指令可以由用户预先定义。指令的阈值大小也可以由用户自行预先定义,阈值的大小可以用来反应触发传感器的灵敏度,过低的阈值有可能造成用户在不经意间移动近眼显示器而错误地触发相应的虚拟现实环境中的操作。
根据本发明的又一实施例,上述设置在近眼显示器中的感测用户位移的传感器也可以设置在控制器或显示器控制单元中,以便当用户通过显示器控制单元连接外部显示器时感测用户的位移运动。
根据本发明的虚拟现实系统可以让用户舒适而自然地操控显示器的控制单元部分,避免了外界线缆对用户头部的限制。单手操作脱出机构,无需用户的双手操作,提高了便利性。根据本发明的虚拟现实系统采用真实的甩动动作来模拟抛掷的用户操作,通过感测用户身体的真实移动来控制虚拟人物的运动,提高了用户的使用体验。
结合这里披露的本发明的说明和实践,本发明的其他实施例对于本领域技术人员都是易于想到和理解的。说明和实施例仅被认为是示例性的,本发明的真正范围和主旨均由权利要求所限定。

Claims (18)

  1. 一种虚拟现实系统,包括近眼显示器、显示器控制单元,以及控制器,所述近眼显示器与所述显示器控制单元是可分离的,其中,
    所述近眼显示器包括第一传感器,用于感测所述近眼显示器的三轴姿态和/或三轴位置;
    所述显示器控制单元包括第一连接件、通信模块和供电模块,其中所述第一连接件用于连接所述控制器,所述通信模块用于以有线或无线方式接收并发送数据/控制信号,并传输至所述近眼显示器以及接收来自所述近眼显示器发送的信号,所述供电模块用于为所述显示器控制单元供电;
    所述控制器包括第二连接件和第二传感器,其中所述第二连接件用于连接所述显示器控制单元,所述第二传感器用于感测所述控制器的三轴姿态和/或三轴位置。
  2. 如权利要求1所述的系统,其中所述近眼显示器与所述显示器控制单元通过连线连接并传输数据和/或供电信号。
  3. 如权利要求1所述的系统,其中所述显示器控制单元包括接口模块用于连接到其他显示终端或外设。
  4. 如权利要求1所述的系统,其中所述第一传感器和第二传感器选自加速度传感器、角速度传感器和磁感应传感器中的至少一种或多种的组合。
  5. 如权利要求4所述的系统,其中所述加速度传感器用于感测所述控制器在惯性系中的线加速度或线速度是否达到一定阈值,是则触发控制信号。
  6. 如权利要求4所述的系统,其中所述角速度传感器用于感测所述控制器在惯性系中的角加速度或角速度是否达到一定阈值,是则触发控制信号。
  7. 如权利要求1所述的系统,其中当所述传感器感测到的运动达到一定阈值后,通过记录触发阈值时刻由传感器感测到的参数作为初始参 量,触发由多个指令构成的事件。
  8. 如权利要求1所述的系统,其中当所述第一传感器或第二传感器感测到用户产生相对于自身坐标轴中第一方向的位移或倾斜,且当所述位移或倾斜超过一定阈值之后,所述传感器触发指令以持续发出所第一述方向上的移动控制信号。
  9. 如权利要求8所述的系统,其中当所述第一传感器或第二传感器感测到用户产生了相对于所述第一方向不同的第二方向的移动或倾斜时,且当该第二方向上的位移或倾斜超过一定阈值之后,触发控制信号以停止之前沿所述第一方向的持续移动,或者触发向该第二方向上的位移控制信号。
  10. 如权利要求1所述的系统,其中所述显示器控制单元和控制器为可分离式组合套件,其组合方式为卡扣式或磁吸式。
  11. 如权利要求10所述的系统,其中所述组合套件为手持枪体形状,所述显示器控制单元为枪管部分,所述控制器为枪柄部分。
  12. 如权利要求10所述的系统,其中还包括脱扣机关,用于在用户的单手手指朝向第一方向的扳动下使所述显示器控制单元与所述控制器脱开。
  13. 如权利要求12所述的系统,其中所述脱扣机关用于在用户朝向与所述第一方向不同的第二方向的扳动下触发所述控制器的操作。
  14. 如权利要求13所述的系统,其中所述脱扣机关触发所述控制器的操作为开启感测用户的抛掷动作。
  15. 如权利要求1所述的系统,其中所述无线方式的所述通信模块选自WiFi、蓝牙、红外线、超声波、Wireless USB、RFID、NFC中的至少一种。
  16. 如权利要求1所述的系统,其中所述供电模块用于为所述显示器控制单元和/或所述控制器进行供电。
  17. 如权利要求1所述的系统,其中所述供电模块将容纳电池的电池槽设置为穿通所述显示器控制单元的机身,以使用户在所述显示器控制单元的机身一侧放入新的电池时,同时将旧的电池顶出机身。
  18. 如权利要求1所述的系统,其中所述显示器控制单元还包括至少一个操控按钮。
PCT/CN2015/075213 2014-04-10 2015-03-27 一种虚拟现实组件系统 WO2015154627A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410143435.1A CN103877726B (zh) 2014-04-10 2014-04-10 一种虚拟现实组件系统
CN201410143435.1 2014-04-10

Publications (1)

Publication Number Publication Date
WO2015154627A1 true WO2015154627A1 (zh) 2015-10-15

Family

ID=50947053

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/075213 WO2015154627A1 (zh) 2014-04-10 2015-03-27 一种虚拟现实组件系统

Country Status (2)

Country Link
CN (1) CN103877726B (zh)
WO (1) WO2015154627A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110833690A (zh) * 2019-10-30 2020-02-25 上海国威互娱文化科技有限公司 一种基于现场互动的虚拟同步的一体机

Families Citing this family (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103877726B (zh) * 2014-04-10 2017-09-26 北京蚁视科技有限公司 一种虚拟现实组件系统
CN105721856B (zh) * 2014-12-05 2017-12-15 北京蚁视科技有限公司 一种用于近眼显示器的远程图像显示方法
US20160378204A1 (en) * 2015-06-24 2016-12-29 Google Inc. System for tracking a handheld device in an augmented and/or virtual reality environment
CN105031918B (zh) * 2015-08-19 2018-02-23 深圳游视虚拟现实技术有限公司 一种基于虚拟现实技术的人机交互系统
CN105159450B (zh) * 2015-08-25 2018-01-05 中国运载火箭技术研究院 一种便携式可交互桌面级虚拟现实系统
CN105353871B (zh) * 2015-10-29 2018-12-25 上海乐相科技有限公司 一种虚拟现实场景中目标物体的控制方法及装置
CN105487670A (zh) * 2015-12-18 2016-04-13 北京诺亦腾科技有限公司 对虚拟现实系统的物理道具的定位方法以及虚拟现实系统
CN105511627A (zh) * 2015-12-18 2016-04-20 北京诺亦腾科技有限公司 一种用于虚拟现实系统的道具
CN105487672A (zh) * 2015-12-18 2016-04-13 北京诺亦腾科技有限公司 一种用于虚拟现实系统的道具
CN105487671A (zh) * 2015-12-18 2016-04-13 北京诺亦腾科技有限公司 一种用于虚拟现实系统的道具
CN105607737A (zh) * 2015-12-18 2016-05-25 北京诺亦腾科技有限公司 对虚拟现实系统的物理道具的定位方法以及虚拟现实系统
CN105528079A (zh) * 2015-12-18 2016-04-27 北京诺亦腾科技有限公司 对虚拟现实系统的物理道具的定位方法以及虚拟现实系统
CN105657408B (zh) 2015-12-31 2018-11-30 北京小鸟看看科技有限公司 虚拟现实场景的实现方法和虚拟现实装置
WO2017120768A1 (zh) * 2016-01-12 2017-07-20 深圳多哚新技术有限责任公司 一种基于vr眼镜手持终端的散热装置、手持终端及vr眼镜
US20170255254A1 (en) * 2016-03-02 2017-09-07 Htc Corporation Tracker device of virtual reality system
CN105920838A (zh) * 2016-06-08 2016-09-07 北京行云时空科技有限公司 一种动作采集操控系统和方法
CN109069927A (zh) * 2016-06-10 2018-12-21 Colopl株式会社 用于提供虚拟空间的方法、用于使计算机实现该方法的程序以及用于提供虚拟空间的系统
WO2018049624A1 (zh) * 2016-09-14 2018-03-22 深圳市柔宇科技有限公司 头戴显示设备
CN106737604B (zh) * 2017-02-03 2024-04-02 释空(上海)品牌策划有限公司 方向控制力臂及虚拟现实体验设备
CN106959762B (zh) * 2017-04-24 2019-12-31 英华达(上海)科技有限公司 虚拟现实系统及方法
CN207924888U (zh) * 2017-07-24 2018-09-28 深圳市大疆创新科技有限公司 遥控器和无人飞行器系统
CN108037827A (zh) * 2017-12-08 2018-05-15 北京凌宇智控科技有限公司 面向虚拟现实环境的虚拟物品抛掷仿真方法及其系统
TWI664995B (zh) 2018-04-18 2019-07-11 鴻海精密工業股份有限公司 虛擬實境多人桌遊互動系統、互動方法及伺服器
CN109213323B (zh) * 2018-08-28 2022-04-29 北京航空航天大学青岛研究院 一种基于眼动交互技术实现屏幕姿态估计的方法
US11789276B1 (en) 2020-04-06 2023-10-17 Apple Inc. Head-mounted device with pivoting connectors

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101024125A (zh) * 2007-03-28 2007-08-29 深圳市飞达荣电子有限公司 多平台无线影音虚拟现实游戏系统
US20080024597A1 (en) * 2006-07-27 2008-01-31 Electronics And Telecommunications Research Institute Face-mounted display apparatus for mixed reality environment
CN102454558A (zh) * 2010-10-25 2012-05-16 通用电气公司 具有塔架支承系统的陆上风力涡轮机
CN102553232A (zh) * 2010-12-07 2012-07-11 鼎亿数码科技(上海)有限公司 人体姿态捕捉装置及其实现方法
CN202724664U (zh) * 2012-05-30 2013-02-13 深圳市宇恒互动科技开发有限公司 一种可以脱离游戏独立使用的游戏枪及标靶系统
CN203075636U (zh) * 2012-11-09 2013-07-24 西安景行数创信息科技有限公司 一种互动飞碟射击游戏系统
CN103877726A (zh) * 2014-04-10 2014-06-25 北京蚁视科技有限公司 一种虚拟现实组件系统

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7699755B2 (en) * 2002-12-04 2010-04-20 Ialabs-Ca, Llc Isometric exercise system and method of facilitating user exercise during video game play
CN200983877Y (zh) * 2005-10-16 2007-12-05 王飞 虚拟现实头盔
CN2892214Y (zh) * 2006-04-30 2007-04-25 吴铁励 一种由人体姿态操作的娱乐机
CN100478990C (zh) * 2007-08-09 2009-04-15 中国科学院计算技术研究所 一种基于运动捕获数据的虚拟角色运动合成方法
EP2281228B1 (en) * 2008-05-26 2017-09-27 Microsoft International Holdings B.V. Controlling virtual reality
CN106131178A (zh) * 2008-12-05 2016-11-16 社会传播公司 实时内核
CN102023700B (zh) * 2009-09-23 2012-06-06 吴健康 一种三维人机交互系统
KR101390383B1 (ko) * 2010-11-16 2014-04-29 한국전자통신연구원 가상현실 기반 훈련 시뮬레이터를 위한 가변형 플랫폼 관리 장치
CN102348068B (zh) * 2011-08-03 2014-11-26 东北大学 一种基于头部姿态控制的随动远程视觉系统
US9255813B2 (en) * 2011-10-14 2016-02-09 Microsoft Technology Licensing, Llc User controlled real object disappearance in a mixed reality display
US9497501B2 (en) * 2011-12-06 2016-11-15 Microsoft Technology Licensing, Llc Augmented reality virtual monitor

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080024597A1 (en) * 2006-07-27 2008-01-31 Electronics And Telecommunications Research Institute Face-mounted display apparatus for mixed reality environment
CN101024125A (zh) * 2007-03-28 2007-08-29 深圳市飞达荣电子有限公司 多平台无线影音虚拟现实游戏系统
CN102454558A (zh) * 2010-10-25 2012-05-16 通用电气公司 具有塔架支承系统的陆上风力涡轮机
CN102553232A (zh) * 2010-12-07 2012-07-11 鼎亿数码科技(上海)有限公司 人体姿态捕捉装置及其实现方法
CN202724664U (zh) * 2012-05-30 2013-02-13 深圳市宇恒互动科技开发有限公司 一种可以脱离游戏独立使用的游戏枪及标靶系统
CN203075636U (zh) * 2012-11-09 2013-07-24 西安景行数创信息科技有限公司 一种互动飞碟射击游戏系统
CN103877726A (zh) * 2014-04-10 2014-06-25 北京蚁视科技有限公司 一种虚拟现实组件系统

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110833690A (zh) * 2019-10-30 2020-02-25 上海国威互娱文化科技有限公司 一种基于现场互动的虚拟同步的一体机

Also Published As

Publication number Publication date
CN103877726A (zh) 2014-06-25
CN103877726B (zh) 2017-09-26

Similar Documents

Publication Publication Date Title
WO2015154627A1 (zh) 一种虚拟现实组件系统
US10936149B2 (en) Information processing method and apparatus for executing the information processing method
US9884248B2 (en) Display control method for head-mounted display (HMD) and image generation device
JP6470796B2 (ja) 情報処理方法、プログラム及びコンピュータ
JP6368411B1 (ja) 仮想体験を提供するためにコンピュータで実行される方法、プログラム及びコンピュータ
JP6248219B1 (ja) 情報処理方法、コンピュータ、および当該情報処理方法をコンピュータに実行させるためのプログラム
JP2018072604A (ja) Vr酔いを抑制するための方法、当該方法をコンピュータに実行させるためのプログラムおよび、情報処理装置
JP2019008751A (ja) 情報処理方法、プログラム及び情報処理装置
JP6646565B2 (ja) 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム
JP2018147465A (ja) 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム
JP6227732B1 (ja) 仮想空間において入力を支援するための方法および装置、ならびに当該方法をコンピュータに実行させるプログラム
JP6479933B1 (ja) プログラム、情報処理装置、および方法
JP2019168962A (ja) プログラム、情報処理装置、及び情報処理方法
JP6458179B1 (ja) プログラム、情報処理装置、および方法
JP6263292B1 (ja) 情報処理方法、コンピュータ、および当該情報処理方法をコンピュータに実行させるためのプログラム
JP2019020832A (ja) 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム
JP2019020836A (ja) 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム
JP6441517B1 (ja) プログラム、情報処理装置、および方法
JP6710731B2 (ja) プログラム、情報処理装置、及び情報処理方法
JP6330072B1 (ja) 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム
JP2019015972A (ja) Vr酔いを抑制するための方法、当該方法をコンピュータに実行させるためのプログラムおよび、情報処理装置
KR20190113453A (ko) 가상현실 시스템에서 머리 장착형 전자장치 및 라이더 장치를 제어하는 방법 및 장치
JP2019179434A (ja) プログラム、情報処理装置、及び情報処理方法
JP6522825B1 (ja) プログラム、情報処理装置、および方法
JP2018032133A (ja) 仮想空間に表示されるオブジェクトを制御するための方法および装置、ならびに、当該方法をコンピュータに実行させるためのプログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15777122

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15777122

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 25/09/2017)

122 Ep: pct application non-entry in european phase

Ref document number: 15777122

Country of ref document: EP

Kind code of ref document: A1