WO2015154627A1 - 一种虚拟现实组件系统 - Google Patents
一种虚拟现实组件系统 Download PDFInfo
- Publication number
- WO2015154627A1 WO2015154627A1 PCT/CN2015/075213 CN2015075213W WO2015154627A1 WO 2015154627 A1 WO2015154627 A1 WO 2015154627A1 CN 2015075213 W CN2015075213 W CN 2015075213W WO 2015154627 A1 WO2015154627 A1 WO 2015154627A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- sensor
- control unit
- controller
- display control
- user
- Prior art date
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/25—Output arrangements for video game devices
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
- A63F13/428—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/80—Special adaptations for executing a specific game genre or game mode
- A63F13/837—Shooting of targets
Definitions
- the present invention relates to the field of virtual reality, and in particular, to a system for implementing a virtual reality scene.
- Virtual Reality is a high-tech technology that has emerged in recent years.
- Virtual reality is the use of computer simulation to generate a virtual world in a three-dimensional space, providing users with simulations of visual, auditory, tactile and other senses, allowing users to observe things in the three-dimensional space in a timely and unrestricted manner.
- Virtual reality technology covers real-time 3D computer graphics technology, wide-angle (wide-view) stereo display technology, tracking sensor technology for observers' heads, eyes and hands, as well as tactile/force feedback, stereo, network transmission, voice input and output.
- a user typically uses a controller and various sensors to track various movements of the head and the body by wearing a head-mounted display, thereby allowing the user to experience an immersive experience.
- the existing product adopts a scheme in which the display portion and the controller portion are detachable, that is, detachable under the operation of the user.
- the board is separately mounted in the control box, and then the controller is connected to the head mounted display by a cable or the like.
- this will bring great inconvenience to the user's use.
- control box Since the control box needs to be placed separately, for example, on a table or placed on the ground, it is not convenient for the user to hold it, so that the scope of the user's activities and its limitations are hindered and affected. In addition, the connection between the control box and the head mounted display also hinders the user's movement, causing inconvenience and reducing the user experience.
- first-person perspective shooting game many games are first-person perspective, such as a first-person perspective shooting game (FPS).
- FPS first-person perspective shooting game
- the existing FPS is not realistic enough in the case of using virtual reality technology to simulate the first person perspective.
- the existing game control controls the front, rear, left and right movement of the character through the button or the joystick, which is far from true in the environment.
- the virtual reality device uses a computer vision method to track users, thereby implementing motion tracking. This solution system is large and complex, and is not easy to promote for civilian use.
- It is an object of the present invention to provide a virtual reality system comprising a near-eye display, a display control unit, and a controller, the near-eye display being separable from the display control unit, wherein the near-eye display comprises a first sensor, a three-axis attitude and/or a three-axis position for sensing the near-eye display;
- the display control unit includes a first connector, a communication module, and a power supply module, wherein the first connector is for connecting the controller
- the communication module is configured to receive and transmit data/control signals in a wired or wireless manner and transmit to the near-eye display and receive signals transmitted from the near-eye display, the power supply module being configured to be the display control unit Powering;
- the controller includes a second connector for connecting the display control unit, and a second sensor for sensing a triaxial attitude of the controller and / or three-axis position.
- the near-eye display and the display control unit are connected by wires and transmit data and/or power supply signals.
- the display control unit includes an interface module for connecting to other display terminals or peripherals.
- the first sensor and the second sensor are selected from the group consisting of at least one or a combination of an acceleration sensor, an angular velocity sensor, and a magnetic induction sensor.
- the acceleration sensor is configured to sense whether the linear acceleration or the linear velocity of the controller in the inertial system reaches a certain threshold, and triggers the control signal.
- the angular velocity sensor is configured to sense whether the angular acceleration or angular velocity of the controller in the inertial frame reaches a certain threshold, and then triggers a control signal.
- an event composed of a plurality of instructions is triggered by recording a parameter sensed by the sensor at the trigger threshold time as an initial parameter.
- the first sensor or the second sensor senses that the user generates a displacement or tilt with respect to a first direction in its own coordinate axis, and after the displacement or tilt exceeds a certain threshold, the sensor triggers an instruction to The motion control signal in the first direction is continuously emitted.
- the control signal is triggered to stop the previous continuous movement in the first direction or to trigger the displacement control signal in the second direction.
- the display control unit and the controller are detachable combination kits, and the combination thereof is a snap type or a magnetic type.
- the combination kit is in the shape of a hand-held gun body, the display control unit is a barrel portion, and the controller is a handle portion.
- a trip mechanism is further included for disengaging the display control unit from the controller under a pull of the user's one-handed finger toward the first direction.
- the tripping mechanism is operative to trigger operation of the controller when the user is biased toward a second direction that is different from the first direction.
- the tripping mechanism triggers an operation of the controller to turn on a throwing action of the sensing user.
- the communication module of the wireless mode is selected from at least one of WiFi, Bluetooth, infrared, ultrasonic, Wireless USB, RFID, and NFC.
- the power supply module is configured to supply power to the display control unit and/or the controller.
- the power supply module sets a battery slot for accommodating the battery to pass through the body of the display control unit, so that when the user puts a new battery on the side of the body of the display control unit, the old one is simultaneously The battery is pushed out of the fuselage.
- the display control unit further comprises at least one manipulation button.
- the virtual reality system according to the present invention allows the user to comfortably and naturally manipulate the display
- the display control unit section avoids the restriction of the external cable to the user's head.
- the one-hand operation of the escape mechanism eliminates the need for the user's hands to operate, improving convenience.
- the virtual reality system according to the present invention simulates the user operation of the throw by using the real swaying action, and controls the movement of the avatar by sensing the real movement of the user's body, thereby improving the user experience.
- FIG. 1 is a schematic block diagram showing a virtual reality system according to the present invention.
- FIG. 2 is a connection form of a display control unit package controller in accordance with one embodiment of the present invention.
- FIG. 3 shows a product structure diagram of a near-eye display in accordance with an embodiment of the present invention.
- FIG. 4 illustrates the process of triggering an instruction when a sensor in a controller senses a user's sway or throw action and reaches a certain threshold in accordance with one embodiment of the present invention.
- FIG. 5 illustrates a process of triggering an instruction when a sensor in a near-eye display or controller senses a user's displacement action and reaches a certain threshold, in accordance with one embodiment of the present invention.
- the virtual reality system 100 of the present invention includes a near-eye display 101, a display control unit 102, and a controller 103, wherein the near-eye display 101 is used to provide display on the user's head.
- Function, near-eye display 101 and display control unit 102 are separate, and data and/or power supply signals can be connected and transmitted via connection 104.
- the control unit 102 is a manipulation part held by the user for manipulating in the experience of the virtual reality, such as manipulating the movement of the training or the character in the game, and the like.
- the display control unit 102 receives a control signal from the controller 103 to generate a display control signal for display by the near-eye display 101.
- display control unit 102 and controller 103 may be provided as a detachable combination kit.
- the combination kit can be configured, for example, in the shape of a hand-held gun body, the display control unit 102 is a barrel portion, and the controller 103 is a handle portion.
- the barrel portion and the shank portion are combined by the connecting members 102a and 103a in the direction of the double arrow as shown in FIG.
- the combination of the connectors 102a and 103a may be, for example, a snap type or a magnetic type.
- a form of a male or female snap as shown in FIG.
- combination kit can also be provided in other ways that are convenient for the user to hold, such as the joystick, handle, steering wheel.
- the near-eye display 101 needs to be worn on the user's head for a long time in use, and therefore needs to be lightweight to ensure comfortable wearing.
- the near-eye display 101 and the display control unit are preferably connected by a wired connection 104.
- the connection 104 can not only transmit data and signals, but also transmit power signals through the power supply module 102e (described below) disposed in the display control unit 102 to avoid the increased weight of the near-eye display 101 requiring the configuration of additional power supply units.
- the controller 103 includes a trip mechanism 103b for controlling tripping of the display control unit 102 and the controller 103.
- the trip mechanism 103b may be provided in the shape of a trigger. It can be used for one-hand control to control the trip of the display control unit 102 and the controller 103 during use, which greatly improves the convenience of use.
- the user can disengage the display control unit 102 from the controller 103 by pulling the trip mechanism 103b in a direction d1 as shown in FIG. 2 with a one-handed finger.
- the tripping mechanism 103b can also be set to move in other directions to trigger an additional control function, for example, the user in a direction d2 different from the direction d1 as shown in FIG.
- the trip mechanism 103b is pulled to trigger other operations of the controller. These operations will be described in detail below.
- the display control unit 102 includes a connector 102a, an interface module 102b, a central processing unit 102c, a communication module 102d, and a power supply module 102e.
- the interface module 102b includes at least one video/audio display interface for connecting to other display terminals.
- the interface module 102b may be selected, for example, from an audio/video display output interface commonly used in the art, and is selected from the group consisting of an HDMI interface, a micro-HDMI interface, a VGA interface, and the like.
- the display control unit 102 itself may transmit the video signal originally transmitted to the near-eye display 101 to an external display device, such as a television, projection, etc., for providing a video output signal thereto.
- the interface module 102b can also be a USB interface for connecting other peripherals such as a charger, an external sensor (such as an infrared or laser ranging sensor, etc.).
- the central processing unit 102e is configured to control the operation of each module of the display control unit 102, and control the communication module 102d to receive and transmit data/control signals from the host in a wired or wireless manner, and transmit the signals to the near-eye display 101 for display. Signals transmitted from the near-eye display 101 (eg, sensor-sensed signals) may be received and processed and sent back to the host for signal interaction.
- the host is, for example, a computer, game machine or the like that can run software or games for human-computer interaction in accordance with the virtual reality system 100 of the present invention.
- the communication module 102d preferably performs signal transmission in a wireless manner for the sake of user manipulation flexibility.
- the wireless communication module 102d may be selected from at least one of WiFi, Bluetooth, infrared, ultrasonic, Wireless USB, RFID, and NFC. It will be understood by those skilled in the art that connection and data signal transmission can also be performed in a wired manner.
- the power supply module 102e is used to power the display control unit 102.
- the near-eye display 101 can also be powered.
- the controller 103 can also be powered.
- the power supply module 102e can adopt various methods such as a rechargeable battery and a disposable battery. It is preferable to use a replaceable battery power supply mode, so that the user can change the battery to extend the use time.
- the battery slot of the power supply module 102e accommodating the battery may be disposed to pass through the body of the display control unit 102, so that when the user replaces the battery, a new battery may be placed on the side of the body of the display control unit 102, and the old battery top is Out of the fuselage, convenient for users to operate with one hand.
- the display control unit 102 can also include at least one manipulation button 102f.
- the manipulation of the virtual reality scene is performed when the manipulation button on the controller 103 is inconvenient to use or when the display control unit 102 is used alone.
- the manipulation button 102f may include up, down, left, and right buttons for controlling the direction keys, and buttons for selecting, starting, switching, and the like.
- the controller 103 includes a connector 103a, a tripping mechanism 103b, and a sensor 103c.
- the sensor 103c is for sensing a motion signal when the user's hand-held controller 103 is moving.
- the motion is one or more of throwing, shaking, swaying, throwing, moving, rotating, and the like.
- the sensor 103c can be used to sense a three-axis (x, y, z) attitude and/or a three-axis (x, y, z) position of the controller 103.
- the sensor 103c may be selected, for example, from at least one or a combination of an acceleration sensor (gravity sensor), an angular velocity sensor (gyro), and a magnetic induction sensor. The sensing operation of the sensor will be described in detail below.
- the controller 103 may further include a communication module 103d and a manipulation button 103e.
- the communication module 102d is configured to receive and transmit data/control signals in a wired or wireless manner with a host or display control unit.
- the communication module 102d preferably performs signal transmission in a wireless manner.
- the wireless communication module 102d may be selected from at least one of WiFi, Bluetooth, infrared, ultrasonic, Wireless USB, RFID, and NFC. It will be understood by those skilled in the art that connection and data signal transmission can also be performed in a wired manner.
- the manipulation button 102f may include, for example, upper, lower, left, and right buttons for controlling the direction keys, and buttons for selecting, starting, switching, and the like. More preferably, the controller 103 can be equipped with a separate power supply unit (not shown) to power the controller 103 when it is independent of the display control unit 102.
- the near-eye display 101 includes a display screen 101a and a sensor 101b.
- the sensor 101b is for sensing a motion signal when the user wears the near-eye display 101 to move.
- the sensor 101b can be used, for example, to sense a three-axis attitude and a three-axis position of the near-eye display 101.
- the sensor 101b may be selected, for example, from at least one of an acceleration sensor, a gyroscope, a gravity sensor, and an angular velocity sensor. The sensing operation of the sensor will be described in detail below.
- FIG. 3 shows a product structure diagram of a near-eye display 300 in accordance with an embodiment of the present invention.
- the near-eye display 101 may be provided as a goggle type, including a display housing 301, a display screen 302 embedded in the display housing 301 on the side close to the human face and corresponding to the left and right eyes of the person, and a display
- the housing 301 is coupled and extends rearwardly to secure the annular headband 303 of the display housing about the user's head.
- the display housing 301 is similar in appearance to a goggle, and is recessed inwardly toward a side of the face by a predetermined depth.
- the recess is provided with a display screen 302.
- the recess is surrounded by a forehead support portion 304 supported on the upper portion of the head and supported on the lower portion of the head.
- the cheek support portion 305, the nose pad 306 supported in the middle of the face, and the head band 303 supported on the left and right side portions of the head are surrounded to form a close fit with the human face.
- the outside visible light is isolated and cannot enter.
- the controller 103 may include a sensor for sensing a motion signal or a human body posture signal when the user's hand-held controller is moving.
- the motion and posture are one or more of throwing, shaking, swaying, throwing, moving, turning, tilting, and the like.
- the sensor may be an acceleration sensor (gravity sensor) that may output a signal indicating an acceleration motion of the controller due to being thrown, shaken, swayed, etc. by the user.
- the controller 103 can monitor the output signal from the acceleration sensor and trigger an instruction when a threshold of the output signal is detected. For example, the threshold is reached when the user pushes the throw with a certain amount of force or the controller 103 reaches a certain acceleration.
- the sensor may also be an angular velocity sensor (gyro) that outputs a signal indicative of a change in angular velocity produced by the controller as a result of being thrown, shaken, swayed, etc. by the user.
- the senor may further comprise a magnetic induction sensor for sensing a change in the magnetic field generated by the motion of the user to perform an offset correction on the acceleration sensor or the angular velocity sensor to make the result of the sensing more accurate.
- a magnetic induction sensor for sensing a change in the magnetic field generated by the motion of the user to perform an offset correction on the acceleration sensor or the angular velocity sensor to make the result of the sensing more accurate.
- the sensor can trigger an instruction when the user action reaches a certain threshold, and the instruction can be an instruction to execute the next step in the process including a series of steps.
- the instruction can be, for example, controlling a virtual grenade in a virtual reality environment game to start to leave the user's hand and move by itself.
- the self-moving trajectory can be obtained by recording the parabolic motion simulation calculation by recording the sensing parameters such as the user's acceleration and the throwing direction sensed by the sensor as the initial parameters at the trigger threshold time.
- the senor may also be an angular velocity sensor for sensing angular velocity, such as a gyroscope.
- the angular velocity sensor 401 senses that the user's throw controller 400 is After the angular acceleration or angular velocity in the inertial system reaches a certain threshold, the command is triggered.
- the sensor employed by the controller 103 according to the present invention is a nine-axis spatially-degree-of-freedom motion sensor integrated with an acceleration sensor, a gyroscope, and a magnetic induction sensor.
- the controller 103 can also cooperate with an external sensor for further accurate correction and measurement.
- an external sensor for further accurate correction and measurement.
- an external range finder infrared, laser
- the controller 103 can also cooperate with an external sensor for further accurate correction and measurement.
- an external range finder infrared, laser
- the controller 103 is connected through the interface of the display control unit 102 to measure the distance between the human body and the fixed marker, thereby more accurately sensing the change of the movement or posture of the human body.
- the instructions may be predefined by the user.
- the threshold size of the command can also be pre-defined by the user.
- the threshold value can be used to reflect the sensitivity of the trigger sensor. Too low a threshold may cause the user to inadvertently move the controller and erroneously trigger the operation in the corresponding virtual reality environment. .
- the controller 400 includes a resilient strap for the user to tie the controller 400 to the wrist through the lanyard to prevent damage caused by the user throwing the controller 400 out of the way during the throwing process.
- a particular change in speed or acceleration is produced, and the change in speed or acceleration can also be used as a trigger control signal for the sensor to trigger subsequent commands or events.
- the triggering of the instructions in the virtual reality environment can also be triggered by the tripping mechanism 103b shown in FIG. 2.
- the triggering of the tripping mechanism 103b is similar to the physical button triggering, and can be implemented by various keystrokes, such as long press, short press, tap, and the like. Different signal functions can be triggered by setting the different durations of the press. For example, when the user is in the virtual reality environment of the battle game, at this time the user holds the controller 400 as a grenade in the virtual game, and the user presses and holds in a certain direction (for example, the direction d2 shown in FIG. 2). The tripping mechanism 103b continues for a specific time, for example, 3-5 seconds. This duration can also be set by the user. After reaching or exceeding the predetermined length of time by the tripping mechanism 103b, the user releases the tripping mechanism 103b, triggering the command to achieve the throwing of the grenade.
- the triggering of the command in the virtual reality environment can also be triggered by the tripping action of the display control unit 102 and the controller 103 by the operation of the tripping mechanism 103b shown in FIG.
- the connection portion of the display control unit 102 and the controller 103 may be provided with at least one electrical contact, and when the user pulls the tripping mechanism 103b in one direction during use, the control unit 102 can be controlled and controlled by one hand.
- the trip of the device 103 at which time the display control unit 102 is disconnected from at least one of the connection portions of the controller 103, triggers an instruction.
- the control controls the trip of the display control unit 102 and the controller 103, resulting in the display control unit 102 and the control.
- Device At least one of the electrical contacts of the connecting portion of 103 is disconnected.
- a trigger signal is generated to turn on the user's throwing action, for example, in the "thundering" mode in the virtual reality game environment, the user can then trigger the motion speed by using the sensor described above, or by long pressing the tripping authority 103b. Subsequent operations.
- a sensor may also be included in the near-eye display for sensing a motion signal when a user's head wears a tilting or directional tilting motion of the near-eye display.
- the sensor may be, for example, an acceleration (gravity) sensor that may output a signal indicative of motion resulting from a movement of the user in position, thereby triggering further instructions.
- the display control unit can monitor a signal sensed by the sensor from the near-eye display and trigger an instruction when a threshold of the output signal is detected. For example, when the user shifts or tilts in a direction from his or her own coordinate axis, and when the displacement or tilt angle of the direction exceeds a certain threshold, the command is triggered.
- the position tracking of the user of the virtual reality system of the present invention can be realized, so that the reality and the virtual synchronization.
- the sensing of the above-mentioned user from a displacement or tilt angle in a certain direction of the own coordinate axis can also be realized by a sensor in the display control unit.
- Fig. 5 schematically shows the process of this position movement tracking.
- the sensor senses whether the user has shifted to a certain extent in any direction, and when the displacement reaches a certain threshold, an instruction may be triggered, which may be performed during a process including a series of steps. The instruction for the next step. For example, when the user is in the process of walking the virtual character in the virtual reality environment of the game, the user wears the near-eye display 501 at this time, and the user moves in any direction as shown in FIG. 5, when the displacement is sensed in the near-eye display 501.
- the sensor senses that the user produces a displacement relative to a direction in its own coordinate axis, and when the displacement in a certain direction exceeds a certain threshold, the sensor triggers an instruction to control the movement of the character in the virtual scene to continue in this direction.
- the control signal that is, begins to walk in the direction in which the user moves.
- the second control signal may be triggered to stop the previous edge
- the continuous movement in one direction or the displacement control signal in the second direction may also be triggered after the displacement in the second direction exceeds another threshold.
- the sensor in the sensor display control unit 502 in the near-eye display 501 can also pass the sensing.
- the user generates an instruction to trigger an inclination relative to a direction in one of its own axes.
- a certain threshold for example, a certain angle
- the sensor triggers an instruction to control the movement control signal in the direction of the character in the virtual scene to start moving in the direction of the user's movement.
- the second control signal may be triggered to stop the previous edge Continuous displacement movement in one direction.
- the displacement control signal to the second direction may also be triggered after the tilt in the second direction exceeds another threshold.
- the angle between the previous direction and the second direction is greater than 90 degrees.
- the next step of the trigger can also be used as a starting signal for a series of subsequent operations. For example, after the trigger starts moving in a certain direction, it can continuously move to the direction itself to reach the predetermined target.
- the trajectory, the initial speed, and the like of the self-moving can be obtained by recording the sensing parameters such as the moving amplitude, the moving acceleration, and the linear velocity sensed by the sensor at the triggering threshold time as the initial parameters, and performing the calculation of the moving motion simulation.
- the instructions may be predefined by the user.
- the threshold size of the command can also be pre-defined by the user.
- the threshold value can be used to reflect the sensitivity of the trigger sensor. Too low a threshold may cause the user to inadvertently move the near-eye display and erroneously trigger the operation in the corresponding virtual reality environment. .
- the senor for sensing the displacement of the user disposed in the near-eye display may also be disposed in the controller or the display control unit to sense the displacement of the user when the user connects the external display through the display control unit. motion.
- the virtual reality system according to the present invention allows the user to comfortably and naturally manipulate the control unit portion of the display, avoiding the restriction of the external cable to the user's head.
- the one-hand operation of the escape mechanism eliminates the need for the user's hands to operate, improving convenience.
- the virtual reality system according to the present invention simulates the user operation of the throw by using the real swaying action, and controls the movement of the avatar by sensing the real movement of the user's body, thereby improving the user experience.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (18)
- 一种虚拟现实系统,包括近眼显示器、显示器控制单元,以及控制器,所述近眼显示器与所述显示器控制单元是可分离的,其中,所述近眼显示器包括第一传感器,用于感测所述近眼显示器的三轴姿态和/或三轴位置;所述显示器控制单元包括第一连接件、通信模块和供电模块,其中所述第一连接件用于连接所述控制器,所述通信模块用于以有线或无线方式接收并发送数据/控制信号,并传输至所述近眼显示器以及接收来自所述近眼显示器发送的信号,所述供电模块用于为所述显示器控制单元供电;所述控制器包括第二连接件和第二传感器,其中所述第二连接件用于连接所述显示器控制单元,所述第二传感器用于感测所述控制器的三轴姿态和/或三轴位置。
- 如权利要求1所述的系统,其中所述近眼显示器与所述显示器控制单元通过连线连接并传输数据和/或供电信号。
- 如权利要求1所述的系统,其中所述显示器控制单元包括接口模块用于连接到其他显示终端或外设。
- 如权利要求1所述的系统,其中所述第一传感器和第二传感器选自加速度传感器、角速度传感器和磁感应传感器中的至少一种或多种的组合。
- 如权利要求4所述的系统,其中所述加速度传感器用于感测所述控制器在惯性系中的线加速度或线速度是否达到一定阈值,是则触发控制信号。
- 如权利要求4所述的系统,其中所述角速度传感器用于感测所述控制器在惯性系中的角加速度或角速度是否达到一定阈值,是则触发控制信号。
- 如权利要求1所述的系统,其中当所述传感器感测到的运动达到一定阈值后,通过记录触发阈值时刻由传感器感测到的参数作为初始参 量,触发由多个指令构成的事件。
- 如权利要求1所述的系统,其中当所述第一传感器或第二传感器感测到用户产生相对于自身坐标轴中第一方向的位移或倾斜,且当所述位移或倾斜超过一定阈值之后,所述传感器触发指令以持续发出所第一述方向上的移动控制信号。
- 如权利要求8所述的系统,其中当所述第一传感器或第二传感器感测到用户产生了相对于所述第一方向不同的第二方向的移动或倾斜时,且当该第二方向上的位移或倾斜超过一定阈值之后,触发控制信号以停止之前沿所述第一方向的持续移动,或者触发向该第二方向上的位移控制信号。
- 如权利要求1所述的系统,其中所述显示器控制单元和控制器为可分离式组合套件,其组合方式为卡扣式或磁吸式。
- 如权利要求10所述的系统,其中所述组合套件为手持枪体形状,所述显示器控制单元为枪管部分,所述控制器为枪柄部分。
- 如权利要求10所述的系统,其中还包括脱扣机关,用于在用户的单手手指朝向第一方向的扳动下使所述显示器控制单元与所述控制器脱开。
- 如权利要求12所述的系统,其中所述脱扣机关用于在用户朝向与所述第一方向不同的第二方向的扳动下触发所述控制器的操作。
- 如权利要求13所述的系统,其中所述脱扣机关触发所述控制器的操作为开启感测用户的抛掷动作。
- 如权利要求1所述的系统,其中所述无线方式的所述通信模块选自WiFi、蓝牙、红外线、超声波、Wireless USB、RFID、NFC中的至少一种。
- 如权利要求1所述的系统,其中所述供电模块用于为所述显示器控制单元和/或所述控制器进行供电。
- 如权利要求1所述的系统,其中所述供电模块将容纳电池的电池槽设置为穿通所述显示器控制单元的机身,以使用户在所述显示器控制单元的机身一侧放入新的电池时,同时将旧的电池顶出机身。
- 如权利要求1所述的系统,其中所述显示器控制单元还包括至少一个操控按钮。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410143435.1A CN103877726B (zh) | 2014-04-10 | 2014-04-10 | 一种虚拟现实组件系统 |
CN201410143435.1 | 2014-04-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015154627A1 true WO2015154627A1 (zh) | 2015-10-15 |
Family
ID=50947053
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2015/075213 WO2015154627A1 (zh) | 2014-04-10 | 2015-03-27 | 一种虚拟现实组件系统 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN103877726B (zh) |
WO (1) | WO2015154627A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110833690A (zh) * | 2019-10-30 | 2020-02-25 | 上海国威互娱文化科技有限公司 | 一种基于现场互动的虚拟同步的一体机 |
Families Citing this family (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103877726B (zh) * | 2014-04-10 | 2017-09-26 | 北京蚁视科技有限公司 | 一种虚拟现实组件系统 |
CN105721856B (zh) * | 2014-12-05 | 2017-12-15 | 北京蚁视科技有限公司 | 一种用于近眼显示器的远程图像显示方法 |
US20160378204A1 (en) * | 2015-06-24 | 2016-12-29 | Google Inc. | System for tracking a handheld device in an augmented and/or virtual reality environment |
CN105031918B (zh) * | 2015-08-19 | 2018-02-23 | 深圳游视虚拟现实技术有限公司 | 一种基于虚拟现实技术的人机交互系统 |
CN105159450B (zh) * | 2015-08-25 | 2018-01-05 | 中国运载火箭技术研究院 | 一种便携式可交互桌面级虚拟现实系统 |
CN105353871B (zh) * | 2015-10-29 | 2018-12-25 | 上海乐相科技有限公司 | 一种虚拟现实场景中目标物体的控制方法及装置 |
CN105487670A (zh) * | 2015-12-18 | 2016-04-13 | 北京诺亦腾科技有限公司 | 对虚拟现实系统的物理道具的定位方法以及虚拟现实系统 |
CN105511627A (zh) * | 2015-12-18 | 2016-04-20 | 北京诺亦腾科技有限公司 | 一种用于虚拟现实系统的道具 |
CN105487672A (zh) * | 2015-12-18 | 2016-04-13 | 北京诺亦腾科技有限公司 | 一种用于虚拟现实系统的道具 |
CN105487671A (zh) * | 2015-12-18 | 2016-04-13 | 北京诺亦腾科技有限公司 | 一种用于虚拟现实系统的道具 |
CN105607737A (zh) * | 2015-12-18 | 2016-05-25 | 北京诺亦腾科技有限公司 | 对虚拟现实系统的物理道具的定位方法以及虚拟现实系统 |
CN105528079A (zh) * | 2015-12-18 | 2016-04-27 | 北京诺亦腾科技有限公司 | 对虚拟现实系统的物理道具的定位方法以及虚拟现实系统 |
CN105657408B (zh) | 2015-12-31 | 2018-11-30 | 北京小鸟看看科技有限公司 | 虚拟现实场景的实现方法和虚拟现实装置 |
WO2017120768A1 (zh) * | 2016-01-12 | 2017-07-20 | 深圳多哚新技术有限责任公司 | 一种基于vr眼镜手持终端的散热装置、手持终端及vr眼镜 |
US20170255254A1 (en) * | 2016-03-02 | 2017-09-07 | Htc Corporation | Tracker device of virtual reality system |
CN105920838A (zh) * | 2016-06-08 | 2016-09-07 | 北京行云时空科技有限公司 | 一种动作采集操控系统和方法 |
CN109069927A (zh) * | 2016-06-10 | 2018-12-21 | Colopl株式会社 | 用于提供虚拟空间的方法、用于使计算机实现该方法的程序以及用于提供虚拟空间的系统 |
WO2018049624A1 (zh) * | 2016-09-14 | 2018-03-22 | 深圳市柔宇科技有限公司 | 头戴显示设备 |
CN106737604B (zh) * | 2017-02-03 | 2024-04-02 | 释空(上海)品牌策划有限公司 | 方向控制力臂及虚拟现实体验设备 |
CN106959762B (zh) * | 2017-04-24 | 2019-12-31 | 英华达(上海)科技有限公司 | 虚拟现实系统及方法 |
CN207924888U (zh) * | 2017-07-24 | 2018-09-28 | 深圳市大疆创新科技有限公司 | 遥控器和无人飞行器系统 |
CN108037827A (zh) * | 2017-12-08 | 2018-05-15 | 北京凌宇智控科技有限公司 | 面向虚拟现实环境的虚拟物品抛掷仿真方法及其系统 |
TWI664995B (zh) | 2018-04-18 | 2019-07-11 | 鴻海精密工業股份有限公司 | 虛擬實境多人桌遊互動系統、互動方法及伺服器 |
CN109213323B (zh) * | 2018-08-28 | 2022-04-29 | 北京航空航天大学青岛研究院 | 一种基于眼动交互技术实现屏幕姿态估计的方法 |
US11789276B1 (en) | 2020-04-06 | 2023-10-17 | Apple Inc. | Head-mounted device with pivoting connectors |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101024125A (zh) * | 2007-03-28 | 2007-08-29 | 深圳市飞达荣电子有限公司 | 多平台无线影音虚拟现实游戏系统 |
US20080024597A1 (en) * | 2006-07-27 | 2008-01-31 | Electronics And Telecommunications Research Institute | Face-mounted display apparatus for mixed reality environment |
CN102454558A (zh) * | 2010-10-25 | 2012-05-16 | 通用电气公司 | 具有塔架支承系统的陆上风力涡轮机 |
CN102553232A (zh) * | 2010-12-07 | 2012-07-11 | 鼎亿数码科技(上海)有限公司 | 人体姿态捕捉装置及其实现方法 |
CN202724664U (zh) * | 2012-05-30 | 2013-02-13 | 深圳市宇恒互动科技开发有限公司 | 一种可以脱离游戏独立使用的游戏枪及标靶系统 |
CN203075636U (zh) * | 2012-11-09 | 2013-07-24 | 西安景行数创信息科技有限公司 | 一种互动飞碟射击游戏系统 |
CN103877726A (zh) * | 2014-04-10 | 2014-06-25 | 北京蚁视科技有限公司 | 一种虚拟现实组件系统 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7699755B2 (en) * | 2002-12-04 | 2010-04-20 | Ialabs-Ca, Llc | Isometric exercise system and method of facilitating user exercise during video game play |
CN200983877Y (zh) * | 2005-10-16 | 2007-12-05 | 王飞 | 虚拟现实头盔 |
CN2892214Y (zh) * | 2006-04-30 | 2007-04-25 | 吴铁励 | 一种由人体姿态操作的娱乐机 |
CN100478990C (zh) * | 2007-08-09 | 2009-04-15 | 中国科学院计算技术研究所 | 一种基于运动捕获数据的虚拟角色运动合成方法 |
EP2281228B1 (en) * | 2008-05-26 | 2017-09-27 | Microsoft International Holdings B.V. | Controlling virtual reality |
CN106131178A (zh) * | 2008-12-05 | 2016-11-16 | 社会传播公司 | 实时内核 |
CN102023700B (zh) * | 2009-09-23 | 2012-06-06 | 吴健康 | 一种三维人机交互系统 |
KR101390383B1 (ko) * | 2010-11-16 | 2014-04-29 | 한국전자통신연구원 | 가상현실 기반 훈련 시뮬레이터를 위한 가변형 플랫폼 관리 장치 |
CN102348068B (zh) * | 2011-08-03 | 2014-11-26 | 东北大学 | 一种基于头部姿态控制的随动远程视觉系统 |
US9255813B2 (en) * | 2011-10-14 | 2016-02-09 | Microsoft Technology Licensing, Llc | User controlled real object disappearance in a mixed reality display |
US9497501B2 (en) * | 2011-12-06 | 2016-11-15 | Microsoft Technology Licensing, Llc | Augmented reality virtual monitor |
-
2014
- 2014-04-10 CN CN201410143435.1A patent/CN103877726B/zh active Active
-
2015
- 2015-03-27 WO PCT/CN2015/075213 patent/WO2015154627A1/zh active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080024597A1 (en) * | 2006-07-27 | 2008-01-31 | Electronics And Telecommunications Research Institute | Face-mounted display apparatus for mixed reality environment |
CN101024125A (zh) * | 2007-03-28 | 2007-08-29 | 深圳市飞达荣电子有限公司 | 多平台无线影音虚拟现实游戏系统 |
CN102454558A (zh) * | 2010-10-25 | 2012-05-16 | 通用电气公司 | 具有塔架支承系统的陆上风力涡轮机 |
CN102553232A (zh) * | 2010-12-07 | 2012-07-11 | 鼎亿数码科技(上海)有限公司 | 人体姿态捕捉装置及其实现方法 |
CN202724664U (zh) * | 2012-05-30 | 2013-02-13 | 深圳市宇恒互动科技开发有限公司 | 一种可以脱离游戏独立使用的游戏枪及标靶系统 |
CN203075636U (zh) * | 2012-11-09 | 2013-07-24 | 西安景行数创信息科技有限公司 | 一种互动飞碟射击游戏系统 |
CN103877726A (zh) * | 2014-04-10 | 2014-06-25 | 北京蚁视科技有限公司 | 一种虚拟现实组件系统 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110833690A (zh) * | 2019-10-30 | 2020-02-25 | 上海国威互娱文化科技有限公司 | 一种基于现场互动的虚拟同步的一体机 |
Also Published As
Publication number | Publication date |
---|---|
CN103877726A (zh) | 2014-06-25 |
CN103877726B (zh) | 2017-09-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015154627A1 (zh) | 一种虚拟现实组件系统 | |
US10936149B2 (en) | Information processing method and apparatus for executing the information processing method | |
US9884248B2 (en) | Display control method for head-mounted display (HMD) and image generation device | |
JP6470796B2 (ja) | 情報処理方法、プログラム及びコンピュータ | |
JP6368411B1 (ja) | 仮想体験を提供するためにコンピュータで実行される方法、プログラム及びコンピュータ | |
JP6248219B1 (ja) | 情報処理方法、コンピュータ、および当該情報処理方法をコンピュータに実行させるためのプログラム | |
JP2018072604A (ja) | Vr酔いを抑制するための方法、当該方法をコンピュータに実行させるためのプログラムおよび、情報処理装置 | |
JP2019008751A (ja) | 情報処理方法、プログラム及び情報処理装置 | |
JP6646565B2 (ja) | 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム | |
JP2018147465A (ja) | 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム | |
JP6227732B1 (ja) | 仮想空間において入力を支援するための方法および装置、ならびに当該方法をコンピュータに実行させるプログラム | |
JP6479933B1 (ja) | プログラム、情報処理装置、および方法 | |
JP2019168962A (ja) | プログラム、情報処理装置、及び情報処理方法 | |
JP6458179B1 (ja) | プログラム、情報処理装置、および方法 | |
JP6263292B1 (ja) | 情報処理方法、コンピュータ、および当該情報処理方法をコンピュータに実行させるためのプログラム | |
JP2019020832A (ja) | 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム | |
JP2019020836A (ja) | 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム | |
JP6441517B1 (ja) | プログラム、情報処理装置、および方法 | |
JP6710731B2 (ja) | プログラム、情報処理装置、及び情報処理方法 | |
JP6330072B1 (ja) | 情報処理方法、装置、および当該情報処理方法をコンピュータに実行させるためのプログラム | |
JP2019015972A (ja) | Vr酔いを抑制するための方法、当該方法をコンピュータに実行させるためのプログラムおよび、情報処理装置 | |
KR20190113453A (ko) | 가상현실 시스템에서 머리 장착형 전자장치 및 라이더 장치를 제어하는 방법 및 장치 | |
JP2019179434A (ja) | プログラム、情報処理装置、及び情報処理方法 | |
JP6522825B1 (ja) | プログラム、情報処理装置、および方法 | |
JP2018032133A (ja) | 仮想空間に表示されるオブジェクトを制御するための方法および装置、ならびに、当該方法をコンピュータに実行させるためのプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15777122 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15777122 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 25/09/2017) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15777122 Country of ref document: EP Kind code of ref document: A1 |