WO2011129543A2 - 가상 세계 처리 장치 및 방법 - Google Patents
가상 세계 처리 장치 및 방법 Download PDFInfo
- Publication number
- WO2011129543A2 WO2011129543A2 PCT/KR2011/002408 KR2011002408W WO2011129543A2 WO 2011129543 A2 WO2011129543 A2 WO 2011129543A2 KR 2011002408 W KR2011002408 W KR 2011002408W WO 2011129543 A2 WO2011129543 A2 WO 2011129543A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- angle value
- sensor
- virtual world
- bending
- sensing position
- Prior art date
Links
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
- A63F13/428—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/60—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
- A63F13/65—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/212—Input arrangements for video game devices characterised by their sensors, purposes or types using sensors worn by the player, e.g. for measuring heart beat or leg activity
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/55—Controlling game characters or game objects based on the game progress
- A63F13/56—Computing the motion of game characters with respect to other game characters, game objects or elements of the game scene, e.g. for simulating the behaviour of a group of virtual soldiers or for path finding
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/70—Game security or game management aspects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/23—Input arrangements for video game devices for interfacing with the game device, e.g. specific interfaces between game controller and console
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/10—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
- A63F2300/1012—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals involving biosensors worn by the player, e.g. for measuring heart beat, limb activity
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/10—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
- A63F2300/1025—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals details of the interface with the game device, e.g. USB version detection
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/50—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers
- A63F2300/55—Details of game data or player data management
- A63F2300/5546—Details of game data or player data management using player registration data, e.g. identification, account, preferences, game history
- A63F2300/5553—Details of game data or player data management using player registration data, e.g. identification, account, preferences, game history user representation in the game field, e.g. avatar
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/6045—Methods for processing data by generating or executing the game program for mapping control signals received from the input arrangement into game commands
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/65—Methods for processing data by generating or executing the game program for computing the condition of a game character
Definitions
- Embodiments of the present invention relate to an apparatus and method for processing a virtual world, and more particularly, to an apparatus and method for applying sensing information measured by a bending sensor to a virtual world.
- the interaction between the real world and the virtual world has two directions. The first is to reflect the data information obtained from the sensor of the real world to the virtual world, and the second is to reflect the data information obtained from the virtual world to the real world through an actuator.
- the virtual world processing apparatus may include a receiver configured to receive an angle value of a sensing position and a sensor characteristic of the bending sensor from a bending sensor, based on the angle value and the sensor characteristic, based on the sensing position. And a processing unit for generating control information for controlling objects of the virtual world, and a transmission unit for transmitting the generated control information to the virtual world.
- the virtual world processing method may include receiving an angle value of a sensing position and a sensor characteristic of a characteristic of the bending sensor from a bending sensor, based on the angle value and the sensor characteristic, based on the sensing position. Generating control information for controlling an object of the virtual world, and transmitting the generated control information to the virtual world.
- Embodiments may implement the interaction between the real world and the virtual world by transmitting the angle value of measuring the angle of the body part of the user of the real world to the virtual world using the sensor characteristic which is information on the characteristic of the bending sensor.
- embodiments are generated based on the sensor characteristics and the angle value of the sensed angle of the body part, to generate control information for controlling the part of the avatar of the virtual world to the body part to pass to the virtual world to the real world and the virtual world Can implement the interaction.
- FIG. 1 is a diagram for exchanging information between a real world and a virtual world, according to an exemplary embodiment. It is a figure which shows the virtual world processing system to control.
- FIG. 2 is a diagram illustrating a configuration of a virtual world processing apparatus according to an embodiment of the present invention.
- FIG. 3 illustrates an operation of measuring a sensing position by a plurality of bending sensors according to an exemplary embodiment of the present invention.
- FIG. 4 illustrates an operation of calculating an angle value by a virtual world processing apparatus according to an embodiment of the present invention.
- FIG. 5 is a diagram illustrating a structure of an avatar control feature point type according to an embodiment of the present invention.
- FIG. 6 is a flowchart illustrating a virtual world processing method according to an embodiment of the present invention.
- FIG. 1 is a diagram illustrating a virtual world processing system that controls information exchange between a real world and a virtual world, according to an exemplary embodiment.
- a virtual world processing system may include a real world 110, a virtual world processing apparatus, and a virtual world 140.
- the real world 110 may represent a sensor that senses information about the real world 110 or a sensory device that implements information about the virtual world 140 in the real world 110.
- the virtual world 140 may represent a sensory media playback device that plays content including sensory effect information that may be implemented in the virtual world 140 itself or the real world 110 implemented by a program.
- a sensor may sense and transmit information on a user's motion, state, intention, shape, etc. of the real world 110 to the virtual world processing device.
- the senor may transmit a sensor capability 101, a sensor adaptation preference 102, and a sensed information 103 to the virtual world processor.
- the sensor characteristic 101 is information about the characteristic of the sensor.
- Sensor adaptation preference 102 is information indicative of the degree to which a user of a sensor prefers to the characteristics of the sensor.
- Sensing information 103 is information that the sensor senses the real world 110.
- the virtual world processing apparatus includes a adaptation real world to virtual world (RV) 120, a virtual world information (VWI) 104, and a adaptation real world to virtual world / Virtual World to Real World) 130.
- RV virtual world
- VWI virtual world information
- VWI adaptation real world to virtual world / Virtual World to Real World
- the detoxification RV 120 is information that can apply the sensing information 103 detected by the sensor to the virtual world 140 to the virtual world 140 based on the sensor characteristic 101 and the sensor decoction preference 102. I can convert it.
- the heroic RV 120 may be implemented as a real world to virtual world engine (RV engine).
- the heroic RV 120 uses the converted sensing information 103.
- VWI Virtual World Information
- the VWI 104 is information about a virtual object of the virtual world 140.
- the Raccoon RV VR 130 may encode the converted VWI 104 to generate a Virtual World Effect Metadata (VWEM) 107, which is metadata about an effect applied to the virtual world 140. have.
- VWEM Virtual World Effect Metadata
- the heroic RV / VR 130 may select the VWEM 107 based on the Virtual World Capabilities (VWC) 10 5 and the Virtual World Preferences (VWP) 106. Can be generated.
- the VWC 105 is information about the characteristics of the virtual world 140.
- the VWP 106 is information indicating the degree of preference of the user with respect to the characteristics of the virtual world 140.
- the heroic RV / VR 130 may transmit the VWEM 107 to the virtual world 140.
- the virtual world 140 is applied to the VWEM 107, so that the effect on the sensing information 103 may be implemented in the virtual world 140.
- an effect event occurring in the virtual world 140 may be driven by a sensory device that is an actuator in the real world 110.
- the virtual world 140 may generate sensory effect metadata (SEM) 111 by encoding sensory effect information, which is information about effect events generated in the virtual world 140.
- SEM sensory effect metadata
- the virtual world 140 may include a sensory media playback device that plays content including sensory effect information.
- the detox RV / VR 130 may generate sensory information 112 based on the SEM (lll).
- Sensory information 112 is information about effect events implemented in sensory devices in the real world 110.
- the detox VR 150 may generate information about a sensory device command (SDCmd) 115 that controls the operation of the sensory device of the real world 110.
- the adaptive VR 150 may be configured based on information on the Sensory Device Capabilities (SDCap) 113 and information on the User Sensory Preference (USP) 114. Information about the SDCmd 115 may be generated.
- the SDCap 113 is information on characteristics of the sensory device.
- the USP 114 is information indicating the degree of preference of the user for the effect implemented in the sensory device.
- FIG. 2 is a diagram showing the configuration of a virtual world processing apparatus according to an embodiment of the present invention.
- the virtual world processing apparatus 200 includes a receiver 210, a processor 220, and a transmitter 230.
- the receiver 210 receives the sensor characteristic of the angle value of the sensing position measured by the bending sensor 260 and the characteristics of the bending sensor 260 from the bending sensor 260.
- the bending sensor 260 is a sensor for measuring the sensed information on the degree of bending of the object in the real world.
- the sensing information is a measure of the degree of bending (ie, the angle) of the sensing position measured by the bending sensor 260.
- the bending sensor 260 may be attached / coupled to the sensing position to measure the degree of bending (ie, the angle value) of the sensing position. For example, when the user 250 attaches the bending sensor 260 to the elbow, the bending sensor 260 may detect the degree of bending of the elbow of the user 250 (ie, the angle of the elbow). In this case, the bending sensor 260 may transmit the angle value of the elbow and the sensor characteristic of the bending sensor 260 to the virtual world processing apparatus 200.
- the bending sensor 260 may measure the degree of bending by being coupled to other joint parts such as an elbow, a finger joint, a neck, and a shoulder of the user 250. Sensor Capability is information on characteristics of the bending sensor 260.
- the sensor characteristics of the bending sensor 260 may include a maximum value (maxValue), a minimum value (mmValue), a number of locations, a distance between location, and a number of axes (numOfAxes). And the number of sensors (numOfSensors).
- the maximum value is the maximum angle value that the bending sensor 260 can measure.
- the minimum value is the minimum angle value that the bending sensor 260 can measure.
- the bending sensor 260 according to an embodiment may measure the angle of the sensing position within a range smaller than the maximum value and larger than the minimum value.
- the number of positions represents the number of sensing positions detected by the bending sensor 260.
- one bending sensor 260 can detect the degree of bending of the plurality of sensing positions at the same time, the number of positions at this time is a plurality of sensing positions measured by one bending sensor 260 It may be the number of.
- the plurality of bending sensors 260 may measure a plurality of sensing positions, and the number of positions may be the number of sensing positions measured by the plurality of bending sensors 260. have.
- the position distance represents the distance between the sensing positions detected by the bending sensor 260.
- the plurality of bending sensors 260 may detect a plurality of detection positions.
- the number of shafts represents the size of the angle value that the bending sensor can detect.
- the number of sensors indicates segments that the plurality of bending sensors 260 can detect when the plurality of bending sensors 260 measures a plurality of sensing positions.
- the sensor characteristic of the bending sensor 260 may further include an accuracy and an offset.
- Accuracy is information about the measurement error of the bending sensor 260.
- the offset is for zero adjustment of the angle value measured by the bending sensor 260.
- Table 1 shows XML syntax (extensible markup language syntax) for a bending sensor capability type according to an embodiment of the present invention.
- the bending sensor characteristic type represents information about basic sensor characteristics of the bending sensor 260.
- Table 2 shows semantics of a bending sensor characteristic type according to an embodiment.
- numOf Locations * 3 Describes the number of locations that a bending sensor can, sense bending angles.
- numOfSensors Describes the number of segments that an array of bending, sensors can perceive. ⁇
- the sensor characteristics may be data encoded in a binary format (binary format).
- the bending sensor encodes the sensor in binary form to generate metadata, and generates the metadata.
- the virtual world processor 200 may transmit the same.
- the receiver 210 may receive a sensor characteristic which is metadata encoded in binary form.
- Table 2-2 shows a binary representation syntax for a bending sensor characteristic type according to an embodiment of the present invention.
- Table 2-3 shows semantics for binary coding of a bending sensor characteristic type according to an embodiment of the present invention.
- attribute.A value of ⁇ implies that the attribute shall be used and a value of "0" implies that the attribute shall not be used.
- n mOfAxesFlag ⁇ This field, which is only present in the binary representation, indicates the presence of the "OUTOQ ⁇ S" attribute. A value of ⁇ "im ⁇ ies that the attribute shall be used and a value (rf" 0 "implies that the attribute shall not be used. * 3
- the type of data is 32 bit fitting valuer
- 2-bit data field is reserved to support up to 3 axes>
- 8-bit data field is reserved to support an array of bend sensors>
- the binary data encoded in the binary format includes data fields for at least one attribute of Num of Locations, NumOf ⁇ xes, and NumOfSensors. Data Field). Therefore, the sensor characteristics, which are metadata encoded in binary form, may include data fields for at least one of the position number, the number of axes, and the number of sensors, thereby limiting the size of the data. Can be. Table 3 shows an example of sensor characteristics for the bending sensor 260.
- the identifier ID of the bending sensor 260 may be 'BS001'.
- the maximum value of the bending degree (that is, the angle) of the sensing position that the bending sensor 260 can measure may be '9 degrees' and the minimum value may be '-30.0 degrees'.
- the number of levels may be '2500 level', and the accuracy of the bending sensor 260 may be ⁇ .1 degree.
- the offset for the bending sensor 260 may be ⁇ .05 degrees.
- Table 4 shows XML syntax for a bending sensor type according to an embodiment of the present invention.
- the bending sensor type indicates information on the sensing information measured by the bending sensor 260.
- the bending sensor type may include a time stamp, a unit, and a-value.
- the time stamp is time information when the bending sensor 260 scales the angle value of the sensing position.
- the time stamp may be information about a time interval at which the bending sensor 260 measures an angle value.
- the unit is the unit for the angle measured by the bending sensor 260.
- the value may include 'degree'.
- the value is the value of the angle measured by the bending sensor 260.
- the angle value for the sensing position is the sum of the plurality of angle values measured by the plurality of bending sensors 260. Can be.
- an operation in which the plurality of bending sensors 260 measures an angle value will be described in detail with reference to FIG. 3.
- FIG 3 illustrates an operation of measuring a sensing position by a plurality of bending sensors according to an embodiment of the present invention.
- a plurality of bending sensors 310 may measure a degree of bending (ie, an angle value) of the sensing position 320.
- the sum of individual angle values measured by the plurality of bending sensors 310 may be an angle value with respect to the sensing position 320.
- the individual angle values measured by the 10 bending sensors 310 are 1 degree, 3 degrees, 5 degrees, 10 For degrees, 12 degrees, 12 degrees, 9 degrees, 5 degrees, 2 degrees, and 1 degree
- the angle value for the sensing position 320 can be 60 degrees which is the sum of the individual angle values measured by the 10 bending sensors 310. have.
- Table 4-2 shows semantics of the bending sensor type according to the first embodiment.
- Table 4-4 shows semantics of binary coding of the bending sensor type according to the first embodiment of the present invention.
- Table 5-2 shows semantics of the bending sensor type according to the second embodiment.
- TimeStamp ⁇ 3 Describes the time that the information is acquired (sensed), unit * -1 Specifies the unit of Hie seised value, if a unit otiier than the default unit is used s as a reference to a classification scheme term provided by UnitTypeCS defined in ⁇ 2 ⁇ of ISO / IEC 23005-6. If the unit is ⁇ t defined here, the default unit is degrees
- Ar rayBend ingV lue ⁇ Describes tiie seised value by bending according to 3 -axis with respect to the default unit or the unit defined in the unil attribute. ⁇
- the Array Value defines by the number of locations and the number of axis. Ten by three matrix indicates that the arr3 ⁇ 4rValue has 10 sensing locations, each of which has 3 -axis bending angles. ⁇
- valuer Describes the sensed value by bending in scalar. This element shall be used, if flie sensor is capable of sensing only a degree ofbending.
- Table 6 shows the X] ⁇ L syntax for the bending sensor type according to the crab 3 embodiment.
- Table 6-2 shows semantics of the bending sensor type according to the third embodiment.
- TimeStan ⁇ Describes the time that the information is acquired (sens!).
- ArrayBendingValu * -1 Describes flie set of sensed values by the bending with respect to flie default unit or the unit defined in the unit attribute on each joints
- AnayValue defines by the number of axes and the number of locations. Three by ten martri indicates that the arrwValue has 10 sensing locations, each of which has 3 -axis bending angles.
- the order of values in each row of the matrix may be started from the fingertip to the palm side
- Table 7 shows a binary representation syntax for a bending sensor type according to an embodiment of the present invention.
- Table 7-2 shows semantics of binary coding of a bending sensor type according to an embodiment of the present invention.
- Table 8 shows an example of sensing information measured by the bending sensor 260.
- the identifier of the sensing information may be ending01 ', and the identifier of the bending sensor 260 may be 3 ⁇ 4endingID01'. That is, the sensing information measured by the bending sensor 260 of 3 ⁇ 4endingIDOr may be 3 ⁇ 4ending01 '.
- the time stamp of the bending sensor 260 may be '60000', which is a value of 100 clock ticks per second.
- the processor 220 may generate control information for controlling the object 280 of the virtual world 270 based on the sensing position based on the angle value and the sensor characteristic received by the receiver 210 from the bending sensor 260. Can be.
- the bending sensor 260 may measure the angle value and transmit it to the virtual world processing apparatus 200.
- the processor 260 controls the object 280 (for example, the avatar) of the virtual world 270 to move the elbow based on the detected angle (ie, the elbow) based on the angle value and the sensor characteristic. Control information can be generated.
- the processor 220 when the angle value measured by the bending sensor 260 is less than or equal to the maximum value that can be measured by the bending sensor 260, and greater than or equal to the minimum value, the processor 220 is configured to control Can be generated.
- the processor 220 may determine the angle value.
- the control information can be generated with the maximum value.
- the processor 220 may generate the control information with the angle value as the minimum value.
- the transmitter 230 may transmit the generated control information to the virtual world 270.
- the transmitter 230 may encode the control information into metadata in the XML format and transmit the encoded control information to the virtual world 270.
- the transmitter 230 may encode the control information into metadata in binary form and transmit the encoded control information to the virtual world 270.
- the transmitter 230 may encode the control information into the first metadata in the XML format, encode the first metadata into the second metadata in the binary format, and transmit the control information to the virtual world 270.
- the virtual world processing apparatus 200 may further include an operation unit 240.
- the calculation unit 240 is configured to determine a third sensing position located between the first sensing position and the second sensing position, based on the first angle value for the first sensing position and the ' second angle value for the second sensing position. We can compute three angle values for.
- the processor 220 may generate control information based on the first angle value, the second angle value, the third angle value, and the sensor characteristic.
- FIG. 4 illustrates an operation of calculating an angle value by a virtual world processing apparatus according to an embodiment of the present invention.
- the bending sensor 410 may measure a first angle value with respect to the first sensing position 401 (eg, the shoulder). Also, bending The sensor 420 may measure a second angle value with respect to the second sensing position 403 (eg, wrist).
- the virtual world processing apparatus is based on the first angle value and the second angle value, nearly 13 angle values of the third sensing position 402 that the bending sensors 410, 420 cannot measure. Can be calculated.
- the virtual world processing apparatus may set the average of the first angle value and the second angle value as the third angle value.
- the virtual world processing apparatus further includes a database that collects and stores a third angle value according to the first angle value and the second angle value, and includes the information, the first angle value, and the second angle stored in the database.
- the third angle value may be calculated based on the value.
- FIG. 5 is a diagram illustrating a structure of an avatar control feature point type according to an embodiment of the present invention.
- the avatar control feature type 510 includes attributes 520, control body features 530, and the like. Control Face Features 540 may be included.
- the avatar control feature point information is information about an feature point of the avatar used to control the avatar of the virtual world.
- the avatar control feature point type 510 is information about a type of avatar control feature point information.
- the attribute 520 is information about an attribute of avatar control feature point information.
- the attribute 520 may include name information on a feature point used to control the avatar.
- the control body feature point 530 is a feature point used to control the avatar and represents a feature point included in the body portion of the avatar.
- control body feature point 530 head skeleton (HeadBones) (531), the upper body armature (UpperBodyBones) (532), the lower body armature (DownBodyBones) (533) and jungche skeleton (MiddleBodyBones) (53 4) Can be classified.
- HeadBones head skeleton
- UpperBodyBones Upper body armature
- DownBodyBones DownBodyBones
- jungche skeleton MiddleBodyBones
- the upper body skeleton 532 may include information about feature points on the hand structure of the avatar.
- the hand structure may include substructures such as a solid neck and a finger joint.
- Table 9 shows a substructure of the hand structure according to one embodiment.
- the virtual world processing apparatus may generate control information for controlling the avatar control feature point based on the sensing position based on the sensing information measured by the bending sensor. That is, the sensing position measured by the bending sensor may be applied to at least one of the avatar control feature points.
- the virtual world processing apparatus may measure the movement of the wrist of the avatar of the virtual world corresponding to the user's wrist based on the measured angle value. Control information to control can be generated.
- FIG. 6 is a flowchart illustrating a virtual world processing method according to an embodiment of the present invention.
- an angle value of a sensing position and sensor characteristics of a bending sensor may be received from a bending sensor (S610).
- a bending sensor is a sensor that measures sensed information about the degree of bending of an object in the real world.
- the sensing information is a measure of the degree of bending (ie, the angle) of the sensing position measured by the bending sensor.
- the bending sensor may be attached / coupled to the sensing position to measure the degree of bending (ie, the angle value) of the sensing position.
- the bending sensor may detect the degree of bending of the user's elbow (ie, the angle of the elbow).
- the virtual world processing method may receive the elbow angle value and the sensor characteristics of the bending sensor from the bending sensor.
- the bending sensor according to an embodiment may measure the degree of bending by being coupled to other joint parts such as an elbow, a finger joint, a neck, and a shoulder of the user.
- Sensor Capability is information about the characteristics of the bending sensor.
- Sensor characteristics for a bending sensor may include a maximum value (maxValue), a minimum value (minValue), a number of locations (Nwn of Locations), a distance between locations, a number of axes (numOfAxes), and a number of sensors (numOfSensors). ) May be included.
- the maximum value is the maximum angle value that the bending sensor can measure.
- the minimum value is the minimum angle that the bending sensor can measure.
- the bending sensor according to an embodiment may measure the angle of the sensing position within a range smaller than the maximum value and larger than the minimum value.
- the number of positions represents the number of sensing positions detected by the bending sensor.
- a bending sensor is at the same time and to detect the "degree of bending of a plurality of detection positions, where the position count will be the number of a plurality of detection positions where the measured one of the bending sensor have.
- the plurality of bending sensors may measure a plurality of sensing positions, and the number of positions at this time may be the number of sensing positions measured by the plurality of bending sensors.
- the distance between positions represents the distance between the sensing positions detected by the bending sensors.
- a plurality of bending sensors may detect a plurality of detection positions.
- the number of shafts represents the size of the angle value that the bending sensor can detect.
- the number of sensors indicates segments that the plurality of bending sensors can detect when the plurality of bending sensors measures the plurality of sensing positions.
- the sensor characteristic of the bending sensor may further include an accuracy and an offset.
- Accuracy is information about the measurement error of the bending sensor.
- the sensor characteristics may be metadata encoded in a binary format (binary format).
- Meta data encoded in binary form includes a data field for at least one attribute of Num of Locations, numOfAxes, and numOfSensors. can do.
- the bending sensor type indicates information on the sensing information measured by the bending sensor. I
- the bending sensor type according to an embodiment may include a time stamp, a unit, and a value.
- the timestamp is time information when the bending sensor measures the angle value of the sensing position.
- the time stamp may be information on a time interval at which the bending sensor measures an angle value.
- the unit is the unit of the angle measured by the bending sensor.
- the value may include 'degree'.
- the value is the value of the angle measured by the bending sensor.
- the angle value for the sensing position may be the sum of the plurality of angle values measured by the plurality of bending sensors.
- the virtual world processing method depends on the angle value and sensor characteristics received from the bending sensor. On the basis of this, control information for controlling an object of the virtual world, which is based on the sensing position, may be generated (S620).
- the bending sensor may measure the angle value.
- the virtual world processing method generates control information for controlling the object of the virtual world (for example, the avatar) to move the elbow based on the angle value and the sensor characteristic. can do.
- the virtual world processing method may generate control information.
- the virtual world processing method may generate control information using the angle value as the maximum value.
- the virtual world processing method may generate control information with the angle value as the minimum value.
- the virtual world processing method may transmit the generated control information to the virtual world (S630).
- the virtual world processing method may encode control information into metadata in an XML format and transmit the control information to the virtual world. Also, the virtual world processing method may encode control information into binary metadata and transmit it to the virtual world. In addition, the virtual world processing method may encode the control information into the first metadata of the XML form, and encode the first metadata into the second metadata of the binary form to transmit to the virtual world.
- the virtual world method is based on a first detection position for the first Based on the angle value and the second angle value for the second sensing position, a third angle value for the third sensing position positioned between the first sensing position and the second sensing position may be calculated.
- the virtual world processing method may generate control information based on the first angle value, the second angle value, the third angle value, and the sensor characteristic.
- Embodiments according to the present invention can be implemented in a program command form that can be executed by various computer means may be recorded on a computer readable medium.
- the computer readable medium may include a program command, a data file, a data structure, etc. alone or in combination.
- Program instructions recorded on the media may be those specially designed and constructed for the present invention, or they may be of the kind well-known and available to those skilled in computer software.
- Examples of computer-readable recording media include magnetic media such as hard disks, floppy disks, and magnetic tape, optical media such as CD-ROMs, DVDs, and floppy disks. Includes: magneto-optical media, and hardware devices specifically configured to store and execute program instructions, such as ROM, RAM, flash memory, and the like.
- program instructions include machine language code, such as produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter.
- the hardware device described above may be configured to operate as one or more software modules to perform the operations of the present invention, and vice versa.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Cardiology (AREA)
- General Health & Medical Sciences (AREA)
- Heart & Thoracic Surgery (AREA)
- Business, Economics & Management (AREA)
- Computer Security & Cryptography (AREA)
- General Business, Economics & Management (AREA)
- User Interface Of Digital Computer (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201180018649.0A CN102858413B (zh) | 2010-04-13 | 2011-04-06 | 虚拟世界处理装置及方法 |
US13/641,026 US9597592B2 (en) | 2010-04-13 | 2011-04-06 | Method and apparatus for processing virtual world |
JP2013504805A JP5801872B2 (ja) | 2010-04-13 | 2011-04-06 | 仮想世界処理装置及び方法 |
EP11769009.9A EP2559463B1 (en) | 2010-04-13 | 2011-04-06 | Device and method for processing a virtual world |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR10-2010-0033947 | 2010-04-13 | ||
KR20100033947 | 2010-04-13 | ||
KR10-2011-0024448 | 2011-03-18 | ||
KR1020110024448A KR101835413B1 (ko) | 2010-04-13 | 2011-03-18 | 가상 세계 처리 장치 및 방법 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2011129543A2 true WO2011129543A2 (ko) | 2011-10-20 |
WO2011129543A3 WO2011129543A3 (ko) | 2012-02-02 |
Family
ID=45029517
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2011/002408 WO2011129543A2 (ko) | 2010-04-13 | 2011-04-06 | 가상 세계 처리 장치 및 방법 |
Country Status (6)
Country | Link |
---|---|
US (1) | US9597592B2 (ko) |
EP (1) | EP2559463B1 (ko) |
JP (2) | JP5801872B2 (ko) |
KR (1) | KR101835413B1 (ko) |
CN (2) | CN104281262B (ko) |
WO (1) | WO2011129543A2 (ko) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105056525A (zh) * | 2015-08-17 | 2015-11-18 | 联想(北京)有限公司 | 一种信息处理方法及电子设备 |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20100138700A (ko) * | 2009-06-25 | 2010-12-31 | 삼성전자주식회사 | 가상 세계 처리 장치 및 방법 |
KR101685922B1 (ko) * | 2010-04-05 | 2016-12-13 | 삼성전자주식회사 | 가상 세계 처리 장치 및 방법 |
US8610663B2 (en) | 2012-02-06 | 2013-12-17 | Lg Electronics Inc. | Portable device and method for controlling the same |
KR101367412B1 (ko) * | 2013-06-18 | 2014-02-24 | 우석 계 | 가상공간에서의 물리적 효과 제어 장치 및 그 방법 |
US20170259167A1 (en) * | 2016-03-14 | 2017-09-14 | Nathan Sterling Cook | Brainwave virtual reality apparatus and method |
US10627909B2 (en) | 2017-01-10 | 2020-04-21 | Disney Enterprises, Inc. | Simulation experience with physical objects |
CN107344019B (zh) * | 2017-07-28 | 2023-11-21 | 歌尔科技有限公司 | 一种虚拟现实系统 |
CN109579689B (zh) * | 2017-09-29 | 2020-12-08 | 西门子公司 | 曲率测量装置 |
GB2569603B (en) | 2017-12-21 | 2020-04-01 | Sony Interactive Entertainment Inc | Position tracking apparatus and method |
JP7070245B2 (ja) * | 2018-08-27 | 2022-05-18 | 富士通株式会社 | 情報処理装置、動作制御プログラム、及び動作制御方法 |
JP7359422B2 (ja) * | 2018-11-07 | 2023-10-11 | 学校法人法政大学 | 仮想体の動作制御システム、人体拘束治具、及び仮想体の動作制御方法 |
CN110822643B (zh) * | 2019-11-25 | 2021-12-17 | 广东美的制冷设备有限公司 | 空调器及其控制方法和计算机存储介质 |
CN110822646B (zh) * | 2019-11-25 | 2021-12-17 | 广东美的制冷设备有限公司 | 空调器的控制方法、空调器及存储介质 |
CN110822661B (zh) * | 2019-11-25 | 2021-12-17 | 广东美的制冷设备有限公司 | 空调器的控制方法、空调器及存储介质 |
CN110822642B (zh) * | 2019-11-25 | 2021-09-14 | 广东美的制冷设备有限公司 | 空调器及其控制方法和计算机存储介质 |
CN110822649B (zh) * | 2019-11-25 | 2021-12-17 | 广东美的制冷设备有限公司 | 空调器的控制方法、空调器及存储介质 |
CN110822641A (zh) * | 2019-11-25 | 2020-02-21 | 广东美的制冷设备有限公司 | 空调器及其控制方法、装置和可读存储介质 |
CN110822645B (zh) * | 2019-11-25 | 2021-12-17 | 广东美的制冷设备有限公司 | 空调器及其控制方法、装置和可读存储介质 |
CN110822644B (zh) * | 2019-11-25 | 2021-12-03 | 广东美的制冷设备有限公司 | 空调器及其控制方法和计算机存储介质 |
CN110822647B (zh) * | 2019-11-25 | 2021-12-17 | 广东美的制冷设备有限公司 | 空调器的控制方法、空调器及存储介质 |
JP2021089691A (ja) * | 2019-12-06 | 2021-06-10 | 未來市股▲ふん▼有限公司 | 動作追跡システム及び動作追跡方法 |
EP4119308A4 (en) * | 2020-03-11 | 2023-08-30 | Sony Group Corporation | INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING METHOD AND INFORMATION PROCESSING PROGRAM |
Family Cites Families (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH03228092A (ja) * | 1990-02-02 | 1991-10-09 | Hitachi Ltd | 文字パターン生成方法 |
JPH09153151A (ja) * | 1995-11-30 | 1997-06-10 | Matsushita Electric Ind Co Ltd | 3次元骨格構造の動き生成装置 |
JP3862348B2 (ja) * | 1997-03-19 | 2006-12-27 | 東京電力株式会社 | モーションキャプチャシステム |
AU7161598A (en) * | 1997-04-21 | 1998-11-13 | Virtual Technologies, Inc. | Goniometer-based body-tracking device and method |
US6049327A (en) * | 1997-04-23 | 2000-04-11 | Modern Cartoons, Ltd | System for data management based onhand gestures |
US5963891A (en) * | 1997-04-24 | 1999-10-05 | Modern Cartoons, Ltd. | System for tracking body movements in a virtual reality system |
JPH1157212A (ja) * | 1997-08-24 | 1999-03-02 | Sony Computer Entertainment:Kk | ゲーム装置、ゲーム機用操作装置、ゲームシステム及びゲーム装置の双方向通信方法 |
US6154199A (en) * | 1998-04-15 | 2000-11-28 | Butler; Craig L. | Hand positioned mouse |
US6304840B1 (en) * | 1998-06-30 | 2001-10-16 | U.S. Philips Corporation | Fingerless glove for interacting with data processing system |
JP2000132305A (ja) * | 1998-10-23 | 2000-05-12 | Olympus Optical Co Ltd | 操作入力装置 |
CN2358435Y (zh) * | 1999-03-10 | 2000-01-12 | 中国科学院自动化研究所 | 柔性角度传感器 |
JP2000329511A (ja) * | 1999-05-19 | 2000-11-30 | Sony Corp | データグローブ及びこれを用いた形状認識方法 |
US6577952B2 (en) * | 2001-01-08 | 2003-06-10 | Motorola, Inc. | Position and heading error-correction method and apparatus for vehicle navigation systems |
US20030062675A1 (en) | 2001-09-28 | 2003-04-03 | Canon Kabushiki Kaisha | Image experiencing system and information processing method |
JP3584229B2 (ja) * | 2001-09-28 | 2004-11-04 | キヤノン株式会社 | 映像体験システムおよび情報処理方法 |
US7033281B2 (en) * | 2002-03-22 | 2006-04-25 | Carnahan James V | Augmented kinematic feedback device and method |
CN1167996C (zh) * | 2002-07-11 | 2004-09-22 | 庄胜雄 | 可形成虚拟追踪球的手套滑鼠 |
JP2004192325A (ja) * | 2002-12-11 | 2004-07-08 | Nippon Telegr & Teleph Corp <Ntt> | ユーザ姿勢情報伝達方法,ユーザ姿勢情報送信装置,ユーザ姿勢情報受信装置,ユーザ姿勢情報送信プログラム,ユーザ姿勢情報受信プログラム,およびそれらのプログラムを記録した記録媒体 |
JP2004213350A (ja) | 2002-12-27 | 2004-07-29 | Seiko Epson Corp | 力覚提示装置及び画像補正方法 |
CN2725994Y (zh) * | 2004-05-13 | 2005-09-14 | 中国科学院自动化研究所 | 一种触觉反馈数据手套 |
JP2005337983A (ja) * | 2004-05-28 | 2005-12-08 | Spice:Kk | モーションキャプチャー方法及びシステム |
JP2009501063A (ja) * | 2005-07-15 | 2009-01-15 | パワーグリッド フィットネス インコーポレイテッド | 等尺性運動システムおよびビデオゲームのプレイ中のユーザーの運動を容易にする方法 |
JP2007041660A (ja) * | 2005-08-01 | 2007-02-15 | Nec Corp | センサネットワーク、センサノード及びそのプログラムと、センサ情報検索方法 |
US7602301B1 (en) * | 2006-01-09 | 2009-10-13 | Applied Technology Holdings, Inc. | Apparatus, systems, and methods for gathering and processing biometric and biomechanical data |
US20080125224A1 (en) | 2006-09-26 | 2008-05-29 | Pollatsek David | Method and apparatus for controlling simulated in flight realistic and non realistic object effects by sensing rotation of a hand-held controller |
JP4647632B2 (ja) * | 2007-03-06 | 2011-03-09 | 日本電信電話株式会社 | センサデータ制御システム及びセンサデータ制御方法 |
US8675017B2 (en) | 2007-06-26 | 2014-03-18 | Qualcomm Incorporated | Real world gaming framework |
US20090221368A1 (en) * | 2007-11-28 | 2009-09-03 | Ailive Inc., | Method and system for creating a shared game space for a networked game |
JP5108032B2 (ja) | 2008-02-01 | 2012-12-26 | 旭光電機株式会社 | 多関節構造体教示装置 |
KR101483713B1 (ko) * | 2008-06-30 | 2015-01-16 | 삼성전자 주식회사 | 모션 캡쳐 장치 및 모션 캡쳐 방법 |
JP5529400B2 (ja) | 2008-07-11 | 2014-06-25 | 任天堂株式会社 | 拡張操作装置 |
-
2011
- 2011-03-18 KR KR1020110024448A patent/KR101835413B1/ko not_active Application Discontinuation
- 2011-04-06 WO PCT/KR2011/002408 patent/WO2011129543A2/ko active Application Filing
- 2011-04-06 US US13/641,026 patent/US9597592B2/en active Active
- 2011-04-06 CN CN201410484088.9A patent/CN104281262B/zh active Active
- 2011-04-06 JP JP2013504805A patent/JP5801872B2/ja active Active
- 2011-04-06 CN CN201180018649.0A patent/CN102858413B/zh active Active
- 2011-04-06 EP EP11769009.9A patent/EP2559463B1/en active Active
-
2015
- 2015-08-27 JP JP2015167901A patent/JP6316251B2/ja active Active
Non-Patent Citations (1)
Title |
---|
"PROJECT NATAL", 2009, article "Press Conference. PROJECT NATAL" |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105056525A (zh) * | 2015-08-17 | 2015-11-18 | 联想(北京)有限公司 | 一种信息处理方法及电子设备 |
Also Published As
Publication number | Publication date |
---|---|
CN104281262B (zh) | 2017-08-01 |
EP2559463A2 (en) | 2013-02-20 |
EP2559463A4 (en) | 2014-02-26 |
JP5801872B2 (ja) | 2015-10-28 |
CN102858413B (zh) | 2014-10-29 |
US9597592B2 (en) | 2017-03-21 |
JP2013535039A (ja) | 2013-09-09 |
CN104281262A (zh) | 2015-01-14 |
EP2559463B1 (en) | 2020-10-07 |
JP6316251B2 (ja) | 2018-04-25 |
CN102858413A (zh) | 2013-01-02 |
US20130093665A1 (en) | 2013-04-18 |
WO2011129543A3 (ko) | 2012-02-02 |
KR101835413B1 (ko) | 2018-03-09 |
KR20110114439A (ko) | 2011-10-19 |
JP2016026342A (ja) | 2016-02-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2011129543A2 (ko) | 가상 세계 처리 장치 및 방법 | |
US9009747B2 (en) | Gesture cataloging and recognition | |
US8781151B2 (en) | Object detection using video input combined with tilt angle information | |
US9393487B2 (en) | Method for mapping movements of a hand-held controller to game commands | |
EP1937380B1 (en) | Wireless video game controller and method for operating a wireless video game controller | |
JP5714695B2 (ja) | 仮想世界処理装置及び方法 | |
CN101438340B (zh) | 用于三维输入控制的系统、方法和设备 | |
US20070265075A1 (en) | Attachable structure for use with hand-held controller having tracking ability | |
JP2016137248A (ja) | 仮想世界処理装置および方法 | |
US20110118021A1 (en) | Scheme for translating movements of a hand-held controller into inputs for a system | |
US20060287085A1 (en) | Inertially trackable hand-held controller | |
US20060256081A1 (en) | Scheme for detecting and tracking user manipulation of a game controller body | |
KR101636734B1 (ko) | 가상 세계 처리 장치 및 방법 | |
WO2007130833A2 (en) | Scheme for detecting and tracking user manipulation of a game controller body and for translating movements thereof into inputs and game commands | |
EP3711828B1 (en) | Scheme for detecting and tracking user manipulation of a game controller body and for translating movements thereof into inputs and game commands | |
USRE48417E1 (en) | Object direction using video input combined with tilt angle information | |
CN102058976A (zh) | 用于跟踪环境中的用户操纵的系统 | |
EP2013864A2 (en) | System, method, and apparatus for three-dimensional input control | |
KR20120106369A (ko) | 가상 세계 처리 장치 및 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180018649.0 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11769009 Country of ref document: EP Kind code of ref document: A2 |
|
ENP | Entry into the national phase |
Ref document number: 2013504805 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011769009 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13641026 Country of ref document: US |