WO2023189422A1 - Control device, control method, haptic presentation system, and program product - Google Patents

Control device, control method, haptic presentation system, and program product Download PDF

Info

Publication number
WO2023189422A1
WO2023189422A1 PCT/JP2023/009309 JP2023009309W WO2023189422A1 WO 2023189422 A1 WO2023189422 A1 WO 2023189422A1 JP 2023009309 W JP2023009309 W JP 2023009309W WO 2023189422 A1 WO2023189422 A1 WO 2023189422A1
Authority
WO
WIPO (PCT)
Prior art keywords
displacement
amount
tactile
control data
tactile presentation
Prior art date
Application number
PCT/JP2023/009309
Other languages
French (fr)
Japanese (ja)
Inventor
修一 赤岩
慶介 木野井
Original Assignee
株式会社栗本鐵工所
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社栗本鐵工所 filed Critical 株式会社栗本鐵工所
Publication of WO2023189422A1 publication Critical patent/WO2023189422A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials

Definitions

  • the present invention relates to a control device, a control method, a tactile presentation system, and a program product that improve the perceptibility of tactile sensations presented by a tactile presentation device.
  • Patent Document 1 Based on knowledge of magnetorheological fluids, the inventors have proposed a tactile presentation device that can be operated by an operator and reproduces different sensations depending on the object (Patent Document 1, etc.).
  • the tactile presentation device includes a displacement section that is a curved plate or the like that can be displaced relative to the base by an operator's operation.
  • the displacement part can be displaced by the operator without resistance when the tactile presentation device is not controlled, but the tactile presentation device under control causes the displacement portion to be displaced according to the amount of displacement (position) of the displacement portion.
  • the tactile sensation is presented by changing the magnitude of the force sense, the output method of the force sense, etc.
  • the manner in which the displacement portion of this tactile presentation device is displaced differs depending on the size of the operator's hands, differences in muscles, etc. Further, depending on the operator, when an operation to displace the displacement portion is performed, the ease of perceiving the sensation as a touch differs depending on the timing at which the force sensation is output from the displacement portion. Therefore, it is desirable to be able to set the tactile presentation device so that each operator can easily perceive the tactile sensation.
  • the present invention has been made in view of the above circumstances, and aims to provide a control device, a control method, a tactile presentation system, and a program product that improve the perceptibility of tactile sensations presented by a tactile presentation device. do.
  • a control device includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to the amount of displacement of the displacement section.
  • a control device for a tactile presentation device that displays a tactile sensation of a displayed object, the control device storing control data for generating the tactile sensation for each of the objects in association with a displacement amount of the displacement portion. and corrects the correspondence between the displacement amount of the control data and the control data based on the settings, and outputs the corrected control data associated with the displacement amount of the displacement portion to the tactile presentation device.
  • a control method includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to a displacement amount of the displacement section.
  • a control method for a tactile presentation device that presents a tactile sensation of a displayed object by causing a computer connected to the tactile presentation device to generate the haptic sensation in association with a displacement amount of the displacement portion.
  • Control data is stored for each object, and based on settings, the correspondence between the displacement amount of the control data and the control data is corrected, and the corrected control is made in correspondence with the displacement amount of the displacement part. It includes a processing unit that outputs data to the tactile presentation device.
  • a program product includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to the amount of displacement of the displacement section.
  • control device when the displacement portion of the tactile presentation device generates a force sensation according to the amount of displacement, the tactile sensation is presented after the displacement portion starts displacement. are corrected individually.
  • a control device receives, as the setting, a shift in the displacement amount for starting to output the control data that continuously changes in ascending order of the displacement amount.
  • control data is set so that the tactile sensation of the object can be continuously reproduced as the displacement amount changes in ascending order.
  • the discrepancies up to that point are corrected individually.
  • a control device receives the settings on a settings screen that includes a slide bar that indicates a position of deviation with respect to the range of the displacement amount.
  • control device of the present disclosure it is possible to set the deviation as a continuous value using a slide bar.
  • a control device receives the settings on a settings screen that selectably displays a plurality of candidates having different amounts of deviation with respect to the range of displacement amounts.
  • the control device of the present disclosure allows selection from multiple candidates.
  • a tactile presentation system includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to the amount of displacement of the displacement section.
  • a tactile presentation device that generates and presents a tactile sensation of a displayed object; and an information processing device that is communicatively connected to the tactile presentation device and includes a display section, and the information processing device is configured to detect a displacement amount of the displacement section.
  • Control data that causes the force sensation and visual data to be displayed on the display unit are stored for each object in association with the control data and the control data and Correct the correspondence with the visual data, output the corrected control data associated with the displacement amount of the displacement part to the tactile presentation device, and calculate the displacement of the displacement part based on the corrected visual data.
  • An image is output to the display section according to the amount.
  • the displacement part of the tactile presentation device when the displacement part of the tactile presentation device generates a sense of force according to the amount of displacement, the time from when the displacement part starts to be displaced until the tactile sense is presented is individual to the operator. It is corrected to
  • the present disclosure it is possible to adjust the tactile presentation by the tactile presentation device according to the operator, and it is possible to improve the degree of tactile perception.
  • FIG. 1 is a schematic diagram showing a tactile presentation system.
  • FIG. 1 is a block diagram showing the configuration of an information processing device. It is an explanatory view showing an example of contents of a sensory DB.
  • FIG. 2 is a block diagram showing the configuration of a tactile presentation device. 2 is a flowchart illustrating an example of a basic processing procedure for tactile presentation in the tactile presentation system.
  • FIG. 2 is a schematic diagram showing an example of tactile presentation in the tactile presentation system.
  • 3 is a flowchart illustrating an example of a setting method. An example of a setting screen displayed on the display unit is shown.
  • 7 is a flowchart illustrating an example of a correction procedure based on a presentation start position. It is an explanatory view showing the result of correction.
  • FIG. 2 is a schematic diagram showing an example of tactile presentation in the tactile presentation system.
  • FIG. 2 is a schematic diagram of a tactile presentation system in a second embodiment. It is a block diagram showing the composition of HMD of a 2nd embodiment.
  • FIG. 1 is a schematic diagram showing a tactile presentation system 100.
  • the tactile presentation system 100 includes an information processing device 1 and a tactile presentation device 2.
  • the information processing device 1 and the tactile presentation device 2 are communicably connected by short-range wireless communication and exchange data with each other.
  • the information processing device 1 uses a smartphone as shown in FIG.
  • the information processing device 1 may be a tablet terminal or a laptop-type PC (Personal Computer) instead of a smartphone.
  • the tactile presentation device 2 is a device that an operator can operate by moving his or her finger while holding it along the displacement portion 202.
  • the tactile presentation device 2 reads the position of the displacement part 202 that is displaced by the operator's finger movement, controls the built-in MRF (Magneto-Rheological Fluid) device 24 according to the position, and moves the displacement part 202 of the operator.
  • MRF Magnetic-Rheological Fluid
  • This is a device that produces a sense of force using a reaction force (rotational resistance) to the operation of the machine, and presents a tactile sensation.
  • the form of the displacement part 202 of the tactile presentation device 2 is not limited to that shown in FIG. 1, but may be stick-like or cushion-like covered with a cover.
  • the tactile presentation device 2 may employ a motor, a piezo element, or the like instead of the MRF device 24, and may generate a force sensation by rotational force or vibration in response to an operation by the operator. It may be combined with something that provides a warm or cold sensation.
  • the structure may be such that it is installed on the ground or a wall and is operated by the operator's palm or foot.
  • the tactile presentation device 2 cooperates with the information processing device 1, and causes the information processing device 1 to output the image and audio of the object, while causing the tactile presentation device 2 to output the tactile sensation of the object.
  • the tactile presentation device 2 becomes squishy or fluffy.
  • the display unit 13 displays an image that changes as if the object is being pushed in, and the audio output unit 14 outputs an audible “squeak” sound that corresponds to the object being pushed in. Output audio.
  • FIG. 2 is a block diagram showing the configuration of the information processing device 1.
  • the information processing device 1 includes a processing section 10, a storage section 11, a communication section 12, a display section 13, an audio output section 14, and an operation section 15.
  • the processing unit 10 is a processor using a CPU (Central Processing Unit) and/or a GPU (Graphics Processing Unit).
  • the processing unit 10 executes a process described below based on a control program P1 for tactile presentation stored in the storage unit 11.
  • the storage unit 11 uses nonvolatile memory such as flash memory and SSD (Solid State Drive).
  • the storage unit 11 stores data referenced by the processing unit 10.
  • the control program P1 program product
  • the control program P1 stored in the storage unit 11 may be the control program P8 stored in the computer-readable storage medium 8, which is read out and stored by the processing unit 10.
  • the storage unit 11 stores a sensory database (DB) 110 that includes tactile data, visual data, and auditory data of the object to be output.
  • the sensory DB 110 stores tactile data, visual data (images, videos), and auditory data to be output at each displacement for each displacement data of a displacement part in the tactile presentation device 2, in association with an object ID that identifies the object. (audio) (see Figure 3).
  • the storage unit 11 further stores setting data for each operator.
  • the communication unit 12 is a communication module for short-range wireless communication, for example, Bluetooth (registered trademark).
  • the processing unit 10 can transmit and receive data to and from the tactile presentation device 2 through the communication unit 12 .
  • the display unit 13 is a display such as a liquid crystal display or an organic EL (Electro Luminescence) display.
  • the display unit 13 is, for example, a display with a built-in touch panel.
  • the processing unit 10 displays, on the display unit 13, an operation screen for causing the tactile sensation presentation device 2 to output a tactile sensation and an image of the target object, based on the control program P1.
  • the audio output unit 14 includes a speaker and the like.
  • the processing unit 10 causes the audio output unit 14 to output the sound of the object, music, etc. based on the control program P1.
  • the operation unit 15 is a user interface that allows input and output to and from the processing unit 10, and is a touch panel built into the display unit 13.
  • the operation unit 15 may be a physical button.
  • the operation unit 15 may be a voice input unit.
  • FIG. 3 is an explanatory diagram showing an example of the contents of the sensory DB 110.
  • the sensory DB 110 stores the current value of the current to the MRF device of the tactile presentation device 2 for each displacement amount (angle) of the displacement part 202 as tactile data in association with the object ID.
  • the objects are plural, and may include real objects that present a tactile sensation, such as still life objects such as balls or balloons, foods such as gummies, vegetables, or fruits, and animals such as dogs, cats, or fish. Characters such as slimes and monsters may also be included.
  • the relationship between the amount of displacement of the displacement section 202 of the sensory DB 110 and the current value is set according to the actual dimensions of the movable range of the displacement section 202 and the dimensions of the object. For example, if the movable range of the displacement unit 202 is 50 mm, and the object is a sphere that will not burst, such as a small gummy bear, and the size is set to 15 mm, the tactile sensation will be reproduced within the first 15 mm, In the remaining 35 mm range, a current value is set that produces a rotational resistance that makes it difficult to push in any further. Alternatively, a large current value that makes it impossible to push in the range of 12 mm to 15 mm may be set, and the current values thereafter may be set to zero.
  • the tactile data may be such as to present the sensation up to 50 mm from the surface of the sofa seat surface.
  • the current value is set. If the target object is a virtual character such as a monster, the current value is set based on the size according to the setting of the character.
  • the sensory DB 110 stores images (frame images) for each position (angle) of the displacement part 202 as visual data in association with the object ID.
  • the frame image is one still image that is recognized as an animation image when displayed continuously.
  • audio data for each position (angle) of the displacement section 202 is stored as auditory data in association with the object ID.
  • the auditory data may be waveform data that differs depending on each angle. It may also be a timestamp of the audio corresponding to each angle.
  • FIG. 4 is a block diagram showing the configuration of the tactile presentation device 2.
  • the tactile presentation device 2 is configured by providing a flat bottomed cylindrical grip 200 with a band-shaped flat plate displacement portion 202 having a curved portion partially along the circumferential direction.
  • the displacement part 202 is made of a flexible material, it may also be made of a highly rigid material and rotatably supported via the grip body 200 and a support shaft.
  • a binding tool 203 in the form of a cloth tape is provided on the outer surface of the distal end of the displacement portion 202 .
  • a link mechanism 204 is provided on the inner surface of the distal end of the displacement portion 202 to connect to the rotating shaft of the rotor of the MRF device 24 housed inside the gripper 200 .
  • the operator uses the gripping body 200 by, for example, holding it with the thumb and middle finger and inserting the index finger into the binding tool 203 with the finger such as the index finger along the displacement part 202.
  • the operator can move the displacement part 202 by pushing in the index finger, or can move the displacement part 202 away from the grip body 200 by extending the index finger.
  • the tactile presentation device 2 includes a gripping body 200 as shown in FIG.
  • the grip body 200 has an MRF device 24 built therein.
  • the control section 20, the storage section 21, the communication section 22, and the power supply section 23 may be provided integrally with the grip body 200, or may be provided as separate bodies connected to the grip body 200 wirelessly or by wire.
  • the control unit 20 includes a processor such as a CPU and an MPU (Micro-Processing Unit), and a memory such as a ROM (Read Only Memory) and a RAM (Random Access Memory).
  • the control unit 20 is, for example, a microcontroller.
  • the control unit 20 controls each component based on a control program P2 stored in the built-in ROM to realize tactile presentation.
  • the storage unit 21 is an auxiliary storage memory for the control unit 20, and stores control data (tactile data) of the MRF device 24 in a rewritable manner.
  • the communication unit 22 is a communication module for short-range wireless communication, for example, Bluetooth (registered trademark).
  • the control unit 20 can transmit and receive data to and from the information processing device 1 through the communication unit 22 .
  • the control unit 20 is connected to the power supply unit 23, the MRF device 24, and the sensor 25 via I/O, and exchanges signals with each other.
  • the power supply unit 23 includes a rechargeable battery.
  • the power supply unit 23 supplies power to each component and the MRF device 24 when turned on.
  • the MRF device 24 has a yoke that sandwiches a disk-shaped rotor with a gap between them, and generates a magnetic field by passing a control current through a coil provided in the yoke, thereby generating a magnetic field sealed in the gap.
  • the viscosity (shear stress) of the viscous fluid is controlled to provide rotational resistance to the rotor.
  • the control unit 20 controls the magnitude of the control current to the MRF device 24, the rotational resistance is immediately changed.
  • the displacement section 202 when the displacement section 202 is operated by the operator, the displacement of the displacement section 202 is caused to move in the direction of rotation of the MRF device 24 toward the rotation axis of the rotor via the link mechanism 204. communicated. Since the rotating shaft rotates freely when the MRF device 24 is not operating, that is, while the control current is zero, the displacement portion 202 fluctuates without resistance. On the other hand, when the MRF device 24 operates and the control current is not zero, the viscosity (shear stress) of the magnetorheological fluid inside the MRF device 24 is changed depending on the magnitude of the current flowing to the MRF device 24. By continuously changing the magnitude of the current to the MRF device 24 or by vibrating the current value at a predetermined frequency, the control unit 20 can change the resistance force applied to the displacement unit 202 and the method of its appearance. .
  • the tactile sensation presentation device 2 varies the resistance (current value) according to the amount of pushing of the displacement section 202 to present a smooth tactile sensation, or increases the resistance as the pushing amount increases to create a tighter feel. It is possible to present a tactile sensation with a certain hardness, or to present a crunchy tactile sensation by repeatedly increasing and decreasing the resistance.
  • FIG. 5 is a flowchart illustrating an example of a basic processing procedure for tactile presentation in the tactile presentation system 100.
  • the processing unit 10 of the information processing device 1 starts the following process in cooperation with the tactile presentation device 2.
  • the processing unit 10 displays on the display unit an operation screen that includes a list of object candidates for which tactile sensation is to be displayed (step S101), and accepts the selection of the object (step S102).
  • the processing unit 10 reads tactile data, image data, and audio data corresponding to the object ID of the selected object from the sensory DB 110 of the storage unit 11 (step S103).
  • the processing unit 10 corrects the read tactile data, image data, and audio data based on the presentation start position set in a process described later, and temporarily stores the corrected data (step S104).
  • the correction method in step S104 will be described in detail later.
  • the processing unit 10 establishes communication between the tactile presentation device 2 and the communication unit 12 (step S105), and transmits the temporarily stored corrected tactile data to the tactile presentation device 2 (step S106).
  • the processing unit 10 causes the display unit 13 to display that tactile presentation is to be started (step S107).
  • the control unit 20 of the tactile presentation device 2 Upon receiving the tactile data (step S201), the control unit 20 of the tactile presentation device 2 stores it in the storage unit 21 (step S202).
  • the control unit 20 samples a signal corresponding to the displacement amount (angle) of the displacement unit 202 output from the sensor 25 (step S203).
  • the control unit 20 transmits the displacement amount obtained by sampling to the information processing device 1 (step S204), and refers to the current value corresponding to the obtained displacement amount from the tactile data stored in the storage unit 21 (step S204).
  • S205 outputs the referenced current to the MRF device 24 (step S206), and transmits the process to step S203.
  • the processing in steps S203 to S206 continues until an operation to end the processing is performed on the information processing device 1 side.
  • the information processing device 1 receives the displacement amount from the tactile presentation device 2 (step S108), and the processing section 10 refers to the image and audio corresponding to the received displacement amount from the visual data and auditory data in the storage section 11 ( Step S109) and output from the display section 13 and audio output section 14 (Step S110), and the process returns to Step S108.
  • the information processing device 1 outputs an image and audio corresponding to the amount of displacement.
  • the processing unit 10 determines whether a termination operation has been performed (step S111). If it is determined that the end operation has not been performed (S111: NO), the process returns to step S108, and the processes of steps S108-S110 are repeated.
  • the processing unit 10 ends the display, disconnects communication with the tactile presentation device 2 (step S112), and ends the process.
  • FIG. 6 is a schematic diagram showing an example of tactile presentation in the tactile presentation system 100.
  • FIG. 6 shows changes in images and sounds displayed on the display unit 13 in response to the amount of displacement in the tactile data transmitted to the tactile presentation device 2.
  • the image changes depending on the amount of depression of the displacement part 202, and the way the depression corresponds to the tactile sensation changes.
  • contact that is, from the initial position (displacement amount is zero) output to the MRF device 24, to the target object in the image.
  • An image of the touching finger may be displayed.
  • the information processing device 1 corrects the tactile data in step S104 of the processing procedure shown in the flowchart of FIG.
  • the output of the device 2 can be tailored to each operator.
  • FIG. 7 is a flowchart illustrating an example of a setting method.
  • the processing procedure shown below is started when the setting menu is selected within the operation screen displayed based on the control program P1.
  • the processing procedure shown in FIG. 7 may be automatically started when the processing based on the control program P1 is executed for the first time. Initially, the amount of displacement corresponding to the presentation start position is zero.
  • the processing unit 10 displays a setting screen including a bar corresponding to the angular range of the displacement amount (step S301).
  • the processing unit 10 receives the selection of the presentation start position using the bar (step S302), determines and stores the amount of displacement (angle) corresponding to the selected presentation start position (step S303).
  • the processing unit 10 determines and stores the position of the displacement amount in the displacement amount range (0 to 90°) from the ratio of the length of the selected position to the length of the bar. That is, when the length of the bar is 100, in the case where the length to the selected position is 5, it is about 5%, so 5° is determined as the displacement amount corresponding to the presentation start position.
  • the processing unit 10 creates tactile data in which the current value up to the amount of displacement (angle) corresponding to the presentation start position is set to zero, and a predetermined current value is set to the amount of displacement greater than or equal to the displacement amount corresponding to the presentation start position. (Step S304).
  • the processing unit 10 transmits the created tactile data to the tactile presentation device 2 (step S305), and causes the display unit 13 to display a message urging you to try it on the setting screen (step S306).
  • the control unit 20 of the tactile presentation device 2 Based on the received tactile data, the control unit 20 of the tactile presentation device 2 causes a current to flow through the MRF device 24 according to the amount of displacement of the displacement portion, so that the amount of displacement of the displacement portion 202 is smaller than the amount of displacement corresponding to the set presentation start position. It is possible to perform a test in which a predetermined reaction force (rotational resistance) is generated only when the holder is pushed in. This allows the operator to test the amount of time (play/displacement) before starting to feel a reaction force.
  • a predetermined reaction force rotational resistance
  • the processing unit 10 displays a determination button on the setting screen (step S307), and accepts whether or not the presentation start position selected in step S302 is acceptable as a result of the trial.
  • the processing unit 10 determines whether the OK button has been selected (step S308). If it is determined that the enter button has not been selected (S308: NO), the processing unit 10 returns the process to step S302 and repeats the process until the enter button is selected.
  • the processing unit 10 stores the displacement amount (angle) determined in step S303 in the storage unit 11 as a correction amount (step S309), and ends the process. do.
  • FIG. 8 shows an example of the setting screen 130 displayed on the display unit 13.
  • the processing unit 10 displays the settings screen 130 on the display unit 13.
  • the settings screen 130 includes a bar interface 131 that includes a control 132 that is slidable over a length that corresponds to a range of displacement amounts. The operator can move the control 132 up and down on the touch panel built into the display section 13.
  • the setting screen 130 includes a test button 133 and a decision button 134.
  • the processing unit 10 detects this using the operation unit 15, and determines the displacement amount in step S303 based on the position of the moved control 132. If the test button 133 is selected, the processing unit 10 executes the processes from step S304 to step S308.
  • the processing unit 10 executes steps S308-S309.
  • FIG. 9 is a flowchart showing an example of a correction procedure based on the presentation start position
  • FIG. 10 is an explanatory diagram showing the results of the correction.
  • the processing unit 10 creates corrected tactile data by adding the displacement amount stored as the correction amount to each displacement amount of the tactile data read in step S104 (step S401). Thereby, it is possible to shift the amount of displacement at which the continuously changing current value in the tactile data starts to be output.
  • the processing unit 10 creates corrected auditory data by adding the displacement amount stored as the correction amount to each displacement amount of the read auditory data (step S402).
  • the processing unit 10 creates corrected visual data by adding the displacement amount stored as the correction amount to each displacement amount of the read visual data (step S403), and ends the correction.
  • the tactile data, auditory data, and visual data become data in which current values, audio data, and image data are associated with each displacement amount after the correction amount is added, as shown in FIG. . That is, the range of the amount of displacement (movable range) of the displacement portion 202 is corrected so as to deviate from the range of the presented sensation.
  • FIG. 11 shows another example of the setting screen 130 displayed on the display unit 13.
  • the processing unit 10 displays the setting screen 130 on the display unit 13.
  • three patterns are displayed as candidates, with respect to the range of displacement amount, with different patterns of how far from the top the tactile presentation should start (tactile presentation position).
  • the setting screen 130 includes a test button 133 and a decision button 134. It is possible to select one by pressing the decision button 134.
  • the center candidate is highlighted by a thick line frame (cursor), indicating that this candidate has been selected by tapping on the operation unit 15. There is.
  • the processing unit 10 detects this using the operation unit 15, and determines the displacement amount in step S303 based on the position of the moved control 132. If the test button 133 is selected, the processing unit 10 executes the processes from step S304 to step S308.
  • FIG. 12 is a schematic diagram showing an example of tactile presentation in the tactile presentation system 100.
  • changes in the image and sound displayed on the display unit 13 are shown in response to the amount of displacement in the corrected tactile data.
  • the state of the tactile data before correction is shown by a broken line.
  • the tactile sensation starts to be presented from the displacement section 202 from a position where the displacement section 202 is pushed in a little (equivalent to 5 degrees), and similarly, images and sounds is output.
  • the storage unit 11 of one information processing device 1 stores correction amounts for different operators, and after receiving the selection of the operator who operates the tactile presentation device 2, the processing unit 10 , the process shown in the flowchart of FIG. 5 may be executed using the set correction amount.
  • FIG. 13 is a schematic diagram of the tactile presentation system 100 in the second embodiment
  • FIG. 14 is a block diagram showing the configuration of the HMD 3.
  • the configuration of the tactile presentation system 100 in the second embodiment is the same as that of the tactile presentation system 100 in the first embodiment, except that the HMD 3 is used and the details of the processing by using the HMD 3 are different. . Therefore, among the configurations of the tactile presentation system 100 of the second embodiment, the same components as those of the tactile presentation system 100 of the first embodiment are given the same reference numerals, and detailed description thereof will be omitted.
  • the HMD 3 includes a display section 31, a motion detection section 32, a space detection section 33, and a connection section 34.
  • the HMD 3 may be provided with a display section 31, a motion detection section 32, a space detection section 33, and a connection section 34 in the main body, or some may be provided separately and mutually exchange control signals via a communication medium. It may be possible to do so.
  • the HMD 3 may have the functions of the information processing device 1 and be configured integrally.
  • the display unit 31 includes, for example, a small liquid crystal display, an optical lens, and an optical system mechanism, and is capable of displaying a three-dimensional image at a viewing angle of 110° or more.
  • the display unit 31 is a transparent or semi-transparent glass-like display, and receives an image signal (including a video signal) output from the information processing device 1 and displays it in a superimposed manner in the operator's actual field of view.
  • the display section 31 is not limited to a glass-like one, and may display an image based on an image signal output from the information processing device 1 superimposed on an image taken of real space by a camera provided facing forward. It's okay.
  • the motion detection unit 32 includes a plurality of triaxial acceleration sensors and gyro sensors installed in various directions on the main body (the cover of the display unit 31 and the attachment belt), and aggregates signals from a group of these sensors. and a control circuit that outputs the output.
  • the movement detection unit 32 detects the movement of the wearer's head.
  • the space detection unit 33 includes two or more infrared cameras arranged outwardly in parallel on the outer surface of the main body, and an infrared LED disposed at an intermediate position between the infrared cameras so as to similarly emit infrared rays outward. use
  • the space detection unit 33 functions as a depth sensor that measures the distance to an object that exists outside the main body of the HMD 3.
  • the space detection unit 33 can measure and output the distance from the HMD 3 main body to a still object such as a wall or floor in the location of the operator wearing the HMD 3.
  • the space detection unit 33 can measure and output the distance from the HMD 3 main body to the operator's arm, hand, and finger.
  • the connection unit 34 is an interface for connecting to the information processing device 1.
  • the HMD 3 outputs signals corresponding to the results measured by the motion detection unit 32 and the space detection unit 33 to the information processing device 1, and also acquires the video signal output from the information processing device 1 and displays it on the display unit. 31.
  • a speaker may be provided in the HMD 3, and the audio output from the audio output unit 14 of the information processing device 1 may be output from the speaker.
  • the information processing device 1 uses the HMD 3 and the distance to the still object in the real space measured by the space detection unit 33 to determine the three-dimensional shape of the virtual object.
  • the original image is superimposed and displayed in accordance with the operator's movement detected by the movement detection unit 32 to realize AR display.
  • the space detection section 33 uses the tactile presentation device 2 to select the coordinates in the three-dimensional image of the virtual object displayed on the display section 31 of the HMD 3 (coordinates that match the real space).
  • the processing unit 10 of the information processing device 1 detects this as being selected.
  • the operator can visually observe an image in which the object is virtually placed in real space, and feel the touch of the object using the tactile presentation device. It is possible to memorize it in 2.
  • the tactile presentation start position in the tactile presentation device 2 can be corrected according to each operator, thereby further improving the degree of perception. .

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The present invention provides a control device, a control method, a haptic presentation system, and a program product for improving the perceptibility of haptic sensations presented by a haptic presentation device. The control device is a control device of a haptic presentation device having a displacement unit provided as to be displaceable in response to an operation by an operator, the haptic presentation device generating a kinesthetic sensation in response to an operation on the displacement unit in accordance with the magnitude of displacement of the displacement unit to present a haptic sensation of a displayed object, wherein control data for generating the kinesthetic sensation is stored on a per-object basis in correspondence with the magnitude of displacement of the displacement unit, and the control device corrects the correspondence relationship between the control data and the magnitude of displacement for the control data on the basis of a setting and outputs corrected control data associated with the magnitude of displacement of the displacement unit to the haptic presentation device.

Description

制御装置、制御方法、触覚提示システム及びプログラム製品Control devices, control methods, tactile presentation systems and program products
 本発明は、触覚提示装置にて提示される触覚の知覚度を向上させる制御装置、制御方法、触覚提示システム及びプログラム製品に関する。 The present invention relates to a control device, a control method, a tactile presentation system, and a program product that improve the perceptibility of tactile sensations presented by a tactile presentation device.
 人間の五感のうち、視覚及び聴覚に係る画像及び音声の伝達技術が高精度化しており、昨今では、触覚を提示する技術(ハプティクス)が種々提案されている。 Of the five human senses, image and audio transmission technologies related to visual and auditory senses are becoming more precise, and recently, various technologies for presenting tactile sensations (haptics) have been proposed.
 触覚の提示技術としては、タッチパネル内蔵のディスプレイに表示された物体の画像に触れた際の振動、主に三種の振動(ERM:Eccentric Rotating Mass、LRA:Linear resonant Actuator、ピエゾ素子)での再現が提案されている。しかしながらタッチパネル上での触覚提示では、より深部での触覚、つまり手指の皮膚のみならず、筋や腱で生じる力覚まで再現することは困難である。 As a tactile presentation technology, it is possible to reproduce the vibrations that occur when an image of an object displayed on a display with a built-in touch panel is touched, mainly using three types of vibrations (ERM: Eccentric Rotating Mass, LRA: Linear resonant actuator, piezo element). Proposed. However, when presenting tactile sensations on a touch panel, it is difficult to reproduce deeper tactile sensations, that is, not only the skin of fingers but also the force sensations generated in muscles and tendons.
 発明者らは、磁気粘性流体に対する知見に基づいて、操作者が操作可能であって、対象物によって異なる感触を再現する触覚提示装置を提案した(特許文献1等)。 Based on knowledge of magnetorheological fluids, the inventors have proposed a tactile presentation device that can be operated by an operator and reproduces different sensations depending on the object (Patent Document 1, etc.).
特許第6906275号Patent No. 6906275
 触覚提示装置は、操作者の操作によって基部に対して変位可能な湾曲板等である変位部を備える。変位部は、触覚提示装置が制御されない状態では、操作者によって抵抗なく変位が可能であるが、制御中の触覚提示装置は、変位部の変位量(ポジション)に応じて変位部に対し生じさせる力覚の大きさ、力覚の出力方法等を変えて感触を提示する。 The tactile presentation device includes a displacement section that is a curved plate or the like that can be displaced relative to the base by an operator's operation. The displacement part can be displaced by the operator without resistance when the tactile presentation device is not controlled, but the tactile presentation device under control causes the displacement portion to be displaced according to the amount of displacement (position) of the displacement portion. The tactile sensation is presented by changing the magnitude of the force sense, the output method of the force sense, etc.
 この触覚提示装置の変位部の変位の仕方は、操作者の手の大きさの相違、筋肉の相違等によって異なる。また、操作者によっては、変位部を変位させる操作をした場合に、変位部において力覚が出力されるタイミングにより、感触としての知覚のし易さが相違する。そこで、触覚提示装置を、個々の操作者にとって触覚を知覚し易く設定できるようにすることが望まれる。 The manner in which the displacement portion of this tactile presentation device is displaced differs depending on the size of the operator's hands, differences in muscles, etc. Further, depending on the operator, when an operation to displace the displacement portion is performed, the ease of perceiving the sensation as a touch differs depending on the timing at which the force sensation is output from the displacement portion. Therefore, it is desirable to be able to set the tactile presentation device so that each operator can easily perceive the tactile sensation.
 本発明は、斯かる事情を鑑みてなされたものであり、触覚提示装置にて提示される触覚の知覚度を向上させる制御装置、制御方法、触覚提示システム及びプログラム製品を提供することを目的とする。 The present invention has been made in view of the above circumstances, and aims to provide a control device, a control method, a tactile presentation system, and a program product that improve the perceptibility of tactile sensations presented by a tactile presentation device. do.
 本開示の一実施形態の制御装置は、操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置の制御装置であって、前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データを前記対象物毎に記憶しておき、設定に基づき、前記制御データの前記変位量と制御データとの対応関係を補正し、前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力する。 A control device according to an embodiment of the present disclosure includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to the amount of displacement of the displacement section. A control device for a tactile presentation device that displays a tactile sensation of a displayed object, the control device storing control data for generating the tactile sensation for each of the objects in association with a displacement amount of the displacement portion. and corrects the correspondence between the displacement amount of the control data and the control data based on the settings, and outputs the corrected control data associated with the displacement amount of the displacement portion to the tactile presentation device.
 本開示の一実施形態の制御方法は、操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置の制御方法であって、前記触覚提示装置に接続されるコンピュータが、前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データを前記対象物毎に記憶しておき、設定に基づき、前記制御データの前記変位量と制御データとの対応関係を補正し、前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力する処理部を含む。 A control method according to an embodiment of the present disclosure includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to a displacement amount of the displacement section. A control method for a tactile presentation device that presents a tactile sensation of a displayed object by causing a computer connected to the tactile presentation device to generate the haptic sensation in association with a displacement amount of the displacement portion. Control data is stored for each object, and based on settings, the correspondence between the displacement amount of the control data and the control data is corrected, and the corrected control is made in correspondence with the displacement amount of the displacement part. It includes a processing unit that outputs data to the tactile presentation device.
 本開示の一実施形態のプログラム製品は、操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置に接続されるコンピュータに、前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データを前記対象物毎に記憶しておき、設定に基づき、前記制御データの前記変位量と制御データとの対応関係を補正し、前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力する処理を実行させる。 A program product according to an embodiment of the present disclosure includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to the amount of displacement of the displacement section. control data for causing the force sensation to be stored in a computer connected to a tactile presentation device that presents a tactile sensation of the displayed object, in association with a displacement amount of the displacement portion, for each of the objects; a process of correcting the correspondence between the displacement amount of the control data and the control data based on the settings, and outputting the corrected control data associated with the displacement amount of the displacement part to the tactile presentation device. Execute.
 本開示の制御装置、制御方法及びプログラム製品では、触覚提示装置が有する変位部が、その変位量に応じて力覚を生じさせる際に、変位部が変位を開始してから触覚が提示されるまでが個別に補正される。 In the control device, control method, and program product of the present disclosure, when the displacement portion of the tactile presentation device generates a force sensation according to the amount of displacement, the tactile sensation is presented after the displacement portion starts displacement. are corrected individually.
 本開示の一実施形態の制御装置は、前記変位量の昇順に連続的に変化する前記制御データを出力し始めるための前記変位量のズレを、前記設定として受け付ける。 A control device according to an embodiment of the present disclosure receives, as the setting, a shift in the displacement amount for starting to output the control data that continuously changes in ascending order of the displacement amount.
 本開示の制御装置では、制御データは、変位量の昇順の変化に対して、連続的にその対象物の触覚を再現できるように設定されるところ、変位部が変位してからその触覚が提示されるまでのズレが個別に補正される。 In the control device of the present disclosure, the control data is set so that the tactile sensation of the object can be continuously reproduced as the displacement amount changes in ascending order. The discrepancies up to that point are corrected individually.
 本開示の一実施形態の制御装置は、前記設定を、前記変位量の範囲に対するズレの位置を示すスライドバーを含む設定画面により受け付ける。 A control device according to an embodiment of the present disclosure receives the settings on a settings screen that includes a slide bar that indicates a position of deviation with respect to the range of the displacement amount.
 本開示の制御装置では、スライドバーで連続的な値としてズレの設定が可能である。 In the control device of the present disclosure, it is possible to set the deviation as a continuous value using a slide bar.
 本開示の一実施形態の制御装置は、前記設定を、前記変位量の範囲に対するズレの量が異なる複数の候補を選択可能に表示する設定画面により受け付ける。 A control device according to an embodiment of the present disclosure receives the settings on a settings screen that selectably displays a plurality of candidates having different amounts of deviation with respect to the range of displacement amounts.
 本開示の制御装置では、複数の候補から選択が可能である。 The control device of the present disclosure allows selection from multiple candidates.
 本開示の一実施形態の触覚提示システムは、操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置と、前記触覚提示装置と通信接続し、表示部を備える情報処理装置とを含み、前記情報処理装置は、前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データと、前記表示部に表示する視覚データとを前記対象物毎に記憶しておき、設定に基づき、前記制御データの前記変位量と制御データ及び視覚データとの対応関係をそれぞれ補正し、前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力し、補正後の視覚データに基づき、前記変位部の変位量に応じて前記表示部に画像を出力する。 A tactile presentation system according to an embodiment of the present disclosure includes a displacement section that is displaceable in response to an operation by an operator, and generates a force sensation in response to an operation on the displacement section according to the amount of displacement of the displacement section. a tactile presentation device that generates and presents a tactile sensation of a displayed object; and an information processing device that is communicatively connected to the tactile presentation device and includes a display section, and the information processing device is configured to detect a displacement amount of the displacement section. Control data that causes the force sensation and visual data to be displayed on the display unit are stored for each object in association with the control data and the control data and Correct the correspondence with the visual data, output the corrected control data associated with the displacement amount of the displacement part to the tactile presentation device, and calculate the displacement of the displacement part based on the corrected visual data. An image is output to the display section according to the amount.
 本開示の触覚提示システムでは、触覚提示装置が有する変位部が、その変位量に応じて力覚を生じさせる際に、変位部が変位を開始してから触覚が提示されるまでが操作者個別に補正される。 In the tactile presentation system of the present disclosure, when the displacement part of the tactile presentation device generates a sense of force according to the amount of displacement, the time from when the displacement part starts to be displaced until the tactile sense is presented is individual to the operator. It is corrected to
 本開示によれば、操作者に応じて触覚提示装置による触覚提示を調整することが可能となり、触覚の知覚度を向上させることができる。 According to the present disclosure, it is possible to adjust the tactile presentation by the tactile presentation device according to the operator, and it is possible to improve the degree of tactile perception.
触覚提示システムを示す模式図である。FIG. 1 is a schematic diagram showing a tactile presentation system. 情報処理装置の構成を示すブロック図である。FIG. 1 is a block diagram showing the configuration of an information processing device. 感覚DBの内容例を示す説明図である。It is an explanatory view showing an example of contents of a sensory DB. 触覚提示装置の構成を示すブロック図である。FIG. 2 is a block diagram showing the configuration of a tactile presentation device. 触覚提示システムにおける触覚提示の基本処理手順の一例を示すフローチャートである。2 is a flowchart illustrating an example of a basic processing procedure for tactile presentation in the tactile presentation system. 触覚提示システムにおける触覚提示の例を示す概要図である。FIG. 2 is a schematic diagram showing an example of tactile presentation in the tactile presentation system. 設定方法の一例を示すフローチャートである。3 is a flowchart illustrating an example of a setting method. 表示部に表示される設定画面の一例を示す。An example of a setting screen displayed on the display unit is shown. 提示開始位置による補正手順の一例を示すフローチャートである。7 is a flowchart illustrating an example of a correction procedure based on a presentation start position. 補正の結果を示す説明図である。It is an explanatory view showing the result of correction. 表示部に表示される設定画面の他の一例を示す。Another example of the setting screen displayed on the display unit is shown. 触覚提示システムにおける触覚提示の例を示す概要図である。FIG. 2 is a schematic diagram showing an example of tactile presentation in the tactile presentation system. 第2実施形態における触覚提示システムの概要図である。FIG. 2 is a schematic diagram of a tactile presentation system in a second embodiment. 第2実施形態のHMDの構成を示すブロック図である。It is a block diagram showing the composition of HMD of a 2nd embodiment.
 本開示をその実施の形態を示す図面を参照して具体的に説明する。以下の実施の形態では、触覚提示システムにおける設定方法の実施について説明する。 The present disclosure will be specifically described with reference to drawings showing embodiments thereof. In the following embodiments, implementation of a setting method in a tactile presentation system will be described.
 (第1実施形態)
 図1は、触覚提示システム100を示す模式図である。触覚提示システム100は、情報処理装置1と、触覚提示装置2とを含む。情報処理装置1と、触覚提示装置2とは近距離無線通信で通信接続され、相互にデータを授受する。
(First embodiment)
FIG. 1 is a schematic diagram showing a tactile presentation system 100. The tactile presentation system 100 includes an information processing device 1 and a tactile presentation device 2. The information processing device 1 and the tactile presentation device 2 are communicably connected by short-range wireless communication and exchange data with each other.
 情報処理装置1は、図1に示すようにスマートフォンを用いる。情報処理装置1は、スマートフォンでなく、タブレット端末であってもよいし、ラップトップ型のPC(Personal Computer)であってもよい。 The information processing device 1 uses a smartphone as shown in FIG. The information processing device 1 may be a tablet terminal or a laptop-type PC (Personal Computer) instead of a smartphone.
 触覚提示装置2は、操作者が指を変位部202に沿えて持ちながら指を動かして操作できる装置である。触覚提示装置2は、操作者が指を動かすことで変位する変位部202の位置を読み取り、内蔵するMRF(Magneto-Rheological Fluid)デバイス24をその位置に応じて制御して操作者の変位部202への操作に対する反力(回転抵抗)により力覚を生じさせ、触覚を提示する装置である。触覚提示装置2の変位部202の態様は、図1に示すようなものに限られず、スティック状であってもよいし、カバーに覆われたクッション状のものであってもよい。触覚提示装置2は、MRFデバイス24に代えて、モータやピエゾ素子などを採用し、操作者の操作に対し回転力や振動により力覚を生じさせるものでもよく、変位部202の他、振動や温感、冷感を提示するものと組み合わされてもよい。地面や壁に設置し、操作者の手のひらや足等で操作される構造でもよい。 The tactile presentation device 2 is a device that an operator can operate by moving his or her finger while holding it along the displacement portion 202. The tactile presentation device 2 reads the position of the displacement part 202 that is displaced by the operator's finger movement, controls the built-in MRF (Magneto-Rheological Fluid) device 24 according to the position, and moves the displacement part 202 of the operator. This is a device that produces a sense of force using a reaction force (rotational resistance) to the operation of the machine, and presents a tactile sensation. The form of the displacement part 202 of the tactile presentation device 2 is not limited to that shown in FIG. 1, but may be stick-like or cushion-like covered with a cover. The tactile presentation device 2 may employ a motor, a piezo element, or the like instead of the MRF device 24, and may generate a force sensation by rotational force or vibration in response to an operation by the operator. It may be combined with something that provides a warm or cold sensation. The structure may be such that it is installed on the ground or a wall and is operated by the operator's palm or foot.
 触覚提示システム100では、触覚提示装置2は、情報処理装置1と連携し、情報処理装置1にて対象物の画像及び音声を出力させながら、対象物の触覚を触覚提示装置2で出力させる。図1に示す例では、情報処理装置1の表示部13に、柔らかな物体を表示させ、操作者が触覚提示装置2の変位部202を指で押し込むと、触覚提示装置2がグニャリ、若しくはフワフワとした触覚を出力させつつ、表示部13に、物体が、押し込まれたように変化する画像を表示させ、音声出力部14にて物体が押し込まれたことに対応する「グニャリ」と聞こえるような音声を出力する。 In the tactile presentation system 100, the tactile presentation device 2 cooperates with the information processing device 1, and causes the information processing device 1 to output the image and audio of the object, while causing the tactile presentation device 2 to output the tactile sensation of the object. In the example shown in FIG. 1, when a soft object is displayed on the display unit 13 of the information processing device 1 and the operator pushes the displacement part 202 of the tactile presentation device 2 with a finger, the tactile presentation device 2 becomes squishy or fluffy. While outputting a tactile sensation, the display unit 13 displays an image that changes as if the object is being pushed in, and the audio output unit 14 outputs an audible “squeak” sound that corresponds to the object being pushed in. Output audio.
 図2は、情報処理装置1の構成を示すブロック図である。情報処理装置1は、処理部10、記憶部11、通信部12、表示部13、音声出力部14及び操作部15を備える。処理部10は、CPU(Central Processing Unit)及び/又はGPU(Graphics Processing Unit)を用いたプロセッサである。処理部10は、記憶部11に記憶されている触覚提示用の制御プログラムP1に基づき、後述する処理を実行する。 FIG. 2 is a block diagram showing the configuration of the information processing device 1. The information processing device 1 includes a processing section 10, a storage section 11, a communication section 12, a display section 13, an audio output section 14, and an operation section 15. The processing unit 10 is a processor using a CPU (Central Processing Unit) and/or a GPU (Graphics Processing Unit). The processing unit 10 executes a process described below based on a control program P1 for tactile presentation stored in the storage unit 11.
 記憶部11は、フラッシュメモリ、SSD(Solid State Drive)等の不揮発性メモリを用いる。記憶部11は、処理部10が参照するデータを記憶する。制御プログラムP1(プログラム製品)は、通信部12を介して情報処理装置1又は他のプログラムサーバ装置からダウンロードして実行可能に記憶したものである。記憶部11に記憶されている制御プログラムP1は、コンピュータから読み取り可能な記憶媒体8に記憶されていた制御プログラムP8を処理部10が読み出して記憶したものであってもよい。 The storage unit 11 uses nonvolatile memory such as flash memory and SSD (Solid State Drive). The storage unit 11 stores data referenced by the processing unit 10. The control program P1 (program product) is downloaded from the information processing device 1 or another program server device via the communication unit 12 and stored in an executable manner. The control program P1 stored in the storage unit 11 may be the control program P8 stored in the computer-readable storage medium 8, which is read out and stored by the processing unit 10.
 記憶部11は、出力する対象物の触覚データ、視覚データ及び聴覚データを含む感覚データベース(DB:Data Base)110を記憶する。感覚DB110は、対象物を識別する対象物IDに対応付けて、触覚提示装置2における変位部の変位データ毎に、その変位において出力すべき触覚データ、視覚データ(画像、動画)、及び聴覚データ(音声)を格納している(図3参照)。記憶部11は更に、操作者個別の設定データを記憶する。 The storage unit 11 stores a sensory database (DB) 110 that includes tactile data, visual data, and auditory data of the object to be output. The sensory DB 110 stores tactile data, visual data (images, videos), and auditory data to be output at each displacement for each displacement data of a displacement part in the tactile presentation device 2, in association with an object ID that identifies the object. (audio) (see Figure 3). The storage unit 11 further stores setting data for each operator.
 通信部12は、近距離無線通信、例えばBluetooth(登録商標)の通信モジュールである。処理部10は、通信部12によって触覚提示装置2との間でデータを送受信できる。 The communication unit 12 is a communication module for short-range wireless communication, for example, Bluetooth (registered trademark). The processing unit 10 can transmit and receive data to and from the tactile presentation device 2 through the communication unit 12 .
 表示部13は、液晶ディスプレイ、有機EL(Electro Luminescence)ディスプレイ等のディスプレイである。表示部13は例えば、タッチパネル内蔵型ディスプレイである。処理部10は、表示部13に、制御プログラムP1に基づき、触覚提示装置2で触覚を出力させるための操作画面や対象物の画像を表示する。 The display unit 13 is a display such as a liquid crystal display or an organic EL (Electro Luminescence) display. The display unit 13 is, for example, a display with a built-in touch panel. The processing unit 10 displays, on the display unit 13, an operation screen for causing the tactile sensation presentation device 2 to output a tactile sensation and an image of the target object, based on the control program P1.
 音声出力部14は、スピーカ等を含む。処理部10は、制御プログラムP1に基づき、対象物の音声、音楽等を音声出力部14から出力させる。 The audio output unit 14 includes a speaker and the like. The processing unit 10 causes the audio output unit 14 to output the sound of the object, music, etc. based on the control program P1.
 操作部15は、処理部10との間で入出力が可能なユーザインタフェースであって、表示部13内蔵のタッチパネルである。操作部15は、物理ボタンであってもよい。操作部15は、音声入力部であってもよい。 The operation unit 15 is a user interface that allows input and output to and from the processing unit 10, and is a touch panel built into the display unit 13. The operation unit 15 may be a physical button. The operation unit 15 may be a voice input unit.
 図3は、感覚DB110の内容例を示す説明図である。感覚DB110は、図3に示すように、対象物IDに対応付けて、変位部202の変位量(角度)毎の触覚提示装置2のMRFデバイスへの電流の電流値を触覚データとして格納している。対象物は複数であり、触覚を提示する物として実在する物体、例えば、ボール又は風船等の静物、グミ、野菜、又は果物等の食べ物、犬、猫又は魚等の動物を含んでもよいし、スライムやモンスター等のキャラクタを含んでもよい。 FIG. 3 is an explanatory diagram showing an example of the contents of the sensory DB 110. As shown in FIG. 3, the sensory DB 110 stores the current value of the current to the MRF device of the tactile presentation device 2 for each displacement amount (angle) of the displacement part 202 as tactile data in association with the object ID. There is. The objects are plural, and may include real objects that present a tactile sensation, such as still life objects such as balls or balloons, foods such as gummies, vegetables, or fruits, and animals such as dogs, cats, or fish. Characters such as slimes and monsters may also be included.
 感覚DB110の変位部202の変位量と、電流値との関係は、変位部202の可動範囲の実際の寸法と、対象物の寸法とに合わせて設定されている。例えば、変位部202の可動範囲が50mmであり、対象物が小さなグミのような破裂しない球体であって大きさが15mmに設定されている場合、触覚は最初の15mm程度の範囲で再現され、残りの35mmの範囲は、それ以上押し込めないほどの感触となるような回転抵抗を生じさせる電流値が設定されている。又は、12mm~15mmの範囲は押し込めなくなる大きな電流値が設定され、それ以降の電流値はゼロに設定されていてもよい。また、対象物がソファの座面のような大きな物体であって大きさが変位部202の可動範囲を大きく超える場合、触覚データでは、ソファの座面表面から50mmまでの感触を提示するような電流値が設定される。対象物がモンスター等の仮想的なキャラクタの場合は、そのキャラクタの設定に応じた大きさに基づいて電流値が設定される。 The relationship between the amount of displacement of the displacement section 202 of the sensory DB 110 and the current value is set according to the actual dimensions of the movable range of the displacement section 202 and the dimensions of the object. For example, if the movable range of the displacement unit 202 is 50 mm, and the object is a sphere that will not burst, such as a small gummy bear, and the size is set to 15 mm, the tactile sensation will be reproduced within the first 15 mm, In the remaining 35 mm range, a current value is set that produces a rotational resistance that makes it difficult to push in any further. Alternatively, a large current value that makes it impossible to push in the range of 12 mm to 15 mm may be set, and the current values thereafter may be set to zero. In addition, if the target object is a large object such as the seat surface of a sofa, and the size greatly exceeds the movable range of the displacement unit 202, the tactile data may be such as to present the sensation up to 50 mm from the surface of the sofa seat surface. The current value is set. If the target object is a virtual character such as a monster, the current value is set based on the size according to the setting of the character.
 感覚DB110は、同様に、対象物IDに対応付けて、変位部202の位置(角度)毎の画像(フレーム画像)を視覚データとして格納している。ここでフレーム画像は、連続して表示することでアニメーション画像として認識される、1つの静止画像である。また同様にして、対象物IDに対応付けて、変位部202の位置(角度)毎の音声を聴覚データとして格納している。聴覚データは、各角度によって異なる波形データであってもよい。各角度に対応する音声のタイムスタンプであってもよい。 Similarly, the sensory DB 110 stores images (frame images) for each position (angle) of the displacement part 202 as visual data in association with the object ID. Here, the frame image is one still image that is recognized as an animation image when displayed continuously. Similarly, audio data for each position (angle) of the displacement section 202 is stored as auditory data in association with the object ID. The auditory data may be waveform data that differs depending on each angle. It may also be a timestamp of the audio corresponding to each angle.
 図4は、触覚提示装置2の構成を示すブロック図である。触覚提示装置2は、図1に示したように、扁平な有底円筒状の把持体200に周方向に一部沿うような湾曲部を有する帯状平板の変位部202を設けて構成される。変位部202は、それ自体が撓むことが可能な素材であるが、剛性の高い素材を採用し、把持体200と支軸を介して回動可能に支持されていてもよい。変位部202の先端の外側の面には、布テープ状の結束具203が設けられている。変位部202の先端の内側面は、把持体200内部に収容されているMRFデバイス24のロータの回転軸と連結するリンク機構204が設けられている。 FIG. 4 is a block diagram showing the configuration of the tactile presentation device 2. As shown in FIG. 1, the tactile presentation device 2 is configured by providing a flat bottomed cylindrical grip 200 with a band-shaped flat plate displacement portion 202 having a curved portion partially along the circumferential direction. Although the displacement part 202 is made of a flexible material, it may also be made of a highly rigid material and rotatably supported via the grip body 200 and a support shaft. A binding tool 203 in the form of a cloth tape is provided on the outer surface of the distal end of the displacement portion 202 . A link mechanism 204 is provided on the inner surface of the distal end of the displacement portion 202 to connect to the rotating shaft of the rotor of the MRF device 24 housed inside the gripper 200 .
 操作者は図1に示したように、把持体200を例えば親指と中指とで把持しつつ、人差し指等の指を変位部202に沿わせて結束具203に人差し指を差し込んで使用する。操作者は人差し指を押し込むように変位部202を動かすことができ、また、人差し指を伸ばして把持体200から変位部202を遠ざけるように動かすことができる。 As shown in FIG. 1, the operator uses the gripping body 200 by, for example, holding it with the thumb and middle finger and inserting the index finger into the binding tool 203 with the finger such as the index finger along the displacement part 202. The operator can move the displacement part 202 by pushing in the index finger, or can move the displacement part 202 away from the grip body 200 by extending the index finger.
 触覚提示装置2は、図1に示すような把持体200と、制御部20、記憶部21、通信部22、電源部23、MRFデバイス24、センサ25を備える。把持体200は、MRFデバイス24を内蔵する。制御部20、記憶部21、通信部22、及び電源部23は把持体200と一体に設けられてもよいし、把持体200と無線又は有線により接続される別体に設けられてもよい。 The tactile presentation device 2 includes a gripping body 200 as shown in FIG. The grip body 200 has an MRF device 24 built therein. The control section 20, the storage section 21, the communication section 22, and the power supply section 23 may be provided integrally with the grip body 200, or may be provided as separate bodies connected to the grip body 200 wirelessly or by wire.
 制御部20は、CPU、MPU(Micro-Processing Unit)等のプロセッサ、ROM(Read Only Memory)、RAM(Random Access Memory)等のメモリを含む。制御部20は、例えばマイクロコントローラである。制御部20は、内蔵ROMに記憶されている制御プログラムP2に基づいて各構成部を制御し、触覚提示を実現する。 The control unit 20 includes a processor such as a CPU and an MPU (Micro-Processing Unit), and a memory such as a ROM (Read Only Memory) and a RAM (Random Access Memory). The control unit 20 is, for example, a microcontroller. The control unit 20 controls each component based on a control program P2 stored in the built-in ROM to realize tactile presentation.
 記憶部21は、制御部20に対する補助記憶メモリであり、MRFデバイス24の制御データ(触覚データ)を書き換え可能に記憶する。 The storage unit 21 is an auxiliary storage memory for the control unit 20, and stores control data (tactile data) of the MRF device 24 in a rewritable manner.
 通信部22は、近距離無線通信、例えばBluetooth(登録商標)の通信モジュールである。制御部20は、通信部22によって情報処理装置1との間でデータを送受信できる。 The communication unit 22 is a communication module for short-range wireless communication, for example, Bluetooth (registered trademark). The control unit 20 can transmit and receive data to and from the information processing device 1 through the communication unit 22 .
 制御部20は、電源部23、MRFデバイス24及びセンサ25とI/Oを介して接続されており、相互に信号を授受する。 The control unit 20 is connected to the power supply unit 23, the MRF device 24, and the sensor 25 via I/O, and exchanges signals with each other.
 電源部23は、充電可能なバッテリを含む。電源部23は、ON状態になると各構成部及びMRFデバイス24へ電力を供給する。 The power supply unit 23 includes a rechargeable battery. The power supply unit 23 supplies power to each component and the MRF device 24 when turned on.
 MRFデバイス24は、円板状のロータを、隙間を開けて挟むようにして設けられたヨークを有し、ヨークに設けられたコイルに制御電流を流して磁界を発生させ、隙間に封入されている磁気粘性流体の粘度(ずり応力)を制御してロータの回転抵抗を与える。制御部20が、MRFデバイス24への制御電流の大きさを制御すると即座に回転抵抗が変更される。 The MRF device 24 has a yoke that sandwiches a disk-shaped rotor with a gap between them, and generates a magnetic field by passing a control current through a coil provided in the yoke, thereby generating a magnetic field sealed in the gap. The viscosity (shear stress) of the viscous fluid is controlled to provide rotational resistance to the rotor. When the control unit 20 controls the magnitude of the control current to the MRF device 24, the rotational resistance is immediately changed.
 センサ25は、変位部202の位置(角度)を測定して制御部20へ出力する。センサ25は変位部202の変位を、角度として測定して出力する。センサ25は、ジャイロセンサ、加速度センサ等の複数のセンサから構成されてもよい。 The sensor 25 measures the position (angle) of the displacement section 202 and outputs it to the control section 20. The sensor 25 measures and outputs the displacement of the displacement portion 202 as an angle. The sensor 25 may include a plurality of sensors such as a gyro sensor and an acceleration sensor.
 上述のように構成される触覚提示装置2では、変位部202が操作者によって操作されると、変位部202の変位がリンク機構204を介してMRFデバイス24のロータの回転軸への回転方向に伝達される。回転軸は、MRFデバイス24が動作していない場合、即ち制御電流がゼロである間は、自由に回転するため、変位部202は抵抗なく変動する。一方で、MRFデバイス24が動作し、制御電流がゼロでない場合には、MRFデバイス24へ流れる電流の大きさに応じてMRFデバイス24内部の磁気粘性流体の粘度(ずり応力)が変更される。制御部20が、MRFデバイス24への電流の大きさを連続的に変更したり、電流値を所定の周波数で振動させたりすることで、変位部202に対する抵抗の力やその出現方法を変更できる。 In the tactile presentation device 2 configured as described above, when the displacement section 202 is operated by the operator, the displacement of the displacement section 202 is caused to move in the direction of rotation of the MRF device 24 toward the rotation axis of the rotor via the link mechanism 204. communicated. Since the rotating shaft rotates freely when the MRF device 24 is not operating, that is, while the control current is zero, the displacement portion 202 fluctuates without resistance. On the other hand, when the MRF device 24 operates and the control current is not zero, the viscosity (shear stress) of the magnetorheological fluid inside the MRF device 24 is changed depending on the magnitude of the current flowing to the MRF device 24. By continuously changing the magnitude of the current to the MRF device 24 or by vibrating the current value at a predetermined frequency, the control unit 20 can change the resistance force applied to the displacement unit 202 and the method of its appearance. .
 このようにして触覚提示装置2は、変位部202の押し込む量に応じて抵抗(電流値)を変動させてヌルリとした触覚を提示したり、押し込む量が大きくなるにつれて抵抗を大きくしてギュッとした固さの触覚を提示したり、抵抗の大小を繰り返してザクザクとした触覚を提示したりすることができる。 In this way, the tactile sensation presentation device 2 varies the resistance (current value) according to the amount of pushing of the displacement section 202 to present a smooth tactile sensation, or increases the resistance as the pushing amount increases to create a tighter feel. It is possible to present a tactile sensation with a certain hardness, or to present a crunchy tactile sensation by repeatedly increasing and decreasing the resistance.
 図5は、触覚提示システム100における触覚提示の基本処理手順の一例を示すフローチャートである。情報処理装置1の処理部10は、操作者が制御プログラムP1を起動させ、触覚提示装置2の電源をONとすると、触覚提示装置2と連携して以下の処理を開始する。 FIG. 5 is a flowchart illustrating an example of a basic processing procedure for tactile presentation in the tactile presentation system 100. When the operator starts the control program P1 and turns on the power of the tactile presentation device 2, the processing unit 10 of the information processing device 1 starts the following process in cooperation with the tactile presentation device 2.
 処理部10は、触覚を表示する対象物の候補のリストを含む操作画面を表示部に表示し(ステップS101)、対象物の選択を受け付ける(ステップS102)。処理部10は、選択された対象物の対象物IDに対応する触覚データ、画像データ及び音声データを記憶部11の感覚DB110から読み出す(ステップS103)。 The processing unit 10 displays on the display unit an operation screen that includes a list of object candidates for which tactile sensation is to be displayed (step S101), and accepts the selection of the object (step S102). The processing unit 10 reads tactile data, image data, and audio data corresponding to the object ID of the selected object from the sensory DB 110 of the storage unit 11 (step S103).
 処理部10は、読み出した触覚データを、後述する処理で設定される提示開始位置に基づき、読み出した触覚データ、画像データ及び音声データを補正して、一時記憶する(ステップS104)。ステップS104における補正の方法については詳細を後述する。 The processing unit 10 corrects the read tactile data, image data, and audio data based on the presentation start position set in a process described later, and temporarily stores the corrected data (step S104). The correction method in step S104 will be described in detail later.
 処理部10は、触覚提示装置2と通信部12による通信を確立させ(ステップS105)、一時記憶した補正後の触覚データを、触覚提示装置2へ送信する(ステップS106)。処理部10は、表示部13に触覚提示を開始することを表示させる(ステップS107)。 The processing unit 10 establishes communication between the tactile presentation device 2 and the communication unit 12 (step S105), and transmits the temporarily stored corrected tactile data to the tactile presentation device 2 (step S106). The processing unit 10 causes the display unit 13 to display that tactile presentation is to be started (step S107).
 触覚提示装置2の制御部20は、触覚データを受信すると(ステップS201)、記憶部21に記憶する(ステップS202)。 Upon receiving the tactile data (step S201), the control unit 20 of the tactile presentation device 2 stores it in the storage unit 21 (step S202).
 制御部20は、センサ25から出力される変位部202の変位量(角度)に対応する信号をサンプリングする(ステップS203)。制御部20は、サンプリングにより得られる変位量を、情報処理装置1へ送信し(ステップS204)、得られた変位量に対応する電流値を、記憶部21に記憶した触覚データから参照し(ステップS205)、参照した電流をMRFデバイス24へ出力し(ステップS206)、処理をステップS203へ送信する。ステップS203-S206の処理は、情報処理装置1側で終了させる操作がされるまで継続する。 The control unit 20 samples a signal corresponding to the displacement amount (angle) of the displacement unit 202 output from the sensor 25 (step S203). The control unit 20 transmits the displacement amount obtained by sampling to the information processing device 1 (step S204), and refers to the current value corresponding to the obtained displacement amount from the tactile data stored in the storage unit 21 (step S204). S205), outputs the referenced current to the MRF device 24 (step S206), and transmits the process to step S203. The processing in steps S203 to S206 continues until an operation to end the processing is performed on the information processing device 1 side.
 情報処理装置1は、触覚提示装置2から変位量を受信し(ステップS108)、処理部10は、受信した変位量に対応する画像及び音声を記憶部11の視覚データ及び聴覚データから参照し(ステップS109)、それぞれを表示部13及び音声出力部14から出力させ(ステップS110)、処理をステップS108へ戻す。触覚提示装置2から変位量が送信される都度、情報処理装置1は、それに対応する画像及び音声を出力する。 The information processing device 1 receives the displacement amount from the tactile presentation device 2 (step S108), and the processing section 10 refers to the image and audio corresponding to the received displacement amount from the visual data and auditory data in the storage section 11 ( Step S109) and output from the display section 13 and audio output section 14 (Step S110), and the process returns to Step S108. Each time the amount of displacement is transmitted from the tactile presentation device 2, the information processing device 1 outputs an image and audio corresponding to the amount of displacement.
 処理部10は、終了操作がされたか否かを判断する(ステップS111)。終了操作がされていないと判断された場合(S111:NO)、処理をステップS108へ戻し、ステップS108-S110の処理を繰り返す。 The processing unit 10 determines whether a termination operation has been performed (step S111). If it is determined that the end operation has not been performed (S111: NO), the process returns to step S108, and the processes of steps S108-S110 are repeated.
 終了操作がされたと判断された場合(S111:YES)、と処理部10は、表示を終了して触覚提示装置2との通信を切断し(ステップS112)、処理を終了する。 If it is determined that the end operation has been performed (S111: YES), the processing unit 10 ends the display, disconnects communication with the tactile presentation device 2 (step S112), and ends the process.
 図6は、触覚提示システム100における触覚提示の例を示す概要図である。図6には、触覚提示装置2に送信される触覚データにおける変位量に対応して表示部13に表示される画像の変化及び音声の変化を示す。図6に示すように、変位部202の押し込み量に応じて、画像が変化し、触覚に対応する凹み方が変わる。また、図6に示すように、画像中に、操作者の指の画像を表示する場合、接触、即ちMRFデバイス24へ出力する初期位置(変位量がゼロ)の場所から、画像における対象物に接触する指の画像が表示されるようにしてもよい。 FIG. 6 is a schematic diagram showing an example of tactile presentation in the tactile presentation system 100. FIG. 6 shows changes in images and sounds displayed on the display unit 13 in response to the amount of displacement in the tactile data transmitted to the tactile presentation device 2. As shown in FIG. 6, the image changes depending on the amount of depression of the displacement part 202, and the way the depression corresponds to the tactile sensation changes. In addition, as shown in FIG. 6, when displaying an image of the operator's finger in the image, contact, that is, from the initial position (displacement amount is zero) output to the MRF device 24, to the target object in the image. An image of the touching finger may be displayed.
 このように触覚を提示する触覚提示システム100では、上述の触覚提示の処理のうち、図5のフローチャートに示した処理手順のステップS104で、情報処理装置1が触覚データを補正するに際し、触覚提示装置2での出力を操作者個々に合わせることができる。 In the tactile presentation system 100 that presents the tactile sense in this way, when the information processing device 1 corrects the tactile data in step S104 of the processing procedure shown in the flowchart of FIG. The output of the device 2 can be tailored to each operator.
 図7は、設定方法の一例を示すフローチャートである。以下に示す処理手順は、制御プログラムP1に基づき表示される操作画面内で、設定メニューを選択した場合に開始される。図7に示す処理手順は、制御プログラムP1に基づく処理を初回に実行する際に自動的に開始されてもよい。初期的には、提示開始位置に対応する変位量は、ゼロである。 FIG. 7 is a flowchart illustrating an example of a setting method. The processing procedure shown below is started when the setting menu is selected within the operation screen displayed based on the control program P1. The processing procedure shown in FIG. 7 may be automatically started when the processing based on the control program P1 is executed for the first time. Initially, the amount of displacement corresponding to the presentation start position is zero.
 処理部10は、変位量の角度範囲に対応するバーを含む設定画面を表示する(ステップS301)。処理部10は、バーによる提示開始位置の選択を受け付け(ステップS302)、選択された提示開始位置に対応する変位量(角度)を決定し、記憶する(ステップS303)。ステップS303において処理部10は、バーの長さに対する選択された位置の長さの割合から、変位量の範囲(0~90°)における変位量の位置を決定して記憶する。つまり、バーの長さを100とした場合に、選択された位置までの長さが5であるケースでは約5%であるから、5°が提示開始位置に対応する変位量として決定される。 The processing unit 10 displays a setting screen including a bar corresponding to the angular range of the displacement amount (step S301). The processing unit 10 receives the selection of the presentation start position using the bar (step S302), determines and stores the amount of displacement (angle) corresponding to the selected presentation start position (step S303). In step S303, the processing unit 10 determines and stores the position of the displacement amount in the displacement amount range (0 to 90°) from the ratio of the length of the selected position to the length of the bar. That is, when the length of the bar is 100, in the case where the length to the selected position is 5, it is about 5%, so 5° is determined as the displacement amount corresponding to the presentation start position.
 処理部10は、提示開始位置に対応する変位量(角度)までの電流値をゼロとし、提示開始位置に対応する変位量以上の変位量に、所定の電流値をセットした触覚データを作成する(ステップS304)。処理部10は、作成した触覚データを触覚提示装置2へ送信し(ステップS305)、設定画面に試すように促すメッセージを表示部13に表示させる(ステップS306)。 The processing unit 10 creates tactile data in which the current value up to the amount of displacement (angle) corresponding to the presentation start position is set to zero, and a predetermined current value is set to the amount of displacement greater than or equal to the displacement amount corresponding to the presentation start position. (Step S304). The processing unit 10 transmits the created tactile data to the tactile presentation device 2 (step S305), and causes the display unit 13 to display a message urging you to try it on the setting screen (step S306).
 触覚提示装置2の制御部20は、受信した触覚データに基づいて、変位部の変位量に応じた電流をMRFデバイス24に流し、設定された提示開始位置に対応する変位量よりも変位部202が押し込まれた場合に初めて、所定の反力(回転抵抗)を発するテストを実行可能とする。これにより、反力を感じ始めるまでの猶予(遊び・ズレ)を、操作者が試すことができる。 Based on the received tactile data, the control unit 20 of the tactile presentation device 2 causes a current to flow through the MRF device 24 according to the amount of displacement of the displacement portion, so that the amount of displacement of the displacement portion 202 is smaller than the amount of displacement corresponding to the set presentation start position. It is possible to perform a test in which a predetermined reaction force (rotational resistance) is generated only when the holder is pushed in. This allows the operator to test the amount of time (play/displacement) before starting to feel a reaction force.
 処理部10は、設定画面に、決定ボタンを表示し(ステップS307)、試した結果、ステップS302で選択した提示開始位置でよいか否かを受け付ける。処理部10は、決定ボタンが選択されたか否かを判断する(ステップS308)。決定ボタンが選択されていないと判断された場合(S308:NO)、処理部10は、処理をステップS302へ戻し、決定ボタンが選択されるまで処理を繰り返す。 The processing unit 10 displays a determination button on the setting screen (step S307), and accepts whether or not the presentation start position selected in step S302 is acceptable as a result of the trial. The processing unit 10 determines whether the OK button has been selected (step S308). If it is determined that the enter button has not been selected (S308: NO), the processing unit 10 returns the process to step S302 and repeats the process until the enter button is selected.
 決定ボタンが選択されたと判断された場合(S308:YES)、処理部10は、ステップS303で決定した変位量(角度)を、補正量として記憶部11に記憶し(ステップS309)、処理を終了する。 If it is determined that the enter button has been selected (S308: YES), the processing unit 10 stores the displacement amount (angle) determined in step S303 in the storage unit 11 as a correction amount (step S309), and ends the process. do.
 図8は、表示部13に表示される設定画面130の一例を示す。操作部15によって設定メニューが選択された場合に、処理部10は設定画面130を表示部13に表示する。設定画面130には、変位量の範囲に対応する長さに対してスライド可能なコントロール132を含むバーインタフェース131が含まれる。操作者は、表示部13内蔵のタッチパネル上で、コントロール132を上下に動かすことができる。 FIG. 8 shows an example of the setting screen 130 displayed on the display unit 13. When the settings menu is selected by the operation unit 15, the processing unit 10 displays the settings screen 130 on the display unit 13. The settings screen 130 includes a bar interface 131 that includes a control 132 that is slidable over a length that corresponds to a range of displacement amounts. The operator can move the control 132 up and down on the touch panel built into the display section 13.
 設定画面130は、テストボタン133及び決定ボタン134を含む。テストボタン133が操作者によって選択されると、処理部10は操作部15によりこれを検知し、動かされたコントロール132の位置によってステップS303において変位量を決定する。テストボタン133が選択された場合、処理部10は、ステップS304からステップS308の処理を実行する。 The setting screen 130 includes a test button 133 and a decision button 134. When the test button 133 is selected by the operator, the processing unit 10 detects this using the operation unit 15, and determines the displacement amount in step S303 based on the position of the moved control 132. If the test button 133 is selected, the processing unit 10 executes the processes from step S304 to step S308.
 決定ボタン134が選択された場合、処理部10は、ステップS308-S309を実行する。 If the decision button 134 is selected, the processing unit 10 executes steps S308-S309.
 これにより、変位部202を操作した場合の提示開始位置の補正量が操作者毎に設定され、図5のフローチャートに示した処理におけるステップS104によって反映される。図9は、提示開始位置による補正手順の一例を示すフローチャートであり、図10は、補正の結果を示す説明図である。 As a result, the amount of correction of the presentation start position when operating the displacement unit 202 is set for each operator, and is reflected in step S104 in the process shown in the flowchart of FIG. FIG. 9 is a flowchart showing an example of a correction procedure based on the presentation start position, and FIG. 10 is an explanatory diagram showing the results of the correction.
 処理部10は、ステップS104において、読み出した触覚データの各変位量に、補正量として記憶した変位量を加算した、補正後の触覚データを作成する(ステップS401)。これにより、触覚データにおける連続的に変化する電流値を出力し始める変位量をずらすことができる。 The processing unit 10 creates corrected tactile data by adding the displacement amount stored as the correction amount to each displacement amount of the tactile data read in step S104 (step S401). Thereby, it is possible to shift the amount of displacement at which the continuously changing current value in the tactile data starts to be output.
 処理部10は、読み出した聴覚データの各変位量に、補正量として記憶した変位量を加算した、補正後の聴覚データを作成する(ステップS402)。 The processing unit 10 creates corrected auditory data by adding the displacement amount stored as the correction amount to each displacement amount of the read auditory data (step S402).
 処理部10は、読み出した視覚データの各変位量に、補正量として記憶した変位量を加算した、補正後の視覚データを作成し(ステップS403)、補正を終了する。 The processing unit 10 creates corrected visual data by adding the displacement amount stored as the correction amount to each displacement amount of the read visual data (step S403), and ends the correction.
 ステップS403において処理部10は、視覚データである画像に、操作者の指の画像を仮想的に重畳して表示する場合、補正量として記憶した変位量と同一の変位量が触覚提示装置2から送信されたタイミングで、指の画像が対象物に接触して見えるように指の画像を表示させる座標を算出してもよい。この場合、処理部10は、提示開始位置と、触覚データにおける電流値がゼロでなくなる変位量(角度)とに基づき、触覚が生じるタイミングと、画像において指の画像が対象物に接触して見えるタイミングとが合致するように、画像を調整してもよい。 In step S403, when displaying the image of the operator's finger virtually superimposed on the image that is the visual data, the processing unit 10 determines that when the image of the operator's finger is virtually superimposed and displayed, the same amount of displacement as the amount of displacement stored as the correction amount is obtained from the tactile presentation device 2. At the transmitted timing, coordinates for displaying the finger image so that the finger image appears to be in contact with the target object may be calculated. In this case, the processing unit 10 determines the timing at which the tactile sensation occurs and the image of the finger appearing in contact with the object based on the presentation start position and the amount of displacement (angle) at which the current value in the tactile data becomes non-zero. The image may be adjusted so that the timing matches.
 補正によって触覚データ、聴覚データ及び視覚データは、図10に示すように、補正量が加算された後の変位量それぞれに対し、電流値、音声データ、及び画像データが対応付けられたデータとなる。つまり、変位部202の変位量の範囲(可動範囲)と、提示される感触の範囲とがずれるように補正される。 Through the correction, the tactile data, auditory data, and visual data become data in which current values, audio data, and image data are associated with each displacement amount after the correction amount is added, as shown in FIG. . That is, the range of the amount of displacement (movable range) of the displacement portion 202 is corrected so as to deviate from the range of the presented sensation.
 図11は、表示部13に表示される設定画面130の他の一例を示す。設定画面130は、図8に示したように、操作部15によって設定メニューが選択された場合に、処理部10は設定画面130を表示部13に表示する。設定画面130には、変位量の範囲に対して、上からどこまで押し込むと触覚提示が始まるか(触覚提示位置)の異なるパターンが3つ候補として表示されている。 FIG. 11 shows another example of the setting screen 130 displayed on the display unit 13. As shown in FIG. 8, when the setting menu is selected by the operation unit 15, the processing unit 10 displays the setting screen 130 on the display unit 13. On the setting screen 130, three patterns are displayed as candidates, with respect to the range of displacement amount, with different patterns of how far from the top the tactile presentation should start (tactile presentation position).
 図11の例でも、設定画面130は、テストボタン133及び決定ボタン134を含む。決定ボタン134により、いずれかを選択することが可能である。図11に示す例では、3つの候補のうち、中央の候補が、太線の枠(カーソル)により強調されており、この候補が操作部15によるタップで選択された状態であることが示されている。 In the example of FIG. 11 as well, the setting screen 130 includes a test button 133 and a decision button 134. It is possible to select one by pressing the decision button 134. In the example shown in FIG. 11, of the three candidates, the center candidate is highlighted by a thick line frame (cursor), indicating that this candidate has been selected by tapping on the operation unit 15. There is.
 テストボタン133が操作者によって選択されると、処理部10は操作部15によりこれを検知し、動かされたコントロール132の位置によってステップS303において変位量を決定する。テストボタン133が選択された場合、処理部10は、ステップS304からステップS308の処理を実行する。 When the test button 133 is selected by the operator, the processing unit 10 detects this using the operation unit 15, and determines the displacement amount in step S303 based on the position of the moved control 132. If the test button 133 is selected, the processing unit 10 executes the processes from step S304 to step S308.
 図11に示す設定画面130の例であっても、決定ボタン134が選択された場合、処理部10は、ステップS308-S309を実行する。 Even in the example of the setting screen 130 shown in FIG. 11, if the enter button 134 is selected, the processing unit 10 executes steps S308-S309.
 図12は、触覚提示システム100における触覚提示の例を示す概要図である。図12に示す例では、図6に示した例と比較して、補正後の触覚データにおける変位量に対応して表示部13に表示される画像の変化及び音声の変化とを示す。図12では、補正前の触覚データにおける状態を破線により示している。図12に示すように、補正後は、図6と比較して、変位部202を少しだけ(5°相当)押し込んだ位置から、変位部202から触覚が提示され始め、同様に、画像及び音声が出力される。 FIG. 12 is a schematic diagram showing an example of tactile presentation in the tactile presentation system 100. In the example shown in FIG. 12, compared with the example shown in FIG. 6, changes in the image and sound displayed on the display unit 13 are shown in response to the amount of displacement in the corrected tactile data. In FIG. 12, the state of the tactile data before correction is shown by a broken line. As shown in FIG. 12, after the correction, compared to FIG. 6, the tactile sensation starts to be presented from the displacement section 202 from a position where the displacement section 202 is pushed in a little (equivalent to 5 degrees), and similarly, images and sounds is output.
 このように、触覚提示装置2を使用する際の触覚の提示開始位置を、操作者の好みや感覚に合わせて調整することが可能になる。触覚提示装置2における触覚の提示の方法を操作者に合わせることにより、操作者による知覚度が向上する。 In this way, it is possible to adjust the presentation start position of the tactile sense when using the tactile sense presentation device 2 according to the operator's preference and sense. By adapting the method of presenting the tactile sensation in the tactile sensation presentation device 2 to the operator, the degree of perception by the operator is improved.
 情報処理装置1に対して、1つの補正量を記憶することとして説明したが、これに限られない。1つの情報処理装置1の記憶部11に、異なる操作者毎の補正量を記憶しておき、触覚提示装置2を操作する操作者が誰であるかの選択を受け付けてから、処理部10は、設定してある補正量を使用して図5のフローチャートに示した処理を実行してもよい。 Although the information processing device 1 has been described as storing one correction amount, the present invention is not limited to this. The storage unit 11 of one information processing device 1 stores correction amounts for different operators, and after receiving the selection of the operator who operates the tactile presentation device 2, the processing unit 10 , the process shown in the flowchart of FIG. 5 may be executed using the set correction amount.
 (第2実施形態)
 第2実施形態では、HMD(Head Mounted Display)を用い、対象物に対する視覚情報を三次元画像で提示する。図13は、第2実施形態における触覚提示システム100の概要図であり、図14は、HMD3の構成を示すブロック図である。第2実施形態における触覚提示システム100の構成は、HMD3を用いることと、HMD3を用いることによる詳細な処理の内容が異なること以外は、第1実施形態における触覚提示システム100と同様の構成である。したがって、第2実施形態の触覚提示システム100の構成のうち、第1実施形態の触覚提示システム100と共通する構成については同一の符号を付して詳細な説明を省略する。
(Second embodiment)
In the second embodiment, an HMD (Head Mounted Display) is used to present visual information regarding a target object as a three-dimensional image. FIG. 13 is a schematic diagram of the tactile presentation system 100 in the second embodiment, and FIG. 14 is a block diagram showing the configuration of the HMD 3. The configuration of the tactile presentation system 100 in the second embodiment is the same as that of the tactile presentation system 100 in the first embodiment, except that the HMD 3 is used and the details of the processing by using the HMD 3 are different. . Therefore, among the configurations of the tactile presentation system 100 of the second embodiment, the same components as those of the tactile presentation system 100 of the first embodiment are given the same reference numerals, and detailed description thereof will be omitted.
 HMD3は、表示部31、動き検知部32、空間検知部33、及び接続部34を備える。HMD3は、本体に、表示部31、動き検知部32、空間検知部33及び接続部34を設けてもよいし、一部は別体に設けられて相互に通信媒体を介して制御信号を授受できるようにしてあってもよい。HMD3は、情報処理装置1の機能を有して一体に構成されてもよい。 The HMD 3 includes a display section 31, a motion detection section 32, a space detection section 33, and a connection section 34. The HMD 3 may be provided with a display section 31, a motion detection section 32, a space detection section 33, and a connection section 34 in the main body, or some may be provided separately and mutually exchange control signals via a communication medium. It may be possible to do so. The HMD 3 may have the functions of the information processing device 1 and be configured integrally.
 表示部31は例えば、小型の液晶ディスプレイ、光学レンズ及び光学系機構を備え、視野角110°以上で三次元画像を表示することが可能である。表示部31は、透明又は半透明のグラス状のディスプレイであり、操作者の実際の視界に、情報処理装置1から出力される画像信号(映像信号を含む)を受け付けて重畳表示させる。表示部31はグラス状のものに限られず、前方に向けて設けられたカメラにて実空間を撮影した画像に、情報処理装置1から出力される画像信号に基づく画像を重畳表示させるものであってもよい。 The display unit 31 includes, for example, a small liquid crystal display, an optical lens, and an optical system mechanism, and is capable of displaying a three-dimensional image at a viewing angle of 110° or more. The display unit 31 is a transparent or semi-transparent glass-like display, and receives an image signal (including a video signal) output from the information processing device 1 and displays it in a superimposed manner in the operator's actual field of view. The display section 31 is not limited to a glass-like one, and may display an image based on an image signal output from the information processing device 1 superimposed on an image taken of real space by a camera provided facing forward. It's okay.
 動き検知部32は、本体(表示部31のカバー及び装着ベルト)の各所に、多様な向きで設けられた複数の三軸加速度センサ及びジャイロセンサと、これらのセンサ群からの信号を集約して出力する制御回路とを備える。動き検知部32により、装着者の頭部の動きが検知される。 The motion detection unit 32 includes a plurality of triaxial acceleration sensors and gyro sensors installed in various directions on the main body (the cover of the display unit 31 and the attachment belt), and aggregates signals from a group of these sensors. and a control circuit that outputs the output. The movement detection unit 32 detects the movement of the wearer's head.
 空間検知部33は、本体の外面に外向きに並設された2つ以上の赤外線カメラと、該赤外線カメラの中間位置に同様に外向きに赤外線を照射するように設けられた赤外線LEDとを用いる。空間検知部33は、HMD3の本体の外側に存在する物体までの距離を測定する深度センサとして機能する。空間検知部33は、HMD3本体から、HMD3を装着する操作者の居場所における壁、床等の静物に対する距離を測り、出力することが可能である。空間検知部33は同様に、HMD3本体から操作者の腕、手、指までの距離を測定し、出力することが可能である。 The space detection unit 33 includes two or more infrared cameras arranged outwardly in parallel on the outer surface of the main body, and an infrared LED disposed at an intermediate position between the infrared cameras so as to similarly emit infrared rays outward. use The space detection unit 33 functions as a depth sensor that measures the distance to an object that exists outside the main body of the HMD 3. The space detection unit 33 can measure and output the distance from the HMD 3 main body to a still object such as a wall or floor in the location of the operator wearing the HMD 3. Similarly, the space detection unit 33 can measure and output the distance from the HMD 3 main body to the operator's arm, hand, and finger.
 接続部34は、情報処理装置1と接続するためのインタフェースである。HMD3は、動き検知部32及び空間検知部33にて測定された結果に対応する信号をそれぞれ、情報処理装置1へ出力すると共に、情報処理装置1から出力される映像信号を取得して表示部31に表示させる。 The connection unit 34 is an interface for connecting to the information processing device 1. The HMD 3 outputs signals corresponding to the results measured by the motion detection unit 32 and the space detection unit 33 to the information processing device 1, and also acquires the video signal output from the information processing device 1 and displays it on the display unit. 31.
 HMD3にスピーカを設け、情報処理装置1の音声出力部14から出力される音声をスピーカにて出力させるようにしてもよい。 A speaker may be provided in the HMD 3, and the audio output from the audio output unit 14 of the information processing device 1 may be output from the speaker.
 上述したような構成の第2実施形態の触覚提示システム100において情報処理装置1は、HMD3を用い、空間検知部33で測定された実空間における静物に対する距離を用い、仮想的な対象物の三次元画像を、動き検知部32にて検知される操作者の動きに合わせて重畳表示し、AR表示を実現する。対象物の選択は、HMD3の表示部31に表示される仮想的対象物の三次元画像の画像内の座標(実空間と合致した座標)に対し、空間検知部33により、触覚提示装置2を把持した手指を重ねたことが検知できた場合に、これを選択したものとして情報処理装置1の処理部10が検知する。 In the tactile presentation system 100 of the second embodiment configured as described above, the information processing device 1 uses the HMD 3 and the distance to the still object in the real space measured by the space detection unit 33 to determine the three-dimensional shape of the virtual object. The original image is superimposed and displayed in accordance with the operator's movement detected by the movement detection unit 32 to realize AR display. To select an object, the space detection section 33 uses the tactile presentation device 2 to select the coordinates in the three-dimensional image of the virtual object displayed on the display section 31 of the HMD 3 (coordinates that match the real space). When it is detected that the gripped fingers are overlapped, the processing unit 10 of the information processing device 1 detects this as being selected.
 HMD3を用いたAR表示と触覚提示装置2とを組み合わせることにより、実空間に対象物を仮想的に配置した画像を、操作者は視覚的に観察しながら、その対象物の感触を触覚提示装置2で覚えることが可能である。その際に、第1実施形態にて説明したものと同様に、触覚提示装置2における触覚提示開始位置を、操作者それぞれに応じて補正することができることで、更に知覚度を向上させることができる。 By combining the AR display using the HMD 3 and the tactile presentation device 2, the operator can visually observe an image in which the object is virtually placed in real space, and feel the touch of the object using the tactile presentation device. It is possible to memorize it in 2. At this time, similar to what was explained in the first embodiment, the tactile presentation start position in the tactile presentation device 2 can be corrected according to each operator, thereby further improving the degree of perception. .
 上述のように開示された実施の形態は全ての点で例示であって、制限的なものではない。本発明の範囲は、請求の範囲によって示され、請求の範囲と均等の意味及び範囲内での全ての変更が含まれる。 The embodiments disclosed above are illustrative in all respects and are not restrictive. The scope of the present invention is indicated by the claims, and includes all changes within the meaning and range equivalent to the claims.
 1 情報処理装置
 10 処理部
 11 記憶部
 13 表示部
 130 設定画面
 P1 制御プログラム
 2 触覚提示装置
 20 制御部
 21 記憶部
 202 変位部
 24 MRFデバイス
 
1 Information processing device 10 Processing unit 11 Storage unit 13 Display unit 130 Setting screen P1 Control program 2 Tactile presentation device 20 Control unit 21 Storage unit 202 Displacement unit 24 MRF device

Claims (7)

  1.  操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置の制御装置であって、
     前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データを前記対象物毎に記憶しておき、
     設定に基づき、前記制御データの前記変位量と制御データとの対応関係を補正し、
     前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力する
     制御装置。
    It has a displacement part that is displaceable in response to an operation by an operator, and generates a force sensation in response to the operation on the displacement part according to the amount of displacement of the displacement part, and presents a tactile sensation of the displayed object. A control device for a tactile presentation device, comprising:
    Control data for generating the force sensation is stored for each object in association with a displacement amount of the displacement portion,
    Correcting the correspondence between the displacement amount of the control data and the control data based on the settings,
    A control device that outputs corrected control data associated with a displacement amount of the displacement portion to the tactile presentation device.
  2.  前記変位量の昇順に連続的に変化する前記制御データを出力し始めるための前記変位量のズレを、前記設定として受け付ける
     請求項1に記載の制御装置。
    The control device according to claim 1, wherein a shift in the amount of displacement for starting to output the control data that changes continuously in ascending order of the amount of displacement is accepted as the setting.
  3.  前記設定を、前記変位量の範囲に対するズレの位置を示すスライドバーを含む設定画面により受け付ける
     請求項2に記載の制御装置。
    The control device according to claim 2, wherein the setting is received on a setting screen including a slide bar indicating a position of deviation with respect to the range of the displacement amount.
  4.  前記設定を、前記変位量の範囲に対するズレの量が異なる複数の候補を選択可能に表示する設定画面により受け付ける
     請求項2に記載の制御装置。
    The control device according to claim 2, wherein the settings are accepted on a settings screen that selectably displays a plurality of candidates having different amounts of deviation from the displacement amount range.
  5.  操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置の制御方法であって、
     前記触覚提示装置に接続されるコンピュータが、
     前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データを前記対象物毎に記憶しておき、
     設定に基づき、前記制御データの前記変位量と制御データとの対応関係を補正し、
     前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力する
     処理部を含む制御方法。
    It has a displacement part that is displaceable in response to an operation by an operator, and generates a force sensation in response to the operation on the displacement part according to the amount of displacement of the displacement part, and presents a tactile sensation of the displayed object. A method for controlling a tactile presentation device, comprising:
    A computer connected to the tactile presentation device,
    Control data for generating the force sensation is stored for each object in association with a displacement amount of the displacement portion,
    Correcting the correspondence between the displacement amount of the control data and the control data based on the settings,
    A control method including a processing section that outputs corrected control data associated with a displacement amount of the displacement section to the tactile presentation device.
  6.  操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置と、
     前記触覚提示装置と通信接続し、表示部を備える情報処理装置と
     を含み、
     前記情報処理装置は、
     前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データと、前記表示部に表示する視覚データとを前記対象物毎に記憶しておき、
     設定に基づき、前記制御データの前記変位量と制御データ及び視覚データとの対応関係をそれぞれ補正し、
     前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力し、
     補正後の視覚データに基づき、前記変位部の変位量に応じて前記表示部に画像を出力する
     触覚提示システム。
    It has a displacement part that is displaceable in response to an operation by an operator, and generates a force sensation in response to the operation on the displacement part according to the amount of displacement of the displacement part, and presents a tactile sensation of the displayed object. a tactile presentation device that
    an information processing device that is communicatively connected to the tactile presentation device and includes a display unit;
    The information processing device includes:
    Control data for causing the force sensation and visual data to be displayed on the display unit are stored for each object in association with the displacement amount of the displacement unit,
    Correcting the correspondence between the displacement amount of the control data and the control data and visual data based on the settings,
    outputting corrected control data associated with the displacement amount of the displacement portion to the tactile presentation device;
    A tactile presentation system that outputs an image to the display unit according to the amount of displacement of the displacement unit based on the corrected visual data.
  7.  操作者の操作に対して変位可能に設けられた変位部を有し、変位部の変位量に応じて、前記変位部への操作に対する力覚を生じさせ、表示される対象物の触覚を提示する触覚提示装置に接続されるコンピュータに、
     前記変位部の変位量に対応付けて、前記力覚を生じさせる制御データを前記対象物毎に記憶しておき、
     設定に基づき、前記制御データの前記変位量と制御データとの対応関係を補正し、
     前記変位部の変位量に対応付けられた補正後の制御データを、前記触覚提示装置へ出力する
     処理を実行させるプログラム製品。
     
    It has a displacement part that is displaceable in response to an operation by an operator, and generates a force sensation in response to the operation on the displacement part according to the amount of displacement of the displacement part, and presents a tactile sensation of the displayed object. a computer connected to a tactile presentation device that
    Control data for generating the force sensation is stored for each object in association with a displacement amount of the displacement portion,
    Correcting the correspondence between the displacement amount of the control data and the control data based on the settings,
    A program product that executes a process of outputting corrected control data associated with a displacement amount of the displacement portion to the tactile presentation device.
PCT/JP2023/009309 2022-03-30 2023-03-10 Control device, control method, haptic presentation system, and program product WO2023189422A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022057116A JP2023148851A (en) 2022-03-30 2022-03-30 Control device, control method, haptic feedback system, and computer program
JP2022-057116 2022-03-30

Publications (1)

Publication Number Publication Date
WO2023189422A1 true WO2023189422A1 (en) 2023-10-05

Family

ID=88200797

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/009309 WO2023189422A1 (en) 2022-03-30 2023-03-10 Control device, control method, haptic presentation system, and program product

Country Status (2)

Country Link
JP (1) JP2023148851A (en)
WO (1) WO2023189422A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011096171A (en) * 2009-11-02 2011-05-12 National Institute Of Information & Communication Technology Multisensory interaction system
WO2017150128A1 (en) * 2016-03-04 2017-09-08 株式会社ソニー・インタラクティブエンタテインメント Control device and control program

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011096171A (en) * 2009-11-02 2011-05-12 National Institute Of Information & Communication Technology Multisensory interaction system
WO2017150128A1 (en) * 2016-03-04 2017-09-08 株式会社ソニー・インタラクティブエンタテインメント Control device and control program

Also Published As

Publication number Publication date
JP2023148851A (en) 2023-10-13

Similar Documents

Publication Publication Date Title
US10509468B2 (en) Providing fingertip tactile feedback from virtual objects
US10133354B2 (en) Gaming device having a haptic-enabled trigger
US10564730B2 (en) Non-collocated haptic cues in immersive environments
JP6591805B2 (en) System and method for a surface element providing an electrostatic haptic effect
US9317123B2 (en) Skin stretch feedback devices, systems, and methods
WO2017150129A1 (en) Control device
US20110148607A1 (en) System,device and method for providing haptic technology
JP5849581B2 (en) Force display device
US20190155387A1 (en) Haptic Accessory Apparatus
KR20200110502A (en) Haptic controller, and System and Method for providing haptic feedback using the haptic controller
JP5352813B2 (en) Non-grounding type haptic device
US20230142242A1 (en) Device for Intuitive Dexterous Touch and Feel Interaction in Virtual Worlds
JP2007026297A (en) Vibration controller, temperature controller and movement controller
WO2023189422A1 (en) Control device, control method, haptic presentation system, and program product
WO2023189423A1 (en) Control device, control method, tactile sense presentation system, and program product
WO2023189425A1 (en) Control device, control method, haptic feedback system, and program product
JP2023148853A (en) Control device, control method, haptic feedback system, and computer program
JP2023148854A (en) Control device, control method, haptic feedback system, and computer program
WO2021176861A1 (en) Information processing device and information processing method, computer program, and augmented reality sensing system
US11550397B1 (en) Systems and methods for simulating a sensation of expending effort in a virtual environment
KR20140106309A (en) Input device for virual reality having the fuction of forth-feedback
WO2024090298A1 (en) Information processing device, information processing method, and input device
WO2024090299A1 (en) Information processing device and information processing method
WO2024090304A1 (en) Input device, control apparatus, control method, information processing apparatus, and information processing method
CN118159933A (en) Information processing device, information processing method, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23779446

Country of ref document: EP

Kind code of ref document: A1