US20190278370A1 - Single actuator haptic effects - Google Patents
Single actuator haptic effects Download PDFInfo
- Publication number
- US20190278370A1 US20190278370A1 US16/276,049 US201916276049A US2019278370A1 US 20190278370 A1 US20190278370 A1 US 20190278370A1 US 201916276049 A US201916276049 A US 201916276049A US 2019278370 A1 US2019278370 A1 US 2019278370A1
- Authority
- US
- United States
- Prior art keywords
- user input
- input element
- actuator
- clutch system
- enabled device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000000694 effects Effects 0.000 title claims abstract description 87
- 238000009877 rendering Methods 0.000 claims description 14
- 238000000034 method Methods 0.000 claims description 10
- 230000008878 coupling Effects 0.000 claims description 8
- 238000010168 coupling process Methods 0.000 claims description 8
- 238000005859 coupling reaction Methods 0.000 claims description 8
- 210000000707 wrist Anatomy 0.000 claims 2
- 230000004044 response Effects 0.000 description 7
- 230000003993 interaction Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000007654 immersion Methods 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 230000002457 bidirectional effect Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 229920001746 electroactive polymer Polymers 0.000 description 2
- 238000010304 firing Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 239000000499 gel Substances 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000000704 physical effect Effects 0.000 description 2
- 229910001285 shape-memory alloy Inorganic materials 0.000 description 2
- 239000002520 smart material Substances 0.000 description 2
- 239000000758 substrate Substances 0.000 description 2
- 239000000725 suspension Substances 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 101150018690 CPS2 gene Proteins 0.000 description 1
- 230000005355 Hall effect Effects 0.000 description 1
- 101100020327 Salvia divinorum KPS gene Proteins 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 239000000090 biomarker Substances 0.000 description 1
- 230000036760 body temperature Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000003155 kinesthetic effect Effects 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 231100000430 skin reaction Toxicity 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000007723 transport mechanism Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/016—Input arrangements with force or tactile feedback as computer generated output to the user
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/25—Output arrangements for video game devices
- A63F13/28—Output arrangements for video game devices responding to control signals received from the game device for affecting ambient conditions, e.g. for vibrating players' seats, activating scent dispensers or affecting temperature or light
- A63F13/285—Generating tactile feedback signals via the game input device, e.g. force feedback
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02N—ELECTRIC MACHINES NOT OTHERWISE PROVIDED FOR
- H02N2/00—Electric machines in general using piezoelectric effect, electrostriction or magnetostriction
- H02N2/0005—Electric machines in general using piezoelectric effect, electrostriction or magnetostriction producing non-specific motion; Details common to machines covered by H02N2/02 - H02N2/16
- H02N2/001—Driving devices, e.g. vibrators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/013—Force feedback applied to a game
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/015—Force feedback applied to a joystick
Abstract
A haptic enabled device that is configured to render one or more haptic effects is provided. The haptic enabled device includes an actuator and a plurality of user input elements. Each of the plurality of user input elements is configured to be selectively coupled to the actuator. In addition, the actuator is configured to be positioned to render the one or more haptic effects at each of the plurality of user input elements.
Description
- This application is a continuation of U.S. patent application Ser. No. 15/665,891, filed on Aug. 1, 2017, which has been incorporated herein by reference in its entirety.
- The embodiments are generally directed to electronic devices, and more particularly, to electronic devices that produce haptic effects.
- Video games and video game systems have become extremely popular. Video game devices or controllers typically use visual and auditory cues to provide feedback to a user. In some interface devices, kinesthetic feedback (e.g., active and resistive force feedback) and/or tactile feedback (e.g., vibration, texture, temperature variation, and the like) may be provided to the user. In general, such feedback is collectively known as “haptic feedback” or “haptic effects.” Haptic feedback provides cues that enhance and simplify a user's interaction with a video game controller, or other electronic device. For example, haptic effects may provide cues to users of video game controllers or other electronic devices to alert the user to specific events, or provide realistic feedback to create greater sensory immersion within a simulated or virtual environment.
- Other devices in which a user interacts with a user input element to cause an action also may benefit from haptic feedback or haptic effects. For example, such devices may include medical devices, automotive controls, remote controls, smartphones, and other similar devices.
- Embodiments of the present invention are directed toward electronic devices configured to produce haptic effects that substantially improve upon the related art.
- Features and advantages of the embodiments are set forth in the description which follows, or will be apparent from the description, or may be learned by practice of the invention.
- In one example, a haptically enabled device that is configured to render one or more haptic effects is provided. The haptically enabled device includes an actuator and a plurality of user input elements. Each of the plurality of user input elements is configured to be selectively coupled to the actuator. In addition, the actuator is configured to be positioned to render the one or more haptic effects at each of the plurality of user input elements.
- It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory and are not intended to limit the invention to the described examples.
- Further embodiments, details, advantages, and modifications will become apparent from the following detailed description of the preferred embodiments, which is to be taken in conjunction with the accompanying drawings.
-
FIG. 1 illustrates a haptically-enabled virtual reality (“VR”) system according to an example embodiment of the invention. -
FIG. 2 illustrates a haptically-enabled controller according to an example embodiment of the present invention. -
FIG. 3 illustrates a structure for rendering haptic effects at multiple user input elements using a single actuator according to an example embodiment of the present invention. -
FIG. 4 illustrates a haptically-enabled controller according to another example embodiment of the present invention. -
FIG. 5 illustrates a multi-directional trigger structure according to an example embodiment of the present invention. -
FIG. 6 illustrates a haptic mouse according to an example embodiment of the present invention. -
FIG. 7 illustrates a flow diagram of a functionality for rendering haptic effects according to an example embodiment of the present invention. - The embodiments of the present invention generally relate to systems and methods that provide haptic effects at different locations, such as different user input elements and/or surfaces, of the same device/system using a single actuator. In the various embodiments, the haptic effects can be independently and/or simultaneously rendered at different locations of the device. The different locations may be adjacent to one another or disposed on different surfaces of the same device.
- Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings. In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it will be apparent to one of ordinary skill in the art that the present invention may be practiced without these specific details. In other instances, well-known methods, procedures, components, and circuits have not been described in detail so as not to unnecessarily obscure aspects of the embodiments. Wherever possible, like reference numbers will be used for like elements.
- In the various embodiments, a variety of user interfaces and methods for using a device are described. In some embodiments, the device is a portable electronic device (e.g., a game controller, wand, console, mobile phone, smartphone, tablet, etc.). It should be understood, however, that the user interfaces and associated methods may be applied to numerous other devices, such as personal computers, smartphones, medical devices, laptops, and the like that may include one or more other physical user-interface devices, such as a keyboard, mouse, trackball, joystick, motion and orientation sensing devices, and the like.
-
FIG. 1 illustrates a haptically-enabled virtual reality (“VR”)system 10 according to an example embodiment of the invention.Example system 10 includes aheadset 11 and ahandheld controller 13.Headset 11 provides audio and visual immersion forVR system 10.Handheld controller 13 provides haptic effects at a variety of user input elements (e.g., triggers, buttons, sticks, etc.) and/or device surfaces in conjunction with the VR immersion experience provided byheadset 11. In one embodiment,system 10 includes two handheld controllers, one for each hand of the user. Although shown inFIG. 1 as a handheld device,handheld controller 13 in other embodiments can be configured to contact other areas of a user's body, and may include other discrete elements for contacting other areas of the user's body.System 10 can also include a gamepad, motion wand, etc. All body contacting elements, includingheadset 11 andhandheld controller 13, are communicatively coupled via a wireless or wired connection. - Internal to
system 10 is a haptic feedback system that generates haptic effects onsystem 10. In one embodiment, the haptic effects are generated usingcontroller 13. The haptic feedback system includes aprocessor 12. Coupled toprocessor 12 are amemory 20 and adrive circuit 16, which is coupled to anactuator 18. -
Processor 12 may comprise one or more general or specific purpose processors configured to perform computation and control functions ofsystem 10.Processor 12 may include an application-specific integrated circuit (“ASIC”), a single integrated circuit, such as a microprocessing device, or may include multiple integrated circuit devices and/or circuit boards working in cooperation to accomplish the functions ofprocessor 12. In addition,processor 12 may execute computer programs, such as an operating system, actuator module, and other applications, stored withinmemory 20. -
Processor 12 can determine what haptic effects are rendered and the order in which the effects are rendered based on high level parameters. In general, the high level parameters that define a particular haptic effect include magnitude, frequency, and duration. Low level parameters, such as streaming motor commands, can also be used to determine a particular haptic effect. A haptic effect may be considered “dynamic” if it includes some variation of these parameters when the haptic effect is generated or a variation of these parameters based on a user's interaction. -
Processor 12 outputs the control signals to drivecircuit 16, which includes electronic components and circuitry used to supplyactuator 18 with the required electrical current and voltage (i.e., “motor signals”) to cause the desired haptic effects to be generated.System 10 may includemultiple actuators 18, and eachactuator 18 may be coupled to multiple user input elements (e.g., buttons, triggers, and other surfaces). In addition, eachactuator 18 may include aseparate drive circuit 16, all coupled to acommon processor 12. In addition, each actuator 18 may be configured to provide haptic effects to two or more user input elements (e.g., triggers, buttons, sticks, etc.) or other surfaces ofhandheld controller 13. -
Memory device 20 can be any type of transitory or non-transitory storage device or computer-readable medium, such as random access memory (“RAM”), dynamic RAM (DRAM), static RAM (SRAM), read only memory (“ROM”), flash memory, cache memory, and/or any other type of computer-readable medium. Communication media may include computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism, and includes any information delivery media.Memory 20 may be located internal toprocessor 12, or any combination of internal and external memory. -
Memory 20 stores instructions that are executed byprocessor 12, such as operating system instructions or haptic instructions. Among the instructions,memory 20 includes a hapticeffect selections module 22 which includes instructions that, when executed byprocessor 12, select one of a plurality of user input elements or device surfaces for rendering haptic effects using a single actuator, as discussed in more detail below. By using a single actuator to render haptic effects at multiple user input elements, the cost and weight ofhandheld device 13 is reduced. In addition, significant the power saving is also achieved. -
Actuator 18 may be any type of device that generates haptic effects, and can be physically located in any area ofsystem 10, typically located withinhandheld controller 13, to be able to generate the desired haptic effect to the desired area of a user's body. In one embodiment,actuator 18 is a haptic output device that generates vibrotactile haptic effects. Actuators used for this purpose may include an electromagnetic actuator such as an Eccentric Rotating Mass (“ERM”) in which an eccentric mass is moved by a motor, a Linear Resonant Actuator (“LRA”) in which a mass attached to a spring is driven back and forth, or a “smart material” such as piezoelectric, electroactive polymers or shape memory alloys. Haptic output devices may also include a device such as an electrostatic friction (“ESF”) device or an ultrasonic surface friction (“USF”) device, or a device that induces acoustic radiation pressure with an ultrasonic haptic transducer. Other devices can use a haptic substrate and a flexible or deformable surface, and devices can provide projected haptic output such as a puff of air using an air jet, etc. Haptic output devices can further include a device that provides thermal haptic effects (e.g., heats up or cools off). -
System 10 further includes asensor 28 coupled toprocessor 12.Sensor 28 can be used to detect any type of properties of the user of system 10 (e.g., a biomarker such as body temperature, heart rate, etc.), or of the context of the user or the current context (e.g., the location of the user, the temperature of the surroundings, etc.). -
Sensor 28 can be configured to detect a form of energy, or other physical property, such as, but not limited to, sound, movement, acceleration, physiological signals, distance, flow, force/pressure/strain/bend, humidity, linear position, orientation/inclination, radio frequency, rotary position, rotary velocity, manipulation of a switch, temperature, vibration, or visible light intensity.Sensor 28 can further be configured to convert the detected energy, or other physical property, into an electrical signal, or any signal that represents virtual sensor information.Sensor 28 can be any device, such as, but not limited to, an accelerometer, an electrocardiogram, an electroencephalogram, an electromyograph, an electrooculogram, an electropalatograph, a galvanic skin response sensor, a capacitive sensor, a hall effect sensor, an infrared sensor, an ultrasonic sensor, a pressure sensor, a fiber optic sensor, a flexion sensor (or bend sensor), a force-sensitive resistor, a load cell, a LuSense CPS2 155, a miniature pressure transducer, a piezo sensor, a strain gage, a hygrometer, a linear position touch sensor, a linear potentiometer (or slider), a linear variable differential transformer, a compass, an inclinometer, a magnetic tag (or radio frequency identification tag), a rotary encoder, a rotary potentiometer, a gyroscope, an on-off switch, a temperature sensor (such as a thermometer, thermocouple, resistance temperature detector, thermistor, or temperature-transducing integrated circuit), a microphone, a photometer, an altimeter, a biological monitor, a camera, or a light-dependent resistor. -
System 10 further includes acommunication interface 25 that allowssystem 10 to communicate over the Internet/cloud 50. Internet/cloud 50 can provide remote storage and processing forsystem 10 and allowsystem 10 to communicate with similar or different types of devices. Further, any of the processing functionality described herein can be performed by a processor/controller remote fromsystem 10 and communicated viainterface 25. The various components ofsystem 10, exceptactuator 18, may be located remote fromcontroller 13, such as in a central console or in an Internet based cloud device. -
FIG. 2 illustrates a haptically-enabledcontroller 200 according to an example embodiment of the present invention. As shown inFIG. 2 , which depicts the capability of rendering of haptic effects at different components of controller 200 (such ascontroller 13 ofFIG. 1 ) in first and second stages,controller 200 may include a variety of components such as ahousing 202, firstuser input element 211 and second user element 212 (e.g., buttons, triggers, sticks, etc.),switchable actuator 221, and clutch system 222 (not labeled in stage 1). In a first stage,clutch system 222 engagesswitchable actuator 221 withgearbox 226 to provide haptic feedback atuser input element 211. In a second stage,clutch system 222 engages switchable motor/actuator 221 withgearbox 227 to provide haptic feedback atuser input element 212. - In an example configuration of the first stage,
switchable actuator 221 laterally (e.g., inwardly) positionsrotating shaft 223 to coupledrive gear 224 withgearbox 226 ofuser input element 211.Shaft 223 is coupled toswitchable actuator 221 at one end and either offirst gearbox 226 orsecond gearbox 227 of respective firstuser input element 211 orsecond user element 212 at the other end. An alignment ofdrive gear 224 withgearbox 226 allowsactuator 221 to render one or more haptic effects (e.g., pushing/pulling, vibrating, etc.) atuser input element 211. Similarly, in an example configuration of the second stage,switchable actuator 221 laterally (e.g., outwardly) movesshaft 223 to coupledrive gear 225 withgearbox 227 ofuser input element 212. As shown inFIG. 2 , an alignment offirst drive gear 224 orsecond drive gear 225 with respectivefirst gear box 226 orsecond gearbox 227 allows asingle actuator 221 to render one or more haptic effects (e.g., pushing/pulling, vibrating, etc.) at firstuser input element 211 andsecond user element 212. In another configuration,shaft 223 may be laterally moved back and forth to render alternating haptic effects at firstuser input element 211 andsecond user element 212. - In some configurations,
shaft 223 may be configured to move relative to an output location, such as firstuser input element 211 or seconduser input element 212. For example, sincesecond user element 212 may be pressed in by the user to activate, a haptic effect may causefirst drive gear 224 to remain at rest so thatsecond gearbox 227 engagesclutch system 222. Here, the lack of a haptic effect call may causefirst drive gear 224 to recede towardswitchable actuator 221 to prevent contact withclutch system 222 if no haptic effect call is made. In addition, gearbox drag is avoided since, at rest,switchable actuator 221 makes contact with no endpoints and can move to avoid contact withclutch system 222. - The various components of
controller 200 may be alternatively configured to independently provide haptic feedback to one of firstuser input element 211 andsecond user element 212 using a singleswitchable actuator 221. For example, incontroller 200,switchable actuator 221 is coupled to one of firstuser input element 211 andsecond user element 212 using any of a variety of drive mechanisms, such as gearboxes, linkages, magnetic couplings, electromagnetic couplings, and the like. For example, electromagnetic clutches operate electrically, but transmit torque mechanically. In another example, the linkages may be magnetic, and accordingly, may not require a physical coupling. In the various configurations, a combination of linkages, levers, drive arms, gearboxes, magnets, electromagnets, locking mechanisms, and the like may be used. - In contrast with conventional systems, in which a respective haptic output device is generally needed for each user input element, embodiments of the present invention provide that the haptic feedback may be rendered at multiple user input elements using a single actuator. Alternatively, or additionally,
controller 200 may be configured to simultaneously render haptic feedback to a subset of a plurality of triggers, such asuser input element 211 and another user input element (not shown) that is disposed opposite ofuser input element 211, using a singleswitchable actuator 221. - In some embodiments, a haptic control signal may be varied to select one of first
user input element 211 andsecond user element 212 for rendering the haptic effect. By applying different haptic control signals toclutch system 222 and/oractuator 221, different user input elements may be targeted. Here, each of firstuser input element 211 andsecond user element 212 or surfaces ofcontroller 200 may be configured to render haptic effects in response to predetermined haptic control signals (e.g., based on amplitude, waveform, frequency, etc.). -
FIG. 3 illustrates astructure 300 for rendering haptic effects at multiple user input elements using a single actuator according to an example embodiment of the present invention. - As shown in
FIG. 3 ,structure 300 may include a variety of components such as firstuser input element 311 and second user element 312 (e.g., triggers, buttons, sticks, and other user input elements or surfaces, etc.), tension springs 313, androtatable actuator 321. In a first configuration, arotatable actuator 321 couples withuser input element 311 to provide haptic feedback (e.g., vibration, push/pull, etc.). In a second configuration,rotatable actuator 321 couples withuser input element 312 to provide haptic feedback. Althoughrotatable actuator 321 is depicted as moving in a clockwise or counter clockwise direction, other configurations are also feasible. For example,actuator 321 may be configured to traverse lateral directions along one or more rails (not shown). In either configuration, tensions springs 313 are configured to maintain a default position of firstuser input element 311 andsecond user element 312. - When a
drive arm 323 ofrotatable actuator 321 is disposed in a central position between firstuser input element 311 andsecond user element 312,actuator mass 322 is also disposed in a central position and is free to move back and forth to render vibration and/or other haptic effects. Whendrive arm 323 turns counter clockwise,drive arm 323 engagesuser input element 311 and one or more forces may be rendered atuser input element 311. Whendrive arm 323 turns clockwise,drive arm 323 engagesuser input element 312 and one or more forces may be rendered atuser input element 312. In some instances, in order to render push and/or pull haptic effects at firstuser input element 311 andsecond user element 312,drive arm 323 may utilize magnets, electromagnets, pin locks, and the like to coupleactuator mass 322 with firstuser input element 311 andsecond user element 312. In addition, magnetic couplings may be used to vary the load ofactuator mass 322 to provide a tunable actuator. - Accordingly, the various components of
structure 300 are configured to independently provide haptic feedback to one of a plurality of user input elements, such as one of firstuser input element 311 andsecond user element 312, using a singlerotatable actuator 321. For example, instructure 300,rotatable actuator 321 may be coupled to one of firstuser input element 311 andsecond user element 312 using any of a variety of drive mechanisms, such as gearboxes, linkages, magnetic couplings, electromagnetic couplings, and the like. - In a gaming example, a user may utilize a controller (e.g., a gun) with a first trigger, such as
user input element 311, and experience a haptic effect when pressing the first trigger asdrive arm 323 has engaged the first trigger. The user then taps the controller on a surface (e.g., a table) and experiences another haptic effect, such as vibration forces, from therotatable actuator 321 oscillating back and forth around the center position. The user then presses a second trigger, such asuser input element 312, on the controller and experiences one or more haptic effects on the second trigger asdrive arm 323 has engaged the second trigger. -
FIG. 4 illustrates a haptically-enabledcontroller 400 according to another example embodiment of the present invention. As shown inFIG. 4 ,controller 400 may include a variety of components such as ahousing 402, firstuser input element 411 andsecond user element 412, and motor/actuator 421.Actuator 421 is configured to independently and/or simultaneously render haptic effects at firstuser input element 411 andsecond user element 412. In some configurations, output by motor/actuator 421 may be selectively applied to one of firstuser input element 411 andsecond user element 412 using a bidirectional actuator. For example, the bidirectional actuator used for motor/actuator 421 may be driven with to produce a haptic effect only in one direction. In this example, with a configurable mechanical assembly, motor/actuator 421 does not make contact with the components in the opposite direction in which it is firing. In other configurations, motor/actuator 421 is disposed on one or more rails and moved by another lower power motor/actuator such that it is in contact with either firstuser input element 411 or seconduser input element 412. Alternatively, a mechanical structure may be configured such that when firstuser input element 411 or seconduser input element 412 is pressed, it comes in contact with motor/actuator 421. Here, if both user input elements are pressed, then both firstuser input element 411 and seconduser input element 412 render haptic effects. - In this embodiment,
actuator 421 is configured to engage either of firstuser input element 411 andsecond user element 412 depending on a haptic control signal (e.g., supplied by the processor, such asprocessor 12 ofFIG. 1 ). For example,actuator 421 may engage with either of firstuser input element 411 andsecond user element 412 according to an input voltage. In another example,user input element 411 may render haptic effects in response to a 3V square signal whileuser input element 412 may render haptic effects in response a 5V pulse signal. Accordingly, the different haptic control signals (e.g., ‘Signal A’ or ‘Signal B’ ofFIG. 4 ) may be used to render haptic effects (e.g., push, pull, vibration, etc.) at respective firstuser input element 411 andsecond user element 412. - In some configurations, first
user input element 411 andsecond user element 412 may include one or more surface ‘smart’ materials or gels having a variety of predetermined electrical properties configured to provide additional haptic effects (e.g., stiffening, loosening, textures, thermal, etc.). Example smart materials include piezo, shape memory alloys, electroactive polymers, and the like. In yet some other configurations, firstuser input element 411 andsecond user element 412 may be separated fromhaptic output device 421 by one or more intermediate layers (e.g., a substrate or gel) that causes firstuser input element 411 andsecond user element 412 to vibrate in a different manner (e.g., frequency, amplitude) in response to the different haptic control signals. Here, the intermediate layers also may act as a variable dampening element. -
FIG. 5 illustrates amulti-directional trigger structure 500 according to an example embodiment of the present invention. As shown inFIG. 5 ,multi-directional trigger structure 500 may include a variety of components such as a first-directional section 511, second-directional section 512,actuator 521, andlinkages actuator 521 may independently and/or simultaneously render one or more haptic effects at first-directional section 511 or second-directional section 512. - In some instances, an
optional hinge 513 may physically join or couple first-directional section 511 and second-directional section 512. By usinghinge 513,linkage 531 may be used to render haptic effects (e.g., push, pull, vibration, etc.) to both first-directional section 511 and second-directional section 512. However,linkage 532 may be used to render haptic effects to only second-directional section 512. Alternatively, or additionally, first-directional section 511 and second-directional section 512 may be comprised of a flexible trigger. The use ofmultiple linkages directional section 511 and second-directional section 512, to sharehaptic output device 521. - The use of
multiple linkages haptic output device 521 may provide a variety of trigger contours or shapes. In other words, the resulting trigger's shape is configurable. In some embodiments, deformable trigger shapes may be used in connection with contextual interactions. For example, contextual interactions may include the holding of an object in virtual reality. Segments of a trigger (or other user input element) may be configured to accurately represent the surface geometry of a virtual reality object. In another example, if in virtual reality, a user's finger is visualized touching a flat virtual surface, the traditional ergonomic trigger shape may remove or reduce the scalloping to render a flat surface. In other words, depending on the context of the event, the trigger can be a different shape due using bifurcated parts. For example, the bifurcated parts can be in line and move together when the trigger is used as the gas pedal for a vehicle. When acting as a trigger, the bifurcated parts can make an “L” shape to represent the trigger of a gun. In addition, the haptic effects may be applied to triggers having a variety of shapes and sizes. Additional end-stop haptic capability is also provided. End-stop haptic capability generally refers to the grounding of the trigger (or button or other user input element) against an end stop. At the end-stop, the user's finger-tip will not be overcome by the amount of force that the trigger can push back with. However, if the haptic effect is rendered immediately below the user's knuckle, the user is able to feel the haptic effect even when pulling the trigger very hard. -
Linkages directional section 511 and second-directional section 512 may be configured to move along multiple axes. - In some embodiments, the shape of
trigger structure 500 may be deformable or semi-deformable. For example,trigger structure 500 may include a switch component (not shown) that may be “toggled” into different positions. This would enable the user to control the shape of thetrigger structure 500, especially if thetrigger structure 500 is composed of a rubberized material. The toggle functionality is particularly useful for devices with a single trigger that may be used in either left or right hands. In addition, the switch component also may be reconfigured in response to each use (e.g., unique trigger shapes for firing individual weapons). -
FIG. 6 illustrates ahaptic mouse 600 according to another example embodiment of the present invention. In addition tohousing 602 androtary wheel 603,haptic mouse 600 includes a plurality ofhaptic regions haptic mouse 600, such asrotary wheel 603 and/orkeypad 604. Subsets of haptic regions 611-615 may render haptic effects using a single haptic output device, as described above. - In addition, any of the user input elements described herein may be configured to provide isolated haptic effects (e.g., vibration and/or deformations applied that are isolated from the device housing). In some embodiments, the user input elements may be indirectly coupled to the housing by one or more suspension flexures (e.g., springs). When a haptic effect is applied to the user input elements, the suspension flexures substantially dampen the haptic effects such that they are not applied to the housing. Additionally, or alternatively, a rubber ring or gasket (not shown) may surround the user input elements to dampen the haptic effects.
-
FIG. 7 illustrates a flow diagram offunctionality 700 for rendering haptic effects according to an example embodiment of the present invention. In some instances, the functionality of the flow diagram ofFIG. 7 is implemented by software stored in memory or other computer readable or tangible media, and executed by a processor. In other instances, the functionality may be performed by hardware (e.g., through the use of an application specific integrated circuit (“ASIC”), a programmable gate array (“PGA”), a field programmable gate array (“FPGA”), etc.), or any combination of hardware and software functionally implemented bysystem 10 ofFIG. 1 . - At the outset,
functionality 700 may optionally initialize the haptic device, such ascontroller 13 ofFIG. 1 , at 710. For example,functionality 700 may determine or otherwise initialize the respective configurations of the various components of the clutch system and/or actuator. In another example, the state of the user input elements may be configured according to one or more user or designer preferences. - Next, at 720, a haptic control signal is received at the clutch system and/or actuator. In response to receiving the haptic control signal,
functionality 700 evaluates the haptic control signal to identify one of a plurality of user input elements configured to be connected to the same actuator for rendering the haptic effect stored within the first haptic control signal, at 730. In response to receiving the haptic control signal,functionality 700 also may reconfigure the components of the clutch system and/or actuator. For example, the actuator may be reconfigured according to an actuator type. The actuator is coupled or otherwise connected to the user input element at 740. In turn, the haptic effect is rendered at the identified haptic output device at 750. - In some embodiments,
functionality 700 may evaluate spatial data (e.g., whether device is in user's left or right hand). One or more sensors (e.g., such assensor 28 ofFIG. 1 ) may be used to determine the controller's orientation to the user or the orientation to other devices. Accordingly, the haptic effects may be modified based on spatial data (e.g., based on determination of finger contact). For example, a haptic effect designed for the first user input element may be rendered at the second input element if the user has no physical contact with the first user input element. - By implementing the various embodiments described herein, enhanced haptic effects may be more efficiently provided by adopting the various embodiments of the present invention. The systems and methods described herein provide multiple haptic effects at different locations of the same device/system using a single actuator. In the various embodiments, the haptic effects are independently or simultaneously rendered at different locations of the device. By reducing the number of haptic output devices needed to drive multiple user input elements, power is saved and cost is reduced.
- One having ordinary skill in the art will readily understand that the invention as discussed above may be practiced with steps in a different order, and/or with elements in configurations which are different than those which are disclosed. Therefore, although the invention has been described based upon these preferred embodiments, it would be apparent to those of skill in the art that certain modifications, variations, and alternative constructions would be apparent, while remaining within the spirit and scope of the invention. In order to determine the metes and bounds of the invention, therefore, reference should be made to the appended claims.
Claims (20)
1. A haptically enabled device configured to render one or more haptic effects, the haptically enabled device comprising:
an actuator;
a first user input element that is configured to be selectively coupled to the actuator, and
a second user input element that is configured to be selectively coupled to the actuator,
wherein the actuator is configured to be positioned to render the one or more haptic effects at each of the first user input element and the second user input element.
2. The haptically enabled device of claim 1 , further comprising:
a clutch system coupled to the actuator via a moveable shaft,
wherein the clutch system comprises a plurality of drive gears,
wherein the actuator is switchable between at least two positions of the clutch system comprising
a first position in which the actuator is coupled to the first user input element via a first drive gear of the clutch system and to a first linkage or gearbox of the first user input element; and
a second position in which the actuator is coupled to the second user input element via a second drive gear of the clutch system and to a second linkage or gearbox of the second user input element.
3. The haptically enabled device of claim 2 , wherein the first drive gear and the second drive gear have different respective sizes.
4. The haptically enabled device of claim 2 , wherein the first drive gear and the second drive gear have different respective diameters.
5. The haptically enabled device of claim 1 , wherein the actuator is configured to move a moveable shaft in a first lateral direction to laterally position the clutch system at a first position, and to move the moveable shaft in a second lateral direction to laterally position the clutch system at a second position.
6. The haptically enabled device of claim 1 , wherein the haptically enabled device comprises a portable or handheld device, wherein the portable or handheld device comprises a toy, gaming console, handheld video game system, gamepad, game controller, desktop computer, portable multifunction device such as a cell phone, smartphone, personal digital assistant (PDA), eReader, portable reading device, handheld reading device, laptop, tablet computer, digital music player, remote control, or medical instrument.
7. The haptically enabled device of claim 1 , wherein the haptically enabled device is embedded in another device.
8. The haptically enabled device of claim 7 , wherein the another device comprises a vehicle or a wearable device.
9. The haptically enabled device of claim 8 , wherein the wearable device comprises a wrist watch, jewelry, arm band, or gloves.
10. A haptically enabled device comprising a clutch system coupled to a single actuator via a shaft, the clutch system configured to selectively position the single actuator at a plurality of positions within a housing, each of the plurality of positions corresponding to a respective one of a plurality of user input elements.
11. The haptically enabled device of claim 10 , wherein the plurality of positions comprises:
a first position in which the clutch system couples the single actuator to a first user input element via a first drive gear of the clutch system and to a first linkage or gearbox of the first user input element; and
a second position in which the clutch system couples the single actuator to a second user input element via a second drive gear of the clutch system and to a second linkage or gearbox of the second user input element.
12. The haptically enabled device of claim 11 , wherein
in the first position, the single actuator is configured to render one or more haptic effects at the first user input element, and
in the second position, the single actuator is configured to render one or more haptic effects at the second user input element.
13. The haptically enabled device of claim 12 , wherein
in the first position, the single actuator is configured to render one or more haptic effects at the first user input element upon a movement of the first user input element with respect to the housing, and
in the second position, the single actuator is configured to render one or more haptic effects at the second user input element upon a movement of the second user input element with respect to the housing.
14. The haptically enabled device of claim 10 , wherein the haptically enabled device comprises a portable or handheld device, wherein the portable or handheld device comprises a toy, gaming console, handheld video game system, gamepad, game controller, desktop computer, portable multifunction device such as a cell phone, smartphone, personal digital assistant (PDA), eReader, portable reading device, handheld reading device, laptop, tablet computer, digital music player, remote control, or medical instrument.
15. The haptically enabled device of claim 10 , wherein the haptically enabled device is embedded in another device.
16. The haptically enabled device of claim 15 , wherein the another device comprises a vehicle or a wearable device, wherein the wearable device comprises a wrist watch, jewelry, arm band, or gloves.
17. A haptic feedback method comprising:
selectively positioning a clutch system using a single actuator at one of a plurality of positions of the single actuator within a housing; and
rendering one or more haptic effects at one or more of a plurality of user input elements at one or more of the plurality of positions.
18. The haptic feedback method of claim 17 , further comprising:
coupling, at a first of the plurality of positions, the clutch system to a first of the plurality of user input elements via a first drive gear of the clutch system, via a first linkage or gearbox of the first user input element; and
coupling, at a second of the plurality of positions, the clutch system to a second of the plurality of user input elements via a second drive gear of the clutch system, via a second linkage or gearbox of the second user input element.
19. The haptic feedback method of claim 18 , further comprising:
rendering, using the single actuator, one or more of the haptic effects at the first user input element when the clutch system is in the first position, and
rendering, using the single actuator, one or more of the haptic effects at the second user input element when the clutch system is in a second position.
20. The haptic feedback method of claim 19 , wherein
in the first position, the rendering of one or more of the haptic effects at the first user input element occurs upon a movement of the first user input element with respect to the housing, and
in the second position, the rendering of one or more of the haptic effects at the second user input element occurs upon a movement of the second user input element with respect to the housing.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/276,049 US20190278370A1 (en) | 2017-08-01 | 2019-02-14 | Single actuator haptic effects |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/665,891 US10241577B2 (en) | 2017-08-01 | 2017-08-01 | Single actuator haptic effects |
US16/276,049 US20190278370A1 (en) | 2017-08-01 | 2019-02-14 | Single actuator haptic effects |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/665,891 Continuation US10241577B2 (en) | 2017-08-01 | 2017-08-01 | Single actuator haptic effects |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190278370A1 true US20190278370A1 (en) | 2019-09-12 |
Family
ID=63077674
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/665,891 Expired - Fee Related US10241577B2 (en) | 2017-08-01 | 2017-08-01 | Single actuator haptic effects |
US16/276,049 Abandoned US20190278370A1 (en) | 2017-08-01 | 2019-02-14 | Single actuator haptic effects |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/665,891 Expired - Fee Related US10241577B2 (en) | 2017-08-01 | 2017-08-01 | Single actuator haptic effects |
Country Status (5)
Country | Link |
---|---|
US (2) | US10241577B2 (en) |
EP (1) | EP3438794A1 (en) |
JP (1) | JP2019029006A (en) |
KR (1) | KR20190013540A (en) |
CN (1) | CN109324682A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10684639B2 (en) * | 2014-11-04 | 2020-06-16 | Alps Alpine Co., Ltd. | Operation device |
US20230116571A1 (en) * | 2017-12-28 | 2023-04-13 | Cilag Gmbh International | Display arrangements for robot-assisted surgical platforms |
WO2023244624A1 (en) * | 2022-06-13 | 2023-12-21 | Meta Platforms Technologies, Llc | Force feedback trigger |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109070138B (en) * | 2016-04-19 | 2021-02-09 | 日本电信电话株式会社 | Simulated force sense generating device |
US11132059B2 (en) * | 2017-09-14 | 2021-09-28 | Logitech Europe S.A. | Input device with haptic interface |
US11150731B2 (en) * | 2018-09-28 | 2021-10-19 | Apple Inc. | Multi-modal haptic feedback for an electronic device using a single haptic actuator |
CN111782046A (en) * | 2020-06-30 | 2020-10-16 | 瑞声新能源发展(常州)有限公司科教城分公司 | Haptic effect acquisition method and system |
US11571617B2 (en) * | 2020-12-18 | 2023-02-07 | Dell Products, Lp | System and method for coordinating gaming haptic response across peripheral device ecosystem |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002196883A (en) | 2000-12-22 | 2002-07-12 | Alps Electric Co Ltd | Manual input device and on-vehicle equipment controller using the manual input device |
US20120092263A1 (en) * | 2009-10-15 | 2012-04-19 | Pacinian Corporation | Haptic keyboard featuring a satisfying tactile keypress experience |
US9293015B2 (en) | 2013-09-09 | 2016-03-22 | Immersion Corporation | Electrical stimulation haptic feedback interface |
US10379614B2 (en) | 2014-05-19 | 2019-08-13 | Immersion Corporation | Non-collocated haptic cues in immersive environments |
US9174134B1 (en) | 2014-11-12 | 2015-11-03 | Immersion Corporation | Peripheral device with haptic diminishment prevention component |
EP3020457A1 (en) * | 2014-11-12 | 2016-05-18 | Immersion Corporation | Haptic controller |
-
2017
- 2017-08-01 US US15/665,891 patent/US10241577B2/en not_active Expired - Fee Related
-
2018
- 2018-06-28 JP JP2018123422A patent/JP2019029006A/en active Pending
- 2018-07-10 EP EP18182711.4A patent/EP3438794A1/en not_active Withdrawn
- 2018-07-24 KR KR1020180085995A patent/KR20190013540A/en unknown
- 2018-07-24 CN CN201810814929.6A patent/CN109324682A/en active Pending
-
2019
- 2019-02-14 US US16/276,049 patent/US20190278370A1/en not_active Abandoned
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10684639B2 (en) * | 2014-11-04 | 2020-06-16 | Alps Alpine Co., Ltd. | Operation device |
US20230116571A1 (en) * | 2017-12-28 | 2023-04-13 | Cilag Gmbh International | Display arrangements for robot-assisted surgical platforms |
WO2023244624A1 (en) * | 2022-06-13 | 2023-12-21 | Meta Platforms Technologies, Llc | Force feedback trigger |
Also Published As
Publication number | Publication date |
---|---|
US20190041986A1 (en) | 2019-02-07 |
EP3438794A1 (en) | 2019-02-06 |
US10241577B2 (en) | 2019-03-26 |
JP2019029006A (en) | 2019-02-21 |
KR20190013540A (en) | 2019-02-11 |
CN109324682A (en) | 2019-02-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10241577B2 (en) | Single actuator haptic effects | |
US10564730B2 (en) | Non-collocated haptic cues in immersive environments | |
US9715278B2 (en) | User interface device provided with surface haptic sensations | |
JP2020072962A (en) | Programmable haptic peripheral | |
Bai et al. | Elastomeric haptic devices for virtual and augmented reality | |
US20180093178A1 (en) | Systems and Methods for Performing Haptic Conversion | |
JP2019121397A (en) | System and method for providing haptic effect related to contacting or grasping virtual object | |
US20200026354A1 (en) | Adaptive haptic effect rendering based on dynamic system identification | |
EP3598272A1 (en) | Trigger button for haptic controller | |
CN113508355A (en) | Virtual reality controller | |
JP2020109622A (en) | Variable curvature interactive device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: IMMERSION CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:RIHN, WILLIAM S.;MOTAMEDI, MOHAMMADREZA;GRANT, DANNY A.;AND OTHERS;SIGNING DATES FROM 20170726 TO 20170801;REEL/FRAME:048336/0771 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE |