WO2018234848A1 - Amplification de son à visée affective pour applications automobiles - Google Patents
Amplification de son à visée affective pour applications automobiles Download PDFInfo
- Publication number
- WO2018234848A1 WO2018234848A1 PCT/IB2017/053672 IB2017053672W WO2018234848A1 WO 2018234848 A1 WO2018234848 A1 WO 2018234848A1 IB 2017053672 W IB2017053672 W IB 2017053672W WO 2018234848 A1 WO2018234848 A1 WO 2018234848A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- vehicle
- soundscape
- sound
- sensor data
- sound cue
- Prior art date
Links
- 230000003416 augmentation Effects 0.000 title description 2
- 238000013507 mapping Methods 0.000 claims abstract description 12
- 230000036651 mood Effects 0.000 claims abstract description 8
- 238000000034 method Methods 0.000 claims description 14
- 230000000694 effects Effects 0.000 claims description 6
- 210000003205 muscle Anatomy 0.000 claims description 3
- 230000001427 coherent effect Effects 0.000 abstract 1
- 230000001133 acceleration Effects 0.000 description 11
- 238000005516 engineering process Methods 0.000 description 5
- 230000006870 function Effects 0.000 description 3
- 230000033001 locomotion Effects 0.000 description 3
- 230000003190 augmentative effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 108091064702 1 family Proteins 0.000 description 1
- 206010048909 Boredom Diseases 0.000 description 1
- 206010041349 Somnolence Diseases 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000003466 anti-cipated effect Effects 0.000 description 1
- 230000003935 attention Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000036461 convulsion Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000002592 echocardiography Methods 0.000 description 1
- 230000002996 emotional effect Effects 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000003387 muscular Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- JTJMJGYZQZDUJJ-UHFFFAOYSA-N phencyclidine Chemical compound C1CCCCN1C1(C=2C=CC=CC=2)CCCCC1 JTJMJGYZQZDUJJ-UHFFFAOYSA-N 0.000 description 1
- 230000008054 signal transmission Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000000153 supplemental effect Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
- G10H1/0025—Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/021—Background music, e.g. for video sequences or elevator music
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2220/00—Input/output interfacing specifically adapted for electrophonic musical tools or instruments
- G10H2220/155—User input interfaces for electrophonic musical instruments
- G10H2220/351—Environmental parameters, e.g. temperature, ambient light, atmospheric pressure, humidity, used as input for musical purposes
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2220/00—Input/output interfacing specifically adapted for electrophonic musical tools or instruments
- G10H2220/155—User input interfaces for electrophonic musical instruments
- G10H2220/351—Environmental parameters, e.g. temperature, ambient light, atmospheric pressure, humidity, used as input for musical purposes
- G10H2220/355—Geolocation input, i.e. control of musical parameters based on location or geographic position, e.g. provided by GPS, WiFi network location databases or mobile phone base station position databases
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2220/00—Input/output interfacing specifically adapted for electrophonic musical tools or instruments
- G10H2220/155—User input interfaces for electrophonic musical instruments
- G10H2220/371—Vital parameter control, i.e. musical instrument control based on body signals, e.g. brainwaves, pulsation, temperature or perspiration; Biometric information
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2220/00—Input/output interfacing specifically adapted for electrophonic musical tools or instruments
- G10H2220/155—User input interfaces for electrophonic musical instruments
- G10H2220/395—Acceleration sensing or accelerometer use, e.g. 3D movement computation by integration of accelerometer data, angle sensing with respect to the vertical, i.e. gravity sensing
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/075—Musical metadata derived from musical analysis or for use in electrophonic musical instruments
- G10H2240/085—Mood, i.e. generation, detection or selection of a particular emotional content or atmosphere in a musical piece
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2250/00—Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
- G10H2250/315—Sound category-dependent sound synthesis processes [Gensound] for musical use; Sound category-specific synthesis-controlling parameters or control means therefor
- G10H2250/371—Gensound equipment, i.e. synthesizing sounds produced by man-made devices, e.g. machines
- G10H2250/381—Road, i.e. sounds which are part of a road, street or urban traffic soundscape, e.g. automobiles, bikes, trucks, traffic, vehicle horns, collisions
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2250/00—Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
- G10H2250/315—Sound category-dependent sound synthesis processes [Gensound] for musical use; Sound category-specific synthesis-controlling parameters or control means therefor
- G10H2250/395—Gensound nature
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2430/00—Signal processing covered by H04R, not provided for in its groups
- H04R2430/01—Aspects of volume control, not necessarily automatic, in sound systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2499/00—Aspects covered by H04R or H04S not otherwise provided for in their subgroups
- H04R2499/10—General applications
- H04R2499/13—Acoustic transducers and sound field adaptation in vehicles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/11—Positioning of individual sound objects, e.g. moving airplane, within a sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/13—Aspects of volume control, not necessarily automatic, in stereophonic sound systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/15—Aspects of sound capture and related signal processing for recording or reproduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
Definitions
- Embodiments of the invention generally relate to in-vehicle audio and, more particularly, to adaptively providing supplemental sounds to enhance drivers' and passengers' awareness of their vehicle's surroundings.
- in-vehicle audio has been unrelated to the ambient conditions of the vehicle.
- the vehicle's sound system may play music or other entertainment for the driver and/or passengers.
- some automotive manufacturers have begun using the vehicle's sound system to augment or otherwise make engine sounds audible in the passenger compartment.
- electric vehicles lack a conventional engine and, as such, may fail to provide the driver with the cues needed to maintain a proper level of attentiveness for vehicle operation.
- the invention includes a system for providing contextual enhancement of audio in a vehicle, comprising one or more vehicle sensors configured to provide sensor data relevant to the vehicle, a processor, one or more computer-storage media storing computer-executable instructions that, when executed by the processor perform steps of determining a vehicle event based on the sensor data, determining an audio cue based on the vehicle event and a mapping of a plurality of vehicle events to a corresponding plurality of sound cues, applying a modulation to the sound cue based on the sensor data, an in-vehicle sound system, programmed to mix the modulated sound cue into a background track and play the mixed sound over in-vehicle speakers.
- the invention includes one or more computer- storage media storing computer-executable instructions that, when execute, perform a method of providing contextual enhancement of audio in a vehicle, the method comprising the steps of determining a soundscape of a plurality of available soundscapes, receiving, from a vehicle sensor, sensor data, determining a vehicle event based on the sensor data, determining a sound cue based on the vehicle event and the determined soundscape, modulating the sound cue based on the sensor data, and mixing the modulated sound cue into the soundscape for playback.
- the invention includes a method of providing contextual enhancement of audio in a vehicle comprising the steps of receiving, from a user, a selection of a soundscape of a plurality of soundscapes, receiving, from the user, an indication of a mapping from a vehicle event of a plurality of vehicle events to a sound cue of a plurality of sound cues, playing a background track of the soundscape, receiving, from a vehicle sensor, sensor data, determining based on the sensor data, that the vehicle event has occurred or is occurring, modulating the sound cue based on the sensor data, and mixing the modulated sound cue into the background track of the soundscape.
- FIG. 1 depicts an exemplary hardware platform for certain embodiments of the invention
- FIG. 2 depicts a block diagram depicting certain components of a system embodying the invention.
- FIG. 3 depicts a flowchart illustrating the operating of a method in accordance with embodiments of the invention.
- embodiments of the invention automatically sequence appropriate audio cues into the sound playback from a vehicle's sound system.
- These cues represent events related to the vehicle or taking place in the vicinity of the vehicle, such as slow speed changes, sudden braking or acceleration, location changes or a variety of other events.
- These sound cues may be represented by prerecorded sound samples of audio from a variety of scenarios and can be adjusted (for example, by altering the playback speed or pitch) based on feedback from sensors attached to the vehicle and various effects (such as, for example, echo or looping) can be added.
- sound cues correspond to vehicle events detected based on ambient conditions. These vehicle events can be user-defined or preprogrammed into the system. The determination of which sound cue is mapped to which vehicle event can also be customized by the user or selected from a set of pre-determined mappings. A variety of soundscapes, with associated sets of sound cues, may be made available to the user and the soundscape for a particular trip can be automatically determined or user-selected.
- references to “one embodiment,” “an embodiment,” or “embodiments” mean that the feature or features being referred to are included in at least one embodiment of the technology.
- references to “one embodiment” “an embodiment”, or “embodiments” in this description do not necessarily refer to the same embodiment and are also not mutually exclusive unless so stated and/or except as will be readily apparent to those skilled in the art from the description.
- a feature, structure, or act described in one embodiment may also be included in other embodiments, but is not necessarily included.
- the technology can include a variety of combinations and/or integrations of the embodiments described herein.
- Computer 102 can be a desktop computer, a laptop computer, a server computer, a mobile device such as a smartphone or tablet, or any other form factor of general- or special-purpose computing device. Depicted with computer 102 are several components, for illustrative purposes. In some embodiments, certain components may be arranged differently or absent. Additional components may also be present. Included in computer 102 is system bus 104, whereby other components of computer 102 can communicate with each other. In certain embodiments, there may be multiple busses or components may communicate with each other directly. Connected to system bus 104 is central processing unit (CPU) 106.
- CPU central processing unit
- graphics card 1 10 Also attached to system bus 104 are one or more random-access memory (RAM) modules. Also attached to system bus 104 is graphics card 1 10. In some embodiments, graphics card 104 may not be a physically separate card, but rather may be integrated into the motherboard or the CPU 106. In some embodiments, graphics card 1 10 has a separate graphics- processing unit (GPU) 1 12, which can be used for graphics processing or for general purpose computing (GPGPU). Also on graphics card 110 is GPU memory 1 14. Connected (directly or indirectly) to graphics card 1 10 is display 1 16 for user interaction. In some embodiments no display is present, while in others it is integrated into computer 102. Similarly, peripherals such as keyboard 1 18 and mouse 120 are connected to system bus 104. Like display 116, these peripherals may be integrated into computer 102 or absent. Also connected to system bus 104 is local storage 122, which may be any form of computer-readable media, and may be internally installed in computer 102 or externally and removeably attached.
- graphics card 1 10 has a separate graphics- processing unit
- Computer-readable media include both volatile and nonvolatile media, removable and nonremovable media, and contemplate media readable by a database.
- computer-readable media include (but are not limited to) RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile discs (DVD), holographic media or other optical disc storage, magnetic cassettes, magnetic tape, magnetic disk storage, and other magnetic storage devices. These technologies can store data temporarily or permanently.
- the term "computer-readable media” should not be construed to include physical, but transitory, forms of signal transmission such as radio broadcasts, electrical signals through a wire, or light pulses through a fiberoptic cable. Examples of stored information include computer-useable instructions, data structures, program modules, and other data representations.
- NIC network interface card
- NIC 124 is also attached to system bus 104 and allows computer 102 to communicate over a network such as network 126.
- NIC 124 can be any form of network interface known in the art, such as Ethernet, ATM, fiber, Bluetooth, or Wi-Fi (i.e., the IEEE 802.1 1 family of standards).
- NIC 124 connects computer 102 to local network 126, which may also include one or more other computers, such as computer 128, and network storage, such as data store 130.
- a data store such as data store 130 may be any repository from which information can be stored and retrieved as needed. Examples of data stores include relational or object oriented databases, spreadsheets, file systems, flat files, directory services such as LDAP and Active Directory, or email storage systems.
- a data store may be accessible via a complex API (such as, for example, Structured Query Language), a simple API providing only read, write and seek operations, or any level of complexity in between. Some data stores may additionally provide management functions for data sets stored therein such as backup or versioning. Data stores can be local to a single computer such as computer 128, accessible on a local network such as local network 126, or remotely accessible over Internet 132. Local network 126 is in turn connected to Internet 132, which connects many networks such as local network 126, remote network 134 or directly attached computers such as computer 136. In some embodiments, computer 102 can itself be directly connected to Internet 132.
- a complex API such as, for example, Structured Query Language
- Some data stores may additionally provide management functions for data sets stored therein such as backup or versioning.
- Data stores can be local to a single computer such as computer 128, accessible on a local network such as local network 126, or remotely accessible over Internet 132. Local network 126 is in turn connected to Internet 132, which connects many
- FIG. 2 a block diagram showing certain components of an environment suitable for practicing embodiments of the invention is depicted and referred to generally by reference numeral 200.
- user 202 is operating vehicle 204.
- Vehicle 204 is one example of a platform suitable for embodiments of the invention.
- vehicle 204 may be a car, truck, sport utility vehicle, or any other form of transportation.
- vehicle 204 may instead be an airplane, train, boat, or other method of transportation.
- any form of public or private conveyance is contemplated as being within the scope of the invention.
- Vehicle 204 may be a conventional (driver-operated) vehicle, or an autonomous (self-driving) vehicle, or a hybrid of the two (for example, a partially autonomous vehicle requiring a driver presence but only limited activity or supervision).
- Vehicle 204 has one or more occupants such as user 202.
- user 202 may be a driver or a passenger of vehicle 204.
- vehicle 204 is autonomous or partially autonomous (and therefore has no driver).
- vehicle 204 has no passengers.
- vehicle 204 has a plurality of passengers.
- embodiments of the invention can be used whenever vehicle 204 has at least one occupant of any type to serve as user 202.
- an occupant of vehicle 204 may be user 202 at a first point in time during a trip and a different occupant at a later point in time.
- an occupant may be in the group of users 202 at a first point in time, leave the group of users 202 at a second point in time, and rejoin the group of users 202 at a third point in time.
- vehicle 204 is a partially autonomous vehicle
- the attention of an occupant designated as driver may be required for some (e.g., the non-highway portion) of a drive.
- the driver can be a member of the group of users 202 during the highway portions of the drive and drop out of the group when their attention is required for driving vehicle 204.
- the term "user" will be employed herein; however, embodiments of the invention contemplate a plurality of users in addition to a single user.
- vehicle 204 includes a sound system 206 controlled by sequencer 208.
- sound system 206 plays audio entertainment (such as, for example, music) for user 202.
- sound system 206 plays an audio component of multimedia entertainment (such as, for example, movies or video games) for user 202.
- sound system 206 augments the soundscape of vehicle 204 by providing or enhancing ambient sounds (such as, for example, engine growl or wind/rain sounds).
- sequencer 208 may control multiple sound systems 206.
- a front seat sound system 206 may provide music to a first user 202, while a rear seat sound system 206 provides the audio component of a movie to another user 202.
- the same or different sound cues can be mixed into each sound system 206.
- sequencer 208 controls sound system 206.
- Sequencer 208 may receive an analog or digital audio signal from one or more other audio components, such as an AM, FM, or Satellite radio receiver, a CD or tape player, a digital audio file player, or auxiliary input jack.
- sequencer 208 may receive an audio or composite signal from a multimedia source such as a movie player or video game console.
- sequencer 208 also functions as an input multiplexer, receiving audio signals from some or all of the above- described sources and selecting one (or mixing two or more) based on user input.
- a separate switcher or mixer provides a single input to sequencer 208.
- sequencer 208 It is a function of sequencer 208 to insert sound cues into the audio output of sound system 206 in order to affect the mood, attention, or emotional state of user 202 based on environmental or user cues.
- the sound cues may also be modulated (for example, by changing the pitch, tempo, and/or volume) and effects (for example, echoes or looping) may be added.
- the sound cues added may be determined based on a user-selected scenario or soundscape, and the precise modulations and effects applied to the sound cues may be altered based on input from the various sensors described below.
- embodiments of the invention include one or more sources of data relating to current or future ambient phenomena.
- the invention includes one or more accelerometers 210 for determining current automotive motion.
- acceleration data is instead gathered via location-determining component 212 (such as a Global Positioning System receiver).
- location-determining component 212 such as a Global Positioning System receiver
- other sensors can collect data on other ambient phenomena in order to add appropriate cues.
- a light sensor could determine the current light levels to determine a simulated time-of-day for the sound cue.
- any form of sensors or transceiver is contemplated as being usable to augment the soundscape of vehicle 204 with appropriate sound cues.
- some embodiments of the invention may integrate with vehicle controls and instruments 214.
- data from vehicle controls and instruments 214 is used to provide data about upcoming movements of the vehicle over a short time horizon.
- different sound cues may be provided based on the current speed of vehicle 204.
- Embodiments of the invention can use information from the car's instruments (here, the speedometer or related component) to determine the current speed and select the appropriate sound cue.
- information on vehicle movements can be collected directly from the controls of vehicle 204. For example, if a driver of vehicle 204 steps on the accelerator, embodiments of the invention can begin to incorporate appropriate sound cues even before accelerometer 210 can detect the resultant acceleration.
- information from sources other than the vehicle itself can also be used to select appropriate sound cues.
- these information sources are also referred to herein as "sensors.”
- weather data 216 in the vicinity of vehicle 204 can be retrieved using a wireless (e.g., cellular) connection to Internet 132. If the current weather is sunny, then a beach soundscape (for example, a jet-ski or surfing scene) might be selected. By contrast, if the current weather includes rain and thunder, then a historical sailing soundscape might be selected, and if it is nighttime, then a muscle car soundscape might be selected.
- mood data for user 202 and/or one or more other occupants of vehicle 204 can be used to determine an appropriate soundscape and select sound cues.
- vehicle 204 could incorporate one or more cameras oriented so as to capture imagery of the face of user 202.
- the steering wheel of vehicle might incorporate heart rate or skin conductivity sensors. From this data, a mood for user 202 can be determined and used to select an appropriate soundscape. For example, if user 202 exhibits signs of stress and/or anger, then a soothing soundscape (for example, stand-up paddle boarding) can be selected.
- an exciting and engaging soundscape for example, jet skiing
- the volume, frequency, and/or urgency of sound cues added to the soundscape can also be affected by this (or any other) source of data.
- traffic data 220 regarding the nearby traffic situation may be used to affect the soundscape.
- congestion and accident data are broadcast over AM or FM radio.
- This data can be decoded by an appropriate receiver and used (alone or in combination with route data from a route planner) to determine that congestion will be encountered during the current trip and select a more sedate soundscape.
- nearby vehicles may form a mesh network to communicate traffic (or other pathing information) with each other.
- This data can be used to determine traffic conditions (as described above) or to communicate other relevant information.
- mood data from nearby vehicle may be transmitted and used to determine an appropriate soundscape and/or relevant sound cues.
- embodiments of the invention may incorporate anticipated and/or predicted phenomena when creating the vehicular soundscape.
- the adaptive cruise control system for a vehicle may include ultrasonic or radar sensors to detect nearby vehicles an automatically apply the brakes.
- embodiments of the invention can anticipate braking before it occurs and incorporate a sound cue corresponding to the resulting acceleration.
- vehicles may communicate (e.g., via a vehicle-to-vehicle network) with other, nearby vehicles about traffic conditions. For example, if vehicle 204 receives such a communication indicating that traffic congestion ahead will require imminent braking, then an appropriate sound cue can be incorporated even before the braking occurs.
- a soundscape and set of sound cue mappings is determined.
- the term "soundscape” refers to the sounds present in a fantasy scenario (as opposed to sounds already present in the vicinity of vehicle 204).
- a soundscape may include background noise and/or one or more specific sound cues corresponding to events in the fantasy scenario.
- a "paddle boarding" soundscape might include a background track of wind, waves, and other beach sounds.
- Sounds cues for the paddle boarding soundscape might include a water-swirl sound, a splashing sound, a cry from another beachgoer, and so forth.
- An alternative soundscape might include a motocross scenario with a background track of engine noise (either synthesized or piped in from the vehicle's actual engine), with sound cues including Doppler-shifted engine noises of other vehicles, tires squealing, and so forth.
- engine noise either synthesized or piped in from the vehicle's actual engine
- Sound cues including Doppler-shifted engine noises of other vehicles, tires squealing, and so forth.
- the soundscape may be determined based on a variety of factors.
- a beach soundscape may be selected if weather information 216 indicates that the current conditions are sunny, while a rain forest soundscape may be selected if weather information 216 indicates that the current conditions are rainy.
- any of the sensors described above can be used to determine a soundscape for a particular trip, as can any other data from sources internal or external to the vehicle.
- the soundscape may be determined based on a mood of the user 202.
- user 202 can specify a soundscape or override an automatically selected one.
- Vehicle events are discussed in greater detail below, but can broadly be any event related to the vehicle or taking place in the vicinity of the vehicle. Some or all vehicle events may have sound cues mapped to them.
- Soundscapes may include a predetermined set of mappings from vehicle events to the sound cues associated with the soundscape. For example, a "surfing" soundscape may map the sound of a surfer paddling to the vehicle event where vehicle 204 accelerates from a stop. Similarly, the vehicle event "driving normally at between 60 and 100 kilometers per hour" may be mapped to a sound cue of a rushing wave.
- User 202 may also change the mappings from vehicle events to sound cues to conform to their preferences, and new sound cues may be added to existing soundscapes.
- processing can proceed to step 304, where data regarding the current vehicle conditions is received from one or more of the sensors associated with the system.
- raw acceleration data may be received from accelerometers 210.
- the sensors may perform some preprocessing of data, such that (for example) a modulus of acceleration is received rather than a set of accelerations for each individual axis.
- speed data may be received from an onboard computer or speedometer of vehicle 204 or by preprocessing successive updates from location-determining component 212.
- step 306 the sensor data is processed to identify one or more vehicle events.
- the acceleration of the vehicle along the longitudinal axis may follow a classic "brake jerk" curve, which could be used to identify a "braking" vehicle event.
- the speed data can be binned to identify the "driving normally at between 60 and 100 kilometers per hour" vehicle event described above, and a location received from location-determining component 212 entering a particular geo-fenced region could be used to identify an "approaching destination" vehicle event.
- vehicle events that can be identified from sensor data include “gradual speed change,” “sudden braking,” “sudden acceleration,” “region change,” “sharp turn,” “other vehicle nearby” “incident ahead” and “collision imminent.”
- processing can proceed to step 308, where the vehicle event is mapped to a corresponding sound cue based on the previously determined mapping. Some vehicle events may not be mapped to a sound cue. Other vehicle events may be mapped to multiple sound cues. Certain sets of sensor data may trigger multiple vehicle events simultaneously, each with an associated sound cue. In some embodiments, user 202 can dynamically adapt the mappings of vehicle events to sound cues while the trip is underway or remove sound cues from particular vehicle events.
- step 310 processing continues at step 310, where the selected sound cue is modulated and effects added by sequencer 208.
- the pitch, tempo, volume, and other properties of the sound cue may be altered based on the context of the sound cue. For example, in a "muscle car" soundscape, the actual sound cue may be sampled from the engine compartment of vehicle 204, but pitch-shifted downward and increased in volume to achieve a more muscular sound.
- a paddling sound cue in a paddle board soundscape can be increased or decreased in tempo based on a modulus of acceleration for vehicle 204.
- short sound cues may be used and looped as long as needed (e.g., as long as the vehicle condition lasts). Other effects can also be added to the sound cue as needed to place it appropriately in the context of the soundscape.
- processing can then proceed to step 312 for mixing.
- multiple sound cues may be played at the same time.
- the soundscapes may include background tracks of sound and/or music.
- multiple channels (for example, three sound-cue channels and two background channels) may be mixed together by the sequencer as well.
- this combined channel is played over a dedicated sound system 206.
- this combined channel is instead played over the conventional sound system of vehicle 204.
- the combined channel may be further mixed with the sound being played over the conventional sound system or sequencer 208 may be integrated into the conventional sound system such that all audio is mixed in a single step.
- the audio is played over sound system 206, and processing can return to step 304, where additional sensor data is processed to identify additional vehicle events mapped to additional sound cues.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Signal Processing (AREA)
- Fittings On The Vehicle Exterior For Carrying Loads, And Devices For Holding Or Mounting Articles (AREA)
Abstract
Afin d'influer sur l'humeur d'un conducteur de véhicule (par exemple, pour augmenter son attention), des modes de réalisation de l'invention ordonnent automatiquement des indicateurs audio appropriés dans le son provenant d'un système sonore du véhicule. Ces indicateurs représentent des événements liés au véhicule ou se produisant à proximité du véhicule, et peuvent être modulés ou des effets peuvent leur être appliqués de manière à mieux informer de l'événement associé au véhicule. Ces indicateurs peuvent également faire partie d'une ambiance sonore cohérente représentant un scénario fictif, tel que faire du motocross, naviguer à la voile ou faire de la planche à rame. L'ambiance sonore tout comme la correspondance entre des événements liés au véhicule et des indicateurs sonores peuvent être configurées par l'utilisateur.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/IB2017/053672 WO2018234848A1 (fr) | 2017-06-20 | 2017-06-20 | Amplification de son à visée affective pour applications automobiles |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/IB2017/053672 WO2018234848A1 (fr) | 2017-06-20 | 2017-06-20 | Amplification de son à visée affective pour applications automobiles |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018234848A1 true WO2018234848A1 (fr) | 2018-12-27 |
Family
ID=64736931
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2017/053672 WO2018234848A1 (fr) | 2017-06-20 | 2017-06-20 | Amplification de son à visée affective pour applications automobiles |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2018234848A1 (fr) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111856960A (zh) * | 2020-07-28 | 2020-10-30 | 北京泽林汇智文创科技有限公司 | 互动声景装置 |
CN114537271A (zh) * | 2022-02-09 | 2022-05-27 | 广州小鹏汽车科技有限公司 | 控制方法、车辆和存储介质 |
CN116880498A (zh) * | 2023-07-31 | 2023-10-13 | 大陆软件系统开发中心(重庆)有限公司 | 驾驶轨迹生成方法、装置、系统、电子设备及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11288291A (ja) * | 1998-04-02 | 1999-10-19 | Sony Corp | 電気自動車 |
US20020041692A1 (en) * | 2000-10-10 | 2002-04-11 | Nissan Motor Co., Ltd. | Audio system and method of providing music |
US20030220722A1 (en) * | 2002-05-27 | 2003-11-27 | Yamaha Corporation | Vehicle capable of auditorily informing its state and method for auditorily informing state of vehicle |
US20080202323A1 (en) * | 2006-12-06 | 2008-08-28 | Yamaha Corporation | Onboard music reproduction apparatus and music information distribution system |
-
2017
- 2017-06-20 WO PCT/IB2017/053672 patent/WO2018234848A1/fr active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11288291A (ja) * | 1998-04-02 | 1999-10-19 | Sony Corp | 電気自動車 |
US20020041692A1 (en) * | 2000-10-10 | 2002-04-11 | Nissan Motor Co., Ltd. | Audio system and method of providing music |
US20030220722A1 (en) * | 2002-05-27 | 2003-11-27 | Yamaha Corporation | Vehicle capable of auditorily informing its state and method for auditorily informing state of vehicle |
US20080202323A1 (en) * | 2006-12-06 | 2008-08-28 | Yamaha Corporation | Onboard music reproduction apparatus and music information distribution system |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111856960A (zh) * | 2020-07-28 | 2020-10-30 | 北京泽林汇智文创科技有限公司 | 互动声景装置 |
CN114537271A (zh) * | 2022-02-09 | 2022-05-27 | 广州小鹏汽车科技有限公司 | 控制方法、车辆和存储介质 |
CN116880498A (zh) * | 2023-07-31 | 2023-10-13 | 大陆软件系统开发中心(重庆)有限公司 | 驾驶轨迹生成方法、装置、系统、电子设备及存储介质 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105270414B (zh) | 可选的自主驾驶模式 | |
US9395702B2 (en) | Safety critical apparatus and method for controlling distraction of an operator of a safety critical apparatus | |
US8606457B2 (en) | System and method for adaptable mobile interface | |
JP4421789B2 (ja) | 移動体用電子システムの制御装置及び制御方法、移動体用電子システム並びにコンピュータプログラム | |
US10068620B1 (en) | Affective sound augmentation for automotive applications | |
US9771082B2 (en) | Reducing cognitive demand on a vehicle operator by generating passenger stimulus | |
CN109791054A (zh) | 信息输出控制装置及信息输出控制方法 | |
GB2553620A (en) | Pedestrian warning system | |
CN103650029A (zh) | 主动声音控制 | |
CN109508403B (zh) | 车载音乐的匹配方法、装置及车载智能控制器 | |
CN114385005B (zh) | 个性化虚拟试驾装置、方法和存储介质 | |
WO2018234848A1 (fr) | Amplification de son à visée affective pour applications automobiles | |
KR100922458B1 (ko) | 차량용 악음 발생 장치, 악음 발생 방법 및 프로그램을기록한 컴퓨터로 판독가능한 기록 매체 | |
US20220091807A1 (en) | Information presentation control device | |
EP3595331A1 (fr) | Dispositif et procédé de traitement d'informations, et support d'enregistrement | |
JP2011155552A (ja) | 映像再生装置及び映像再生方法 | |
JP2018119842A (ja) | 車両用情報出力システム | |
EP4115415A1 (fr) | Dispositif électronique, procédé et programme informatique | |
US11537358B1 (en) | Sound experience generator | |
CN108454513B (zh) | 用于动态发动机声音增强的系统和方法 | |
JP2010032533A (ja) | 移動体用電子システムの制御装置及び制御方法、移動体用電子システム並びにコンピュータプログラム | |
US10560735B2 (en) | Media augmentation through automotive motion | |
JP2007261526A (ja) | 車両用制御装置、及び、情報通信システム | |
CN111791893B (zh) | 疲劳处理方法、装置、车辆及计算机可读存储介质 | |
US11499627B2 (en) | Advanced vehicle transmission control unit based on context |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17915028 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17915028 Country of ref document: EP Kind code of ref document: A1 |