EP3542545A1 - Improved audio headphones device - Google Patents
Improved audio headphones deviceInfo
- Publication number
- EP3542545A1 EP3542545A1 EP17808108.9A EP17808108A EP3542545A1 EP 3542545 A1 EP3542545 A1 EP 3542545A1 EP 17808108 A EP17808108 A EP 17808108A EP 3542545 A1 EP3542545 A1 EP 3542545A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- user
- sound
- environment
- signals
- loudspeaker
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000012545 processing Methods 0.000 claims abstract description 41
- 230000005236 sound signal Effects 0.000 claims abstract description 16
- 238000000034 method Methods 0.000 claims description 26
- 230000000694 effects Effects 0.000 claims description 21
- 238000001914 filtration Methods 0.000 claims description 13
- 238000004458 analytical method Methods 0.000 claims description 11
- 238000000926 separation method Methods 0.000 claims description 10
- 238000004590 computer program Methods 0.000 claims description 3
- 239000000203 mixture Substances 0.000 description 14
- 238000009877 rendering Methods 0.000 description 12
- 230000015572 biosynthetic process Effects 0.000 description 7
- 238000003786 synthesis reaction Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 6
- 230000000007 visual effect Effects 0.000 description 5
- 230000008447 perception Effects 0.000 description 4
- 230000002123 temporal effect Effects 0.000 description 4
- 238000012546 transfer Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 230000036651 mood Effects 0.000 description 3
- 230000035790 physiological processes and functions Effects 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 230000003595 spectral effect Effects 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 208000032041 Hearing impaired Diseases 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000003321 amplification Effects 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- SBPBAQFWLVIOKP-UHFFFAOYSA-N chlorpyrifos Chemical compound CCOP(=S)(OCC)OC1=NC(Cl)=C(Cl)C=C1Cl SBPBAQFWLVIOKP-UHFFFAOYSA-N 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 230000002996 emotional effect Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000009532 heart rate measurement Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 239000013028 medium composition Substances 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000001020 rhythmical effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 235000019640 taste Nutrition 0.000 description 1
- 230000017105 transposition Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1083—Reduction of ambient noise
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/08—Mouthpieces; Microphones; Attachments therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/10—Earpieces; Attachments therefor ; Earphones; Monophonic headphones
- H04R1/1091—Details not provided for in groups H04R1/1008 - H04R1/1083
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/20—Arrangements for obtaining desired frequency or directional characteristics
- H04R1/22—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired frequency characteristic only
- H04R1/28—Transducer mountings or enclosures modified by provision of mechanical or acoustic impedances, e.g. resonator, damping means
- H04R1/2803—Transducer mountings or enclosures modified by provision of mechanical or acoustic impedances, e.g. resonator, damping means for loudspeaker transducers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2410/00—Microphones
- H04R2410/05—Noise reduction with a separate noise microphone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2420/00—Details of connection covered by H04R, not provided for in its groups
- H04R2420/01—Input selection or mixing for amplifiers or loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2460/00—Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
- H04R2460/01—Hearing devices using active noise cancellation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/13—Aspects of volume control, not necessarily automatic, in stereophonic sound systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/01—Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]
Definitions
- the invention relates to a portable sound listening device.
- This may be an audio headset with left and right headphones, or left and right hand-held headphones.
- Noise-canceling audio headphones are known, based on a pickup by a microphone array of the user's sound environment.
- these devices seek to build, in real time, the optimal filter to minimize the contribution of the sound environment in the sound signal perceived by the user.
- a filter of the surrounding noise may be a function of the type of environment provided by the user himself, who can then select different modes of noise cancellation (office, outside, etc.).
- the "outside" mode in this case provides a reinjection of the surrounding signal (but at a much lower level than without a filter, and this so as to allow the user to remain aware of its environment).
- It can be audio headphones, configurable via a smartphone application. Speech amplification is possible in a noisy environment, where speech is usually located in front of the user.
- the methods implemented by certain hearing aids to improve the hearing-impaired user's experience propose axes of innovation such as the improvement of the spatial selectivity (according to the direction of the eyes of the user for example) .
- these different existing achievements do not allow:
- the noise-canceling headphones are based on a sound-only multichannel capture of the user's environment. They seek to reduce overall its contribution to the signal perceived by the user regardless of the nature of the environment, even if it contains potentially interesting information. These devices therefore tend to isolate the user from his environment.
- the selective headphones prototypes allow the user to configure his sound environment for example by applying equalization filters or by increasing the intelligibility of speech. These devices make it possible to improve the perception of the environment of the user but do not really make it possible to modify the broadcasted content according to the state of the user or the classes of sounds present in the environment. In this configuration, the user listening to music with a loud volume is always isolated from his environment and the need a device allowing the user to capture the relevant information in his environment is always present.
- the headphones and interactive earphones can be equipped with sensors to load and broadcast content associated with a place (as part of a tourist visit for example) or an activity (game, sports training). If some devices even have inertial or physiological sensors to monitor the activity of the user and if the dissemination of certain content may then depend on the results of the analysis of the signals from these sensors, the content broadcast does not result from an automatic generation process taking into account the analysis of the sound scene surrounding the user and do not automatically select the components of this environment relevant to the user. Furthermore, the operating modes are static, and do not automatically follow the evolution over time of the sound environment, and even less other evolutionary parameters such as a physiological state for example of the user.
- the present invention improves the situation. It proposes for this purpose a method implemented by computer means, data processing for a sound reproduction on a sound reproduction device, headset or earphones, portable by a user in an environment, the device comprising:
- the processing circuit comprising:
- an input interface for receiving signals from at least the microphone
- a processing unit for reading at least one audio content to be reproduced on the loudspeaker, and an output interface for delivering at least audio signals to be reproduced by the loudspeaker.
- processing unit is further arranged to implement the steps:
- the device comprises a plurality of microphones and the analysis of the signals from the microphones further comprises a sound source separation process in the environment applied to the signals from the microphones.
- the selected sound can be:
- the device comprises at least two loudspeakers and the reproduction of the signals on the loudspeakers applies a 3D sound effect
- a sound source position detected in the environment and emitting a selected sound, can be taken into account. to apply a sound spatialization effect of the source in the mix.
- the device may further comprise a connection to a human-machine interface available to a user for entering preferences for selecting sounds from the environment (in the general sense, as will be seen below) and the criterion The user preference is then determined by learning a history of preferences entered by the user and stored in memory.
- the device may further comprise a connection to a database of user preferences and the user preference criterion is then determined by analyzing the content of said database.
- the device may further include a connection to one or more state sensors of a user of the device, such that the user preference criterion takes into account a current state of the user, thereby contributing to a definition. of the "environment" of the user, in the general sense.
- the device may comprise a connection to a mobile terminal available to the user of the device, this terminal advantageously comprising one or more state sensors of the user.
- the processing unit may be further arranged to select a content to be read from among a plurality of contents, depending on the state of the user.
- the predetermined target sound classes may include at least speech sounds whose voiceprints are pre-recorded.
- step a) may optionally include at least one of the following operations:
- 'interest for the user of the device extracting parameters specific to these sources of interest with a view to a subsequent rendering of the sounds picked up and coming from these sources of interest in a spatialized audio mix;
- a classification system for example by deep neural networks
- known sound classes speech, music, noise, etc.
- possible identification by d other techniques for classifying the sound stage for example, sound recognition of a desk, an outdoor street, transportation, etc.
- step c) may optionally include at least one of the following operations:
- temporal filtering for example Wiener filtering, and / or Duet algorithm
- spectral filtering and / or spatial filtering for example Wiener filtering, and / or Duet algorithm
- HRTF Head Related Transfer Functions
- the present invention also relates to a computer program comprising instructions for implementing the above method, when this program is executed by a processor.
- the invention also relates to a sound reproduction device, of the headset or earphone type, portable by a user in an environment, the device comprising:
- the processing circuit comprising:
- an input interface for receiving signals from at least the microphone, a processing unit for reading at least one audio content to be reproduced on the loudspeaker, and
- an output interface for delivering at least audio signals to be reproduced by the loudspeaker.
- the processing unit is further arranged for:
- the invention thus proposes a system including an intelligent audio device, integrating for example a network of sensors, at least one loudspeaker and a terminal (e.g. smartphoned).
- a terminal e.g. smartphoned.
- the originality of this system is to be able to automatically generate, in real time, the "optimal soundtrack" of the user, that is to say the multimedia content best suited to its environment and its state clean.
- the user's own state can be defined by: i) a set of preferences (type of music, sound classes of interest, etc.); (ii) his activity (rest, office, sports training, etc.); iii) its physiological states (stress, fatigue, stress, etc.) and / or socio-emotional states (personality, mood, emotions, etc.).
- the multimedia content generated may comprise a main audio content (to be broadcast in the headset) and possibly secondary multimedia contents (texts, images, video) that can be broadcast via the smartphoned terminal.
- the different content elements include both the elements of the user's content base (music, video, etc., hosted on the terminal or in the cloud), the result of captures made by a sensor network that includes the system and synthetic elements generated by the system (notifications, "jingles” sound or text, noise comfort, etc.).
- the system can automatically analyze the user's environment and predict the components potentially of interest to the user in order to restore them in an increased and controlled manner, superimposing them optimally on the contents consumed by the user (typical the music he listens to).
- state-of-the-art devices do not make it possible to automatically identify each class of sound present in the user's environment in order to associate with each of them a treatment that meets the expectations of the user. the user (for example a highlighting of a sound, or on the contrary a reduction, the generation of an alert), according to its identification in the environment.
- the state of the art does not use analysis of the sound stage, nor the state of the user or his activity to calculate the sound reproduction.
- FIG. 1 illustrates a device according to the invention, in a first embodiment
- FIG. 2 illustrates a device according to the invention, in a second embodiment, here connected to a mobile terminal
- FIG. 3 illustrates the steps of a method according to one embodiment of the invention
- FIG. 4 specifies steps of the method of FIG. 3, according to a particular embodiment.
- a sound reproduction device DIS (of the headset or earpiece type), worn for example by a user in an environment ENV, comprises at least:
- one (or two, in the example represented) speakers HP at least one sensor, for example a microphone MIC (or a row of microphones in the example shown to capture a directivity sounds from the environment) and - a connection to a processing circuit.
- a microphone MIC or a row of microphones in the example shown to capture a directivity sounds from the environment
- the processing circuit can be integrated directly into the headphones and housed in a loudspeaker enclosure (as illustrated in FIG. 1), or can, in the variant illustrated in FIG. 2, be implemented in a TER terminal of the user, by example a smartphone-type mobile terminal, or be distributed between several terminals of the user (a smartphone, and a connected object possibly including other sensors).
- the connection between the headset (or headsets) and the dedicated processing circuit of the terminal is performed by a USB connection or short-range radio frequency (for example by Bluetooth or other) and the headset (or headsets) is equipped with a transmitter / receiver BT1, communicating with a transceiver BT2 that includes the terminal TER.
- a hybrid solution in which the processing circuit is distributed between the speaker of the headset and a terminal is also possible.
- the processing circuit comprises:
- an input interface IN for receiving signals coming from at least the microphone MIC, a processing unit typically comprising a processor PROC and a memory MEM, for interpreting, with respect to the environment ENV, the signals coming from the microphone by learning (for example by classification, or by "matching" type "finger printing” for example),
- an output interface OUT for delivering at least audio signals that are functions of the environment and to be reproduced by the loudspeaker.
- the memory MEM can store instructions of a computer program within the meaning of the present invention, and possibly temporary data (calculation or otherwise), as well as durable data, such as the user's preferences, or even model definition data or other, as will be discussed later.
- the input interface IN is, in a sophisticated embodiment, connected to a network of microphones, as well as to an inertial sensor (provided on the headphones or in the terminal) and the definition of user preferences.
- the user preference data may be stored locally in the MEM memory, as indicated above. As a variant, they can be stored, possibly with other data, in a remote database DB accessible via a communication via a local or extended NW network.
- An LP communication module with such a network suitable for this purpose can be provided in the headset or in the TER terminal.
- a man / machine interface can allow the user to define and update his preferences.
- the man / machine interface can simply correspond to a touch screen of the smartphone TER for example. Otherwise, it can be provided such an interface directly on the headset.
- additional sensors in the TER terminal to enrich the definition of the environment of the user, in the general sense.
- additional sensors may be physiological sensors specific to the user (measurement of electroencephalogram, heart rate measurement, pedometer, etc.) or any other sensor to improve the knowledge of the environment / current state of the user.
- this definition can directly include the notification by the user himself of his activity, his own condition and his environment.
- the definition of the environment can take into account:
- Metadata for example the genre, the listening occurrences per piece
- metadata may also be associated; - Moreover, the browsing history and applications of his smartphone; the history of its consumption of streaming content (via a service provider) or locally;
- the input interface may, in the general sense, be connected to a set of sensors, and also include connection modules (including the LP interface) for characterizing the user's environment, but also habits and preferences (history of content consumption, streaming activities and / or social networks).
- connection modules including the LP interface
- habits and preferences history of content consumption, streaming activities and / or social networks.
- multimedia output is implemented by automatic extraction, via signal processing and artificial intelligence modules, in particular machine learning modules (represented by step S7 in FIG. important for creating the output media stream.
- PI, P2, ... in the figures can typically be environment parameters which must be taken into account for the reproduction on loudspeakers. For example, if a sound picked up in the environment is identified as a speech signal to be played back:
- a first set of parameters may be coefficients of an optimal filter (Wiener filter type) making it possible to enhance the speech signal to increase its intelligibility;
- a second parameter is the directivity of the sound captured in the environment and to be rendered for example by means of a binaural rendering (rendering technique using transfer functions of HRTF type);
- the processing unit requests the input interface to collect the signals from the microphone or microphone array MIC that the DIS carries.
- sensors in the terminal TER in step S2, or elsewhere in step S3 (connected heart rate sensors, EEG, etc.), can communicate their signals to the processing unit.
- information data other than captured signals can be transmitted. by the memory MEM and / or by the database BD to the processing unit.
- step S4 all these data and signals specific to the environment and the state of the user (hereinafter referred to generically as "environment”) are collected and interpreted by the implementation, at the step S7, a computer module for decoding the environment by artificial intelligence.
- this decoding module can use a learning base which can, for example, be remote and requested in step S8 via the network NW (and the communication interface LP), in order to extract parameters relevant PI, P2, P3, at step S9 which model the environment in general.
- the sound scene to be rendered is generated in step S 10 and transmitted in the form of audio signals to the loudspeakers HP at step SU.
- This sound scene may possibly be accompanied by graphic information, for example metadata, to be displayed on the screen of the terminal TER in step S12.
- an analysis of the environmental signals is carried out, with: an identification of the environment with a view to estimating prediction models making it possible to characterize the user's environment and his or her own state (these models being used with a recommendation engine as will be seen below with reference to the figure 4), and
- a fine acoustic analysis to generate more precise parameters and used to manipulate the audio content to be restored (separation / enhancement of particular sound sources, sound effects, mixing, spatialization, or other).
- the identification of the environment makes it possible to characterize, by automatic learning, the environment / state pair of the user. It is mainly:
- the classes of its target may be defined, one by one, by the user via his terminal or by using predefined operating modes;
- the fine acoustic analysis makes it possible to calculate the acoustic parameters which are used for the audio reproduction (for example in 3D restitution).
- a recommendation engine is used to receive the descriptors of the "environment", in particular the classes of identified sound events (parameters PI, P2, etc.). , and provide on this basis a recommendation model (or a combination of models) at step S 19.
- the recommendation engine may use the characterization of the user's contents and their similarity to external contents as well as user preferences, which have been recorded in a learning base in step S 15, and / or the standard preferences of other users in step S18.
- the user can also intervene at this step with his terminal to enter a preference at step S24, for example with respect to a content or a list of contents. to play.
- a recommendation model is chosen that is relevant to the environment and the state of the user (for example, in the group of rhythmic music, in a situation of movement of the apparently user in a gym).
- a composition engine is then implemented in step S20, which combines the parameters PI, P2 ..., with the recommendation model, to develop a composition program in step S21. This is a routine that suggests for example:
- composition engine To mix to the content, according to a sound level and a spatial rendering (3D audio) that has been defined by the composition engine.
- the synthesis engine strictly speaking, of the audible signal intervenes in step S22, to elaborate the signals to be restored to the steps SI 1 and S 12, from:
- step S25 (as a substep of step S6), of course, one of the contents having been selected in step S21 by the composition engine,
- the stream generated is adapted to the expectations of the user and optimized according to the context of its distribution, according to three main steps in a particular embodiment:
- the generated multimedia stream includes at least audio signals but potentially textual, haptic and / or visual notifications.
- the audio signals include a mix:
- a content selected in the user's content database music, video, etc.
- entered as a preference by the user in step S24 or recommended directly by the recommendation engine according to the state of the user and the environment
- sounds picked up by the MIC sensor network selected in the sound environment (thus filtered), enhanced (for example by source separation techniques) and processed so that they are of frequency texture, intensity and spatialisation, adjusted to be injected into the mix in a timely fashion, and
- the recommendation engine is based on:
- the models are updated continuously over time to adapt to the evolution of the user.
- composition engine plans The composition engine plans:
- the time at which each piece of content must be played including the order in which the user's contents are presented (for example, the order of the music in a playlist), and the external moments or sounds or the notifications are broadcast: in real time or delayed (for example between two pieces of a playlist) so as not to disturb the listening or the current activity of the user at an inconvenient time;
- Planning is based on templates and rules built from decoding the user's environment and their own state. For example, the spatial position of a sound event captured by the pickups and the gain level associated with it depend on the result of the sound source localization detection performed by the decoding of the environment in step S7 of the figure 3.
- the synthesis engine is based on signal processing techniques, natural languages and images, respectively for the synthesis of audio, textual and visual outputs (images or videos), and jointly for the generation of multimedia outputs, for example video.
- the synthesis is first performed locally on short time windows and the signal is reconstructed by addition-overlap before being transmitted to at least two speakers (one for each ear).
- Different gains (power levels) and audio effects are applied to the different content elements as provided by the composition engine.
- the processing applied by windows may include a filtering (for example of Wiener) making it possible to enhance, from one or more of the audio streams captured, a particular sound source (as provided by the composition engine).
- the processing may include 3D audio rendering, possibly using HRTF filtering techniques (HRTF transfer functions for "Head Related Transfer Functions").
- the description of the environment of the user is limited to his sound environment; the user's own state is limited to his preferences: class of his target, notifications he wishes to receive, these preferences being defined by the user using his terminal; the device (possibly in cooperation with the terminal) is equipped with inertial sensors (accelerometer, gyroscope and magnetometer); playback parameters are automatically changed when a target sound class is detected in the user's environment; - messages of short duration can be recorded; notifications can be sent to the user to warn him of the detection of an event of interest.
- inertial sensors accelerelerometer, gyroscope and magnetometer
- the captured signals are analyzed to determine:
- a set of sensors including a network of microphones, a video camera, a pedometer, inertial sensors (accelerometers, gyroscopes, magnetometers), physiological sensors can capture the visual and sound environment of the user (microphones and camera), the data characterizing his movement (inertial sensors, pedometer) and his physiological parameters (EEG, ECG, EMG, electrodermal) as well as all the contents he is consulting (music , radio broadcasts, videos, browsing history and applications of his smartphone).
- inertial sensors accelerometers, gyroscopes, magnetometers
- physiological sensors can capture the visual and sound environment of the user (microphones and camera), the data characterizing his movement (inertial sensors, pedometer) and his physiological parameters (EEG, ECG, EMG, electrodermal) as well as all the contents he is consulting (music , radio broadcasts, videos, browsing history and applications of his smartphone).
- a musical stream adapted to the environment and to the user's own state can be generated (for example a playlist of which each piece is selected according to its musical tastes, its stride and its state of fatigue). While all sound sources are canceled in the user's headset, the voice of a coach ("coach") near the user, when it is identified (voice record previously recorded), is mixed at stream and rendered spatially using binaural rendering techniques (by HRTF for example).
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FR1661324A FR3059191B1 (en) | 2016-11-21 | 2016-11-21 | PERFECTLY AUDIO HELMET DEVICE |
PCT/FR2017/053183 WO2018091856A1 (en) | 2016-11-21 | 2017-11-20 | Improved audio headphones device |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3542545A1 true EP3542545A1 (en) | 2019-09-25 |
Family
ID=58347514
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP17808108.9A Withdrawn EP3542545A1 (en) | 2016-11-21 | 2017-11-20 | Improved audio headphones device |
Country Status (5)
Country | Link |
---|---|
US (1) | US20200186912A1 (en) |
EP (1) | EP3542545A1 (en) |
FR (1) | FR3059191B1 (en) |
TW (1) | TW201820315A (en) |
WO (1) | WO2018091856A1 (en) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10361673B1 (en) * | 2018-07-24 | 2019-07-23 | Sony Interactive Entertainment Inc. | Ambient sound activated headphone |
TWI671738B (en) * | 2018-10-04 | 2019-09-11 | 塞席爾商元鼎音訊股份有限公司 | Sound playback device and reducing noise method thereof |
US10575094B1 (en) * | 2018-12-13 | 2020-02-25 | Dts, Inc. | Combination of immersive and binaural sound |
US11221820B2 (en) * | 2019-03-20 | 2022-01-11 | Creative Technology Ltd | System and method for processing audio between multiple audio spaces |
US11252497B2 (en) * | 2019-08-09 | 2022-02-15 | Nanjing Zgmicro Company Limited | Headphones providing fully natural interfaces |
TWI731472B (en) * | 2019-11-14 | 2021-06-21 | 宏碁股份有限公司 | Electronic device and automatic adjustment method for volume |
TWI740374B (en) * | 2020-02-12 | 2021-09-21 | 宏碁股份有限公司 | Method for eliminating specific object voice and ear-wearing audio device using same |
CN113347519B (en) * | 2020-02-18 | 2022-06-17 | 宏碁股份有限公司 | Method for eliminating specific object voice and ear-wearing type sound signal device using same |
TWI768589B (en) * | 2020-12-10 | 2022-06-21 | 國立勤益科技大學 | Deep learning rhythm practice system |
US11307825B1 (en) * | 2021-02-28 | 2022-04-19 | International Business Machines Corporation | Recording a separated sound from a sound stream mixture on a personal device |
US20220312126A1 (en) * | 2021-03-23 | 2022-09-29 | Sonova Ag | Detecting Hair Interference for a Hearing Device |
CN113301466A (en) * | 2021-04-29 | 2021-08-24 | 南昌大学 | Adjustable active noise reduction earphone with built-in noise monitoring device |
CN114067832A (en) * | 2021-11-11 | 2022-02-18 | 中国科学院声学研究所 | Head-related transfer function prediction method and device and electronic equipment |
WO2024010501A1 (en) * | 2022-07-05 | 2024-01-11 | Telefonaktiebolaget Lm Ericsson (Publ) | Adjusting an audio experience for a user |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8155334B2 (en) * | 2009-04-28 | 2012-04-10 | Bose Corporation | Feedforward-based ANR talk-through |
FR2983605A1 (en) * | 2011-12-05 | 2013-06-07 | France Telecom | DEVICE AND METHOD FOR SELECTING AND UPDATING USER PROFILE. |
US10038952B2 (en) * | 2014-02-04 | 2018-07-31 | Steelcase Inc. | Sound management systems for improving workplace efficiency |
US9344793B2 (en) * | 2013-02-11 | 2016-05-17 | Symphonic Audio Technologies Corp. | Audio apparatus and methods |
US9508335B2 (en) * | 2014-12-05 | 2016-11-29 | Stages Pcs, Llc | Active noise control and customized audio system |
-
2016
- 2016-11-21 FR FR1661324A patent/FR3059191B1/en active Active
-
2017
- 2017-11-20 EP EP17808108.9A patent/EP3542545A1/en not_active Withdrawn
- 2017-11-20 WO PCT/FR2017/053183 patent/WO2018091856A1/en active Application Filing
- 2017-11-20 US US16/462,691 patent/US20200186912A1/en not_active Abandoned
- 2017-11-21 TW TW106140244A patent/TW201820315A/en unknown
Also Published As
Publication number | Publication date |
---|---|
TW201820315A (en) | 2018-06-01 |
WO2018091856A1 (en) | 2018-05-24 |
US20200186912A1 (en) | 2020-06-11 |
FR3059191A1 (en) | 2018-05-25 |
FR3059191B1 (en) | 2019-08-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
FR3059191B1 (en) | PERFECTLY AUDIO HELMET DEVICE | |
US10187740B2 (en) | Producing headphone driver signals in a digital audio signal processing binaural rendering environment | |
US9959783B2 (en) | Converting audio to haptic feedback in an electronic device | |
KR102436168B1 (en) | Systems and methods for creating listening logs and music libraries | |
CN105637903B (en) | System and method for generating sound | |
KR20180137490A (en) | Personal emotion-based computer-readable cognitive memory and cognitive insights for memory and decision making | |
JP7167910B2 (en) | Information processing device, information processing method, and program | |
CN106488311B (en) | Sound effect adjusting method and user terminal | |
McGill et al. | Acoustic transparency and the changing soundscape of auditory mixed reality | |
US10466955B1 (en) | Crowdsourced audio normalization for presenting media content | |
EP2920979B1 (en) | Acquisition of spatialised sound data | |
WO2017162980A1 (en) | Method and device for controlling the setting of at least one audio and/or video parameter, corresponding terminal and computer program | |
CN110493635A (en) | Video broadcasting method, device and terminal | |
WO2005017892A2 (en) | Method for reproducing audio documents with the aid of an interface comprising document groups and associated reproducing device | |
WO2021129444A1 (en) | File clustering method and apparatus, and storage medium and electronic device | |
CN113039815B (en) | Sound generating method and device for executing the same | |
EP3777247B1 (en) | Method and system for broadcasting a multichannel audio stream to terminals of spectators attending a sporting event | |
US20230267942A1 (en) | Audio-visual hearing aid | |
EP2206236A1 (en) | Audio or audio-video player including means for acquiring an external audio signal | |
WO2022178852A1 (en) | Listening assisting method and apparatus | |
JP2016131329A (en) | Image and sound recording device, image and sound recording method, and image and sound recording program | |
CN115767407A (en) | Sound generating method and device for executing the same | |
FR2921746A1 (en) | Portable musical signal listening device e.g. MPEG-1 audio layer 3 walkman, for e.g. car, has transferring stage transferring external audio signal to musical signal listening unit, and processor applying processing function to audio signal | |
FR2921747A1 (en) | Portable audio signal i.e. music, listening device e.g. MPEG-1 audio layer 3 walkman, for e.g. coach, has analyzing and transferring unit transferring external audio signal that informs monitoring of sound event to user, to listening unit |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20190515 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20210205 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
RAP3 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: INSTITUT MINES TELECOM Owner name: BLOUET, RAPHAEL |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20221022 |