WO2020096406A1 - Procédé de génération de son et dispositifs réalisant ledit procédé - Google Patents

Procédé de génération de son et dispositifs réalisant ledit procédé Download PDF

Info

Publication number
WO2020096406A1
WO2020096406A1 PCT/KR2019/015145 KR2019015145W WO2020096406A1 WO 2020096406 A1 WO2020096406 A1 WO 2020096406A1 KR 2019015145 W KR2019015145 W KR 2019015145W WO 2020096406 A1 WO2020096406 A1 WO 2020096406A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
real
sounds
processor
virtual
Prior art date
Application number
PCT/KR2019/015145
Other languages
English (en)
Korean (ko)
Inventor
배영식
Original Assignee
주식회사 후본
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020190140346A external-priority patent/KR102379734B1/ko
Priority claimed from KR1020190140347A external-priority patent/KR102322120B1/ko
Application filed by 주식회사 후본 filed Critical 주식회사 후본
Priority to CN202211375358.3A priority Critical patent/CN115767407A/zh
Priority to US17/292,447 priority patent/US20220021998A1/en
Priority to CN201980073791.1A priority patent/CN113039815B/zh
Publication of WO2020096406A1 publication Critical patent/WO2020096406A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/10Earpieces; Attachments therefor ; Earphones; Monophonic headphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S3/00Systems employing more than two channels, e.g. quadraphonic
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S5/00Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control

Definitions

  • the embodiments below relate to a sound generating method and apparatuses performing the same.
  • 3D audio-only contents can be provided to a user.
  • separately recorded 3D sound may be output to provide 3D audio-only content to a user.
  • the content for exclusive use of 3D audio may be various content generated by using 3D sound recorded separately as described above as 3D sound. For example, recently, a variety of 3D audio-only contents are generated by performing directionality and automation calculation on a general 2D sound (or sound).
  • the content for exclusive use of 3D audio may be sound content to which general 3D sound conversion and output technology is applied.
  • 3D audio-only content has the advantage of reproducing high immersion and realism than existing 3D sound content.
  • Embodiments may provide a technique for combining a real sound generated in a real space and a virtual sound generated in a virtual space to generate a combined sound generated in a mixed reality in which the real space and the virtual space are mixed.
  • embodiments may provide a technique for converting each of a plurality of 2D object sounds of a 2D sound track into a 3D object sound to generate a 3D sound track reflecting the 3D object sound.
  • the sound generation method may include acquiring a real sound generated in a real space and a play sound generated in a virtual space, and combining the real sound and the play sound to mix the real space and the virtual space And generating a combination sound generated in mixed reality.
  • the generating step includes selecting at least one real object sound among a plurality of real object sounds included in the real sound, and generating at least one virtual object sound among a plurality of virtual object sounds included in the play sound.
  • the method may include selecting and combining the at least one real object sound and the at least one virtual object sound to generate the combined sound.
  • the plurality of real object sounds may be sounds generated from a plurality of real objects located in the real space.
  • the plurality of virtual object sounds may be sounds generated from a plurality of virtual objects located in the virtual space.
  • the step of selecting the at least one real object sound may include recognizing the plurality of real object sounds based on characteristics of the object sound, and the at least one of the plurality of real object sounds based on a real sound selection condition. And selectively extracting a real object sound.
  • the recognizing step includes removing the noise sound from the real sound based on a noise filtering technique, and the plurality of real objects from the real sound from which the noise sound is removed based on at least one of a frequency and a volume of the object sound. And recognizing sounds.
  • the generating may include adjusting the volume of the at least one real object sound based on the position of the real object corresponding to the at least one real object sound, and adjusting the volume of the at least one real object sound and the at least And combining one virtual object sound to generate the combined sound.
  • the adjusting may include determining a position of the real object in the real space based on a sound acquisition time of the at least one real object sound, and based on a separation distance between the user's position and the position of the real object. And adjusting the volume of the at least one real object sound.
  • An apparatus includes a memory including instructions, and a processor for executing the instructions, wherein the processor acquires real sound generated in real space and play sound generated in virtual space, and the real sound And combining the play sounds to generate a combination sound generated in a mixed reality in which the real space and the virtual space are mixed.
  • the processor selects at least one real object sound from among a plurality of real object sounds included in the real sound, selects at least one virtual object sound from among a plurality of virtual object sounds included in the play sound, and the The combination sound may be generated by combining at least one real object sound and the at least one virtual object sound.
  • the plurality of real object sounds may be sounds generated from a plurality of real objects located in the real space.
  • the plurality of virtual object sounds may be sounds generated from a plurality of virtual objects located in the virtual space.
  • the processor may recognize the plurality of real object sounds based on the characteristics of the object sound, and selectively extract the at least one real object sound among the plurality of real object sounds based on a real sound selection condition.
  • the processor removes the noise sound from the real sound based on a noise filtering technique, and recognizes the plurality of real object sounds from the real sound from which the noise sound is removed based on at least one of the frequency and volume of the object sound. Can be.
  • the processor adjusts the volume of the at least one real object sound based on the position of the real object corresponding to the at least one real object sound, and adjusts the volume of the at least one real object sound and the at least one virtual object Sounds can be combined to generate the combined sound.
  • the processor determines the location of the real object in the real space based on the sound acquisition time of the at least one real object sound, and based on the separation distance between the user's location and the location of the real object You can adjust the volume of the actual object sound.
  • a sound generating method includes extracting a plurality of 2D object sounds included in a 2D sound track, and applying a plurality of binaural effects to each of the plurality of 2D object sounds to apply the plurality of sounds. And converting the two-dimensional object sounds into a plurality of three-dimensional object sounds, and generating a three-dimensional sound track based on the plurality of three-dimensional object sounds.
  • the plurality of 2D object sounds may be sounds separated by any one of frequency and object in the 2D sound track.
  • the extracting may include extracting the plurality of two-dimensional object sounds by separating the two-dimensional sound track for each frequency band using an equalizer effect.
  • the extracting may include extracting the plurality of two-dimensional object sounds by separating the two-dimensional sound track for each object using sound detecting.
  • the converting may include generating a first 3D object sound by applying a first binaural effect to a first 2D object sound among the plurality of 2D object sounds, and the plurality of 2D object sounds
  • the method may include generating a second 3D object sound by applying a second binaural effect to the second 2D object sound.
  • the first binaural effect and the second binaural effect may be different from each other or the same binaural effect.
  • Generating the first three-dimensional object sound may include determining a first three-dimensional positioning for the first two-dimensional object sound, and the first three-dimensional positioning on the first two-dimensional object sound. And generating the first 3D object sound by applying the first binaural effect.
  • the generating of the second three-dimensional object sound may include determining a second three-dimensional positioning for the second two-dimensional object sound differently from the first three-dimensional positioning, and the second two And generating the second three-dimensional object sound by applying the second three-dimensional positioning and the second binaural effect to the dimensional object sound.
  • the generating may include generating the 3D sound track by integrating the plurality of 3D object sounds.
  • a sound generating apparatus includes a memory including instructions, and a processor for executing the instructions, wherein the processor extracts a plurality of two-dimensional object sounds included in a two-dimensional sound track, and the plurality of Each of the two-dimensional object sounds is applied to each of a plurality of binaural effects to convert the plurality of two-dimensional object sounds into a plurality of three-dimensional object sounds, and a three-dimensional sound based on the plurality of three-dimensional object sounds You can create tracks.
  • the plurality of 2D object sounds may be sounds separated by any one of frequency and object in the 2D sound track.
  • the processor may extract the plurality of two-dimensional object sounds by separating the two-dimensional sound track for each frequency band using an equalizer effect.
  • the processor may extract the plurality of two-dimensional object sounds by separating the two-dimensional sound track for each object using sound detecting.
  • the processor generates a first 3D object sound by applying a first binaural effect to a first 2D object sound among the plurality of 2D object sounds, and a second of the plurality of 2D object sounds
  • a second 3D object sound may be generated by applying a second binaural effect to the 2D object sound of.
  • the first binaural effect and the second binaural effect may be different from each other or the same binaural effect.
  • the processor determines the first three-dimensional positioning of the first two-dimensional object sound, and applies the first three-dimensional positioning and the first binaural effect to the first two-dimensional object sound to produce the first three-dimensional object sound. You can create a 3D object sound of 1.
  • the processor determines a second three-dimensional positioning for the second two-dimensional object sound differently from the first three-dimensional positioning, and the second three-dimensional positioning and the second two-dimensional object sound.
  • the second 3D object sound may be generated by applying the second binaural effect.
  • the processor may generate the 3D sound track by integrating the plurality of 3D object sounds.
  • FIG. 1 shows a schematic block diagram of a sound generation system according to an embodiment.
  • FIG. 2 shows a schematic block diagram of the sound generating apparatus shown in FIG. 1.
  • FIG. 3 shows an example for describing the sound providing apparatus illustrated in FIG. 1.
  • FIG. 4 shows an example for describing the sound providing apparatus illustrated in FIG. 1 or an example for describing the first providing apparatus illustrated in FIG. 3.
  • FIG. 5 shows an example for describing the sound generating apparatus illustrated in FIG. 1 or an example for describing the second providing apparatus illustrated in FIG. 3.
  • FIG. 6 shows an example for describing the sound output device illustrated in FIG. 1.
  • FIG. 7 shows another example for describing the sound output device shown in FIG. 1.
  • FIG. 8 shows an example for describing a sound output device that is an in-ear type earphone.
  • FIG 9 shows another example for describing a sound output device that is an in-ear type earphone.
  • FIG. 10 shows an example for explaining a combination sound according to an embodiment.
  • FIG. 11 is a flowchart illustrating an operation of the sound generating apparatus shown in FIG. 1.
  • FIG. 12 shows a sound generating system according to another embodiment.
  • FIG. 13 shows an example for explaining the operation of the sound generating apparatus shown in FIG. 12.
  • FIG. 14 is a flowchart for describing the operation of the processor illustrated in FIG. 13.
  • first or second may be used to describe various components, but the components should not be limited by the terms. The terms are for the purpose of distinguishing one component from another component, for example, without departing from the scope of rights according to the concept of the embodiment, the first component may be referred to as the second component, and similarly The second component may also be referred to as the first component.
  • a module in the present specification may mean hardware capable of performing functions and operations according to each name described in the present specification, or computer program code capable of performing specific functions and operations. Or, it may mean an electronic recording medium on which computer program code capable of performing a specific function and operation is mounted, for example, a processor or a microprocessor.
  • the module may mean a functional and / or structural combination of hardware for performing the technical idea of the present invention and / or software for driving the hardware.
  • FIG. 1 shows a schematic block diagram of a sound generation system according to an embodiment.
  • the sound generating system 10 includes a sound providing device 100 and a sound generating device 300.
  • the sound providing device 100 may generate (or record) a play sound to be provided to a user (or a listener), and then provide the play sound to the sound generating device 300.
  • the play sound may be various, such as a 3D sound source and a 3D virtual reality (VR) sound content.
  • the play sound may be a sound generated in virtual space.
  • the virtual space may be a 3D virtual space (or 3D virtual reality) implemented to provide a 3D sound reflecting a sense of space and realism.
  • the sound providing device 100 may provide the 2D sound track to the sound generating device 300.
  • the two-dimensional sound track is a stereo or mono type unidirectional sound track, which may be audible to a listener.
  • the 2D sound track may be various, such as a 2D sound source, a 2D voice, and a 2D VR (virtual reality) sound.
  • the two-dimensional sound track may include a plurality of object sounds.
  • Each of the plurality of object sounds is a two-dimensional sound, and may be an object sound generated from each of the plurality of objects.
  • the sound generating device 300 combines a real sound generated in a real space with a virtual sound generated in a virtual space to generate a combined sound (or mixed reality sound) generated in a mixed reality in which the real space and the virtual space are mixed. Can be created.
  • the sound generating device 300 may provide a sound with a high immersion feeling that allows a user to recognize that the user is located in a mixed space in which a real space and a virtual space are mixed.
  • the sound generating apparatus 300 may provide a personalized (or personalized) three-dimensional sound to the user by providing various three-dimensional sounds by selectively combining real and virtual sounds.
  • the sound generating apparatus 300 may provide the user with the real sound and the sound generated by mixing the real sound and the play sound without completely removing the sound, and secure the user's safety.
  • the sound generating apparatus 300 may convert each of a plurality of 2D object sounds of the 2D sound track into a 3D object sound to generate a 3D sound track reflecting the 3D object sound.
  • the sound generating apparatus 300 may provide various immersive 3D sounds (or 3D contents) in various forms by reflecting 3D directionality according to 3D effects to each of the 2D object sounds.
  • the sound generating apparatus 300 can easily generate a 3D sound track by generating a 3D sound track using only the 2D sound track.
  • the sound generating device 300 may generate a 3D sound track that can be used in tinnitus treatment or tinnitus diagnosis by reproducing the direction of a 3D sound track that is a characteristic of 3D sound (or 3D audio).
  • the sound generating apparatus 300 may generate a 3D sound track that can be used for practical tinnitus treatment and tinnitus diagnosis by reflecting the position reproduction according to the frequency band in the 3D sound track.
  • the sound output device 500 may acquire a real sound generated in real space.
  • the actual space may be a space where the user is located to listen to the combined sound.
  • the sound output device 500 may detect (or sense, acquire) the user's head direction (or the user's point of view, the user's gaze) by trekking (or head tracking) the user's head. Can be.
  • the sound output device 500 may acquire 3D real sound according to the direction of the user's head through the microphones of the sound output devices 500 located in both directions of the user. Both directions may be a right ear direction and a left ear direction in a direction corresponding to both ears of the user.
  • the sound output device 500 may transmit the actual sound, sound acquisition information about the real sound, and / or user's head direction information to the sound generating device 300.
  • the sound output device 500 may receive (or provide) the combined sound transmitted from the sound generating device 300 to the user.
  • the user can listen to the combined sound generated in the mixed reality in which the real space and the virtual space are mixed through the sound output device 500.
  • the sound providing device 100, the sound generating device 300, and the sound output device 500 are configured to be independently distinguished, but are not limited thereto.
  • the sound providing device 100 may be implemented by being included in the sound generating device 300
  • the sound generating device 300 may be implemented by being included in the sound output device 500.
  • FIG. 2 shows a schematic block diagram of the sound generating apparatus shown in FIG. 1.
  • the sound generating device 300 may include a communication module 310, a memory 330 and a processor 350.
  • the communication module 310 may receive the 2D sound track or play sound transmitted from the sound providing device 100 and transmit it to the processor 350.
  • the communication module 310 may receive and transmit actual sound transmitted from the sound output device 100, sound acquisition time information for the real sound, and / or information about the direction of the user's head to the processor 350.
  • the communication module 310 may receive the combined sound transmitted from the processor 350 and transmit it to the sound output device 100.
  • the memory 330 may store instructions (or programs) executable by the processor 350.
  • instructions may include instructions for performing the operation of processor 350 and / or the operation of each component of processor 350.
  • the processor 350 may process data stored in the memory 330.
  • the processor 350 may execute computer readable code (eg, software) stored in the memory 330 and instructions caused by the processor 350.
  • the processor 350 may be a data processing device embodied in hardware having circuits having a physical structure for performing desired operations.
  • desired operations may include code or instructions included in a program.
  • the data processing device implemented in hardware includes a microprocessor, a central processing unit, a processor core, a multi-core processor, and a multiprocessor.
  • a microprocessor e.g., a central processing unit, a processor core, a multi-core processor, and a multiprocessor.
  • ASIC Application-Specific Integrated Circuit
  • FPGA Field Programmable Gate Array
  • the processor 350 may control the overall operation of the sound generating device 300.
  • the processor 350 may control the operation of each component 310 and 330 of the sound generating device 300.
  • the processor 350 may acquire real sound transmitted from the sound output device 500, sound acquisition time information for the real sound, and / or information about the direction of the user's head.
  • the real sound may include a plurality of real object sounds.
  • Each of the plurality of real object sounds may be a sound generated from each of the plurality of real objects located in the real space.
  • the sound generated from the real object may be an object sound corresponding to the real object.
  • the actual object may be various, such as a person, an animal, or an object located in the real space.
  • the object sound corresponding to the real object may be various, such as the voice of a person located in a real space, the cry and foot sounds of an animal, and the horn of a vehicle.
  • the sound acquisition time information for the actual sound may include a sound acquisition time of each of a plurality of real object sounds.
  • the sound acquisition time of the real object sound acquired in the right ear direction and the real object sound acquired in the left ear direction may be different from each other.
  • the processor 350 may acquire play sound transmitted from the sound providing device 100.
  • the play sound may include a plurality of virtual object sounds.
  • Each of the plurality of virtual object sounds may be a sound generated from each of the plurality of virtual objects disposed in the virtual space.
  • the sound generated from the virtual object may be an object sound that has already been recorded and / or already generated as an object sound corresponding to the virtual object.
  • the virtual object When the play sound is a 3D sound source, the virtual object may be various objects constituting a sound source such as drum, guitar, bass, and vocal.
  • the play sound is a 3D VR sound content
  • the virtual object may be various objects constituting 3D VR sound content such as people, animals, and objects included in the 3D virtual reality corresponding to the 3D VR sound content.
  • the object sound corresponding to the virtual object may be various sounds constituting a sound source such as a pre-recorded drum sound, guitar sound, bass sound, and vocal sound.
  • the object sound corresponding to the virtual object may be various sounds constituting the 3D VR sound content, such as a recorded human voice, animal cry and footsteps, and a vehicle horn sound. .
  • the processor 350 may generate a combination sound by selectively combining actual sound and play sound.
  • the processor 350 may select at least one real object sound from a plurality of real object sounds included in the real sound.
  • the processor 350 may recognize a plurality of real object sounds included in the real sound based on the characteristics of the pre-stored object sound.
  • the characteristics of the object sound may be frequency characteristics and volume characteristics of the object sound.
  • the processor 350 may remove the noise sound from the actual sound based on the noise filtering technique. For example, the processor 350 may analyze the noise generated in the real space to remove the sound corresponding to the general noise sound.
  • the noise sound may be a sound corresponding to general noise.
  • the noise sound may be a sound significantly higher than the sound corresponding to the normal audible frequency.
  • the processor 350 may recognize a plurality of real object sounds from the real sound from which the noise sound is removed based on the frequency and / or volume of pre-stored object sounds. For example, the processor 350 may detect a sound corresponding to the frequency and / or volume of pre-stored object sounds from the real sound from which the noise sound has been removed, and recognize the detected sound as a plurality of real object sounds.
  • the processor 350 may select at least one virtual object sound from among a plurality of virtual object sounds included in the play sound.
  • the processor 350 may select all or a plurality of virtual object sounds based on a user's motion.
  • the user's motion may vary, such as the number of rotations of the user's head and the rotational speed of the head.
  • the processor 350 may select all of the plurality of virtual object sounds.
  • the processor 350 may select some of the plurality of virtual object sounds.
  • the processor 350 may select a virtual object sound by applying a method of selecting all or part of the plurality of virtual object sounds opposite to each other. For example, the processor 350 may select all of the plurality of virtual object sounds when the number of head rotations is less than the critical rotation number and / or the head rotation speed is less than the critical rotation speed. The processor 350 may select some of the plurality of virtual object sounds when the number of head rotations is greater than or equal to the critical rotation speed and / or when the head rotation speed is greater than or equal to the critical rotation speed.
  • the processor 350 may select a virtual object sound corresponding to the virtual object located in the user's head direction from among the plurality of virtual object sounds based on the user's head direction. .
  • the processor 350 may generate a combined sound by combining at least one real object sound and at least one virtual object sound.
  • the processor 350 may adjust the volume of the at least one real object sound based on the position of the real object corresponding to the at least one real object sound.
  • the processor 350 may determine the location of the real object corresponding to the at least one real object sound in the real space based on the sound acquisition time for the at least one real object sound.
  • the processor 350 may adjust the volume of at least one real object sound based on the distance between the user's location and the location of the real object.
  • the processor 350 may adjust the volume of the at least one real object sound based on the threshold volume corresponding to the separation distance and the volume of the at least one real object sound.
  • the critical volume corresponding to the separation distance may be preset.
  • the critical volume is a volume range set for each separation distance between the user and the object, and may be a volume range that is not dangerous to the user.
  • the processor 350 may adjust the volume of the at least one real object sound to be lower than the threshold volume range.
  • the processor 350 may adjust the volume of the at least one real object sound to a high level within the threshold volume range.
  • the processor 350 may generate a combined sound by combining at least one real object sound and at least one virtual object sound whose volume is adjusted.
  • the processor 350 may acquire a 2D sound track through the communication module 310 and then extract a plurality of 2D object sounds included in the 2D sound track.
  • the plurality of 2D object sounds may be a sound separated by any one of frequency and object in a 2D sound track.
  • the processor 350 may extract a plurality of two-dimensional object sounds included in the two-dimensional sound track by separating the two-dimensional sound track for each frequency band using an equalizer effect (EQ).
  • EQ equalizer effect
  • the processor 350 may extract a plurality of two-dimensional object sounds included in the two-dimensional sound track by separating the two-dimensional sound track for each object using sound detecting.
  • the processor 350 may convert a plurality of two-dimensional object sounds into a plurality of three-dimensional object sounds by applying each of a plurality of binaural effects to each of the plurality of two-dimensional object sounds.
  • Each of the plurality of 3D object sounds may be a 3D binaural sound in which a 2D object sound is converted into a 3D object sound.
  • the processor 350 may generate first 3D object sounds by applying a first binaural effect to the first 2D object sound among a plurality of 2D object sounds.
  • the processor 350 may determine the first 3D positioning for the first 2D object sound.
  • the processor 350 may generate the first 3D object sound by applying the first 3D positioning and the first binaural effect to the first 2D object sound.
  • the first 3D object sound may be a 3D sound obtained by converting the first 2D object sound into a 3D sound.
  • the processor 350 may generate a second 3D object sound by applying a second binaural effect to the second 2D object sound among a plurality of 2D object sounds.
  • the processor 350 may determine the second three-dimensional positioning for the second two-dimensional object sound differently from the first three-dimensional positioning.
  • the processor 350 may generate a second 3D object sound by applying a second 3D positioning and a second binaural effect to the second 2D object sound.
  • the second 3D object sound may be a 3D sound obtained by converting a 2D object sound into a 3D sound.
  • the first binaural effect and the second binaural effect described above may be different from each other or may be the same binaural effect.
  • the processor 350 may generate a 3D sound track based on a plurality of 3D object sounds.
  • the 3D sound track may be a sound track obtained by converting a 2D sound of a 2D sound track into a 3D sound.
  • the processor 350 may generate a 3D sound track in which a plurality of 3D object sounds are integrated by muxing a plurality of 3D object sounds.
  • the sound generating device 300 is implemented in the sound output device 500 and the two-dimensional sound track is assumed to be a two-dimensional sound source.
  • FIG. 3 shows an example for explaining the sound providing apparatus illustrated in FIG. 1
  • FIG. 4 shows an example for describing the first providing apparatus illustrated in FIG. 3
  • FIG. 5 is an example illustrated in FIG. 3 2 shows an example for explaining the providing device.
  • the sound providing device 100 may be an MP3 player as an electronic device that generates play sound, which is tangible 3D sound content, and provides it to the sound generating device 300 used by the user.
  • the play sound may be various, such as a 3D sound source, a 3D voice, and a 3D VR (virtual reality) sound.
  • the electronic device may be various devices such as a personal computer (PC), a data server, or a portable electronic device.
  • Portable electronic devices include laptop computers, mobile phones, smart phones, tablet PCs, mobile internet devices (MIDs), personal digital assistants (PDAs), and enterprise digital assistants (EDAs). ), Digital still camera, digital video camera, portable multimedia player (PMP), personal navigation device or portable navigation device (PND), handheld game console, e-book (e-book), may be implemented as a smart device (smart device). At this time, the smart device may be implemented as a smart watch or a smart band.
  • the sound providing device 100 includes a first providing device 110 and a second providing device 130.
  • the first providing device 110 may provide the second providing device 130 with object sound already recorded or general 2D audio-only sound.
  • the first providing device 110 may provide the object providing sound or general 2D audio-only sound to the second providing device 130 in a wired and / or wireless manner.
  • the general 2D audio-only sound may be general mono or stereo and multi-channel audio.
  • the wired method may be various wired methods such as USB, display port, and HDMI.
  • the wireless method may be various wireless methods such as Wi-Fi and Bluetooth.
  • the second providing device 130 may generate the play sound by reflecting the object sound in a 3D virtual space or converting a general 2D audio-only sound into a 3D sound.
  • the second providing device 130 may generate a play sound using a binaural recoding technique and / or a binaural effect technique.
  • the binaural recording technology may be a technology for recording 3D sound using a 3D microphone.
  • the 3D microphone may be various, such as a 360-degree microphone and a plurality of microphones.
  • the binaural effect may be a technology that generates a 3D sound through a stereo speaker based on a spatial recognition technology in a sound direction through a sound transmission difference according to the position of both ears of a human.
  • the second providing device 130 may place the virtual object in the 3D virtual space using the sound information of the compass mems and accelator mems so that the object sound corresponding to the virtual object is reflected in the 3D virtual space.
  • the second providing device 110 may generate play sound generated in the 3D virtual space in which the object sound is reflected.
  • the second providing device 130 may convert a general 2D audio-only sound into a 3D audio-only sound to generate a play sound converted into a 3D audio-only sound.
  • the play sound converted to a 3D audio-only sound is a 3D sound (or a multi-channel sound) such as a non-three-dimensional sound (or a non-three-dimensional audio, a non-three-dimensional sound source) or a 5.1 channel.
  • the 3D content may be various, such as a 3D 5.1 channel and a 3D 10.1 channel.
  • the second providing device 130 may provide play sound in various ways.
  • the second providing device 130 may provide play sound to the sound output device 500 used by each of a plurality of users in a 1: N manner.
  • the 1: N method may be a broadcast type method in which play sound is provided to a plurality of users.
  • the second providing device 130 may selectively provide a plurality of play sounds to the sound output device 500 used by each of a plurality of users in an N: N manner.
  • the N: N method may be a customized method that allows a plurality of play sounds to be selectively provided to a plurality of users.
  • the second providing device 130 may provide all of a plurality of play sounds to the sound output device 500 used by a single user in an N: 1 manner.
  • the N: 1 method may be a service-focused multi-access method in which a plurality of play sounds are provided to a single user.
  • the second providing device 130 may provide play sound to the sound output device 500 in the above-described wired and / or wireless manner.
  • FIG. 6 shows an example for explaining the sound output device shown in FIG. 1
  • FIG. 7 shows another example for explaining the sound output device shown in FIG. 1.
  • the sound output device 500 may be a device used by a user to listen to play sounds or combination sounds.
  • the sound output device 500 may be implemented as a wearable type, an in-ear type, an on-ear type, and a brain trans type.
  • the sound output device 500 may be an MP3 player implemented in a wearable type, an in-ear type, an on-ear type, and a brain trans type.
  • the MP3 player may include a wireless communication and a processor with a built-in battery operated independently.
  • the wearable type may be a type combined with a product that is conveniently worn by a user.
  • the wearable type may be a hair band, a shoulder-mounted device, a jumper and / or a jacket-like device such as a space suit, goggles and glasses, and the like.
  • the in-ear type may be an earphone.
  • the on-ear type may be a headphone and a helmet.
  • the brain trans type may be an EEG transmission device.
  • the sound output device 500 may be implemented in an HMD, a smart glass, a see-thru display device, a haptic device using a multi-modal (eg, 5 sense sensing), and a bone conduction audio device.
  • a multi-modal eg, 5 sense sensing
  • the sound output device 500 may directly output a play sound or a combination sound to a user's ear to allow the user to directly listen.
  • the sound output device 500 When the sound output device 500 is a wearable type or a brain trans type, the sound output device 500 detects the position of the user's ear and indirectly outputs a play sound or a combination sound to the user's ear to allow the user to indirectly listen Can be.
  • the sound output device 500 can accurately acquire the actual sound generated in the real space in three dimensions by accurately tracking the user's head using a Compass Mems, Gyroscope, and Accelator Mems.
  • the sound output device 500 may provide various functions such as an energy harvesting function and a black box function as additional functions.
  • the sound output device 500 converts heat from the user's contact area into electrical energy or ambient radio frequency (RF), converts ambient sound into electrical energy, and converts kinetic energy of the listener's movement into electrical energy. It has a harvesting function and can be operated without a separate energy source.
  • RF ambient radio frequency
  • the black box When the black box function is provided to the sound output device 500, the black box may be implemented inside and / or outside of a substantial storage location.
  • the black box can store data through various methods such as internal memory storage and / or external storage and storage using blockchain.
  • the external storage may be external storage such as a cloud connection.
  • the black box can use a security key such as a PKI for access rights for security.
  • the black box may be a camera built-in black box and / or an audio black box and / or a black box based on a physical sensor.
  • the audio black box may store and / or store ambient sound in real time and / or transmit / receive audio data while driving, and read sound by location when storing.
  • the audio black box may be a sound-based black box through audio-based location reading, including 3D audio storage and object-specific location information storage, which are easy to analyze after risk and danger.
  • the black box may be a black box having various functions.
  • the black box may have a real-time storage function.
  • the black box may include a real-time call, a real-time stream, a real-time surrounding recording function, and a function capable of playing when necessary.
  • the black box may further include a function of storing and storing real-time information.
  • the black box may detect an ambient sound event and store it in real time at a specific time.
  • the black box can detect a conversation, an important conversation, an accident, etc., and store (or record) data a few minutes before and after the detection occurrence time point.
  • the black box may perform storage based on an event, not a method of always storing.
  • the black box may store object location information.
  • the black box can detect objects in a specific space or objects with sound, animals and / or people, read them as objects, and store the location information of the objects as 3D information.
  • the black box may be stored by reflecting a few minutes and a specific time before and after the starting point of the detection occurrence time.
  • the black box may store audio data and information data, such as a sound source, a stream audio, and the contents of a call, a transmission / reception call while driving, in real time or a specific time.
  • audio data and information data such as a sound source, a stream audio, and the contents of a call, a transmission / reception call while driving, in real time or a specific time.
  • the black box may utilize recognition through sound, such as voice recognition, as an interface such as spatial storage or indicator control by reading object recognition based on 3D location through spatial and object recognition.
  • recognition through sound such as voice recognition
  • the black box can perform 3D spatial recognition based on sound and control recognition by determining an indicator because it is difficult for a user to recognize who is the user.
  • the black box can store the space in 3D in real time and use it for object-specific storage.
  • the sound output device 500 is an in-ear type earphone.
  • FIG. 8 shows an example for describing a sound output device that is an in-ear type earphone
  • FIG. 9 shows another example for describing a sound output device that is an in-ear type earphone.
  • the sound output device 500 may be a plurality of earphones 510 and 530 worn by a user.
  • the first earphone 510 is an earphone worn on the left ear of the user, and may include a first microphone 511, a first speaker 513, and a first processor 515.
  • the second earphone 530 is an earphone worn on the user's right ear, and may include a second microphone 531, a second speaker 533, and a second processor 535.
  • the first earphone 510 and the second earphone 530 may include a sound generating device 300.
  • the first processor 515 and the second processor 535 may share data with each other.
  • the first processor 515 and the second processor 535 may filter the noise of the real space through the first microphone 511 and the second microphone 531 to obtain a real sound.
  • the first processor 515 and the second processor 535 may analyze noise information around the user to obtain a real sound from which noise is removed through a noise canceling function.
  • the sound acquisition time of the actual sound acquired through the first microphone 511 and the sound acquisition time of the actual sound acquired through the second microphone 513 may be different.
  • the first processor 515 and the second processor 535 correspond to the actual sound based on the difference in sound acquisition time of the actual sound acquired through the first microphone 511 and the second microphone 531, and correspond to the actual space. It can be recognized as a 3D real sound.
  • the first processor 515 and the second processor 535 may acquire the direction of the user's head by trekking the user's head using compass mems. At this time, the first processor 515 and the second processor 535 may perform more precise head tracking using a gyroscope, an accelator mems, etc. in addition to the compass mems.
  • the first processor 515 and the second processor 535 are the actual sound, the sound acquisition time of the actual sound acquired through the first microphone 511, and the sound acquisition time of the actual sound acquired through the second microphone 513 And transmitting information about the direction of the user's head to the sound generating device 300.
  • the first processor 515 and the second processor 535 may output play sound or combination sound through the first speaker 513 and the second speaker 533.
  • the first processor 515 and the second processor 535 acquire the play sound through the MP3 player which is the sound providing device 100, and then apply the 3D sound effect through 3D audio convert Play sound can be output.
  • the play sound to which the 3D sound effect is applied may be a sound to which a 3D audio effect such as a binaural effect is applied.
  • the sound to which the 3D audio effect is applied may be a sound in which a multi-channel 3D audio effect is reflected according to the number of speakers.
  • the sound reflecting the multi-channel 3D audio effect may be various, such as 5.1-channel sound, 7.1-channel sound and 10.1-channel sound.
  • two microphones 511 and 531 and speakers 531 and 533 are illustrated, but are not limited thereto.
  • a plurality of microphones may be implemented to obtain actual sound.
  • the speaker may be implemented in a plurality to output play sound or combination sound.
  • FIG. 10 shows an example for explaining a combination sound according to an embodiment.
  • the combined sound may be a 3D sound generated in a mixed space in which virtual space and real space are mixed.
  • the virtual space may be varied such as a street mode, a forest, a travel destination, a space with memories, or an outer space.
  • the actual space may be a space where listeners are currently located, such as cafes and restaurants. Mixed spaces can range from cafes on the streets to restaurants in the woods.
  • the mixed space may be a space in which the cafe and the outer space are mixed.
  • the mixed space may be a mixed space of the restaurant and Hawaii.
  • the virtual object sound generated in the virtual space may be a 3D sound reflecting the distance (sound size), position (sound direction), and movement (change in sound size and direction) of the virtual object located in the virtual space.
  • the real object sound generated in the real space may be a 3D sound reflecting the distance (sound size), position (sound direction), and movement (change in sound size and direction) of the real object in the real space.
  • FIG. 11 is a flowchart illustrating an operation of the sound generating apparatus shown in FIG. 1.
  • the processor 350 may acquire actual sound through the first microphone 511 and the second microphone 531 of the sound output device 500 (1110).
  • the processor 350 may filter a plurality of real object sounds from the real sound based on a filter (1120).
  • the filter may be an audio filter of various methods such as a real time filter.
  • the processor 350 may remove the noise sound from the actual sound based on the noise filtering technique (1130).
  • the processor 350 may detect a sound corresponding to the frequency and volume of the object sound from the actual sound from which noise is removed based on at least one of the frequency and volume of the object sound (1140).
  • the processor 350 may recognize the detected sound as a plurality of real object sounds (1140).
  • the processor 350 may acquire the play sound transmitted from the sound providing device 500 (1160).
  • the processor 350 may generate a combined sound by combining at least one real object sound among a plurality of real object sounds of the real sound and at least one virtual object sound among a plurality of virtual object sounds of the play sound (1170). ).
  • the processor 350 may provide the combined sound to the user through the first speaker 513 and the second speaker 533 of the sound output device 500 (1180).
  • FIG. 12 shows a sound generating system according to another embodiment.
  • FIGS. 1 to 12 apply equally to each configuration of FIGS. 12 to 14.
  • the sound generating system 20 includes a sound providing device 100 and a sound generating device 300.
  • the sound providing device 100 may provide a 2D sound track, which is a general 2D audio-only sound, to the sound generating device 300.
  • the general 2D audio-only sound may be a 2D sound that is not recorded by applying a 3D sound effect.
  • the sound providing apparatus 100 may generate a two-dimensional sound track composed of a plurality of object sounds.
  • the sound providing apparatus 100 may generate a two-dimensional sound track by recording sound played by each of various instruments.
  • the sound providing apparatus 100 may generate a two-dimensional sound track by combining object sounds for each instrument that has already been recorded (or already generated).
  • the sound providing device 100 may transmit a two-dimensional sound track to the sound generating device 300 by a wired communication method and / or a wireless communication method.
  • the wired communication method may be a communication method using various wired communication methods such as USB, display port, and HDMI.
  • the wireless communication method may be a communication method using various wireless communication methods such as Wi-Fi and Bluetooth.
  • the sound generating apparatus 300 may convert a 2D sound track into a 3D sound track reflecting a 3D virtual space by arranging each of a plurality of 2D object sounds in a 3D virtual space using a binaural effect technology. .
  • the binaural effect may be a technology that generates a 3D sound through a stereo speaker based on a spatial recognition technology in a sound direction through a sound transmission difference according to the position of both ears of a human.
  • the 3D sound track may be a 3D audio-only sound reflecting the sense of space, realism, and directionality of the sound.
  • the 3D audio-only sound is a 3D sound obtained by converting a non-three-dimensional sound (or a non-three-dimensional audio, a non-three-dimensional sound source) or a multi-channel two-dimensional sound (or multi-channel two-dimensional audio) such as 5.1 channels into 3D. Can be.
  • the 3D audio-only sound may be 3D sound of various channels such as 3D 2 channel, 3D 5.1 channel, and 3D 10.1 channel.
  • the sound generating device 300 may provide the 3D sound track to the electronic device in various ways.
  • the electronic device may be various devices such as a sound output device, a personal computer (PC), a data server, or a portable electronic device.
  • Portable electronic devices include laptop computers, mobile phones, smart phones, tablet PCs, mobile internet devices (MIDs), personal digital assistants (PDAs), and enterprise digital assistants (EDAs). ), Digital still camera, digital video camera, portable multimedia player (PMP), personal navigation device or portable navigation device (PND), handheld game console, e-book (e-book), may be implemented as a smart device (smart device). At this time, the smart device may be implemented as a smart watch or a smart band.
  • the sound generating device 300 may provide a 3D sound track to a sound output device used by each of a plurality of listeners in a 1: N manner.
  • the 1: N method may be a broadcast type method in which a 3D sound track is provided to a plurality of listeners.
  • the sound generating device 300 may selectively provide a 3D sound track to a sound output device used by each of a plurality of listeners in an N: N manner.
  • the N: N method may be a custom method in which a plurality of 3D sound tracks are selectively provided to a plurality of listeners.
  • the sound generating device 300 may provide all of a plurality of three-dimensional sound tracks to a sound output device used by a single listener in an N: 1 manner.
  • the N: 1 method may be a service-focused multi-access method in which a plurality of 3D sound tracks are provided to a single listener.
  • the sound generating device 300 may provide the 3D sound track to the sound output device by the wired communication method and / or the wireless communication method described above.
  • the sound output device described above may be implemented as a wearable type, an in-ear type, an on-ear type, and a brain trans type.
  • the wearable type may be a type coupled (or mounted) to a product that the listener conveniently wears.
  • the wearable type may be a hair band, a shoulder-mounted device, a jumper and / or a jacket-like device such as a space suit, goggles and glasses, and the like.
  • the in-ear type may be an earphone.
  • the on-ear type may be a headphone and a helmet.
  • the brain trans type may be an EEG transmission device.
  • the sound output device may be implemented in an HMD, a smart glass, a see-thru display device, a haptic device using a multi-modal (eg, 5 sense sensing), and a bone conduction audio device.
  • a multi-modal eg, 5 sense sensing
  • FIG. 13 shows an example for explaining the operation of the sound generating apparatus shown in FIG. 12.
  • the sound generating device 300 may include a communication module 310, a memory 330 and a processor 350.
  • the basic technical details for each of the components 310, 330, and 350 are substantially the same as those described in FIG. 3.
  • the processor 350 may acquire a 2D sound track.
  • the 2D sound track may be a 2 channel stereo or 1 channel mono type sound.
  • the processor 350 may separate a 2D sound track for each object and frequency, and extract a plurality of 2D object sounds included in the 2D sound track.
  • the processor 350 may detect a plurality of 2D object sounds included in a 2D sound track using sound decoding.
  • the processor 350 separates the 2D sound track for each object corresponding to the detected 2D object sound, and extracts each of a plurality of 2D object sounds, which are sound for each object.
  • the plurality of two-dimensional object sounds may be various musical instrument sounds such as violin sound, drum sound, guitar sound, bass sound, keyboard sound, and trumpet sound.
  • the processor 350 may index (or store) a name (or name) of an object corresponding to each of the plurality of 2D object sounds to each of the plurality of 2D object sounds and manage (or store) the sound track (or audio track).
  • the processor 350 may index the violin to the violin sound and manage the first two-dimensional track.
  • the processor 350 may index the drum to the drum sound and manage it as a second two-dimensional track.
  • the processor 350 may index the guitar to the guitar sound and manage it as a third 2D track.
  • the processor 350 may index the bass to the bass sound and manage it as a fourth 2D track.
  • the processor 350 may manage the keyboard as a fifth two-dimensional track by indexing the keyboard to the keyboard sound.
  • the processor 350 may index the trumpet into the trumpet sound and manage the sixth two-dimensional track.
  • the processor 350 may determine the 3D positioning of the first 2D track to the 6th 2D track differently.
  • the processor 350 applies different 3D positioning and binaural effects to the first 2D track to the 6D 2D track to apply the first 2D track to the 6D 2D track from the first 3D track to the 6D 3D track. Can be converted to tracks. At this time, the processor 350 separates the first 2D track to the 6D 2D track into a plurality of channels, and then applies a binaural effect to the first 2D track to the 6D 2D track divided into a plurality of channels. (Or rendering).
  • the processor 350 may generate a 3D sound track by integrating the first 3D track to the 6th 3D track.
  • the 3D sound track may be a multi-channel 3D sound to which a binaural effect is applied by converting a stereo or mono type 2D sound track.
  • FIG. 14 is a flowchart for describing the operation of the processor illustrated in FIG. 13.
  • the processor 350 may acquire a 2D sound track transmitted from the sound providing device 100 (610).
  • the processor 350 may extract a plurality of 2D object sounds included in the 2D sound track by separating the 2D sound track for each frequency and / or object using the equalizer effect and / or sound detecting technology (630). ).
  • the processor 350 may convert a plurality of 2D object sounds into a plurality of 3D object sounds by applying each of a plurality of binaural effects to each of the plurality of 2D object sounds through a binaural effect technology ( 650).
  • the processor 350 may generate a 3D sound track in which a 2D sound track is converted to 3D sound by integrating a plurality of 3D object sounds (670).
  • the method according to the embodiment may be implemented in the form of program instructions that can be executed through various computer means and recorded on a computer-readable medium.
  • the computer-readable medium may include program instructions, data files, data structures, or the like alone or in combination.
  • the program instructions recorded in the medium may be specially designed and configured for the embodiments or may be known and usable by those skilled in computer software.
  • Examples of computer-readable recording media include magnetic media such as hard disks, floppy disks, and magnetic tapes, optical media such as CD-ROMs, DVDs, and magnetic media such as floptical disks.
  • -Hardware devices specifically configured to store and execute program instructions such as magneto-optical media, and ROM, RAM, flash memory, and the like.
  • program instructions include high-level language codes that can be executed by a computer using an interpreter, etc., as well as machine language codes produced by a compiler.
  • the hardware device described above may be configured to operate as one or more software modules to perform the operations of the embodiments, and vice versa.
  • the software may include a computer program, code, instruction, or a combination of one or more of these, and configure the processing device to operate as desired, or process independently or collectively You can command the device.
  • Software and / or data may be interpreted by a processing device, or to provide instructions or data to a processing device, of any type of machine, component, physical device, virtual equipment, computer storage medium or device. , Or may be permanently or temporarily embodied in the transmitted signal wave.
  • the software may be distributed over networked computer systems, and stored or executed in a distributed manner.
  • Software and data may be stored in one or more computer-readable recording media.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)

Abstract

L'invention concerne un procédé pour générer un son et des et dispositifs réalisant ledit procédé. Le procédé de génération d'un son, selon un mode de réalisation, comprend les étapes consistant à : acquérir un son réel généré dans un espace réel, et un son de lecture généré dans un espace virtuel ; et générer, en combinant le son réel et le son de lecture, un son combiné généré dans une réalité mixte dans laquelle l'espace réel et l'espace virtuel sont mélangés.
PCT/KR2019/015145 2018-11-09 2019-11-08 Procédé de génération de son et dispositifs réalisant ledit procédé WO2020096406A1 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN202211375358.3A CN115767407A (zh) 2018-11-09 2019-11-08 声音生成方法及执行其的装置
US17/292,447 US20220021998A1 (en) 2018-11-09 2019-11-08 Method for generating sound and devices for performing same
CN201980073791.1A CN113039815B (zh) 2018-11-09 2019-11-08 声音生成方法及执行其的装置

Applications Claiming Priority (14)

Application Number Priority Date Filing Date Title
KR20180136955 2018-11-09
KR10-2018-0136955 2018-11-09
KR20190052338 2019-05-03
KR10-2019-0052338 2019-05-03
KR10-2019-0092596 2019-07-30
KR20190092596 2019-07-30
KR10-2019-0128553 2019-10-16
KR20190128553 2019-10-16
KR10-2019-0129756 2019-10-18
KR20190129756 2019-10-18
KR1020190140346A KR102379734B1 (ko) 2018-11-09 2019-11-05 사운드 생성 방법 및 이를 수행하는 장치들
KR10-2019-0140347 2019-11-05
KR1020190140347A KR102322120B1 (ko) 2018-11-09 2019-11-05 사운드 생성 방법 및 이를 수행하는 장치들
KR10-2019-0140346 2019-11-05

Publications (1)

Publication Number Publication Date
WO2020096406A1 true WO2020096406A1 (fr) 2020-05-14

Family

ID=70611928

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2019/015145 WO2020096406A1 (fr) 2018-11-09 2019-11-08 Procédé de génération de son et dispositifs réalisant ledit procédé

Country Status (1)

Country Link
WO (1) WO2020096406A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2602464A (en) * 2020-12-29 2022-07-06 Nokia Technologies Oy A method and apparatus for fusion of virtual scene description and listener space description
GB2619513A (en) * 2022-06-06 2023-12-13 Nokia Technologies Oy Apparatus, method, and computer program for rendering virtual reality audio

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20010091173A (ko) * 2000-03-13 2001-10-23 박종권 가상현실 시스템
KR20050061808A (ko) * 2003-12-18 2005-06-23 한국전자통신연구원 가상 입체 음향 생성 장치 및 그 방법
KR20060050420A (ko) * 2004-08-17 2006-05-19 하르만 인터내셔날 인더스트리즈, 인코포레이티드 차량에서의 오디오 신호의 구성을 위한 사운드 프로세싱시스템
KR101535013B1 (ko) * 2014-02-27 2015-07-07 현대자동차주식회사 차량의 실내외 가상 사운드 시스템
KR20180109910A (ko) * 2016-02-04 2018-10-08 매직 립, 인코포레이티드 증강 현실 시스템에서 오디오를 지향시키기 위한 기법

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20010091173A (ko) * 2000-03-13 2001-10-23 박종권 가상현실 시스템
KR20050061808A (ko) * 2003-12-18 2005-06-23 한국전자통신연구원 가상 입체 음향 생성 장치 및 그 방법
KR20060050420A (ko) * 2004-08-17 2006-05-19 하르만 인터내셔날 인더스트리즈, 인코포레이티드 차량에서의 오디오 신호의 구성을 위한 사운드 프로세싱시스템
KR101535013B1 (ko) * 2014-02-27 2015-07-07 현대자동차주식회사 차량의 실내외 가상 사운드 시스템
KR20180109910A (ko) * 2016-02-04 2018-10-08 매직 립, 인코포레이티드 증강 현실 시스템에서 오디오를 지향시키기 위한 기법

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2602464A (en) * 2020-12-29 2022-07-06 Nokia Technologies Oy A method and apparatus for fusion of virtual scene description and listener space description
GB2619513A (en) * 2022-06-06 2023-12-13 Nokia Technologies Oy Apparatus, method, and computer program for rendering virtual reality audio

Similar Documents

Publication Publication Date Title
WO2018056780A1 (fr) Procédé et appareil de traitement de signal audio binaural
WO2020141824A2 (fr) Procédé de traitement de signal audio et dispositif électronique pour le prendre en charge
WO2018199390A1 (fr) Dispositif électronique
WO2016089180A1 (fr) Procédé et appareil de traitement de signal audio destiné à un rendu binauriculaire
WO2012157792A1 (fr) Dispositif électronique
EP3248383A1 (fr) Appareil et procédé de correction de contenu
WO2019107868A1 (fr) Appareil et procédé de sortie de signal audio, et appareil d'affichage l'utilisant
WO2014175669A1 (fr) Procédé de traitement de signaux audio pour permettre une localisation d'image sonore
WO2019139301A1 (fr) Dispositif électronique et procédé d'expression de sous-titres de celui-ci
WO2016154868A1 (fr) Système de vol, appareil volant, appareil sonore et procédé de traitement de son
WO2016032021A1 (fr) Appareil et procédé de reconnaissance de commandes vocales
WO2020096406A1 (fr) Procédé de génération de son et dispositifs réalisant ledit procédé
WO2021060680A1 (fr) Procédés et systèmes d'enregistrement de signal audio mélangé et de reproduction de contenu audio directionnel
WO2016114432A1 (fr) Procédé de traitement de sons sur la base d'informations d'image, et dispositif correspondant
WO2017171137A1 (fr) Aide auditive, dispositif portatif et procédé de commande associé
WO2015147619A1 (fr) Procédé et appareil pour restituer un signal acoustique, et support lisible par ordinateur
WO2017204498A1 (fr) Terminal mobile
WO2017065444A1 (fr) Dispositif électronique et procédé de commande de dispositif électronique
WO2015125993A1 (fr) Terminal mobile et son procédé de commande
WO2019031652A1 (fr) Procédé de lecture audio tridimensionnelle et appareil de lecture
WO2021060575A1 (fr) Serveur à intelligence artificielle et procédé de fonctionnement associé
WO2021010562A1 (fr) Appareil électronique et procédé de commande associé
WO2016167508A1 (fr) Procédé et dispositif d'annulation de bruit pour un dispositif sonore mobile
WO2020040541A1 (fr) Dispositif électronique, procédé de commande associé et support d'enregistrement
WO2020141643A1 (fr) Serveur de synthèse vocale et terminal

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19883100

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19883100

Country of ref document: EP

Kind code of ref document: A1