WO2016148553A2 - Procédé et dispositif de modification et de reproduction d'un son tridimensionnel - Google Patents

Procédé et dispositif de modification et de reproduction d'un son tridimensionnel Download PDF

Info

Publication number
WO2016148553A2
WO2016148553A2 PCT/KR2016/002826 KR2016002826W WO2016148553A2 WO 2016148553 A2 WO2016148553 A2 WO 2016148553A2 KR 2016002826 W KR2016002826 W KR 2016002826W WO 2016148553 A2 WO2016148553 A2 WO 2016148553A2
Authority
WO
WIPO (PCT)
Prior art keywords
source
sound
parameters
channel
data
Prior art date
Application number
PCT/KR2016/002826
Other languages
English (en)
Korean (ko)
Other versions
WO2016148553A3 (fr
Inventor
구본희
이종석
김대진
김동준
Original Assignee
(주)소닉티어랩
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020160032468A external-priority patent/KR20160113036A/ko
Application filed by (주)소닉티어랩 filed Critical (주)소닉티어랩
Publication of WO2016148553A2 publication Critical patent/WO2016148553A2/fr
Publication of WO2016148553A3 publication Critical patent/WO2016148553A3/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S3/00Systems employing more than two channels, e.g. quadraphonic
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S5/00Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation 
    • H04S5/02Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation  of the pseudo four-channel type, e.g. in which rear channel signals are derived from two-channel stereo signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control

Definitions

  • the following embodiments relate to three-dimensional sound, and more particularly, a method and apparatus for providing editing of a three-dimensional sound and providing data of the three-dimensional sound generated in accordance with the edit are disclosed.
  • the three-dimensional sound is a sound that makes the user feel as if he or she is actually in the situation of the image as the intensity and direction of the sound is set to correspond to the image, more than simply output from the speaker.
  • a sound image externalization technology is required to cause sound images to form on the outside of the head of a user seated in a movie theater.
  • HRTF Head Related Transfer Function
  • an environment for forming and controlling a sound field using a plurality of speakers has been provided. Multiple channels are used for the formation and control of the sound field.
  • the sound field is becoming more sophisticated as more channels are used. For example, beyond the existing 5.1 and 7.1 channels, 11.1 channels, 15.1 channels and 31.1 channels are used for the formation of the sound field.
  • the image externalization technology is creating an environment in which a user can enjoy a three-dimensional sound without a speaker through a plurality of channels through the headphones.
  • Prior art still reproduces only the ambient sound as the surround sound based on the speaker.
  • Prior arts unlike the three-dimensional image of a three-dimensional movie, have limitations in bringing the sound image to the front or back of the screen.
  • the sound image unlike a three-dimensional image protruding to the front of the screen, the sound image does not protrude to the front of the screen, thereby providing a complete three-dimensional sound.
  • One embodiment may provide a method and apparatus for providing editing of three-dimensional sound using parameters.
  • One embodiment may provide a method and apparatus for adjusting a parameter of three-dimensional sound via a graphical interface.
  • the method comprises: determining values of one or more parameters for a three-dimensional sound; Generating data of the three-dimensional sound based on the one or more parameters; And outputting a signal comprising the data, wherein the one or more parameters are provided to provide a three dimensional sound associated with a source of sound located in a three dimensional space.
  • the one or more parameters may include a maximum distance parameter indicative of a maximum distance of an area behind the screen and a maximum threshold parameter indicative of a maximum distance of an area ahead of the screen.
  • the one or more parameters may include a damper parameter representing a damper of the source.
  • the one or more parameters may include a source start point parameter indicative of the position of the start point of the source and a source end point parameter indicative of the position of the end point of the source.
  • the one or more parameters may include a source trace line parameter indicating a line through which the source from the start point to the end point travels.
  • the source trace line parameter may indicate a plurality of lines connecting the start point and the end point.
  • the one or more parameters may include speed and acceleration parameters indicative of the speed and acceleration of the source.
  • the one or more parameters may have different values over time.
  • the data of the 3D sound may include data of a plurality of channels.
  • the one or more parameters may be used for each channel of a plurality of channels.
  • the values of the one or more parameters can be edited via a graphical interface.
  • Determining unit for determining the values of one or more parameters for the three-dimensional sound;
  • a generator configured to generate data of the 3D sound based on the one or more parameters;
  • an output unit for outputting a signal including the data, wherein the one or more parameters are provided with a three-dimensional sound providing apparatus associated with a source of sound located in a three-dimensional space.
  • the one or more parameters may include a maximum distance parameter indicative of a maximum distance of an area behind the screen and a maximum threshold parameter indicative of a maximum distance of an area ahead of the screen.
  • the one or more parameters may include a damper parameter representing a damper of the source.
  • the one or more parameters may include a source start point parameter indicative of the position of the start point of the source and a source end point parameter indicative of the position of the end point of the source.
  • the one or more parameters may include a source trace line parameter indicating a line through which the source from the start point to the end point travels.
  • the source trace line parameter may indicate a plurality of lines connecting the start point and the end point.
  • the one or more parameters may include speed and acceleration parameters indicative of the speed and acceleration of the source.
  • the one or more parameters may have different values over time.
  • the data of the 3D sound may include data of a plurality of channels.
  • the one or more parameters may be used for each channel of a plurality of channels.
  • the values of the one or more parameters can be edited via a graphical interface.
  • a method and apparatus are provided for providing editing of three-dimensional sound using parameters.
  • Methods and apparatus are provided for adjusting parameters of three-dimensional sound through a graphical interface.
  • FIG. 1 is a structural diagram of a 3D sound reproducing apparatus according to an example.
  • FIG. 2 is a flowchart illustrating a 3D sound reproduction method according to an example.
  • FIG. 3 is a block diagram illustrating a primary signal processor according to an example.
  • FIG. 5 is a structural diagram of a 3D sound providing apparatus according to an embodiment.
  • FIG. 6 is a flowchart of a 3D sound providing method according to an exemplary embodiment.
  • FIG 8 shows a first process of a mixing method according to an example.
  • FIG. 10 shows a third process of a mixing method according to an example.
  • 11 illustrates an overall interface for determining values of one or more parameters for three-dimensional sound according to an example.
  • FIG. 13 illustrates a source setting interface according to an example.
  • FIG. 16 illustrates editing a source trace line according to an example.
  • FIG. 17 illustrates a screen interface according to an example.
  • 20 illustrates a locator area of a project according to one embodiment.
  • FIG. 21 illustrates an enlarged speed graph according to an example.
  • FIG. 22 illustrates an electronic device implementing a 3D sound reproducing apparatus according to an embodiment.
  • FIG. 23 is a diagram illustrating an electronic device that implements a 3D sound providing apparatus according to an embodiment.
  • first and second may be used to describe various components, but the above components should not be limited by the above terms. The above terms are used to distinguish one component from another component.
  • first component may be referred to as the second component, and similarly, the second component may also be referred to as the first component.
  • each component shown in the embodiments are shown independently to represent different characteristic functions, and do not mean that each component is composed of only separate hardware or one software component unit. That is, each component is listed as each component for convenience of description. For example, at least two of the components may be combined into one component. In addition, one component may be divided into a plurality of components. The integrated and separated embodiments of each of these components are also included in the scope of the present invention without departing from the essence.
  • components may not be essential components for performing essential functions, but may be optional components for improving performance.
  • Embodiments may be implemented including only components necessary to implement the nature of the embodiments, and structures including the optional components, such as, for example, components used only for performance improvement, are also included in the scope of rights.
  • FIG. 1 is a structural diagram of a 3D sound reproducing apparatus according to an example.
  • the three-dimensional sound reproducing apparatus 100 includes a signal receiver 105, a signal detector 110, a primary signal processor 120, a channel allocator 130, a secondary signal processor 140, and a function adjuster 150.
  • the sound image externalization implementer 160, the bypass adjuster 170, the bypass switch 171, the far-range speaker sensing / reproducing unit 180, and the proximity speaker sensing / reproducing unit 190 may be included.
  • the functions and operations of the bypass adjusting unit 170, the bypass switching switch 171, the remote speaker sensing / reproducing unit 180, and the proximity speaker sensing / reproducing unit 190 will be described in detail below.
  • the 3D sound reproducing apparatus 100 may include a remote speaker 181 and a proximity speaker 191. Alternatively, the 3D sound reproducing apparatus 100 and the remote speaker 181 may be connected by wire or wirelessly. The 3D sound reproducing apparatus 100 and the proximity speaker 191 may be connected by wire or wirelessly.
  • the remote speaker 181 may be a speaker that is not used for sound externalization.
  • the remote speaker 181 may be a speaker physically separated from the listener of the sound generated by the 3D sound reproducing apparatus 100.
  • the remote speaker 181 may be a loud speaker.
  • the remote speaker 181 may be a main speaker, a surround speaker, or a sealing speaker connected to a receiver.
  • the remote speaker 181 may be a speaker attached to a TV or a desktop speaker attached to a computer. Alternatively, the remote speaker 181 may be a speaker for transative reproduction.
  • the proximity speaker 191 may be a speaker used for sound externalization.
  • the proximity speaker 191 may be a speaker in close contact with the listener's ear of the sound generated by the 3D sound reproducing apparatus 100.
  • the proximity speaker 191 may be headphones or earphones.
  • the proximity speaker 191 may be in-ear headphones, on-ear headphones, or over-ear headphones.
  • FIG. 2 is a flowchart illustrating a 3D sound reproduction method according to an example.
  • the signal receiver 105 may receive a signal transmitted from a sound transmission device.
  • the signal detector 110 may detect a type of the received signal.
  • the signal detector 110 may detect whether the received signal is in a format that supports 3D sound, and may detect whether the received signal includes object data.
  • the signal detector 110 may detect that the received signal is a signal or content such as mono, stereo, 5.1 channel, and 7.1 channel.
  • the primary signal processor 120 may determine how to perform processing on the transmitted signal.
  • the signal detector 110 may detect the type of the received signal using the number of audio channels of the received signal.
  • the signal detector 110 may detect that the received signal is a stereo signal. If the received signal has six audio channels, the signal detector 110 may detect that the received 5.1 channel signal is received. When the received signal is a signal using a codec, the signal detector 110 may check the information of the channel using the information of the head of the signal, and the type of the received signal using the confirmed channel information. Can be detected.
  • the signal detector 110 may determine whether to bypass the received signal. Alternatively, the signal detector 110 may determine whether to generate 3D sound or stereo sound. When bypassing the received signal, stereo sound may be produced. If the received signal is not bypassed, three-dimensional sound may be generated. The stereo sound may be a general 2D sound rather than a 3D sound.
  • the signal detector 110 may determine whether to generate three-dimensional sound or stereo sound.
  • the signal detector 110 may determine to generate a stereo sound using the bypass function.
  • the signal detector 110 may determine whether to generate 3D sound or stereo sound based on the setting of the bypass switch 171.
  • the user of the 3D sound reproducing apparatus 100 may determine whether to generate 3D sound or stereo sound by setting the bypass switch 171.
  • the bypass changeover switch 171 may be set to on or off.
  • the signal detector 110 may generate stereo sound.
  • the bypass switching switch 171 is set to off, a three-dimensional sound to be described later may be generated.
  • bypass changeover switch 171 may be set to "far output”, “close output” or off.
  • the signal detector 110 may generate stereo sound.
  • the bypass switching switch 171 is set to off, a three-dimensional sound to be described later may be generated.
  • step 235 If bypass of the received signal is determined (ie, it is determined that stereo sound is to be produced), step 235 can be performed. If it is determined not to bypass the received signal (ie, it is determined that three-dimensional sound is to be produced), step 245 may be performed.
  • the bypass adjuster 170 may select a speaker to which the received signal is immediately transmitted among the remote speaker 181 and the proximity speaker 191.
  • the bypass adjuster 170 may directly transmit the received signal to one of the one or more remote speakers and the adjacent speaker 191.
  • bypass adjuster 170 may select the far speaker 181 and transmit the received signal directly to the far speaker 181.
  • the bypass switching switch 171 is set to "proximity output”
  • the bypass adjuster 170 may select the proximity speaker 191 and transmit the received signal directly to the proximity speaker 191.
  • the function adjusting unit 150 may perform setting according to the taste of the listener.
  • the setting may affect the operation of the primary signal processor 120.
  • the function adjuster 150 may perform settings for the primary signal processor 120.
  • the function adjusting unit 150 may set a downmix according to the taste of the listener.
  • the function adjusting unit 150 may set an upmix according to the taste of the listener.
  • the primary signal processor 120 may generate one or more separated signals by performing channel separation and object separation on the signal detected by the signal detector 110.
  • the primary signal processor 120 may extract channel data and object data from the detected signal by separating the channel data and the object data from the detected signal.
  • Channel data and object data may constitute three-dimensional data.
  • the 3D data may include channel data and object data.
  • the channel data may be data of channel components recognized by the primary signal processor 120 among the detected signals.
  • the object data may be data of an object component recognized by the primary signal processor 120 among the detected signals.
  • the object data may include location information about the space, direction information about the motion, size information of the sound, and the like.
  • the primary signal processor 120 may separate the channel data of the detected signal according to the surround standard.
  • the primary signal processor 120 may separate and extract automation data of each object for each object of one or more objects.
  • the primary signal processor 120 may apply processing according to the sensed signal type.
  • the primary signal processor 120 may perform channel separation and object separation on the detected signal according to the mode of processing determined by the user.
  • the mode may include a stereo mode, a multichannel surround mode, and a three dimensional audio mode including an object.
  • the primary signal processor 120 may provide channel data of the sensed signal to the channel allocator 130 without additional processing.
  • the primary signal processor 120 may separate the channel data and the object data using the object extraction method, and provide the separated channel data and the object data to the channel allocator 130. Can be.
  • the primary signal processor 120 may separate the channel data and the object data according to the channel information and the object information of the detected signal, and separate the separated channel data and the object data.
  • the channel allocator 130 may provide the channel allocation unit 130.
  • the primary signal processor 120 may separate and extract channel data, object data, and spatial image data by using a decoder when the sensed signal corresponds to 3D sound.
  • One or more separated signals may include at least some of channel data, object data, and spatial image data.
  • the primary signal processor 120 may transmit one or more separated signals to the channel allocator 130.
  • the channel allocator 130 may acquire information about the remote speaker 181 and the proximity speaker 191.
  • the channel allocator 130 may detect whether the remote speaker 181 is connected to the 3D sound reproducing apparatus 100, and determine the number of remote speakers 181 connected to the 3D sound reproducing apparatus 100. Can be.
  • the remote speaker detecting / reproducing unit 180 may detect whether the remote speaker 181 is connected or detect the number of the remote speakers 181 connected to the 3D sound reproducing apparatus 100.
  • the remote speaker detecting / reproducing unit 180 may detect the remote speaker 181 actually connected to the 3D sound reproducing apparatus 100.
  • the remote speaker detecting / reproducing unit 180 may provide the number of the detected remote speakers 181 to the channel allocator 130.
  • the remote speaker detection / reproducing unit 180 may detect the number of remote speakers 181 currently connected to the preamplifier and / or the power amplifier.
  • the remote speaker detecting / reproducing unit 180 may provide the channel allocating unit 130 with the number of detected remote speakers 181 set by the user.
  • the channel allocator 130 may detect whether the proximity speaker 191 is connected to the 3D sound reproducing apparatus 100.
  • the proximity speaker detecting / reproducing unit 190 may detect the proximity speaker 191 actually connected to the 3D sound reproducing apparatus 100.
  • the proximity speaker detecting / reproducing unit 190 may provide the channel allocating unit 130 with information indicating whether the proximity speaker 191 is actually connected to the 3D sound reproducing apparatus 100.
  • step 255 allocation of each signal of one or more separate signals may be performed.
  • the channel allocator 130 may allocate each signal of one or more separated signals according to the number of detected one or more remote speakers and whether the proximity speaker 191 is connected. .
  • the assignment may be to classify the signal into one of channel data and object data.
  • the channel allocator 130 may classify each signal of one or more separated signals into one of channel data and object data.
  • the channel allocator 130 may classify each signal of the one or more separated signals into one of channel data and object data according to the number of detected one or more remote speakers and whether the proximity speaker 191 is connected.
  • the channel allocator 130 may allocate channel information according to a channel supported by the 3D sound reproducing apparatus 100 and arrange object data.
  • the channel allocator 130 may separate the channel data into one or more remote speaker channels of the one or more remote speakers, and may separate the object data into a sound externalization channel of the proximity speaker 191.
  • the channel allocator 130 converts each signal of one or more separated signals according to the number of one or more remote speakers into one or more channels of one or more remote speaker channels of the one or more remote speakers and the sound externalization channel of the proximity speaker 191. Can be sent to.
  • the remote speaker channel can provide separate data in surround formats such as 5.1 channel, 7.1 channel, 8.1 channel, 9.1 channel, 10.2 channel, 11.1 channel, 13.1 channel, 14.2 channel, 15.1 channel, 22.2 channel, 30.2 channel and 31.1 channel. Can be.
  • the remote speaker channel can support all existing surround formats.
  • the acoustic externalization channel may provide information about data, spatial coordinates, vectors, levels, and the like of the object.
  • the remote speaker channel and the sound image externalization channel may be exchanged with each other.
  • the channel allocator 130 may use the remote speaker channel as a channel for reproducing object data, and use the sound externalization channel as a channel for reproducing channel data, if necessary.
  • the channel allocator 130 may transmit one or more separated signals to one or more remote speaker channels using the following allocation rule.
  • N may represent the number of one or more remote speakers, and n may represent the number of one or more effective channels of one or more separate signals.
  • the channel allocator 130 may assign one or more effective channels to one or more remote speaker channels one-to-one.
  • the channel allocator 130 assigns some effective channels as many as one or more remote speakers among one or more effective channels to one or more remote speaker channels of the one or more remote speakers, respectively. And the remaining valid channels that are not assigned to one or more far channels of the one or more effective channels can be assigned to the audio externalization channel.
  • the channel allocator 130 may perform downmixing of one or more effective channels according to a user's setting.
  • the channel allocator 130 may perform downmixing on a plurality of effective channels according to the number of one or more remote speakers.
  • the channel allocator 130 may allocate all of the one or more effective channels to only one or more remote speaker channels of the one or more remote speakers through the downmix, but may not allocate the sound externalization channel.
  • the channel allocator 130 may allocate all of one or more effective channels only to the sound externalization channel through the downmix, but may not assign the one or more remote speaker channels.
  • the channel allocator 130 assigns one or more effective channels to some remote speaker channels as many as one or more effective channels of one or more remote speaker channels of the one or more remote speakers, respectively.
  • the effective channel may not be allocated to the remaining remote speaker channel to which each effective channel of the one or more effective channels of the one or more remote speaker channels is not assigned.
  • the channel allocator 130 may perform upmix on one or more effective channels according to a user's setting.
  • the channel allocator 130 may perform upmix on one or more effective channels according to the number of one or more remote speakers.
  • the channel allocator 130 may allocate all of one or more effective channels to one or more remote speaker channels of the one or more remote speakers through the upmix, but may not allocate the sound externalization channel.
  • the channel allocator 130 may assign all of one or more effective channels to only the sound externalization channel, but may not assign to one or more remote speaker channels.
  • the function adjusting unit 150 may perform a setting for the secondary signal processing unit 140.
  • the function adjuster 150 may set a parameter used by the secondary signal processor 140 to perform processing on one or more separated signals.
  • the function adjusting unit 150 may receive a user input for setting a parameter.
  • the function adjusting unit 150 may adjust the size of the space according to the environment of the listener, and may adjust the size of the virtual space according to the taste of the listener.
  • the function adjusting unit 150 may set sound effects such as an equalizer, a compressor, and a delay for correcting the listener's environment.
  • the function adjuster 150 may set an adjustment of the level of each signal of one or more separate signals for correction of the listener's listening environment.
  • the function adjuster 150 may set positions of one or more remote speakers for correction of the listener's listening environment.
  • the secondary signal processor 140 may perform processing on one or more separated signals.
  • the secondary signal processor 140 may set the tone and level of each signal of one or more separated signals according to the taste of the listener through processing, and may add a virtual space image to each signal.
  • the secondary signal processor 140 may process each signal of one or more separated signals in accordance with a space according to the environment of the listener or a virtual space according to the taste of the listener.
  • Each signal may be one of channel data and object data.
  • the secondary signal processor 140 may mix object data into channel data.
  • the secondary signal processor 140 may give a spatial image to each signal of one or more separated signals.
  • the secondary signal processor 140 may perform a function of a room simulator through the provision of a spatial image.
  • the secondary signal processor 140 may apply a sound effect to each signal of one or more separated signals according to the listener's listening environment.
  • the secondary signal processor 140 may adjust the level and tone of each signal of one or more separated signals according to the listener's listening environment.
  • the secondary signal processor 140 may adjust the levels of each of the one or more far speakers and the near speaker of the one or more far speakers.
  • the secondary signal processor 140 may adjust positions of one or more remote speakers.
  • the secondary signal processor 140 may adjust an artificial distance between one or more remote speakers and a proximity speaker.
  • the secondary signal processor 140 may mix levels of one or more remote speakers and a proximity speaker.
  • the remote speaker sensing / reproducing unit 180 may output the channel data processed by the secondary signal processing unit 140 to one or more remote speakers.
  • the sound externalization implementer 160 may implement sound externalization of the object data processed by the secondary signal processor 140.
  • the sound externalization implementer 160 may generate data on which sound externalization is implemented by implementing sound externalization.
  • the sound externalization implementer 150 may implement sound externalization by rearranging information represented by the object data processed by the secondary signal processor 140 according to the spatial image represented by the spatial image data.
  • the channel of the object data may not be limited.
  • the object data may be input to the sound externalization implementer 160 through one or more infinite channels.
  • Sound externalization data generated by the sound externalization implementation unit 160 may include two or more channels.
  • the proximity speaker detection / reproducing unit 190 may reproduce data in which sound image externalization is implemented using the proximity speaker 191.
  • the proximity speaker detection / reproducing unit 190 may output data on which sound image externalization is implemented to the proximity speaker 191.
  • the proximity speaker sensing / reproducing unit 190 may process the processing of the sound externalization of the data on which the sound externalization is implemented.
  • the proximity speaker sensing / reproducing unit 190 may output a result of the processing of the external sound image to the proximity speaker 191.
  • FIG. 3 is a block diagram illustrating a primary signal processor according to an example.
  • the primary signal processor 120 includes a channel detector 310, a level detector 320, a threshold controller 330, a channel comparator 340, an object data allocation controller 350, and channel data.
  • the generator 360 may include an object data generator 370, a channel mixer 380, and an object data mixer 390.
  • Channel detector 310 level detector 320, threshold controller 330, channel comparator 340, object data allocation controller 350, channel data generator 360, object data generator 370 ), Functions and operations of the channel mix unit 380 and the object data mix unit 390 will be described in detail below.
  • Step 410, 415, 420, 425, 430, 440, 445, 450 and 455 the signal detector 110 detects a general surround signal rather than a 3D audio signal, and the bypass adjuster 170. ) May be performed when the bypass function is not used.
  • Step 245 described above with reference to FIG. 2 may include steps 410, 415, 420, 425, 430, 440, 445, 450, and 455.
  • the primary signal processor 120 may provide compatibility with existing content. Can be extracted.
  • the channel detector 310 may divide the signal detected by the signal detector 110 into one or more channels according to the surround format of the detected signal. Each channel of the one or more channels may have a unique channel number.
  • the channel detector 310 may separate the detected signal into one or more channels according to the surround format of the surround channel.
  • the surround channel may include a 5.1 channel, a 7.1 channel, and the like.
  • a surround channel may include all channels that do not have a height component.
  • the surround channel may include 6.1 channels, 8.1 channels, and 9.1 channels.
  • the channel detector 310 may bypass the detected signal to the channel allocator 130. For example, if the sensed signal is a signal of a stereo channel, the sensed signal may not include a surround channel.
  • the threshold controller 330 may determine the threshold level.
  • the level detector 320 may detect levels of one or more channels, and may divide one or more channels into one or more high level channels and one or more low level channels.
  • Each channel of the one or more high level channels may be a channel having a level above the threshold level.
  • Each channel of the one or more low level channels may be a channel having a level less than the threshold level.
  • the level detector 320 may classify a channel having a level higher than or equal to a threshold level among the one or more channels as a high level channel, and classify a channel having a level smaller than the threshold level as a low level channel.
  • high level channels can be used as object data and low level channels can be used as channel data.
  • the level detector 320 may transmit one or more high level channels and one or more low level channels to the channel comparator 340.
  • the object data allocation controller 350 may set whether to use dialogue as channel data or object data.
  • the channel comparator 340 may set whether to use data of each channel of one or more channels as channel data or object data.
  • the channel comparator 340 may set one or more low level channels to be used as channel data.
  • the channel comparator 340 may finally determine whether to use data of one or more high level channels as object data or channel data through channel comparison with respect to one or more high level channels.
  • the channel comparator 340 may be configured to use the dialogue channel among the one or more high level channels as one of the object data and the channel data according to the setting of the object data allocation controller 350. In other words, the channel comparator 340 may determine which channel among the one or more high level channels is used as the remote speaker channel or the audio externalization channel according to the setting of the object data allocation controller 350.
  • the channel comparator 340 may set the channel of dialogue among the one or more high level channels to be used as channel data.
  • the channel comparator 340 may set the dialogue channel among the one or more high level channels to be used as the object data.
  • the center channel may be a metabolic channel with a probability of 90% or more.
  • the channel comparator 340 may set the plurality of channels from which the same data is extracted to be used as object data.
  • the extracted data when data for one channel is extracted from a channel other than the center channel among the one or more high level channels, the extracted data may be a special effect.
  • the special effect may be object data. Therefore, when data for one channel is extracted from a channel other than the center channel among one or more high level channels, the channel comparator 340 may set the channel from which the data for one channel is extracted as object data. Can be.
  • the extracted data when data is extracted only from a left channel and a right channel of one or more high level channels, the extracted data may be music data or ambience with a large volume. Music data or ambience with large volume may be channel data. Therefore, when data is extracted only from the left channel and the right channel among the one or more high level channels, the channel comparator 340 may set the left channel and the right channel to be used as the channel data.
  • the channel data generator 360 may generate channel data using data of a channel set to be used as channel data among one or more channels.
  • the channel data generator 360 may prevent data of a channel having a level higher than or equal to the threshold level set by the threshold controller 330 from being reproduced in the remote speaker 181.
  • the channel data generator 360 may allow the data of the channel set to be used as the channel data to be reproduced through the remote speaker channel according to the setting method of the channel comparator 340.
  • the channel mixer 380 may perform level compensation on the channel data, which is reduced according to the extraction of the object data.
  • the channel mix unit 380 may apply a release parameter in level compensation. By applying the release parameter, a break between the object data and the sound in the threshold region may be prevented.
  • the object data generator 370 may generate object data using data of a channel set to be used as object data among one or more channels.
  • the object data generator 370 may allow only the data of the channel having a level equal to or greater than the threshold level set by the threshold controller 330 to be reproduced in the proximity speaker 191.
  • the channel data generator 360 may allow the data of the channel set to be used as the object data to be reproduced through the sound externalization channel according to the setting method of the channel comparator 340.
  • the object data mixing unit 390 may perform level compensation on the object data, which is reduced according to the extraction of the object data.
  • Object data may be extracted from channel components. Accordingly, the object data mixing unit 390 may adjust the object data such that only pure object data can be reproduced by using a parameter such as a gate.
  • the object data mix unit 390 may apply a gate time to adjust the object data so that a mix between the channel data and the object data occurs naturally.
  • FIG. 5 is a structural diagram of a 3D sound providing apparatus according to an embodiment.
  • the 3D sound providing apparatus 500 may include a setting unit 510, a generating unit 520, and an output unit 530. Functions and operations of the setting unit 510, the generating unit 520, and the output unit 530 will be described in detail below.
  • FIG. 6 is a flowchart of a 3D sound providing method according to an exemplary embodiment.
  • the setting unit 510 may determine values of one or more parameters for the 3D sound.
  • One or more parameters may be associated with a source of sound located in three-dimensional space.
  • the values of the one or more parameters can be edited via the graphical interface described below.
  • Step 610 may be performed repeatedly as the values of one or more parameters change.
  • the generator 520 may generate data of the 3D sound based on one or more parameters.
  • the generator 520 may generate data of the 3D sound by reflecting one or more parameters.
  • the data of the 3D sound may include data of a plurality of channels. One or more parameters may be used for each channel of the plurality of channels.
  • the output unit 530 may output a signal including data of 3D sound.
  • the output signal may be received and used by the 3D sound reproducing apparatus 100.
  • Data of the 3D sound of the output signal may correspond to the object data described above with reference to FIGS. 2 to 4.
  • the one or more parameters may correspond to positional information on the space of the object data, direction information on the movement, and size information of the sound, respectively.
  • one of the one or more parameters related to the location of the source may correspond to location information about the space.
  • the parameter related to the movement of the source among the one or more parameters may correspond to the direction information about the movement.
  • parameters representing damper, mix, diffuse, gain, and tail of the source may correspond to loudness information.
  • the threshold may represent a bulge limit line.
  • D may represent a difference between a general mixing line and a mixing line according to an example of the present invention.
  • the screen may be a location where the image is formed.
  • a mix based on the screen can be made.
  • the delay at the point of the screen may be zero.
  • the setting unit 510 may set the threshold d as a parameter.
  • the generator 520 may apply the delay d as much as the threshold d to the entire signal of the three-dimensional sound so that the three-dimensional sound is represented by the distance corresponding to the threshold d.
  • the delay cannot be pulled forward.
  • the viewer's hearing recognizes the distance d '
  • the viewer may recognize the sound according to the position of the screen according to the psychological factor of the viewer. That is to say, as a result, the viewer can perceive the position of the screen as if a sound was produced.
  • the sound can be projected toward the seat where the viewer is located by the distance of D.
  • the source of the sound may have a velocity component.
  • the source may represent a source of sound.
  • the source may correspond to the above-described object.
  • FIG 8 shows a first process of a mixing method according to an example.
  • the generation unit 520 may enlarge the spatial image by applying a threshold d delay to the entire signal of the 3D sound.
  • the generation unit 520 may reduce the pre-delay value by a distance d corresponding to the increased space in the reverb process. By reducing the value of the pre-delay, the generator 520 may perform mixing to prevent the viewer from recognizing the increased space.
  • the delay may not be applied to the rear of the screen in order to minimize distortion of the space and recognition of the distortion of the viewer.
  • the delay applied by d may be ignored and a sound image may be formed relative to the screen.
  • FIG. 10 shows a third process of a mixing method according to an example.
  • the sound protruding from the screen can be expressed as a result.
  • a graphical interface for front screen panning is described.
  • the parameters can be edited in nonlinear form by the graphical interface below.
  • the speed and acceleration of the sound source can be imparted by the graphic interface below, and sound effects on the sound source can be processed. Sound effects can include Doppler processing.
  • 11 illustrates an overall interface for determining values of one or more parameters for three-dimensional sound according to an example.
  • the overall interface 1100 may include at least some of the distance setting interface 1110, the source setting interface 1120, the distance interface 1130, the screen interface 1140, the speed interface 1150, and the sequence interface 1160. Can be.
  • the starting point of the source can be represented by S.
  • the end point of the sound source may be represented by E.
  • the starting point of the source may be outlined as the starting point.
  • the end point of the source can be outlined as the end point.
  • the distance setting interface 1110 includes a distance knob 1210, a distance value 1215, a threshold knob 1220, a threshold value 1225, a damper knob 1230, a damper value 1235, and a mix knob 1240. , A mix value 1245, and a limit reset button 1250.
  • the sound source may be formed behind the screen via the distance setting interface 1210.
  • the distance value 1215 can represent the maximum distance of the area behind the screen.
  • the distance knob 1210 can adjust the distance value 1215.
  • One or more parameters may include a maximum distance parameter that indicates a maximum distance of the area behind the screen.
  • the maximum distance parameter may have a value of 0 to 1000.
  • the default value of the maximum distance parameter may be zero.
  • the range of values of the start point and the end point of the source may be changed according to the maximum distance parameter. For example, when the start and end points of the source are located behind the screen, the distance from the screen of the start and end points of the source cannot be greater than the value of the maximum distance parameter.
  • the threshold value 1225 may represent the distance of the area in front of the screen. Threshold knob 1220 can adjust threshold value 1225.
  • One or more parameters may include a maximum threshold parameter that indicates a maximum distance of the area in front of the screen.
  • the maximum threshold parameter may have a value between 0 and 30.
  • the default value of the maximum threshold parameter may be zero.
  • the range of values of the start point and the end point of the source may be changed according to the maximum threshold parameter. For example, when the start and end points of the source are located in front of the screen, the distance from the screen of the start and end points of the source cannot be greater than the value of the maximum threshold parameter.
  • the damper value 1235 may represent the damper of the source.
  • the damper knob 1230 can adjust the damper value 1235.
  • the damper may exhibit a velocity effect.
  • One or more parameters may include a damper parameter indicating a damper of the source.
  • the damper parameter may have a value of 0 to 1.
  • the default value of the damper parameter may be zero.
  • the generator 520 may maximize the level effect and the reverberation effect of the 3D sound in consideration of the speed of the sound by using the damper parameter.
  • the generator 520 may perform the compression of the 3D sound by using the damper parameter, and may harden the 3D sound through the compression.
  • the generator 520 may cut the upper part and the lower part of the 3D sound by applying a filter to the 3D sound by using the damper parameter.
  • the size of the damper may be indicated as one or more concentric circles at speed interface 1150.
  • the mix value 1245 may represent a balance between the sound generated by the processing and the original sound.
  • the mix knob 1240 may adjust the mix value 1235.
  • One or more parameters may include a mix parameter indicative of a balance between the sound produced by the processing and the original sound.
  • the mix parameter may have a value between 0 and 100.
  • the default value of the damper parameter may be 100.
  • the mix parameter having a minimum value may indicate that only the sound generated by the processing is output.
  • a mix parameter having a maximum value may indicate that only the original, unprocessed sound is output. That is to say that the mix parameter has a maximum value may indicate that the sound is bypassed.
  • the generator 520 may adjust a balance between the sound generated by the processing and the original sound according to the value of the mix parameter.
  • the limit reset button may be used to return one or more parameters of the distance setting interface 1110 to the default value.
  • FIG. 13 illustrates a source setting interface according to an example.
  • the source setting interface 1120 includes a source start point knob 1310, a source start point value 1315, a source start point diffuse knob 1320, a source start point diffuse value 1325, a source end point knob 1330, a source It may include an endpoint value 1335, a source endpoint diffuse knob 1340, a source endpoint diffuse value 1345, and a source reset button 1350.
  • the position of the start point of the source, the position of the end point of the source, the diffuse of the start point of the source and the diffuse of the end point of the source may be determined.
  • the diffuse may indicate the degree of diffusion.
  • the value of diffuse can represent the radius of diffusion.
  • the source start point value 1315 may indicate the position of the start point of the source.
  • the source start point knob 1310 may adjust the source start point value 1315.
  • One or more parameters may include a source start point parameter that indicates the location of the start point of the source.
  • the location of the start point of the source may include the distance from the listening area to the start point of the source and / or the coordinates of the start point of the source.
  • the source start point parameter can optionally be activated.
  • the source start point knob 1310, the source start point value 1315, and the source start point parameter may be activated when the source start point is generated.
  • the starting point of the source may be generated by user manipulation at one of the distance interface 1130, the screen interface 1140, and the speed interface 1150.
  • the minimum value of the source start point parameter may be zero.
  • the maximum value of the source start point parameter may be the sum of the maximum distance parameter and the maximum threshold parameter.
  • the source start point diffuse value 1325 may represent the diffuse of the source start point.
  • the source start point diffuse knob 1320 may adjust the source start point diffuse value 1325.
  • One or more parameters may include a source start point diffuse parameter that indicates a diffuse of the start point of the source.
  • the diffuse at the beginning of the source may represent the degree of dispersion of the sound at the beginning of the source.
  • the source start point diffuse parameter can optionally be activated.
  • the source start point diffuse knob 1320, the source start point diffuse value 1325, and the source start point diffuse parameters may be activated when the source start point is generated.
  • the starting point of the source may be generated by user manipulation at one of the distance interface 1130, the screen interface 1140, and the speed interface 1150.
  • the source start point diffuse parameter may have a value of 0 to 100.
  • the default value of the source start point diffuse parameter may be zero.
  • the generator 520 may set the degree of dispersion of the sound at the start point of the source using the source start point diffuse parameter. The same delay can be applied to the distributed sound. The distributed sound may be reproduced by the proximity speaker 191.
  • the degree of dispersion by the source start point diffuse parameter may be displayed in the screen interface 1140.
  • the source endpoint value 1335 may indicate the location of the endpoint of the source.
  • the source endpoint knob 1330 may adjust the source endpoint value 1335.
  • One or more parameters may include a source endpoint parameter that indicates the location of the endpoint of the source.
  • the location of the end point of the source may include the distance from the listening area to the end point of the source and / or the coordinates of the end point of the source.
  • the source endpoint parameter can optionally be activated.
  • the source endpoint knob 1330, the source endpoint value 1335, and the source endpoint parameter may be activated when the endpoint of the source is generated.
  • the end point of the source may be generated by user manipulation at one of the distance interface 1130, the screen interface 1140, and the speed interface 1150.
  • the minimum value of the source endpoint parameter may be zero.
  • the maximum value of the source endpoint parameter may be the sum of the maximum distance parameter and the maximum threshold parameter.
  • the source endpoint diffuse value 1345 may represent the diffuse of the source endpoint.
  • the source endpoint diffuse knob 1340 may adjust the source endpoint diffuse value 1345.
  • One or more parameters may include a source endpoint diffuse parameter that indicates a diffuse of the endpoint of the source.
  • the diffuse at the end of the source may indicate the degree of dispersion of the sound at the end of the source.
  • the source endpoint diffuse parameter can optionally be activated.
  • the source endpoint diffuse knob 1340, the source endpoint diffuse value 1345, and the source endpoint diffuse parameters can be activated when the source endpoint is generated.
  • the end point of the source may be generated by user manipulation at one of the distance interface 1130, the screen interface 1140, and the speed interface 1150.
  • the source endpoint diffuse parameter may have a value between 0 and 100.
  • the default value of the source endpoint diffuse parameter may be zero.
  • the generator 520 may set the degree of dispersion of the sound at the end point of the source using the source end point diffuse parameter. The same delay can be applied to the distributed sound. The distributed sound may be reproduced by the proximity speaker 191.
  • the degree of dispersion by the source endpoint diffuse parameter may be displayed in the screen interface 1140.
  • the distance interface 1130 includes a maximum distance value 1410, a distance value control 1415, a maximum threshold value 1420, a threshold value control 1425, a unit 1430, a source start point 1440, a source start It may include a point distance 1445, a source end point 1450, a source end point distance 1455, a source trace line 1460, a screen line 1470, and a listening area 1480.
  • the distance interface 1130 may visually show the parameters set in the distance setting interface 1110.
  • the distance interface 1130 may reflect the increase or decrease of the unit without changing the appearance.
  • the maximum distance value 1410 may represent a maximum distance value inside the screen.
  • the inside of the screen may represent an area behind the screen based on the viewer.
  • the maximum distance value 1410 may represent a value of the maximum distance parameter.
  • the distance value into the interior of the screen of the source start point 1240 and the distance value into the interior of the screen of the source end point 1250 may be limited to below the maximum distance value 1410.
  • the distance interface 110 may increase or decrease the unit of the maximum distance value 1410.
  • the unit of the maximum distance value 1410 may be meters or feet.
  • the distance value control 1415 may be appropriately spaced scales generated according to the maximum distance value 1410.
  • the maximum threshold value 1420 may represent a maximum distance value outside of the screen.
  • the outside of the screen may represent an area in front of the screen based on the viewer.
  • the maximum threshold value 1420 may represent a value of the maximum threshold parameter.
  • the distance value to the outside of the screen of the source start point 1440 and the distance value to the outside of the screen of the source end point 1450 may be limited to below the maximum distance value 1420.
  • the distance interface 1130 may increase or decrease the unit of the maximum threshold value 1420.
  • the unit of the maximum threshold value 1420 may be meters or feet.
  • Threshold value control 1425 may be appropriately spaced scales generated according to maximum threshold value 1420.
  • the unit 1430 may indicate a value displayed on the distance interface 1130 or a unit of parameters displayed on the distance interface 1130. Through a global setting, one of the meters and feet can be selected as the unit. The default value of the unit may be meters.
  • the source start point 1440 may indicate the start point of the source.
  • the source start point 1440 may indicate a value of a source start point parameter.
  • the user may generate the source start point 1440 by manipulating the input device within the dotted line of the area of the distance interface 1130.
  • the location specified by the manipulation of the input device may represent the source start point 1440.
  • a starting point of the source may be generated, and a source starting point knob 1310, a source starting point value 1315, a source starting point diffuse knob 1320, a source starting point diffuse value 1325 and the source start point diffuse parameter may be activated.
  • the input device can include a keyboard and / or a mouse. Operation of the input device may include clicking, double-clicking, dragging, and dragging and dropping of a mouse, and may include pressing a specific key of the keyboard.
  • the source start point 1440 may display an X coordinate and a Y coordinate among coordinates of the start point of the source.
  • the value of the Z coordinate of the start point of the source may not be reflected at the source start point 1440.
  • the user may move the source start point 1440 to a desired position through manipulation of the input device.
  • the user may delete the source start point 1440 through an operation of the input device.
  • the source start point distance 1445 can represent the distance between the screen and the source start point 1440.
  • the source start point distance 1445 may represent a value of a source start point parameter.
  • the source start point distance 1445 may be displayed as a horizontal line and may be displayed as a value located at the end of the horizontal line.
  • the source endpoint 1450 may represent an endpoint of the source.
  • the source endpoint 1450 may represent a value of the source endpoint parameter.
  • the user may generate the source endpoint 1450 by manipulating the input device within a dashed line of the area of the distance interface 1130.
  • the location specified by the manipulation of the input device may represent the source end point 1450.
  • Source end point 1450 may be generated after generation of source start point 1440.
  • the user may generate a source start point 1440 and a source end point 1450 through manipulation of the input device.
  • the source end point 1450 may be automatically generated. In this case, the location of the generated source endpoint 1450 may be the center of the screen line 1470.
  • an end point of the source may be generated, the source end point knob 1330, the source end point value 1335, the source end point diffuse knob 1340, and the source end point diffuse value. 1345 and the source endpoint diffuse parameter may be activated.
  • the source end point 1450 may display an X coordinate and a Y coordinate among coordinates of the end point of the source.
  • the value of the Z coordinate of the end point of the source may not be reflected in the source end point 1450.
  • the user may move the source end point 1450 to a desired position through manipulation of the input device.
  • the user may delete the source endpoint 1450 through manipulation of the input device.
  • the source endpoint distance 1455 may represent the distance between the screen and the source endpoint 1450.
  • the source endpoint distance 1455 may represent a value of the source endpoint parameter.
  • the source end point distance 1455 may be displayed as a horizontal line and may be displayed as a value located at the end of the horizontal line.
  • the source trace line 1460 may represent a line along which the source moves from the start point of the source to the end point of the source.
  • One or more parameters may include a source trace line parameter indicating the line through which the source travels from the start point of the source to the end point of the source.
  • the line through which the source moves may include a plurality of lines.
  • the source trace line parameter may indicate a plurality of lines connecting the start point of the source and the end point of the source.
  • the generator 520 may move the position of the source using the source trace line parameter.
  • the user may change the shape of the source trace line 1460 through manipulation of the input device. For example, the user may divide one line of the source trace line 1460 into two connected lines through manipulation of the input device.
  • the distance interface 1130 may indicate a point moving along the source trace line 1460 according to the processing length set in the sequence interface 1160.
  • the screen line 1470 may indicate a point where a screen on which an image is displayed is located.
  • the screen line 1470 may be a reference for the distance of the source start point 1440 and the distance of the source end point 1450.
  • the listening area 1480 may represent an area where the viewer is located. Also, the listening area may indicate the direction of the viewer.
  • the user can select one point of source trace line 1460.
  • the selected point 1510 may represent a point selected by the user. When a point is selected by the user, a distance value of the selected point may be displayed.
  • the distance value 1520 may represent a distance value of a point selected by the user. In addition, the distance value may represent the distance between the selected point and the screen.
  • FIG. 16 illustrates editing a source trace line according to an example.
  • a user can edit source trace line 1460.
  • the value of the source trace line parameter may be set by editing the source trace line 1160.
  • the user may move the position of the selected point 1510 through manipulation of the input device.
  • the moved selected point 1610 is shown.
  • the source trace line 1460 may change.
  • the line in which the selected point 1510 of one or more lines of the source trace line 1460 is located is two lines in which the moved selected point 1610 is the end point as the selected point 1510 moves.
  • the first of the two lines may be a line from the start point of the line where the selected point 1510 is located to the selected point 1610 moved.
  • the second of the two lines may be a line from the moved selected point 1610 to the end point of the line where the selected point 1510 is located.
  • the setting unit 510 may set the value of the source trace line parameter to reflect the division of the line.
  • FIG. 17 illustrates a screen interface according to an example.
  • the screen interface 1140 may provide editing of the value of the X coordinate and the value of the Y coordinate for the start point and the end point of the source.
  • the screen interface 1140 may display the value of the X coordinate and the value of the Y coordinate among the coordinates of the start point of the source, and display the value of the X coordinate and the value of the Y coordinate among the coordinates of the end point of the source. have.
  • the screen interface 1140 may include a panning area 1710, a source start point 1720, a source end point 1730, a diffuse circle 1740, and a source trace line 1750.
  • the panning area 1710 may represent an area required for displaying the source start point 1720 and the source end point 1730.
  • the source start point 1720 may represent the coordinates of the start point of the source.
  • the source start point 1720 may indicate an X coordinate and a Y coordinate among coordinates of the start point of the source.
  • the function of the source start point 1720 may correspond to the function of the source start point 1440 described above.
  • the source end point 1730 may represent the coordinates of the end point of the source.
  • the source end point 1730 may indicate an X coordinate and a Y coordinate among coordinates of the end point of the source.
  • the function of the source endpoint 1730 may correspond to the function of the source endpoint 1450 described above.
  • the radius of the diffuse circle 1740 may represent the value of the diffuse of the start point of the source and / or the end point of the source.
  • the radius of the diffuse circle 1740 may represent only the value of the diffuse of the start point of the source.
  • the value of the diffuse of the start point of the source may represent the value of the source start point diffuse parameter.
  • the radius of the diffuse circle 1740 may represent only the value of the diffuse of the end point of the source.
  • the value of the diffuse of the end point of the source may represent the value of the source endpoint diffuse parameter.
  • the source trace line 1750 may indicate a line from which the source moves from the start point of the source to the end point of the source.
  • the function of the source trace line 1750 may correspond to the function of the source trace line 1460 described above.
  • Velocity interface 1150 may provide editing of the value of the Y coordinate and the value of the Z coordinate for the start point of the source and the end point of the source.
  • the velocity interface 1150 may display the value of the Y coordinate and the value of the Z coordinate among the coordinates of the start point of the source, and display the value of the Y coordinate and the value of the Z coordinate among the coordinates of the end point of the source. have.
  • Speed interface 1150 includes screen line 1810, maximum distance value 1820, distance value control 1825, maximum threshold value 1830, threshold value control 1835, listening area 1840, damper value 1850, source start point 1860, source end point 1870, and source trace line 1880.
  • Screen line 1810 may correspond to screen line 1470 described above. However, screen line 1470 may be displayed for the X and Y coordinates, while screen line 1810 may be displayed for the Y and Z coordinates.
  • the maximum distance value 1820 may correspond to the maximum distance value 1410 described above. However, the maximum distance value 1410 may be displayed for the X coordinate and the Y coordinate, while the maximum distance value 1820 may be displayed for the Y coordinate and the Z coordinate.
  • Distance value control 1825 may correspond to distance value control 1415 described above. However, distance value control 1415 may be displayed for X and Y coordinates, while distance value control 1825 may be displayed for Y and Z coordinates.
  • the maximum threshold value 1830 may correspond to the maximum threshold value 1420 described above. However, the maximum threshold value 1420 may be displayed for the X coordinate and the Y coordinate, while the maximum threshold value 1830 may be displayed for the Y coordinate and the Z coordinate.
  • Threshold value control 1835 may correspond to threshold value control 1425 described above. However, the threshold value control 1425 can be displayed for the X and Y coordinates, while the threshold value control 1835 can be displayed for the Y and Z coordinates.
  • the listening area 1840 may correspond to the aforementioned listening area 1480. However, while the listening area 1480 is displayed with respect to the X coordinate and the Y coordinate, the listening area 1840 may be displayed with respect to the Y coordinate and the Z coordinate.
  • the damper display unit 1850 may indicate the above-described damper value.
  • the shape of the damper display unit may indicate a value of the damper parameter.
  • the damper indicator 1850 may include one or more concentric circles. One or more concentric circles may be shown only in part. The number, shape and radius of one or more concentric circles may represent a damper applied to the source and may vary depending on the value of the damper parameter. For example, an ellipse may be formed based on the threshold line as the value of the damper parameter increases.
  • the source start point 1860 may correspond to the source start point 1440 described above. However, the source start point 1440 may be displayed for the X and Y coordinates, while the source start point 1860 may be displayed for the Y and Z coordinates.
  • Source endpoint 1870 may correspond to source endpoint 1450 described above. However, the source end point 1450 may be displayed for the X and Y coordinates, while the source end point 1870 may be displayed for the Y and Z coordinates.
  • Source trace line 1880 may correspond to source trace line 1460 described above. However, the source trace line 1460 may be displayed for the X and Y coordinates, while the source trace line 1880 may be displayed for the Y and Z coordinates.
  • the sequence interface 1160 may provide editing of the value of the Y coordinate and the value of the Z coordinate for the start point and the end point of the source.
  • the sequence interface 1160 may be linked with the source start point 1440 and the source end point 1450 of the distance interface 1130.
  • the information about the source displayed in the sequence interface 1160 may be displayed from the Y coordinate and the Z coordinate of the source start point 1440, and may be displayed from the Y coordinate and the Z coordinate of the source end point 1450. .
  • sequence interface 1160 may indicate a damper at an end point of the source.
  • the sequence interface 1160 includes a timeline 1910, a project cursor 1915, a sequence marker 1920, a time stretch switch 1925, a speed length 1930, a time stretch grid line 1935, a speed graph 1940.
  • Gain graph (1945), tail (1950), tail (1950), velocity length (1955), timeline speed (1960), timeline gain (1965), tail length (1970), tail end A point 1975 and sequence editing area 1980 may be included.
  • the timeline 1910 may represent a timeline according to a length set in the sequence interface 1160.
  • the timeline may represent the flow of time from the time of the start point of the source to the time of the end point of the source.
  • the start point and the end point of the timeline 1910 may be set to be the same as the start point and the end point of the locator area of the projector.
  • Timeline 1910 may represent an application for pre-roll and / or post-roll.
  • the time at which the preroll and / or postroll is applied can be set globally.
  • the project cursor 1915 may operate in synchronization with the cursor of the host program. For example, when the cursor of the host program moves, the project cursor 2015 may also move along the cursor of the host program.
  • One or more parameters may have different values over time.
  • the point indicated by the project cursor 1915 in the timeline 1910 may indicate a reference time in displaying a parameter value.
  • the sequence marker 1920 may be a button for setting a start point and an end point of the timeline 1910.
  • the timeline 1910 may be set in the same manner as the locator of the project.
  • the time of the start point locator may be set to the start point of the timeline 1910
  • the time of the end point locator may be set to the end point of the timeline 1910.
  • sequence marker 1920 may be deactivated.
  • the time stretch switch 1925 can create an additional timeline on the line of velocity length 1930.
  • the user can set the time stretch switch 1925 to ON through manipulation of the input device. If time stretch switch 1925 is set to on, an additional time line may be created in the line of speed length 1930.
  • the user can adjust the interval of the additional timeline by operating the input device. As the interval of additional timelines becomes wider, the passage of time can be faster. As the intervals of additional timelines become narrower, the passage of time may be slower.
  • the project cursor 1915 can reflect the passage of time based on the additional timeline.
  • the user may set the time stretch switch 1925 to OFF through manipulation of the input device. If time stretch switch 1925 is set to off, an additional time line may disappear from the line of speed length 1930. If the additional timeline disappears, the project cursor 1915 may reflect the passage of time relative to the timeline 1910.
  • the velocity length 1930 may represent the length of time that the process takes.
  • the length of velocity length 1930 may be the same as the length of timeline 1910.
  • the starting point of velocity length 1930 may correspond to the starting point locator of the project.
  • the end point of velocity length 1930 may correspond to the end point locator of the project.
  • the velocity length 1930 may have a value corresponding to the length generated by the sequence marker 1920.
  • the time stretch grid line 1935 can indicate the extent of stretch in time.
  • the time stretch grid line 1935 may be generated by manipulation of a user's input device.
  • the user may manipulate the input device to change the position of the points of the time stretch grid line 1935. As the location of the point changes, the spacing of the grid lines may change.
  • the speed graph 1940 may represent the speed of the source and the acceleration of the source along the timeline.
  • the speed graph 1940 may represent a change in the speed of the source from the start point of the source to the end point of the source.
  • One or more parameters may include a speed parameter indicative of the speed of the source and may include an acceleration parameter indicative of the acceleration of the source.
  • the speed parameter may indicate the speed of the source over time.
  • the acceleration parameter may represent the acceleration of the source over time.
  • the one or more parameters may include speed and acceleration parameters indicative of the speed and acceleration of the source.
  • the generator 520 may generate the 3D sound by reflecting the speed of the source using the speed parameter. In addition, the generator 520 may generate the 3D sound by reflecting the acceleration of the source using the acceleration parameter. Alternatively, the generator 520 may generate the 3D sound by reflecting the speed and acceleration of the source using the speed and acceleration parameters.
  • the user may enlarge the speed graph 2040 by manipulating the input device.
  • the speed graph 1940 may automatically calculate the speed of the source linearly over time once the start and end points are determined, and display the calculated speed of the source as a graph.
  • the height of a point may represent the speed of the source at the time corresponding to the point.
  • the slope of a point may represent the acceleration of the source at the time corresponding to the point.
  • the user may move one point on the speed graph to another position by manipulating the input device.
  • the shape of the graph may change, and the speed of the source may change according to the changed shape of the speed graph.
  • the gain graph 1945 may represent a change in gain of the source along the timeline.
  • the gain graph 1945 may represent a change in gain from the start point of the source to the end point of the source.
  • One or more parameters may include a gain parameter indicative of the gain of the source.
  • the gain parameter may represent a gain of a source over time.
  • the generator 520 may generate the 3D sound by reflecting the gain of the source using the gain parameter.
  • the gain graph 1945 may automatically calculate the gain of the source linearly over time when the start point and the end point are determined, and display the calculated gain of the source as a graph.
  • the height of a point may represent the gain of the source at the time corresponding to the point.
  • the user can apply the value of the speed of the source to the gain of the source through the manipulation of the input device.
  • the tail graph 1950 may represent the tail of the source.
  • the tail of the source may occur after the end point of the source. Basically, 3 seconds of reverberation can occur as the tail.
  • One or more parameters may include a tail parameter indicating a tail of the source.
  • the tail parameter may indicate the tail of the source.
  • the generator 520 may generate the 3D sound by reflecting the tail of the source using the tail parameter.
  • the tail graph 1950 may operate regardless of the project cursor 1915.
  • the user may change the shape of the tail through manipulation of the input device.
  • the tail may change cyclically into a plurality of predefined shapes.
  • the setting unit 510 may set a value of the tail parameter according to the changed shape.
  • the velocity length value 1955 may represent the time used for the representation of velocity.
  • the timeline speed value 1945 may indicate a speed at a point indicated by the project cursor 1915 of the timeline 1910.
  • Timeline Gain Value 1965 The gain at the point indicated by the project cursor 1915 of the timeline 1910 may be represented.
  • Tail length 1970 may represent the reverberation time of the tail.
  • Tail end point 1975 may represent the end point of the tail.
  • the user may move the position of the end point of the tail through manipulation of the input device.
  • the setting unit 510 may set the value of the tail parameter by reflecting the position of the moved end point.
  • the length of the tail may change depending on the position of the end point to which the tail is moved.
  • the changed tail length can be indicated at tail length 1970.
  • the sequence editing area 1980 may be an area that provides editing for the sequence.
  • the user can manipulate the input device to select one of speed, gain and tail. As one of the speed, the gain and the tail is selected, the size of the edit window for editing the selected object may increase. If a specific object is not selected, an editing window of the same size may be provided for the objects.
  • 20 illustrates a locator area of a project according to one embodiment.
  • the locator area 2000 may include a start point locator 2010, an end point locator 2020, and a project cursor 2040.
  • the time of the start point locator may be set to the start point of the timeline 1910 and the time of the end point locator may be set to the end point of the timeline 1910.
  • FIG. 21 illustrates an enlarged speed graph according to an example.
  • an enlarged speed graph 2100 is shown enlarged in the sequence editing area 1980.
  • the point 2110 selected as the object of editing due to the manipulation of the input device by the user is shown, and the current speed 2120 is shown.
  • the current speed 2120 can be displayed as the height of the graph at the point where the project cursor 1915 is located. Also, the current speed 2120 may be displayed at the speed length value 1955.
  • FIG. 22 illustrates an electronic device implementing a 3D sound reproducing apparatus according to an embodiment.
  • the 3D sound reproducing apparatus 100 may be implemented as the electronic device 2200 illustrated in FIG. 22.
  • the electronic device 2200 may be a general purpose computer system that operates as the 3D sound reproducing apparatus 100.
  • the electronic device 2200 may include at least a portion of a processor 2221, a network interface 2229, a memory 2223, a storage 2228, and a bus 2222.
  • Components of the electronic device 2200 such as the processor 2221, the network interface 2229, the memory 2223, the storage 2228, and the like, may communicate with each other through the bus 2222.
  • the processor 2221 may be a semiconductor device that executes processing instructions stored in the memory 2223 or the storage 2228.
  • the processor 2221 may process a task required for the operation of the electronic device 2200.
  • the processor 2221 may execute code of an operation or step of the processor 2221 described in the embodiments.
  • the network interface 2229 may be connected to the network 2230.
  • the network interface 2229 may receive data or information required for the operation of the electronic device 2200, and may transmit data or information required for the operation of the electronic device 2200.
  • the network interface 2229 may transmit data to and receive data from other devices via the network 2230.
  • the network interface 2229 may be a network chip or port.
  • Memory 2223 and storage 2228 may be various forms of volatile or nonvolatile storage media.
  • the memory 2223 may include at least one of a ROM 2224 and a RAM 2225.
  • the storage 2228 may include internal storage media such as RAM, flash memory, hard disk, and the like, and may include removable storage media such as a memory card.
  • the electronic device 2200 may further include a user interface (UI) input device 2226 and a UI output device 2227.
  • the UI input device 2226 may receive a user input required for the operation of the electronic device 2200.
  • the UI output device 2227 may output information or data according to the operation of the electronic device 2200.
  • the function or operation of the electronic device 2200 may be performed as the processor 2221 executes at least one program module.
  • the memory 2223 and / or the storage 2228 may store at least one program module.
  • At least one program module may be configured to be executed by the processor 2221.
  • At least one program module may include a signal detector 110, a primary signal processor 120, a channel allocator 130, a secondary signal processor 140, a function controller 150, and a sound image externalization implementer 160. ), A bypass adjusting unit 170, a remote speaker detection / playback unit 180 and a proximity speaker detection / playback unit 190 may be included.
  • the UI input device 2226 can include a bypass switch 171.
  • the network interface 2229 may include a signal receiver 105.
  • FIG. 23 is a diagram illustrating an electronic device that implements a 3D sound providing apparatus according to an embodiment.
  • the 3D sound providing apparatus 500 may be implemented as the electronic device 2300 illustrated in FIG. 23.
  • the electronic device 2300 may be a general purpose computer system that operates as the 3D sound providing device 500.
  • the electronic device 2300 may include at least a portion of a processor 2321, a network interface 2333, a memory 2323, a storage 2328, and a bus 2322.
  • Components of the electronic device 2300 such as the processor 2321, the network interface 2329, the memory 2323, the storage 2328, and the like, may communicate with each other through the bus 2232.
  • the processor 2321 may be a semiconductor device that executes processing instructions stored in the memory 2323 or the storage 2328.
  • the processor 2321 may process a task required for the operation of the electronic device 2300.
  • the processor 2321 may execute code of an operation or step of the processor 2321 described in the embodiments.
  • the network interface 2329 may be connected to the network 2330.
  • the network interface 2329 may receive data or information required for the operation of the electronic device 2 # 00, and may transmit data or information required for the operation of the electronic device 2300.
  • the network interface 2329 may transmit data to and receive data from other devices through the network 2330.
  • the network interface 2329 may be a network chip or a port.
  • the memory 2323 and the storage 2328 may be various forms of volatile or nonvolatile storage media.
  • the memory 2323 may include at least one of a ROM 2324 and a RAM 2325.
  • the storage 2328 may include built-in storage media such as RAM, flash memory, hard disk, and the like, and may include removable storage media such as a memory card.
  • the electronic device 2300 may further include a user interface (UI) input device 2326 and a UI output device 2327.
  • the UI input device 2326 may receive a user input required for the operation of the electronic device 2300.
  • the UI output device 2327 may output information or data according to the operation of the electronic device 2300.
  • the function or operation of the electronic device 2300 may be performed as the processor 2321 executes at least one program module.
  • the memory 2323 and / or the storage 2328 may store at least one program module.
  • At least one program module may be configured to be executed by the processor 2321.
  • At least one program module may include a setting unit 510 and a generating unit 520.
  • the network interface 2329 may include an output unit 530.
  • the apparatus described above may be implemented as a hardware component, a software component, and / or a combination of hardware components and software components.
  • the devices and components described in the embodiments may be, for example, processors, controllers, arithmetic logic units (ALUs), digital signal processors, microcomputers, field programmable arrays (FPAs), It may be implemented using one or more general purpose or special purpose computers, such as a programmable logic unit (PLU), microprocessor, or any other device capable of executing and responding to instructions.
  • the processing device may execute an operating system (OS) and one or more software applications running on the operating system.
  • the processing device may also access, store, manipulate, process, and generate data in response to the execution of the software.
  • OS operating system
  • the processing device may also access, store, manipulate, process, and generate data in response to the execution of the software.
  • processing device includes a plurality of processing elements and / or a plurality of types of processing elements. It can be seen that it may include.
  • the processing device may include a plurality of processors or one processor and one controller.
  • other processing configurations are possible, such as parallel processors.
  • the software may include a computer program, code, instructions, or a combination of one or more of the above, and configure the processing device to operate as desired, or process it independently or collectively. You can command the device.
  • Software and / or data may be any type of machine, component, physical device, virtual equipment, computer storage medium or device in order to be interpreted by or to provide instructions or data to the processing device. Or may be permanently or temporarily embodied in a signal wave to be transmitted.
  • the software may be distributed over networked computer systems so that they may be stored or executed in a distributed manner.
  • Software and data may be stored on one or more computer readable recording media.
  • the method according to the embodiment may be embodied in the form of program instructions that can be executed by various computer means and recorded in a computer readable medium.
  • the computer readable medium may include program instructions, data files, data structures, etc. alone or in combination.
  • the program instructions recorded on the media may be those specially designed and constructed for the purposes of the embodiments, or they may be of the kind well-known and available to those having skill in the computer software arts.
  • Examples of computer-readable recording media include magnetic media such as hard disks, floppy disks, and magnetic tape, optical media such as CD-ROMs, DVDs, and magnetic disks, such as floppy disks.
  • Examples of program instructions include not only machine code generated by a compiler, but also high-level language code that can be executed by a computer using an interpreter or the like.
  • the hardware device described above may be configured to operate as one or more software modules to perform the operations of the embodiments, and vice versa.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)

Abstract

L'invention concerne un procédé et un dispositif de modification et de reproduction d'un son tridimensionnel. Un dispositif de reproduction d'un son tridimensionnel dispose d'une fonction de modification associée à des valeurs d'un ou plusieurs paramètres de génération du son tridimensionnel. Un utilisateur du dispositif de reproduction d'un son tridimensionnel peut régler les valeurs des paramètres par l'intermédiaire d'une interface graphique. Une propriété d'une source d'un son positionnée dans un espace tridimensionnel peut être réglée par l'intermédiaire des valeurs des paramètres. Le dispositif de reproduction d'un son tridimensionnel génère des données d'un son tridimensionnel sur la base des valeurs des paramètres.
PCT/KR2016/002826 2015-03-19 2016-03-21 Procédé et dispositif de modification et de reproduction d'un son tridimensionnel WO2016148553A2 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR20150038372 2015-03-19
KR10-2015-0038372 2015-03-19
KR1020160032468A KR20160113036A (ko) 2015-03-19 2016-03-18 3차원 사운드를 편집 및 제공하는 방법 및 장치
KR10-2016-0032468 2016-03-18

Publications (2)

Publication Number Publication Date
WO2016148553A2 true WO2016148553A2 (fr) 2016-09-22
WO2016148553A3 WO2016148553A3 (fr) 2016-11-10

Family

ID=56919009

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2016/002826 WO2016148553A2 (fr) 2015-03-19 2016-03-21 Procédé et dispositif de modification et de reproduction d'un son tridimensionnel

Country Status (1)

Country Link
WO (1) WO2016148553A2 (fr)

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20080093422A (ko) * 2006-02-09 2008-10-21 엘지전자 주식회사 오브젝트 기반 오디오 신호의 부호화 및 복호화 방법과 그장치
US8965000B2 (en) * 2008-12-19 2015-02-24 Dolby International Ab Method and apparatus for applying reverb to a multi-channel audio signal using spatial cue parameters
US8396576B2 (en) * 2009-08-14 2013-03-12 Dts Llc System for adaptively streaming audio objects
KR102548756B1 (ko) * 2011-07-01 2023-06-29 돌비 레버러토리즈 라이쎈싱 코오포레이션 향상된 3d 오디오 오서링과 렌더링을 위한 시스템 및 툴들
US9118999B2 (en) * 2011-07-01 2015-08-25 Dolby Laboratories Licensing Corporation Equalization of speaker arrays

Also Published As

Publication number Publication date
WO2016148553A3 (fr) 2016-11-10

Similar Documents

Publication Publication Date Title
WO2016024847A1 (fr) Procédé et dispositif de génération et de lecture de signal audio
WO2015199508A1 (fr) Procédé et dispositif permettant de restituer un signal acoustique, et support d'enregistrement lisible par ordinateur
WO2016104988A9 (fr) Terminal mobile, dispositif de sortie audio et système de sortie audio constitué de ces derniers
WO2009131391A1 (fr) Procédé de génération et de lecture de contenus audio basés sur un objet et support d'enregistrement lisible par ordinateur pour l'enregistrement de données présentant une structure de format fichier pour un service audio basé sur un objet
WO2016017945A1 (fr) Dispositif mobile et son procédé d'appariement à un dispositif électronique
WO2010008234A2 (fr) Procédé et appareil de représentation d'effets sensoriels, et support d'enregistrement lisible par ordinateur sur lequel sont enregistrées des métadonnées concernant la performance d'un dispositif sensoriel
WO2019103584A1 (fr) Dispositif de mise en oeuvre de son multicanal utilisant des écouteurs à oreille ouverte et procédé associé
WO2017010651A1 (fr) Système d'affichage
WO2010120137A2 (fr) Procédé et appareil de fourniture de métadonnées pour un support d'enregistrement lisible par ordinateur à effet sensoriel sur lequel sont enregistrées des métadonnées à effet sensoriel, et procédé et appareil de reproduction sensorielle
WO2010008235A2 (fr) Procédé et appareil d'expression d'effets sensoriels, et support d'enregistrement lisible par ordinateur sur lequel sont enregistrées des métadonnées concernant la commande d'un dispositif sensoriel
WO2013022222A2 (fr) Procédé de commande d'appareil électronique basé sur la reconnaissance de mouvement, et appareil appliquant ce procédé
WO2010008232A2 (fr) Procédé d'expression d'effets sensoriels et appareil associé, et support d'enregistrement lisible par ordinateur sur lequel des métadonnées associées à un effet sensoriel sont enregistrées
WO2010008233A2 (fr) Procédé d'expression d'effets sensoriels et appareil associé, et support d'enregistrement lisible par ordinateur sur lequel des métadonnées associées à des informations d'environnement d'utilisateur sont enregistrées
WO2016093510A1 (fr) Appareil d'affichage et procédé d'affichage
WO2020231202A1 (fr) Dispositif électronique comprenant une pluralité de haut-parleurs et son procédé de commande
WO2016104952A1 (fr) Appareil d'affichage et procédé d'affichage
WO2016182133A1 (fr) Dispositif d'affichage et son procédé de fonctionnement
WO2020145659A1 (fr) Dispositif de traitement de signal et appareil d'affichage d'image le comprenant
WO2019117409A1 (fr) Serveur central et système de représentation théâtrale comprenant celui-ci
WO2018139884A1 (fr) Procédé de traitement audio vr et équipement correspondant
WO2019050200A1 (fr) Appareil et procédé de traitement d'image
WO2019031652A1 (fr) Procédé de lecture audio tridimensionnelle et appareil de lecture
WO2019125036A1 (fr) Procédé de traitement d'image et appareil d'affichage associé
WO2023163322A1 (fr) Dispositif d'affichage d'image et serveur
WO2016182124A1 (fr) Dispositif d'affichage et procédé de fonctionnement correspondant

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16765309

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase in:

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16765309

Country of ref document: EP

Kind code of ref document: A2