CN106060758A - Processing method for virtual reality sound field metadata - Google Patents

Processing method for virtual reality sound field metadata Download PDF

Info

Publication number
CN106060758A
CN106060758A CN201610391252.0A CN201610391252A CN106060758A CN 106060758 A CN106060758 A CN 106060758A CN 201610391252 A CN201610391252 A CN 201610391252A CN 106060758 A CN106060758 A CN 106060758A
Authority
CN
China
Prior art keywords
audio object
point
moment
virtual reality
processing method
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610391252.0A
Other languages
Chinese (zh)
Other versions
CN106060758B (en
Inventor
杨洪旭
孙学京
张晨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Tuoling Inc
Original Assignee
Beijing Tuoling Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Tuoling Inc filed Critical Beijing Tuoling Inc
Priority to CN201610391252.0A priority Critical patent/CN106060758B/en
Publication of CN106060758A publication Critical patent/CN106060758A/en
Application granted granted Critical
Publication of CN106060758B publication Critical patent/CN106060758B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S5/00Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation 
    • H04S5/005Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation  of the pseudo five- or more-channel type, e.g. virtual surround

Abstract

The invention discloses a processing method for virtual reality sound field metadata. The processing method comprises the following steps of judging a movement mode of an audio object, and if the movement mode of the audio object is linear movement, setting the parameter m of the movement mode to be 0; if the movement mode of the audio object is curvilinear movement, setting the parameter m of the movement mode to be 1; when m is equal to 0, using a rectangular coordinate system to express the azimuth information of the audio object, and meanwhile processing the movement track of the audio object in the rectangular coordinate system; when m is equal to 1, using a polar coordinate system to express the azimuth information of the audio object, and meanwhile processing the movement track of the audio object in the polar coordinate system; and generating virtual surround according to location information corresponding to the movement track of the audio object. According to the processing method, the movement direction and track of the audio object can be more perfectly displayed when the complex object simultaneously performing linear movement and curvilinear movement is processed. The sense of immersion is improved in the virtual reality, and the sound is more real, more specific and more vivid.

Description

The processing method of virtual reality sound field metadata
Technical field
The present invention relates to signal processing technology field, be specifically related to the processing method of a kind of virtual reality sound field metadata.
Background technology
When presenting content with virtual reality helmet (Head-Mounted Display, HMD) to user, in audio frequency Hold and play to user by stereophone.At this moment need to face the problem how improving virtual surrounding sound effect.Virtual existing In real application, when playing audio content by stereophone, the purpose of virtual 3D audio frequency is intended to reach a kind of effect, allows User is just as listening with loudspeaker array (such as 5.1 or 7.1), even true as listening the sound in reality.
When making virtual reality audio content, generally there are some sources of sound to be distributed in different positions and constitute together with environment Sound field.Follow the tracks of user's headwork (head tracking), sound is processed accordingly.Such as, if original sound quilt User is perceived as from dead ahead, when, after user's rotary head to the left 90 degree, sound should be processed so that user's perception sound is from just 90 degree, right.
Here virtual reality device can have many types, the display device that such as headed is followed the tracks of, or simply one The stereophone of portion's headed tracking transducer.
Realize head tracking and also have multiple method.Relatively common is to use multiple sensors.Motion sensor external member is led to Often include accelerometer, gyroscope and magnetometric sensor.In terms of motion tracking and absolute direction, every kind of sensor has oneself Intrinsic strong point and weakness.Therefore practices well is to use sensor " to merge " (sensor fusion), will be from each sensor Signal combine, produce a more accurate motion detection result.
After obtaining end rotation angle, need sound is changed accordingly.Generate virtual reality sound field and have several Method, a kind of way is to audio object, according to its positional information, uses corresponding HRTF (Head Related Transfer Function, head related transfer function) wave filter is filtered, and obtains virtual surround sound.HRTF is in the name corresponding to time domain Title is HRIR (Head Related Impulse Response).Another way is that sound forwards to the B in ambisonic territory Format signal, is converted to virtual speaker array signal by this B format signal, is filtered by HRTF by virtual speaker array signal Ripple device is filtered, and obtains virtual surround sound.
It will be seen that both approaches is required for definition, designs and process audio object azimuth information in space.One As way be by the position rectangular coordinate system in space (x, y, z) represent.Assume that whole sound field space is with a limit A length of 2 cubes represent (2 × 2 × 2), and the head of hearer is at cubical central point, and supposes that this central point is at rectangular coordinate The coordinate of system is (0,0,0), as shown in Figure 1, it is assumed that the coordinate of the audio object in this sound field is (xi, yi, zi), then understands such as Lower example:
The audio object coordinate at sound field edge, positive left side is (-1,0,0);
The audio object coordinate at dead ahead sound field edge is (0,1,0);
The audio object coordinate at sound field edge, surface is (0,0,1);
The audio object coordinate at sound field edge, right front is (1,1,0);
As in figure 2 it is shown, wherein three components of coordinate must meet:
-1≤xi≤1
-1≤yi≤1
-1≤zi≤1
The definition method of this audio object azimuth information is very suitable for the object of linearly orbiting motion, such as object (x2, y2, the z2) in t2 moment it is linearly moved to, then any moment ti pair between t1 and t2 from (x1, y1, the z1) in t1 moment The position Point_i of the object answered can be expressed as by linear interpolation:
Point_i=Point_1+ (Point_2-Point_1) × (ti-t1)/(t2-t1);
Wherein Point_i is defined as: (xi, yi, zi);Point_1 is: (x1, y1, z1);Point_2 is: (x2, y2, z2);
But, if audio object is riding, the metadata of this azimuth information based on rectangular coordinate system is fixed Justice and interpolation method are not the most the most suitable.Such as audio object rotates a circle around central point, respectively at labelling all around 1,2,3,4 totally 4 azimuthal points (see Fig. 3):
The most front Point_1 (0,1,0)
The rightest Point_2 (1,0,0)
Point_3 (0 ,-1,0) after just
The most left Point_4 (-1,0,0)
The movement locus that actual interpolation obtains not is a circular motion, but a broken line motion, track is One rhombus, as it is shown on figure 3, the difference of this movement locus can cause the distortion of virtual reality audio content, affects audio frequency effect Really.
For solving this technical problem in prior art, use on time dimension, take the sampled point than comparatively dense, Carry out approximate simulation curvilinear motion.But disadvantage of this is that the curve of track simply approximation, and imperfect, and the most intensive adopts Sampling point causes amount of metadata too big, occupies more bandwidth.
Summary of the invention
It is an object of the invention to provide the processing method of a kind of virtual reality sound field metadata, be used for describing virtual reality Audio object curvilinear motion track in sound field, makes the in hgher efficiency of metadata, makes audio object do curve fortune in space simultaneously Time dynamic, track is smoother, uses hybrid (mixing) mode further, compatible with straight line or the curve of audio object simultaneously Motion.
For achieving the above object, the processing method of virtual reality sound field metadata of the present invention comprises the following steps:
Judge the motor pattern of audio object, if the motor pattern of audio object is linear motion, make motor pattern join Number m=0;If the motor pattern of audio object is curvilinear motion, make motor pattern parameter m=1;
As m=0, with the azimuth information of this audio object of right-angle coordinate representation, the simultaneously movement locus of audio object Also process in rectangular coordinate system;
As m=1, representing the azimuth information of this audio object by polar coordinate system, the movement locus of audio object is also simultaneously Polar coordinate system processes;
The positional information that movement locus according to audio object is corresponding generates virtual surround sound.
As m=0, obtain the audio object rectangular coordinate Point_1 in the t1 moment, obtain audio object in the t2 moment Rectangular coordinate Point_2, the rectangular coordinate making ti moment audio object corresponding is Point_i, wherein the ti moment be the t1 moment with Any time between the t2 moment;
The particular location coordinate of Point_i is obtained by linear interpolation, and the formula of linear interpolation is:
Point_i=Point_1+ (Point_2-Point_1) × (ti-t1)/(t2-t1).
As m=1, acquisition audio object, in polar coordinate Point ' _ 1 in t1 moment, obtains audio object in the t2 moment Polar coordinate Point ' _ 2, the polar coordinate making ti moment audio object corresponding are Point ' _ i, and wherein the ti moment is t1 moment and t2 Any time between moment;
The particular location coordinate of Point ' _ i is obtained by linear interpolation, and the formula of linear interpolation is:
Point ' _ i=Point ' _ 1+ (Point ' _ 2-Point ' _ 1) × (ti-t1)/(t2-t1).
Concretely comprising the following steps of the positional information virtual surround sound of generation that the described movement locus according to audio object is corresponding:
If m=1, then obtain polar coordinate Point ' _ i that ti moment audio object is corresponding;
If m=0, then rectangular coordinate Point_i corresponding for the ti moment audio object of linear interpolation acquisition is converted into Polar coordinate Point ' _ i that ti moment audio object is corresponding;
According to polar coordinate Point ' _ i that ti moment audio object is corresponding, audio object is encoded to high-order B-form letter Number;Described B-format signal is converted into virtual speaker array signal;
The described virtual speaker array signal of audio object is carried out ears transcoding based on binaural room impulse response, Ears to audio object export virtual ring around acoustical signal.
Described audio object can be one or more.
Described binaural room impulse response off-line generates, and uses true measurement or by Software Create.
Present invention have the advantage that the processing method of virtual reality sound field metadata of the present invention is processing straight line fortune Dynamic and curvilinear motion and deposit complex object time, can more ideally represent the direction of motion and the track of audio object.Virtual Reality improves feeling of immersion, more truly, more specifically, vivider, and make the object handles of movement locus complexity get up simpler Single quick, also improve work efficiency while improve feeling of immersion.
Accompanying drawing explanation
Fig. 1 is with the schematic diagram in right-angle coordinate representation sound field space in prior art.
Fig. 2 is the schematic diagram of three coordinate components in the rectangular coordinate system shown in Fig. 1.
Fig. 3 is the audio object movement locus schematic diagram obtained by rectangular coordinate system interpolation in prior art.
Fig. 4 is the schematic diagram in present invention polar coordinate representation sound field space.
Fig. 5 is the schematic diagram of three coordinate components in the polar coordinate shown in Fig. 4.
Fig. 6 is the audio object movement locus schematic diagram that present invention polar coordinate interpolation obtains.
Detailed description of the invention
Following example are used for illustrating the present invention, but are not limited to the scope of the present invention.
As Figure 4-Figure 6, the present invention provides the processing method of a kind of virtual reality sound field metadata, comprises the following steps:
Judge the motor pattern of audio object, if the motor pattern of audio object is linear motion, make motor pattern join Number m=0;If the motor pattern of audio object is curvilinear motion, make motor pattern parameter m=1;
As m=0, with the azimuth information of this audio object of right-angle coordinate representation, the simultaneously movement locus of audio object Also process in rectangular coordinate system;
As m=1, representing the azimuth information of this audio object by polar coordinate system, the movement locus of audio object is also simultaneously Polar coordinate system processes;
The positional information that movement locus according to audio object is corresponding generates virtual surround sound.
As m=0, obtain the audio object rectangular coordinate Point_1 in the t1 moment, obtain audio object in the t2 moment Rectangular coordinate Point_2, the rectangular coordinate making ti moment audio object corresponding is Point_i, wherein the ti moment be the t1 moment with Any time between the t2 moment;
The particular location coordinate of Point_i is obtained by linear interpolation, and the formula of linear interpolation is:
Point_i=Point_1+ (Point_2-Point_1) × (ti-t1)/(t2-t1).
Wherein the rectangular coordinate of Point_1 is defined as (x1, y1, z1), the rectangular coordinate of Point_2 be defined as (x2, y2, Z2), the rectangular coordinate of Point_i is defined as (xi, yi, zi).
As m=1, acquisition audio object, in polar coordinate Point ' _ 1 in t1 moment, obtains audio object in the t2 moment Polar coordinate Point ' _ 2, the polar coordinate making ti moment audio object corresponding are Point ' _ i, and wherein the ti moment is t1 moment and t2 Any time between moment;
The particular location coordinate of Point ' _ i is obtained by linear interpolation, and the formula of linear interpolation is:
Point ' _ i=Point ' _ 1+ (Point ' _ 2-Point ' _ 1) × (ti-t1)/(t2-t1).
Wherein the polar coordinate of Point ' _ 1 are defined as (ρ 1, θ 1, φ 1);The polar coordinate of Point ' _ 2 are defined as (ρ 2, θ 2, φ 2);The polar coordinate of Point ' _ i are defined as (ρ i, θ i, φ i).
In polar coordinate, the position of the centre of sphere is defined as (0,0,0), ρ 1 represent t1 moment audio object distance the centre of sphere away from From;θ 1 represents the t1 moment audio object horizontal angle relative to the centre of sphere;φ 1 represents t1 moment audio object facing upward relative to the centre of sphere Angle.ρ 2 represents the distance of the t2 moment audio object distance centre of sphere;θ 2 represents the t2 moment audio object horizontal angle relative to the centre of sphere; φ 2 represents the t2 moment audio object elevation angle relative to the centre of sphere.ρ i represents the distance of the ti moment audio object distance centre of sphere;θ i table Show the ti moment audio object horizontal angle relative to the centre of sphere;φ i represents the ti moment audio object elevation angle relative to the centre of sphere.
Use polar providing the benefit that, it is assumed that audio object rotates a circle around central point, respectively all around With 4 azimuthal points of polar coordinate labelling:
The most front Point_1 (1,0,0)
The rightest Point_2 (1,90,0)
Point_3 (1,180,0) after just
The most left Point_4 (1,270,0)
The movement locus that so interpolation obtains is exactly a perfect circular motion, as shown in Figure 6.
When processing linear motion and curvilinear motion the complex object deposited, the present invention can more ideally represent audio frequency pair The direction of motion of elephant and track.Feeling of immersion is improved in virtual reality, more truly, more specifically, vivider, and make movement locus Complicated object handles get up simpler quickly, also improve work efficiency while improve feeling of immersion.
The definition specification (as a example by 2 audio objects) of the virtual reality sound field sound intermediate frequency object metadata of present invention design As follows:
Concretely comprising the following steps of the positional information virtual surround sound of generation that the described movement locus according to audio object is corresponding:
If m=1, then obtain polar coordinate Point ' _ i that ti moment audio object is corresponding;
If m=0, then rectangular coordinate Point_i corresponding for the ti moment audio object of linear interpolation acquisition is converted into Polar coordinate Point ' _ i that ti moment audio object is corresponding;
Rectangular coordinate (xi, yi, zi) is converted into the formula of polar coordinate (ρ i, θ i, φ i):
ρ i = X 2 + Y 2 + Z 2
θ i=arccos (z/ ρ i)
φ i=arctan (y/sqrt (x*x+z*z))
Wherein sqrt represents sqrt computing.
According to polar coordinate Point ' _ i that ti moment audio object is corresponding, audio object is encoded to high-order (preferably 2 rank Or 3 rank) B-format signal;Described B-format signal is converted into virtual speaker array signal;With a single order B-form letter Number [W1 X1 Y1 Z1]TAs a example by, it is converted into virtual speaker array signal [L1 L2…LN]TProcess be just by following computing (wherein X1=cos θ1cosφ1;Y1=sin θ1cosφ1;Z1=sin φ1):
L 1 L 2 .. L N = G w 1 G x 1 G y 1 G z 1 G w 2 G x 2 G y 2 G z 2 . . . . . . . . G w N G x N G y N G z N W 1 X 1 Y 1 Z 1 = G W 1 X 1 Y 1 Z 1 .
Wherein, N is the number of the virtual speaker that virtual speaker topological structure includes.G matrix used in above formula For ambisonic decoding matrix, can draw by seeking pseudo inverse matrix.
The described virtual speaker array signal of audio object is carried out ears based on binaural room impulse response (BRIR) Transcoding (typically 3-dimensional i.e. comprises elevation information), obtains the ears output virtual ring of audio object around acoustical signal.Specifically: from Virtual speaker signal forwards the two road stereo BRIR matrixes that earphone signal is corresponding to, the stereo matrix in Jiang Gai bis-road and virtual raise Sound device array signal carries out matrix multiplication, obtains virtual surround sound.BRIR matrix isThe most virtual surround sound For
Described audio object can be one or more.
Described binaural room impulse response is preferably off-line and generates, and can use true measurement or raw by special software Become, therefore not necessarily like needing when using online generating mode under prior art to store substantial amounts of BRIR, decrease memory consumption.
Describe below and how audio object is encoded to ambisonic territory.
Audio object is encoded to single order ambisonic signal:
W = 1 k Σ i = 1 k s i [ 1 2 ] ;
X = 1 k Σ i = 1 k s i [ cosθ i cosφ i ] ;
Y = 1 k Σ i = 1 k s i [ sinθ i cosφ i ] ;
Z = 1 k Σ i = 1 R s i [ sinφ i ] ;
siIt is i-th audio object, i=1 ... k, k are the numbers of audio object.θiIt it is the angle (orientation in plane Angle), φiThe angle being vertically oriented.W sound channel signal represents omnirange sound wave, X sound channel signal, Y sound channel signal and Z sound channel Signal represents the sound wave of orthogonal orientation X, Y, Z along three, space respectively.
Single order B-format signal is expressed as
In like manner, audio object is encoded to 2 rank or 3 rank B-format signals is preferably carried out according to lower table definition:
Trigonometric function in upper table is even function for azimuth angle theta, then the respective component of corresponding B-format signal is left Right symmetry, if the trigonometric function in upper table is odd function for azimuth angle theta, then the respective component of corresponding B-format signal is Heterochiral.As a example by single order B-format signal, from the point of view of physical significance and coordinate, w, x, z are regardless of left and right, if so listened The position of person is symmetrical, and supposes that corresponding HRTF coefficient also approximates symmetrical, then the ears output that w, x, z are corresponding Component for output left and right passage be identical.And y is the most reverse for left and right.So y corresponding ears output point Amount is contrary for left and right passage.For having symmetric component, can use in fast algorithm, i.e. calculating process is right The optimization of title property, can reduce operand further.
The alleged left and right of the present invention, level, the orientation such as vertically are all the perspective definition from hearer (i.e. user).
Although, the present invention is described in detail to have used general explanation and specific embodiment, but at this On the basis of invention, can make some modifications or improvements it, this will be apparent to those skilled in the art.Therefore, These modifications or improvements without departing from theon the basis of the spirit of the present invention, belong to the scope of protection of present invention.

Claims (6)

1. the processing method of a virtual reality sound field metadata, it is characterised in that the place of described virtual reality sound field metadata Reason method comprises the following steps:
Judge the motor pattern of audio object, if the motor pattern of audio object is linear motion, make motor pattern parameter m= 0;If the motor pattern of audio object is curvilinear motion, make motor pattern parameter m=1;
As m=0, by the azimuth information of this audio object of right-angle coordinate representation, the movement locus of audio object also exists simultaneously Rectangular coordinate system processes;
As m=1, representing the azimuth information of this audio object by polar coordinate system, the movement locus of audio object is also in pole simultaneously Coordinate system processes;
The positional information that movement locus according to audio object is corresponding generates virtual surround sound.
2. the processing method of virtual reality sound field metadata as claimed in claim 1, it is characterised in that as m=0, obtain sound Frequently object is at the rectangular coordinate Point_1 in t1 moment, obtains the audio object rectangular coordinate Point_2 in the t2 moment, when making ti The rectangular coordinate carving audio object corresponding is Point_i, and wherein the ti moment is any time between t1 moment and t2 moment;
The particular location coordinate of Point_i is obtained by linear interpolation, and the formula of linear interpolation is:
Point_i=Point_1+ (Point_2-Point_1) × (ti-t1)/(t2-t1).
3. the processing method of virtual reality sound field metadata as claimed in claim 1, it is characterised in that as m=1, obtain sound Frequently object is in polar coordinate Point ' _ 1 in t1 moment, and acquisition audio object, in polar coordinate Point ' _ 2 in t2 moment, makes the ti moment The polar coordinate that audio object is corresponding are Point ' _ i, and wherein the ti moment is any time between t1 moment and t2 moment;
The particular location coordinate of Point ' _ i is obtained by linear interpolation, and the formula of linear interpolation is:
Point ' _ i=Point ' _ 1+ (Point ' _ 2-Point ' _ 1) × (ti-t1)/(t2-t1).
4. the processing method of as claimed in claim 1 virtual reality sound field metadata, it is characterised in that described according to audio frequency pair The positional information that the movement locus of elephant is corresponding generates concretely comprising the following steps of virtual surround sound:
If m=1, then obtain polar coordinate Point ' _ i that ti moment audio object is corresponding;
If m=0, then when being converted into ti by rectangular coordinate Point_i corresponding for the ti moment audio object of linear interpolation acquisition Carve polar coordinate Point ' _ i that audio object is corresponding;
According to polar coordinate Point ' _ i that ti moment audio object is corresponding, audio object is encoded to high-order B-format signal;Will Described B-format signal is converted into virtual speaker array signal;
The described virtual speaker array signal of audio object is carried out ears transcoding based on binaural room impulse response, obtains sound Frequently the ears output virtual ring of object is around acoustical signal.
5. the processing method of virtual reality sound field metadata as claimed in claim 4, it is characterised in that described audio object is Individual or multiple.
6. the processing method of virtual reality sound field metadata as claimed in claim 5, it is characterised in that described binaural room impulse Response off-line generates, and uses true measurement or by Software Create.
CN201610391252.0A 2016-06-03 2016-06-03 The processing method of virtual reality sound field metadata Active CN106060758B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610391252.0A CN106060758B (en) 2016-06-03 2016-06-03 The processing method of virtual reality sound field metadata

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610391252.0A CN106060758B (en) 2016-06-03 2016-06-03 The processing method of virtual reality sound field metadata

Publications (2)

Publication Number Publication Date
CN106060758A true CN106060758A (en) 2016-10-26
CN106060758B CN106060758B (en) 2018-03-23

Family

ID=57169737

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610391252.0A Active CN106060758B (en) 2016-06-03 2016-06-03 The processing method of virtual reality sound field metadata

Country Status (1)

Country Link
CN (1) CN106060758B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106604200A (en) * 2017-01-19 2017-04-26 浪潮(苏州)金融技术服务有限公司 Audio data processing method and apparatus
CN109462811A (en) * 2018-11-23 2019-03-12 武汉轻工大学 Sound field rebuilding method, equipment, storage medium and device based on non-central point
GB2567244A (en) * 2017-10-09 2019-04-10 Nokia Technologies Oy Spatial audio signal processing
CN109618275A (en) * 2018-12-29 2019-04-12 武汉轻工大学 Non-central sound field rebuilding method, equipment, storage medium and the device of multichannel
CN112584297A (en) * 2020-12-01 2021-03-30 中国电影科学技术研究所 Audio data processing method and device and electronic equipment
CN114020235A (en) * 2021-09-29 2022-02-08 北京城市网邻信息技术有限公司 Audio processing method in real scene space, electronic terminal and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101521843A (en) * 2008-02-27 2009-09-02 索尼株式会社 Head-related transfer function convolution method and head-related transfer function convolution device
CN101874414A (en) * 2007-10-30 2010-10-27 索尼克埃莫申股份公司 Method and device for improved sound field rendering accuracy within a preferred listening area
WO2013067712A1 (en) * 2011-11-12 2013-05-16 Liv Runchun Method for establishing 5.1 sound track on headset
CN104284291A (en) * 2014-08-07 2015-01-14 华南理工大学 Headphone dynamic virtual replaying method based on 5.1 channel surround sound and implementation device thereof
CN105075293A (en) * 2013-03-29 2015-11-18 三星电子株式会社 Audio apparatus and audio providing method thereof
CN105376690A (en) * 2015-11-04 2016-03-02 北京时代拓灵科技有限公司 Method and device of generating virtual surround sound
CN105451152A (en) * 2015-11-02 2016-03-30 上海交通大学 Hearer-position-tracking-based real-time sound field reconstruction system and method

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101874414A (en) * 2007-10-30 2010-10-27 索尼克埃莫申股份公司 Method and device for improved sound field rendering accuracy within a preferred listening area
CN101521843A (en) * 2008-02-27 2009-09-02 索尼株式会社 Head-related transfer function convolution method and head-related transfer function convolution device
WO2013067712A1 (en) * 2011-11-12 2013-05-16 Liv Runchun Method for establishing 5.1 sound track on headset
CN105075293A (en) * 2013-03-29 2015-11-18 三星电子株式会社 Audio apparatus and audio providing method thereof
CN104284291A (en) * 2014-08-07 2015-01-14 华南理工大学 Headphone dynamic virtual replaying method based on 5.1 channel surround sound and implementation device thereof
CN105451152A (en) * 2015-11-02 2016-03-30 上海交通大学 Hearer-position-tracking-based real-time sound field reconstruction system and method
CN105376690A (en) * 2015-11-04 2016-03-02 北京时代拓灵科技有限公司 Method and device of generating virtual surround sound

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
郭龙: "坐标系在大学物理运动学中作用的探究", 《基础科学》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106604200A (en) * 2017-01-19 2017-04-26 浪潮(苏州)金融技术服务有限公司 Audio data processing method and apparatus
GB2567244A (en) * 2017-10-09 2019-04-10 Nokia Technologies Oy Spatial audio signal processing
CN109462811A (en) * 2018-11-23 2019-03-12 武汉轻工大学 Sound field rebuilding method, equipment, storage medium and device based on non-central point
CN109462811B (en) * 2018-11-23 2020-11-17 武汉轻工大学 Sound field reconstruction method, device, storage medium and device based on non-central point
CN109618275A (en) * 2018-12-29 2019-04-12 武汉轻工大学 Non-central sound field rebuilding method, equipment, storage medium and the device of multichannel
CN109618275B (en) * 2018-12-29 2020-11-17 武汉轻工大学 Multi-channel non-center point sound field reconstruction method, equipment, storage medium and device
CN112584297A (en) * 2020-12-01 2021-03-30 中国电影科学技术研究所 Audio data processing method and device and electronic equipment
CN114020235A (en) * 2021-09-29 2022-02-08 北京城市网邻信息技术有限公司 Audio processing method in real scene space, electronic terminal and storage medium
CN114020235B (en) * 2021-09-29 2022-06-17 北京城市网邻信息技术有限公司 Audio processing method in live-action space, electronic terminal and storage medium

Also Published As

Publication number Publication date
CN106060758B (en) 2018-03-23

Similar Documents

Publication Publication Date Title
CN106060758B (en) The processing method of virtual reality sound field metadata
CN107205207B (en) A kind of virtual sound image approximation acquisition methods based on middle vertical plane characteristic
CN106210990B (en) A kind of panorama sound audio processing method
US20180220253A1 (en) Differential headtracking apparatus
CN102572676A (en) Real-time rendering method for virtual auditory environment
Mehra et al. Source and listener directivity for interactive wave-based sound propagation
CN105376690A (en) Method and device of generating virtual surround sound
CN106331977B (en) A kind of virtual reality panorama acoustic processing method of network K songs
CN106454686A (en) Multi-channel surround sound dynamic binaural replaying method based on body-sensing camera
CN105682000B (en) A kind of audio-frequency processing method and system
CN102333265B (en) Replay method of sound fields in three-dimensional local space based on continuous sound source concept
CN101184349A (en) Three-dimensional ring sound effect technique aimed at dual-track earphone equipment
Hogg et al. HRTF upsampling with a generative adversarial network using a gnomonic equiangular projection
Kaneko et al. Ear shape modeling for 3D audio and acoustic virtual reality: The shape-based average HRTF
Liu et al. A sound source localization method based on microphone array for mobile robot
Moldoveanu et al. TRAINING SYSTEM FOR IMPROVING SPATIAL SOUND LOCALIZATION.
CN109598785A (en) A kind of three-dimensional grid model view conversion method
Rumsey Binaural audio and virtual acoustics
CN113347530A (en) Panoramic audio processing method for panoramic camera
O'Dwyer et al. A machine learning approach to detecting sound-source elevation in adverse environments
Sakamoto et al. SENZI and ASURA: New high-precision sound-space sensing systems based on symmetrically arranged numerous microphones
Klein et al. Sonification of three-dimensional vector fields
Salvati et al. Sound spatialization control by means of acoustic source localization system
Wickert A Real-Time 3D Audio Simulator for Cognitive Hearing Science.
Geronazzo et al. Use of personalized binaural audio and interactive distance cues in an auditory goal-reaching task

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant