CN116724351A - Quantization coding method, device, equipment and storage medium - Google Patents

Quantization coding method, device, equipment and storage medium Download PDF

Info

Publication number
CN116724351A
CN116724351A CN202380008765.7A CN202380008765A CN116724351A CN 116724351 A CN116724351 A CN 116724351A CN 202380008765 A CN202380008765 A CN 202380008765A CN 116724351 A CN116724351 A CN 116724351A
Authority
CN
China
Prior art keywords
sound source
bit number
source object
listening
sound
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202380008765.7A
Other languages
Chinese (zh)
Inventor
胡晨昊
史润宇
王宾
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Xiaomi Mobile Software Co Ltd
Original Assignee
Beijing Xiaomi Mobile Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Xiaomi Mobile Software Co Ltd filed Critical Beijing Xiaomi Mobile Software Co Ltd
Publication of CN116724351A publication Critical patent/CN116724351A/en
Pending legal-status Critical Current

Links

Landscapes

  • Stereophonic System (AREA)

Abstract

The present disclosure proposes a quantization encoding method, apparatus, device, and storage medium, the method comprising: determining a target quantization bit number of sound source orientation information based on at least one of sound cone information of a sound source object, and an area in which the sound source object is located with respect to a listening object; and carrying out quantization coding on the sound source orientation information based on the target quantization bit number to obtain a code stream signal. The method and the device can pertinently select the proper quantization coding mode based on the current perception sensitivity of the listening object to the change of the direction of the sound source object, and have high flexibility.

Description

Quantization coding method, device, equipment and storage medium
Technical Field
The disclosure relates to the technical field of communication, and in particular relates to a quantization coding method, a quantization coding device, quantization coding equipment and a storage medium.
Background
The spatial audio frequency can bring realistic spatial sense and azimuth sense to the user, so the spatial audio frequency processing technology is widely applied. Among other things, one effective way to implement spatial audio is object-based spatial audio technology. In the object-based spatial audio technology, in addition to transmitting the encoded audio signal to the decoding device, the encoding device needs to determine sound source orientation information (such as a horizontal angle and/or a height angle of a sound source object relative to a listening object), and quantitatively encode the sound source orientation information and transmit the encoded sound source orientation information to the decoding device. And the decoding equipment can decode and inverse-quantize the received information to obtain sound source orientation information and an audio signal, render the audio signal based on the sound source orientation information and play the audio signal to a listening object so as to restore the space sense and the azimuth sense of the audio signal.
Disclosure of Invention
The present disclosure proposes a quantization encoding method, apparatus, device, and storage medium.
In a first aspect, an embodiment of the present disclosure provides a quantization encoding method, including:
determining a target quantization bit number of sound source orientation information based on at least one of sound cone information of a sound source object, and an area in which the sound source object is located with respect to a listening object;
and carrying out quantization coding on the sound source orientation information based on the target quantization bit number to obtain a code stream signal.
In a second aspect, embodiments of the present disclosure provide a quantization encoding method, including:
receiving a code stream signal sent by coding equipment; the code stream signal is: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
and decoding and inversely quantizing the code stream signal to obtain the sound source orientation information.
In a third aspect, embodiments of the present disclosure provide a communication apparatus, including:
a processing module for determining a target quantization bit number of the sound source orientation information based on at least one of sound cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
The processing module is further configured to perform quantization encoding on the sound source orientation information based on the target quantization bit number, and obtain a code stream signal.
In a fourth aspect, embodiments of the present disclosure provide a communication apparatus, including:
the receiving and transmitting module is used for receiving the code stream signal sent by the encoding equipment; the code stream signal is: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
and the processing module is used for decoding and inverse quantizing the code stream signal to obtain the sound source orientation information.
In a fifth aspect, an embodiment of the present disclosure provides a communication device, including a processor, which when calling a computer program in a memory, performs the method of the first or second aspect.
In a sixth aspect, embodiments of the present disclosure provide a communication apparatus comprising a processor and a memory, the memory having a computer program stored therein; the processor executes the computer program stored in the memory to cause the communication device to perform the method of the first or second aspect described above.
In a seventh aspect, embodiments of the present disclosure provide a communications apparatus comprising a processor and interface circuitry for receiving code instructions and transmitting to the processor, the processor being configured to execute the code instructions to cause the apparatus to perform the method of the first or second aspects described above.
In an eighth aspect, embodiments of the present disclosure provide a communication system, which includes the communication device according to the third aspect to the fourth aspect, or which includes the communication device according to the fifth aspect, or which includes the communication device according to the sixth aspect, or which includes the communication device according to the seventh aspect.
In a ninth aspect, embodiments of the present disclosure provide a computer readable storage medium storing instructions for use by a network device as described above, which when executed cause the terminal to perform the method of the first or second aspect.
In a tenth aspect, the present disclosure also provides a computer program product comprising a computer program which, when run on a computer, causes the computer to perform the method of the first or second aspect described above.
In an eleventh aspect, the present disclosure provides a chip system comprising at least one processor and an interface for supporting a network device to implement the functions involved in the method of the first or second aspect, e.g. to determine or process at least one of data and information involved in the above method. In one possible design, the system-on-chip further includes a memory to hold the necessary computer programs and data for the source and secondary nodes. The chip system can be composed of chips, and can also comprise chips and other discrete devices.
In a twelfth aspect, the present disclosure provides a computer program which, when run on a computer, causes the computer to perform the method of the first or second aspect described above.
Drawings
The foregoing and/or additional aspects and advantages of the present disclosure will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings, in which:
fig. 1 is a schematic architecture diagram of some communication systems provided in embodiments of the present disclosure;
FIG. 2a is a flowchart of a quantization encoding method according to another embodiment of the present disclosure;
FIG. 2b is a schematic diagram of an inner cone angle, an outer cone angle, and an attenuation amplitude of an audio signal according to an embodiment of the present disclosure;
FIG. 3a is a flowchart of a quantization encoding method according to still another embodiment of the present disclosure;
fig. 3b is a schematic structural diagram of a sound source object according to an embodiment of the present disclosure in different areas with respect to a listening object;
FIG. 4 is a flowchart of a quantization encoding method according to still another embodiment of the present disclosure;
FIG. 5 is a flowchart of a quantization encoding method according to still another embodiment of the present disclosure;
FIG. 6 is a flowchart of a quantization encoding method according to still another embodiment of the present disclosure;
FIG. 7 is a flowchart of a quantization encoding method according to still another embodiment of the present disclosure;
FIG. 8 is a flowchart of a quantization encoding method according to still another embodiment of the present disclosure;
FIG. 9 is an interactive flow chart of a quantization encoding method provided in a further embodiment of the present disclosure;
fig. 10 is a schematic structural diagram of a communication device according to an embodiment of the present disclosure;
fig. 11 is a schematic structural diagram of a communication device according to an embodiment of the present disclosure;
FIG. 12 is a block diagram of a communication device provided by one embodiment of the present disclosure;
fig. 13 is a schematic structural diagram of a chip according to an embodiment of the disclosure.
Detailed Description
Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numbers in different drawings refer to the same or similar elements, unless otherwise indicated. The implementations described in the following exemplary embodiments do not represent all implementations consistent with the embodiments of the present disclosure. Rather, they are merely examples of apparatus and methods consistent with aspects of embodiments of the present disclosure as detailed in the accompanying claims.
The terminology used in the embodiments of the disclosure is for the purpose of describing particular embodiments only and is not intended to be limiting of the embodiments of the disclosure. As used in this disclosure of embodiments and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that the term "and/or" as used herein refers to and encompasses any or all possible combinations of one or more of the associated listed items.
It should be understood that although the terms first, second, third, etc. may be used in embodiments of the present disclosure to describe various information, these information should not be limited to these terms. These terms are only used to distinguish one type of information from another. For example, the first information may also be referred to as second information, and similarly, the second information may also be referred to as first information, without departing from the scope of embodiments of the present disclosure. The words "if" and "if" as used herein may be interpreted as "at … …" or "at … …" or "in response to a determination", depending on the context.
Embodiments of the present disclosure are described in detail below, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the like or similar elements throughout. The embodiments described below by referring to the drawings are exemplary and intended for the purpose of explaining the present disclosure and are not to be construed as limiting the present disclosure.
In the related art, a unified quantization coding mode is adopted to perform unified quantization coding on sound source orientation information of different sound source objects at different moments. The method in the related art generally causes waste of the coding rate or insufficient quantization coding precision to affect the rendering effect of the subsequent audio signal, and the specific reasons are as follows:
in different situations, the perceived sensitivity of the listening object to the audio signal will also be different when the sound source object is oriented differently. For example, in some cases, the perceived sensitivity of the listening object to the change in the orientation of the sound source object is low, at which point the listening object cannot easily perceive the change in the orientation of the sound source object; in other cases, the listening object has a high perceived sensitivity to changes in the orientation of the sound source object, where the listening object can easily perceive changes in the orientation of the sound source object. Wherein when the perceived sensitivity of the listening object to the change in the direction of the sound source object is low, since the listening object cannot easily perceive the change in the direction of the sound source object, the encoding apparatus can appropriately perform rough rendering without fine rendering in rendering the audio object based on the sound source direction information later. The encoding device can also correspondingly adopt a coarse quantization encoding mode to perform quantization encoding when the sound source orientation information is subjected to quantization encoding, so that the encoding code stream can be reduced on the premise of not influencing the subsequent rendering effect. And when the perceived sensitivity of the listening object to the change in the direction of the sound source object is high, since the listening object can easily perceive the change in the direction of the sound source object, the encoding apparatus should be finely rendered later in rendering the audio object based on the sound source direction information. The encoding device also needs to perform quantization encoding by adopting a fine quantization encoding mode when performing quantization encoding on the sound source orientation information so as to ensure the subsequent rendering effect.
Based on this, when the sound source direction information is quantized by the method in the related art, that is, the unified quantization encoding method is used to perform unified quantization encoding on the sound source direction information of different sound source objects at different moments. If the unified quantization coding scheme is a fine quantization coding scheme, there is a possibility that "when the perceived sensitivity of the listening object to the change in the direction of the sound source object is low," the fine quantization coding scheme is used to quantization-code the sound source direction information, which may result in unnecessary waste of the coding rate. If the unified quantization coding scheme is a coarse quantization coding scheme, there is a possibility that "when the perceived sensitivity of the listening object to the change in the direction of the sound source object is high," the coarse quantization coding scheme is used to perform quantization coding on the sound source direction information, and the accuracy of quantization coding is insufficient, so that the rendering effect of the subsequent audio signal is affected.
Accordingly, there is a need for a quantization encoding method for selectively selecting an appropriate quantization encoding method for quantization encoding sound source orientation information based on the perceived sensitivity of a listening object to the change of the orientation of a sound source object in different situations. The present disclosure thus provides a quantization encoding method.
In order to better understand a quantization encoding method disclosed in the embodiments of the present disclosure, a communication system to which the embodiments of the present disclosure are applied will be described first.
Referring to fig. 1, fig. 1 is a schematic architecture diagram of a communication system according to an embodiment of the disclosure. The communication system may include, but is not limited to, at least one encoding device and at least one decoding device, wherein the encoding device may be a network device or a terminal device, and the encoding device may also be a network device or a terminal device. And the number and form of the devices shown in fig. 1 are only for example and not to limit the embodiments of the disclosure, and one or more encoding devices, or one or more terminal devices may be included in the application. The communication system shown in fig. 1 includes an encoding device, which is a network device, and a decoding device, which is a terminal device, for example.
It should be noted that the technical solution of the embodiment of the present disclosure may be applied to various communication systems. For example: a long term evolution (long term evolution, LTE) system, a fifth generation (5th generation,5G) mobile communication system, a 5G New Radio (NR) system, or other future new mobile communication systems, etc.
The network device (e.g., the first network device or the second network device) in the embodiments of the present disclosure is an entity on the network side for transmitting or receiving signals. For example, the network device may be an evolved NodeB (eNB), a transmission and reception point (transmission reception point, TRP), a remote radio head (Radio Remote Head, RRH), a next generation NodeB (gNB) in an NR system, a base station in other future mobile communication systems, or an access node in a wireless fidelity (wireless fidelity, wiFi) system, etc. The embodiments of the present disclosure do not limit the specific technology and specific device configuration adopted by the base station. The base station provided in the embodiments of the present disclosure may be composed of a Central Unit (CU) and a Distributed Unit (DU), where the CU may also be referred to as a control unit (control unit), and the structure of the CU-DU may be used to split the protocol layers of the base station, for example, the base station, where functions of part of the protocol layers are placed in the CU for centralized control, and functions of part or all of the protocol layers are distributed in the DU for centralized control of the DU by the CU.
The terminal device in the embodiments of the present disclosure may be an entity on the user side for receiving or transmitting signals, such as a mobile phone. The terminal device may also be referred to as a terminal (terminal), a User Equipment (UE), a Mobile Station (MS), a Mobile Terminal (MT), etc. The terminal device may be an automobile with communication function, a smart car, a mobile phone (mobile phone), a wearable device, a tablet computer (Pad), a computer with wireless transceiving function, a Virtual Reality (VR) terminal, an augmented reality (augmented reality, AR) terminal, a wireless terminal in industrial control (industrial control), a wireless terminal in unmanned (self-driving), a wireless terminal in teleoperation (remote medical surgery), a wireless terminal in smart grid (smart grid), a wireless terminal in transportation security (transportation safety), a wireless terminal in smart city (smart city), a wireless terminal in smart home (smart home), or the like. The embodiment of the present disclosure does not limit the specific technology and the specific device configuration adopted by the terminal device.
It may be understood that, the communication system described in the embodiments of the present disclosure is for more clearly describing the technical solutions of the embodiments of the present disclosure, and is not limited to the technical solutions provided in the embodiments of the present disclosure, and those skilled in the art can know that, with the evolution of the system architecture and the appearance of new service scenarios, the technical solutions provided in the embodiments of the present disclosure are equally applicable to similar technical problems.
Further, in order to facilitate understanding of the embodiments of the present disclosure, the following description is made.
In the first and the second embodiments, each step in any implementation manner or embodiment may be implemented as an independent embodiment without contradiction, and the steps may be arbitrarily combined, for example, a scheme after removing part of the steps in a certain implementation manner or embodiment may also be implemented as an independent embodiment, and the order of the steps in a certain implementation manner or embodiment may be arbitrarily exchanged, and in addition, an optional manner or optional embodiment in a certain implementation manner or embodiment may be arbitrarily combined; furthermore, various embodiments or examples may be arbitrarily combined, for example, some or all steps of different embodiments or examples may be arbitrarily combined, and a certain embodiment or example may be arbitrarily combined with alternative modes or alternative examples of other embodiments or examples.
Second, the expressions "a or B", "a and/or B", "at least one of a and B", "a in one case a", "B in another case B", etc. in relation to the present disclosure may include at least one of the following aspects according to circumstances: a is performed independently of B, i.e., a in some embodiments; b is performed independently of a, i.e., in some embodiments B; A. b is selectively performed, i.e., in some embodiments selected from a and B; A. b are all performed, i.e., a and B in some embodiments.
Third, each element, each row, or each column in the tables to which the present disclosure relates may be implemented as a separate embodiment, and any combination of elements, any rows, or any columns may also be implemented as a separate embodiment.
Fourth, in some embodiments or examples, "comprising a", "containing a", "for indicating a", "carrying a" in the present disclosure may be interpreted as carrying a directly, or as indicating a indirectly.
Fifth, in some implementations or embodiments, "responsive to … …," "in the case of … …," "at … …," "when … …," "if … …," "if … …," and the like in the present disclosure may be replaced with each other.
Fig. 2a is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by an encoding device, and as shown in fig. 2a, the quantization encoding method may include the following steps:
step 201, determining a target quantization bit number of sound source orientation information based on at least one of sound cone information of the sound source object and an area where the sound source object is located relative to the listening object.
Alternatively, the quantization encoding method of the present disclosure may be applied to quantization encoding of an "object-based audio signal". The object-based audio signal may for example comprise at least one of the following: the audio to be played of the terminal equipment, the audio to be played of the home theater, and the audio to be played of the Virtual Reality (VR) equipment. That is, the method of the present disclosure may be applied to scenes such as "terminal device plays audio, home theater plays movie, and uses VR device for game audio-visual entertainment".
Optionally, in an embodiment of the present disclosure, the cone information of the sound source object may include at least one of:
the sound cone inner angle of the sound source object;
the cone outer angle of the sound source object;
maximum attenuation amplitude of an audio signal emitted from a sound source object.
Optionally, the angle of the sound cone inner angle means: the audio signal emitted by the sound source object and lying within the cone interior angle range of the sound source object is considered to be free of attenuation due to orientation, i.e. the intensity of the audio signal lying within the cone interior angle range remains unchanged.
Optionally, the cone outer angle means: an audio signal emitted by a sound source object and located between an inside cone angle edge and an outside cone angle edge of the sound source object is believed to have attenuation due to orientation, wherein the intensity of the audio signal located between the inside cone angle edge and the outside cone angle edge decays linearly with the exit angle of the audio signal, the exit angle of the audio signal being: the angle between the audio signal and the orientation (e.g. absolute orientation) of the sound source object.
Alternatively, the maximum attenuation amplitude of the audio signal emitted by the sound source object may be: the attenuation amplitude of the audio signal emitted by the sound source object and lying outside the angular range of the sound cone outside the sound source object.
Optionally, fig. 2b is a schematic diagram of an inside angle of sound cone, an outside angle of sound cone, and an attenuation amplitude of an audio signal according to an embodiment of the present disclosure, where fig. 2a (1) is a schematic diagram of a structure of the inside angle of sound cone (i.e., inside angle of sound cone in the drawing) and the outside angle of sound cone (i.e., outside angle of sound cone in the drawing), and fig. 2a (2) is a correspondence between an exit angle of an audio signal (i.e., angle in fig. 2a (2)) and an attenuation amplitude of an audio signal (i.e., gain value in fig. 2a (2)). As shown in fig. 2b (1), the sound source object has an inner cone angle of 30 ° and an outer cone angle of 90 °, and as shown in fig. 2b (2), the maximum attenuation amplitude of the audio signal emitted from the sound source object is 0.6. Alternatively, in fig. 2b (1), the intensity of the sound emitted by the sound source object in the region included in the angle edge of the sound cone is kept constant, so that the attenuation amplitude of the audio signal having the emission angle between [ -15, 15] in fig. 2b (2) is 1. In fig. 2b (1), the intensity of the audio signal emitted by the sound source object in the region between the inner angle edge of the sound cone and the outer angle edge of the sound cone is linearly attenuated, so that the attenuation amplitude of the audio signal in fig. 2b (2) with the emission angle between [ -45, -15] and [ -15, 45] is linearly varied with the emission angle of the audio signal. In fig. 2b (1), the intensity of the audio signal emitted from the sound source object in the region outside the angle edge of the cone has been attenuated to the maximum attenuation amplitude, and it is considered that the intensity of the sound is not changed any more, so that the attenuation amplitude of the corresponding region in fig. 2b (2) is 0.6. It should be noted that the data for the inside cone angle, the outside cone angle, and the maximum attenuation magnitude shown in fig. 2b are for illustration only and are not fixed.
Alternatively, in one embodiment of the present disclosure, the cone information corresponding to the different sound source objects may be the same or different, e.g., the cone inner angle of the different sound source objects may be the same or different, the cone outer angle of the different sound source objects may be the same or different, and the maximum attenuation magnitudes of the different sound source objects may be the same or different.
Alternatively, the cone information of the sound source object may be predetermined by the encoding device, and alternatively, the encoding device may determine the cone information of the sound source object by using an existing method, which will not be described in detail in this disclosure.
Alternatively, in one embodiment of the present disclosure, the region where the sound source object is located with respect to the listening object may be determined by the encoding device based on at least one of an orientation of the listening object, a relative position of the sound source object with respect to the listening object. Alternatively, the method of determining the region in which the sound source object is located with respect to the listening object by the encoding apparatus may comprise the steps of:
first, the orientation of the listening object is determined.
Optionally, in one embodiment of the present disclosure, the orientation of the listening object may be an absolute orientation of the listening object, which may be, for example: forward south orientation; alternatively, the orientation of the listening object may be the relative orientation of the listening object with respect to the sound source object, alternatively the relative orientation of the listening object with respect to the sound source object may be, for example: the listening object is offset by 30 ° southeast relative to the sound source object.
Alternatively, in one embodiment of the present disclosure, the orientation of the listening object may be sent by the decoding device to the encoding device. Alternatively, when the encoding device needs to determine the absolute orientation of the listening object, the decoding device may send the absolute orientation of the listening object directly to the encoding device. When the encoding device needs to determine the relative orientation of the listening object, the decoding device may send the relative orientation of the listening object directly to the sound source object, or the decoding device may send the absolute orientation of the listening object to the encoding device, after which the encoding device determines the relative orientation of the listening object based on the absolute orientation of the sound source object itself and the absolute orientation of the listening object it receives.
Alternatively, in another embodiment of the present disclosure, the orientation of the listening object may also be manually entered into the encoding device.
And a second step of determining the relative position of the sound source object with respect to the listening object.
Alternatively, in one embodiment of the present disclosure, the encoding device may determine the absolute position of the sound source object and the absolute position of the listening object first, and then determine the relative position of the sound source object with respect to the listening object based on the absolute position of the sound source object and the absolute position of the listening object. Alternatively, the absolute position of the sound source object may be autonomously determined by the encoding device, or may be manually input into the encoding device. The absolute position of the listening object may be sent by the decoding device to the encoding device or may be manually entered into the encoding device.
Alternatively, in another embodiment of the present disclosure, the encoding device may not need to determine the absolute position of the sound source object and the absolute position of the listening object, but may directly receive the relative position of the sound source object with respect to the listening object sent by the decoding device, or the relative position of the sound source object with respect to the listening object may be manually input to the encoding device.
It should be noted that the above-mentioned determination methods of "the orientation of the listening object, the relative position of the sound source object with respect to the listening object" are merely exemplary descriptions of the present disclosure. It should be understood that other methods for determining the orientation of the listening object, the relative position of the sound source object with respect to the listening object are within the scope of the present disclosure.
It can be understood that the first step and the second step in the embodiment of the present disclosure define only different operations, and do not limit the sequence of the two operations, that is, the operation of determining the relative position of the sound source object with respect to the listening object may be performed first, and then the operation of determining the orientation of the listening object may be performed.
And a third step of determining an area where the sound source object is located relative to the listening object based on at least one of an orientation of the listening object and a relative position of the sound source object relative to the listening object.
Optionally, the area where the sound source object is located relative to the listening object may be: after the sound source object is projected to the horizontal plane where the sound listening object is located, the projection position of the sound source object on the horizontal plane where the sound listening object is located is opposite to the area where the sound listening object is located; for example, the region in which the sound source object is located relative to the listening object may be: the sound source object (or referred to as the projection position of the sound source object at the horizontal plane where the sound object is located) is located in a front or rear or left or right area of the sound object.
Alternatively, in one embodiment of the present disclosure, when the encoding device obtains the relative orientation of the listening object with respect to the sound source object in the first step, the second step may be not performed, and the area where the sound source object is located with respect to the listening object may be determined directly based on the relative orientation of the listening object with respect to the sound source object.
Alternatively, in another embodiment of the present disclosure, when the encoding device obtains the absolute orientation of the listening object in the first step, a second step is required to determine the relative position of the sound source object with respect to the listening object, and then determine the area where the sound source object is located with respect to the listening object based on the absolute orientation of the listening object and the relative position of the sound source object with respect to the listening object.
Optionally, in one embodiment of the disclosure, the above sound source orientation information may include at least one of:
a horizontal angle between the sound source object and the listening object;
the altitude angle between the sound source object and the listening object.
Alternatively, the horizontal angle between the sound source object and the listening object described above may be understood as, for example: the relative deflection angle in the horizontal plane between the sound source object and the listening object, the above-mentioned altitude angle between the sound source object and the listening object can be understood as, for example: the relative deflection angle between the sound source object and the listening object in the vertical plane.
Alternatively, the target quantization bit number may be a bit number of the quantized encoded sound source orientation information. Optionally, when the target quantization bit number is greater, the quantization coding mode for indicating the sound source orientation information is: fine quantization coding, when the target quantization bit number is smaller, the quantization coding mode for describing the sound source orientation information is as follows: and (5) coarse quantization coding.
The following describes an example of the reason why the target quantization bit number of the sound source orientation information is determined based on at least one of the sound cone information of the sound source object, the region in which the sound source object is located with respect to the listening object in this step 201.
As can be seen from the foregoing description of the embodiment of fig. 1, the method of the present disclosure is mainly for implementing that the encoding device can selectively perform quantization encoding on the sound source orientation information by using a proper quantization encoding mode (such as a coarse quantization encoding mode (i.e., a smaller number of target quantization bits) or a fine quantization encoding mode (i.e., a larger number of target quantization bits) based on the perceived sensitivity of the listening object to the change of the orientation of the sound source object under different conditions.
Based on this, since the above-mentioned sound cone information of the sound source object and the region where the sound source object is located relative to the sound source object may represent the perceived sensitivity of the sound source object to the change in the direction of the sound source object, in the embodiment of the present disclosure, the target quantization bit number of the sound source direction information is determined based on at least one of the sound cone information of the sound source object and the region where the sound source object is located relative to the sound source object, so that the above-mentioned encoding device can selectively perform quantization encoding on the sound source direction information in a proper quantization encoding manner based on the perceived sensitivity of the sound source object to the change in the direction of the sound source object under different conditions, thereby improving the flexibility of quantization encoding.
Optionally, the cone information of the sound source object may embody a perceived sensitivity of the listening object to a change in orientation of the sound source object may specifically include: as can be seen from the foregoing description of the sound cone information, the sound cone information may represent a sound intensity change rate of the sound source object, where the sound intensity change rate is: the rate of change of the intensity of an audio signal emitted by a sound source object with the angle of emergence of the audio signal can be understood as the sound attenuation amplitude. When the sound cone information of the sound source object indicates a larger sound intensity change rate, it indicates that when the direction of the sound source object changes, the intensity change and the attenuation amplitude of the audio signal emitted by the sound source object are larger, and the listening object is more sensitive (i.e. more perceptible) to the audio signal with larger intensity change or attenuation amplitude, then it indicates that: when the sound cone information of a sound source object indicates a larger rate of change in sound intensity, the perceived sensitivity of the listening object to changes in the orientation of the sound source object is higher. And when the sound cone information of the sound source object indicates that the sound intensity change rate is smaller, the intensity change and the attenuation amplitude of the audio signal emitted by the sound source object are smaller when the direction of the sound source object is changed, and the listener is insensitive (i.e. is not easy to perceive) to the audio signal with smaller intensity change or smaller attenuation amplitude, then the following is indicated: when the sound cone information of a sound source object indicates a smaller sound intensity change rate, the perceived sensitivity of the listening object to changes in the orientation of the sound source object is lower.
Therefore, the sound intensity change rate of the sound source object can be determined based on the sound cone information, and the perception sensitivity of the listening object to the change of the direction of the sound source object can be further known.
Optionally, the region where the sound source object is located relative to the listening object may represent a perceived sensitivity of the listening object to a change in orientation of the sound source object may include: when the sound source object is located in different areas of the sound source object, the perception sensitivity of the sound source object to the audio signal sent by the sound source object is different, so that the perception sensitivity of the sound source object to the change of the direction of the sound source object in the different areas is different. For example, when a sound source object is located in a left or right region of a sound object, the perceived sensitivity of the sound object to changes in the direction of the sound source object is low, and when a sound source object is located in a front or rear region of the sound object, the perceived sensitivity of the sound object to changes in the direction of the sound source object is high.
Optionally, in an embodiment of the present disclosure, the method of determining the target quantization bit number of the sound source orientation information based on at least one of the sound cone information of the sound source object and the region where the sound source object is located with respect to the listening object may include: the target quantization bit number of the sound source orientation information is determined based on sound cone information of the sound source object and an area where the sound source object is located with respect to the listening object. Alternatively, in another embodiment of the present disclosure, when the encoding apparatus cannot know the region in which the sound source object is located with respect to the listening object, the encoding apparatus may determine the target quantization bit number of the sound source orientation information based on the sound cone information of the sound source object. Here, an introduction about how the target quantization bit number of the sound source orientation information is determined in detail in this step 201 will be described in the following embodiments.
Based on the foregoing, it can be seen that in the embodiments of the present disclosure, the target quantization bit number of the sound source orientation information is determined based on at least one of the sound cone information of the sound source object and the region where the sound source object is located with respect to the listening object. The cone information of the sound source object and the region where the sound source object is located relative to the listening object can show the perception sensitivity of the listening object to the change of the direction of the sound source object, and the target quantization bit number is: the bit number of the sound source orientation information after quantization coding can represent a specific quantization coding mode, for example, when the determined target quantization bit number is more, the corresponding quantization coding mode is described as follows: the fine quantization coding scheme, when the number of the determined target quantization bits is smaller, indicates that the corresponding quantization coding scheme is: coarse quantization coding scheme. It can be seen that, in the present disclosure, when determining the target quantization bit number of the sound source orientation information based on at least one of the sound cone information of the sound source object and the region where the sound source object is located relative to the listening object, an appropriate quantization coding manner is determined pertinently based on the current perceived sensitivity of the listening object to the change of the orientation of the sound source object, so that when at least one of the sound cone information of the sound source object and the region where the sound source object is located relative to the listening object indicates: when the perception sensitivity of the listening object to the change of the direction of the sound source object is low, the determined target quantization bit number is small, and the sound source direction information is quantized and encoded in a coarse quantization encoding mode, so that the encoding code stream is ensured to be reduced on the premise that the subsequent rendering effect is not affected, and the resource waste is avoided. And when at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object indicates: when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, the determined target quantization bit number is more, so that the sound source direction information is quantized and encoded in a fine quantization encoding mode, and the subsequent rendering effect can be ensured.
And 202, carrying out quantization coding on sound source orientation information based on the target quantization bit number to obtain a code stream signal.
Optionally, in one embodiment of the disclosure, the code stream signal may include at least one of:
an encoded audio signal;
quantizing the encoded sound source orientation information;
a target quantization bit number;
cone information of the sound source object;
an indication index corresponding to the sound cone information;
absolute position of sound source object;
an absolute position of the listening object;
the relative position of the sound source object with respect to the listening object.
The detailed method of "quantization encoding sound source orientation information based on the target quantization bit number and obtaining a code stream signal" will be described in the following embodiments.
In one embodiment of the present disclosure, after the sound source direction information of the sound source object is quantized and encoded for the first time, if the sound source direction information of the sound source object is not changed, the sound source direction information may be encoded repeatedly, and the instruction information for indicating that the sound source direction information is not changed may be sent to the decoding device; when the sound source direction information of the sound source object is changed compared with the previous sound source direction information, the changed sound source direction information is quantized and encoded, so that the encoding equipment can be prevented from performing unnecessary repeated quantization and encoding on the same sound source direction information, resources are saved, and the quantization and encoding efficiency is improved.
In summary, in the quantization encoding method provided in the embodiments of the present disclosure, the encoding device may determine the target quantization bit number of the sound source orientation information based on at least one of the sound cone information of the sound source object and the region where the sound source object is located relative to the listening object; and carrying out quantization coding on sound source orientation information based on the target quantization bit number to obtain a code stream signal. The cone information of the sound source object and the region where the sound source object is located relative to the listening object may represent the perceived sensitivity of the listening object to the change of the direction of the sound source object, and the target quantization bit number may represent a specific quantization coding mode (such as a fine quantization coding mode or a coarse quantization coding mode). Therefore, the encoding device in the disclosure can pertinently select a proper quantization encoding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is lower, a coarse quantization encoding mode is selected to perform quantization encoding on the sound source direction information, thereby ensuring that the encoding code stream is reduced on the premise of not affecting the subsequent rendering effect and avoiding resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 3a is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by an encoding device, and as shown in fig. 3a, the quantization encoding method may include the following steps:
step 301, determining a first quantization bit number based on sound cone information of a sound source object.
Alternatively, the embodiment of fig. 3a is used to introduce a method of determining the target quantization bit number of sound source orientation information based on cone information of a sound source object and the area where the sound source object is located relative to a listening object.
Alternatively, in one embodiment of the present disclosure, the first quantization bit number may be calculated using the following formula; the formula may include:
wherein in angle Represents the angle of the inner angle of the sound cone, out angle Represents the angle of the outer angle of the sound cone, A max Representing the maximum attenuation amplitude, gap represents the rate of change of sound intensity between the sound cone inner angle and the sound cone outer angle, which is: the rate of change of the intensity of an audio signal emitted by a sound source object with the angle of emergence of the audio signal, percep representing the minimum rate of change of sound intensity perceived by the human ear, square int Representing quantization interval, range represents quantization range, alternatively, quantization range of horizontal angle may be [0 °,360 ] ]The quantization range of the altitude angle can be [0 DEG, 180 DEG ]],quan bit Representing the determined first quantization bit number, ceil represents a round-up function.
Step 302, determining a target quantization bit number based on the region where the sound source object is located relative to the listening object and the first quantization bit number.
Optionally, in an embodiment of the present disclosure, the determining the target quantization bit number based on the region where the sound source object is located with respect to the listening object and the first quantization bit number may include: determining a first quantization bit number as a target quantization bit number when the sound source object is in a first region of the listening object; when the sound source object is in the second region of the listening object, determining a difference between the first quantization bit number and a preset value (the preset value may be 1, for example) as a target quantization bit number; alternatively, the perceptual sensitivity of the listening object to the audio signal of the first region is higher than the perceptual sensitivity of the listening object to the audio signal of the second region, in other words, the first region may be a region where the perceptual sensitivity of the listening object to the audio signal is higher, for example, the first region may be a front region and a rear region of the listening object, and the second region may be a region where the perceptual sensitivity of the listening object to the audio signal is lower, for example, the second region may be a left region and a right region of the listening object.
For example, fig. 3b is a schematic structural diagram of a sound source object in a different area relative to a listening object according to an embodiment of the present disclosure, as shown in fig. 3b, if the first quantization bit #1 is determined based on the sound cone information of the sound source object 1 (i.e. the sound source 1 in fig. 3 b) in the above step 301, and if the first quantization bit #2 is determined based on the sound cone information of the sound source object 2 (i.e. the sound source 2 in fig. 3 b) in the above step 301, wherein, because the sound source object 1 is located in a less sensitive right area (i.e. the area 2 in fig. 3 b) of the listening object (e.g. the human ear), the value after the first quantization bit #1 may be determined as the target quantization bit; and since the sound source object 2 is located in a more sensitive front area of the listening object, e.g. the human ear (i.e. area 3 of fig. 3 b), the first quantization bit number can be determined directly as the target quantization bit number.
As can be seen from the foregoing, in the embodiments of the present disclosure, after the encoding apparatus determines the first quantization bit based on the cone information of the sound source object, the first quantization bit may be further updated based on whether the area where the sound source object is located relative to the listening object is the "perception sensitive area of the listening object to the audio signal", for example, when the area where the sound source object is located relative to the listening object is not the perception sensitive area of the listening object to the audio signal, the difference obtained by subtracting the preset value from the first quantization bit may be determined as the target quantization bit, so as to reduce the target quantization bit, so that the subsequent rendering effect is not affected, and meanwhile, the encoded bitstream is saved to the greatest extent; when the region where the sound source object is located relative to the listening object is a perceptually sensitive region of the listening object to the audio signal, then, to ensure a subsequent rendering effect, the first quantization bit number may be directly determined as the target quantization bit number.
In summary, in the quantization coding method provided by the embodiment of the present disclosure, the coding device may pertinently select an appropriate quantization coding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is low, a coarse quantization coding mode is selected to perform quantization coding on the sound source direction information, so as to ensure that the coding code stream is reduced without affecting the subsequent rendering effect, and avoid resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 4 is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by an encoding device, and as shown in fig. 4, the quantization encoding method may include the following steps:
step 401, determining a target quantization bit number based on sound cone information of a sound source object.
Alternatively, the embodiment of fig. 4 is used to introduce a method of "how the target quantization bit number of the sound source orientation information is determined based on the sound cone information of the sound source object, in particular, when the encoding apparatus cannot acquire the region where the sound source object is located with respect to the listening object".
Alternatively, in one embodiment of the present disclosure, the first quantization bit number may be calculated using the following formula; the formula may include:
wherein in angle Represents the angle of the inner angle of the sound cone, out angle Represents the angle of the outer angle of the sound cone, A max Represents the maximum attenuation amplitude, gap represents the angle of the interior angle of the sound coneA sound intensity rate of change with an angle of an outer cone angle, the sound intensity rate of change being: the rate of change of the intensity of an audio signal emitted by a sound source object with the angle of emergence of the audio signal, percep representing the minimum rate of change of sound intensity perceived by the human ear, square int Representing quantization interval, range represents quantization range, alternatively, quantization range of horizontal angle may be [0 °,360 ]]The quantization range of the altitude angle can be [0 DEG, 180 DEG ]],quan bit Representing the determined target quantization bit number, ceil represents a round-up function.
In summary, in the quantization coding method provided by the embodiment of the present disclosure, the coding device may pertinently select an appropriate quantization coding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is low, a coarse quantization coding mode is selected to perform quantization coding on the sound source direction information, so as to ensure that the coding code stream is reduced without affecting the subsequent rendering effect, and avoid resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 5 is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by an encoding device, and as shown in fig. 5, the quantization encoding method may include the following steps:
step 501, determining a corresponding codebook based on the target quantization bit number.
Alternatively, the embodiment of fig. 5 is used to introduce a method of how the encoding apparatus specifically performs quantization encoding on sound source orientation information based on the target quantization bit number.
Optionally, in one embodiment of the disclosure, different codebooks correspond to different quantization bits, respectively. Optionally, at least one codeword may be included in each codebook, where each codeword corresponds to a bit value.
Alternatively, in one embodiment of the present disclosure, the encoding device may determine a codebook corresponding to a target quantization bit number for subsequent quantization of encoded sound source orientation information based on the codebook.
And 502, performing quantization coding on the sound source orientation information based on the codebook to obtain the quantized coded sound source orientation information.
Optionally, the above-mentioned performing quantization encoding on the sound source orientation information based on the codebook to obtain the quantized encoded sound source orientation information may include at least one of the following methods:
First, the quantization interval corresponding to the target quantization bit number is determined (i.e., the square in the foregoing embodiment ibt ) And quantizing (e.g., uniformly quantizing) the quantization range of the sound source orientation information based on the quantization interval to obtain at least one quantized value, determining a quantized value closest to the value of the sound source orientation information in the at least one quantized value as a target quantized value, and determining a bit value corresponding to a codeword closest to the target quantized value in the determined codebook as quantized encoded sound source orientation information.
Alternatively, the "closest" may be, for example: the absolute value of the difference is the smallest.
For example, assume that the sound source orientation information is 35 ° in the altitude of the sound source object, and that the quantization interval determined in the above step 301 is 20 °, the codebook determined in the above step 501 is: (0 °,45 °,90 °,135 °,180 °), wherein bit values corresponding to each codeword in the codebook are in turn: 000. 001, 010, 011, 100. The process of the above-mentioned first method for quantitatively encoding the altitude angle 35 ° of the sound source object may be: firstly, based on the quantization interval 20 degrees, the quantization range [0 degrees, 180 degrees ] of the height angle is quantized to obtain nine quantized values, wherein the nine quantized values are respectively as follows: 0 °, 20 °, 40 °, 60 °, 80 °, 100 °, 120 °, 140 °, 160 °,180 °. The quantized value closest to the sound source object at a distance of 35 ° from the altitude angle of the sound source object among the nine quantized values is: 40 °, i.e., the target quantization value is 40 °. And, the codeword closest to the target quantized value 40 ° in the codebook (0 °,45 °,90 °,135 °,180 °) is 45 °, whereby the bit value 001 corresponding to the codeword 45 ° can be determined as quantized encoded sound source orientation information.
And second, determining a bit value corresponding to a codeword with the closest value distance to the sound source orientation information in the determined codebook as quantized encoded sound source orientation information.
By way of example, assume that the sound source orientation information is 35 ° of the altitude of the sound source object, and assume that the codebook determined in step 501 is: (0 °,45 °,90 °,135 °,180 °), wherein bit values corresponding to each codeword in the codebook are in turn: 000. 001, 010, 011, 100. The process of the 35 DEG quantization coding of the altitude angle of the sound source object by the second method can be as follows: from the codebook (0 °,45 °,90 °,135 °,180 °), it is determined that the codeword closest to the altitude angle 35 ° of the sound source object is 45 °, whereby the bit value 001 corresponding to the codeword 45 ° can be determined as quantized encoded sound source orientation information.
In summary, in the quantization coding method provided by the embodiment of the present disclosure, the coding device may pertinently select an appropriate quantization coding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is low, a coarse quantization coding mode is selected to perform quantization coding on the sound source direction information, so as to ensure that the coding code stream is reduced without affecting the subsequent rendering effect, and avoid resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 6 is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by an encoding device, and as shown in fig. 6, the quantization encoding method may include the following steps:
step 601, the code stream signal is sent to a decoding device.
Optionally, in one embodiment of the disclosure, the code stream signal may include at least one of:
an encoded audio signal;
quantizing the encoded sound source orientation information;
a target quantization bit number;
cone information of the sound source object;
an indication index corresponding to the sound cone information;
absolute position of sound source object;
an absolute position of the listening object;
the relative position of the sound source object with respect to the listening object.
Alternatively, in one embodiment of the present disclosure, after the encoding device obtains the bitstream signal, the bitstream signal may be transmitted to the decoding device, so that the decoding device may decode and inverse-quantize the bitstream signal to obtain sound source orientation information and a decoded audio signal, and further render and play the decoded audio signal to a listening object based on the sound source orientation information.
In summary, in the quantization coding method provided by the embodiment of the present disclosure, the coding device may pertinently select an appropriate quantization coding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is low, a coarse quantization coding mode is selected to perform quantization coding on the sound source direction information, so as to ensure that the coding code stream is reduced without affecting the subsequent rendering effect, and avoid resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 7 is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by a decoding device, and as shown in fig. 7, the quantization encoding method may include the following steps:
step 701, receiving a code stream signal sent by an encoding device.
Alternatively, the code stream signal may be: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object.
The code stream signal may include at least one of:
an encoded audio signal;
quantizing the encoded sound source orientation information;
a target quantization bit number;
cone information of the sound source object;
an indication index corresponding to the sound cone information;
absolute position of sound source object;
an absolute position of the listening object;
the relative position of the sound source object with respect to the listening object.
Wherein a detailed description of step 701 may be described with reference to the previous embodiments.
Step 702, decoding and inverse-quantizing the code stream signal to obtain sound source orientation information.
Optionally, in one embodiment of the disclosure, the method of decoding and inverse quantizing a bitstream signal to obtain sound source orientation information may include at least one of:
In the first method, when the code stream signal acquired by the decoding device includes the target quantization bit number, the decoding device may directly decode and inverse-quantize the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number to restore the sound source orientation information.
Optionally, the method for decoding and inverse-quantizing the quantized coded sound source orientation information in the bitstream signal based on the target quantization bit number may include: and determining a corresponding codebook based on the target quantization bit number, and decoding and inverse-quantizing the quantized and encoded sound source orientation information based on the codebook.
It should be noted that, the process of decoding and inverse-quantizing the quantized and encoded sound source orientation information in the bitstream signal by the decoding device based on the target quantization bit number is substantially the inverse process of quantizing and encoding the sound source orientation information by the encoding device based on the target quantization bit number, and the detailed flow of this portion may be described with reference to the foregoing embodiment, which is not repeated herein.
In the second method, when the code stream signal acquired by the decoding device does not include the target quantization bit number, the decoding device may determine the target quantization bit number first, and then decode and inverse-quantize the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number to restore the sound source orientation information.
Optionally, the method for determining the target quantization bit number by the decoding device may include: firstly, determining sound cone information of a sound source object based on information (such as sound cone information of the sound source object and/or indication index corresponding to the sound cone information) included in a code stream signal; then determining the area of the sound source object relative to the listening object; and determining a target quantization bit number of the sound source orientation information based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object.
Optionally, the method for determining the area where the sound source object is located relative to the listening object by the decoding device may include at least one of the following:
the decoding device determines an area in which the sound source object is located with respect to the listening object based on information included in the code stream signal (e.g., at least one of an absolute position of the sound source object, an absolute position of the listening object, a relative position of the sound source object with respect to the listening object) and an orientation of the listening object;
the decoding device autonomously determines the region in which the sound source object is located relative to the listening object. For example, the decoding device autonomously determines at least one of an absolute position of the sound source object, an absolute position of the listening object, a relative position of the sound source object with respect to the listening object, and determines an area in which the sound source object is located with respect to the listening object in combination with an orientation of the listening object.
Alternatively, the above detailed process of "determining the target quantization bit number of the sound source orientation information based on at least one of the sound cone information of the sound source object, the region in which the sound source object is located with respect to the listening object", and "decoding and inverse-quantizing the quantized encoded sound source orientation information in the bitstream signal based on the target quantization bit number" to restore the sound source orientation information "may be described with reference to the foregoing embodiments.
In summary, in the quantization encoding method provided in the embodiments of the present disclosure, the decoding device receives the code stream signal sent by the encoding device; the code stream signal is: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object; the decoding device then decodes and inverse quantizes the code stream signal to derive sound source orientation information. Therefore, the encoding device in the disclosure can pertinently select a proper quantization encoding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is lower, a coarse quantization encoding mode is selected to perform quantization encoding on the sound source direction information, thereby ensuring that the encoding code stream is reduced on the premise of not affecting the subsequent rendering effect and avoiding resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 8 is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by a decoding device, and as shown in fig. 8, the quantization encoding method may include the following steps:
step 801, when the target quantization bit number is included in the code stream signal, decoding and inverse-quantizing the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number.
Wherein a detailed description of step 801 may be described with reference to the previous embodiments.
In summary, in the quantization coding method provided by the embodiment of the present disclosure, the coding device may pertinently select an appropriate quantization coding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is low, a coarse quantization coding mode is selected to perform quantization coding on the sound source direction information, so as to ensure that the coding code stream is reduced without affecting the subsequent rendering effect, and avoid resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 9 is a flowchart of a quantization encoding method provided in an embodiment of the present disclosure, where the method is performed by a decoding device, and as shown in fig. 9, the quantization encoding method may include the following steps:
step 901, when the target quantization bit number is not included in the code stream signal, determining sound cone information of the sound source object based on information included in the code stream signal, determining an area where the sound source object is located relative to the listening object, and determining the target quantization bit number of the sound source orientation information based on at least one of the sound cone information of the sound source object and the area where the sound source object is located relative to the listening object.
And step 902, decoding and inverse-quantizing the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number.
Wherein a detailed description of steps 901-902 may be described with reference to the previous embodiments.
In summary, in the quantization coding method provided by the embodiment of the present disclosure, the coding device may pertinently select an appropriate quantization coding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is low, a coarse quantization coding mode is selected to perform quantization coding on the sound source direction information, so as to ensure that the coding code stream is reduced without affecting the subsequent rendering effect, and avoid resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Fig. 10 is a schematic structural diagram of a communication device according to an embodiment of the disclosure, where, as shown in fig. 10, the device may include:
a processing module for determining a target quantization bit number of the sound source orientation information based on at least one of sound cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
the processing module is further configured to perform quantization encoding on the sound source orientation information based on the target quantization bit number, and obtain a code stream signal.
In summary, in the communication apparatus provided in the embodiments of the present disclosure, the encoding device may determine the target quantization bit number of the sound source orientation information based on at least one of the sound cone information of the sound source object and the region where the sound source object is located with respect to the listening object; and carrying out quantization coding on sound source orientation information based on the target quantization bit number to obtain a code stream signal. The cone information of the sound source object and the region where the sound source object is located relative to the listening object may represent the perceived sensitivity of the listening object to the change of the direction of the sound source object, and the target quantization bit number may represent a specific quantization coding mode (such as a fine quantization coding mode or a coarse quantization coding mode). Therefore, the encoding device in the disclosure can pertinently select a proper quantization encoding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is lower, a coarse quantization encoding mode is selected to perform quantization encoding on the sound source direction information, thereby ensuring that the encoding code stream is reduced on the premise of not affecting the subsequent rendering effect and avoiding resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Optionally, in one embodiment of the disclosure, the cone information of the sound source object includes at least one of:
the sound cone inner angle of the sound source object;
the cone outer angle of the sound source object;
maximum attenuation amplitude of an audio signal emitted from a sound source object.
Optionally, in one embodiment of the disclosure, the sound source orientation information includes at least one of:
a horizontal angle between the sound source object and the listening object;
the altitude angle between the sound source object and the listening object.
Optionally, in one embodiment of the disclosure, the apparatus is further configured to:
determining an orientation of the listening object;
determining a relative position of the sound source object with respect to the listening object;
an area in which the sound source object is located relative to the listening object is determined based on at least one of the orientation of the listening object and the relative position.
Optionally, in one embodiment of the disclosure, the processing module is further configured to:
determining a first quantization bit number based on sound cone information of the sound source object;
the target quantization bit number is determined based on the region in which the sound source object is located relative to the listening object and the first quantization bit number.
Optionally, in one embodiment of the disclosure, the processing module is further configured to:
a target quantization bit number is determined based on the sound cone information of the sound source object.
Optionally, in one embodiment of the disclosure, the processing module is further configured to:
calculating the first quantization bit number or the target quantization bit number using the following formula; the formula includes:
wherein in angle Represents the angle of the inner angle of the sound cone, out angle Represents the angle of the outer angle of the sound cone, A max Representing the maximum attenuation amplitude, gap represents the rate of change of sound intensity between the sound cone inner angle and the sound cone outer angle, the rate of change of sound intensity being: the change rate of the intensity of the audio signal emitted by the sound source object along with the emergent angle of the audio signal, wherein the percentage represents the minimum sound intensity change rate perceived by human ears, and square int Represents quantization interval, range represents quantization range, square bit Representing the determined first quantization bit number or target quantization bit number, ceil represents a round-up function.
Optionally, in one embodiment of the disclosure, the processing module is further configured to:
determining the first quantization bit number as the target quantization bit number when the sound source object is in a first region of the listening object;
When the sound source object is in the second area of the listening object, determining the difference value between the first quantization bit number and a preset value as the target quantization bit number;
wherein the perceptual sensitivity of the listening object to the audio signals of the first region is higher than the perceptual sensitivity of the listening object to the audio signals of the second region.
Optionally, in one embodiment of the disclosure, the processing module is further configured to:
determining a corresponding codebook based on the target quantization bit number;
and carrying out quantization coding on the sound source orientation information based on the codebook to obtain the sound source orientation information after quantization coding.
Optionally, in one embodiment of the disclosure, the code stream signal includes at least one of:
quantizing the encoded sound source orientation information;
the target quantization bit number;
sound cone information of the sound source object;
an indication index corresponding to the sound cone information;
an absolute position of the sound source object;
an absolute position of the listening object;
the relative position of the sound source object with respect to the listening object.
Optionally, in one embodiment of the disclosure, the apparatus is further configured to:
and transmitting the code stream signal to a decoding device.
Fig. 11 is a schematic structural diagram of a communication device according to an embodiment of the present disclosure, where, as shown in fig. 11, the device may include:
the receiving and transmitting module is used for receiving the code stream signal sent by the encoding equipment; the code stream signal is: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
and the processing module is used for decoding and inverse quantizing the code stream signal to obtain the sound source orientation information.
In summary, in the communication apparatus provided in the embodiments of the present disclosure, the decoding device receives the code stream signal sent by the encoding device; the code stream signal is: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object; the decoding device then decodes and inverse quantizes the code stream signal to derive sound source orientation information. Therefore, the encoding device in the disclosure can pertinently select a proper quantization encoding mode based on the current perceived sensitivity of the listening object to the change of the direction of the sound source object, so that when the perceived sensitivity of the listening object to the change of the direction of the sound source object is lower, a coarse quantization encoding mode is selected to perform quantization encoding on the sound source direction information, thereby ensuring that the encoding code stream is reduced on the premise of not affecting the subsequent rendering effect and avoiding resource waste; when the perception sensitivity of the listening object to the change of the direction of the sound source object is higher, a fine quantization coding mode is selected to perform quantization coding on the sound source direction information so as to ensure the subsequent rendering effect. Meanwhile, the quantization coding method disclosed by the invention has higher flexibility.
Optionally, in one embodiment of the disclosure, the code stream signal includes at least one of:
quantizing the encoded sound source orientation information;
the target quantization bit number;
sound cone information of the sound source object;
an indication index corresponding to the sound cone information;
an absolute position of the sound source object;
an absolute position of the listening object;
the relative position of the sound source object with respect to the listening object.
Optionally, in one embodiment of the disclosure, when the target quantization bit number is included in the code stream signal, the processing module is further configured to:
and decoding and inverse-quantizing the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number.
Optionally, in one embodiment of the disclosure, when the target quantization bit number is not included in the code stream signal, the processing module is further configured to:
determining cone information of the sound source object based on information included in the code stream signal;
determining the region of the sound source object relative to the listening object;
determining a target quantization bit number of sound source orientation information based on at least one of sound cone information of the sound source object and an area where the sound source object is located relative to the listening object;
And decoding and inverse-quantizing the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number.
Optionally, in one embodiment of the disclosure, the processing module is further configured to at least one of:
determining an area where the sound source object is located relative to the listening object based on information included in the code stream signal and an orientation of the listening object;
the decoding device autonomously determines the region in which the sound source object is located relative to the listening object.
Optionally, in one embodiment of the disclosure, the processing module is further configured to:
determining a corresponding codebook based on the target quantization bit number;
and decoding and inverse-quantizing the quantized and encoded sound source orientation information based on the codebook.
Referring to fig. 12, fig. 12 is a schematic structural diagram of a communication device 1200 according to an embodiment of the disclosure. The communication device 1200 may be a base station, a terminal, a chip system, a processor, or the like that supports the base station to implement the above method, or a chip, a chip system, a processor, or the like that supports the terminal to implement the above method. The device can be used for realizing the method described in the method embodiment, and can be particularly referred to the description in the method embodiment.
The communications apparatus 1200 can include one or more processors 1201. The processor 1201 may be a general purpose processor, a special purpose processor, or the like. For example, a baseband processor or a central processing unit. The baseband processor may be used to process communication protocols and communication data, and the central processor may be used to control communication devices (e.g., base stations, baseband chips, terminals, terminal chips, DUs or CUs, etc.), execute computer programs, and process data of the computer programs.
Optionally, the communication device 1200 may further include one or more memories 1202, on which a computer program 1204 may be stored, and the processor 1201 executes the computer program 1204, so that the communication device 1200 performs the method described in the above method embodiments. Optionally, the memory 1202 may also store data. The communication device 1200 and the memory 1202 may be provided separately or may be integrated.
Optionally, the communication device 1200 may further include a transceiver 1205, an antenna 1206. The transceiver 1205 may be referred to as a transceiver unit, transceiver circuitry, or the like, for implementing a transceiver function. The transceiver 1205 may include a receiver, which may be referred to as a receiver or a receiving circuit, etc., for implementing a receiving function; the transmitter may be referred to as a transmitter or a transmitting circuit, etc., for implementing a transmitting function.
Optionally, one or more interface circuits 1207 may also be included in the communications device 1200. The interface circuit 1207 is configured to receive code instructions and transmit the code instructions to the processor 1201. The processor 1201 executes the code instructions to cause the communication device 1200 to perform the method described in the method embodiments described above.
In one implementation, a transceiver for implementing the receive and transmit functions may be included in the processor 1201. For example, the transceiver may be a transceiver circuit, or an interface circuit. The transceiver circuitry, interface or interface circuitry for implementing the receive and transmit functions may be separate or may be integrated. The transceiver circuit, interface or interface circuit may be used for reading and writing codes/data, or the transceiver circuit, interface or interface circuit may be used for transmitting or transferring signals.
In one implementation, the processor 1201 may store a computer program 1203, where the computer program 1203 runs on the processor 1201, and may cause the communication apparatus 1200 to perform the method described in the above method embodiment. The computer program 1203 may be solidified in the processor 1201, in which case the processor 1201 may be implemented in hardware.
In one implementation, the communication apparatus 1200 may include circuitry that may implement the functions of transmitting or receiving or communicating in the foregoing method embodiments. The processors and transceivers described in this disclosure may be implemented on integrated circuits (integrated circuit, ICs), analog ICs, radio frequency integrated circuits RFICs, mixed signal ICs, application specific integrated circuits (application specific integrated circuit, ASIC), printed circuit boards (printed circuit board, PCB), electronic devices, and the like. The processor and transceiver may also be fabricated using a variety of IC process technologies such as complementary metal oxide semiconductor (complementary metal oxide semiconductor, CMOS), N-type metal oxide semiconductor (NMOS), P-type metal oxide semiconductor (positive channel metal oxide semiconductor, PMOS), bipolar junction transistor (bipolar junction transistor, BJT), bipolar CMOS (BiCMOS), silicon germanium (SiGe), gallium arsenide (GaAs), etc.
The communication apparatus in the above embodiment description may be a base station or a terminal, but the scope of the communication apparatus described in the present disclosure is not limited thereto, and the structure of the communication apparatus may not be limited by fig. 12. The communication means may be a stand-alone device or may be part of a larger device. For example, the communication device may be:
(1) A stand-alone integrated circuit IC, or chip, or a system-on-a-chip or subsystem;
(2) A set of one or more ICs, optionally including storage means for storing data, a computer program;
(3) An ASIC, such as a Modem (Modem);
(4) Modules that may be embedded within other devices;
(5) Receivers, terminals, smart terminals, cellular telephones, wireless devices, handsets, mobile units, vehicle devices, base stations, cloud devices, artificial intelligence devices, etc.;
(6) Others, and so on.
For the case where the communication device may be a chip or a chip system, reference may be made to the schematic structural diagram of the chip shown in fig. 13. The chip shown in fig. 13 includes a processor 1301 and an interface 1302. Wherein the number of processors 1301 may be one or more, and the number of interfaces 1302 may be a plurality.
Optionally, the chip further comprises a memory 1303, the memory 1303 being configured to store necessary computer programs and data.
Those of skill in the art will further appreciate that the various illustrative logical blocks (illustrative logical block) and steps (step) described in connection with the embodiments of the disclosure may be implemented by electronic hardware, computer software, or combinations of both. Whether such functionality is implemented as hardware or software depends upon the particular application and design requirements of the overall system. Those skilled in the art may implement the described functionality in varying ways for each particular application, but such implementation is not to be understood as beyond the scope of the embodiments of the present disclosure.
The present disclosure also provides a readable storage medium having instructions stored thereon which, when executed by a computer, perform the functions of any of the method embodiments described above.
The present disclosure also provides a computer program product which, when executed by a computer, performs the functions of any of the method embodiments described above.
In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product comprises one or more computer programs. When the computer program is loaded and executed on a computer, the flow or functions described in accordance with the embodiments of the present disclosure are produced in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable apparatus. The computer program may be stored in or transmitted from one computer readable storage medium to another, for example, by wired (e.g., coaxial cable, optical fiber, digital subscriber line (digital subscriber line, DSL)) or wireless (e.g., infrared, wireless, microwave, etc.) means from one website, computer, server, or data center. The computer readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that contains an integration of one or more available media. The usable medium may be a magnetic medium (e.g., a floppy disk, a hard disk, a magnetic tape), an optical medium (e.g., a high-density digital video disc (digital video disc, DVD)), or a semiconductor medium (e.g., a Solid State Disk (SSD)), or the like.
Those of ordinary skill in the art will appreciate that: the various numbers of first, second, etc. referred to in this disclosure are merely for ease of description and are not intended to limit the scope of embodiments of this disclosure, nor to indicate sequencing.
At least one of the present disclosure may also be described as one or more, a plurality may be two, three, four or more, and the present disclosure is not limited. In the embodiment of the disclosure, for a technical feature, the technical features in the technical feature are distinguished by "first", "second", "third", "a", "B", "C", and "D", and the technical features described by "first", "second", "third", "a", "B", "C", and "D" are not in sequence or in order of magnitude.
The correspondence relationships shown in the tables in the present disclosure may be configured or predefined. The values of the information in each table are merely examples, and may be configured as other values, and the present disclosure is not limited thereto. In the case of the correspondence between the configuration information and each parameter, it is not necessarily required to configure all the correspondence shown in each table. For example, in the table in the present disclosure, the correspondence shown by some rows may not be configured. For another example, appropriate morphing adjustments, e.g., splitting, merging, etc., may be made based on the tables described above. The names of the parameters indicated in the tables may be other names which are understood by the communication device, and the values or expressions of the parameters may be other values or expressions which are understood by the communication device. When the tables are implemented, other data structures may be used, for example, an array, a queue, a container, a stack, a linear table, a pointer, a linked list, a tree, a graph, a structure, a class, a heap, a hash table, or a hash table.
Predefined in this disclosure may be understood as defining, predefining, storing, pre-negotiating, pre-configuring, curing, or pre-sintering.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present disclosure.
It will be clear to those skilled in the art that, for convenience and brevity of description, specific working procedures of the above-described systems, apparatuses and units may refer to corresponding procedures in the foregoing method embodiments, and are not repeated herein.
The foregoing is merely specific embodiments of the disclosure, but the protection scope of the disclosure is not limited thereto, and any person skilled in the art can easily think about changes or substitutions within the technical scope of the disclosure, and it is intended to cover the scope of the disclosure. Therefore, the protection scope of the present disclosure shall be subject to the protection scope of the claims.

Claims (22)

1. A quantization encoding method, the method performed by an encoding device, the method comprising:
determining a target quantization bit number of sound source orientation information based on at least one of sound cone information of a sound source object, and an area in which the sound source object is located with respect to a listening object;
and carrying out quantization coding on the sound source orientation information based on the target quantization bit number to obtain a code stream signal.
2. The method of claim 1, wherein the cone information of the sound source object comprises at least one of:
the sound cone inner angle of the sound source object;
the cone outer angle of the sound source object;
maximum attenuation amplitude of an audio signal emitted from a sound source object.
3. The method of claim 1, wherein the sound source orientation information comprises at least one of:
a horizontal angle between the sound source object and the listening object;
the altitude angle between the sound source object and the listening object.
4. The method of claim 1, wherein the method further comprises:
determining an orientation of the listening object;
determining a relative position of the sound source object with respect to the listening object;
an area in which the sound source object is located relative to the listening object is determined based on at least one of the orientation of the listening object and the relative position.
5. The method of any of claims 1-4, wherein determining the target quantization bit number of the sound source orientation information based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object, comprises:
determining a first quantization bit number based on sound cone information of the sound source object;
the target quantization bit number is determined based on the region in which the sound source object is located relative to the listening object and the first quantization bit number.
6. The method of any of claims 1-4, wherein determining the target quantization bit number of the sound source orientation information based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object, comprises:
a target quantization bit number is determined based on the sound cone information of the sound source object.
7. The method of claim 5 or 6, wherein the determining a first quantization bit number or a target quantization bit number based on the sound cone information of the sound source object comprises:
calculating the first quantization bit number or the target quantization bit number using the following formula; the formula includes:
wherein in angle Represents the angle of the inner angle of the sound cone, out angle Represents the angle of the outer angle of the sound cone, A max Representing the maximum attenuation amplitude, gap represents the rate of change of sound intensity between the sound cone inner angle and the sound cone outer angle, the rate of change of sound intensity being: the change rate of the intensity of the audio signal emitted by the sound source object along with the emergent angle of the audio signal, wherein the percentage represents the minimum sound intensity change rate perceived by human ears, and square int Represents quantization interval, range represents quantization range, square bit Representing the determined first quantization bit number or target quantization bit number, ceil represents a round-up function.
8. The method of claim 5, wherein the determining the target quantization bit number based on the region in which the sound source object is located relative to a listening object and the first quantization bit number comprises:
determining the first quantization bit number as the target quantization bit number when the sound source object is in a first region of the listening object;
when the sound source object is in the second area of the listening object, determining the difference value between the first quantization bit number and a preset value as the target quantization bit number;
wherein the perceptual sensitivity of the listening object to the audio signals of the first region is higher than the perceptual sensitivity of the listening object to the audio signals of the second region.
9. The method according to any one of claims 1-8, wherein said quantizing said sound source orientation information based on said target quantization bit number comprises:
determining a corresponding codebook based on the target quantization bit number;
and carrying out quantization coding on the sound source orientation information based on the codebook to obtain the sound source orientation information after quantization coding.
10. The method of any of claims 1-9, wherein the code stream signal comprises at least one of:
quantizing the encoded sound source orientation information;
the target quantization bit number;
sound cone information of the sound source object;
an indication index corresponding to the sound cone information;
an absolute position of the sound source object;
an absolute position of the listening object;
the relative position of the sound source object with respect to the listening object.
11. The method of any one of claims 1-10, wherein the method further comprises:
and transmitting the code stream signal to a decoding device.
12. A quantization encoding method, the method performed by a decoding device, the method comprising:
receiving a code stream signal sent by coding equipment; the code stream signal is: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
And decoding and inversely quantizing the code stream signal to obtain the sound source orientation information.
13. The method of claim 12, wherein the code stream signal comprises at least one of:
quantizing the encoded sound source orientation information;
the target quantization bit number;
sound cone information of the sound source object;
an indication index corresponding to the sound cone information;
an absolute position of the sound source object;
an absolute position of the listening object;
the relative position of the sound source object with respect to the listening object.
14. The method of claim 12 or 13, wherein said decoding and inverse quantizing the code stream signal when the target quantization bit number is included in the code stream signal comprises:
and decoding and inverse-quantizing the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number.
15. The method of claim 12 or 13, wherein said decoding and inverse quantizing the code stream signal when the target quantization bit number is not included in the code stream signal comprises:
determining cone information of the sound source object based on information included in the code stream signal;
Determining the region of the sound source object relative to the listening object;
determining a target quantization bit number of sound source orientation information based on at least one of sound cone information of the sound source object and an area where the sound source object is located relative to the listening object;
and decoding and inverse-quantizing the quantized and encoded sound source orientation information in the code stream signal based on the target quantization bit number.
16. The method of claim 15, wherein the determining the region in which the sound source object is located relative to the listening object comprises at least one of:
determining an area where the sound source object is located relative to the listening object based on information included in the code stream signal and an orientation of the listening object;
the decoding device autonomously determines the region in which the sound source object is located relative to the listening object.
17. The method according to claim 14 or 15, wherein said decoding and inverse-quantizing the quantized coded sound source orientation information in the bitstream signal based on the target quantization bit number comprises:
determining a corresponding codebook based on the target quantization bit number;
and decoding and inverse-quantizing the quantized and encoded sound source orientation information based on the codebook.
18. A communication apparatus, comprising:
a processing module for determining a target quantization bit number of the sound source orientation information based on at least one of sound cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
the processing module is further configured to perform quantization encoding on the sound source orientation information based on the target quantization bit number, and obtain a code stream signal.
19. A communication apparatus, comprising:
the receiving and transmitting module is used for receiving the code stream signal sent by the encoding equipment; the code stream signal is: the sound source orientation information is obtained after quantization coding is carried out on the basis of the target quantization bit number; the target quantization bit number is: based on at least one of cone information of the sound source object, an area in which the sound source object is located with respect to the listening object;
and the processing module is used for decoding and inverse quantizing the code stream signal to obtain the sound source orientation information.
20. A communication device, characterized in that the device comprises a processor and a memory, wherein the memory has stored therein a computer program, which processor executes the computer program stored in the memory to cause the device to perform the method according to any one of claims 1 to 11, or which processor executes the computer program stored in the memory to cause the device to perform the method according to any one of claims 12 to 17.
21. A communication device, comprising: processor and interface circuit, wherein
The interface circuit is used for receiving code instructions and transmitting the code instructions to the processor;
the processor being configured to execute the code instructions to perform the method of any one of claims 1 to 11 or to execute the code instructions to perform the method of any one of claims 12 to 17.
22. A computer readable storage medium storing instructions which, when executed, cause the method of any one of claims 1 to 11 to be implemented or which, when executed, cause the method of any one of claims 12 to 17 to be implemented.
CN202380008765.7A 2023-03-27 2023-03-27 Quantization coding method, device, equipment and storage medium Pending CN116724351A (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2023084194 2023-03-27

Publications (1)

Publication Number Publication Date
CN116724351A true CN116724351A (en) 2023-09-08

Family

ID=87866587

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202380008765.7A Pending CN116724351A (en) 2023-03-27 2023-03-27 Quantization coding method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN116724351A (en)

Similar Documents

Publication Publication Date Title
CN104980788B (en) Video encoding/decoding method and device
US11496238B2 (en) Data transmission method and communications device
KR20190092067A (en) Apparatus and method for constructing polar code for multiple input multiple output channel
CN116348952A (en) Audio signal processing device, equipment and storage medium
CN115552518A (en) Signal encoding and decoding method and device, user equipment, network side equipment and storage medium
KR20220062621A (en) Spatial audio parameter encoding and related decoding
CN116724351A (en) Quantization coding method, device, equipment and storage medium
CN116368460A (en) Audio processing method and device
CN117158089A (en) Method and device for transmitting and receiving uplink channel
CN114125457A (en) Signaling of rectangular stripes in VVC video compression standards
WO2024108449A1 (en) Signal quantization method, apparatus, device, and storage medium
US20240072827A1 (en) Encoding/decoding method, communication apparatus, and system
KR20230084587A (en) Communication method and device
WO2024050673A1 (en) Audio signal frequency band extension method and apparatus, device, and storage medium
WO2023197187A1 (en) Channel state information processing methods and apparatuses
WO2024082195A1 (en) Ai model-based terminal positioning method and apparatus
WO2024098221A1 (en) Audio signal rendering method, apparatus, device, and storage medium
WO2024082196A1 (en) Terminal positioning method and apparatus based on ai model
WO2024031713A1 (en) Uplink 8-port codebook generation method and apparatus, device, and storage medium
WO2023245587A1 (en) Validation time acquisition method and apparatus
WO2022051897A1 (en) Encoding method and apparatus
WO2024020904A1 (en) Method for sending phase shift configuration of intelligent reflecting surface (irs), method for receiving phase shift configuration of irs, and apparatus
CN117836785A (en) Model generation method and device
CN118120189A (en) Audio quality testing method, device, equipment and storage medium
CN115956390A (en) Information transmission method, device, equipment and chip system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination