CN117479076A - Sound effect adjusting method and device, electronic equipment and storage medium - Google Patents

Sound effect adjusting method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN117479076A
CN117479076A CN202311482036.3A CN202311482036A CN117479076A CN 117479076 A CN117479076 A CN 117479076A CN 202311482036 A CN202311482036 A CN 202311482036A CN 117479076 A CN117479076 A CN 117479076A
Authority
CN
China
Prior art keywords
sound
sound effect
user
module
emotion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311482036.3A
Other languages
Chinese (zh)
Inventor
徐海
冯少瑜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Desam Audio Co ltd
Original Assignee
Guangzhou Desam Audio Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Desam Audio Co ltd filed Critical Guangzhou Desam Audio Co ltd
Priority to CN202311482036.3A priority Critical patent/CN117479076A/en
Publication of CN117479076A publication Critical patent/CN117479076A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/20Arrangements for obtaining desired frequency or directional characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/65Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/27Regression, e.g. linear or logistic regression
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Signal Processing (AREA)
  • Acoustics & Sound (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Stereophonic System (AREA)

Abstract

The invention relates to the technical field of sound effect adjusting methods, in particular to a sound effect adjusting method, a sound effect adjusting device, electronic equipment and a storage medium, which comprise the following steps: the sound system acquires the environmental acoustic characteristics through an environmental perception technology. According to the invention, through the sensing sound effect technology and the emotion analysis algorithm, the sound system can more accurately identify the emotion state and the perception requirement of the user, the user can better establish emotion connection with the audio content, the infectivity of the audio experience is enhanced, the inter-sound modeling technology enables the audio to present the sensation of a physical object in the auditory space, the user can freely move in an unrestricted environment, and more dispersed and omnibearing audio perception is obtained, so that more immersive and immersive audio experience is brought to the user, and through the intelligent sound effect adaptability based on the environment and the audio content, the sound system can automatically adjust sound effect parameters to adapt to different audio contents and user preferences, so that the optimal audio experience is provided.

Description

Sound effect adjusting method and device, electronic equipment and storage medium
Technical Field
The present invention relates to the technical field of sound effect adjustment methods, and in particular, to a sound effect adjustment method, a sound effect adjustment device, an electronic device, and a storage medium.
Background
The sound effect adjusting method includes adjusting frequency through an equalizer, selecting an appropriate sound field mode, adjusting volume balance, setting surround sound effect, using sound effect devices, and the like. Bass or treble effects can be enhanced by the equalizer, surround sound effects can be changed by sound stage mode selection, volume balance adjustment ensures proper proportions of the individual sound elements, and surround sound settings can increase the stereo perception of music or movies. In addition, sound effects can also be adjusted using sound effect devices such as reverberators and sound processors.
In the existing sound effect adjusting method, the conventional method generally adopts a fixed sound effect setting or preset mode, and cannot meet the personalized requirements of users. Different users have different preferences and demands for sound effects, and the traditional method cannot provide personalized adjustment options. The system sound effect adjustment method may have limitations in providing surround sound effects and realism. The traditional method is difficult to realize real sound scene reproduction, and limits the immersion and the immersive experience of the user. In conventional approaches, the user is required to manually adjust the volume, sound setting, and sound effect parameters, which may require a certain skill and experience. It may be confusing and inconvenient for a non-professional user. The conventional method is poorly adaptable to different audio contents and environments. The user experience may be affected because of the inability to automatically adjust to the characteristics of the audio content and the acoustic characteristics of the environment.
Disclosure of Invention
The invention aims to solve the defects in the prior art and provides a sound effect adjusting method, a sound effect adjusting device, electronic equipment and a storage medium.
In order to achieve the above purpose, the present invention adopts the following technical scheme: an acoustic effect adjusting method, comprising the steps of:
the sound system acquires environmental acoustic characteristics through an environmental perception technology, and recognizes the emotion state and perception requirement of a user by using an emotion analysis algorithm;
based on the acoustic characteristics, the emotion states and the perception requirements as input items, the sound system automatically adjusts sound effect parameters;
the sound system establishes a sound scene model according to the positioning and simulation of the audio source by means of a space sound modeling technology;
the sound system automatically adjusts sound parameters based on the intelligent sound adaptability of the audio content and the user demand;
the user interacts with the sound system to adjust sound parameters or to select different sound modes.
As a further scheme of the invention, the sound system acquires environmental acoustic characteristics through an environmental perception technology, and the steps of identifying the emotion state and the perception requirement of the user by using an emotion analysis algorithm are specifically as follows:
the sound system captures sound data in the environment by using microphone equipment, the audio data comprise music, human voice and noise, and the audio data are subjected to audio enhancement processing comprising denoising, echo suppression and gain control to obtain processed audio data;
acquiring spectral features from the processed audio data by using a digital signal processing technology, particularly a fast Fourier transform, calculating the spectral features based on Mel spectral coefficients, capturing human ear perception characteristics of the audio, and calculating the environmental acoustic features;
extracting emotion characteristics related to emotion based on the processed audio data, establishing an emotion classifier based on the emotion characteristics by utilizing a random forest, and performing emotion classification on the extracted emotion characteristics by using the emotion classifier to obtain the emotion state;
collecting feedback data, history records and other contextual information of a user, and arranging emotion feedback, use preference and use scenes;
based on the emotion feedback, the use preference and the use scene, adopting a decision tree algorithm to establish an emotion model, and based on expert domain knowledge and the emotion model, formulating a set of rule expression to correlate the emotion state with the perception requirement;
and acquiring the perception requirement based on the emotion state comparison rule expression.
As a further scheme of the present invention, the step of automatically adjusting the sound effect parameter by the sound system specifically includes:
collecting a data set containing emotion states, perception requirements and corresponding sound effect parameter adjustment through expert labeling, user feedback and experiment collection methods;
extracting acoustic features, emotion classification features and sound effect parameter features from the data set;
training a mapping model through the acoustic features, the emotion classification features and the sound effect parameter features of the data set by using a machine learning algorithm, and mapping the emotion states and the perception requirements to a sound effect parameter adjustment target;
and importing the input item into a mapping model, converting the emotion state and the perception requirement into corresponding sound effect parameter adjustment targets, and automatically adjusting the sound effect parameters to meet the emotion state and the perception requirement of the user, wherein the sound effect parameters comprise volume, tone and equalizer settings.
As a further aspect of the present invention, the step of establishing a sound scene by the sound system according to the positioning and simulation of the audio source by means of the spatial sound modeling technology specifically includes:
acquiring sound signals from different directions by using a microphone array by adopting a cross correlation method and a beam forming technology, wherein the audio source positioning information;
based on the audio source positioning information, combining a sound signal of an audio source with the environmental acoustic features to establish a sound scene model, wherein the sound scene model comprises the distance, direction, room characteristics, reflection and absorption information of a sound source;
based on the sound scene model, simulating sound effects including echo, reverberation and attenuation in the environment, and generating an environment sound effect simulation result;
based on the sound scene model and the environmental sound effect simulation result, performing spatial sound processing on the audio signal to generate a spatial sound processing result, wherein the spatial sound processing comprises wave field synthesis, a stereo processing algorithm and self-adaptive filtering;
and automatically adjusting sound effect parameters based on the spatial sound processing result and the sound scene model.
As a further scheme of the present invention, the steps of automatically adjusting the sound parameters of the sound system based on the audio content and the intelligent sound adaptability required by the user are specifically as follows:
collecting preference data including user feedback, historical data or user configured preference settings using a questionnaire, user evaluation, observation experiment mode;
extracting preference features including extracting emotion and preference scores from the preference data based on feature extraction;
based on the preference characteristics, dividing users into groups with similar sound effect preferences by using a clustering algorithm, and analyzing personalized sound effect preferences by taking each group as the user cluster to generate a clustering result;
generating association rules based on the clustering result and personalized sound effect preference of the user cluster, wherein the sound system provides real-time sound effect, and automatically adjusts sound effect parameters to adapt to different audio contents and user preference;
and displaying the real-time sound effect to the user, collecting real-time sensory experience and satisfaction evaluation of the user, and fine-tuning the real-time sound effect.
As a further aspect of the present invention, the step of the user interacting with the sound system to adjust sound parameters or select different sound modes specifically includes:
providing a user interface comprising a touch screen, a remote controller and voice control;
based on the user interface, acquiring the requirements of sound effect parameter adjustment or sound effect mode selection of a user as adjustment item data;
and adjusting the sound effect parameters and the sound effect modes based on the adjustment item data.
An acoustic effect adjusting device is composed of an environment sensing module, an automatic adjusting module, a space sound modeling module, an intelligent acoustic effect adaptability module and a user interaction module, wherein the acoustic effect adjusting device is used for executing the acoustic effect adjusting method of claims 1-6.
As a further scheme of the invention, the environment sensing module provides environment acoustic characteristics, and transmits the environment acoustic characteristics to the automatic adjusting module for automatic adjustment of sound effect parameters so as to adapt to emotion and sensing requirements of users;
the environment perception module provides positioning information of the environment acoustic characteristics and the audio source, transmits the positioning information to the space sound modeling module, and generates a sound scene model to realize stereo effect and environment sound effect;
the intelligent sound effect adaptability module generates personalized sound effect preference by collecting user feedback data and historical records and provides the personalized sound effect preference for the automatic adjustment module to carry out fine adjustment;
the user interaction module is responsible for adjusting sound effect parameters and selecting sound effect modes, and transmits user requirements to the automatic adjustment module to realize personalized sound effect experience.
As a further scheme of the invention, the environment sensing module comprises a sound collecting module, a sound analyzing module and a positioning module;
the automatic adjusting module comprises an emotion analyzing module, an audio content analyzing module, an audio parameter adjusting module and a dynamic range control module;
the spatial sound modeling module comprises a stereo effect module and an environment sound effect module;
the intelligent sound effect adaptability module comprises a user preference collection module, a personalized sound effect generation module and a history analysis module;
the user interaction module comprises an audio parameter adjusting interface, a voice command analyzing module and a data display module.
An electronic device for sound effect adjustment, which is specifically a sound controller, which is responsible for carrying the sound effect adjustment device of claims 7-9, and a storage medium, which is an electronic storage device, specifically a hard disk drive, for storing data and information required for sound effect adjustment.
Compared with the prior art, the invention has the advantages and positive effects that:
according to the invention, through the sensing sound effect technology and the emotion analysis algorithm, the sound system can more accurately identify the emotion state and the perception requirement of the user, so that the sound effect parameters are regulated to provide the audio experience with more emotion resonance. The user can better establish emotional connection with the audio content, and the infectivity of the audio experience is enhanced. The inter-sound modeling technique enables audio to present the sensation of a physical object in auditory space, a user can freely move in an unrestricted environment, and a more diffuse, omnidirectional perception of audio is obtained. This brings the user with a more immersive and immersive audio experience. By intelligent sound adaptation based on the environment and audio content, the sound system can automatically adjust sound parameters to accommodate different audio content and user preferences to provide an optimal audio experience.
Drawings
Fig. 1 is a schematic workflow diagram of a sound effect adjusting method, a sound effect adjusting device, an electronic device and a storage medium according to the present invention;
fig. 2 is a detailed flowchart of step 1 of providing a method, an apparatus, an electronic device and a storage medium for adjusting sound effects according to the present invention;
FIG. 3 is a flowchart of a step 2 refinement of a method, an apparatus, an electronic device, and a storage medium for adjusting sound effects according to the present invention;
FIG. 4 is a flowchart illustrating the fine step 3 of the method, apparatus, electronic device and storage medium for adjusting sound effects according to the present invention;
FIG. 5 is a flowchart illustrating the fine step 4 of the method, apparatus, electronic device and storage medium for adjusting sound effects according to the present invention;
FIG. 6 is a flowchart illustrating a method, an apparatus, an electronic device and a storage medium for adjusting sound effects in step 5;
fig. 7 is a flowchart of an audio effect adjusting method, an audio effect adjusting device, an electronic device and a storage medium.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the description of the present invention, it should be understood that the terms "length," "width," "upper," "lower," "front," "rear," "left," "right," "vertical," "horizontal," "top," "bottom," "inner," "outer," and the like indicate orientations or positional relationships based on the orientation or positional relationships shown in the drawings, merely to facilitate describing the present invention and simplify the description, and do not indicate or imply that the devices or elements referred to must have a specific orientation, be configured and operated in a specific orientation, and therefore should not be construed as limiting the present invention. Furthermore, in the description of the present invention, the meaning of "a plurality" is two or more, unless explicitly defined otherwise.
Example 1
Referring to fig. 1, the present invention provides a technical solution: an acoustic effect adjusting method, comprising the steps of:
the sound system acquires environmental acoustic characteristics through an environmental perception technology, and recognizes the emotion state and perception requirement of a user by using an emotion analysis algorithm;
based on the acoustic characteristics, the emotion states and the perception requirements as input items, the sound system automatically adjusts sound effect parameters;
by means of a space sound modeling technology, a sound system establishes a sound scene model according to the positioning and simulation of an audio source;
the sound system automatically adjusts sound parameters based on the intelligent sound adaptability of the audio content and the user demand;
the user interacts with the sound system to adjust the sound parameters or to select different sound modes.
Firstly, environmental acoustic characteristics and emotion analysis algorithms are obtained through an environmental perception technology, and the emotion state and the perception requirement of a user are identified. Then, the sound system automatically adjusts the sound parameters to provide personalized sound experience according to the acoustic characteristics, the emotional state and the perception requirements. Then, by means of space sound modeling technology, the system builds a sound scene model to simulate vivid and stereo sound effect. In addition, the sound system automatically adjusts sound parameters based on the intelligent sound adaptability of the audio content and the user requirements, so that the sound better meets the user requirements. Finally, the user can adjust the sound effect parameters or select different sound effect modes in an interactive mode. In general, the implementation of the method can provide personalized and intelligent sound effect adjusting functions, enhance the hearing experience of users and improve interaction convenience.
Referring to fig. 2, the steps of the sound system acquiring the environmental acoustic features through the environmental perception technology and identifying the emotion state and the perception requirement of the user by using the emotion analysis algorithm are specifically as follows:
the sound system captures sound data in the environment by using microphone equipment, the audio data comprise music, human voice and noise, audio enhancement processing comprising denoising, echo suppression and gain control is carried out on the audio data, and processed audio data are obtained;
the method comprises the steps of acquiring spectrum characteristics from processed audio data by using a digital signal processing technology, particularly a fast Fourier transform, calculating the spectrum characteristics based on a Mel spectrum coefficient, capturing human ear perception characteristics of audio, and calculating environmental acoustic characteristics;
extracting emotion characteristics related to emotion based on the processed audio data, establishing an emotion classifier based on the emotion characteristics by utilizing a random forest, and performing emotion classification on the extracted emotion characteristics by using the emotion classifier to obtain an emotion state;
collecting feedback data, history records and other contextual information of a user, and arranging emotion feedback, use preference and use scenes;
based on emotion feedback, use preference and use scene, adopting a decision tree algorithm to establish an emotion model, and based on expert domain knowledge and the emotion model, formulating a set of rule expression to associate emotion states with perception demands;
and obtaining the perception requirement based on the emotion state comparison rule expression.
First, the sound system captures sound data in the environment using a microphone device and performs audio enhancement processing such as denoising, echo suppression, gain control, and the like on the sound data to improve sound quality and reduce interference. Then, the spectral features are extracted by digital signal processing technology, and environmental acoustic features are calculated by means of mel spectral coefficients and the like. Meanwhile, extracting features related to emotion from the processed audio data, and establishing an emotion classifier through algorithms such as random forests to identify the emotion state of the user. And collecting feedback data, historical records and other context information of the user, establishing an emotion model based on a decision tree algorithm, formulating a rule expression, and associating the emotion state with the perception requirement. Finally, through comparing rule expression, the perception requirement matched with the emotion state can be obtained. Benefits of such sound system implementations include personalized sound experience, automated operation to reduce user burden, immersive audio experience enhancement, and improved user satisfaction. Through environment perception and emotion recognition, the sound system can automatically adjust sound effect parameters according to the requirements and emotion states of the user, and personalized and intelligent sound effect experience is provided, so that the hearing satisfaction and immersion of the user are enhanced.
Referring to fig. 3, the steps of the sound system for automatically adjusting the sound parameters are specifically:
collecting a data set containing emotion states, perception requirements and corresponding sound effect parameter adjustment through expert labeling, user feedback and experiment collection methods;
extracting acoustic features, emotion classification features and sound effect parameter features from the data set;
training a mapping model by using a machine learning algorithm through acoustic features, emotion classification features and sound effect parameter features of the data set, and mapping emotion states and perception requirements to a sound effect parameter adjustment target;
and importing the input item into a mapping model, converting the emotion state and the perception requirement into corresponding sound effect parameter adjustment targets, and automatically adjusting the sound effect parameters to meet the emotion state and the perception requirement of the user, wherein the sound effect parameters comprise volume, tone and equalizer settings.
And collecting a data set containing emotion states, perception requirements and corresponding sound effect parameter adjustment through methods such as expert annotation, user feedback, experiment collection and the like. The data may include the user's sound preferences and needs under different emotional states, as well as corresponding sound parameter adjustment values. Acoustic features, emotion classification features, and sound effect parameter features are extracted from the dataset. The acoustic features may include spectral features, time domain features, etc. of the audio. The emotion classification feature may be an emotion tag or feature associated with an emotion state. The sound effect parameter characteristics comprise the current value or the adjusting range of the sound effect parameters and other information. A mapping model is trained using acoustic features, emotion classification features, and sound effect parameter features in the dataset using a machine learning algorithm, such as a regression model or neural network. The goal of this model is to map the input emotional state and perceived need to the corresponding sound effect parameter adjustment targets. And importing the emotion states and the perception requirements of the user as input items into a trained mapping model. The model converts the emotional state and the perception requirement into corresponding sound effect parameter adjustment targets. For example, the model may translate into corresponding volume, tone, and equalizer settings depending on the user's relaxed emotion and need for immersion. And according to the sound effect parameter adjustment target output by the model, the sound system automatically adjusts the sound effect parameter to meet the emotion state and perception requirements of the user. The system can automatically adjust sound effect parameters such as volume, tone, equalizer settings and the like to provide sound effects that meet the user's expectations.
Referring to fig. 4, with the help of spatial sound modeling technology, the sound system establishes a sound scene according to the positioning and simulation of the audio source, specifically:
the method comprises the steps of obtaining sound signals from different directions and audio source positioning information by a microphone array by adopting a cross correlation method and a beam forming technology;
based on the audio source positioning information, combining the sound signals of the audio source with the environmental acoustic characteristics to establish a sound scene model, wherein the sound scene model comprises the distance, direction, room characteristics, reflection and absorption information of a sound source;
based on the sound scene model, simulating sound effects including echo, reverberation and attenuation in the environment, and generating an environment sound effect simulation result;
based on the sound scene model and the environmental sound effect simulation result, performing spatial sound processing on the audio signal to generate a spatial sound processing result, wherein the spatial sound processing comprises wave field synthesis, a stereo processing algorithm and self-adaptive filtering;
and automatically adjusting sound effect parameters based on the spatial sound processing result and the sound scene model.
Through spatial sound modeling techniques, sound systems are able to build realistic sound scenes from localization and simulation of audio sources. The specific steps include that the microphone array collects sound signals from different directions, performs audio source positioning, then combines the sound signals of the audio sources with environmental acoustic features, and builds a sound scene model, wherein the sound scene model comprises distance, direction, room characteristics and reflection and absorption information of sound sources. Next, according to the sound scene model, sound effects including echo, reverberation, attenuation, and the like are simulated, and an environmental sound effect simulation result is generated. The audio signal is processed by spatial sound processing techniques, such as wave field synthesis, stereo processing, adaptive filtering, etc., to further enhance the realism and immersion of sound. Finally, according to the spatial sound processing result and the sound scene model, the sound system automatically adjusts sound effect parameters such as volume, tone quality and equalizer settings to meet the user demands. By the implementation method, the sound system can provide personalized and realistic sound experience, improve the hearing satisfaction of users and create an immersive audio effect for the users.
Referring to fig. 5, the steps of the sound system for automatically adjusting sound parameters based on the audio content and the intelligent sound adaptability required by the user are specifically as follows:
collecting preference data including user feedback, historical data or user configured preference settings using a questionnaire, user evaluation, observation experiment mode;
extracting preference features including extracting emotion and preference scores from the preference data based on the feature extraction;
based on preference characteristics, dividing users into groups with similar sound effect preferences by using a clustering algorithm, and analyzing personalized sound effect preferences by taking each group as a user cluster to generate a clustering result;
based on the clustering result and the personalized sound effect preference of the user cluster, generating association rules, and automatically adjusting sound effect parameters to adapt to different audio contents and user preference, wherein the sound system provides real-time sound effect;
and displaying the real-time sound effect to the user, collecting real-time sensory experience and satisfaction evaluation of the user, and fine-tuning the real-time sound effect.
Preference data including user feedback, historical data, or user configured preference settings is collected by way of questionnaires, user ratings, and observation experiments, etc. The data may include information about the user's preference, score, preference, etc. for different sound parameters. Features are extracted from the preference data, including mainly emotional features and features of preference scores. Emotional characteristics may be extracted by an emotion analysis algorithm, such as emotion classification or emotion expressed characteristics. The preference scoring feature is then a representation of the user's score or degree of preference for different sound effects. Users are divided into groups with similar sound effect preferences based on preference characteristics using a clustering algorithm. Each group is a user cluster for analyzing personalized sound preferences. By clustering the results, the commonality preference of users in different groups can be known. And generating association rules according to the clustering result and the personalized sound effect preference of the user cluster. These rules describe the relationship between different audio content and user preferences. The sound system uses these rules to provide real-time sound effects, automatically adjusting sound parameters to accommodate different audio content and user preferences. And providing real-time sound effects by the sound system according to the generated association rules, and displaying the real-time sound effects to a user. At the same time, the user's real-time sensory experience and satisfaction ratings are collected. And fine tuning is performed on the real-time sound effect through user feedback and evaluation so as to further adapt to user requirements and improve user satisfaction.
Referring to fig. 6, the steps of the user interacting with the sound system to adjust the sound parameters or select different sound modes are as follows:
providing a user interface comprising a touch screen, a remote controller and voice control;
based on a user interface, acquiring the requirements of sound effect parameter adjustment or sound effect mode selection of a user as adjustment item data;
and adjusting the sound effect parameters and the sound effect modes based on the adjustment item data.
First, the sound system should provide a variety of user interface options such as touch screens, remote controls, voice controls, etc. to provide a convenient way of operation. Secondly, accurately acquiring requirements of sound effect parameter adjustment or sound effect mode selection of a user through a user interface, and converting the requirements into adjustment item data. Then, according to the adjustment item data, the sound system can flexibly adjust sound effect parameters including volume, bass, treble and sound field effects and the like so as to meet the personalized requirements of users. In addition, the system should also provide different preset sound effects modes for users to select, such as a movie mode, a music mode, a game mode, etc., and the system adjusts the sound effect parameters accordingly to provide a sound effect experience conforming to the mode. By the implementation of these steps, the following beneficial effects can be achieved: the user can enjoy individualized sound effect experience, operation becomes more convenient and flexible, sound effect adaptability is enhanced, the user can select proper sound effect modes according to different scenes and requirements, and flexible adjustment of sound effect parameters can provide accurate sound effect, so that satisfaction of the user on a sound system is enhanced. In summary, by providing a user interface, obtaining user requirements, adjusting sound parameters according to the requirements, and selecting sound modes, a personalized, convenient and adaptable sound experience can be achieved.
Referring to fig. 7, an audio effect adjusting apparatus is composed of an environment sensing module, an automatic adjusting module, a spatial sound modeling module, an intelligent audio effect adaptability module, and a user interaction module, and is used for executing the audio effect adjusting method of claims 1-6.
The environment sensing module provides environment acoustic characteristics, and transmits the environment acoustic characteristics to the automatic adjusting module for automatic adjustment of sound effect parameters so as to adapt to emotion and sensing requirements of users;
the environment sensing module provides positioning information of the environment acoustic characteristics and the audio source, and transmits the positioning information to the space sound modeling module to generate a sound scene model so as to realize stereo effect and environment sound effect;
the automatic adjusting module adjusts sound effect parameters according to the audio content and the user requirements, and the intelligent sound effect adaptability module generates personalized sound effect preference by collecting user feedback data and historical records and provides the personalized sound effect preference for the automatic adjusting module to carry out fine adjustment;
the user interaction module is responsible for adjusting sound effect parameters and selecting sound effect modes, and transmitting user requirements to the automatic adjustment module to realize personalized sound effect experience.
The sound effect adjusting device consists of an environment sensing module, an automatic adjusting module, a space sound modeling module, an intelligent sound effect adaptability module and a user interaction module and is used for executing a sound effect adjusting method. The environment sensing module provides the environment acoustic characteristics and the audio source positioning information, and transmits the information to the automatic adjusting module and the space sound modeling module. And the automatic adjusting module adjusts the sound effect parameters according to the audio content and the user requirements so as to adapt to the emotion and perception requirements of the user. The intelligent sound effect adaptability module generates personalized sound effect preference by collecting user feedback and history records, and provides the personalized sound effect preference for the automatic adjustment module to carry out fine adjustment. The user interaction module is responsible for interaction between a user and the device and realizes sound effect parameter adjustment and sound effect mode selection. The device can automatically adjust sound effect parameters, generate stereo sound effect and environment sound effect, and provide personalized sound effect experience. In a combined view, the integration and the function of the sound effect adjusting device bring high-quality sound experience, meet the personalized requirements of users and enhance the satisfaction degree of the users to the device.
Referring to fig. 7, the environment sensing module includes a sound collecting module, a sound analyzing module, and a positioning module;
the automatic adjusting module comprises an emotion analyzing module, an audio content analyzing module, an audio parameter adjusting module and a dynamic range control module;
the space sound modeling module comprises a stereo effect module and an environment sound effect module;
the intelligent sound effect adaptability module comprises a user preference collection module, a personalized sound effect generation module and a history analysis module;
the user interaction module comprises an audio parameter adjusting interface, a voice command analyzing module and a data display module.
The sound effect adjusting device can achieve a plurality of beneficial effects by integrating an environment sensing module (sound collection, sound analysis and positioning), an automatic adjusting module (emotion analysis, audio content analysis, sound effect parameter adjustment and dynamic range control), a spatial sound modeling module (stereo effect and environment sound effect), an intelligent sound effect adaptability module (user preference collection, personalized sound effect generation and history analysis) and a user interaction module (sound effect parameter adjusting interface, voice command analysis and data display). Firstly, the environment sensing module provides accurate environment acoustic characteristics and audio source positioning information, and basic data is provided for subsequent sound effect adjustment. And secondly, the automatic adjusting module carries out intelligent adjustment and dynamic range control on sound effect parameters according to emotion requirements and audio content, so that the sound effect better accords with user expectations. Meanwhile, the space sound modeling module provides a more realistic and immersive sound experience by simulating stereo effects and environmental sound effects. The intelligent sound effect adaptability module generates personalized sound effect preference according to user feedback and history records, and provides personalized adjustment basis for the automatic adjustment module. Finally, the user interaction module enables the user to conveniently adjust sound effect parameters and select sound effect modes, and personalized sound experience is achieved. In summary, the integrated sound effect adjusting device provides accurate environment perception, intelligent sound effect adjustment, enhanced stereo effect and environment sound effect, personalized sound effect adaptability and convenient user interaction experience, and meets the requirements of users on high quality and customized sound effect.
An electronic device for sound effect adjustment, in particular a sound controller, which is responsible for carrying the sound effect adjustment device of claims 7-9, and a storage medium, in particular a hard disk drive, for storing data and information required for sound effect adjustment.
Working principle: and the environment sensing module collects sound signals in the environment through the sound collecting module, and then processes and analyzes the sound through the sound analyzing module to extract the acoustic characteristics of the environment. Meanwhile, the positioning module is used for determining the position information of the audio source. And an emotion analysis module in the automatic adjustment module analyzes the emotion state and the perception requirement of the user by using an algorithm, and an audio content analysis module analyzes the audio content. According to the emotion state, the perception requirement and the audio content, the sound effect parameter adjusting module intelligently adjusts the sound effect parameters so that the sound effect accords with the expectations of the user. The dynamic range control module is used for balancing the dynamic range of the audio and ensuring the definition and fidelity of the audio. The stereo effect module in the space sound modeling module simulates the stereo effect by utilizing corresponding algorithm and technology, so that the sound has sense of direction and sense of positioning in the sense of hearing. The environmental sound effect module can simulate different environmental sounds, such as a concert, a natural landscape, and the like, so that the immersion and the realism of the sound are enhanced. The intelligent sound effect adaptability module is used for collecting preference and preference information of different sound effects of a user, the personalized sound effect generation module generates personalized sound effects according to the collected user preference information, and the history analysis module analyzes by using user history data to provide personalized sound effect recommendation of the user. The sound effect parameter adjusting interface in the user interaction module provides an intuitive interface for a user, so that the user can conveniently adjust sound effect parameters such as volume, equalizer and the like. The voice command parsing module allows a user to adjust the sound effects or select a sound effect mode through voice instructions. The data display module is used for displaying the information of the current sound effect parameters and the sound effect modes so that a user can know the current setting and adjusting states.
The present invention is not limited to the above embodiments, and any equivalent embodiments which can be changed or modified by the technical disclosure described above can be applied to other fields, but any simple modification, equivalent changes and modification made to the above embodiments according to the technical matter of the present invention will still fall within the scope of the technical disclosure.

Claims (10)

1. A sound effect adjusting method, comprising the steps of:
the sound system acquires environmental acoustic characteristics through an environmental perception technology, and recognizes the emotion state and perception requirement of a user by using an emotion analysis algorithm;
based on the acoustic characteristics, the emotion states and the perception requirements as input items, the sound system automatically adjusts sound effect parameters;
the sound system establishes a sound scene model according to the positioning and simulation of the audio source by means of a space sound modeling technology;
the sound system automatically adjusts sound parameters based on the intelligent sound adaptability of the audio content and the user demand;
the user interacts with the sound system to adjust sound parameters or to select different sound modes.
2. The sound effect adjusting method according to claim 1, wherein the sound system obtains the environmental acoustic characteristics through the environmental perception technology, and the steps of identifying the emotion state and the perception requirement of the user by using the emotion analysis algorithm specifically include:
the sound system captures sound data in the environment by using microphone equipment, the audio data comprise music, human voice and noise, and the audio data are subjected to audio enhancement processing comprising denoising, echo suppression and gain control to obtain processed audio data;
acquiring spectral features from the processed audio data by using a digital signal processing technology, particularly a fast Fourier transform, calculating the spectral features based on Mel spectral coefficients, capturing human ear perception characteristics of the audio, and calculating the environmental acoustic features;
extracting emotion characteristics related to emotion based on the processed audio data, establishing an emotion classifier based on the emotion characteristics by utilizing a random forest, and performing emotion classification on the extracted emotion characteristics by using the emotion classifier to obtain the emotion state;
collecting feedback data, history records and other contextual information of a user, and arranging emotion feedback, use preference and use scenes;
based on the emotion feedback, the use preference and the use scene, adopting a decision tree algorithm to establish an emotion model, and based on expert domain knowledge and the emotion model, formulating a set of rule expression to correlate the emotion state with the perception requirement;
and acquiring the perception requirement based on the emotion state comparison rule expression.
3. The sound effect adjusting method according to claim 1, wherein the step of automatically adjusting the sound effect parameter by the sound system specifically comprises:
collecting a data set containing emotion states, perception requirements and corresponding sound effect parameter adjustment through expert labeling, user feedback and experiment collection methods;
extracting acoustic features, emotion classification features and sound effect parameter features from the data set;
training a mapping model through the acoustic features, the emotion classification features and the sound effect parameter features of the data set by using a machine learning algorithm, and mapping the emotion states and the perception requirements to a sound effect parameter adjustment target;
and importing the input item into a mapping model, converting the emotion state and the perception requirement into corresponding sound effect parameter adjustment targets, and automatically adjusting the sound effect parameters to meet the emotion state and the perception requirement of the user, wherein the sound effect parameters comprise volume, tone and equalizer settings.
4. The sound effect adjusting method according to claim 1, wherein the sound system establishes a sound scene based on the positioning and simulation of the audio source by means of the spatial sound modeling technique specifically comprises:
acquiring sound signals from different directions by using a microphone array by adopting a cross correlation method and a beam forming technology, wherein the audio source positioning information;
based on the audio source positioning information, combining a sound signal of an audio source with the environmental acoustic features to establish a sound scene model, wherein the sound scene model comprises the distance, direction, room characteristics, reflection and absorption information of a sound source;
based on the sound scene model, simulating sound effects including echo, reverberation and attenuation in the environment, and generating an environment sound effect simulation result;
based on the sound scene model and the environmental sound effect simulation result, performing spatial sound processing on the audio signal to generate a spatial sound processing result, wherein the spatial sound processing comprises wave field synthesis, a stereo processing algorithm and self-adaptive filtering;
and automatically adjusting sound effect parameters based on the spatial sound processing result and the sound scene model.
5. The sound effect adjusting method according to claim 1, wherein the sound system automatically adjusts sound effect parameters based on the intelligent sound effect adaptability of the audio content and the user's demand, specifically comprising the steps of:
collecting preference data including user feedback, historical data or user configured preference settings using a questionnaire, user evaluation, observation experiment mode;
extracting preference features including extracting emotion and preference scores from the preference data based on feature extraction;
based on the preference characteristics, dividing users into groups with similar sound effect preferences by using a clustering algorithm, and analyzing personalized sound effect preferences by taking each group as the user cluster to generate a clustering result;
generating association rules based on the clustering result and personalized sound effect preference of the user cluster, wherein the sound system provides real-time sound effect, and automatically adjusts sound effect parameters to adapt to different audio contents and user preference;
and displaying the real-time sound effect to the user, collecting real-time sensory experience and satisfaction evaluation of the user, and fine-tuning the real-time sound effect.
6. The sound effect adjustment method according to claim 1, wherein the step of the user interacting with the sound system to adjust sound effect parameters or to select different sound effect modes is specifically:
providing a user interface comprising a touch screen, a remote controller and voice control;
based on the user interface, acquiring the requirements of sound effect parameter adjustment or sound effect mode selection of a user as adjustment item data;
and adjusting the sound effect parameters and the sound effect modes based on the adjustment item data.
7. An acoustic effect adjusting device, characterized in that the acoustic effect adjusting device is composed of an environment sensing module, an automatic adjusting module, a space sound modeling module, an intelligent acoustic effect adaptability module and a user interaction module, and the acoustic effect adjusting device is used for executing the acoustic effect adjusting method of claims 1-6.
8. The sound effect adjusting device according to claim 7, wherein the environment sensing module provides environment acoustic characteristics, and transmits the environment acoustic characteristics to the automatic adjusting module for automatic adjustment of sound effect parameters so as to adapt to emotion and perception requirements of a user;
the environment perception module provides positioning information of the environment acoustic characteristics and the audio source, transmits the positioning information to the space sound modeling module, and generates a sound scene model to realize stereo effect and environment sound effect;
the intelligent sound effect adaptability module generates personalized sound effect preference by collecting user feedback data and historical records and provides the personalized sound effect preference for the automatic adjustment module to carry out fine adjustment;
the user interaction module is responsible for adjusting sound effect parameters and selecting sound effect modes, and transmits user requirements to the automatic adjustment module to realize personalized sound effect experience.
9. The sound effect adjustment device of claim 7, wherein the environment sensing module comprises a sound collection module, a sound analysis module, a positioning module;
the automatic adjusting module comprises an emotion analyzing module, an audio content analyzing module, an audio parameter adjusting module and a dynamic range control module;
the spatial sound modeling module comprises a stereo effect module and an environment sound effect module;
the intelligent sound effect adaptability module comprises a user preference collection module, a personalized sound effect generation module and a history analysis module;
the user interaction module comprises an audio parameter adjusting interface, a voice command analyzing module and a data display module.
10. An electronic device for adjusting sound effects and a storage medium, wherein the electronic device for adjusting sound effects is specifically a sound controller, the sound controller is responsible for carrying the device for adjusting sound effects according to claims 7-9, and the storage medium is an electronic storage device, specifically a hard disk drive, and is used for storing data and information required for adjusting sound effects.
CN202311482036.3A 2023-11-09 2023-11-09 Sound effect adjusting method and device, electronic equipment and storage medium Pending CN117479076A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311482036.3A CN117479076A (en) 2023-11-09 2023-11-09 Sound effect adjusting method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311482036.3A CN117479076A (en) 2023-11-09 2023-11-09 Sound effect adjusting method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN117479076A true CN117479076A (en) 2024-01-30

Family

ID=89625271

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311482036.3A Pending CN117479076A (en) 2023-11-09 2023-11-09 Sound effect adjusting method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN117479076A (en)

Similar Documents

Publication Publication Date Title
US11468750B2 (en) Authoring an immersive haptic data file using an authoring tool
US10645518B2 (en) Distributed audio capture and mixing
US20200186912A1 (en) Audio headset device
CN109076305A (en) The rendering of augmented reality earphone environment
US11915687B1 (en) Systems and methods for generating labeled data to facilitate configuration of network microphone devices
JP6485711B2 (en) Sound field reproduction apparatus and method, and program
US20220101623A1 (en) Room Acoustics Simulation Using Deep Learning Image Analysis
CN108269578A (en) For handling the method and apparatus of information
JP7453712B2 (en) Audio reproduction method, device, computer readable storage medium and electronic equipment
KR20220044204A (en) Acoustic Echo Cancellation Control for Distributed Audio Devices
GB2582991A (en) Audio generation system and method
CN110580914A (en) Audio processing method and equipment and device with storage function
US20240098416A1 (en) Audio enhancements based on video detection
WO2022170716A1 (en) Audio processing method and apparatus, and device, medium and program product
KR102114335B1 (en) Audio amplifier with sound tuning system using artificial intelligence model
KR102098886B1 (en) Control method, apparatus and system for artificial intelligence sound tuning system including multichannel audio mixer
CN117479076A (en) Sound effect adjusting method and device, electronic equipment and storage medium
WO2023211385A1 (en) Soundscape augmentation system and method of forming the same
KR102065030B1 (en) Control method, apparatus and program of audio tuning system using artificial intelligence model
Lorho Perceptual evaluation of mobile multimedia loudspeakers
US9055362B2 (en) Methods, apparatus and systems for individualizing audio, music and speech adaptively, intelligently and interactively
US20240233711A1 (en) Systems and Methods for Generating Labeled Data to Facilitate Configuration of Network Microphone Devices
WO2024079910A1 (en) Content reproduction device, vibration control signal generation device, server device, vibration control signal generation method, content reproduction system, and design assistance device
US20230267942A1 (en) Audio-visual hearing aid
WO2023076823A1 (en) Mapping of environmental audio response on mixed reality device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination