CN112581924A - Audio processing method and device based on point-to-sing equipment, storage medium and equipment - Google Patents

Audio processing method and device based on point-to-sing equipment, storage medium and equipment Download PDF

Info

Publication number
CN112581924A
CN112581924A CN201910943095.3A CN201910943095A CN112581924A CN 112581924 A CN112581924 A CN 112581924A CN 201910943095 A CN201910943095 A CN 201910943095A CN 112581924 A CN112581924 A CN 112581924A
Authority
CN
China
Prior art keywords
sound
audio
original
processing
file
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910943095.3A
Other languages
Chinese (zh)
Inventor
张辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Aimyunion Network Technology Co ltd
Original Assignee
Guangzhou Aimyunion Network Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Aimyunion Network Technology Co ltd filed Critical Guangzhou Aimyunion Network Technology Co ltd
Priority to CN201910943095.3A priority Critical patent/CN112581924A/en
Publication of CN112581924A publication Critical patent/CN112581924A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/02Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/003Changing voice quality, e.g. pitch or formants
    • G10L21/007Changing voice quality, e.g. pitch or formants characterised by the process used

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Reverberation, Karaoke And Other Acoustics (AREA)

Abstract

The present application relates to the field of audio processing technologies, and in particular, to an audio processing method and apparatus based on a karaoke device, a storage medium, and a device, where the audio processing method based on the karaoke device includes: acquiring a plurality of original voice frequencies picked up by a point-singing device; the original sound frequency is the recording sound frequency of a plurality of different sound pickup devices of the requesting equipment in the same scene and the same space; respectively carrying out digital processing on each original sound audio frequency and calling an effector algorithm to generate corresponding effect sound, and respectively carrying out synthesis processing on each original sound audio frequency and the corresponding effect sound to generate mutually independent recording files; and mixing at least two sound recording files to form a combined audio file. The scheme provided by the application can improve the fidelity rate of the sound recording file corresponding to the single audio and the fidelity rate of the audio file combined.

Description

Audio processing method and device based on point-to-sing equipment, storage medium and equipment
Technical Field
The present application relates to the field of audio processing technologies, and in particular, to an audio processing method and apparatus, a storage medium, and a device based on a point-to-sing device.
Background
Along with the continuous development of public entertainment demands, the mode of communication by using music is more and more popular. At present, a singing ordering device is widely used, a user can conveniently and individually perform a singing ordering experience, the singing ordering device is a singing device for the user to sing according to the selection, when multiple users perform chorus or PK on the singing ordering device in the same scene or in the same space, when the subsequent chorus or PK performance of singing on the singing songs is processed, the chorus or PK sound mixing information is often processed through a processor, and the scoring processing cannot be performed on the recordings of the single users.
Under a general condition, when the recorded mixed audio needs to be processed, a scheme of splitting the mixed audio is generally adopted, but the fidelity of audio information is reduced in the process of splitting the mixed audio, so that the accuracy of subsequently scoring the audio file is influenced.
Disclosure of Invention
The application provides an audio processing method and device based on a point-to-sing device, a computer readable storage medium and a computer device, so as to improve the fidelity rate of a recording file corresponding to a single audio and the fidelity rate of a combined audio file.
The embodiment of the application firstly provides an audio processing method based on a point-to-sing device, which comprises the following steps:
acquiring a plurality of original voice frequencies picked up by a point-singing device; the original sound frequency is the recording sound frequency of a plurality of different sound pickup devices of the requesting equipment in the same scene and the same space;
respectively carrying out digital processing on each original sound audio frequency and calling an effector algorithm to generate corresponding effect sound, and respectively carrying out synthesis processing on each original sound audio frequency and the corresponding effect sound to generate mutually independent recording files;
and mixing at least two sound recording files to form a combined audio file.
In one embodiment, the step of respectively performing digital processing on each original sound audio and invoking an effector algorithm to generate corresponding effect sounds, and respectively performing synthesis processing on each original sound audio and the corresponding effect sounds to generate mutually independent recording files includes:
sequentially carrying out signal amplification processing and analog-to-digital conversion processing on the original audio to obtain corresponding digital signals;
inputting the digital signal into a mainboard through a sound card chip;
calling effect sounds generated by an effector algorithm in the main board, and synthesizing the effect sounds and the digital signals into corresponding recording files according to a preset rule.
In one embodiment, after the signal amplification processing and the analog-to-digital conversion processing are sequentially performed on the original audio to obtain the corresponding digital signal, the method further includes:
inputting the digital signal into a digital signal processing circuit, and dynamically processing the digital signal;
superposing and mixing the dynamically processed digital signal and an environment sound signal output by a main board to obtain sound mixing information;
and D/A conversion processing, pre-stage amplification processing and player amplification processing are carried out on the sound mixing information, and the processed sound mixing information is respectively pushed to corresponding players for feedback monitoring.
In one embodiment, the step of invoking an effect sound generated by an effector algorithm comprises:
and determining the effect sound corresponding to each original sound audio according to the recording requirement of the recording file.
In one embodiment, the audio processing method based on the point-to-sing device further includes:
receiving a request sent by a user terminal for acquiring the sound recording file or the combined audio file corresponding to any user;
and responding to the request to call the corresponding sound recording file or combine the audio file and send the sound recording file or the combined audio file to the user terminal.
In one embodiment, the step of mixing at least two of the audio records into a merged audio file comprises:
acquiring time marks on all the sound recording files;
and according to the time marks, synchronously mixing the sound of the sound recording files to generate a combined audio file.
Further, the present application also provides a singing scoring system, comprising:
receiving a singing song and a singing cooperation request selected by a first user, and sending the singing song and the singing cooperation request to related users;
receiving a response of a second user, starting a singing link, picking up original voice frequencies corresponding to the first user and the second user by using different sound pickups of the ordering device, and processing the original voice frequencies by using the audio processing method based on the ordering device in any one of the technical schemes to obtain a recording file and a combined audio file corresponding to the original voice frequencies;
and scoring the sound recording file and/or the combined audio file.
Correspondingly, the embodiment of the present application further provides an audio processing apparatus based on a jukebox device, including:
the original sound audio acquisition module is used for acquiring a plurality of original sound audios picked up by the singing equipment; the original sound frequency is the recording sound frequency of a plurality of different sound pickup devices of the requesting equipment in the same scene and the same space;
a record file generation module, configured to perform digital processing on each original audio frequency and call an effector algorithm to generate corresponding effect sounds, and perform synthesis processing on each original audio frequency and the corresponding effect sounds to generate independent record files;
and the merged audio file obtaining module is used for mixing at least two sound recording files into a merged audio file.
Further, an embodiment of the present application also provides a computer-readable storage medium, which is used for storing computer instructions, and when the computer-readable storage medium is run on a computer, the computer is enabled to execute the steps of the audio processing method based on a point-to-sing device according to any one of the above technical solutions.
Still further, an embodiment of the present application further provides a computer device, where the computer device includes:
one or more processors;
a storage device for storing one or more programs,
when the one or more programs are executed by the one or more processors, the one or more processors implement the steps of the audio processing method based on a karaoke apparatus according to any one of the above technical solutions.
Compared with the prior art, the scheme provided by the application at least has the following advantages:
according to the audio processing method based on the requesting equipment, a plurality of original sound audios are picked up through a plurality of different sound pickups of the requesting equipment, digital processing and effect sound synthesis processing are sequentially carried out on the original sound audios, and a plurality of independent recording files are obtained. According to the scheme provided by the application, one sound recording file corresponds to a single audio file of a user, effect sounds are mixed in the sound recording file, and the sound recording file is a complete file with sound effects; moreover, the method and the device can directly process the original audio frequency corresponding to a single user, can avoid the damage of the audio mixing splitting process to the fidelity ratio of the single audio frequency information, and are favorable for obtaining lossless recording files.
The audio recording file and the merged audio file provided based on the application are beneficial to improving the accuracy of scoring the files.
Drawings
Fig. 1 is an environment diagram of an implementation of an audio processing method based on a karaoke apparatus according to an embodiment of the present application;
fig. 2 is a flowchart of an audio processing method based on a karaoke apparatus according to an embodiment of the present application;
FIG. 3 is a flowchart illustrating the generation of audio files that are independent of one another according to an embodiment of the present application;
fig. 4 is a flowchart illustrating obtaining a feedback monitoring signal based on the digital information obtained in S310 according to an embodiment of the present application;
fig. 5 is a block diagram of a processing procedure for obtaining a feedback monitor signal based on two original audio frequencies according to an embodiment of the present application;
fig. 6 is a schematic structural diagram of an audio processing apparatus based on a jukebox device according to an embodiment of the present application;
fig. 7 is a schematic structural diagram of a computer device according to an embodiment of the present application.
Detailed Description
Reference will now be made in detail to embodiments of the present application, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are exemplary only for the purpose of explaining the present application and are not to be construed as limiting the present application.
Fig. 1 is a diagram of an implementation environment of an audio processing method according to an embodiment, where the implementation environment includes a server, a user terminal, and a singing apparatus, where the singing apparatus is configured with at least one sound pickup, and as shown in fig. 1, the singing apparatus is configured with 2 sound pickups, which are respectively a sound pickup 1 and a sound pickup 2. The audio processing method based on the singing requesting equipment can be executed in a server section, can be executed in the singing requesting equipment terminal, and can also be executed in a user terminal which establishes an association relation with the singing requesting equipment and has a data processing function. The following embodiments of the present application mainly use a user terminal as an execution end to describe the scheme, but do not limit the present application.
With reference to fig. 1, the implementation process of the scheme provided by the present application when the user terminal is used is as follows:
the method comprises the steps that a user terminal obtains a plurality of original voice frequencies picked up by a requesting device, wherein the original voice frequencies are recorded voice frequencies of different sound pickups of the requesting device in the same scene and the same space; the user terminal respectively carries out operations such as digitalization processing and synthesis processing with the effect sound on the obtained original sound audio, before the step of synthesizing with the effect sound, the method also comprises the step of calling the effect sound generated by the effector algorithm, wherein the effect sound can be special effect sound, scene sound and the like, then the digitized original sound audio frequency and the corresponding effect sound are synthesized to generate mutually independent recording files, the recording file is the complete recorded audio of the user, in the singing scene, the recording file is the song file with the complete tune, that is, the sound recording file contains not only the original audio of the user, but also the tune of the song, and even additionally added other effect sounds, based on the obtained sound recording file, and carrying out sound mixing processing on the sound recording files to obtain a combined audio file, wherein the combined audio file is a chorus file of at least two users in the singing field.
It should be noted that the user terminal may be a smart phone, a tablet computer, a notebook computer, a desktop computer, and the like, and the server side may be implemented by a computer device having a processing function, but is not limited thereto. The server and the user terminal may be connected to each other through bluetooth, USB (Universal Serial Bus), or other communication connection methods, which is not limited herein.
Fig. 2 is a flowchart of an audio processing method based on a jukebox device according to an embodiment of the present application, including the following steps:
step S210, acquiring a plurality of original voice frequencies picked up by the singing-clicking equipment; the original sound frequency is the recording sound frequency of a plurality of different sound pickup devices of the requesting equipment in the same scene and the same space;
step S220, respectively carrying out digital processing on each original sound audio and calling an effector algorithm to generate corresponding effect sounds, respectively carrying out synthesis processing on each original sound audio and the corresponding effect sounds, and generating mutually independent recording files;
step S230, mixing at least two of the sound recording files to form a merged audio file.
The song is not limited to the form of a song, a drama and the like, the song ordering device can also provide the services of melody, accompaniment, lyric prompt, video recording and the like for the user, subsequently, the audio information recorded by a sound pickup matched with the song ordering device is synthesized with the background music of the song, and subsequently, the synthesized file can be stored, played, transmitted and the like.
Step S210, acquiring a plurality of original audio picked up by the karaoke apparatus; the original sound audio is the recorded audio of a plurality of different sound pick-up devices of the singing equipment in the same scene and at the same time.
Here, the apparatus for requesting music may be one or more apparatuses, and the different sound pickup apparatuses may be sound pickup apparatuses on the same apparatus for requesting music, or sound pickup apparatuses on different apparatuses for requesting music, that is, different users may record original sound through one or more apparatuses for requesting music, but original sound of different users is obtained through different sound pickup apparatuses.
The sound pickup records the recording audio in the same scene, the same space or the same time, wherein the same scene can be in the same song scene.
If different users record original sound in the same space of the same scene, sound mixing is inevitably generated, that is, the original sound needs to be processed to obtain pure single-person recording audio, and for the recording audio in the same space of the same scene, the sound mixing of other users in the recording audio can be removed by using any one or more of audio, tone and timbre, so that each sound pickup acquires the recording audio of one user.
If the sound pickup picks up the recording audio in the same scene and at the same time, the same time can be the same time interval or the same time, namely the recording audio is in the recording time interval of one song at the same time, so that the time synchronization of the acquired reason audio is ensured, and the recording audio acquired in the scene is the audio information of an individual user and is not doped with the audio information of other users.
No matter which kind of above-mentioned situation obtains the recording audio frequency, after obtaining a plurality of original sound audio frequencies that equipment of singing point picked up, all need remove dry processing to the original sound audio frequency, can adopt modes such as Fourier transform, sampling to remove dry etc. to carry out the audio frequency and remove dry, through taking a sample to the waveform sample of noise, then the waveform and the sampling noise sample analysis of original sound audio frequency, the noise automatic removal obtains the single user corresponding and removes dry audio frequency. In step S220, the original audio frequencies are respectively digitized and an effector algorithm is called to generate corresponding effect sounds, and the original audio frequencies and the corresponding effect sounds are respectively synthesized to generate independent recording files.
In this embodiment, the user terminal sequentially performs operations such as digitizing the acquired original audio and synthesizing the acquired original audio with the effect audio, and before the step of synthesizing the acquired original audio with the effect audio, the method further includes: calling an effect sound generated by an effector algorithm, wherein the effect sound can be at least one of special effect sound and scene sound, and then synthesizing the digitized original sound audio and the corresponding effect sound to generate mutually independent recording files.
The user terminal obtains the independent recording files of each user, performs the operation of combining the audio files in step S230 for the independent recording files, and combines at least two recording files obtained in the above manner to obtain combined audio files.
The merged audio file is obtained by mixing the sound of the sound files of at least two users to form a sound-mixed audio file, and in a chorus scene, the mixed audio file is chorus audio, and the merged audio file is obtained based on the individual sound files of the users, so that the merged audio file with high fidelity ratio and high clarity and integrity can be obtained.
When the audio processing scheme based on the jukebox is executed at the user terminal, the user terminal may further upload the independent recording files and the combined audio files corresponding to the users to the server for storage and subsequent processing.
When the audio processing scheme based on the point-singing equipment is executed at the server end, the server stores the obtained recording file and the combined audio file so as to carry out subsequent processing.
The audio processing method based on the requesting device picks up a plurality of original sound audios through a plurality of different sound pickups of the requesting device, and sequentially carries out digital processing and effect sound synthesis processing on the original sound audios to obtain a plurality of independent recording files, wherein one recording file corresponds to an independent audio file of a user, the recording files are mixed with effect sounds, and the recording files are complete files with sound effects; moreover, the method and the device can directly process the original audio frequency corresponding to a single user, can avoid the damage of the audio mixing splitting process to the fidelity ratio of the single audio frequency information, and are favorable for obtaining lossless recording files.
The audio mixing is carried out on the basis of the obtained recording file to form a combined audio file, on the basis of the recording file with high fidelity rate, the obtained combined audio file is also an audio file with high fidelity rate and complete sound effect, the recording file and the combined audio file provided by the application are favorable for improving the accuracy of scoring the files, and the user experience is improved.
In order to make clearer the audio processing scheme based on the point-singing device and the technical effects thereof provided by the present application, specific schemes thereof are explained in detail with a plurality of embodiments.
In one embodiment, the step of performing digital processing on each original audio and invoking an effector algorithm to generate corresponding effect sounds in step S220, and performing synthesis processing on each original audio and the corresponding effect sounds to generate mutually independent recording files may be implemented in the following manner, where a flowchart is shown in fig. 3, and includes the following sub-steps:
s310, sequentially carrying out signal amplification processing and analog-to-digital conversion processing on the original audio to obtain a corresponding digital signal;
s320, inputting the digital signal into a mainboard through a sound card chip;
s330, calling effect sounds generated by an effector algorithm in the mainboard, and synthesizing the effect sounds and the digital signals into corresponding recording files according to a preset rule.
Specifically, the original sound audio picked up by the sound pick-up is amplified, the amplified original sound audio signal is input into an ADC analog-to-digital circuit, the analog signal is converted into a digital signal, and the reason audio is digitized, so that the audio signal can be processed by a subsequent mainboard.
And then inputting the converted digital signal into the main board through the USB sound card chip, calling an effector algorithm in the main board to generate corresponding effect sound, and mixing the effect sound with the digital signal obtained in the step S310 to obtain a corresponding recording file.
According to the embodiment of the application, the original sound audio is subjected to amplification, analog-to-digital processing and audio mixing processing at a host end, the recording file corresponding to the original sound audio is obtained, the obtaining scheme is simple and easy to operate, too many operations which damage the tone quality of the original sound audio are not executed in the process, and the fidelity ratio of the recording file is guaranteed to the greatest extent.
In one embodiment, the step of invoking the effect sound generated by the effector algorithm of S330 includes: and determining the effect sound corresponding to each original sound audio according to the recording requirement of the recording file.
The effect sounds may include: for example, sound effects such as electric sound, praise and call in the song singing field are selected in advance according to the requirement information of the recording file, the parameter configuration of the effector algorithm can be adjusted to obtain effect sound meeting the requirement information, and different original sound audio frequencies can correspond to the same or different effect sound. Different effect sounds are correspondingly set for different original sound frequencies, so that an individualized recording file can be obtained, and the interestingness of the recording file is improved.
In addition to obtaining a corresponding recording file based on the original audio, the present application can also obtain a feedback monitoring signal based on the digital information obtained in S310, and a flowchart of the process is shown in fig. 4, specifically as follows:
after obtaining the corresponding digital information in step S310, the method further includes the following sub-steps:
s410, inputting the digital signal into a digital signal processing circuit, and dynamically processing the digital signal;
s420, superposing and mixing the dynamically processed digital signal and an environment sound signal output by the main board to obtain sound mixing information;
and S430, performing digital-to-analog conversion processing, pre-stage amplification processing and player amplification processing on the sound mixing information, and respectively pushing the processed sound mixing information to corresponding players for feedback monitoring.
Specifically, the original audio is amplified by a signal amplifier and is converted from analog to digital to obtain a digital signal, and the digital signal is input into a digital signal processor for dynamic processing, wherein the dynamic processing comprises: high level compression, overall compression, loudness enhancement limit, volume adjustment, loudness adjustment, equalization processing, effect sound addition and the like, so that the output audio signal meets the output requirement.
After dynamic processing, the audio signals are superposed and mixed with the environment sound signals of the main board to obtain mixed sound signals, the mixed sound signals are input into a digital-analog conversion circuit to obtain analog signals, the analog signals are input into a pre-stage amplifying circuit and a player amplifying circuit to be processed, and all the audio signals are respectively pushed into corresponding players to be fed back and monitored.
The feedback monitoring signal acquired by the embodiment of the application can be used for monitoring the singing effect so as to improve the processing of each link, such as instructing a user to sing in the original voice frequency input process or adjusting parameters in the amplification, conversion and sound mixing processes.
Taking two original audio frequencies as an example, fig. 5 shows a processing block diagram for obtaining feedback monitoring signals based on the two original audio frequencies, where the microphone 1 and the microphone 2 are two pickups, the microphone 1 picks up a signal of the original audio frequency 1, and the microphone 2 picks up a signal of the original audio frequency 2, the two pickups may be from the same karaoke apparatus or different karaoke apparatuses, the two microphones obtain original audio frequencies of two users, the original audio frequencies are amplified by a microphone amplifier, the amplified original audio signals are input to an ADC analog-to-digital circuit, the analog signals are converted into digital signals for data processing, the converted digital signals are input to a main board via a USB sound card chip, mixing of effect sounds and generation of a recording file are performed in the main board, and the recording file is uploaded.
The other signal processing in fig. 5 is as follows: the original sound audio frequency is amplified by respective signal amplifier, the amplified original sound audio frequency is transmitted to another two ADC analog-to-digital circuits, the converted digital signals are respectively sent to a DSP digital signal processor, each path of audio signal is subjected to the adjustment of the audio signal, finally, the audio signal is superposed and mixed with the environmental sound signal from a mainboard through a USB sound card chip, the audio signal is processed through a DAC digital-to-analog conversion circuit, a pre-stage amplification circuit and an earphone amplification circuit, each path of audio signal is respectively pushed to a corresponding earphone for feedback monitoring, namely, the original sound audio signal 1 collected by the microphone 1 is pushed to the earphone 1 through the processing, the original sound audio signal collected by the microphone 2 is pushed to the earphone 2 through the series of processing, and the obtained feedback monitoring signal can be used for monitoring the singing effect to improve the processing of each link, such as instructing the user to sing in the process of inputting the original sound and audio, or parameters in the processes of amplifying, converting and mixing.
The scheme provided by the embodiment obtains the recording file corresponding to a single user, and based on the recording file, the method and the device for obtaining the audio-mixed combined audio file can also obtain the audio-mixed combined audio file, which are specifically as follows:
the step of mixing at least two of the sound recording files to form a merged audio file in step S230 includes:
a1, acquiring a time mark on a sound recording file to be mixed;
and A2, synchronously mixing the sound of the sound recording files according to the time marks to generate a combined audio file.
In the embodiment, the recording files are aligned according to the time marks on the recording files, so that the recording files with the synchronized time are synchronously mixed, the problem of time overlapping of a plurality of audios in the audio file combination is avoided, and the accuracy of obtaining the combined audio file is further improved.
In the foregoing embodiment, the audio processing method based on the pointing device further includes:
b1, receiving a request sent by a user terminal for acquiring any one of the single sound recording files or the combined audio file;
and B2, responding to the request, calling the corresponding sound recording file or merging the audio file and sending the audio file to the user terminal.
Based on the obtained sound recording file and the merging request file, the user terminal and the server terminal can respond to the request of any file and send the corresponding file to the user terminal, and currently, the user terminal and the server terminal can also respond to other processing requests of any file such as editing of the file, and the like, so that the operation flexibility of the sound recording file and the merging audio file is enhanced.
Further, based on the audio processing method based on the pointing device provided by the above embodiment, the application also provides a singing scoring system, including:
c1, receiving the singing song and the singing cooperation request selected by the first user, and sending the singing song and the singing cooperation request to the related user;
c2, receiving the response of the second user, starting a singing link, picking up the original voice frequencies corresponding to the first user and the second user by using different sound pickups of the ordering device, and processing the original voice frequencies by using the audio processing method based on the ordering device to obtain the recording files and the combined audio files corresponding to the original voice frequencies;
and C3, scoring the sound recording files and/or the merged audio files.
The system defines the user who initiates the cooperation invitation as a first user, defines the user who accepts the cooperation invitation as a second user, and can continue to define a third user and a fourth user … … Nth user if the users who accept the cooperation invitation are multiple.
The singing system provided by the embodiment of the application can be an application on the user terminal. The first user confirms the singing song at the singing ordering device end, and then issues the selected song and the cooperation invitation to the singing system, the singing system issues the song selected by the user and the transmitted PK or the chorus invitation to the current singing system, and if at least one user in the system accepts the invitation, the chorus or the PK of the song can be performed.
The singing scoring system performs audio processing by using the audio processing method of the singing ordering device, obtains the mutually independent recording files and combined audio files corresponding to a plurality of users in the system, scores by using the recording files and the combined audio files, scores based on the recording files and the combined audio files with high fidelity ratio, can obtain accurate scoring results, and improves user experience.
The audio processing method of the ordering device can be applied to the singing devices of mini KTVs and common KTV rooms, and can also be applied to mobile phones loaded with ordering software.
The above embodiments of the audio processing method based on the jukebox device provided by the present application are directed to the above methods, and the following describes embodiments of an audio processing apparatus corresponding to the above methods.
An embodiment of the present application further provides an audio processing apparatus based on a jukebox device, a schematic structural diagram of which is shown in fig. 6, and the audio processing apparatus includes: the module for acquiring original sound audio 610, the module for generating recording file 620, and the module for acquiring merged audio file 630 are as follows:
an original audio obtaining module 610, configured to obtain multiple original audio frequencies picked up by the karaoke apparatus; the original sound frequency is the recording sound frequency of a plurality of different sound pickup devices of the requesting equipment in the same scene and the same space;
a record file generating module 620, configured to perform digital processing on each original audio frequency and call an effector algorithm to generate corresponding effect sounds, and perform synthesis processing on each original audio frequency and the corresponding effect sounds to generate independent record files;
a merge audio file module 630, configured to mix at least two of the sound recording files into a merge audio file.
With regard to the audio processing apparatus based on the jukebox device in the above-mentioned embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
Further, an embodiment of the present application also provides a computer-readable storage medium, on which computer instructions are stored, and the computer instructions, when executed by a processor, implement the steps of the audio processing method based on a karaoke apparatus according to any one of the above items. The storage medium includes, but is not limited to, any type of disk including floppy disks, hard disks, optical disks, CD-ROMs, and magneto-optical disks, ROMs (Read-Only memories), RAMs (Random AcceSS memories), EPROMs (EraSable Programmable Read-Only memories), EEPROMs (Electrically EraSable Programmable Read-Only memories), flash memories, magnetic cards, or optical cards. That is, a storage medium includes any medium that stores or transmits information in a form readable by a device (e.g., a computer). Which may be a read-only memory, magnetic or optical disk, or the like.
Still further, an embodiment of the present application further provides a computer device, where the computer device includes:
one or more processors;
a storage device for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to implement the steps of the method for audio processing based on a jukebox device of any one of the preceding claims.
FIG. 7 is a block diagram illustrating a computer device 700 according to an example embodiment. For example, the computer device 700 may be provided as a user terminal. Referring to fig. 7, computer device 700 includes a processing component 722 that further includes one or more processors, and memory resources, represented by memory 732, for storing instructions, e.g., applications, that are executable by processing component 722. The application programs stored in memory 732 may include one or more modules that each correspond to a set of instructions. Further, the processing component 722 is configured to execute instructions to perform the steps of the audio processing method described above.
The computer device 700 may also include a power component 726 configured to perform power management of the computer device 700, a wired or wireless network interface 750 configured to connect the computer device 700 to a network, and an input output (I/O) interface 758. The computer device 700 may operate based on an operating system stored in memory 732, such as Windows Server, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM, or the like. It should be understood that, although the steps in the flowcharts of the figures are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and may be performed in other orders unless explicitly stated herein. Moreover, at least a portion of the steps in the flow chart of the figure may include multiple sub-steps or multiple stages, which are not necessarily performed at the same time, but may be performed at different times, which are not necessarily performed in sequence, but may be performed alternately or alternately with other steps or at least a portion of the sub-steps or stages of other steps.
It should be understood that each functional unit in the embodiments of the present application may be integrated into one processing module, each unit may exist alone physically, or two or more units may be integrated into one module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode.
The foregoing is only a partial embodiment of the present application, and it should be noted that, for those skilled in the art, several modifications and decorations can be made without departing from the principle of the present application, and these modifications and decorations should also be regarded as the protection scope of the present application.

Claims (10)

1. An audio processing method based on a point singing device is characterized by comprising the following steps:
acquiring a plurality of original voice frequencies picked up by a point-singing device; the original sound frequency is the recording sound frequency of a plurality of different sound pickup devices of the requesting equipment in the same scene and the same space;
respectively carrying out digital processing on each original sound audio frequency and calling an effector algorithm to generate corresponding effect sound, and respectively carrying out synthesis processing on each original sound audio frequency and the corresponding effect sound to generate mutually independent recording files;
and mixing at least two sound recording files to form a combined audio file.
2. The audio processing method according to claim 1, wherein the step of digitizing each of the original audio frequencies and invoking an effector algorithm to generate corresponding effect sounds, and synthesizing each of the original audio frequencies and the corresponding effect sounds to generate independent recording files comprises:
sequentially carrying out signal amplification processing and analog-to-digital conversion processing on the original audio to obtain corresponding digital signals;
inputting the digital signal into a mainboard through a sound card chip;
calling effect sounds generated by an effector algorithm in the main board, and synthesizing the effect sounds and the digital signals into corresponding recording files according to a preset rule.
3. The audio processing method based on the karaoke apparatus as claimed in claim 2, wherein after the signal amplification processing and the analog-to-digital conversion processing are sequentially performed on the original audio to obtain the corresponding digital signals, the method further comprises:
inputting the digital signal into a digital signal processing circuit, and dynamically processing the digital signal;
superposing and mixing the dynamically processed digital signal and an environment sound signal output by a main board to obtain sound mixing information;
and D/A conversion processing, pre-stage amplification processing and player amplification processing are carried out on the sound mixing information, and the processed sound mixing information is respectively pushed to corresponding players for feedback monitoring.
4. The apparatus for audio processing based on a karaoke device according to claim 1, wherein the step of invoking an effect sound generated by an effector algorithm comprises:
and determining the effect sound corresponding to each original sound audio according to the recording requirement of the recording file.
5. The method for audio processing based on a point-to-sing device according to claim 1, further comprising:
receiving a request sent by a user terminal for acquiring the sound recording file or the combined audio file corresponding to any user;
and responding to the request to call the corresponding sound recording file or combine the audio file and send the sound recording file or the combined audio file to the user terminal.
6. The apparatus for audio processing based on a karaoke as claimed in claim 1, wherein said step of mixing at least two of said audio records into a merged audio file comprises:
acquiring time marks on all the sound recording files;
and according to the time marks, synchronously mixing the sound of the sound recording files to generate a combined audio file.
7. A singing scoring system, comprising:
receiving a singing song and a singing cooperation request selected by a first user, and sending the singing song and the singing cooperation request to related users;
receiving a response of a second user, starting a singing link, picking up original voice frequencies corresponding to the first user and the second user by using different sound pickups of a singing pointing device, and processing the original voice frequencies by using the audio processing method based on the singing pointing device in any one of claims 1 to 6 to obtain a recording file and a combined audio file corresponding to the original voice frequencies;
and scoring the sound recording file and/or the combined audio file.
8. An audio processing apparatus based on a point-to-sing device, comprising:
the original sound audio acquisition module is used for acquiring a plurality of original sound audios picked up by the singing equipment; the original sound frequency is the recording sound frequency of a plurality of different sound pickup devices of the requesting equipment in the same scene and the same space;
a record file generation module, configured to perform digital processing on each original audio frequency and call an effector algorithm to generate corresponding effect sounds, and perform synthesis processing on each original audio frequency and the corresponding effect sounds to generate independent record files;
and the merged audio file obtaining module is used for mixing at least two sound recording files into a merged audio file.
9. A computer-readable storage medium for storing computer instructions which, when executed on a computer, cause the computer to perform the steps of the method for audio processing based on a pointing device according to any one of the preceding claims 1 to 7.
10. A computer device, characterized in that the computer device comprises:
one or more processors;
a storage device for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to implement the steps of the method of audio processing based on a pointing device as claimed in any one of claims 1 to 7.
CN201910943095.3A 2019-09-30 2019-09-30 Audio processing method and device based on point-to-sing equipment, storage medium and equipment Pending CN112581924A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910943095.3A CN112581924A (en) 2019-09-30 2019-09-30 Audio processing method and device based on point-to-sing equipment, storage medium and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910943095.3A CN112581924A (en) 2019-09-30 2019-09-30 Audio processing method and device based on point-to-sing equipment, storage medium and equipment

Publications (1)

Publication Number Publication Date
CN112581924A true CN112581924A (en) 2021-03-30

Family

ID=75116506

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910943095.3A Pending CN112581924A (en) 2019-09-30 2019-09-30 Audio processing method and device based on point-to-sing equipment, storage medium and equipment

Country Status (1)

Country Link
CN (1) CN112581924A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113470613A (en) * 2021-07-16 2021-10-01 北京达佳互联信息技术有限公司 Chorus sound mixing method and device, electronic equipment and storage medium
CN116994545A (en) * 2023-09-25 2023-11-03 苏州至盛半导体科技有限公司 Dynamic original sound adjusting method and device for K song system

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103021401A (en) * 2012-12-17 2013-04-03 上海音乐学院 Internet-based multi-people asynchronous chorus mixed sound synthesizing method and synthesizing system
CN103474082A (en) * 2013-09-16 2013-12-25 无敌科技(西安)有限公司 Multi-microphone vocal accompaniment marking system and method thereof
CN104902012A (en) * 2015-05-11 2015-09-09 福建星网视易信息系统有限公司 Method for carrying out singing contest through network and singing contest system
CN204795576U (en) * 2015-07-30 2015-11-18 徐文波 Audio system
CN106375905A (en) * 2016-11-30 2017-02-01 北京酷我科技有限公司 Method for synthesizing and adjusting multichannel audio
CN107221340A (en) * 2017-05-31 2017-09-29 福建星网视易信息系统有限公司 Real-time methods of marking, storage device and application based on MCVF multichannel voice frequency
CN108109652A (en) * 2017-12-27 2018-06-01 北京酷我科技有限公司 A kind of method of K songs chorus recording
CN108269560A (en) * 2017-01-04 2018-07-10 北京酷我科技有限公司 A kind of speech synthesizing method and system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103021401A (en) * 2012-12-17 2013-04-03 上海音乐学院 Internet-based multi-people asynchronous chorus mixed sound synthesizing method and synthesizing system
CN103474082A (en) * 2013-09-16 2013-12-25 无敌科技(西安)有限公司 Multi-microphone vocal accompaniment marking system and method thereof
CN104902012A (en) * 2015-05-11 2015-09-09 福建星网视易信息系统有限公司 Method for carrying out singing contest through network and singing contest system
CN204795576U (en) * 2015-07-30 2015-11-18 徐文波 Audio system
CN106375905A (en) * 2016-11-30 2017-02-01 北京酷我科技有限公司 Method for synthesizing and adjusting multichannel audio
CN108269560A (en) * 2017-01-04 2018-07-10 北京酷我科技有限公司 A kind of speech synthesizing method and system
CN107221340A (en) * 2017-05-31 2017-09-29 福建星网视易信息系统有限公司 Real-time methods of marking, storage device and application based on MCVF multichannel voice frequency
CN108109652A (en) * 2017-12-27 2018-06-01 北京酷我科技有限公司 A kind of method of K songs chorus recording

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113470613A (en) * 2021-07-16 2021-10-01 北京达佳互联信息技术有限公司 Chorus sound mixing method and device, electronic equipment and storage medium
CN116994545A (en) * 2023-09-25 2023-11-03 苏州至盛半导体科技有限公司 Dynamic original sound adjusting method and device for K song system
CN116994545B (en) * 2023-09-25 2023-12-08 苏州至盛半导体科技有限公司 Dynamic original sound adjusting method and device for K song system

Similar Documents

Publication Publication Date Title
US8019094B2 (en) Audio processing system with function of automatic gain control and method thereof
US7945342B2 (en) Audio processing apparatus for automatic gain control
US20130144626A1 (en) Rap music generation
CN102169705A (en) Tone reproduction apparatus and method
CN110390925B (en) Method for synchronizing voice and accompaniment, terminal, Bluetooth device and storage medium
CN201229768Y (en) Electronic piano
CN103915086A (en) Information processing method, device and system
CN110047497B (en) Background audio signal filtering method and device and storage medium
CN112581924A (en) Audio processing method and device based on point-to-sing equipment, storage medium and equipment
WO2023221559A1 (en) Karaoke audio processing method and apparatus, and computer-readable storage medium
CN111341289A (en) Method and device for intelligent sound box Karaoke
CN108510997A (en) Electronic equipment and echo cancel method applied to electronic equipment
TW201118719A (en) System and method for modifying an audio signal
CN113225574B (en) Signal processing method and device
JP2014066922A (en) Musical piece performing device
CN112927666A (en) Audio processing method and device, electronic equipment and storage medium
JP7143607B2 (en) MUSIC PLAYBACK SYSTEM, TERMINAL DEVICE, MUSIC PLAYBACK METHOD, AND PROGRAM
KR102070071B1 (en) Song playing method, karaoke system and karaoke able to share and replay song performed
CN111147655B (en) Model generation method and device
CN111883090A (en) Method and device for making audio file based on mobile terminal
US11740862B1 (en) Method and system for accelerated decomposing of audio data using intermediate data
US20230395053A1 (en) Sound processing component and string instrument employing component
CN106878841B (en) Microphone assembly
EP4375984A1 (en) Method and system for accelerated decomposing of audio data using intermediate data
CN113611266B (en) Audio synchronization method, device and storage medium suitable for multi-user K songs

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination