WO2016188211A1 - 音频处理方法、装置及系统 - Google Patents

音频处理方法、装置及系统 Download PDF

Info

Publication number
WO2016188211A1
WO2016188211A1 PCT/CN2016/077218 CN2016077218W WO2016188211A1 WO 2016188211 A1 WO2016188211 A1 WO 2016188211A1 CN 2016077218 W CN2016077218 W CN 2016077218W WO 2016188211 A1 WO2016188211 A1 WO 2016188211A1
Authority
WO
WIPO (PCT)
Prior art keywords
file
lyric
audio
audio file
song
Prior art date
Application number
PCT/CN2016/077218
Other languages
English (en)
French (fr)
Inventor
赵伟峰
李纯
刘翔
徐豪杰
张德文
郑飞
曹利
李天文
陈向文
王宝华
王美星
刘智勇
黄斯亮
宋波
王伟龙
邓小康
谭升民
傅鸿城
张庆
石玉磊
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN201510280492.9A external-priority patent/CN105023559A/zh
Priority claimed from CN201510280087.7A external-priority patent/CN104966527B/zh
Priority claimed from CN201510278911.5A external-priority patent/CN105006234B/zh
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Priority to EP16799100.9A priority Critical patent/EP3306606A4/en
Priority to JP2017560972A priority patent/JP2018519536A/ja
Priority to KR1020177037253A priority patent/KR20180012800A/ko
Publication of WO2016188211A1 publication Critical patent/WO2016188211A1/zh
Priority to US15/819,734 priority patent/US10403255B2/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/38Chord
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H7/00Instruments in which the tones are synthesised from a data store, e.g. computer organs
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/005Musical accompaniment, i.e. complete instrumental rhythm synthesis added to a performed melody, e.g. as output by drum machines

Definitions

  • the present invention relates to the field of audio processing, and in particular, to an audio processing method, apparatus, and system.
  • K song application is gradually moving towards people's daily life, and the function of K song application is increasingly rich.
  • the K song application is a combination of a music player and a recording software. It can play the original sing, record the user's song, and mix the recorded song with the accompaniment to get the music file. The user can also upload the music file to the music file. Network so that more people can hear their own voices.
  • the K song application will use the accompaniment audio file to provide the accompaniment for the user's singing voice recording.
  • the user can sing a single song and cannot sing with others (such as stars). K song experience.
  • Embodiments of the present invention provide an audio processing method, apparatus, and system, which can implement a K-song experience that is sung with others.
  • a first aspect of the embodiments of the present invention provides an audio processing method, including:
  • the audio data of the user and the audio data of the display time corresponding to the first lyric portion of the audio file are subjected to a mixing process.
  • a second aspect of embodiments of the present invention provides an audio processing apparatus including at least one processor and a memory, the memory storing program instructions, when the program instructions are executed by a processor, configuring the audio processing apparatus to perform A method of any of the embodiments of the invention.
  • a third aspect of an embodiment of the present invention provides an audio processing system comprising a server and at least one client, the server and/or the client being configured to perform a method according to any embodiment of the present invention.
  • a fourth aspect of the embodiments of the present invention provides an audio processing apparatus, including:
  • a first obtaining unit configured to acquire an audio file of the song and a lyric file of the song
  • a first processing unit configured to play the audio file at a display time corresponding to the first lyric portion of the lyric file and a display time corresponding to the second lyric portion, and display the corresponding lyric portion of the lyric file Time recording user's audio data;
  • a first mixing unit configured to perform mixing processing on the audio data of the user and the audio data of the display time corresponding to the first lyric portion of the audio file.
  • an audio file for example, an accompaniment audio file
  • an audio data of a recording user are played at a display time corresponding to the first lyric portion
  • an audio file is played at a display time corresponding to the second lyric portion (for example, Originally sings the audio file), thereby, it is possible to realize that the song is sung by the user in the display time of the first lyric part, and the lyrics of the second lyric part are sung by others or stars, thereby realizing the karaoke experience of singing with others or stars. .
  • FIG. 1 is a schematic diagram of an audio processing method according to an exemplary embodiment of the present invention
  • FIG. 2 is a schematic diagram of an audio processing method according to another exemplary embodiment of the present invention.
  • FIG. 3 is a schematic diagram of an audio processing method according to still another exemplary embodiment of the present invention.
  • FIG. 4 is a schematic diagram of an audio processing method according to still another exemplary embodiment of the present invention.
  • FIG. 5 is a schematic diagram of an audio processing method according to still another exemplary embodiment of the present invention.
  • FIG. 6 is a schematic diagram of an audio processing device according to an exemplary embodiment of the present invention.
  • FIG. 7 is a schematic diagram of an audio processing device according to another exemplary embodiment of the present invention.
  • FIG. 8 is a schematic diagram of an audio processing device according to an exemplary embodiment of the present invention.
  • FIG. 9 is a schematic diagram of an audio processing system according to an exemplary embodiment of the present invention.
  • FIG. 10 is a schematic diagram of an audio processing system according to another exemplary embodiment of the present invention.
  • Embodiments of the present invention provide an audio processing method, apparatus, and system, which can implement a K-song experience sung with others (such as a star), which will be described in detail below.
  • An embodiment of the present invention provides an audio processing method, including: acquiring an audio file of a song and a lyric file of the song; playing the audio file and recording the audio of the user at a display time corresponding to the first lyric portion of the lyric file Data; playing the audio file at a display time corresponding to the second lyric portion of the lyric file; mixing the audio data of the user with the audio data of the audio file at a display time corresponding to the first lyric portion Tone processing.
  • the audio file may be played at a display time corresponding to the second lyric portion of the lyrics file, but the user's audio data is not recorded.
  • the audio file records the user's audio data, but does not use or remove the audio data of the user recorded at the display time corresponding to the second lyric portion of the lyric file when mixing or generating a chorus file.
  • the audio processing method of the embodiments of the present invention may be performed by any computing device, including but not limited to a mobile phone, a personal computer (PC, Personal Computer), and/or a tablet computer.
  • the audio processing method of the embodiment of the present invention may also be executed by a server or a client in a system including a server and a client or by a server and a client in cooperation.
  • An example of the audio processing method of the embodiment of the present invention is a K song processing method, and those skilled in the art can understand that the audio processing method can have other applications.
  • the lyrics file can be replaced by any other text file or picture, and the audio of the song can be replaced by other audio without departing from the principles of the invention.
  • the client may specifically include a mobile phone, a personal computer (PC, Personal Computer), and/or a tablet computer.
  • FIG. 1 illustrates an audio processing method according to an exemplary embodiment of the present invention.
  • the client is described as an execution body of the audio processing method, and those skilled in the art can understand that the method of the embodiment can be executed by other subjects.
  • the client receives the request of the user.
  • the user when the user wants to select a song for K song, the user needs to input a K song request to trigger the K song processing flow.
  • the client obtains an audio file of the song and a corresponding lyric file according to the request.
  • the audio file of the song may include, for example, an accompaniment audio file of the song, an original singer audio file of the song, and the like.
  • the audio file of the client acquiring the song may include the accompaniment audio file of the client acquiring the song, the original singer audio file of the song, and the like.
  • the client may determine the song selected by the user according to the K song request, and further acquire the accompaniment audio file, the original singer audio file, and the corresponding lyric file of the song.
  • the client can obtain, for example, an accompaniment audio file of the song, an original singer audio file, and a corresponding lyric file from the server.
  • the accompaniment audio file and the original singer audio file can be, for example, bare audio data in the actual application process. If the acquired accompaniment audio file or the original singer audio file is compressed audio data, such as mp3 format, the server first decompresses the compressed format audio data into bare audio data.
  • the lyric file may include, for example, time information for identifying the lyrics display time, so that the lyric file is synchronized with the accompaniment audio file and the original singer audio file according to the time information.
  • the lyrics file may include an LRC lyrics file that identifies the lyrics display time by the time stamp it contains.
  • the client plays the audio file in the display time corresponding to the first lyric portion of the lyric file and the display time corresponding to the second lyric portion, and simultaneously records the audio data of the user at the display time corresponding to the first lyric portion of the lyric file.
  • an audio file is played at a display time corresponding to a second lyric portion of the lyrics file, but the user's audio data is not recorded, for example, a microphone for recording audio data of the user is closed.
  • an audio file is played at the display time corresponding to the second lyric portion of the lyric file, and the user's audio data is also recorded, but is removed (or not used) in the mixing process to be described below or when the chorus file is generated.
  • the accompaniment audio file is played at the display time corresponding to the first lyric portion of the lyric file, in the lyric file
  • the original singer audio file is played at the display time corresponding to the second lyric portion.
  • the lyric file is divided and processed at the client to obtain the first lyric portion and the second lyric portion.
  • the client can divide the lyric file into a plurality of implementation manners to obtain a first lyric portion and a second lyric portion.
  • the following examples are respectively illustrated:
  • the first is that, before the step 103, the client receives the user's division instruction, and then divides the lyric file according to the division instruction to obtain the first lyric portion and the second lyric portion.
  • a plurality of lyrics division processing strategies may be preset on the client side, and the lyrics file is divided into two lyrics portions by the lyrics division processing strategy described above.
  • the client determines the paragraph information of the lyrics file according to the time information in the lyric file, and then divides the lyric file into two lyric parts according to the paragraph information.
  • the time information of the lyric file may reflect the start point and the end point of the display time of each paragraph of the corresponding song, whereby the client may determine, according to the time information, that the lyric file includes several paragraphs and the display time corresponding to each paragraph The starting point and the end point.
  • the server may divide the preset at least one paragraph into one lyric part, and the other part is divided into another lyric part.
  • the number of paragraphs of each lyric part and the specific paragraph or paragraphs are not limited herein.
  • the structure of paragraph information can be expressed as:
  • the role id is used to indicate whether the song corresponds to a female singing part or a male singing part.
  • the paragraph corresponding to the female singing part can be set as a lyric part, and the corresponding paragraph of the male singing part is set as another lyric part.
  • the client determines the main song part and the chorus part of the lyric file, and determines the main song part as one lyric part and the chorus part as another lyric part.
  • the client can use the difference between the main song and the chorus to divide the lyrics, so as to obtain a lyric part corresponding to the main song, corresponding to another lyric part of the chorus, so that the difference between the main song and the chorus will be
  • the lyrics file is divided into two lyric parts.
  • the client may pre-define a plurality of templates for dividing the lyrics, and then divide the lyric file according to the customized template to obtain two lyric portions.
  • a fixed N sentence lyrics or N words are set as a lyric part.
  • there is no limitation on how to set a customized template Those skilled in the art can set a plurality of different templates according to actual conditions.
  • the client when receiving the division instruction of the user, the client performs a division process according to the division instruction by using a corresponding lyric division processing strategy, thereby obtaining a corresponding first lyric portion and The second lyrics section.
  • the client may further provide a selection module for the user to select the two lyric portions obtained by the division to specify which part is used as the first lyric part or which part is the second part. For example, by selecting a module, the user may choose to use the male singing part as the first lyric part for singing, or the male singing part as the second lyric part for singing by others or by the star.
  • the client may play the accompaniment audio file and record the user's audio data at the display time corresponding to the first lyric portion, in the second
  • the lyrics part of the display time plays the original singer audio file.
  • the second is that the client divides the lyric file according to whether there is sound input, and obtains the first lyric part and the second lyric part.
  • the client detects whether there is a voice input, and if so, divides the lyrics portion of the lyric file into the first lyric portion in the corresponding display time, and plays the accompaniment audio file and switches to the microphone input mode to record the user's audio data. If not, the lyrics portion of the lyrics file in the corresponding display time is divided into the second lyrics portion, and the original singer audio file is played.
  • the division processing operation of the lyric file is synchronously performed with the operations of playing, recording, and the like, and the client does not need to wait for the first lyric portion and the second lyric portion to be played, and then the client plays the display time corresponding to the first lyric portion.
  • the accompaniment audio file and the audio data of the recording user are played, and the original singer audio file is played at the display time of the second lyric portion.
  • the client can combine the above-mentioned dividing processing manner. If you use it, you can use more other division processing methods, which are not limited here.
  • the client can further identify the lyrics displayed on the client, thereby preferably, this embodiment can also The method further includes: setting, by the client, a first identifier on a first lyric portion displayed on the client, and setting a second identifier on a second lyric portion displayed on the client.
  • the identifier may be a color identifier, for example, the first lyric portion displays one color, the second lyric portion displays another color, or the identifier is a highlighted state identifier, for example, the first lyric portion is displayed as high. In the bright state, the second lyrics part is displayed as a non-highlighted state, specifically
  • the specific implementation form of the identifier is not limited, and those skilled in the art can obtain more other identification forms according to requirements.
  • the client performs mixing processing on the audio data of the user and the audio data of the audio file at the display time corresponding to the first lyric portion to obtain a mixed audio file.
  • the client after obtaining audio data of the user (for example, audio data collected by a microphone), the client performs sound mixing processing on the audio data of the user and the audio data of the audio file at the display time corresponding to the first lyric portion.
  • the mixed audio file is obtained, whereby the mixed audio file can embody all the audio data of the user's singing part, including the user's voice and the corresponding accompaniment.
  • the specific process of the mixing process can be referred to the existing mixing algorithm, for example, directly adding the corresponding sampling point data by two.
  • the client Under the condition that the audio file includes the accompaniment audio file and the original singer audio file, the client performs a mixing process on the audio data of the user and the audio data of the accompaniment audio file at the display time corresponding to the first lyric portion.
  • the client after the client performs audio mixing processing on the audio data of the user and the audio data of the display time corresponding to the first lyric portion, the client outputs the display time corresponding to the first lyric portion (or The sound data file is selected and the audio file is output (or selected) at the display time corresponding to the second lyric portion to generate a chorus file.
  • the audio file includes the accompaniment audio file and the original singer audio file
  • the original singer audio file is output (or selected) at the display time corresponding to the second lyric portion.
  • the audio data of the mixed audio file in the corresponding time period is obtained, and if the display time corresponding to the second lyric part is obtained, the original vocal audio is obtained.
  • the audio data of the file in the corresponding time period is combined with the audio data of each time point into an audio file in chronological order, and the synthesized audio file is a chorus file. It should be emphasized that in the above process of generating the chorus file, the audio data of the specific output target audio file and the original singer audio file is determined by the display time of the lyrics.
  • the obtained chorus file is displayed at the display time corresponding to the first lyric portion, and the user sing portion corresponding to the corresponding time period is output, and the original sing portion corresponding to the corresponding time period is output at the display time corresponding to the second lyric portion.
  • the client performs balancing processing on the audio data of the switching file at the switching point, where the switching point refers to switching the display time corresponding to the first lyric portion to the corresponding corresponding to the second lyric portion.
  • the time point at which the time is displayed or the display time corresponding to the second lyric portion is switched to the time point of the display time corresponding to the first lyric portion.
  • the specific balancing processing manner can be referred to the existing audio splicing technology, such as averaging filtering, superimposing processing, and the like.
  • the user may also upload the chorus file to the network through the client to share the chorus with others or stars.
  • the audio file described above for the song acquired by the client may include the accompaniment audio file of the song and the original singer audio file of the song.
  • the audio file of the song acquired by the client may be the target audio file to be described below, and therefore, the audio file of the client acquiring the song may include, for example, a target audio file transmitted by the client receiving server.
  • the lyric file of the client acquiring the song may include, for example, a target lyric file sent by the client receiving server.
  • the specific process for the server to send the target audio file and/or the target lyric file will be described in detail below. In the case that the client receives the target audio file and/or the target lyric file sent by the server, in the above step 103, the client may not perform the dividing process on the lyric file.
  • the accompaniment audio file and the audio data of the recording user are played only at the display time corresponding to the first lyric portion, and the original singer audio file is played at the display time corresponding to the second lyric portion. It is possible to realize that the song is sung by the user in the display time of the first lyric part, and the lyrics of the second lyric part are originally sung by others or stars, thereby realizing the karaoke experience of chorus with others or stars. Moreover, in this embodiment, the K-song experience of sing with other people or stars can be realized directly by the client using the ordinary accompaniment audio file.
  • FIG. 2 illustrates an audio processing method according to another embodiment of the present invention.
  • the server is described as an execution body of the audio processing method, and those skilled in the art can understand that the method of the embodiment can be executed by other subjects.
  • the server acquires an accompaniment audio file of the song, an original singer audio file, and a corresponding lyric file.
  • the accompaniment audio file, the original singer audio file, and the corresponding lyric file have been described in detail above, and will not be repeated here for brevity.
  • the server divides the lyric file to obtain a first lyric portion and a second lyric portion, and generates lyric split information correspondingly.
  • the server may divide the lyric file into a plurality of implementation manners to obtain a first lyric portion and a second lyric portion.
  • the following is respectively exemplified by the following: the server determines the paragraph information of the lyric file according to the time information, and then according to the The paragraph information is divided into the lyric file to obtain a first lyric portion and a second lyric portion.
  • the second type is: the server determines the main song part and the chorus part of the lyric file, and determines the main song part as the first lyric part, the chorus part as the second lyric part, or the chorus part as the first In the lyrics section, the main song portion is determined as the second lyric portion.
  • the third type is: the server divides the lyric file according to the customized template to obtain the first lyric part and the second lyric part.
  • the method may further include: receiving, by the client, a partitioning instruction, where the dividing instruction is used to indicate a dividing processing manner specified by the user, and the server may divide the lyric file according to the dividing instruction.
  • the first lyric portion and the second lyric portion are obtained, thereby enabling the user to select how to divide the lyrics and obtain the first lyric portion and the second lyric portion in accordance with the user's wishes.
  • the server further records the corresponding dividing processing result to generate lyrics splitting information, wherein the lyrics splitting information includes Identification information of the first lyric portion.
  • the lyrics split information may further include identification information for identifying the second lyric portion.
  • the server outputs (or selects) an accompaniment audio file at a display time corresponding to the first lyric portion, and outputs (or selects) the original sing audio file at a display time of the second lyric portion to generate a target audio file.
  • the audio data of the accompaniment audio file in the corresponding time period is acquired, and if the display time corresponding to the second lyric portion is obtained, the original vocal audio is acquired.
  • the audio data of the file in the corresponding time period, followed by time The audio data of each time point is synthesized into an audio file, and the synthesized audio file is the target audio file. It should be emphasized that in the above process of generating the target audio file, the audio data of the specific output accompaniment audio file and the original singer audio file is determined by the display time of the lyrics.
  • the target audio file when the target audio file is played, the accompaniment of the corresponding time period is output at the display time corresponding to the first lyric portion, and the original sing of the corresponding time period is output at the display time corresponding to the second lyric portion, that is,
  • the target audio file enables automatic switching between the accompaniment and the original vocals.
  • the method further includes: performing, by the server, balancing processing on the audio data of the target audio file at the switching point, where the switching point refers to switching the display time corresponding to the first lyric portion to the display corresponding to the second lyric portion.
  • the time point of time or the display time corresponding to the second lyric portion is switched to the time point of the display time corresponding to the first lyric portion.
  • the specific balancing processing manner can be referred to the existing audio splicing technology, such as averaging filtering, superimposing processing, and the like.
  • the server sends the target audio file and the target lyric file to the client.
  • the server Upon receiving the request sent by the client (for example, a K song request), the server sends the target audio file and the target lyric file to the client, so that the client uses the target audio file as the accompaniment audio and the target lyric file as the lyrics to be played to the user.
  • the target lyrics file is obtained by carrying the lyrics split information in the lyric file.
  • the client when playing the target audio data, the client outputs the accompaniment of the corresponding time period at the display time of the first lyric portion, and the display time corresponding to the second lyric portion is outputted for the corresponding time period.
  • the original sing by which, the display time of the first lyric part can be sung by the user, and the lyric part of the second lyric part can be sung by others (or sung by the star), thereby realizing the karaoke experience of chorus with others or stars.
  • the target audio file and the target lyric file are generated by the server, and the target audio file and the target lyric file are sent to the client, so that the K-song experience with other people or stars can be realized. There is no need to change the client logic.
  • the target audio file sent by the server to the client corresponds to the accompaniment in the display time of the first lyric portion, and the display time in the second lyric portion corresponds to the original sing
  • the embodiment of the present invention can realize that the display time of the first lyric part is sung by the user, and the display time of the second lyric part is displayed. Singing by others or stars to achieve a K-song experience with others or stars.
  • FIG. 3 illustrates an audio processing method according to another exemplary embodiment of the present invention, including:
  • the server acquires an accompaniment audio file of the song, an original singer audio file, and a corresponding lyric file;
  • the server divides the lyric file to obtain a first lyric portion and a second lyric portion.
  • the server outputs an accompaniment audio file at a display time corresponding to the first lyric portion, and outputs an original sing audio file at a display time of the second lyric portion to generate a target audio file.
  • the server sends the target audio file and the target lyric file to the client.
  • steps 301 to 304 is the same as the process of step 201 to step 204 in the embodiment shown in FIG. 2, and details are not described herein again.
  • the client plays the target audio file and the target lyric file.
  • the client's K song application After receiving the target audio file and the target lyrics file sent by the server, the client's K song application plays the target audio file and the target lyric file, wherein the target audio file is played as the accompaniment audio.
  • the client records the audio data of the user according to the lyrics division information at a display time corresponding to the first lyric portion.
  • the client may identify the display time corresponding to the first lyric portion and the display time corresponding to the second lyric portion according to the lyric division information, and record the user's audio data at the display time corresponding to the first lyric portion.
  • the display time corresponding to the second lyric part stops recording.
  • the target file played by the client corresponds to the original sing of the other person or the star, thereby realizing that the display time of the first lyric part is sung by the user, and the display time of the second lyric part is realized. Singing by others or stars to achieve a K-song experience with others or stars.
  • the client performs mixing processing on the user's audio data and the target audio file to obtain a chorus file.
  • the obtained chorus file is the audio data sung by the user in the first lyric part, and the audio data of the original singer in the second lyric part is displayed (ie, the original singer audio file is in the Corresponding time period audio data).
  • the user may also upload the chorus file to the network through the client, so as to share the other with others or The chorus of the star.
  • FIG. 4 illustrates an audio processing method according to still another embodiment of the present invention, including:
  • the first client receives the first request of the first user.
  • the first request is, for example, a first chorus request for requesting initiation of a chorus.
  • the first client may set a button corresponding to the first chorus request, and when the first user presses the button, the first client receives the first chorus request.
  • the first client acquires an accompaniment audio file and a lyric file of the song from the server according to the first request, so that the first user performs recording based on the partial lyrics by using the accompaniment audio file and the lyric file to obtain the first recorded file.
  • the accompaniment audio file, the original singer audio file, and the corresponding lyric file have been described in detail above, and will not be repeated here for brevity.
  • the client After acquiring the accompaniment audio file and the lyric file, the client plays the accompaniment audio file and the lyric file, and the first user can select any lyric to sing. Of course, the first user does not sing all the lyrics, and the client first The user performs recording using the accompaniment audio file and the lyrics file based on the lyrics of the partial lyrics to obtain the first recorded file.
  • the first client determines a second lyric portion based on the first recording file, and generates segment information correspondingly.
  • the second lyrics part refers to the part of the lyrics sung by the first user, that is, the lyrics part that the first client recognizes based on the first recorded file as the part of the lyrics sung by the first user, and correspondingly generates paragraph information, the paragraph information Identification information for identifying the second lyric portion is included to record the portion of the lyrics sung by the first user.
  • the identification information may include a display time corresponding to the second lyric portion, for example, the start time of the second lyric portion is 10S, the end time is 15S, etc., or the identification information may include a sentence identifier of the second lyric portion, for example
  • the second lyrics part is the first, third, and fifth sentences, etc., and the specific form of the identification information is not limited herein, as long as the first client can identify the second lyric part according to the identifier.
  • the first client may use multiple implementation manners to determine the second lyrics portion. For example, the first client may determine the lyric portions of the lyrics that the first user sang for the validity of the sound in the first recorded file. include:
  • the first client determines whether the voice recorded by the first user in a certain lyric portion is valid, and if so, Then, the corresponding lyric portion is determined to be the second lyric portion.
  • the first client determines whether the volume of the sound is greater than the first preset value, and if yes, determines that the sound recorded by the first user in a certain lyric portion is valid.
  • the VAD algorithm or its improved algorithm.
  • the first client further determines whether the time occupied by the recorded sound corresponds to the display time occupied by the lyric part, and if yes, determines the lyrics.
  • Part is the second lyrics section.
  • the sentence part of the lyrics part after determining that the sound input by a certain lyrics is valid, if the sound recorded in the lyrics is occupied by the display time is 20S-30S, and the lyrics occupied by the sentence are occupied.
  • the display time is also 20 to 30, then it can be determined that the lyrics of the sentence is the second lyrics part, and a threshold value can be set here, when the occlusion time occupied by the lyrics and the display time occupied by the lyrics are greater than the display time At the threshold, it is determined that the display time occupied by the lyrics satisfying the sentence corresponds to the display time occupied by the lyrics of the sentence.
  • the first client may introduce an audio real-time detection module including the above determined manner to determine the second lyric portion.
  • the first client performs a mixing process on the first recording file and the accompaniment audio file to obtain a target audio file.
  • the specific process of the mixing process can be referred to the existing mixing algorithm, for example, directly adding the corresponding sampling point data by two.
  • the method further includes: before the step 404 is performed, in order to prevent the lyrics portion of the first lyric file from being included in the lyrics portion of the second lyric portion.
  • the first client clears the audio data corresponding to the lyric portion of the first recorded file other than the second lyric portion according to the paragraph information, thereby ensuring that the obtained first recorded file is only displayed at the second lyric portion Contains audio data, which is equivalent to only containing the effective sound of the first user singing.
  • the first client sends the target audio file and the paragraph information to the server, so that the second client obtains the target audio file and the paragraph information from the server, and uses the target audio file as the accompaniment audio, and provides the second user according to the paragraph information.
  • K song service The first client sends the target audio file and the paragraph information to the server, so that the second client obtains the target audio file and the paragraph information from the server, and uses the target audio file as the accompaniment audio, and provides the second user according to the paragraph information.
  • K song service K song service.
  • the second client obtains the target audio file and the paragraph information from the server, and may provide the karaoke service to the second user according to the paragraph information by using the target audio file as the accompaniment audio. Specifically, the second user can determine which lyrics have been sung according to the paragraph information, and sing in the lyrics portion that has not been sung.
  • the embodiment may further include: the first client generates a configuration file, where the configuration file includes a sound effect parameter adopted by the first user; and the first client sends the configuration file to the server, so that the second client The target audio file is used as the accompaniment audio, and the karaoke service is provided to the second user according to the paragraph information and the configuration file.
  • the sound effect parameters include a reverberation parameter, a variable pitch parameter, a sound beautification parameter, and the like.
  • the target audio file in the embodiment of the present invention is sung by the first user only at the display time of the second lyric portion, and the lyric portion outside the second lyric portion is accompaniment, thereby, when the second client
  • the lyrics portion other than the second lyric portion may be sung according to the paragraph information, so that the display time of the second lyric portion is sung by the first user, and the second lyric portion is not.
  • the display time of the lyrics portion is sung by the second user, thereby realizing the K-song experience of sing with others.
  • FIG. 5 illustrates another embodiment of an audio processing method according to an embodiment of the present invention, including:
  • the first client receives a first request of the first user.
  • the first client acquires an accompaniment audio file and a lyric file of the song from the server according to the first request, so that the first user performs recording based on the partial lyrics by using the accompaniment audio file and the lyric file to obtain the first recorded file.
  • the first client determines a second lyric part based on the first recording file, and generates segment information correspondingly.
  • the first client performs a mixing process on the first recording file and the accompaniment audio file to obtain a target audio file.
  • the first client sends the target audio file and the paragraph information to the server, so that the second client obtains the target audio file and the paragraph information from the server, and uses the target audio file as the accompaniment audio, and provides the second user according to the paragraph information.
  • K song service The first client sends the target audio file and the paragraph information to the server, so that the second client obtains the target audio file and the paragraph information from the server, and uses the target audio file as the accompaniment audio, and provides the second user according to the paragraph information.
  • K song service K song service.
  • steps 501 to 505 are the same as the processes of steps 401 to 405 in the embodiment shown in FIG. 4, and details are not described herein again.
  • the second client receives the second request sent by the second user.
  • the second request is for requesting to join the chorus to trigger the chorus joining process.
  • the second client may set a button corresponding to the second chorus request, and when the second user presses the button, the second client receives the second chorus request.
  • the second client acquires a target audio file, a lyric file, and a paragraph information of the song from the server according to the second request, so that the second user uses the target audio file as the accompaniment audio and based on the paragraph information based on all of the second lyric portion or Part of the lyrics is recorded to get the second recorded file.
  • the second client Upon receiving the second request, the second client first obtains the target audio file, the lyric file, and the paragraph information from the server, and then plays the target audio file and the lyric file, and the second user is based on all or part of the lyrics except the second lyric portion. Singing, while the second client records the second user with the target audio file as the accompaniment audio and based on the paragraph information based on all or part of the lyrics other than the second lyric portion, to obtain the second recorded file. It should be emphasized that if the second user sings based on part of the lyrics other than the second lyric part, instead of singing based on all the lyric parts except the second lyric part, the song has a part of the lyrics that can be used by more other users. Join the chorus.
  • the step 507 may include: the second client obtains the target audio file, the lyric file, and the paragraph information of the song from the server. And a configuration file, so that the second user records the target audio file as the accompaniment audio and based on the paragraph information and the configuration file based on part or all of the lyrics portion other than the second lyric portion, to obtain the second recording file.
  • the client performs corresponding configuration according to the sound effect parameter in the configuration file, and sends it to the second user for selection by the second user. The user can also modify the corresponding parameter configuration according to actual needs.
  • the second client performs a mixing process on the second recording file and the target audio file to obtain a mixed audio file.
  • the specific process of the mixing process can be referred to the existing mixing algorithm, for example, directly adding the corresponding sampling point data by two.
  • the mixed audio file is sung by the first user in the second lyric part, and part or all of the lyric part outside the second lyric part is sung by the second user, thereby achieving two The karaoke experience of the user chorus.
  • the sound is determined to be invalid, and the second client may be further included before the step 508 is performed.
  • the terminal clears the audio data corresponding to the first lyric portion in the first recorded file according to the paragraph information.
  • FIG. 6 is an audio processing device according to an exemplary embodiment of the present invention, which includes the first The receiving unit 601, the first obtaining unit 602, the first processing unit 603, and the first mixing unit 604.
  • the first receiving unit 601 is configured to receive a request of the user.
  • the first obtaining unit 602 is configured to acquire an audio file of the song and a corresponding lyric file according to the request.
  • the lyric file includes time information for identifying a lyric display time.
  • the audio file of the song may include, for example, an accompaniment audio file of the song, an original singer audio file of the song, and the like.
  • the first processing unit 603 is configured to play the audio file at a display time corresponding to the first lyric portion and a display time corresponding to the second lyric portion, and simultaneously record the audio of the user at a display time corresponding to the first lyric portion of the lyric file. data.
  • the first processing unit 603 may be further configured to perform a dividing process on the lyric file to obtain a first lyric portion and a second lyric portion.
  • the first processing unit 603 plays the accompaniment audio file and the audio data of the recorded user at the display time corresponding to the first lyric portion.
  • the original singer audio file is played at the display time of the second lyric portion.
  • the first mixing unit 604 is configured to perform mixing processing on the audio data of the user and the audio data of the display time corresponding to the first lyric portion of the audio file to obtain a mixed audio file.
  • the first mixing unit 604 performs a mixing process on the audio data of the user and the audio data of the accompaniment audio file at the display time corresponding to the first lyric portion, under the condition that the audio file includes the accompaniment audio file and the original singer audio file.
  • the first receiving unit 601 receives the K song request input by the user; the first obtaining unit 602 according to the The K song requests to acquire the accompaniment audio file of the song, the original singer audio file, and the corresponding lyric file, wherein the lyric file includes time information for identifying the lyrics display time; the first processing unit 603 divides the lyric file Processing, obtaining a first lyric portion and a second lyric portion, and playing the accompaniment audio file and recording the user's audio data at a display time corresponding to the first lyric portion, and playing the original at the display time of the second lyric portion Sing an audio file.
  • the first acquiring unit 602 of the client acquires the accompaniment audio file and the original singer audio file of the song according to the K song request. And a corresponding lyric file, wherein the lyric file includes time information for identifying the lyrics display time; then the lyric file is divided by the first processing unit 603 to obtain the first lyric portion and the second lyric portion, and The lyrics portion corresponding to the display time plays the accompaniment audio file and the audio data of the recorded user, and the original singer audio file is played at the display time of the second lyric portion; and the audio data and the accompaniment audio file of the recorded user are in the first
  • the audio data of the display time corresponding to a lyric part is subjected to mixing processing to obtain a mixed audio file, and the mixed audio file is outputted at the display time corresponding to the first lyric portion, and the original singed audio is outputted at the display time corresponding to the second ly
  • the accompaniment audio file and the recording are played only at the display time corresponding to the first lyric portion.
  • the user's audio data is played, and the original singer audio file is played at the display time corresponding to the second lyric portion.
  • a part of the display time of the first lyric part can be sung by the user, and the display time of the second lyric part is by the original star. Sing, thus achieving the K-song experience with the stars.
  • the audio processing apparatus may further include:
  • the first generating unit 605 is configured to output (or select) the mixed audio file at a display time corresponding to the first lyric portion, and output (or select) the audio at a display time corresponding to the second lyric portion.
  • File to generate a chorus file In the case that the audio file includes the accompaniment audio file and the original singer audio file, in order to generate the chorus file, the original singer audio file is output (or selected) at the display time corresponding to the second lyric portion.
  • the first receiving unit 601 is further configured to receive a dividing instruction of the user, where the first processing unit 603 is specifically configured to perform, according to the dividing instruction, a division process on the lyric file to obtain a first A lyric part and a second lyric part.
  • the first processing unit 603 is specifically configured to detect whether there is a voice input, and if yes, divide the lyrics portion of the lyric file in the corresponding display time into a first lyric portion, and play the accompaniment audio.
  • the file and the audio data of the recording user if not, dividing the lyrics portion of the lyrics file into the second lyric portion in the corresponding display time, and playing the original singer audio file.
  • the audio processing device may further include:
  • an identifier unit configured to set a first identifier on a first lyric portion displayed on the client, and a second identifier on a second lyric portion displayed on the client.
  • the audio processing device may further include:
  • a first balancing processing unit configured to perform balancing processing on the audio data of the chorus file at the switching point, wherein the switching point refers to switching from a display time corresponding to the first lyric portion to a display corresponding to the second lyric portion The time point of time or the display time corresponding to the second lyric portion is switched to the time point of the display time corresponding to the first lyric portion.
  • FIG. 7 illustrates an audio processing apparatus according to still another embodiment of the present invention, which includes:
  • the second obtaining unit 701 is configured to acquire an accompaniment audio file of the song, an original singer audio file, and a corresponding lyric file, wherein the lyric file includes time information for identifying a lyric display time;
  • the second processing unit 702 is configured to perform a dividing process on the lyric file to obtain a first lyric portion and a second lyric portion, and correspondingly generate lyrics splitting information, where the lyrics splitting information includes Identification information of the lyrics part;
  • a second generating unit 703 configured to output (or select) the accompaniment audio file at a display time corresponding to the first lyric portion, and output (or select) the original sing audio file at a display time of the second lyric portion to generate Target audio file;
  • a first sending unit 704 configured to send the target audio file and the target lyrics file to the client (for example, when receiving a K song request sent by a client), so that the client uses the target
  • the audio file is provided as an accompaniment audio
  • the lyrics file is provided as a lyrics to be played to the user, wherein the target lyrics file is obtained by carrying the lyrics split information in the lyric file.
  • the target audio file sent by the audio processing device in the server to the client in the embodiment of the present invention corresponds to the accompaniment in the display time of the first lyric portion, and the display time in the second lyric portion corresponds to the original sing, and Compared with providing accompaniment audio files directly to the client in the prior art,
  • the embodiment of the invention can realize that the display time of the first lyric part is sung by the user, and the display time of the second lyric part is sung by the star, thereby realizing the karaoke experience of chorus with the star.
  • the second processing unit 702 may determine the paragraph information of the lyric file according to the time information, and divide the lyric file according to the paragraph information to obtain a first lyric part and a second lyric section.
  • the second processing unit 702 may determine a main song portion and a chorus portion of the lyric file, and determine the main song portion as a first lyric portion, and determine the chorus portion as a first And the lyrics portion is determined as the first lyric portion, and the main song portion is determined as the second lyric portion.
  • the second processing unit 702 may perform the dividing process on the lyric file according to the customized template to obtain the first lyric portion and the second lyric portion.
  • the audio processing device may further include:
  • a second balancing processing unit configured to perform balancing processing on the audio data of the target audio file at the switching point, wherein the switching point refers to switching from a display time corresponding to the first lyric portion to a corresponding second lyric portion The time point at which the time is displayed or the display time corresponding to the second lyric portion is switched to the time point of the display time corresponding to the first lyric portion.
  • FIG. 8 illustrates an audio processing apparatus according to another embodiment of the present invention, including:
  • a second receiving unit configured to receive a first request of the first user
  • the 802 third obtaining unit is configured to acquire an accompaniment audio file and a lyric file of the song from the server according to the first request, so that the first user performs recording based on the partial lyrics by using the accompaniment audio file and the lyric file to obtain the first recorded file;
  • a third processing unit 803, configured to determine a second lyric portion based on the first recording file, and correspondingly generate paragraph information, where the paragraph information includes at least identifier information for identifying the second lyric portion;
  • the second mixing unit is configured to perform mixing processing on the first recording file and the accompaniment audio file to obtain a target audio file.
  • the second sending unit is configured to send the target audio file and the paragraph information, so that the second client obtains the target audio file and the paragraph information from the server, and provides the second user with the target audio file as the accompaniment audio according to the paragraph information. K song service.
  • FIG. 9 illustrates an audio processing system including a server 901 and a client 902, according to an exemplary embodiment of the present invention.
  • the audio processing method of the above-described embodiments of the present invention may be executed by a server and/or a client of the audio processing system.
  • the server 901 includes an audio processing device as described in the embodiment shown in FIG. 7;
  • the client 902 includes an audio processing device as described in the embodiment of FIG.
  • FIG. 10 illustrates an audio processing system including a first client 1001, a second client 1002, and a server 1003, according to another exemplary embodiment of the present invention.
  • the first client 1001 is configured to receive a first request of the first user, where the first request is used, for example, to request to initiate a chorus; and acquire an accompaniment audio file and a lyric file of the song from the server 1003 according to the first request, So that the first user uses the accompaniment audio file and the lyrics file to perform recording based on partial lyrics to obtain a first recorded file; determining a second lyric portion based on the first recorded file, and correspondingly generating paragraph information, wherein the a second lyric portion refers to a lyric portion sung by the first user, the paragraph information includes identification information for identifying the second lyric portion; and the first recorded file and the accompaniment audio file are mixed Processing, obtaining a target audio file; and transmitting the target audio file and paragraph information to the server 1003, so that the second client obtains the target audio file and the paragraph information from the server 1003, and the target
  • the audio file serves as accompaniment audio, and provides a karaoke service to the second user based on the paragraph information.
  • the second client 1002 is configured to acquire the target audio file and the paragraph information from the server 1003, and provide the karaoke service to the second user according to the paragraph information by using the target audio file as the accompaniment audio.
  • the first client 1001 is specifically configured to determine whether the voice recorded by the first user in a certain lyric portion is valid, and if yes, determine that the corresponding lyric portion is the second lyric portion.
  • the first client 1001 is specifically configured to determine whether the volume of the sound is greater than a preset value, and if yes, determine that the sound recorded by the first user in a certain lyric portion is valid.
  • the first client 1001 may be specifically configured to determine whether the voice recorded by the first user in a certain lyric portion is valid, and if yes, further determine the voice of the input. Whether the occupied time corresponds to the display time occupied by the lyric portion, and if so, it is determined that the lyric portion is the second lyric portion.
  • the first client 1001 is further configured to perform, after the first client, performs a mixing process on the first recording file and the accompaniment audio file to obtain a target audio file, according to the The paragraph information clears the audio data corresponding to the lyric portion of the first recorded file other than the second lyric portion.
  • the identifier information includes a display time corresponding to the second lyric portion or a sentence identifier of the second lyric portion.
  • the first client 1001 may be further configured to: before the first client sends the target audio file and the paragraph information to the server, generate a configuration file, where the configuration file includes the first a sound effect parameter employed by a user; and for transmitting the configuration file to the server.
  • the second client 1002 is specifically configured to receive a second request sent by the second user, where the second request is used to request to join the chorus; and the song is obtained from the server according to the second request.
  • a target audio file, a lyric file, and paragraph information so that the second user uses the target audio file as the accompaniment audio and records based on the paragraph information based on all or part of the lyrics except the second lyric portion, to obtain a second Recording a file; and mixing the second recorded file and the target audio file to obtain a mixed audio file.
  • the second client 1002 is further configured to perform mixing processing on the second recording file and the target audio file in the second client, before obtaining the mixed audio file, according to the paragraph
  • the information clears the audio data corresponding to the second lyric portion of the first recorded file.
  • the second client 1002 is specifically configured to: when the server further includes the configuration file, obtain a target audio file, a lyric file, a paragraph information, and a configuration file of the song from the server, so that The second user uses the target audio file as the accompaniment audio and performs recording based on the segment information and the configuration file based on part or all of the lyrics portion other than the second lyric portion, to obtain a second recorded file.
  • the second client is specifically configured to receive a third request sent by the second user, where the third request is used to request to join the chorus and indicate only part of the song other than the second lyric part
  • the word is recorded;
  • the target audio file, the lyric file, and the paragraph information of the song are obtained from the server according to the third request, so that the second user uses the target audio file as the accompaniment audio and based on the paragraph information Part of the lyrics portion other than the second lyric portion is recorded to obtain a third recording file; determining a third lyric portion, and adding identification information for identifying the third lyric portion to the paragraph information, wherein a third lyrics portion refers to a portion of the lyrics sung by the second user; performing a mixing process on the third recorded file and the target audio file to obtain a third target file; and transmitting the third to the server An object file and paragraph information, so that the third client passes the server for the third user to use the third object file as the accompaniment audio and according to the paragraph information outside the second lyric portion and the
  • the second client 1002 of FIG. 10 is configured to perform the audio processing method of the exemplary embodiment of the present invention described with reference to FIG. 1, the server 1003 for performing the present invention described with reference to FIG.
  • An audio processing method of an exemplary embodiment the first client 1001 is configured to perform an audio processing method of an exemplary embodiment of the present invention described with reference to FIG. 4,
  • an exemplary embodiment of the present invention provides an audio processing device including at least one processor and a memory, the memory storing program instructions that, when executed by a processor, configure the audio processing device to perform The method described in any of the above embodiments.
  • an exemplary embodiment of the present invention provides a computer readable storage medium storing program instructions that, when executed by a processor of a computing device, configure the computing device to perform according to any of the above embodiments Methods.
  • the disclosed system, apparatus, and method may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be indirect coupling through some interface, device or unit.
  • a communication connection which may be in electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the integrated unit if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium.
  • the technical solution of the present invention which is essential or contributes to the prior art, or all or part of the technical solution, may be embodied in the form of a software product stored in a storage medium.
  • a number of instructions are included to cause a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present invention.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like. .

Abstract

一种音频处理方法、装置和系统,可以实现与他人合唱的K歌体验。该方法包括:获取歌曲的音频文件和所述歌曲的歌词文件;在所述歌词文件的第一歌词部分对应的显示时间播放所述音频文件并录制用户的音频数据;在所述歌词文件的第二歌词部分对应的显示时间播放所述音频文件;对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。

Description

音频处理方法、装置及系统
相关申请的交叉引用
本申请要求2015年5月27日提交中国专利局、申请号为201510278911.5、发明名称为“一种K歌处理方法及装置”,以及2015年5月27提交中国专利局、申请号为201510280087.7、发明名称为“K歌处理方法、装置以及K歌处理系统”以及2015年5月27提交中国专利局、申请号为201510280492.9发明名称为“K歌处理方法及系统”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本发明涉及音频处理领域,尤其涉及一种音频处理方法、装置以及系统。
背景技术
随着人们生活质量的提高,K歌应用正逐步走向人们的日常生活,K歌应用具有的功能也日益丰富。
K歌应用是音乐播放器和录音软件的结合体,既可以播放原唱,也可以录制用户的歌声,并将录制的歌声与伴奏进行混音得到音乐文件,用户还可以将上述音乐文件上传到网络,以便让更多的人听到自己的歌声。
通常,在用户录制自己的歌声时,K歌应用会使用伴奏音频文件为用户的歌声录制提供伴奏,然而,这过程中,只能用户一人单独K歌,无法达到与他人(例如明星)合唱的K歌体验。
发明内容
本发明实施例提供了一种音频处理方法、装置和系统,可以实现与他人合唱的K歌体验。
本发明实施例的第一方面提供一种音频处理方法,包括:
获取歌曲的音频文件和所述歌曲的歌词文件;
在所述歌词文件的第一歌词部分对应的显示时间播放所述音频文件并录 制用户的音频数据;
在所述歌词文件的第二歌词部分对应的显示时间播放所述音频文件;
对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。
本发明实施例第二方面提供了一种音频处理装置,包括至少一个处理器和存储器,所述存储器存储有程序指令,所述程序指令当由处理器执行时,配置所述音频处理装置执行根据本发明任意实施例的方法。
本发明实施例第三方面提供了一种音频处理系统,包括服务器和至少一个客户端,所述服务器和/或所述客户端被配置为执行根据本发明任意实施例的方法。
本发明实施例第四方面提供了一种音频处理装置,包括:
第一获取单元,用于获取歌曲的音频文件以及所述歌曲的歌词文件;
第一处理单元,用于在所述歌词文件的第一歌词部分对应的显示时间和第二歌词部分对应的显示时间播放所述音频文件,并在所述歌词文件的第一歌词部分对应的显示时间录制用户的音频数据;
第一混音单元,用于对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。
相对于现有技术,本发明实施例中在第一歌词部分对应的显示时间播放音频文件(例如伴奏音频文件)和录制用户的音频数据,在第二歌词部分对应的显示时间播放音频文件(例如原唱音频文件),由此,可以实现歌曲在第一歌词部分的显示时间由用户演唱,在第二歌词部分的显示时间由他人或者明星原唱,从而实现与他人或者明星合唱的K歌体验。
附图说明
图1为根据本发明一个示例实施例的音频处理方法示意图;
图2为根据本发明另一示例实施例的音频处理方法示意图;
图3为根据本发明又一示例实施例的音频处理方法示意图;
图4为根据本发明又一示例实施例的音频处理方法示意图;
图5为根据本发明又一示例实施例的音频处理方法示意图;
图6为根据本发明一个示例实施例的音频处理装置示意图;
图7为根据本发明另一示例实施例的音频处理装置示意图;
图8为根据本发明一个示例实施例的音频处理装置示意图;
图9为根据本发明一个示例实施例的音频处理系统示意图;
图10为根据本发明另一示例实施例的音频处理系统示意图。
具体实施方式
本发明实施例提供了一种音频处理方法、装置以及系统,可以实现与他人(例如明星)合唱的K歌体验,以下分别进行详细说明。
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。
本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”、“第三”“第四”等(如果存在)是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的实施例能够以除了在这里图示或描述的内容以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。
本发明实施例提供一种音频处理方法,包括:获取歌曲的音频文件和所述歌曲的歌词文件;在所述歌词文件的第一歌词部分对应的显示时间播放所述音频文件并录制用户的音频数据;在所述歌词文件的第二歌词部分对应的显示时间播放所述音频文件;对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。可选地,可以在所述歌词文件的第二歌词部分对应的显示时间播放所述音频文件,但不录制用户的音频数据。或者,在所述歌词文件的第二歌词部分对应的显示时间播放 所述音频文件并录制用户的音频数据,但在混音处理或生成合唱文件时不使用或去除在所述歌词文件的第二歌词部分对应的显示时间录制的所述用户的音频数据。本发明实施例的音频处理方法可以由任意计算装置执行,所述计算装置包括但不限于:手机、个人计算机(PC,Personal Computer)和/或平板电脑等。本发明实施例的音频处理方法也可以由包括服务器和客户端的系统中的服务器或者客户端执行或者由服务器与客户端协作执行。
本发明实施例的音频处理方法的一个例子是K歌处理方法,本领域的技术人员可以理解音频处理方法可以有其它应用。此外,歌词文件可以由其它任意文本文件或者图片代替,歌曲的音频可以由其他音频代替,而不会脱离本发明的原理。
需要说明的是,客户端具体可以包括手机、个人计算机(PC,Personal Computer)和/或平板电脑等设备。具体实现时,可以通过这些客户端登录网页,或者在这些客户端上加载程序,或者使用各移动平台发布的应用程序(App,Application)来访问K歌服务。
请参阅图1,图1示出了根据本发明一个示例实施例的音频处理方法。在该实施例中以客户端作为音频处理方法的执行主体进行描述,本领域的技术人员可以理解该实施例的方法可以由其它主体执行。
101、客户端接收用户的请求。
例如,当用户想选取歌曲进行K歌时,需向客户端输入K歌请求,以触发K歌处理流程。
102、客户端根据该请求获取歌曲的音频文件以及对应的歌词文件。歌曲的音频文件例如可以包括歌曲的伴奏音频文件以及歌曲的原唱音频文件等。由此,客户端获取歌曲的音频文件可以包括客户端获取歌曲的伴奏音频文件以及歌曲的原唱音频文件等。
在接收到用户输入的K歌请求时,客户端可以根据该K歌请求确定用户选取的歌曲,并进一步获取该歌曲的伴奏音频文件、原唱音频文件以及对应的歌词文件。客户端例如可以从服务器获取歌曲的伴奏音频文件、原唱音频文件以及对应的歌词文件。
伴奏音频文件和原唱音频文件例如可以是裸音频数据,在实际应用过程 中,如果获取的伴奏音频文件或原唱音频文件为压缩格式的音频数据,如mp3格式,则服务器先将该压缩格式的音频数据解压缩为裸音频数据。
需要说明的是,歌词文件例如可以包括用于标识歌词显示时间的时间信息,以便歌词文件根据该时间信息与伴奏音频文件、原唱音频文件同步。例如该歌词文件可以包括LRC歌词文件,该LRC歌词文件通过其包含的时间标签来标识歌词显示时间。
103、客户端在歌词文件的第一歌词部分对应的显示时间和第二歌词部分对应的显示时间播放音频文件,同时在歌词文件的第一歌词部分对应的显示时间录制用户的音频数据。
根据本发明的一个实施例在歌词文件的第二歌词部分对应的显示时间播放音频文件,但不录制用户的音频数据,例如关上用于录制用户的音频数据的麦克风。根据本发明的另一实施例在歌词文件的第二歌词部分对应的显示时间播放音频文件同时也录制用户的音频数据,但是在下文要描述的混音处理或生成合唱文件时去除(或者不使用)在歌词文件的第二歌词部分对应的显示时间录制的用户的音频数据。
根据本发明的实施例,在获取的歌曲的音频文件包括歌曲的伴奏音频文件以及歌曲的原唱音频文件时,在歌词文件的第一歌词部分对应的显示时间播放伴奏音频文件,在歌词文件的第二歌词部分对应的显示时间播放原唱音频文件。
可选地播放音频文件和录制用户的音频数据之前,在客户端对该歌词文件进行划分处理,以得到第一歌词部分和第二歌词部分。
客户端可以采用多种实施方式对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,下面分别举例说明:
第一种是,在步骤103之前,客户端接收用户的划分指令,然后根据该划分指令对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
在本实施例中,可以在客户端预置多种歌词划分处理策略并通过上述歌词划分处理策略将歌词文件划分为两个歌词部分,下面进行举例:
(1)、客户端根据歌词文件中的时间信息确定该歌词文件的段落信息,再根据该段落信息将该歌词文件划分为两个歌词部分。
在本实施例中,该歌词文件的时间信息可以反映相应歌曲各个段落显示时间的起点和终点,由此,客户端可以根据该时间信息确定歌词文件包括几个段落且每个段落对应的显示时间的起点和终点。再根据上述段落信息对歌词文件进行划分时,服务器可以将预设的至少一个段落划为一个歌词部分,则其他部分划为另一个歌词部分。具体此处对每个歌词部分的段落数量以及具体为哪个或哪几个段落不设限定。例如,段落信息的结构体可以表示为:
typdef struct section_info{
int section_id;//段落id
int start_ms;//开始时间ms
int end_ms;//结束时间ms
int role_id;//角色id
}Ssection,*PSection;
其中,上述角色id用于表示该歌曲对应为女声演唱部分还是男声演唱部分,作为优选,可以将女声演唱部分对应的段落设置为一个歌词部分,将男声演唱部分对应的段落设置为另一个歌词部分,以便利用男女角色的区别来划分歌词,得到不同角色的两个歌词部分。
(2)、客户端确定该歌词文件的主歌部分和副歌部分,并将主歌部分确定为一个歌词部分,将副歌部分确定为另一个歌词部分。
在本实施例中,客户端可以利用主歌和副歌的区别来划分歌词,以便得到对应主歌的一个歌词部分,对应副歌的另一个歌词部分,从而通过主歌和副歌的区别将歌词文件划分为两个歌词部分。
(3)、客户端根据自定义的模板对该歌词文件进行划分处理,得到两个歌词部分;
在本实施例中,客户端可以预先自定义多个划分歌词的模板,再根据上述自定义的模板对该歌词文件进行划分处理,得到两个歌词部分。例如,将固定的N句歌词或者N个字设置为一个歌词部分,具体此处对于如何设置自定义的模板不做限定,本领域技术人员可以根据实际情况设置多种不同模板。
需要说明的是,客户端在接收到用户的划分指令时,根据该划分指令采用相应的歌词划分处理策略进行划分处理,从而得到对应的第一歌词部分和 第二歌词部分。可选地,在实际应用过程中,客户端还可以提供一选取模块,供用户对划分得到的两个歌词部分进行选取,以指定哪一部分作为第一歌词部分或者哪一步部分作为第二部分。例如,通过选取模块,用户可以选择将男声演唱部分作为第一歌词部分,以供自己演唱,或者将男声演唱部分作为第二歌词部分,以由他人演唱或由明星原唱。
在客户端对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分之后,客户端可以在第一歌词部分对应的显示时间播放伴奏音频文件和录制用户的音频数据,在第二歌词部分的显示时间播放原唱音频文件。
第二种是,客户端根据是否有声音输入来对歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
具体包括:客户端检测是否有声音输入,若是,则将歌词文件在对应显示时间内的歌词部分划分为第一歌词部分,并播放伴奏音频文件和切换到麦克风输入模式以录制用户的音频数据,若否,则将歌词文件在对应显示时间内的歌词部分划分为第二歌词部分,并播放原唱音频文件。可以理解的是,此处歌词文件的划分处理操作和播放、录制等操作同步进行,而无需等得到第一歌词部分和第二歌词部分之后,客户端再在第一歌词部分对应的显示时间播放伴奏音频文件和录制用户的音频数据,在第二歌词部分的显示时间播放原唱音频文件。
在本实施例中,上面仅以几个例子说明了如何对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,在实际应用过程中,客户端可以对上述划分处理方式进行结合使用,还可以采用更多其他的划分处理方式,具体此处不做限定。
为方便用户快速识别哪部分为自己需要演唱的歌词,哪部分为他人或者明星演唱的歌词,客户端可以进一步对显示在客户端上的歌词进行标识,由此,作为优选,本实施例还可以进一步包括:客户端对显示在客户端的第一歌词部分设置第一标识,对显示在客户端的第二歌词部分设置第二标识。可以理解的是,上述标识可以为颜色标识,例如,第一歌词部分显示一种颜色,第二歌词部分显示另一个颜色;或者上述标识为高亮状态标识,例如,第一歌词部分显示为高亮状态,第二歌词部分显示为非高亮状态,具体此处对采 用标识的具体实现形式不做限定,本领域技术人员可以根据需求获取更多其他的标识形式。
104、客户端对该用户的音频数据和音频文件在第一歌词部分对应的显示时间的音频数据进行混音处理,得到混音音频文件。
根据本发明的实施例,在得到用户的音频数据(例如麦克风采集的音频数据)后,客户端对该用户的音频数据和音频文件在第一歌词部分对应的显示时间的音频数据进行混音处理,得到混音音频文件,由此,该混音音频文件可以体现用户演唱部分的全部音频数据,包括用户的声音以及对应的伴奏。
混音处理的具体过程可参见现有的混音算法,例如,直接将对应采样点数据相加除以二。
在音频文件包括伴奏音频文件以及原唱音频文件的条件下,客户端对该用户的音频数据和伴奏音频文件在第一歌词部分对应的显示时间的音频数据进行混音处理。
根据本发明的可选实施例,客户端对该用户的音频数据和音频文件在第一歌词部分对应的显示时间的音频数据进行混音处理之后,在第一歌词部分对应的显示时间输出(或选取)混音音频文件,在第二歌词部分对应的显示时间输出(或选取)音频文件,以生成合唱文件。而在音频文件包括伴奏音频文件以及原唱音频文件的条件下,为生成合唱文件,在第二歌词部分对应的显示时间输出(或选取)的是原唱音频文件。
在生成合唱文件的过程中,若为第一歌词部分对应的显示时间,则获取混音音频文件在对应时间段的音频数据,若为第二歌词部分对应的显示时间,则获取该原唱音频文件在对应时间段的音频数据,再按照时间先后顺序将各个时间点的音频数据合成一个音频文件,合成的该音频文件即为合唱文件。需要强调的是,在上述生成合唱文件的过程中,由歌词的显示时间确定具体输出目标音频文件和原唱音频文件在哪个时间段的音频数据。由此,得到的合唱文件在第一歌词部分对应的显示时间,输出的为相应时间段的用户演唱部分,在第二歌词部分对应的显示时间,输出的为相应时间段的原唱部分。
可选地,客户端对上述合唱文件在切换点的音频数据进行平衡处理,其中,该切换点指由第一歌词部分对应的显示时间切换至第二歌词部分对应的 显示时间的时间点或者由第二歌词部分对应的显示时间切换至第一歌词部分对应的显示时间的时间点。此处,由于混音音频文件由两个音频文件拼接而成,在拼接在切换点进行平衡处理可以保证混音音频文件的音频质量。在本实施例中,具体的平衡处理方式可参见现有的音频拼接技术,比如均值滤波、叠加处理等。
可选地,在本实施例中,用户还可以通过客户端将该合唱文件上传到网络,以便与其他人分享自己与他人或者明星的合唱。
上文描述了客户端获取的歌曲的音频文件可以包括歌曲的伴奏音频文件以及歌曲的原唱音频文件。可替代地,客户端获取的歌曲的音频文件可以是下文将要描述的目标音频文件,因此,客户端获取歌曲的音频文件例如可以包括客户端接收服务器发送的目标音频文件。此外,客户端获取歌曲的歌词文件例如可以包括客户端接收服务器发送的目标歌词文件。关于服务器发送目标音频文件和/或目标歌词文件的具体过程将在下文进行详细描述。在客户端接收服务器发送的目标音频文件和/或目标歌词文件的情况下,在上述步骤103中,客户端可以不对歌词文件进行划分处理。
相对于现有技术,本发明实施例中仅在第一歌词部分对应的显示时间播放伴奏音频文件和录制用户的音频数据,而在第二歌词部分对应的显示时间播放原唱音频文件,由此,可以实现歌曲在第一歌词部分的显示时间由用户演唱,在第二歌词部分的显示时间由他人或者明星原唱,从而实现与他人或者明星合唱的K歌体验。而且,在本实施例可以直接在客户端利用普通的伴奏音频文件便可实现与他人或者明星合唱的K歌体验。
请参阅图2,图2示出根据本发明另一实施例的音频处理方法。在该实施例中以服务器作为音频处理方法的执行主体进行描述,本领域的技术人员可以理解该实施例的方法可以由其它主体执行。
201、服务器获取歌曲的伴奏音频文件、原唱音频文件以及对应的歌词文件。关于伴奏音频文件、原唱音频文件以及对应的歌词文件在上文已有详细描述,为了简明这里不再重复。
202、服务器对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,并对应生成歌词划分信息。
服务器可以采用多种实施方式对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,下面分别举例说明第一种是:服务器根据该时间信息确定该歌词文件的段落信息,再根据该段落信息对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
第二种是:服务器确定该歌词文件的主歌部分和副歌部分,并将主歌部分确定为第一歌词部分,将副歌部分确定为第二歌词部分,或者将副歌部分确定为第一歌词部分,将主歌部分确定为第二歌词部分。
第三种是:服务器根据自定义的模板对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
需要说明的是,上面描述了如何对该歌词文件进行划分处理的几个例子,为了简明,这里对这些例子没有详细说明,这些例子的详细描述可以参照上文步骤中103的相关描述。
可选地,在服务器执行步骤202之前,还可以包括:接收客户端发送的划分指令,该划分指令用于指示用户指定的划分处理方式,则服务器将根据该划分指令对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,由此,可以实现让用户选择如何划分歌词,并得到符合用户意愿的第一歌词部分和第二歌词部分。
在本实施例中,在经划分处理得到第一歌词部分和第二歌词部分后,服务器还对相应的划分处理结果进行记录,以对应生成歌词划分信息,其中,该歌词划分信息包括用于标识该第一歌词部分的标识信息。此处,由于歌词总共由两部分构成,在具有标识该第一歌词部分的标识信息后,服务器或者客户端等都可以识别对应的第二歌词部分。当然,在其他一些实施例中,该歌词划分信息还可以包括用于标识该第二歌词部分的标识信息。
203、服务器在第一歌词部分对应的显示时间输出(或选取)伴奏音频文件,在第二歌词部分的显示时间输出(或选取)原唱音频文件,以生成目标音频文件。
在生成目标音频文件的过程中,若为第一歌词部分对应的显示时间,则获取伴奏音频文件在对应时间段的音频数据,若为第二歌词部分对应的显示时间,则获取该原唱音频文件在对应时间段的音频数据,再按照时间先后顺 序将各个时间点的音频数据合成一个音频文件,合成的该音频文件即为目标音频文件。需要强调的是,在上述生成目标音频文件的过程中,由歌词的显示时间确定具体输出伴奏音频文件和原唱音频文件在哪个时间段的音频数据。由此,在播放目标音频文件时,在第一歌词部分对应的显示时间,输出的为相应时间段的伴奏,在第二歌词部分对应的显示时间,输出的为相应时间段的原唱,即目标音频文件能够实现在伴奏与原唱之间自动切换。
可选地,所述方法还包括:服务器对该目标音频文件在切换点的音频数据进行平衡处理,其中,该切换点指由第一歌词部分对应的显示时间切换至第二歌词部分对应的显示时间的时间点或者由第二歌词部分对应的显示时间切换至第一歌词部分对应的显示时间的时间点。此处,由于目标音频文件由两个音频文件拼接而成,在拼接在切换点进行平衡处理可以保证目标音频文件的音频质量。在本实施例中,具体的平衡处理方式可参见现有的音频拼接技术,比如均值滤波、叠加处理等。
204、服务器向客户端发送目标音频文件和目标歌词文件。
在接收到客户端发送的请求(例如K歌请求)时,服务器向客户端发送目标音频文件和目标歌词文件,以便客户端以目标音频文件作为伴奏音频、以目标歌词文件作为待播放歌词向用户提供K歌服务。其中,该目标歌词文件通过在上述歌词文件中携带该歌词划分信息得到。在本实施例中,客户端在播放目标音频数据时,在第一歌词部分的显示时间,输出的为相应时间段的伴奏,在第二歌词部分对应的显示时间,输出的为相应时间段的原唱,由此,可以实现在第一歌词部分的显示时间由用户演唱,在第二歌词部分的显示时间由他人演唱(或由明星原唱),从而实现与他人或明星合唱的K歌体验。可以理解的是,本实施例中仅通过服务器生成目标音频文件和目标歌词文件、并向客户端上述目标音频文件和目标歌词文件,便可实现与他人或明星合唱的K歌体验,此过程中无需改变客户端逻辑。
相对于现有技术,本发明实施例中服务器向客户端发送的目标音频文件在第一歌词部分的显示时间对应为伴奏,在第二歌词部分的显示时间对应为原唱,与现有技术中直接向客户端提供伴奏音频文件相比,本发明实施例可以实现在第一歌词部分的显示时间由用户演唱,在第二歌词部分的显示时间 由他人或者明星原唱,从而实现与他人或者明星合唱的K歌体验。
请参阅图3,图3示出根据本发明另一示例实施例的音频处理方法,包括:
301、服务器获取歌曲的伴奏音频文件、原唱音频文件以及对应的歌词文件;
302、服务器对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分;
303、服务器在第一歌词部分对应的显示时间输出伴奏音频文件,在第二歌词部分的显示时间输出原唱音频文件,以生成目标音频文件;
304、服务器向客户端发送目标音频文件和目标歌词文件;
步骤301至304的过程与图2所示实施例中的步骤201至步骤204的过程相同,此处不再赘述。
305、客户端播放目标音频文件和目标歌词文件。
客户端在接收到服务器发送的目标音频文件和目标歌词文件后,客户端的K歌应用播放上述目标音频文件和目标歌词文件,其中,该目标音频文件作为伴奏音频进行播放。
306、客户端根据歌词划分信息在第一歌词部分对应的显示时间录制用户的音频数据。
在本实施例中,客户端根据歌词划分信息可以识别第一歌词部分对应的显示时间以及第二歌词部分对应的显示时间,并在第一歌词部分对应的显示时间录制用户的音频数据,在第二歌词部分对应的显示时间停止录制,此时客户端播放的目标文件对应他人或者明星原唱,由此,可实现在第一歌词部分的显示时间由用户演唱,在第二歌词部分的显示时间由他人或者明星原唱,从而实现与他人或者明星合唱的K歌体验。
307、客户端对用户的音频数据和目标音频文件进行混音处理,得到合唱文件。
在本实施例中,得到的合唱文件在第一歌词部分的显示时间为用户演唱的音频数据,在第二歌词部分的显示时间为他人或者明星原唱的音频数据(即该原唱音频文件在对应时间段的音频数据)。可选地,在本实施例中,用户还可以通过客户端将该合唱文件上传到网络,以便与其他人分享自己与他人或 者明星的合唱。
请参阅图4,图4示出根据本发明又一实施例的音频处理方法,包括:
401、第一客户端接收第一用户的第一请求。
该第一请求例如是用于请求发起合唱的第一合唱请求。例如,在实际应用过程中,第一客户端可以设置对应于该第一合唱请求的按钮,在第一用户按下该按钮时,第一客户端接收第一合唱请求。
402、第一客户端根据该第一请求从服务器获取歌曲的伴奏音频文件和歌词文件,以便第一用户利用伴奏音频文件和歌词文件基于部分歌词进行录音,得到第一录制文件。关于伴奏音频文件、原唱音频文件以及对应的歌词文件在上文已有详细描述,为了简明这里不再重复。
在获取伴奏音频文件和歌词文件后,客户端播放上述伴奏音频文件和歌词文件,第一用户可以选择任意的歌词进行演唱,当然,第一用户不对全部歌词进行演唱,同时客户端对该第一用户利用上述伴奏音频文件和歌词文件基于部分歌词进行的演唱进行录音,以得到第一录制文件。
403、第一客户端基于所述第一录制文件确定第二歌词部分,并对应生成段落信息。
该第二歌词部分是指第一用户演唱过的歌词部分,即第一客户端基于第一录制文件识别出哪些歌词部分为第一用户演唱过的歌词部分,并对应生成段落信息,该段落信息包括用于标识该第二歌词部分的标识信息,以便对第一用户演唱过的歌词部分进行记载。
可选地,标识信息可以包括第二歌词部分所对应的显示时间,例如,第二歌词部分的开始时间是10S,结束时间是15S等,或者标识信息可以包括第二歌词部分的句子标识,例如,第二歌词部分为第1、3、5句等,具体此处对标识信息的具体形式不做限定,只要第一客户端能根据该标识识别出第二歌词部分即可。
第一客户端可以采用多种实施方式来确定第二歌词部分,例如,第一客户端可以对第一录制文件中声音的有效性来确定哪些歌词部分为第一用户演唱过的歌词部分,具体包括:
第一客户端判断第一用户在某一歌词部分录入的声音是否有效,若是, 则确定对应的歌词部分为第二歌词部分。
可选地,第一客户端判断声音的音量是否大于第一预设值,若是,则确定第一用户在某一歌词部分录入的声音有效。具体此处可以参考VAD算法或者其改进算法。
作为优选,在确定第一用户在某一歌词部分录入的声音有效后,第一客户端进一步判断录入的声音所占用的时间与该歌词部分所占用的显示时间是否对应,若是,则确定该歌词部分为第二歌词部分。
例如,此处以歌词部分中的句子为单位,在确定在某一句歌词录入的声音有效后,如果在该句歌词录入的声音所占用的显示时间为20S-30S,而在该句歌词所占用的显示时间也为20至30,则可以确定该句歌词为第二歌词部分,具体此处可以设置一个阈值,当该句歌词所占用的显示时间与该句歌词所占用的显示时间的重合比例大于该阈值时,确定满足该句歌词所占用的显示时间与该句歌词所占用的显示时间对应。
可以理解的是,在本实施例中,还可以采用更多的实施方式来确定第二歌词部分,具体此处不做限定。在实际应用过程中,第一客户端可以引入包含上述确定方式的音频实时检测模块来确定第二歌词部分。
404、第一客户端对所述第一录制文件和所述伴奏音频文件进行混音处理,得到目标音频文件。
混音处理的具体过程可参见现有的混音算法,例如,直接将对应采样点数据相加除以二。
可选地,为防止第一录制文件在第二歌词部分以外的歌词部分还包括一些噪声或者被判断为无效的声音,则在执行步骤404之前还可以包括:
第一客户端根据段落信息对第一录制文件在第二歌词部分以外的歌词部分所对应的音频数据进行清零,由此,可以确保得到的第一录制文件仅在第二歌词部分的显示时间包含音频数据,相当于仅包含第一用户演唱的有效声音。
405、第一客户端向服务器发送目标音频文件和段落信息,以便第二客户端从服务器获取该目标音频文件和段落信息,并以该目标音频文件作为伴奏音频、根据段落信息向第二用户提供K歌服务。
第二客户端在从服务器获取该目标音频文件和段落信息,可以以该目标音频文件作为伴奏音频、根据段落信息向第二用户提供K歌服务。具体地,第二用户根据该段落信息可以确定哪些歌词部分已经被演唱,并在没有被演唱过的歌词部分进行演唱。
可选地,在本实施例还可以包括:该第一客户端生成配置文件,该配置文件包括第一用户所采用的音效参数;第一客户端向服务器发送该配置文件,以便第二客户端以该目标音频文件作为伴奏音频、根据段落信息和配置文件向第二用户提供K歌服务。可选地,上述音效参数包括混响参数、变声变调参数、声音美化参数等。
相对于现有技术,本发明实施例中的目标音频文件仅在第二歌词部分的显示时间由第一用户演唱,而在第二歌词部分以外的歌词部分为伴奏,由此,当第二客户端的以该目标音频文件作为伴奏音频时,可以根据段落信息在第二歌词部分以外的歌词部分进行演唱,以实现在第二歌词部分的显示时间由第一用户演唱,在第二歌词部分以外的歌词部分的显示时间由第二用户演唱,从而实现与他人合唱的K歌体验。
请参阅图5,图5示出本发明实施例中音频处理方法另一实施例,包括:
501、第一客户端接收第一用户的第一请求;
502、第一客户端根据该第一请求从服务器获取歌曲的伴奏音频文件和歌词文件,以便第一用户利用伴奏音频文件和歌词文件基于部分歌词进行录音,得到第一录制文件;
503、第一客户端基于所述第一录制文件确定第二歌词部分,并对应生成段落信息;
504、第一客户端对所述第一录制文件和所述伴奏音频文件进行混音处理,得到目标音频文件;
505、第一客户端向服务器发送目标音频文件和段落信息,以便第二客户端从服务器获取该目标音频文件和段落信息,并以该目标音频文件作为伴奏音频、根据段落信息向第二用户提供K歌服务。
步骤501至505的过程与图4所示实施例中的步骤401至步骤405的过程相同,此处不再赘述。
506、第二客户端接收第二用户发送的第二请求。
该第二请求用于请求加入合唱,以触发合唱加入流程。例如,在实际应用过程中,第二客户端可以设置对应于该第二合唱请求的按钮,在第二用户按下该按钮时,第二客户端接收第二合唱请求。
507、第二客户端根据该第二请求从服务器获取歌曲的目标音频文件、歌词文件以及段落信息,以便第二用户以目标音频文件作为伴奏音频并根据段落信息基于第二歌词部分以外的全部或部分歌词进行录音,得到第二录制文件。
在接收到第二请求时,第二客户端首先从服务器获取目标音频文件、歌词文件以及段落信息,然后播放上述目标音频文件和歌词文件,第二用户基于第二歌词部分以外的全部或部分歌词进行演唱,同时第二客户端对该第二用户以目标音频文件作为伴奏音频并根据段落信息基于第二歌词部分以外的全部或部分歌词进行的演唱进行录音,以得到第二录制文件。需要强调的是,如果第二用户基于第二歌词部分以外的部分歌词部分进行演唱,而不是基于第二歌词部分以外的全部歌词部分进行演唱,该歌曲还有一部分歌词可以由更多其他的用户加入合唱。
可选地,当第一客户端向服务器发送过配置文件,在服务器上还包括该配置文件时,步骤507具体可以包括:第二客户端从服务器获取歌曲的目标音频文件、歌词文件、段落信息以及配置文件,以便第二用户以目标音频文件作为伴奏音频并根据段落信息和配置文件基于第二歌词部分以外的部分或全部歌词部分进行录音,得到第二录制文件。在实际应用过程中,客户端根据配置文件中的音效参数进行相应配置,并下发给第二用户供该第二用户选择,用户也可以根据实际需要修改相应的参数配置。
508、第二客户端对第二录制文件和目标音频文件进行混音处理,得到混音音频文件。
混音处理的具体过程可参见现有的混音算法,例如,直接将对应采样点数据相加除以二。
可以理解的是,混音音频文件在第二歌词部分由第一用户演唱,在第二歌词部分以外的部分或全部歌词部分由第二用户演唱,由此,可以实现两个 用户合唱的K歌体验。
可选地,在本实施例中,为防止第二录制文件在第二歌词部分以外的歌词部分还包括一些噪声或者被判断为无效的声音,则在执行步骤508之前还可以包括:第二客户端根据段落信息对第一录制文件在第二歌词部分所对应的音频数据进行清零。
上面对本发明实施例中的音频处理方法进行了描述,下面对本发明实施例中的音频处理装置进行描述,请参阅图6,图6是根据本发明示例实施例的音频处理装置,其包括第一接收单元601、第一获取单元602、第一处理单元603和第一混音单元604。
第一接收单元601用于接收用户的请求。
第一获取单元602用于根据所述请求获取歌曲的音频文件以及对应的歌词文件。其中,所述歌词文件包括用于标识歌词显示时间的时间信息。歌曲的音频文件例如可以包括歌曲的伴奏音频文件以及歌曲的原唱音频文件等。
第一处理单元603用于在所述第一歌词部分对应的显示时间和第二歌词部分对应的显示时间播放所述音频文件,同时在歌词文件的第一歌词部分对应的显示时间录制用户的音频数据。第一处理单元603,例如还可以用于对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
在获取的歌曲的音频文件包括歌曲的伴奏音频文件以及歌曲的原唱音频文件时,第一处理单元603在所述第一歌词部分对应的显示时间播放所述伴奏音频文件和录制用户的音频数据,在第二歌词部分的显示时间播放所述原唱音频文件。
第一混音单元604用于对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理,得到混音音频文件。在音频文件包括伴奏音频文件以及原唱音频文件的条件下,第一混音单元604对该用户的音频数据和伴奏音频文件在第一歌词部分对应的显示时间的音频数据进行混音处理。
为便于理解,下面以一具体应用场景为例,对本实施例中的音频处理装置的内部运作流程进行描述:
第一接收单元601接收用户输入的K歌请求;第一获取单元602根据所 述K歌请求获取歌曲的伴奏音频文件、原唱音频文件以及对应的歌词文件,其中,所述歌词文件包括用于标识歌词显示时间的时间信息;第一处理单元603对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,并在所述第一歌词部分对应的显示时间播放所述伴奏音频文件和录制用户的音频数据,在第二歌词部分的显示时间播放所述原唱音频文件。
本发明实施例提供的技术方案中,客户端的第一获取单元602在通过第一接收单元601接收到用户输入的K歌请求时,根据该K歌请求获取歌曲的伴奏音频文件、原唱音频文件以及对应的歌词文件,其中,该歌词文件包括用于标识歌词显示时间的时间信息;然后由第一处理单元603对该歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,并在第一歌词部分对应的显示时间播放伴奏音频文件和录制用户的音频数据,在第二歌词部分的显示时间播放原唱音频文件;再对上述经录制得到的用户的音频数据和伴奏音频文件在第一歌词部分对应的显示时间的音频数据进行混音处理,得到混音音频文件,并在第一歌词部分对应的显示时间输出混音音频文件,在第二歌词部分对应的显示时间输出原唱音频文件,以生成合唱文件。因此相对于现有技术,本发明实施例中客户端在对歌词文件进行划分处理,得到第一歌词部分和第二歌词部分后,仅在第一歌词部分对应的显示时间播放伴奏音频文件和录制用户的音频数据,而在第二歌词部分对应的显示时间播放原唱音频文件,由此,可以实现一部分在第一歌词部分的显示时间由用户演唱,在第二歌词部分的显示时间由明星原唱,从而实现与明星合唱的K歌体验。
可选地,在本实施例中,所述音频处理装置还可以包括:
第一生成单元605,用于在所述第一歌词部分对应的显示时间输出(或选取)所述混音音频文件,在所述第二歌词部分对应的显示时间输出(或选取)所述音频文件,以生成合唱文件。而在音频文件包括伴奏音频文件以及原唱音频文件的条件下,为生成合唱文件,在第二歌词部分对应的显示时间输出(或选取)的是原唱音频文件。
可选地,所述第一接收单元601,还用于接收所述用户的划分指令;所述第一处理单元603,具体用于根据所述划分指令对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
可选地,所述第一处理单元603,具体用于检测是否有声音输入,若是,则将所述歌词文件在对应显示时间内的歌词部分划分为第一歌词部分,并播放所述伴奏音频文件和录制用户的音频数据,若否,则将所述歌词文件在对应显示时间内的歌词部分划分为第二歌词部分,并播放所述原唱音频文件。
可选地,所述音频处理装置还可以包括:
标识单元,用于对显示在所述客户端的第一歌词部分设置第一标识,对显示在所述客户端的第二歌词部分设置第二标识。
可选地,所述音频处理装置还可以包括:
第一平衡处理单元,用于对所述合唱文件在切换点的音频数据进行平衡处理,其中,所述切换点指由所述第一歌词部分对应的显示时间切换至第二歌词部分对应的显示时间的时间点或者由所述第二歌词部分对应的显示时间切换至所述第一歌词部分对应的显示时间的时间点。
请参阅图7,图7示出根据本发明又一实施例的音频处理装置,其包括:
第二获取单元701,用于获取歌曲的伴奏音频文件、原唱音频文件以及对应的歌词文件,其中,所述歌词文件包括用于标识歌词显示时间的时间信息;
第二处理单元702,用于对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,并对应生成歌词划分信息,其中,所述歌词划分信息包括用于标识所述第一歌词部分的标识信息;
第二生成单元703,用于在第一歌词部分对应的显示时间输出(或选取)所述伴奏音频文件,在第二歌词部分的显示时间输出(或选取)所述原唱音频文件,以生成目标音频文件;
第一发送单元704,用于向所述客户端发送所述目标音频文件和所述目标歌词文件(例如,在接收到客户端发送的K歌请求时),以便所述客户端以所述目标音频文件作为伴奏音频、以所述目标歌词文件作为待播放歌词向用户提供K歌服务,其中,所述目标歌词文件通过在所述歌词文件中携带所述歌词划分信息得到。
相对于现有技术,本发明实施例服务器中的音频处理装置向客户端发送的目标音频文件在第一歌词部分的显示时间对应为伴奏,在第二歌词部分的显示时间对应为原唱,与现有技术中直接向客户端提供伴奏音频文件相比, 本发明实施例可以实现在第一歌词部分的显示时间由用户演唱,在第二歌词部分的显示时间由明星原唱,从而实现与明星合唱的K歌体验。
可选地,所述第二处理单元702,可以根据所述时间信息确定所述歌词文件的段落信息,根据所述段落信息对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
可选地,所述第二处理单元702,可以确定所述歌词文件的主歌部分和副歌部分,并将所述主歌部分确定为第一歌词部分,将所述副歌部分确定为第二歌词部分,或者将所述副歌部分确定为第一歌词部分,将所述主歌部分确定为第二歌词部分。
可选地,所述第二处理单元702,可以根据自定义的模板对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
可选地,,所述音频处理装置还可以包括:
第二平衡处理单元,用于对所述目标音频文件在切换点的音频数据进行平衡处理,其中,所述切换点指由所述第一歌词部分对应的显示时间切换至第二歌词部分对应的显示时间的时间点或者由所述第二歌词部分对应的显示时间切换至所述第一歌词部分对应的显示时间的时间点。
请参阅图8,图8示出根据本发明另一实施例的音频处理装置,包括:
801第二接收单元,用于接收第一用户的第一请求;
802第三获取单元,用于根据该第一请求从服务器获取歌曲的伴奏音频文件和歌词文件,以便第一用户利用伴奏音频文件和歌词文件基于部分歌词进行录音,得到第一录制文件;
803第三处理单元,用于基于所述第一录制文件确定第二歌词部分,并对应生成段落信息,所述段落信息至少包括用于标识所述第二歌词部分的标识信息;
804第二混音单元,用于对所述第一录制文件和所述伴奏音频文件进行混音处理,得到目标音频文件;
805第二发送单元,用于发送目标音频文件和段落信息,以便第二客户端从服务器获取该目标音频文件和段落信息,并以该目标音频文件作为伴奏音频、根据段落信息向第二用户提供K歌服务。
请参阅图9,图9示出根据本发明示例实施例的音频处理系统,其包括服务器901和客户端902。本发明上述实施例的音频处理方法可以由音频处理系统的服务器和/或客户端执行。
根据本发明的一个可选实施例,所述服务器901包括如图7所示实施例中描述的音频处理装置;
根据本发明的一个可选实施例,所述客户端902,包括如图6所示实施例中描述的音频处理装置。
请参阅图10,图10示出根据本发明另一示例实施例的音频处理系统,其包括第一客户端1001、第二客户端1002和服务器1003。
所述第一客户端1001,用于接收第一用户的第一请求,所述第一请求例如用于请求发起合唱;根据所述第一请求从服务器1003获取歌曲的伴奏音频文件和歌词文件,以便所述第一用户利用所述伴奏音频文件和歌词文件基于部分歌词进行录音,得到第一录制文件;基于所述第一录制文件确定第二歌词部分,并对应生成段落信息,其中,所述第二歌词部分指所述第一用户演唱过的歌词部分,所述段落信息包括用于标识所述第二歌词部分的标识信息;对所述第一录制文件和所述伴奏音频文件进行混音处理,得到目标音频文件;以及用于向所述服务器1003发送所述目标音频文件和段落信息,以便第二客户端从所述服务器1003获取所述目标音频文件和段落信息,并以所述目标音频文件作为伴奏音频、根据所述段落信息向第二用户提供K歌服务。
所述第二客户端1002,用于从所述服务器1003获取所述目标音频文件和段落信息,并以所述目标音频文件作为伴奏音频、根据所述段落信息向第二用户提供K歌服务。
可选地,所述第一客户端1001,具体可以用于判断所述第一用户在某一歌词部分录入的声音是否有效,若是,则确定对应的歌词部分为第二歌词部分。
可选地,所述第一客户端1001,具体可以用于判断所述声音的音量是否大于预设值,若是,则确定所述第一用户在某一歌词部分录入的声音有效。
可选地,所述第一客户端1001,具体可以用于判断判断所述第一用户在某一歌词部分录入的声音是否有效,若是,则进一步判断所述录入的声音所 占用的时间与所述歌词部分所占用的显示时间是否对应,若是,则确定所述歌词部分为第二歌词部分。
可选地,所述第一客户端1001,还可以用于在所述第一客户端对所述第一录制文件和所述伴奏音频文件进行混音处理,得到目标音频文件之前,根据所述段落信息对所述第一录制文件在所述第二歌词部分以外的歌词部分所对应的音频数据进行清零。
可选地,所述标识信息包括第二歌词部分所对应的显示时间或者所述第二歌词部分的句子标识。
可选地,所述第一客户端1001,还可以用于在所述第一客户端向所述服务器发送所述目标音频文件和段落信息之前,生成配置文件,所述配置文件包括所述第一用户所采用的音效参数;以及用于向所述服务器发送所述配置文件。
可选地,所述第二客户端1002,具体可以用于接收第二用户发送的第二请求,所述第二请求用于请求加入合唱;根据所述第二请求从所述服务器获取歌曲的目标音频文件、歌词文件以及段落信息,以便所述第二用户以所述目标音频文件作为伴奏音频并根据所述段落信息基于所述第二歌词部分以外的全部或部分歌词进行录音,得到第二录制文件;以及对所述第二录制文件和所述目标音频文件进行混音处理,得到混音音频文件。
可选地,所述第二客户端1002,还可以用于在第二客户端对所述第二录制文件和所述目标音频文件进行混音处理,得到混音音频文件之前,根据所述段落信息对所述第一录制文件在所述第二歌词部分所对应的音频数据进行清零。
可选地,所述第二客户端1002,具体可以用于当所述服务器上还包括所述配置文件时,从所述服务器获取歌曲的目标音频文件、歌词文件、段落信息以及配置文件,以便所述第二用户以所述目标音频文件作为伴奏音频并根据所述段落信息和配置文件基于所述第二歌词部分以外的部分或全部歌词部分进行录音,得到第二录制文件。
可选地,所述第二客户端,具体可以用于接收第二用户发送的第三请求,所述第三请求用于请求加入合唱且指示仅在所述第二歌词部分以外的部分歌 词进行录音;根据所述第三请求从所述服务器获取歌曲的目标音频文件、歌词文件以及段落信息,以便所述第二用户以所述目标音频文件作为伴奏音频并根据所述段落信息基于所述第二歌词部分以外的部分歌词部分进行录音,得到第三录制文件;确定第三歌词部分,并在所述段落信息中添加用于标识所述第三歌词部分的标识信息,其中,所述第三歌词部分指所述第二用户演唱过的歌词部分;对所述第三录制文件和所述目标音频文件进行混音处理,得到第三目标文件;以及向所述服务器发送所述第三目标文件和段落信息,以便第三客户端通过所述服务器供第三用户以所述第三目标文件作为伴奏音频并根据所述段落信息在所述第二歌词部分和所述第三歌词部分以外的歌词部分进行录音。
在本发明一个可选实施例中,图10中第二客户端1002用于执行本发明参照图1描述的示例实施例的音频处理方法,所述服务器1003用于执行本发明参照图2描述的示例实施例的音频处理方法,所述第一客户端1001用于执行本发明参照图4描述的示例实施例的音频处理方法,
此外,本发明一个示例实施例提供一种音频处理装置,包括至少一个处理器和存储器,所述存储器存储有程序指令,所述程序指令当由处理器执行时,配置所述音频处理装置执行根据上述任一实施例所述的方法。
此外,本发明一个示例实施例提供一种计算机可读存储介质,存储有程序指令,所述程序指令当由计算装置的处理器执行时,配置所述计算装置执行根据上述任一实施例所述的方法。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统,装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请所提供的几个实施例中,应该理解到,所揭露的系统,装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合 或通信连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本发明各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。
以上所述,以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。

Claims (31)

  1. 一种音频处理方法,包括:
    获取歌曲的音频文件和所述歌曲的歌词文件;
    在所述歌词文件的第一歌词部分对应的显示时间播放所述音频文件并录制用户的音频数据;
    在所述歌词文件的第二歌词部分对应的显示时间播放所述音频文件;
    对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。
  2. 如权利要求1所述的方法,其中,在所述歌词文件的第二歌词部分对应的显示时间不录制用户的音频数据。
  3. 如权利要求1所述的方法,
    其中,获取歌曲的音频文件和所述歌曲的歌词文件之前,所述方法还包括:接收用户的请求,
    其中,获取歌曲的音频文件和所述歌曲的歌词文件包括:根据所述请求获取歌曲的伴奏音频文件、原唱音频文件以及所述歌曲的歌词文件。
  4. 如权利要求3所述的方法,
    其中,获取歌曲的音频文件和所述歌曲的歌词文件之后,所述方法还包括:对所述歌词文件进行划分处理,以得到所述第一歌词部分和所述第二歌词部分,
    其中,在所述歌词文件的第一歌词部分对应的显示时间播放所述音频文件并录制用户的音频数据包括:在所述第一歌词部分对应的显示时间播放所述伴奏音频文件并录制用户的音频数据,
    其中,在所述歌词文件的第二歌词部分对应的显示时间播放所述音频文件包括:在所述第二歌词部分对应的显示时间播放所述原唱音频文件。
  5. 如权利要求3所述的方法,其中,对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理包括:对所述用户的音频数据和所述伴奏音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。
  6. 如权利要求3所述的方法,其中所述混音处理生成混音音频文件,所述方法还包括:
    根据第一歌词部分对应的显示时间的所述混音音频文件,以及第二歌词部分对应的显示时间的所述原唱音频文件,生成合唱文件。
  7. 如权利要求1所述的方法,还包括:
    接收所述用户的划分指令;
    根据所述划分指令对所述歌词文件进行划分处理,以得到第一歌词部分和第二歌词部分。
  8. 如权利要求1所述的方法,还包括:
    检测是否有声音输入,若是,则将所述歌词文件在对应显示时间内的歌词部分划分为第一歌词部分,若否,则将所述歌词文件在对应显示时间内的歌词部分划分为第二歌词部分。
  9. 如权利要求1所述的方法,其中所述混音处理生成混音音频文件,所述方法还包括:
    根据第一歌词部分对应的显示时间的所述混音音频文件,以及第二歌词部分对应的显示时间的所述歌曲的所述音频文件,生成合唱文件。
  10. 如权利要求9所述的方法,还包括:
    对所述合唱文件在切换点的音频数据进行平衡处理,其中,所述切换点指由所述第一歌词部分对应的显示时间切换至第二歌词部分对应的显示时间的时间点或者由所述第二歌词部分对应的显示时间切换至所述第一歌词部分对应的显示时间的时间点。
  11. 如权利要求1所述的方法,其中在获取歌曲的音频文件和所述歌曲的歌词文件之前,还包括:
    获取歌曲的伴奏音频文件、原唱音频文件以及所述歌曲的歌词文件;
    对所述歌词文件进行划分处理,以得到第一歌词部分和第二歌词部分,并对应生成歌词划分信息,其中,所述歌词划分信息至少包括用于标识所述第一歌词部分的标识信息;
    根据第一歌词部分对应的显示时间的伴奏音频文件,以及第二歌词部分对应的显示时间的原唱音频文件,生成所述歌曲的所述音频文件。
  12. 如权利要求11所述的方法,
    其中,获取歌曲的音频文件和所述歌曲的歌词文件还包括获取所述歌词划分信息,
    其中,在所述歌词文件的第一歌词部分对应的显示时间播放所述音频文件并录制用户的音频数据包括:根据所述歌词划分信息在所述第一歌词部分对应的显示时间录制用户的音频数据。
  13. 如权利要求1所述的方法,其中,所述歌词文件包括用于标识歌词显示时间的时间信息,所述方法还包括:
    根据所述时间信息确定所述歌词文件的段落信息;
    根据所述段落信息对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
  14. 如权利要求1所述的方法,还包括:
    确定所述歌词文件的主歌部分和副歌部分,并将所述主歌部分确定为第一歌词部分,将所述副歌部分确定为第二歌词部分,或者将所述副歌部分确定为第一歌词部分,将所述主歌部分确定为第二歌词部分。
  15. 如权利要求1所述的方法,还包括:
    根据自定义的模板对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分。
  16. 如权利要求1所述的方法,其中在获取歌曲的音频文件和所述歌曲的歌词文件之前,还包括:
    接收另一用户的请求;
    根据所述请求获取所述歌曲的伴奏音频文件和歌词文件;
    利用所述伴奏音频文件和歌词文件基于部分歌词对所述另一用户进行录音,得到第一录制文件;
    基于所述第一录制文件确定所述第二歌词部分,并对应生成段落信息,其中,所述段落信息至少包括用于标识所述第二歌词部分的标识信息;
    对所述第一录制文件和所述伴奏音频文件进行混音处理,以得到所述歌曲的所述音频文件。
  17. 如权利要求16所述的方法,其中,基于所述第一录制文件确定所述 第二歌词部分包括:
    判断所述另一用户在某一歌词部分录入的声音是否有效,若是,则确定对应的歌词部分为第二歌词部分。
  18. 如权利要求16所述的方法,还包括生成配置文件,所述配置文件包括所述另一用户所采用的音效参数。
  19. 如权利要求18所述的方法,其中在所述歌词文件的第一歌词部分对应的显示时间播放所述音频文件并录制用户的音频数据包括:
    以所述歌曲的所述音频文件作为伴奏音频并根据所述段落信息和配置文件基于所述第二歌词部分以外的歌词部分进行录音。
  20. 一种音频处理装置,包括至少一个处理器和存储器,所述存储器存储有程序指令,所述程序指令当由处理器执行时,配置所述音频处理装置执行根据权利要求1-19中任一项所述的方法。
  21. 一种音频处理系统,包括服务器和至少一个客户端,所述服务器和/或所述客户端被配置为执行根据权利要求1-19中任一项所述的方法。
  22. 如权利要求21所述的系统,其中至少一个客户端包括第一客户端,所述第一客户端被配置为执行根据权利要求1-10中任一项所述的方法。
  23. 如权利要求22所述的系统,其中所述服务器被配置为:
    获取歌曲的伴奏音频文件、原唱音频文件以及所述歌曲的歌词文件,
    对所述歌词文件进行划分处理,以得到第一歌词部分和第二歌词部分,并对应生成歌词划分信息,其中,所述歌词划分信息包括至少用于标识所述第一歌词部分的标识信息,
    根据第一歌词部分对应的显示时间的伴奏音频文件,以及第二歌词部分对应的显示时间的原唱音频文件,生成所述歌曲的所述音频文件。
  24. 如权利要求23所述的系统,其中所述至少一个客户端还包括第二客户端,所述第二客户端被配置为:
    接收另一用户的请求;
    根据所述请求获取所述歌曲的伴奏音频文件和歌词文件;
    利用所述伴奏音频文件和歌词文件基于部分歌词对所述另一用户进行录音,得到第一录制文件;
    基于所述第一录制文件确定所述第二歌词部分,并对应生成段落信息,其中,所述段落信息包括用于标识所述第二歌词部分的标识信息;
    对所述第一录制文件和所述伴奏音频文件进行混音处理,以得到所述歌曲的所述音频文件。
  25. 一种音频处理装置,包括:
    第一获取单元,用于获取歌曲的音频文件以及所述歌曲的歌词文件;
    第一处理单元,用于在所述歌词文件的第一歌词部分对应的显示时间和第二歌词部分对应的显示时间播放所述音频文件,并在所述歌词文件的第一歌词部分对应的显示时间录制用户的音频数据;
    第一混音单元,用于对所述用户的音频数据和所述音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。
  26. 如权利要求25所述的装置,还包括第一接收单元,用于接收用户的请求,
    其中所述第一获取单元进一步用于根据所述请求获取歌曲的伴奏音频文件、原唱音频文件以及所述歌曲的歌词文件。
  27. 如权利要求26所述的装置,其中,所述第一处理单元进一步用于:
    对所述歌词文件进行划分处理,以得到第一歌词部分和第二歌词部分;
    在所述第一歌词部分对应的显示时间播放所述伴奏音频文件并录制用户的音频数据,
    在所述第二歌词部分对应的显示时间播放所述原唱音频文件。
  28. 如权利要求26所述的装置,
    其中所述第一混音单元进一步用于对所述用户的音频数据和所述伴奏音频文件在所述第一歌词部分对应的显示时间的音频数据进行混音处理。
  29. 如权利要求25所述的装置,其中所述混音处理生成混音音频文件,所述装置还包括
    第一生成单元,用于根据第一歌词部分对应的显示时间的所述混音音频文件,以及所述第二歌词部分对应的显示时间的所述原唱音频文件,生成合唱文件。
  30. 如权利要求25所述的装置,还包括:
    第二获取单元,用于获取歌曲的伴奏音频文件、原唱音频文件以及所述歌曲的歌词文件;
    第二处理单元,用于对所述歌词文件进行划分处理,得到第一歌词部分和第二歌词部分,并对应生成歌词划分信息,其中,所述歌词划分信息至少包括用于标识所述第一歌词部分的标识信息;
    第二生成单元,用于根据第一歌词部分对应的显示时间的伴奏音频文件,以及第二歌词部分对应的显示时间的原唱音频文件,生成所述歌曲的所述音频文件。
  31. 如权利要求25所述的装置,还包括:
    第二接收单元,用于接收另一用户的请求;
    第三获取单元,用于根据所述请求获取所述歌曲的伴奏音频文件和歌词文件,并利用所述伴奏音频文件和歌词文件基于部分歌词对所述另一用户进行录音,得到第一录制文件;
    第三处理单元,用于基于所述第一录制文件确定所述第二歌词部分,并对应生成段落信息,所述段落信息至少包括用于标识所述第二歌词部分的标识信息;
    第二混音单元,用于对所述第一录制文件和所述伴奏音频文件进行混音处理,以得到所述歌曲的所述音频文件。
PCT/CN2016/077218 2015-05-27 2016-03-24 音频处理方法、装置及系统 WO2016188211A1 (zh)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP16799100.9A EP3306606A4 (en) 2015-05-27 2016-03-24 METHOD, APPARATUS AND SYSTEM FOR AUDIO PROCESSING
JP2017560972A JP2018519536A (ja) 2015-05-27 2016-03-24 オーディオ処理方法、装置、およびシステム
KR1020177037253A KR20180012800A (ko) 2015-05-27 2016-03-24 오디오 처리 방법, 장치 및 시스템
US15/819,734 US10403255B2 (en) 2015-05-27 2017-11-21 Audio processing method, apparatus and system

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
CN201510280492.9A CN105023559A (zh) 2015-05-27 2015-05-27 K歌处理方法及系统
CN201510280087.7A CN104966527B (zh) 2015-05-27 2015-05-27 K歌处理方法、装置以及k歌处理系统
CN201510280087.7 2015-05-27
CN201510278911.5A CN105006234B (zh) 2015-05-27 2015-05-27 一种k歌处理方法及装置
CN201510278911.5 2015-05-27
CN201510280492.9 2015-05-27

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/819,734 Continuation US10403255B2 (en) 2015-05-27 2017-11-21 Audio processing method, apparatus and system

Publications (1)

Publication Number Publication Date
WO2016188211A1 true WO2016188211A1 (zh) 2016-12-01

Family

ID=57392491

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/077218 WO2016188211A1 (zh) 2015-05-27 2016-03-24 音频处理方法、装置及系统

Country Status (5)

Country Link
US (1) US10403255B2 (zh)
EP (1) EP3306606A4 (zh)
JP (1) JP2018519536A (zh)
KR (1) KR20180012800A (zh)
WO (1) WO2016188211A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107993637A (zh) * 2017-11-03 2018-05-04 厦门快商通信息技术有限公司 一种卡拉ok歌词分词方法与系统
CN110660376A (zh) * 2019-09-30 2020-01-07 腾讯音乐娱乐科技(深圳)有限公司 音频处理方法、装置及存储介质

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104966527B (zh) * 2015-05-27 2017-04-19 广州酷狗计算机科技有限公司 K歌处理方法、装置以及k歌处理系统
CN105023559A (zh) * 2015-05-27 2015-11-04 腾讯科技(深圳)有限公司 K歌处理方法及系统
CN108831425B (zh) * 2018-06-22 2022-01-04 广州酷狗计算机科技有限公司 混音方法、装置及存储介质
CN111046226B (zh) * 2018-10-15 2023-05-05 阿里巴巴集团控股有限公司 一种音乐的调音方法及装置
WO2021041393A1 (en) 2019-08-25 2021-03-04 Smule, Inc. Short segment generation for user engagement in vocal capture applications
KR102259488B1 (ko) * 2019-08-26 2021-06-02 주식회사 이멜벤처스 노래방 어플리케이션에서의 듀엣 노래부르기 제공 방법 및 시스템
CN111125028B (zh) * 2019-12-25 2023-10-24 腾讯音乐娱乐科技(深圳)有限公司 识别音频文件的方法、装置、服务器及存储介质
CN111402844B (zh) * 2020-03-26 2024-04-09 广州酷狗计算机科技有限公司 歌曲合唱的方法、装置及系统

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005010639A (ja) * 2003-06-20 2005-01-13 Yamaha Corp カラオケ装置
JP2009075611A (ja) * 2008-12-08 2009-04-09 Yamaha Corp 合唱合成装置、合唱合成方法およびプログラム
TW200923675A (en) * 2007-11-19 2009-06-01 Inventec Besta Co Ltd Network chorusing system and method thereof
CN102456340A (zh) * 2010-10-19 2012-05-16 盛大计算机(上海)有限公司 基于互联网的卡拉ok对唱方法及系统
TW201228290A (en) * 2010-12-28 2012-07-01 Tse-Ming Chang Networking multi-person asynchronous chorus audio/video works system
CN103295568A (zh) * 2013-05-30 2013-09-11 北京小米科技有限责任公司 一种异步合唱方法和装置
CN104966527A (zh) * 2015-05-27 2015-10-07 腾讯科技(深圳)有限公司 K歌处理方法、装置以及k歌处理系统
CN105006234A (zh) * 2015-05-27 2015-10-28 腾讯科技(深圳)有限公司 一种k歌处理方法及装置
CN105023559A (zh) * 2015-05-27 2015-11-04 腾讯科技(深圳)有限公司 K歌处理方法及系统

Family Cites Families (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3241372B2 (ja) * 1990-11-27 2001-12-25 パイオニア株式会社 カラオケ演奏方法
JPH04199096A (ja) * 1990-11-29 1992-07-20 Pioneer Electron Corp カラオケ演奏装置
JPH0561491A (ja) * 1991-09-02 1993-03-12 Sanyo Electric Co Ltd カラオケ装置及びその記録媒体
JPH0744182A (ja) 1993-07-21 1995-02-14 Hitachi Ltd 音楽再生装置
KR0129964B1 (ko) * 1994-07-26 1998-04-18 김광호 악기선택 가능한 영상노래반주장치
JP2838977B2 (ja) * 1995-01-17 1998-12-16 ヤマハ株式会社 カラオケ装置
JP2820052B2 (ja) 1995-02-02 1998-11-05 ヤマハ株式会社 コーラス効果付与装置
JP3662969B2 (ja) 1995-03-06 2005-06-22 富士通株式会社 カラオケシステム
JP3587916B2 (ja) * 1995-10-31 2004-11-10 ブラザー工業株式会社 映像音声データ供給装置
US6231347B1 (en) * 1995-11-20 2001-05-15 Yamaha Corporation Computer system and karaoke system
JP3353595B2 (ja) * 1996-03-21 2002-12-03 ヤマハ株式会社 自動演奏装置およびカラオケ装置
JP3871382B2 (ja) * 1996-09-11 2007-01-24 ブラザー工業株式会社 楽音再生装置
JP3709631B2 (ja) * 1996-11-20 2005-10-26 ヤマハ株式会社 カラオケ装置
JP4013281B2 (ja) * 1997-04-18 2007-11-28 ヤマハ株式会社 カラオケデータ伝送方法、カラオケ装置およびカラオケデータ記録媒体
JP4029442B2 (ja) * 1997-07-18 2008-01-09 ヤマハ株式会社 通信カラオケシステム
KR100270340B1 (ko) 1998-08-17 2000-12-01 김대기 이동전화망을 이용한 노래반주 서비스 시스템 및 그 구현방법
JP4236024B2 (ja) * 1999-03-08 2009-03-11 株式会社フェイス データ再生装置および情報端末機
JP4172610B2 (ja) * 1999-08-03 2008-10-29 株式会社タイトー プレイヤー参加型通信カラオケシステム
US6740804B2 (en) * 2001-02-05 2004-05-25 Yamaha Corporation Waveform generating method, performance data processing method, waveform selection apparatus, waveform data recording apparatus, and waveform data recording and reproducing apparatus
WO2003001477A1 (en) * 2001-06-18 2003-01-03 First International Digital, Inc. Music player with synchronized recording and display
TWI244838B (en) 2002-01-07 2005-12-01 Compal Electronics Inc Method of karaoke by network system
JP4188013B2 (ja) * 2002-07-17 2008-11-26 株式会社第一興商 ファイル保管サーバに預託した歌唱トラック録音ファイルをカラオケ演奏装置に取り寄せて合唱することができる通信カラオケシステム、カラオケ演奏装置
JP2004240065A (ja) * 2003-02-04 2004-08-26 Konami Co Ltd カラオケ装置、音声出力制御方法、ならびに、プログラム
US7134876B2 (en) * 2004-03-30 2006-11-14 Mica Electronic Corporation Sound system with dedicated vocal channel
JP4382786B2 (ja) 2006-08-22 2009-12-16 株式会社タイトー 音声ミックスダウン装置、音声ミックスダウンプログラム
US20080184870A1 (en) * 2006-10-24 2008-08-07 Nokia Corporation System, method, device, and computer program product providing for a multiple-lyric karaoke system
KR20080082019A (ko) 2006-12-29 2008-09-11 고리텍 주식회사 원음 노래방 서비스를 제공하는 휴대 노래방 시스템 및 그방법
US8138409B2 (en) * 2007-08-10 2012-03-20 Sonicjam, Inc. Interactive music training and entertainment system
CN101131816B (zh) 2007-09-30 2012-01-04 炬力集成电路设计有限公司 一种音频文件生成方法、装置及数码播放器
JP2010014823A (ja) * 2008-07-01 2010-01-21 Nippon Telegr & Teleph Corp <Ntt> 楽曲情報制御装置
JP5331494B2 (ja) * 2009-01-19 2013-10-30 株式会社タイトー カラオケサービスシステム、端末装置
CN101483536B (zh) 2009-02-10 2010-11-03 腾讯科技(深圳)有限公司 传输用于实现网络卡拉ok的数据的方法、系统和客户端
CN101859561B (zh) 2009-04-07 2012-12-26 林文信 伴唱歌词自动显示方法
US20110126103A1 (en) * 2009-11-24 2011-05-26 Tunewiki Ltd. Method and system for a "karaoke collage"
JP2011191357A (ja) 2010-03-12 2011-09-29 Yamaha Corp カラオケシステム
JP5429495B2 (ja) * 2010-07-28 2014-02-26 ブラザー工業株式会社 録画装置及び録画プログラム
CN102497448A (zh) 2011-12-26 2012-06-13 深圳市五巨科技有限公司 一种移动终端在线音乐播放的方法及系统
JP2014006480A (ja) 2012-06-27 2014-01-16 Sony Corp 情報処理装置、情報処理方法及びプログラム
CN103165119B (zh) 2013-01-31 2015-12-09 华为技术有限公司 K歌方法及系统
US9224374B2 (en) * 2013-05-30 2015-12-29 Xiaomi Inc. Methods and devices for audio processing
CN103337240B (zh) 2013-06-24 2016-03-30 华为技术有限公司 处理语音数据的方法、终端、服务器及系统
CN103841278B (zh) 2014-03-11 2017-11-28 华为技术有限公司 歌曲点唱的处理方法、装置及系统
CN104392711A (zh) 2014-08-27 2015-03-04 贵阳朗玛信息技术股份有限公司 一种实现卡拉ok功能的方法及装置

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005010639A (ja) * 2003-06-20 2005-01-13 Yamaha Corp カラオケ装置
TW200923675A (en) * 2007-11-19 2009-06-01 Inventec Besta Co Ltd Network chorusing system and method thereof
JP2009075611A (ja) * 2008-12-08 2009-04-09 Yamaha Corp 合唱合成装置、合唱合成方法およびプログラム
CN102456340A (zh) * 2010-10-19 2012-05-16 盛大计算机(上海)有限公司 基于互联网的卡拉ok对唱方法及系统
TW201228290A (en) * 2010-12-28 2012-07-01 Tse-Ming Chang Networking multi-person asynchronous chorus audio/video works system
CN103295568A (zh) * 2013-05-30 2013-09-11 北京小米科技有限责任公司 一种异步合唱方法和装置
CN104966527A (zh) * 2015-05-27 2015-10-07 腾讯科技(深圳)有限公司 K歌处理方法、装置以及k歌处理系统
CN105006234A (zh) * 2015-05-27 2015-10-28 腾讯科技(深圳)有限公司 一种k歌处理方法及装置
CN105023559A (zh) * 2015-05-27 2015-11-04 腾讯科技(深圳)有限公司 K歌处理方法及系统

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3306606A4 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107993637A (zh) * 2017-11-03 2018-05-04 厦门快商通信息技术有限公司 一种卡拉ok歌词分词方法与系统
CN110660376A (zh) * 2019-09-30 2020-01-07 腾讯音乐娱乐科技(深圳)有限公司 音频处理方法、装置及存储介质
CN110660376B (zh) * 2019-09-30 2022-11-29 腾讯音乐娱乐科技(深圳)有限公司 音频处理方法、装置及存储介质

Also Published As

Publication number Publication date
EP3306606A4 (en) 2019-01-16
US20180090116A1 (en) 2018-03-29
US10403255B2 (en) 2019-09-03
JP2018519536A (ja) 2018-07-19
KR20180012800A (ko) 2018-02-06
EP3306606A1 (en) 2018-04-11

Similar Documents

Publication Publication Date Title
WO2016188211A1 (zh) 音频处理方法、装置及系统
WO2016188322A1 (zh) K歌处理方法、装置以及k歌处理系统
TWI576822B (zh) K歌處理方法及系統
CN106652997B (zh) 一种音频合成的方法及终端
WO2020107626A1 (zh) 歌词显示处理方法、装置、电子设备及计算机可读存储介质
CN110992970B (zh) 音频合成方法及相关装置
JP6669883B2 (ja) 音声データ処理方法及び装置
US11120782B1 (en) System, method, and non-transitory computer-readable storage medium for collaborating on a musical composition over a communication network
WO2016112841A1 (zh) 一种信息处理方法及客户端、计算机存储介质
CN104361897B (zh) 一种制作铃音的方法及装置
JP6452229B2 (ja) カラオケ効果音設定システム
CN110428798B (zh) 人声与伴奏同步方法、蓝牙设备、终端及存储介质
JP5986387B2 (ja) 情報処理プログラム、情報処理装置、歌詞表示方法、及び通信システム
JP5311071B2 (ja) 楽曲再生装置及び楽曲再生プログラム
EP3203468B1 (en) Acoustic system, communication device, and program
KR101573868B1 (ko) 노래 가사 자동 디스플레이 방법, 노래 가사를 인식하는 서버 및 이 서버를 포함하는 노래 가사 자동 디스플레이 시스템
JP7117228B2 (ja) カラオケシステム、カラオケ装置
JP2023144076A (ja) プログラム、情報処理方法及び情報処理装置
JP6182493B2 (ja) 音楽再生システム、サーバ、及びプログラム
JP4462368B2 (ja) 楽曲再生装置及び楽曲再生プログラム
KR101458526B1 (ko) 공동음원 생성 서비스 시스템 및 그 방법, 그리고 이에 적용되는 장치

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16799100

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2017560972

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 20177037253

Country of ref document: KR

Kind code of ref document: A