CN104572882A - Audio data management method, server and client - Google Patents

Audio data management method, server and client Download PDF

Info

Publication number
CN104572882A
CN104572882A CN201410808946.0A CN201410808946A CN104572882A CN 104572882 A CN104572882 A CN 104572882A CN 201410808946 A CN201410808946 A CN 201410808946A CN 104572882 A CN104572882 A CN 104572882A
Authority
CN
China
Prior art keywords
voice data
audio
data
server
mark
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410808946.0A
Other languages
Chinese (zh)
Other versions
CN104572882B (en
Inventor
敖绍青
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Kugou Computer Technology Co Ltd
Original Assignee
Guangzhou Kugou Computer Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Kugou Computer Technology Co Ltd filed Critical Guangzhou Kugou Computer Technology Co Ltd
Priority to CN201410808946.0A priority Critical patent/CN104572882B/en
Publication of CN104572882A publication Critical patent/CN104572882A/en
Application granted granted Critical
Publication of CN104572882B publication Critical patent/CN104572882B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content

Abstract

The invention discloses an audio data management method, a server and a client, and belongs to the technical field of internet technology. The audio data management method comprises the steps: receiving an audio search request transmitted by a first client; acquiring first audio data and a first identifier of the first audio data; transmitting first audio data and the first identifier to the first client; correspondently storing the second audio data according to the second audio data generated by the first client according to the third audio data, the first audio data and the first identifier. The first audio data and the first identifier are returned to the first client, and after the first client submits the second audio data, the first identifier and the second audio data are correspondently stored. Since the audio data is managed by referring to the first identifier, so that the audio data management way is relatively orderly. When the second audio data generated by other users according to the same audio data is searched by the user, the second audio data can be searched according to the first identifier, so that the search process is time-saving.

Description

Audio data management method, server and client
Technical field
The present invention relates to Internet technical field, particularly a kind of audio data management method, server and client.
Background technology
Along with developing rapidly of Internet technology, occur that various K sings software.Sing software by K, user can realize K song whenever and wherever possible.User is when using K to sing software K song, and sing software by K and provide accompaniment, user follows accompaniment and sings, and K sings the song of software to user and records, and this song and accompaniment are synthesized voice data.
When a user selects the accompaniment of a song to sing, K sings the user song of software according to the accompaniment of this song and recording, generates a voice data, therefore, when multiple user selects the accompaniment of same song to sing, K sings software for the accompaniment of this song, can generate multiple voice data.
In this case, if each voice data stores in a jumble because of the rule of not fixing when storing, be then not easy to K and sing each voice data of software administration.In addition, if some users will search for the voice data that other users generate according to same accompaniment, compare with to carry out K song, then because each voice data stores in a jumble, make the voice data being not easy to search user's needs, cause search procedure more time-consuming.To sum up, the management method of voice data is most important.
Summary of the invention
In order to solve the problem of correlation technique, embodiments provide a kind of audio data management method, server and client.Described technical scheme is as follows:
First aspect, provide a kind of audio data management method, described method comprises:
Receive the audio search requests that the first client sends, described audio search requests carries audio keyword;
According to described audio keyword, obtain the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
Described first voice data and described first mark is sent to described first client;
When receiving the second audio data that described first client sends, according to described first mark, second audio data described in corresponding stored, the 3rd voice data, described first voice data and described first mark that described second audio data is recorded according to user by described first client generate.
In conjunction with first aspect, in the first possible implementation of first aspect, described before described first client described first voice data of transmission and described first mark, also comprise:
Obtain voice data;
According to the audio-frequency information of described voice data, judge whether this locality has stored described voice data;
If this locality does not store described voice data, then store described voice data, described voice data comprises described first voice data;
For described voice data generates the audio identification being used for voice data described in unique identification.
In conjunction with the first possible implementation of first aspect, in the implementation that the second of first aspect is possible, described acquisition voice data, comprising:
Every the first preset duration, obtain first kind voice data from the first audio data service device, described first kind voice data is that tonequality is higher than the voice data of specifying tonequality standard;
Every the second preset duration, obtain Equations of The Second Kind voice data from second audio data server, described Equations of The Second Kind voice data is the voice data of tonequality lower than described appointment tonequality standard;
Every the 3rd preset duration, the 3rd class audio frequency data are obtained from the 3rd audio data service device, described 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of described 3rd class audio frequency data is lower than the tonequality of described Equations of The Second Kind voice data;
Receive the 4th class audio frequency data that the second client sends, described 4th class audio frequency data are obtained from third-party server search by described second client.
In conjunction with the implementation that the second of first aspect is possible, in the third possible implementation of first aspect, described every the first preset duration, after the first audio data service device obtains first kind voice data, also comprise:
At the audio-frequency information according to described first voice data, after determining that described first kind voice data comprises described first voice data, the first voice data in described first kind voice data is used to replace local the first voice data stored;
Described every the second preset duration, after second audio data server obtains Equations of The Second Kind voice data, also comprise:
At the audio-frequency information according to described first voice data, after determining that described Equations of The Second Kind voice data comprises described first voice data, judge whether local the first voice data stored derives from described 3rd audio data service device;
If local the first audio data source stored is in described 3rd audio data service device, then the first voice data in described Equations of The Second Kind voice data is used to replace local the first voice data stored;
If local the first voice data stored does not derive from described 3rd audio data service device, then ignore the first voice data in described Equations of The Second Kind voice data;
Described every the 3rd preset duration, after the 3rd audio data service device obtains the 3rd class audio frequency data, also comprise:
At the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, ignore the first voice data in the 3rd class audio frequency data got;
After the 4th class audio frequency data that described reception second client sends, also comprise:
At the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, after determining that the tonequality of the first voice data in described 4th class audio frequency data is better than the tonequality of local the first voice data stored, upgrade local the first voice data stored;
After the content of the audio-frequency information of the first voice data stored more than this locality in the content of the audio-frequency information determining the first voice data in described 4th class audio frequency data, upgrade the audio-frequency information of local the first voice data stored.
In conjunction with first aspect to any one the possible implementation in the third possible implementation of the first aspect of first aspect, in the 4th kind of possible implementation of first aspect, described first voice data is accompaniment data.
Second aspect, provide a kind of audio data management method, described method comprises:
Send audio search requests, described audio search requests carries audio keyword;
Obtain the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
According to the 3rd voice data, described first voice data and described first mark that user records, generate second audio data;
Described second audio data is sent to server, makes described server according to described first mark, second audio data described in corresponding stored.
In conjunction with second aspect, in the first possible implementation of second aspect, described transmission audio search requests, comprising:
Audio search requests is sent respectively to described server and third-party server;
Described acquisition first voice data and the first mark for the first voice data described in unique identification, comprising:
Receive the first voice data that described server returns and described server is the first mark that described first voice data generates;
Or, receive at least one the 4th voice data that described third-party server returns; After detecting that user selects the first voice data from least one the 4th voice data described, the first voice data selected is committed to described server, makes described server be that described first voice data generates the first mark; Receiving described server is the first mark that described first voice data generates;
Or, receive the first voice data that described server returns and described server is the first mark that described first voice data generates; Receive at least one the 4th voice data that described third-party server returns; After the first voice data that user selects described server to return being detected, the first mark returned when described server is returned described first voice data is as the first mark got; After detecting that user selects a voice data at least one the 4th voice data described as the first voice data, the first voice data selected is committed to described server, after the first voice data being described selection by described server generates the first mark, receive the first mark that described server is the first voice data generation of described selection, using the receive first mark as the first mark got.
In conjunction with the first possible implementation of second aspect, in the implementation that the second of second aspect is possible, after at least one the 4th voice data that the described third-party server of described reception returns, also comprise:
Whether the audio-frequency information of the first voice data returned according to described server and the audio-frequency information of at least one the 4th voice data described, exist the first voice data that described server returns at least one the 4th voice data described in determining;
If there is the first voice data that described server returns at least one the 4th voice data described, from at least one the 4th voice data described, then delete the first voice data that described server returns, obtain at least one the 4th voice data after upgrading;
Show at least one the 4th voice data after described first voice data and described renewal.
In conjunction with the implementation that the second of second aspect is possible, in the third possible implementation of second aspect, at least one the 4th voice data after described first voice data of described display and described renewal, comprising:
Different classes of according at least one the 4th voice data after described first voice data and described renewal, shows at least one the 4th voice data after described first voice data and described renewal in the zones of different of display screen.
In conjunction with second aspect to any one the possible implementation in the third possible implementation of second aspect, in the 4th kind of possible implementation of second aspect, described first voice data is accompaniment data.
The third aspect, provides a kind of server, and described server comprises:
Receiver module, for receiving the audio search requests that the first client sends, described audio search requests carries audio keyword;
First acquisition module, for according to described audio keyword, obtain the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
Sending module, for sending described first voice data and described first mark to described first client;
First memory module, for when receiving the second audio data that described first client sends, according to described first mark, second audio data described in corresponding stored, the 3rd voice data, described first voice data and described first mark that described second audio data is recorded according to user by described first client generate.
In conjunction with the third aspect, in the first possible implementation of the third aspect, described server also comprises:
Second acquisition module, for obtaining voice data;
Judge module, for the audio-frequency information according to described voice data, judges whether this locality has stored described voice data;
Second memory module, for when this locality does not store described voice data, stores described voice data, and described voice data comprises described first voice data;
Generation module, for generating for described voice data the audio identification being used for voice data described in unique identification.
In conjunction with the first possible implementation of the third aspect, in the implementation that the second of the third aspect is possible, described second acquisition module comprises:
First acquiring unit, for every the first preset duration, obtains first kind voice data from the first audio data service device, and described first kind voice data is that tonequality is higher than the voice data of specifying tonequality standard;
Second acquisition unit, for every the second preset duration, obtains Equations of The Second Kind voice data from second audio data server, and described Equations of The Second Kind voice data is the voice data of tonequality lower than described appointment tonequality standard;
3rd acquiring unit, for every the 3rd preset duration, the 3rd class audio frequency data are obtained from the 3rd audio data service device, described 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of described 3rd class audio frequency data is lower than the tonequality of described Equations of The Second Kind voice data;
Receiving element, for receiving the 4th class audio frequency data that the second client sends, described 4th class audio frequency data are obtained from third-party server search by described second client.
In conjunction with the implementation that the second of the third aspect is possible, in the third possible implementation of the third aspect, described second acquisition module also comprises:
First replacement unit, for at the audio-frequency information according to described first voice data, after determining that described first kind voice data comprises described first voice data, the first voice data in described first kind voice data is used to replace local the first voice data stored;
Judging unit, at the audio-frequency information according to described first voice data, after determining that described Equations of The Second Kind voice data comprises described first voice data, judges whether local the first voice data stored derives from described 3rd audio data service device;
Second replacement unit, for when local the first audio data source stored is in described 3rd audio data service device, uses the first voice data in described Equations of The Second Kind voice data to replace local the first voice data stored;
First ignores unit, for when local the first voice data stored does not derive from described 3rd audio data service device, ignores the first voice data in described Equations of The Second Kind voice data;
Second ignores unit, for when at the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, ignores the first voice data in the 3rd class audio frequency data got;
First updating block, for at the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, after determining that the tonequality of the first voice data in described 4th class audio frequency data is better than the tonequality of local the first voice data stored, upgrade local the first voice data stored;
Second updating block, for after the content of the audio-frequency information of the first voice data that stores more than this locality in the content of the audio-frequency information determining the first voice data in described 4th class audio frequency data, upgrades the audio-frequency information of local the first voice data stored.
In conjunction with the third aspect to any one the possible implementation in the third possible implementation of the third aspect, in the 4th kind of possible implementation of the third aspect, described first voice data is accompaniment data.
Fourth aspect, provides a kind of client, and described client comprises:
First sending module, for sending audio search requests, described audio search requests carries audio keyword;
Acquisition module, for obtaining the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
Generation module, for the 3rd voice data recorded according to user, described first voice data and described first mark, generates second audio data;
Second sending module, for described second audio data is sent to server, makes described server according to described first mark, second audio data described in corresponding stored.
In conjunction with fourth aspect, in the first possible implementation of fourth aspect, described first sending module, for sending audio search requests respectively to described server and third-party server;
Described acquisition module, for
Receive the first voice data that described server returns and described server is the first mark that described first voice data generates;
Or, receive at least one the 4th voice data that described third-party server returns; After detecting that user selects the first voice data from least one the 4th voice data described, the first voice data selected is committed to described server, makes described server be that described first voice data generates the first mark; Receiving described server is the first mark that described first voice data generates;
Or, receive the first voice data that described server returns and described server is the first mark that described first voice data generates; Receive at least one the 4th voice data that described third-party server returns; After the first voice data that user selects described server to return being detected, the first mark returned when described server is returned described first voice data is as the first mark got; After detecting that user selects a voice data at least one the 4th voice data described as the first voice data, the first voice data selected is committed to described server, after the first voice data being described selection by described server generates the first mark, receive the first mark that described server is the first voice data generation of described selection, using the receive first mark as the first mark got.
In conjunction with the first possible implementation of fourth aspect, in the implementation that the second of fourth aspect is possible, described client also comprises:
Determination module, for the audio-frequency information of the first voice data that returns according to described server and the audio-frequency information of at least one the 4th voice data described, determine whether to exist at least one the 4th voice data described the first voice data that described server returns;
Removing module, for when there is the first voice data that described server returns at least one the 4th voice data described, from at least one the 4th voice data described, delete the first voice data that described server returns, obtain at least one the 4th voice data after upgrading;
Display module, for showing at least one the 4th voice data after described first voice data and described renewal.
In conjunction with the implementation that the second of fourth aspect is possible, in the third possible implementation of fourth aspect, described display module, different classes of for according at least one the 4th voice data after described first voice data and described renewal, shows at least one the 4th voice data after described first voice data and described renewal in the zones of different of display screen.
In conjunction with fourth aspect to any one the possible implementation in the third possible implementation of fourth aspect, in the 4th kind of possible implementation of fourth aspect, described first voice data is accompaniment data.
The beneficial effect that the technical scheme that the embodiment of the present invention provides is brought is:
By returning the first mark of the first voice data and this first voice data to the first client, guarantee after the first client submits second audio data to, server can by the first mark and second audio data corresponding stored.During owing to managing voice data, be designated reference with first, thus make audio data management Method compare orderly.In addition, when user will search for the second audio data of other users according to same first voice data generation, during to carry out the competition of K song, can search for according to the first mark, search procedure be compared and saves time.
Accompanying drawing explanation
In order to be illustrated more clearly in the technical scheme in the embodiment of the present invention, below the accompanying drawing used required in describing embodiment is briefly described, apparently, accompanying drawing in the following describes is only some embodiments of the present invention, for those of ordinary skill in the art, under the prerequisite not paying creative work, other accompanying drawing can also be obtained according to these accompanying drawings.
Fig. 1 is the implementation environment schematic diagram involved by a kind of audio data management method that one embodiment of the invention provides;
Fig. 2 is the flow chart of a kind of audio data management method that another embodiment of the present invention provides;
Fig. 3 is the flow chart of a kind of audio data management method that another embodiment of the present invention provides;
Fig. 4 is the flow chart of a kind of audio data management method that another embodiment of the present invention provides;
Fig. 5 is the structural representation of a kind of server that another embodiment of the present invention provides;
Fig. 6 is the structural representation of a kind of server that another embodiment of the present invention provides;
Fig. 7 is the structural representation of a kind of client that another embodiment of the present invention provides;
Fig. 8 is the structural representation of a kind of terminal that another embodiment of the present invention provides.
Embodiment
For making the object, technical solutions and advantages of the present invention clearly, below in conjunction with accompanying drawing, embodiment of the present invention is described further in detail.
As shown in Figure 1, it illustrates the implementation environment schematic diagram involved by a kind of audio data management method that the embodiment of the present invention provides.As shown in Figure 1, this implementation environment comprises a server cluster, at least one client and at least one third-party server.Wherein, this server cluster comprises server 101 and each audio data service device, and each audio data service device is respectively the first audio data service device 102, second audio data server 103 and the 3rd audio data service device 104.First audio data service device 102, second audio data server 103 and the 3rd audio data service device 104 can process respectively and obtain first kind voice data, Equations of The Second Kind voice data and the 3rd class audio frequency data.Server 101 can provide accompanying song service, i.e. karaoke service, after each accessing server by customer end 101, can be sung by the voice data searched.
Particularly, server 101 obtains voice data every preset duration from each audio data service device, integrates with the voice data obtained each audio data service device process, and guaranteeing can provide the voice data of high-quality for the user of each client.At least one client comprises first client 105, first client 105 can search for voice data to server 101, and the voice data using server 101 to return is sung for user.Alternatively, the first client 105 can also search for voice data to third-party server 106, and the voice data using third-party server 106 to return is sung.Wherein, the first client 105 is the 4th class audio frequency data from the voice data that third-party server 106 searches.In addition, at least one second client 107 is also comprised in this at least one client, after the second client 107 searches the 4th class audio frequency data from third-party server 106, be committed to server 101, this voice data of server stores, to guarantee, when the first client 105 searches for voice data, this voice data to be obtained from server 101.
As shown in Figure 1, be connected respectively by network between server 101 with each audio data service device, first client 105 and the second client 107 are connected by network with server 101 respectively, and the first client 105 and the second client 107 are connected by network with third-party server 106 respectively.This network can be cable network, also can be wireless network.
Wherein, first client 105 and the terminal corresponding to the second client 107 can be smart mobile phone, Intelligent bracelet, wearable device, panel computer, E-book reader, MP3 (Moving Picture ExpertsGroup Audio Layer III, dynamic image expert compression standard audio frequency aspect 3) player, MP4 (MovingPicture Experts Group Audio Layer IV, dynamic image expert compression standard audio frequency aspect 3) player, pocket computer on knee and desktop computer etc.
Concrete audio data management method refers to each embodiment following:
Implementation environment schematic diagram shown in composition graphs 1, Fig. 2 provides a kind of flow chart of audio data management method according to an exemplary embodiment.For the method that the server execution embodiment of the present invention provides, see Fig. 2, the method flow that the embodiment of the present invention provides comprises:
201, receive the audio search requests that the first client sends, wherein, audio search requests carries audio keyword.
202, according to audio keyword, obtain the first voice data and the first mark for unique identification first voice data, wherein, the audio-frequency information of the first voice data comprises audio keyword.
203, the first voice data and the first mark is sent to the first client.
204, when the second audio data that reception first client sends, according to the first mark, corresponding stored second audio data, wherein, the 3rd voice data, the first voice data and the first mark that second audio data is recorded according to user by the first client generate.
The method that the embodiment of the present invention provides, by returning the first mark of the first voice data and this first voice data to the first client, guarantee after the first client submits second audio data to, server can by the first mark and second audio data corresponding stored.During owing to managing voice data, be designated reference with first, thus make audio data management Method compare orderly.In addition, user will search for the second audio data that other users generate according to same first voice data, during to carry out the competition of K song, can search for, search procedure is compared and saves time according to the first mark.
In another embodiment, before sending the first voice data and the first mark to the first client, also comprise:
Obtain voice data;
According to the audio-frequency information of voice data, judge whether this locality stores voice data;
If this locality does not store voice data, then stores audio data, wherein, voice data comprises the first voice data;
For voice data generates the audio identification being used for unique identification voice data.
In another embodiment, obtain voice data, comprising:
Every the first preset duration, obtain first kind voice data from the first audio data service device, wherein, first kind voice data is that tonequality is higher than the voice data of specifying tonequality standard;
Every the second preset duration, obtain Equations of The Second Kind voice data from second audio data server, wherein, Equations of The Second Kind voice data is that tonequality is lower than the voice data of specifying tonequality standard;
Every the 3rd preset duration, the 3rd class audio frequency data are obtained from the 3rd audio data service device, wherein, the 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of the 3rd class audio frequency data is lower than the tonequality of Equations of The Second Kind voice data;
Receive the 4th class audio frequency data that the second client sends, wherein, the 4th class audio frequency data are obtained from third-party server search by the second client.
In another embodiment, every the first preset duration, after the first audio data service device obtains first kind voice data, also comprise:
At the audio-frequency information according to the first voice data, after determining that first kind voice data comprises the first voice data, the first voice data in first kind voice data is used to replace local the first voice data stored;
Every the second preset duration, after second audio data server obtains Equations of The Second Kind voice data, also comprise:
At the audio-frequency information according to the first voice data, after determining that Equations of The Second Kind voice data comprises the first voice data, judge whether local the first voice data stored derives from the 3rd audio data service device;
If local the first audio data source stored is in the 3rd audio data service device, then the first voice data in Equations of The Second Kind voice data is used to replace local the first voice data stored;
If local the first voice data stored does not derive from the 3rd audio data service device, then ignore the first voice data in Equations of The Second Kind voice data;
Every the 3rd preset duration, after the 3rd audio data service device obtains the 3rd class audio frequency data, also comprise:
At the audio-frequency information according to the first voice data, after determining that this locality has stored the first voice data, ignore the first voice data in the 3rd class audio frequency data got;
After receiving the 4th class audio frequency data of the second client transmission, also comprise:
At the audio-frequency information according to the first voice data, after determining that this locality has stored the first voice data, after determining that the tonequality of the first voice data in the 4th class audio frequency data is better than the tonequality of local the first voice data stored, upgrade local the first voice data stored;
After the content of the audio-frequency information of the first voice data stored more than this locality in the content of the audio-frequency information determining the first voice data in the 4th class audio frequency data, upgrade the audio-frequency information of local the first voice data stored.
In another embodiment, the first voice data is accompaniment data.
Above-mentioned all alternatives, can adopt and combine arbitrarily formation optional embodiment of the present invention, this is no longer going to repeat them.
The content of embodiment corresponding to implementation environment schematic diagram shown in composition graphs 1 and Fig. 2, Fig. 3 provides a kind of flow chart of audio data management method according to an exemplary embodiment.For the method that the first client executing embodiment of the present invention provides, see Fig. 3, the method flow that the embodiment of the present invention provides comprises:
301, send audio search requests, wherein, audio search requests carries audio keyword.
302, obtain the first voice data and the first mark for unique identification first voice data, wherein, the audio-frequency information of the first voice data comprises audio keyword.
303, according to the 3rd voice data, the first voice data and the first mark that user records, second audio data is generated.
304, second audio data is sent to server, makes server according to the first mark, corresponding stored second audio data.
The method that the embodiment of the present invention provides, first mark of the first voice data returned by reception server and this first voice data, guarantee to sing according to the first voice data as user, after obtaining second audio data, submit second audio data to server, server can be identified first and second audio data corresponding stored.Because server is when managing voice data, be designated reference with first, thus make the way to manage of server to voice data more orderly.In addition, when user will search for the second audio data of other users according to same first voice data generation to carry out the competition of K song, server can be searched for according to the first mark, search procedure is compared and saves time.
In another embodiment, send audio search requests, comprising:
Audio search requests is sent respectively to server and third-party server;
Obtain the first voice data and the first mark for unique identification first voice data, comprising:
The first voice data that reception server returns and server are the first mark that the first voice data generates;
Or, receive at least one the 4th voice data that third-party server returns; After detecting that user selects the first voice data from least one the 4th voice data, the first voice data selected is committed to server, makes server be that the first voice data generates the first mark; Reception server is the first mark that the first voice data generates;
Or the first voice data that reception server returns and server are the first mark that the first voice data generates; Receive at least one the 4th voice data that third-party server returns; After the first voice data that user selects server to return being detected, the first mark returned when server being returned the first voice data is as the first mark got; After detecting that user selects a voice data at least one the 4th voice data as the first voice data, the first voice data selected is committed to server, be after the first voice data generation first mark selected by server, reception server is the first mark that the first voice data selected generates, using the receive first mark as the first mark got.
In another embodiment, after at least one the 4th voice data that reception third-party server returns, also comprise:
The audio-frequency information of the first voice data returned according to server and the audio-frequency information of at least one the 4th voice data, determine the first voice data that at least one the 4th voice data, whether presence server returns;
If the first voice data that at least one the 4th voice data, presence server returns, then from least one the 4th voice data, delete the first voice data that server returns, obtain at least one the 4th voice data after upgrading;
Show the first voice data and at least one the 4th voice data after upgrading.
In another embodiment, show the first voice data and at least one the 4th voice data after upgrading, comprising:
According to the first voice data and at least one the 4th voice data different classes of after upgrading, in the zones of different of display screen, the first voice data and at least one the 4th voice data after upgrading are shown.
In another embodiment, the first voice data is accompaniment data.
Above-mentioned all alternatives, can adopt and combine arbitrarily formation optional embodiment of the present invention, this is no longer going to repeat them.
The content of embodiment corresponding to implementation environment schematic diagram shown in composition graphs 1 and Fig. 2 or Fig. 3, Fig. 4 provides a kind of flow chart of audio data management method according to an exemplary embodiment.Perform for server and the first client the method that the embodiment of the present invention provides alternately, see Fig. 4, the method flow that the embodiment of the present invention provides comprises:
401, the first user end to server sends audio search requests, and this audio search requests carries audio keyword.
In embodiments of the present invention, server provides K to sing service, and voice data required when providing K to sing.When the user of the first client carries out K song, can the first user end to server request voice data be passed through, and according to asking the voice data obtained to be sung.When the first client detects the K song operation of user, and when detecting that user has carried out the operation of the voice data of a certain song of search, trigger the first user end to server and send audio search requests.Such as, when detecting that user have input audio keyword in the audio service interface that the first client provides, and after detecting that search option in audio service interface is selected further, triggering and sending audio search requests to server.
Wherein, audio search requests carries audio keyword, and this audio keyword is the word of the feature of voice data that can identify the first client and ask or word.Such as, this audio keyword can be song title and/or singer's name.By carrying audio keyword in audio search requests, server can be known, and the first client needs the voice data of search.
402, after server receives the audio search requests of the first client transmission, according to audio keyword, obtain the first voice data and the first mark for unique identification first voice data, wherein, the audio-frequency information of the first voice data comprises audio keyword.
After server receives the audio search requests that the first client sends, in order to return voice data needed for the first client to the first client, server, according to audio keyword, obtains the first voice data.Wherein, the first voice data is the voice data corresponding with audio keyword that server gets.
In addition, in embodiments of the present invention, server is for the ease of management by the voice data for the actual performance synthesized by audio accompaniment data, and each voice data that can store for server this locality generates the unique mark of an overall situation.In this step, the second audio data that server generates according to the first voice data for the ease of managing follow-up first client, also needs the first mark obtaining this first voice data.
The mode of the first voice data is obtained about server, include but not limited to: according to audio keyword, the local all voice datas prestored of inquiry, obtain the voice data mated with this audio keyword, using the voice data that mates with this audio keyword as the first voice data.
Such as, if audio keyword is singer A and song B, then server first can inquire about all voice datas corresponding to singer A from the voice data that this locality stores, and inquire about song B from all voice datas that singer A is corresponding, thus the voice data of singer A and song B will be met as the first voice data simultaneously.
Further, server, before acquisition first identifies, needs in advance for the local each voice data stored generates the unique mark of an overall situation.Wherein, server is when for each voice data generating identification, can realize according to the audio-frequency information of each voice data, such as, if the audio-frequency information of the local voice data stored comprises the source of song title, singer's name and voice data, then server can generate the mark of each voice data according to the source of the song title of each voice data, singer's name and voice data.
About the type of the first voice data, the embodiment of the present invention does not do concrete restriction.During concrete enforcement, because user is when K sings, often need to realize according to certain accompaniment.Therefore, the first voice data comprises but is not limited to accompaniment data.
403, server sends the first voice data and the first mark to the first client.
In order to enable the user of the first client sing according to the first voice data, server sends the first voice data to the first client.Manage the second audio data that the first client makes according to the first voice data for the ease of follow-up, the first mark is also sent to the first client by server.
Wherein, before server sends the first voice data and the first mark to the first client; in order to ensure meeting the demand of the first client to various voice data; first should obtain various voice data in advance; and generate unique mark for each voice data got, guarantee the first voice data that the voice data got comprises the first client and asks.Voice data is obtained about server, and the voice data generation uniquely identified mode for getting, include but not limited to the following two kinds:
First kind of way: server obtains voice data from the audio data service device server cluster, and the voice data generation unique identification for getting from audio data service device.
Composition graphs 1, audio data service device is herein the first audio data service device, second audio data server and the 3rd audio data service device in server cluster.The specific implementation process of this kind of mode includes but not limited to that 4031 to step 4034 realizes as follows:
4031, server is every preset duration, obtains voice data from audio data service device.
About the concrete numerical value of preset duration, the embodiment of the present invention does not do concrete restriction.During concrete enforcement, can set as required.Such as, this preset duration can be one week, two weeks etc.Composition graphs 1, in embodiments of the present invention, server cluster comprises different audio data service devices, and each audio data service device can process and obtains dissimilar voice data.Therefore, for the ease of managing the voice data got from each audio data service device, server, when obtaining voice data from each audio data service device, can adopt the different time intervals.
In addition, server is for the ease of distinguishing never with the voice data that voice data server gets, in embodiments of the present invention, the voice data that server gets from the first audio data service device is defined as first kind voice data, and namely the first audio data service device can process and obtain first kind voice data; Server is defined as Equations of The Second Kind voice data from the voice data that second audio data server gets, and namely second audio data server can process and obtain Equations of The Second Kind voice data; Server is defined as the 3rd class audio frequency data from the voice data that the 3rd audio data service device gets, and namely the 3rd audio data service device can process and obtain the 3rd class audio frequency data.
Because different audio data service devices can process the voice data obtaining different tonequality, therefore, the tonequality of first kind voice data, Equations of The Second Kind voice data and the 3rd class audio frequency data is different.Particularly, in embodiments of the present invention, the tonequality of first kind voice data is higher than appointment tonequality standard; The tonequality of Equations of The Second Kind voice data is lower than appointment tonequality standard; 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of the 3rd class audio frequency data is lower than the tonequality of Equations of The Second Kind voice data.Wherein, appointment tonequality standard can be the numerical value of the tonequality for identifying different voice data that industry is general.
In conjunction with the explanation of the above-mentioned tonequality to first kind voice data, Equations of The Second Kind voice data and the 3rd class audio frequency data, first kind voice data can referred to as good quality audio data, Equations of The Second Kind voice data can referred to as ordinary audio data, and the 3rd class audio frequency data can referred to as noise reduction voice data.
In addition, in order to more easily distinguish first kind voice data, Equations of The Second Kind voice data and the 3rd class audio frequency data, other adeditive attribute can also be set respectively for first kind voice data, Equations of The Second Kind voice data and the 3rd class audio frequency data.Such as, first kind voice data is made by professional and has carried out quality testing, and each first kind voice data has all bound the lyrics.Equations of The Second Kind voice data is the voice data that volumes of searches is greater than the song of the first assign thresholds.3rd class audio frequency data are the voice data that volumes of searches is greater than the song of the second assign thresholds.Wherein, the first assign thresholds is less than the second assign thresholds, and as the first assign thresholds, to be the 1,000,000, second assign thresholds be 2,000,000 etc.
Wherein, server can perform following process when obtaining voice data from each audio data service device: every the first preset duration, obtain first kind voice data from the first audio data service device; Every the second preset duration, obtain Equations of The Second Kind voice data from second audio data server; Every the 3rd preset duration, obtain the 3rd class audio frequency data from the 3rd audio data service device.About the concrete numerical value of the first preset duration, the second preset duration and the 3rd preset duration, the embodiment of the present invention does not do concrete restriction equally.
As, server a bit can obtain first kind voice data from the first audio data service device on every Mondays morning; Equations of The Second Kind voice data is obtained two o'clock in the morning on every Mondays from second audio data server; The 3rd class audio frequency data etc. are obtained 3 o'clock mornings on every Mondays from the 3rd audio data service device.
Obtaining the mode of voice data about server from audio data service device, including but not limited to the Interface realization by being connected with audio data service device.Such as, server by the interface be connected with the first audio data service device, can obtain first kind voice data from the first audio data service device; By the interface be connected with second audio data server, obtain Equations of The Second Kind voice data from second audio data server; By the interface be connected with the 3rd audio data service device, obtain the 3rd class audio frequency data from the 3rd audio data service device.
4032, server is according to the audio-frequency information of the voice data got, and judges whether this locality has stored this voice data, if this locality does not store this voice data, then performs step 4033 and 4034; If this locality has stored the voice data got, then perform step 4035 to step 4037.
Wherein, audio-frequency information is can the information of identification audio data characteristics, such as, industry is when making voice data, usually be directed to each voice data, a cryptographic Hash can be generated, the different voice datas of same type can be distinguished according to cryptographic Hash, therefore, audio-frequency information can be the cryptographic Hash of voice data.In addition, audio-frequency information can also be the song title, Ge Shouming, size of data, performance duration etc. that voice data is corresponding.
Unnecessary memory space is taken because storing repeat tone audio data in order to avoid server is local, server is after getting voice data from audio data service device, according to the audio-frequency information of the voice data got, judge whether this locality has stored this voice data.Wherein, server is at the audio-frequency information according to the voice data got, when judging whether this locality has stored this voice data, include but not limited to: according to the cryptographic Hash of the voice data got, the cryptographic Hash of the local voice data stored of inquiry, if the cryptographic Hash that there is the voice data got in the cryptographic Hash of voice data that stored of this locality, then determines that this locality has stored this voice data, otherwise, determine that this locality does not store this voice data.
4033, this voice data is stored.
When this locality do not store get voice data time, this voice data got of server stores.Wherein, server, when storing the voice data got, can be realized by the form of list, namely be directed to each voice data, a store items in corresponding lists.Therefore, server, when storing this voice data, can realize by adding a new store items at list end.
In addition, due in embodiments of the present invention, server is dissimilar voice data from the voice data that different audio data service devices gets, for the ease of distinguishing dissimilar voice data, server, when storing the voice data got, can distribute different memory spaces for the voice data of every type.Such as, for first kind voice data distributes the first memory space, for Equations of The Second Kind voice data distributes the second memory space, it is the 3rd class audio frequency data allocations the 3rd memory space.Now, after getting voice data, server can store this voice data according to the type of the voice data got.Such as, if this voice data got is first kind voice data, then this voice data is stored in pre-assigned first memory space by server.
Wherein, server when the type stores audio data according to the voice data that gets, because dissimilar audio data source is in different audio data service devices, therefore, server first according to the source of this voice data, can determine the type of this voice data.Such as, if this audio data source is in the first audio data service device, then determine that this voice data is first kind voice data.
4034, server is the audio identification that this voice data generates for this voice data of unique identification.
About the mode that server is this voice data generation audio identification, the embodiment of the present invention does not do concrete restriction, ensures that this mark uniquely can distinguish this voice data and other voice data.
Above-mentioned steps 4031 to step 4034 has been carried out explanation for the local situation not storing the voice data got of server and has been illustrated.Further, when having stored the arbitrary voice data in the voice data got for server this locality, the processing mode of server has referred to following content.For convenience of explanation, this sentences the voice data that server gets and comprises the first voice data, and server this locality has stored the first voice data is that example is described.
Particularly, due to for dissimilar voice data, its tonequality is different, and therefore, when the voice data that server gets is dissimilar voice data, the processing mode of server is not identical, refers to following step 4035 to step 4037.
4035, when the voice data that server gets is first kind voice data, if server is according to the audio-frequency information of the first voice data, determine that the first kind voice data got comprises the first voice data, then server uses the first voice data in the first kind voice data got to replace the local first kind voice data stored.
Due in embodiments of the present invention, server is determined that the first audio data service device can process in real time and is obtained tonequality higher than the first kind voice data of specifying tonequality standard, namely server determines that the first kind voice data that the first audio data service device process obtains is top quality voice data, therefore, in order to continue to optimize the tonequality of the local voice data stored, server uses the first voice data in the first kind voice data got to replace the local first kind voice data stored.
4036, when the voice data that server gets is Equations of The Second Kind voice data, if server is according to the audio-frequency information of the first voice data, after determining that the Equations of The Second Kind voice data got comprises the first voice data, server judges whether local the first voice data stored derives from the 3rd audio data service device further; If local the first audio data source stored is in the 3rd audio data service device, then the first voice data in the Equations of The Second Kind voice data got is used to replace local the first voice data stored; If local the first voice data stored does not derive from the 3rd audio data service device, then ignore the first voice data of the Equations of The Second Kind voice data kind got.
Particularly, because the tonequality of Equations of The Second Kind voice data is higher than the tonequality of the 3rd class audio frequency data, and the 3rd class audio frequency data are obtained by the 3rd audio data service device process, namely the 3rd class audio frequency is data from the 3rd audio data service device, therefore, if local the first audio data source stored is in the 3rd audio data service device, the tonequality of the first voice data then in order to Optimized Service device stores, the first voice data that server uses the first voice data in the Equations of The Second Kind voice data got to replace server this locality to store.If the first voice data that server this locality stores does not derive from the 3rd audio data service device, then in order to avoid this first voice data of repeated storage, server ignores the first voice data in the Equations of The Second Kind voice data got.
Such as, if the first voice data that server gets is the voice data of the song A got from second audio data server, the audio data source of the song A that server this locality stores is in the 3rd audio data service device, voice data due to the song A got from second audio data server is Equations of The Second Kind voice data, its tonequality is higher than the voice data got from the 3rd audio data service device, therefore, server uses the voice data of the song A got to replace the voice data of the local song A stored.
Wherein, server, when using the first voice data in the Equations of The Second Kind voice data got to replace the first voice data of server this locality storage, can use the first voice data that first voice data covering server this locality in the Equations of The Second Kind voice data got stores.Server, when ignoring the first voice data in the Equations of The Second Kind voice data got, directly can delete the first voice data in this Equations of The Second Kind voice data got.
If this locality has stored the first voice data, then by ignoring the first voice data in this Equations of The Second Kind voice data got, the unnecessary memory space taking server because of repeated storage first voice data can be avoided, make it possible to the memory space of Optimized Service device.
4037, when the voice data that server gets is the 3rd class audio frequency data, if server is according to the audio-frequency information of the first voice data, determine that the 3rd class audio frequency data got comprise the first voice data, then server ignores the first voice data in the 3rd class audio frequency data got.
If this locality has stored the first voice data, then by ignoring the first voice data in these the 3rd class audio frequency data got, the unnecessary memory space taking server because of repeated storage first voice data can be avoided, make it possible to the memory space of Optimized Service device.Wherein, server, when ignoring the first voice data in the 3rd class audio frequency data got, also directly can delete the first voice data in these the 3rd class audio frequency data got.
It should be noted that, when the voice data that the step mark of above-mentioned steps 4035 to step 4037 only gets for identification server is from different audio data service devices, when having stored the first voice data for server this locality, server is to the different disposal situation of the first voice data got, and this step identifies and is not used in the sequencing limited between step.
In addition, server, after getting each voice data from each audio data service device, regularly can also upgrade the audio-frequency information of each voice data.Such as, if audio-frequency information comprises lyrics mark, then server every the 4th preset duration, can upgrade the lyrics mark of each voice data, whether has bound the lyrics with the voice data that identification server this locality stores.Whether binding the lyrics by identifying the local voice data stored, when making to return the first voice data to the first client, mark first voice data can have been carried and whether bound the information of the lyrics.Now, if the user of the first client needs the first voice data having bound the lyrics, and the lyrics mark of the first voice data that server returns shows that this first voice data does not bind the lyrics, then the first client can not select this first voice data, thus avoid the first client to waste data on flows because of the first voice data of having downloaded this and not meeting the demands, thus the flow rate of the first client can be saved, bring good experience to user.
The second way: server receives the 4th class audio frequency data that the second client sends, using the 4th class audio frequency data of reception as the voice data got.Wherein, the 4th class audio frequency data are obtained from third-party server search by the second client.
Particularly, the 4th class audio frequency data are uploaded to third-party server after being obtained by user's making of other client.After the second client searches the 4th class audio frequency data from third-party server, the 4th class audio frequency data can be committed to server.
Because the user making of the 4th class audio frequency data by each client obtains, therefore, the tonequality of the 4th class audio frequency data is unstable, may very well, and also possibility is very bad.In addition, the user of each client, when making the 4th class audio frequency data, can also add individual element according to hobby, e.g., add DJ (Disk Jockey, Disc Jocker) element etc. in the 4th class audio frequency data in the 4th class audio frequency data.In addition, because the 4th class audio frequency data may have a lot of different user to obtain in corresponding client making, therefore, the abundant species of the song that the 4th class audio frequency data may comprise.Such as, the voice data of song C is not all comprised in first kind voice data, Equations of The Second Kind voice data and the 3rd class audio frequency data, then some clients can make a voice data for this song C, and namely the voice data of the song C that this client makes belongs to the 4th class audio frequency data.
The processing mode after voice data is got from audio data service device in conjunction with above-mentioned server, after server receives the 4th class audio frequency data, also need the audio-frequency information of each voice data comprised according to the 4th class audio frequency data, judge the local voice data whether having stored the 4th class audio frequency data and comprised; If server this locality does not store any one voice data that the 4th class audio frequency data comprise, then store all voice datas that the 4th class audio frequency data comprise, and all voice datas comprised for the 4th class audio frequency data generate unique audio identification.The principle of this process is consistent to the principle of step 4034 with step 4031, specifically see the content of above-mentioned steps 4031 to step 4034, can not repeat them here.
Further, when server determines arbitrary voice data that local the 4th class audio frequency data having stored reception comprise, if server is according to the audio-frequency information of the first voice data, judge that server this locality has stored this first voice data, then owing to being directed to the voice data of same song, each client is when making the voice data of this song, make the voice data tonequality that obtains and/or audio-frequency information may and different, therefore, server needs the content of the tonequality of the first voice data and audio-frequency information in the 4th class audio frequency data received and the local tonequality of the first voice data that stores and the content of audio-frequency information to be compared further.After the tonequality of the first voice data determined in the 4th class audio frequency data received when server is better than the tonequality of local the first voice data stored, the first voice data in the 4th class audio frequency data received is used to upgrade local the first voice data stored.
Such as, be directed to song C, if the tonequality of the voice data of the song C that server this locality stores is lower than the tonequality of the voice data of song C in the 4th class audio frequency data received, then server uses the voice data of song C in the 4th class audio frequency data received to upgrade the voice data of the local song C stored.
In addition, if it is more than the audio-frequency information of local the first voice data stored that server determines the content of the audio-frequency information of the first voice data in the 4th class audio frequency data received, then server can use the audio-frequency information of the first voice data in the 4th class audio frequency data of reception to upgrade the audio-frequency information of local the first voice data stored.
Such as, if the first voice data in the 4th class audio frequency data is the voice data of song C, and the audio-frequency information of the song C received comprises the lyrics of song C, and the audio-frequency information of the local song C stored does not comprise the lyrics of song C, then the lyrics of song C can be added in the audio-frequency information of the local song C stored by server, thus realize the audio-frequency information upgrading the local song C stored.
To sum up, by receiving the 4th class audio frequency data from the second client, and use the first voice data in the 4th class audio frequency data to upgrade local the first voice data stored, or upgrade the audio-frequency information of local the first voice data stored, or upgrade the first voice data and the audio-frequency information of local storage simultaneously, make server when upgrading the first voice data and the audio-frequency information thereof of local storage, tonequality and the audio-frequency information of the voice data can submitted to according to the second client realize, thus local the first voice data stored can be continued to optimize, when guaranteeing first this first voice data of user end to server request, server can provide optimum voice data for it.
404, after the first voice data and first that the first client reception server returns identifies, according to the 3rd voice data, the first voice data and the first mark that user records, generate second audio data, and second audio data is sent to server.
After the first voice data that first client reception server returns and first identifies, this first voice data and the first mark can be shown.Wherein, the first voice data, when display the first voice data, can be presented to user with Protobuf data transmission format by the first client.
Particularly, the user of the first client is according to the first voice data, the song that this first voice data is corresponding can be sung, the 3rd voice data that first client recording user generates according to this first voice data, and by the 3rd voice data, the first voice data and the first mark synthesis second audio data.Further, for the ease of the second audio data that server admin generates, this second audio data is sent to server by the first client.
Such as, if this first voice data is the accompaniment data of song A, then the user of the first client can to give song recitals A according to this accompaniment data, because accompaniment data and first identifies corresponding, therefore, the second audio data comprising accompaniment data, the 3rd voice data and the first mark can be obtained.
405, after server receives the second audio data of the first client transmission, according to the first mark, corresponding stored second audio data.
The client of carrying out singing due to the first voice data selecting server to provide has a lot of, when the user of different clients selects the first voice data to sing, for this first voice data, a client can make and obtain a second audio data.For the ease of managing the second audio data of each client according to the first sound data creating, server identifies corresponding stored second audio data according to first, under being all stored in this first mark by each client according to the second audio data that the first voice data makes.
On this basis, if user can obtain the second audio data of other users, then can trigger this user and another user and carry out K and sing and compare with, thus increase the K song enthusiasm of these two users.Therefore, under different clients is all stored in this first mark according to the second audio data that the first voice data generates, each client follow-up can search the second audio data that other users generate easily, thus the K between activated user sings passion competition, and increase the customer volume that K sings software further.In addition, compared with by K song, can interaction between adding users, if the type of the voice data of different user hobby is similar, user then can be made more to get to know other users with identical hobby, thus user is reached with the object of singing party friend, and then improve Consumer's Experience.
Above-mentioned steps 401 to step 405 is with the first client when obtaining voice data, and the voice data returned by server is example as the first voice data, has carried out detailed explanation to the management method of the voice data that the embodiment of the present invention provides.Alternatively, in step 401, the first client, when sending audio search requests to server, can also send audio search requests, to obtain the voice data of needs from third-party server to third-party server simultaneously.When server sends audio search requests to server in step 401, have sent audio search requests to third-party server, the method that server carries out audio data management is as follows simultaneously:
First client sends audio search requests to third-party server, and after third-party server receives this audio search requests, third-party server searches for its audio database according to the audio keyword in audio search requests, obtains the voice data meeting audio search requests.Because third-party server may obtain multiple voice data meeting search condition according to this audio search requests, in embodiments of the present invention, the voice data meeting audio search requests searched by third-party server is defined as at least one the 4th voice data.Wherein, the first client see the content in above-mentioned steps 401, can not repeat them here to the mode that third-party server sends the mode of audio search requests, third-party server obtains at least one the 4th voice data.
After the first client sends audio search requests to server and third-party server simultaneously, because server and third-party server can respond the audio search requests of the first client simultaneously, therefore, server returns the first voice data to the first client, and third-party server returns at least one the 4th voice data to the first client.Now, the first voice data that the first client may select server to return, also may select a voice data as the first voice data from least one the 4th voice data that third-party server returns.On this basis, the first client obtains the mode of the first voice data and the first mark, can have following several situation:
The first situation: the first client is when detecting that the first voice data that user selects server to return is as the first voice data got, and the first mark returned when server being returned the first voice data identifies as get first.
The second situation: the first client is after detecting that user selects a voice data from least one the 4th voice data, and voice data user selected is as the first voice data got.Further, in order to obtain the first mark, the first voice data selected is committed to server by the first client; Server is, after the first voice data of this selection generates the first mark, return the first mark to the first client; First client reception server is the first mark that the first voice data of this selection generates, using this first mark as the first mark got.
Wherein, after the first client detects that user selects a voice data at least one the 4th voice data as the first voice data, by the voice data of this selection is committed to server, not only so that server expands the content of audio database, and can guarantee that server returns the first mark for this first voice data of unique identification.
Server is when returning the first mark, if server this locality has stored this first voice data, server can directly return the first mark; If server does not store this first voice data, then server needs, first for after this first voice data generation first mark, to return this first mark.
The third situation: the first client have received the first voice data that server returns and at least one the 4th voice data that third-party server returns simultaneously.Now, after the first client detects the first voice data that user selects server to return, the first mark returned when server being returned the first voice data is as the first mark got; After the first client detects that user selects a voice data at least one the 4th voice data as the first voice data, the first voice data selected is committed to server by the first client, be after the first voice data generation first mark selected by server, first client reception server is the first mark that the first voice data selected generates, using the receive first mark as the first mark got.
Alternatively, when after the first client at least one the 4th voice data that simultaneously the first voice data of returning of reception server and third-party server return, identical voice data may be there is at least one the 4th voice data that the first voice data returned due to server and third-party server return, in order to avoid the voice data that display repeats, first client further according to the audio-frequency information of the first voice data and the audio-frequency information of at least one the 4th voice data, can also determine whether there is the first voice data at least one the 4th voice data; If there is the first voice data at least one the 4th voice data, then the first client can delete the first voice data that server returns from least one the 4th voice data, obtain the 4th voice data after upgrading, and the first voice data of returning of display server and at least one the 4th voice data after upgrading.
Wherein, first client is according to the audio-frequency information of the audio-frequency information of the first voice data and at least one the 4th voice data, when determining whether there is the first voice data at least one the 4th voice data, can be realized by the cryptographic Hash of the first voice data and at least one the 4th voice data.If the cryptographic Hash that there is a voice data at least one the 4th voice data is identical with the cryptographic Hash of the first voice data, then determine to there is the first voice data at least one the 4th voice data; If there is not the cryptographic Hash identical with the cryptographic Hash of the first voice data at least one the 4th voice data, then the first client determines there is not the first voice data at least one the 4th voice data.
Further, first client is when display first voice data and at least one the 4th voice data after upgrading, because the first voice data and at least one the 4th voice data after upgrading are the voice data got from different server, and the first voice data that server returns may not be identical with the type of at least one the 4th voice data after renewal, as, the first voice data that server returns is Equations of The Second Kind voice data, and at least one the 4th voice data after upgrading is the 4th class volume data, therefore, first client can according to the first voice data and at least one the 4th voice data different classes of after upgrading, the first voice data returned server in the zones of different of display screen and at least one the 4th voice data after upgrading show.
Such as, display screen can be divided into upper and lower two regions by the first client, thus when the first voice data that display server returns and at least one the 4th voice data after upgrading, first voice data that can return at the upper half area display server of display screen, at least one the 4th voice data after second region display update of display.
First client is by the first voice data of returning at the different viewing areas display server of display screen and at least one the 4th voice data after upgrading, make user when selecting voice data, can the obvious voice data that returns of the voice data that returns of Differentiated Services device and third-party server, not only make the voice data of display very clear, and the selection mode being supplied to user is more diversified, thus good operating experience can be brought to user.
The method that the embodiment of the present invention provides, returned the first mark of the first voice data and this first voice data to the first client by server, guarantee after the first user end to server submits second audio data to, server can by the first mark and second audio data corresponding stored.Because server is when managing voice data, is be designated reference with first, thus makes audio data management Method compare orderly.In addition, when user will search for the second audio data of other users according to same first voice data generation, during to carry out the competition of K song, server can be searched for according to the first mark, search procedure is compared and saves time.
Fig. 5 is the structural representation of a kind of server provided according to an exemplary embodiment, and this server is for performing the function in embodiment corresponding to above-mentioned Fig. 2 to Fig. 4 in any embodiment performed by server.See Fig. 5, this server comprises:
Receiver module 501, for receiving the audio search requests that the first client sends, wherein, audio search requests carries audio keyword;
First acquisition module 502, for according to audio keyword, obtain the first voice data and the first mark for unique identification first voice data, wherein, the audio-frequency information of the first voice data comprises audio keyword;
Sending module 503, for sending the first voice data and the first mark to the first client;
First memory module 504, for when the second audio data that reception first client sends, according to the first mark, corresponding stored second audio data, wherein, second audio data is recorded according to user by the first client the 3rd voice data, the first voice data and the first mark generate.
The server that the embodiment of the present invention provides, by returning the first mark of the first voice data and this first voice data to the first client, guarantee after the first client submits second audio data to, server can by the first mark and second audio data corresponding stored.Because server is when managing voice data, is be designated reference with first, thus makes audio data management Method compare orderly.In addition, when user will search for the second audio data of other users according to same first voice data generation to carry out the competition of K song, can search for according to the first mark, search procedure be compared and saves time.
In another embodiment, server also comprises:
Second acquisition module, for obtaining voice data;
Judge module, for the audio-frequency information according to voice data, judges whether this locality stores voice data;
Second memory module, for when this locality does not store voice data, stores audio data, voice data comprises the first voice data;
Generation module, for generating for voice data the audio identification being used for unique identification voice data.
In another embodiment, the second acquisition module comprises:
First acquiring unit, for every the first preset duration, obtains first kind voice data from the first audio data service device, and wherein, first kind voice data is that tonequality is higher than the voice data of specifying tonequality standard;
Second acquisition unit, for every the second preset duration, obtains Equations of The Second Kind voice data from second audio data server, and wherein, Equations of The Second Kind voice data is that tonequality is lower than the voice data of specifying tonequality standard;
3rd acquiring unit, for every the 3rd preset duration, obtain the 3rd class audio frequency data from the 3rd audio data service device, wherein, 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of the 3rd class audio frequency data is lower than the tonequality of Equations of The Second Kind voice data;
Receiving element, for receiving the 4th class audio frequency data that the second client sends, wherein, the 4th class audio frequency data are obtained from third-party server search by the second client.
In another embodiment, the second acquisition module also comprises:
First replacement unit, at the audio-frequency information according to the first voice data, after determining that first kind voice data comprises the first voice data, uses the first voice data in first kind voice data to replace local the first voice data stored;
Judging unit, at the audio-frequency information according to the first voice data, after determining that Equations of The Second Kind voice data comprises the first voice data, judges whether local the first voice data stored derives from the 3rd audio data service device;
Second replacement unit, for when local the first audio data source stored is in the 3rd audio data service device, uses the first voice data in Equations of The Second Kind voice data to replace local the first voice data stored;
First ignores unit, for when local the first voice data stored does not derive from the 3rd audio data service device, ignores the first voice data in Equations of The Second Kind voice data;
Second ignores unit, for when at the audio-frequency information according to the first voice data, after determining that this locality has stored the first voice data, ignores the first voice data in the 3rd class audio frequency data got;
First updating block, for at the audio-frequency information according to the first voice data, after determining that this locality has stored the first voice data, after determining that the tonequality of the first voice data in the 4th class audio frequency data is better than the tonequality of local the first voice data stored, upgrade local the first voice data stored;
Second updating block, for after the content of the audio-frequency information of the first voice data that stores more than this locality in the content of the audio-frequency information determining the first voice data in the 4th class audio frequency data, upgrades the audio-frequency information of local the first voice data stored.
In another embodiment, the first voice data is accompaniment data.
Above-mentioned all alternatives, can adopt and combine arbitrarily formation optional embodiment of the present invention, this is no longer going to repeat them.
Fig. 6 is the structural representation of a kind of server according to an exemplary embodiment.With reference to Fig. 6, server 600 comprises processing components 622, and it comprises one or more processor further, and the memory resource representated by memory 632, can such as, by the instruction of the execution of processing components 622, application program for storing.The application program stored in memory 632 can comprise each module corresponding to one group of instruction one or more.In addition, processing components 622 is configured to perform instruction, to perform the audio data management method that any embodiment in embodiment corresponding to above-mentioned Fig. 2 to Fig. 4 provides.
Server 600 can also comprise the power management that a power supply module 626 is configured to perform server 600, and a wired or wireless network interface 650 is configured to server 600 to be connected to network, and input and output (I/O) interface 658.Server 600 can operate the operating system based on being stored in memory 632, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or similar.
Wherein, more than one or one program is stored in memory, and is configured to be performed by more than one or one processor, and more than one or one program package is containing the instruction for carrying out following operation:
Receive the audio search requests that the first client sends, wherein, audio search requests carries audio keyword;
According to audio keyword, obtain the first voice data and the first mark for unique identification first voice data, wherein, the audio-frequency information of the first voice data comprises audio keyword;
The first voice data and the first mark is sent to the first client;
When the second audio data that reception first client sends, according to the first mark, corresponding stored second audio data, wherein, the 3rd voice data, the first voice data and the first mark that second audio data is recorded according to user by the first client generate.
Suppose that above-mentioned is the first possible execution mode, then, in the execution mode that the second provided based on the execution mode that the first is possible is possible, in the memory of server, also comprise the instruction for performing following operation:
Before sending the first voice data and the first mark to the first client, also comprise:
Obtain voice data;
According to the audio-frequency information of voice data, judge whether this locality stores voice data;
If this locality does not store voice data, then stores audio data, voice data comprises the first voice data;
For voice data generates the audio identification being used for unique identification voice data.
In the third the possible execution mode provided based on the execution mode that the second is possible, in the memory of server, also comprise the instruction for performing following operation:
Obtain voice data, comprising:
Every the first preset duration, obtain first kind voice data from the first audio data service device, wherein, first kind voice data is that tonequality is higher than the voice data of specifying tonequality standard;
Every the second preset duration, obtain Equations of The Second Kind voice data from second audio data server, wherein, Equations of The Second Kind voice data is that tonequality is lower than the voice data of specifying tonequality standard;
Every the 3rd preset duration, the 3rd class audio frequency data are obtained from the 3rd audio data service device, wherein, the 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of the 3rd class audio frequency data is lower than the tonequality of Equations of The Second Kind voice data;
Receive the 4th class audio frequency data that the second client sends, wherein, the 4th class audio frequency data are obtained from third-party server search by the second client.
In the 4th kind of possible execution mode provided based on the execution mode that the third is possible, in the memory of server, also comprise the instruction for performing following operation:
Every the first preset duration, after the first audio data service device obtains first kind voice data, also comprise:
At the audio-frequency information according to the first voice data, after determining that first kind voice data comprises the first voice data, the first voice data in first kind voice data is used to replace local the first voice data stored;
Every the second preset duration, after second audio data server obtains Equations of The Second Kind voice data, also comprise:
At the audio-frequency information according to the first voice data, after determining that Equations of The Second Kind voice data comprises the first voice data, judge whether local the first voice data stored derives from the 3rd audio data service device;
If local the first audio data source stored is in the 3rd audio data service device, then the first voice data in Equations of The Second Kind voice data is used to replace local the first voice data stored;
If local the first voice data stored does not derive from the 3rd audio data service device, then ignore the first voice data in Equations of The Second Kind voice data;
Every the 3rd preset duration, after the 3rd audio data service device obtains the 3rd class audio frequency data, also comprise:
At the audio-frequency information according to the first voice data, after determining that this locality has stored the first voice data, ignore the first voice data in the 3rd class audio frequency data got;
After receiving the 4th class audio frequency data of the second client transmission, also comprise:
At the audio-frequency information according to the first voice data, after determining that this locality has stored the first voice data, after determining that the tonequality of the first voice data in the 4th class audio frequency data is better than the tonequality of local the first voice data stored, upgrade local the first voice data stored;
After the content of the audio-frequency information of the first voice data stored more than this locality in the content of the audio-frequency information determining the first voice data in the 4th class audio frequency data, upgrade the audio-frequency information of local the first voice data stored.
In the 5th kind of possible execution mode that the first provides to any one the possible execution mode in the 4th kind of possible execution mode, in the memory of server, also comprise the instruction for performing following operation: the first voice data is accompaniment data.
The server that the embodiment of the present invention provides, by returning the first mark of the first voice data and this first voice data to the first client, guarantee after the first client submits second audio data to, server can by the first mark and second audio data corresponding stored.Because server is when managing voice data, is be designated reference with first, thus makes audio data management Method compare orderly.In addition, when user will search for the second audio data of other users according to same first voice data generation, during to carry out the competition of K song, can search for according to the first mark, search procedure be compared and saves time.
Fig. 7 is the structural representation of a kind of client provided according to an exemplary embodiment, and this client is for performing the function in embodiment corresponding to above-mentioned Fig. 2 to Fig. 4 in any embodiment performed by the first client.See Fig. 7, this client comprises:
First sending module 701, for sending audio search requests, wherein, audio search requests carries audio keyword;
Acquisition module 702, for obtaining the first voice data and the first mark for unique identification first voice data, wherein, the audio-frequency information of the first voice data comprises audio keyword;
Generation module 703, for the 3rd voice data, the first voice data and the first mark recorded according to user, generates second audio data;
Second sending module 704, for second audio data is sent to server, makes server according to the first mark, corresponding stored second audio data.
The client that the embodiment of the present invention provides, first mark of the first voice data returned by reception server and this first voice data, guarantee to sing according to the first voice data as user, after obtaining second audio data, submit second audio data to server, server can be identified first and second audio data corresponding stored.Because server is when managing voice data, be designated reference with first, thus make the way to manage of server to voice data more orderly.In addition, when user will search for the second audio data of other users according to same first voice data generation, during to carry out the competition of K song, server can be searched for according to the first mark, search procedure is compared and saves time.
In another embodiment, the first sending module 701, for sending audio search requests respectively to server and third-party server;
Acquisition module 702, for
The first voice data that reception server returns and server are the first mark that the first voice data generates;
Or, receive at least one the 4th voice data that third-party server returns; After detecting that user selects the first voice data from least one the 4th voice data, the first voice data selected is committed to server, makes server be that the first voice data generates the first mark; Reception server is the first mark that the first voice data generates;
Or the first voice data that reception server returns and server are the first mark that the first voice data generates; Receive at least one the 4th voice data that third-party server returns; After the first voice data that user selects server to return being detected, the first mark returned when server being returned the first voice data is as the first mark got; After detecting that user selects a voice data at least one the 4th voice data as the first voice data, the first voice data selected is committed to server, be after the first voice data generation first mark selected by server, reception server is the first mark that the first voice data selected generates, using the receive first mark as the first mark got.
In another embodiment, client also comprises:
Determination module, for the audio-frequency information of the first voice data that returns according to server and the audio-frequency information of at least one the 4th voice data, determines the first voice data that at least one the 4th voice data, whether presence server returns;
Removing module, for when the first voice data that presence server at least one the 4th voice data returns, deletes the first voice data that server returns from least one the 4th voice data, obtains at least one the 4th voice data after upgrading;
Display module, for showing the first voice data and at least one the 4th voice data after upgrading.
In another embodiment, display module, for according to the first voice data and at least one the 4th voice data different classes of after upgrading, shows the first voice data and at least one the 4th voice data after upgrading in the zones of different of display screen.
In another embodiment, the first voice data is accompaniment data.
Above-mentioned all alternatives, can adopt and combine arbitrarily formation optional embodiment of the present invention, this is no longer going to repeat them.
Please refer to Fig. 8, it illustrates the structural representation of the terminal involved by the embodiment of the present invention, this terminal can comprise the first client, and this first client may be used for implementing the audio data management method that any embodiment in embodiment corresponding to above-mentioned Fig. 2 to Fig. 4 provides.Specifically:
Terminal 800 can comprise RF (Radio Frequency, radio frequency) circuit 110, the memory 120 including one or more computer-readable recording mediums, input unit 130, display unit 140, transducer 150, voicefrequency circuit 160, WiFi (Wireless Fidelity, Wireless Fidelity) module 170, include the parts such as processor 180 and power supply 190 that more than or processes core.It will be understood by those skilled in the art that the restriction of the not structure paired terminal of the terminal structure shown in Fig. 8, the parts more more or less than diagram can be comprised, or combine some parts, or different parts are arranged.Wherein:
RF circuit 110 can be used for receiving and sending messages or in communication process, the reception of signal and transmission, especially, after being received by the downlink information of base station, transfer to more than one or one processor 180 to process; In addition, base station is sent to by relating to up data.Usually, RF circuit 110 includes but not limited to antenna, at least one amplifier, tuner, one or more oscillator, subscriber identity module (SIM) card, transceiver, coupler, LNA (Low Noise Amplifier, low noise amplifier), duplexer etc.In addition, RF circuit 110 can also by radio communication and network and other devices communicatings.Described radio communication can use arbitrary communication standard or agreement, include but not limited to GSM (Global System of Mobile communication, global system for mobile communications), GPRS (General Packet Radio Service, general packet radio service), CDMA (Code Division Multiple Access, code division multiple access), WCDMA (Wideband CodeDivision Multiple Access, Wideband Code Division Multiple Access (WCDMA)), LTE (Long Term Evolution, Long Term Evolution), Email, SMS (Short Messaging Service, Short Message Service) etc.
Memory 120 can be used for storing software program and module, and processor 180 is stored in software program and the module of memory 120 by running, thus performs the application of various function and data processing.Memory 120 mainly can comprise storage program district and store data field, and wherein, storage program district can storage operation system, application program (such as sound-playing function, image player function etc.) etc. needed at least one function; Store data field and can store the data (such as voice data, phone directory etc.) etc. created according to the use of terminal 800.In addition, memory 120 can comprise high-speed random access memory, can also comprise nonvolatile memory, such as at least one disk memory, flush memory device or other volatile solid-state parts.Correspondingly, memory 120 can also comprise Memory Controller, to provide the access of processor 180 and input unit 130 pairs of memories 120.
Input unit 130 can be used for the numeral or the character information that receive input, and produces and to arrange with user and function controls relevant keyboard, mouse, action bars, optics or trace ball signal and inputs.Particularly, input unit 130 can comprise Touch sensitive surface 131 and other input equipments 132.Touch sensitive surface 131, also referred to as touch display screen or Trackpad, user can be collected or neighbouring touch operation (such as user uses any applicable object or the operations of annex on Touch sensitive surface 131 or near Touch sensitive surface 131 such as finger, stylus) thereon, and drive corresponding jockey according to the formula preset.Optionally, Touch sensitive surface 131 can comprise touch detecting apparatus and touch controller two parts.Wherein, touch detecting apparatus detects the touch orientation of user, and detects the signal that touch operation brings, and sends signal to touch controller; Touch controller receives touch information from touch detecting apparatus, and converts it to contact coordinate, then gives processor 180, and the order that energy receiving processor 180 is sent also is performed.In addition, the polytypes such as resistance-type, condenser type, infrared ray and surface acoustic wave can be adopted to realize Touch sensitive surface 131.Except Touch sensitive surface 131, input unit 130 can also comprise other input equipments 132.Particularly, other input equipments 132 can include but not limited to one or more in physical keyboard, function key (such as volume control button, switch key etc.), trace ball, mouse, action bars etc.
Display unit 140 can be used for the various graphical user interface showing information or the information being supplied to user and the terminal 800 inputted by user, and these graphical user interface can be made up of figure, text, icon, video and its combination in any.Display unit 140 can comprise display floater 141, optionally, the form such as LCD (Liquid Crystal Display, liquid crystal display), OLED (Organic Light-Emitting Diode, Organic Light Emitting Diode) can be adopted to configure display floater 141.Further, Touch sensitive surface 131 can cover display floater 141, when Touch sensitive surface 131 detects thereon or after neighbouring touch operation, send processor 180 to determine the type of touch event, on display floater 141, provide corresponding vision to export with preprocessor 180 according to the type of touch event.Although in fig. 8, Touch sensitive surface 131 and display floater 141 be as two independently parts realize input and input function, in certain embodiments, can by Touch sensitive surface 131 and display floater 141 integrated and realize input and output function.
Terminal 800 also can comprise at least one transducer 150, such as optical sensor, motion sensor and other transducers.Particularly, optical sensor can comprise ambient light sensor and proximity transducer, and wherein, ambient light sensor the light and shade of environmentally light can regulate the brightness of display floater 141, proximity transducer when terminal 800 moves in one's ear, can cut out display floater 141 and/or backlight.As the one of motion sensor, Gravity accelerometer can detect the size of all directions (are generally three axles) acceleration, size and the direction of gravity can be detected time static, can be used for the application (such as horizontal/vertical screen switching, dependent game, magnetometer pose calibrating) of identification terminal attitude, Vibration identification correlation function (such as pedometer, knock) etc.; As for terminal 800 also other transducers such as configurable gyroscope, barometer, hygrometer, thermometer, infrared ray sensor, do not repeat them here.
Voicefrequency circuit 160, loud speaker 161, microphone 162 can provide the audio interface between user and terminal 800.Voicefrequency circuit 160 can by receive voice data conversion after the signal of telecommunication, be transferred to loud speaker 161, by loud speaker 161 be converted to voice signal export; On the other hand, the voice signal of collection is converted to the signal of telecommunication by microphone 162, voice data is converted to after being received by voicefrequency circuit 160, after again voice data output processor 180 being processed, through RF circuit 110 to send to such as another terminal, or export voice data to memory 120 to process further.Voicefrequency circuit 160 also may comprise earphone jack, to provide the communication of peripheral hardware earphone and terminal 800.
WiFi belongs to short range wireless transmission technology, and by WiFi module 170, terminal 800 can help that user sends and receive e-mail, browsing page and access streaming video etc., and its broadband internet wireless for user provides is accessed.Although Fig. 8 shows WiFi module 170, be understandable that, it does not belong to must forming of terminal 800, can omit in the scope of essence not changing invention as required completely.
Processor 180 is control centres of terminal 800, utilize the various piece of various interface and the whole terminal of connection, software program in memory 120 and/or module is stored in by running or performing, and call the data be stored in memory 120, perform various function and the deal with data of terminal 800, thus integral monitoring is carried out to terminal.Optionally, processor 180 can comprise one or more process core; Preferably, processor 180 accessible site application processor and modem processor, wherein, application processor mainly processes operating system, user interface and application program etc., and modem processor mainly processes radio communication.Be understandable that, above-mentioned modem processor also can not be integrated in processor 180.
Terminal 800 also comprises the power supply 190 (such as battery) of powering to all parts, preferably, power supply can be connected with processor 180 logic by power-supply management system, thus realizes the functions such as management charging, electric discharge and power managed by power-supply management system.Power supply 190 can also comprise one or more direct current or AC power, recharging system, power failure detection circuit, power supply changeover device or the random component such as inverter, power supply status indicator.
Although not shown, terminal 800 can also comprise camera, bluetooth module etc., does not repeat them here.Specifically in the present embodiment, the display unit of terminal is touch-screen display, and terminal also includes memory, and one or more than one program, one of them or more than one program are stored in memory, and are configured to be performed by more than one or one processor.Described more than one or one program package is containing the instruction for performing following operation:
Send audio search requests, wherein, audio search requests carries audio keyword;
Obtain the first voice data and the first mark for unique identification first voice data, wherein, the audio-frequency information of the first voice data comprises audio keyword;
According to the 3rd voice data, the first voice data and the first mark that user records, generate second audio data;
Second audio data is sent to server, makes server according to the first mark, corresponding stored second audio data.
Suppose that above-mentioned is the first possible execution mode, in the execution mode that the second then provided based on the execution mode that the first is possible is possible, the instruction for performing following operation is also comprised: send audio search requests, comprising in the memory of terminal:
Audio search requests is sent respectively to server and third-party server;
Obtain the first voice data and the first mark for unique identification first voice data, comprising:
The first voice data that reception server returns and server are the first mark that the first voice data generates;
Or, receive at least one the 4th voice data that third-party server returns; After detecting that user selects the first voice data from least one the 4th voice data, the first voice data selected is committed to server, makes server be that the first voice data generates the first mark; Reception server is the first mark that the first voice data generates;
Or the first voice data that reception server returns and server are the first mark that the first voice data generates; Receive at least one the 4th voice data that third-party server returns; After the first voice data that user selects server to return being detected, the first mark returned when server being returned the first voice data is as the first mark got; After detecting that user selects a voice data at least one the 4th voice data as the first voice data, the first voice data selected is committed to server, be after the first voice data generation first mark selected by server, reception server is the first mark that the first voice data selected generates, using the receive first mark as the first mark got.
In the third the possible execution mode provided based on the execution mode that the second is possible, the instruction for performing following operation is also comprised: after at least one the 4th voice data that reception third-party server returns, also comprise in the memory of terminal:
The audio-frequency information of the first voice data returned according to server and the audio-frequency information of at least one the 4th voice data, determine the first voice data that at least one the 4th voice data, whether presence server returns;
If the first voice data that at least one the 4th voice data, presence server returns, then from least one the 4th voice data, delete the first voice data that server returns, obtain at least one the 4th voice data after upgrading;
Show the first voice data and at least one the 4th voice data after upgrading.
In the 4th kind of possible execution mode provided based on the execution mode that the third is possible, in the memory of terminal, also comprise the instruction for performing following operation: show the first voice data and at least one the 4th voice data after upgrading, comprising:
According to the first voice data and at least one the 4th voice data different classes of after upgrading, in the zones of different of display screen, the first voice data and at least one the 4th voice data after upgrading are shown.
In the 5th kind of possible execution mode that the first provides to any one the possible execution mode in the 4th kind of possible execution mode, also comprise the instruction for performing following operation in the memory of terminal: the first voice data is accompaniment data.
The terminal that the embodiment of the present invention provides, first mark of the first voice data returned by reception server and this first voice data, guarantee to sing according to the first voice data as user, after obtaining second audio data, submit second audio data to server, server can be identified first and second audio data corresponding stored.Because server is when managing voice data, be designated reference with first, thus make the way to manage of server to voice data more orderly.In addition, when user will search for the second audio data of other users according to same first voice data generation, during to carry out the competition of K song, server can be searched for according to the first mark, search procedure is compared and saves time.
It should be noted that: the server that above-described embodiment provides and client are when performing audio data management method, only be illustrated with the division of above-mentioned each functional module, in practical application, can distribute as required and by above-mentioned functions and be completed by different functional modules, internal structure by equipment is divided into different functional modules, to complete all or part of function described above.In addition, the server that above-described embodiment provides and client and audio data management embodiment of the method belong to same design, and its specific implementation process refers to embodiment of the method, repeats no more here.
One of ordinary skill in the art will appreciate that all or part of step realizing above-described embodiment can have been come by hardware, the hardware that also can carry out instruction relevant by program completes, described program can be stored in a kind of computer-readable recording medium, the above-mentioned storage medium mentioned can be read-only memory, disk or CD etc.
The foregoing is only preferred embodiment of the present invention, not in order to limit the present invention, within the spirit and principles in the present invention all, any amendment done, equivalent replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (20)

1. an audio data management method, is characterized in that, described method comprises:
Receive the audio search requests that the first client sends, described audio search requests carries audio keyword;
According to described audio keyword, obtain the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
Described first voice data and described first mark is sent to described first client;
When receiving the second audio data that described first client sends, according to described first mark, second audio data described in corresponding stored, the 3rd voice data, described first voice data and described first mark that described second audio data is recorded according to user by described first client generate.
2. method according to claim 1, is characterized in that, described before described first client described first voice data of transmission and described first mark, also comprises:
Obtain voice data;
According to the audio-frequency information of described voice data, judge whether this locality has stored described voice data;
If this locality does not store described voice data, then store described voice data, described voice data comprises described first voice data;
For described voice data generates the audio identification being used for voice data described in unique identification.
3. method according to claim 2, is characterized in that, described acquisition voice data, comprising:
Every the first preset duration, obtain first kind voice data from the first audio data service device, described first kind voice data is that tonequality is higher than the voice data of specifying tonequality standard;
Every the second preset duration, obtain Equations of The Second Kind voice data from second audio data server, described Equations of The Second Kind voice data is the voice data of tonequality lower than described appointment tonequality standard;
Every the 3rd preset duration, the 3rd class audio frequency data are obtained from the 3rd audio data service device, described 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of described 3rd class audio frequency data is lower than the tonequality of described Equations of The Second Kind voice data;
Receive the 4th class audio frequency data that the second client sends, described 4th class audio frequency data are obtained from third-party server search by described second client.
4. method according to claim 3, is characterized in that, described every the first preset duration, after the first audio data service device obtains first kind voice data, also comprises:
At the audio-frequency information according to described first voice data, after determining that described first kind voice data comprises described first voice data, the first voice data in described first kind voice data is used to replace local the first voice data stored;
Described every the second preset duration, after second audio data server obtains Equations of The Second Kind voice data, also comprise:
At the audio-frequency information according to described first voice data, after determining that described Equations of The Second Kind voice data comprises described first voice data, judge whether local the first voice data stored derives from described 3rd audio data service device;
If local the first audio data source stored is in described 3rd audio data service device, then the first voice data in described Equations of The Second Kind voice data is used to replace local the first voice data stored;
If local the first voice data stored does not derive from described 3rd audio data service device, then ignore the first voice data in described Equations of The Second Kind voice data;
Described every the 3rd preset duration, after the 3rd audio data service device obtains the 3rd class audio frequency data, also comprise:
At the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, ignore the first voice data in the 3rd class audio frequency data got;
After the 4th class audio frequency data that described reception second client sends, also comprise:
At the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, after determining that the tonequality of the first voice data in described 4th class audio frequency data is better than the tonequality of local the first voice data stored, upgrade local the first voice data stored;
After the content of the audio-frequency information of the first voice data stored more than this locality in the content of the audio-frequency information determining the first voice data in described 4th class audio frequency data, upgrade the audio-frequency information of local the first voice data stored.
5. the method according to claim arbitrary in Claims 1-4, is characterized in that, described first voice data is accompaniment data.
6. an audio data management method, is characterized in that, described method comprises:
Send audio search requests, described audio search requests carries audio keyword;
Obtain the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
According to the 3rd voice data, described first voice data and described first mark that user records, generate second audio data;
Described second audio data is sent to server, makes described server according to described first mark, second audio data described in corresponding stored.
7. method according to claim 6, is characterized in that, described transmission audio search requests, comprising:
Audio search requests is sent respectively to described server and third-party server;
Described acquisition first voice data and the first mark for the first voice data described in unique identification, comprising:
Receive the first voice data that described server returns and described server is the first mark that described first voice data generates;
Or, receive at least one the 4th voice data that described third-party server returns; After detecting that user selects the first voice data from least one the 4th voice data described, the first voice data selected is committed to described server, makes described server be that described first voice data generates the first mark; Receiving described server is the first mark that described first voice data generates;
Or, receive the first voice data that described server returns and described server is the first mark that described first voice data generates; Receive at least one the 4th voice data that described third-party server returns; After the first voice data that user selects described server to return being detected, the first mark returned when described server is returned described first voice data is as the first mark got; After detecting that user selects a voice data at least one the 4th voice data described as the first voice data, the first voice data selected is committed to described server, after the first voice data being described selection by described server generates the first mark, receive the first mark that described server is the first voice data generation of described selection, using the receive first mark as the first mark got.
8. method according to claim 7, is characterized in that, after at least one the 4th voice data that the described third-party server of described reception returns, also comprises:
Whether the audio-frequency information of the first voice data returned according to described server and the audio-frequency information of at least one the 4th voice data described, exist the first voice data that described server returns at least one the 4th voice data described in determining;
If there is the first voice data that described server returns at least one the 4th voice data described, from at least one the 4th voice data described, then delete the first voice data that described server returns, obtain at least one the 4th voice data after upgrading;
Show at least one the 4th voice data after described first voice data and described renewal.
9. method according to claim 8, is characterized in that, at least one the 4th voice data after described first voice data of described display and described renewal, comprising:
Different classes of according at least one the 4th voice data after described first voice data and described renewal, shows at least one the 4th voice data after described first voice data and described renewal in the zones of different of display screen.
10. the method according to claim arbitrary in claim 6 to 9, is characterized in that, described first voice data is accompaniment data.
11. 1 kinds of servers, is characterized in that, described server comprises:
Receiver module, for receiving the audio search requests that the first client sends, described audio search requests carries audio keyword;
First acquisition module, for according to described audio keyword, obtain the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
Sending module, for sending described first voice data and described first mark to described first client;
First memory module, for when receiving the second audio data that described first client sends, according to described first mark, second audio data described in corresponding stored, the 3rd voice data, described first voice data and described first mark that described second audio data is recorded according to user by described first client generate.
12. servers according to claim 11, is characterized in that, described server also comprises:
Second acquisition module, for obtaining voice data;
Judge module, for the audio-frequency information according to described voice data, judges whether this locality has stored described voice data;
Second memory module, for when this locality does not store described voice data, stores described voice data, and described voice data comprises described first voice data;
Generation module, for generating for described voice data the audio identification being used for voice data described in unique identification.
13. servers according to claim 12, is characterized in that, described second acquisition module comprises:
First acquiring unit, for every the first preset duration, obtains first kind voice data from the first audio data service device, and described first kind voice data is that tonequality is higher than the voice data of specifying tonequality standard;
Second acquisition unit, for every the second preset duration, obtains Equations of The Second Kind voice data from second audio data server, and described Equations of The Second Kind voice data is the voice data of tonequality lower than described appointment tonequality standard;
3rd acquiring unit, for every the 3rd preset duration, the 3rd class audio frequency data are obtained from the 3rd audio data service device, described 3rd class audio frequency data are carry out to original audio the voice data that noise reduction process obtains, and the tonequality of described 3rd class audio frequency data is lower than the tonequality of described Equations of The Second Kind voice data;
Receiving element, for receiving the 4th class audio frequency data that the second client sends, described 4th class audio frequency data are obtained from third-party server search by described second client.
14. servers according to claim 13, is characterized in that, described second acquisition module also comprises:
First replacement unit, for at the audio-frequency information according to described first voice data, after determining that described first kind voice data comprises described first voice data, the first voice data in described first kind voice data is used to replace local the first voice data stored;
Judging unit, at the audio-frequency information according to described first voice data, after determining that described Equations of The Second Kind voice data comprises described first voice data, judges whether local the first voice data stored derives from described 3rd audio data service device;
Second replacement unit, for when local the first audio data source stored is in described 3rd audio data service device, uses the first voice data in described Equations of The Second Kind voice data to replace local the first voice data stored;
First ignores unit, for when local the first voice data stored does not derive from described 3rd audio data service device, ignores the first voice data in described Equations of The Second Kind voice data;
Second ignores unit, for when at the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, ignores the first voice data in the 3rd class audio frequency data got;
First updating block, for at the audio-frequency information according to described first voice data, after determining that this locality has stored described first voice data, after determining that the tonequality of the first voice data in described 4th class audio frequency data is better than the tonequality of local the first voice data stored, upgrade local the first voice data stored;
Second updating block, for after the content of the audio-frequency information of the first voice data that stores more than this locality in the content of the audio-frequency information determining the first voice data in described 4th class audio frequency data, upgrades the audio-frequency information of local the first voice data stored.
15. according to claim 11 to the server described in arbitrary claim in 14, and it is characterized in that, described first voice data is accompaniment data.
16. 1 kinds of clients, is characterized in that, described client comprises:
First sending module, for sending audio search requests, described audio search requests carries audio keyword;
Acquisition module, for obtaining the first voice data and the first mark for the first voice data described in unique identification, the audio-frequency information of described first voice data comprises described audio keyword;
Generation module, for the 3rd voice data recorded according to user, described first voice data and described first mark, generates second audio data;
Second sending module, for described second audio data is sent to server, makes described server according to described first mark, second audio data described in corresponding stored.
17. clients according to claim 16, is characterized in that, described first sending module, for sending audio search requests respectively to described server and third-party server;
Described acquisition module, for
Receive the first voice data that described server returns and described server is the first mark that described first voice data generates;
Or, receive at least one the 4th voice data that described third-party server returns; After detecting that user selects the first voice data from least one the 4th voice data described, the first voice data selected is committed to described server, makes described server be that described first voice data generates the first mark; Receiving described server is the first mark that described first voice data generates;
Or, receive the first voice data that described server returns and described server is the first mark that described first voice data generates; Receive at least one the 4th voice data that described third-party server returns; After the first voice data that user selects described server to return being detected, the first mark returned when described server is returned described first voice data is as the first mark got; After detecting that user selects a voice data at least one the 4th voice data described as the first voice data, the first voice data selected is committed to described server, after the first voice data being described selection by described server generates the first mark, receive the first mark that described server is the first voice data generation of described selection, using the receive first mark as the first mark got.
18. clients according to claim 17, is characterized in that, described client also comprises:
Determination module, for the audio-frequency information of the first voice data that returns according to described server and the audio-frequency information of at least one the 4th voice data described, determine whether to exist at least one the 4th voice data described the first voice data that described server returns;
Removing module, for when there is the first voice data that described server returns at least one the 4th voice data described, from at least one the 4th voice data described, delete the first voice data that described server returns, obtain at least one the 4th voice data after upgrading;
Display module, for showing at least one the 4th voice data after described first voice data and described renewal.
19. clients according to claim 18, it is characterized in that, described display module, different classes of for according at least one the 4th voice data after described first voice data and described renewal, shows at least one the 4th voice data after described first voice data and described renewal in the zones of different of display screen.
20. according to claim 16 to the client described in arbitrary claim in 19, and it is characterized in that, described first voice data is accompaniment data.
CN201410808946.0A 2014-12-19 2014-12-19 Audio data management method, server and client Active CN104572882B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410808946.0A CN104572882B (en) 2014-12-19 2014-12-19 Audio data management method, server and client

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410808946.0A CN104572882B (en) 2014-12-19 2014-12-19 Audio data management method, server and client

Publications (2)

Publication Number Publication Date
CN104572882A true CN104572882A (en) 2015-04-29
CN104572882B CN104572882B (en) 2019-03-26

Family

ID=53088944

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410808946.0A Active CN104572882B (en) 2014-12-19 2014-12-19 Audio data management method, server and client

Country Status (1)

Country Link
CN (1) CN104572882B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111309963A (en) * 2020-01-22 2020-06-19 百度在线网络技术(北京)有限公司 Audio file processing method and device, electronic equipment and readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101980198A (en) * 2010-11-01 2011-02-23 福州星网视易信息系统有限公司 Method for carrying karaoke
CN103400592A (en) * 2013-07-30 2013-11-20 北京小米科技有限责任公司 Recording method, playing method, device, terminal and system
CN103902728A (en) * 2014-04-14 2014-07-02 北京君正集成电路股份有限公司 Method and device for storing voice signals of intelligent watch
CN104157292A (en) * 2014-08-20 2014-11-19 杭州华为数字技术有限公司 Anti-howling audio signal processing method and device thereof
CN104168433A (en) * 2014-08-28 2014-11-26 广州华多网络科技有限公司 Media content processing method and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101980198A (en) * 2010-11-01 2011-02-23 福州星网视易信息系统有限公司 Method for carrying karaoke
CN103400592A (en) * 2013-07-30 2013-11-20 北京小米科技有限责任公司 Recording method, playing method, device, terminal and system
CN103902728A (en) * 2014-04-14 2014-07-02 北京君正集成电路股份有限公司 Method and device for storing voice signals of intelligent watch
CN104157292A (en) * 2014-08-20 2014-11-19 杭州华为数字技术有限公司 Anti-howling audio signal processing method and device thereof
CN104168433A (en) * 2014-08-28 2014-11-26 广州华多网络科技有限公司 Media content processing method and system

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111309963A (en) * 2020-01-22 2020-06-19 百度在线网络技术(北京)有限公司 Audio file processing method and device, electronic equipment and readable storage medium

Also Published As

Publication number Publication date
CN104572882B (en) 2019-03-26

Similar Documents

Publication Publication Date Title
CN106210755B (en) A kind of methods, devices and systems playing live video
CN105554522A (en) Method for playing audio in group, server and terminal
CN103596017B (en) Video downloading method and system
CN106231433B (en) A kind of methods, devices and systems playing network video
CN103678605B (en) A kind of method of information transfer, device and terminal device
CN105187733A (en) Video processing method, device and terminal
CN108184143B (en) Method and device for acquiring resources
CN103475914A (en) Video playing method, video playing device, terminal equipment and server
CN106131930A (en) A kind of WiFi network connection control method, device and terminal
CN104519404A (en) Graphics interchange format file playing method and device
CN105606117A (en) Navigation prompting method and navigation prompting apparatus
CN103294442B (en) A kind of method of playing alert tones, device and terminal device
CN103945241A (en) Streaming data statistical method, system and related device
CN106210919A (en) A kind of main broadcaster of broadcasting sings the methods, devices and systems of video
CN104735132A (en) Information inquiry method, servers and terminal
CN108388451A (en) Method for starting up mobile terminal, device and mobile terminal
CN103607431B (en) Mobile terminal resource processing method, device and equipment
CN105207880A (en) Group recommendation method and apparatus
CN105047185B (en) A kind of methods, devices and systems obtaining audio accompaniment
CN104539571A (en) Information interaction method, identity authentication method, server and terminal
CN105530239A (en) Multimedia data obtaining method and device
CN105185401A (en) Method and device for synchronizing multi-media file list
CN105303591A (en) Method for superposing place information in puzzle, terminal and server
CN105550316A (en) Pushing method and device of audio list
CN105302804A (en) Service bill display method, terminal, and server

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 510660 Guangzhou City, Guangzhou, Guangdong, Whampoa Avenue, No. 315, self - made 1-17

Applicant after: Guangzhou KuGou Networks Co., Ltd.

Address before: 510000 B1, building, No. 16, rhyme Road, Guangzhou, Guangdong, China 13F

Applicant before: Guangzhou KuGou Networks Co., Ltd.

GR01 Patent grant
GR01 Patent grant