WO2023160713A1 - 音乐生成方法、装置、设备、存储介质及程序 - Google Patents

音乐生成方法、装置、设备、存储介质及程序 Download PDF

Info

Publication number
WO2023160713A1
WO2023160713A1 PCT/CN2023/078545 CN2023078545W WO2023160713A1 WO 2023160713 A1 WO2023160713 A1 WO 2023160713A1 CN 2023078545 W CN2023078545 W CN 2023078545W WO 2023160713 A1 WO2023160713 A1 WO 2023160713A1
Authority
WO
WIPO (PCT)
Prior art keywords
track
period
multimedia
music
audio
Prior art date
Application number
PCT/CN2023/078545
Other languages
English (en)
French (fr)
Inventor
陈昊楠
冯成立
王维斯
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Publication of WO2023160713A1 publication Critical patent/WO2023160713A1/zh

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0025Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]

Definitions

  • the embodiments of the present disclosure relate to the technical field of the Internet, and in particular to a music generation method, device, device, storage medium, computer program product, and computer program.
  • terminal devices can be installed with karaoke software, and the background music of music works can be played through the karaoke software, and users can perform karaoke along with the background music to realize music creation or music performance.
  • Embodiments of the present disclosure provide a music generation method, device, device, storage medium, computer program product, and computer program to solve the problem of poor cooperation and interaction in the music creation/performance process.
  • an embodiment of the present disclosure provides a method for generating music, including:
  • the music template includes a plurality of tracks, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the multimedia segment corresponding to each period of each track is obtained in the database; wherein, the multimedia segment corresponding to each period of each track includes: the user plays the track melody corresponding to the period in the track
  • the obtained audio information the database includes a plurality of multimedia clips performed by a plurality of users;
  • the audio information in the multimedia segment corresponding to each period of each track is spliced respectively to obtain the track audio corresponding to each track;
  • the track audio corresponding to each track is merged to generate the target music.
  • an embodiment of the present disclosure provides a method for generating music, including:
  • each track is divided into at least one period, and the plurality of tracks are arrangement tracks of music
  • the first time period In response to the selection operation of the first time period, displaying the multimedia recording page corresponding to the first time period, the first time period being any one of the time periods corresponding to the plurality of tracks;
  • Record multimedia information within the recording period corresponding to the multimedia recording page obtain a first multimedia segment corresponding to the first period, and send the first multimedia segment to the server.
  • an embodiment of the present disclosure provides a music generating device, including:
  • a determining module configured to determine a music template, the music template including a plurality of tracks, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the obtaining module is used to obtain the multimedia segment corresponding to each period of each track in the database according to the music template; wherein, the multimedia segment corresponding to each period of each track includes: the user for the period in the track
  • the audio information obtained by performing the corresponding track melody, the database includes multiple multimedia clips played by multiple users;
  • the processing module is used to splicing the audio information in the multimedia clips corresponding to each time period of each track respectively to obtain the track audio corresponding to each track;
  • the generating module is configured to merge track audio corresponding to each track to generate target music.
  • an embodiment of the present disclosure provides a music generating device, including:
  • the display module is used to display a plurality of tracks corresponding to the music template, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the display module is further configured to display a multimedia recording page corresponding to the first time period in response to a selection operation of the first time period, where the first time period is any one of the time periods corresponding to the plurality of tracks;
  • a generating module configured to record multimedia information within a recording period corresponding to the multimedia recording page, to obtain a first multimedia segment corresponding to the first period;
  • a sending module configured to send the first multimedia segment to a server.
  • an embodiment of the present disclosure provides an electronic device, including: a processor and a memory;
  • the memory stores computer-executable instructions
  • the processor executes the computer-executed instructions to implement the music generation method in the first aspect and various possible implementations of the first aspect, or implement the second aspect and various possible implementations of the second aspect music generation method.
  • the embodiments of the present disclosure provide a computer-readable storage medium, the computer-readable storage medium stores computer-executable instructions, and when the processor executes the computer-executable instructions, the above first aspect and the first
  • the music generation method in various possible implementation manners of the aspect, or implement the second aspect and the music generation method in various possible implementation manners of the second aspect.
  • an embodiment of the present disclosure provides a computer program product, including a computer program.
  • the computer program When the computer program is executed by a processor, it implements the music generation method in the first aspect and in various possible implementation manners of the first aspect, Or, realize the music generating method in the second aspect and various possible implementation manners of the second aspect.
  • an embodiment of the present disclosure provides a computer program.
  • the computer program When the computer program is executed by a processor, it implements the music generation method in the first aspect and various possible implementation manners of the first aspect, or implements the method as in The second aspect and the music generation method in various possible implementation manners of the second aspect.
  • Embodiments of the present disclosure provide a music generation method, device, device, storage medium, computer program product and computer program.
  • the method comprises: determining a music template, the music template comprising a plurality of tracks, each track being divided into at least one period, and according to the music template, obtaining a multimedia segment corresponding to each period of each track in a database; wherein , the multimedia segment corresponding to each period of each track includes: the audio information obtained by the user playing the melody of the track corresponding to the period in the track, and the database includes multiple multimedia segments played by multiple users, respectively for each
  • the audio information in the multimedia segment corresponding to each period of each track is spliced to obtain the track audio corresponding to each track, and the track audio corresponding to each track is merged to generate the target music.
  • the target music generated by the server according to the multimedia segment in the database is generated by the cooperative performance of multiple users.
  • the target music is generated by adopting a cooperative performance method of multiple users, which enhances cooperation and interaction between users, increases the interest of users in creating music or playing music, and improves user experience.
  • FIG. 1 is a schematic diagram of music tracks and periods provided by an embodiment of the present disclosure
  • FIG. 2 is a schematic diagram of an application scenario provided by an embodiment of the present disclosure
  • FIG. 3 is a schematic flowchart of a music generation method provided by an embodiment of the present disclosure
  • FIG. 4 is a schematic diagram of a group of display pages provided by an embodiment of the present disclosure.
  • FIG. 5 is a schematic diagram of another group of display pages provided by an embodiment of the present disclosure.
  • FIG. 6 is a schematic diagram of another group of display pages provided by an embodiment of the present disclosure.
  • FIG. 7 is a schematic flowchart of another music generation method provided by an embodiment of the present disclosure.
  • FIG. 8 is a schematic diagram of another display page provided by an embodiment of the present disclosure.
  • FIG. 9 is a schematic diagram of a music generation process provided by an embodiment of the present disclosure.
  • FIG. 10 is a schematic diagram of a video template provided by an embodiment of the present disclosure.
  • FIG. 11 is a schematic structural diagram of a music generating device provided by an embodiment of the present disclosure.
  • FIG. 12 is a schematic structural diagram of another music generating device provided by an embodiment of the present disclosure.
  • FIG. 13 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the ways for users to create or play music usually include: listening to songs and karaoke.
  • the terminal device installs K song software, and plays the background music of the music work through the K song software, and the user performs K songs along with the background music.
  • the present disclosure provides a music generation method, device, device, storage medium and program, which can generate target music by means of cooperative performance by multiple users.
  • the manner in which multiple users cooperate to play means that the music template is divided into multiple parts, each user plays one or more parts, and the parts played by multiple users are combined to obtain the target music.
  • the music template may include multiple tracks.
  • the track in the embodiment of the present disclosure refers to the arrangement track of music.
  • the track type corresponding to each track may be one of melody track type, accompaniment track type and percussion track type.
  • the melody track refers to the track corresponding to the sound of singing lyrics
  • the accompaniment track can be the track corresponding to the sound of instruments such as guitar, piano, erhu, and violin
  • the percussion track can be the sound of percussion instruments such as drums and tambourines. the corresponding track.
  • the number of tracks included in the music template is not limited.
  • some music templates include a melody track, a piano accompaniment track, an erhu accompaniment track, and a drum kit percussion track.
  • Some music templates may include a melody track, a guitar accompaniment track, and a tambourine hit track.
  • Fig. 1 is a schematic diagram of music tracks and periods provided by an embodiment of the present disclosure.
  • the target music can be generated in the following way of cooperative performance: user 1 plays track A to get the multimedia information corresponding to track A, user 1 2 playing track B to obtain the multimedia information corresponding to track B, and user 3 playing track C to obtain the multimedia information corresponding to track C. Merge the multimedia information corresponding to the track A, track B, and track C to generate the target music.
  • the above-mentioned multimedia information includes audio information and/or video information.
  • each track can also be divided into multiple time periods from the time dimension. Each period corresponds to a certain duration, and there is no overlap between different periods. For example, in beat order, every 4-8 beats can be divided into a period, or in time order, every 10 seconds can be divided into a period.
  • track A is divided into three periods, namely period A1, period A2, and period A3; track B is divided into three periods, respectively, period B1, period B2, and period B3; track C is divided into three periods
  • the time periods are respectively time period C1, time period C2, and time period C3.
  • the target music can be generated by means of cooperative performance as follows: each user plays part of the period A1, period A2, period A3, period B1, period B2, period B3, period C1, period C2, and period C3.
  • user 1 plays period A1 to get the multimedia segment corresponding to period A1
  • user 2 plays period A2 to obtain the multimedia segment corresponding to period A2
  • user 3 plays period A3 to obtain the multimedia segment corresponding to period A3
  • user 4 plays period B1 to obtain the period
  • For the multimedia clip corresponding to B1 user 4 plays the time period B2 to obtain the multimedia fragment corresponding to time period B2
  • user 5 plays the time period B3 to obtain the multimedia fragment corresponding to time period B3
  • user 6 plays the time period C1 to obtain the multimedia fragment corresponding to time period C1
  • user 7 plays the time period C2
  • the multimedia segment corresponding to the period C2 is obtained
  • the user 8 plays the period C3 to obtain the multimedia segment corresponding to the period C3.
  • the multimedia fragments corresponding to the period A1, period A2 and period A3 are spliced to obtain the multimedia information corresponding to the track A; the multimedia fragments corresponding to the period B1, period B2 and period B3 are spliced to obtain the multimedia information corresponding to the track B Information; splicing the multimedia segments corresponding to the period C1, the period C2, and the period C3 to obtain the multimedia information corresponding to the track C. Furthermore, the above-mentioned multimedia information corresponding to Track A, Track B, and Track C are merged to generate target music.
  • FIG. 2 is a schematic diagram of an application scenario provided by an embodiment of the present disclosure.
  • the application scenario includes: a terminal device, a database, and a server.
  • the terminal device may be any electronic device with multimedia interaction function, including but not limited to: smart phone, tablet computer, notebook computer, smart TV, smart wearable device, smart home device, smart vehicle device, smart speaker, etc.
  • the server can be a cloud server, also known as a cloud computing server or a cloud host, which is a host product in the cloud computing service system.
  • the server can also be a server of a distributed system, or a server combined with a blockchain.
  • the music generating method provided in this embodiment may include two processes, namely: a process of generating a multimedia segment and a process of generating target music.
  • the generating process of the multimedia segment can be executed by the terminal device, and the generating process of the target music can be executed by the server.
  • the above two processes are respectively introduced below in conjunction with FIG. 1 .
  • multiple users may participate in the process of generating multimedia clips.
  • Each user can record one or more multimedia clips through the terminal device.
  • each user can choose to perform a certain period of time in a certain track, and record the performance process to obtain a multimedia clip.
  • the terminal device uploads the multimedia segment to the database.
  • the database is used to store multimedia segments corresponding to each time period in each track.
  • the server may acquire multimedia clips corresponding to each period in each track from the database, and merge these multimedia clips to generate the target music.
  • each multimedia clip stored in the database is obtained by playing by multiple users, therefore, the target music generated by the server according to the multimedia clip in the database is generated by cooperative performance of multiple users .
  • multiple users cooperate to play to generate target music, which enhances the cooperation and interaction between users, increases the interest of users in creating or playing music, and improves user experience.
  • it can also enrich the network user generated content (User Generated Content, UGC), and promote the popularization and dissemination of music performance.
  • UGC User Generated Content
  • Fig. 3 is a schematic flowchart of a method for generating music provided by an embodiment of the present disclosure.
  • the method in this embodiment can be executed by a terminal device.
  • This embodiment describes the process of generating a multimedia segment.
  • the method of this embodiment includes:
  • S301 Display multiple tracks of a music template, each track is divided into at least one period, and the multiple tracks are music arrangement tracks.
  • the music template may also be referred to as music to be played.
  • the music template has been divided into multiple tracks, and each track is divided into multiple time periods.
  • the terminal device may display a selection page for the user to select which period in which track to perform.
  • FIG. 4 is a schematic diagram of a group of display pages provided by an embodiment of the present disclosure.
  • the terminal device displays a page 401 , and multiple pieces of music are displayed on the page 401 .
  • the user can select one of the music to perform on the page 401 .
  • the music selected by the user is used as the music template.
  • the terminal device displays the selection page corresponding to "Music 1", that is, page 402, in which multiple tracks included in Music 1 are displayed, and each track is At least one period of division.
  • music 1 includes 3 tracks and each track includes 3 time periods as an example.
  • the user may not be able to accurately determine which part of the target music the time period corresponds to according to the time period identifier. Therefore, in some possible implementations, when the user selects an identifier of a certain time period, the description information corresponding to the time period may also be displayed on the page 402, for example, the lyrics, musical notes, time intervals, etc. corresponding to the time period are displayed. This embodiment does not limit the display position and display manner of the description information. In one example, when a marker for a time period is selected, the The identification of the time period is replaced and displayed as the description information corresponding to the time period. In this way, the user can accurately know the specific content to be played according to the description information of the time period, which is convenient for the user to make a selection.
  • the user can choose one of the periods to perform according to his/her will. For example, suppose the user wants to play the period A2 corresponding to track A, and the user selects the period A2 corresponding to track A and clicks OK After the control, the terminal device is triggered to execute S302 of this embodiment.
  • each track by dividing each track into multiple time periods, the user can select a time period to perform, which can reduce the difficulty of the user's performance.
  • the user can also select multiple time periods to play.
  • S302 In response to a selection operation on a first time period, display a multimedia recording page corresponding to the first time period, where the first time period is any one of the time periods corresponding to the plurality of tracks.
  • S303 Record multimedia information within a recording period corresponding to the multimedia recording page, obtain a first multimedia segment corresponding to the first period, and send the first multimedia segment to a server.
  • the time period selected by the user is referred to as the first time period.
  • the first time period may be any time period in the page 402 . For example, if the user selects time period A2 corresponding to track A on page 402 in FIG. 4 , then time period A2 corresponding to track A is the first time period.
  • the terminal device displays a multimedia recording page corresponding to the first time period, and the user can record multimedia information on the multimedia recording page.
  • the user may play the music content corresponding to the first time period.
  • the terminal device records his own performance process to obtain the first multimedia segment corresponding to the first time period.
  • FIG. 5 is a schematic diagram of another group of display pages provided by an embodiment of the present disclosure.
  • the terminal device may display a page 501 as shown in FIG. 5 .
  • Page 501 is the multimedia recording page. The user can record multimedia information on the page 501 to generate multimedia clips.
  • the multimedia recording page includes a start control and a stop control
  • the terminal device starts recording multimedia information in response to a click operation on the start control until the click operation on the stop control is received or At the end of the recording period, stop recording the multimedia information to obtain the first multimedia segment.
  • the multimedia recording page may further include an information prompt area.
  • the terminal device acquires the recording progress of the multimedia information, and displays prompt information in the information prompt area according to the recording progress, and the prompt information includes at least one of the following: lyrics, musical notes, and the like.
  • the dotted box on the page 501 marks the information prompt area.
  • lyrics indicated by XXX
  • musical notes indicated by a waveform composed of black rectangular frames
  • the user after the user clicks the start control on page 501, the user starts to perform the first period, and records the performance process through the audio collection device and/or video collection device of the terminal device.
  • the recording process is shown on page 502. During the recording process, information such as lyrics and notes to be played can be prompted in the information prompt area.
  • the first multimedia segment corresponding to the first time period is obtained.
  • the terminal device displays a page 503, on which the user can play back and watch the first multimedia segment, and determine whether to upload the first multimedia segment to the server.
  • the terminal device may also determine the similarity between the first multimedia segment and the original song, and determine the corresponding score of the first multimedia segment according to the similarity , the score can be displayed on the page 503 . In this way, the user can determine whether to perform re-recording according to the score. If the user clicks to re-record, the terminal device displays page 501 so that the user can start recording again. If the user clicks OK, the terminal device sends the first multimedia segment to the server.
  • the server after receiving the first multimedia segment, stores the first multimedia segment in the database.
  • the target music is generated by adopting multiple users to play cooperatively, which enhances the cooperation and interaction between users, increases the interest of users in creating or playing music, and improves user experience.
  • multiple users adopt an "asynchronous cooperation" method, which does not require multiple users to perform at the same time, reducing the difficulty of mutual cooperation.
  • it also avoids the problem that users cannot find someone to cooperate with.
  • a cappella is a cappella, sung a cappella, without the use of instruments.
  • candidate songs for example, songs with the right to adapt, or songs created by the user
  • the arranged song has multiple tracks, each in the form of a cappella. As the user plays each track, they perform it a cappella with a human voice. In this way, even if users do not know how to play musical instruments, they can participate in multi-person cooperation and reduce the difficulty of music performance.
  • a teaching video may be recorded in advance.
  • professionals can demonstrate the singing of this period, or professionals can explain the singing precautions of this period to help users master the correct pitch and rhythm. The process of using the teaching video will be described below with reference to FIG. 6 .
  • FIG. 6 is a schematic diagram of another group of display pages provided by an embodiment of the present disclosure.
  • the terminal device may first display page 601 as shown in FIG. 6 .
  • Page 601 is the teaching video page.
  • the user can click the start control to start playing the teaching video corresponding to the first period.
  • the teaching video page may include an information prompt area.
  • the information prompt information may display prompt information such as lyrics and characters. The playing process of the teaching video is shown in page 602,
  • the user by providing the user with a teaching video corresponding to each time period, the user can learn before playing, further reducing the difficulty of playing for the user.
  • the foregoing embodiments describe the process of generating a multimedia segment corresponding to one period of target music. It should be understood that after the multimedia segments corresponding to all time periods in the target music are generated, the multimedia segments corresponding to each time period may be merged to generate the target music. The following describes in detail how to combine multiple multimedia clips to generate target music in combination with several embodiments.
  • Fig. 7 is a schematic flowchart of another method for generating music provided by an embodiment of the present disclosure.
  • the method in this embodiment can be executed by a server.
  • This embodiment describes the generation process of target music.
  • the method of this embodiment includes:
  • S701 Determine a music template, where the music template includes multiple tracks, each track is divided into at least one period, and the multiple tracks are musical arrangement tracks.
  • the execution timing of this embodiment is when the server determines that the target music needs to be generated according to the music template, and there are two triggering methods as follows.
  • Method 1 The server automatically triggers.
  • the server detects that the multimedia segment corresponding to the music template has been collected, it triggers the execution of this embodiment.
  • the music template includes 3 tracks, track A, track B, and track C respectively
  • track A is divided into 3 time periods, namely time period A1, time period A2, and time period A3, and track B is divided into 3 time periods
  • Time period B1, time period B2, and time period B3 track C is divided into three time periods, namely time period C1, time period C2, and time period C3, then the above time period A1, time period A2, time period A3, time period B1, time period B2, and time period B3
  • the method of this embodiment may be executed to generate target music.
  • Method 2 Manually triggered by the user.
  • the terminal device may also display performance information corresponding to each time period, for example, whether the time period has been played. If it has already been played, it means that the multimedia segment corresponding to this time period already exists in the server. If the user determines that each period of the target music has been played according to the performance information of each period displayed on the page 401, the generation process of the target music can be manually triggered.
  • the terminal device detects that the user performs a music synthesis operation
  • the terminal device sends a music generation request to the server, and the server executes the method of this embodiment after receiving the music generation request.
  • the multimedia segment corresponding to each period of each track includes: the user corresponds to the track melody of the period in the track
  • the database includes multiple multimedia clips performed by multiple users.
  • each multimedia segment in the database may be generated by using the embodiment shown in FIG. 3 above.
  • Each multimedia segment stored in the database may be played and uploaded by multiple users. That is to say, the database includes multiple multimedia clips performed by multiple users. For example, there are 10 multimedia clips stored in the database, wherein a part of the multimedia clips is performed and uploaded by user 1, another part of the multimedia clips is played and uploaded by user 2, and another part of the multimedia clips is played and uploaded by user 3.
  • the multimedia fragments corresponding to each period of each track can be obtained respectively, so as to obtain the multimedia fragments required for generating the target music.
  • the multimedia segment corresponding to each period of each track includes: audio information obtained by the user playing the melody of the track corresponding to the period in the track.
  • audio information obtained by the user playing the melody of the track corresponding to the period in the track For how the user plays the track melody corresponding to each period of each track, refer to the relevant description of the foregoing embodiments, and details are not repeated here.
  • the database when the database stores multimedia clips, it can record the relationship among music templates, tracks, time periods and multimedia clips.
  • the content stored in the database is shown in Table 1.
  • the server can obtain from the database the multimedia clips needed to generate the target music according to the multiple tracks included in the music template and the time periods corresponding to each track.
  • a certain time period may be performed by multiple users, therefore, multiple multimedia clips corresponding to this time period may be stored in the database.
  • multiple multimedia clips corresponding to this time period may be stored in the database.
  • the server can determine a multimedia segment among the multiple multimedia segments to be used to generate the target music according to the attribute information of a plurality of multimedia segments corresponding to a time period.
  • the attribute information may include at least one of the following: performer information, rating, upload time, ranking, popularity value, and the like.
  • the server may select the multimedia fragment ranked first, or select the multimedia fragment with the highest popularity value, or select, the most recently uploaded multimedia fragment, Alternatively, select the multimedia segment with the highest score, or select a multimedia segment performed by a specified performer.
  • the user may specify which multimedia segment is used to generate the target music among multiple multimedia segments corresponding to a time period.
  • FIG. 8 is a schematic diagram of another display page provided by an embodiment of the present disclosure.
  • the example is a musical work generation page, in which, a plurality of multimedia clips corresponding to the time period A1 and obtained by playing by multiple users respectively, and a plurality of multimedia clips corresponding to the time period A2 respectively played by multiple users are displayed.
  • multimedia clips, and a plurality of multimedia clips corresponding to the time period A3 which are respectively performed by multiple users.
  • the user can select which multimedia clips are used to generate the target music on the page shown in FIG. 8 .
  • the user may select the multimedia segment corresponding to the period A1 performed by user 1, the multimedia segment corresponding to the period A2 performed by user 4, and the multimedia segment corresponding to the period A3 performed by user 2.
  • the server will use these three multimedia fragments as the multimedia fragment corresponding to track A.
  • track A is used as an example in FIG. 8 , and similar implementation manners may be adopted for other tracks.
  • S703 Concatenate the audio information in the multimedia clips corresponding to each period of each track to obtain the track audio corresponding to each track.
  • S704 Merge track audio corresponding to each track to generate the target music.
  • track A includes three time periods, namely time period A1, time period A2, and time period A3, the multimedia segment corresponding to time period A1, the multimedia fragment corresponding to time period A2, and the time period corresponding to time period A3 can be The multimedia clips are spliced to obtain the track audio corresponding to track A.
  • the processing of track B and track C is similar to that of track A, and will not be repeated here.
  • the following processing manner may be adopted: determine the track type of the track, and the track type is a melody track type, an accompaniment track type or a percussion track type.
  • the audio information in at least one multimedia segment corresponding to the track is processed to obtain the track audio corresponding to the track. That is to say, the processing methods corresponding to different track types can be different, so that different track types can be satisfied.
  • the sound effect characteristics of the channel can improve the effect of the target music.
  • the specific processing manner for each track is not limited here, and reference may be made to the detailed description of subsequent embodiments.
  • the track audio corresponding to track A, the track audio corresponding to track B, and the track audio corresponding to track C are merged to generate target music.
  • the music generation method provided by this embodiment includes: determining a music template, the music template includes a plurality of tracks, each track is divided into at least one period, and according to the music template, the corresponding period of each track is obtained in the database.
  • the multimedia segment corresponding to each period of each track includes: the audio information obtained by the user performing the melody of the track corresponding to the period in the track, and the database includes multiple multimedia segments performed by multiple users. Segment: Splicing the audio information in the multimedia segment corresponding to each period of each track to obtain the track audio corresponding to each track, and merging the track audio corresponding to each track to generate the target music.
  • the target music generated by the server according to the multimedia segment in the database is generated by the cooperative performance of multiple users.
  • the target music is generated by adopting a manner of cooperative performance by multiple users, which enhances cooperation and interaction between users, increases the interest of users in creating music or playing music, and improves user experience.
  • Fig. 9 is a schematic diagram of a music generation process provided by an embodiment of the present disclosure.
  • the target music includes Track A, Track B, and Track C.
  • the track type of track A is the melody track type
  • the track type of track B is the accompaniment track type
  • the track type of track C is the percussion track type.
  • Track A is divided into 3 periods, namely period A1, period A2, and period A3, track B is divided into 3 periods, respectively, period B1, period B2, and period B3, and track C is divided into 3 periods, respectively It is period C1, period C2 and period C3.
  • the server obtains the multimedia segments corresponding to the time period A1, time period A2, time period A3, time period B1, time period B2, time period B3, time period C1, time period C2, and time period C3.
  • the server performs the following processing on the above multimedia clips to generate target music.
  • the first processing is carried out to the audio information in the multimedia segment corresponding to each period of track A, to obtain the corresponding intermediate multimedia segment of this period, and to the corresponding intermediate multimedia segment of each period of track A
  • the audio information is spliced to obtain the track audio corresponding to track A.
  • the first processing includes at least one of the following: treble correction processing and reverberation processing.
  • the pitch correction process refers to modifying the pitch of the audio information in the multimedia clip according to the pitch of the original song corresponding to the music template, so that the audio information in the multimedia clip is more consistent with the original song.
  • Adding obfuscation processing refers to adding an obfuscation effect to the audio information in the multimedia segment, so that the audio signal has a more sense of space.
  • Reverberation is an acoustic phenomenon in which the sound continues to exist and bounces after the sound source stops speaking in a space.
  • the performance effect of the generated target music can be improved by performing pitch correction processing and/or adding reverberation processing to the melody track.
  • the second processing includes at least one of the following: treble correction processing, reverberation processing, and low frequency enhancement processing.
  • the effective sound of the accompaniment track is mainly in the low-frequency part
  • in addition to performing pitch modification processing and adding reverberation processing on the accompaniment track it is also possible to perform enhanced low-frequency processing on the accompaniment track, namely , the loudness of the low-frequency part of the audio information in the multimedia clip is enhanced, so that the accompaniment is thicker and can support the entire target music work.
  • the third processing includes at least one of the following: high-pass filtering processing and adding reverberation processing.
  • the effective sound of the strike track is mainly in the high-frequency part
  • high-frequency filtering can be performed on the strike track to eliminate low-frequency noise, so that the sound of the high-frequency part is more clearly identifiable.
  • the volume equalization processing may be performed on the track audio corresponding to each track, so that the volume of the track corresponding to each track is relatively balanced.
  • volume equalization processing by performing volume equalization processing on the track audio corresponding to each track, on the one hand, the volume between the tracks is relatively balanced, and on the other hand, the volume between the time periods in the same track is also relatively balanced.
  • the volume of the original song corresponding to the music template may be determined as the preset volume, so that the performance effect of the target music generated in this embodiment is closer to the performance effect of the original song.
  • track audio of each track may be merged first to obtain the merged audio, and then noise reduction processing may be performed on the merged audio to obtain the target music. It is also possible to perform noise reduction processing on the track audio of each track separately first to obtain the noise reduction audio corresponding to each track, and then perform merging processing on the noise reduction audio corresponding to each track to obtain the target music. By performing noise reduction processing, the background noise in the target music can be eliminated, and the performance effect of the target music can be improved.
  • the multimedia segment includes audio information
  • the target music is generated by combining the audio information in the multimedia segment.
  • the target music is music in audio form.
  • video information may also be recorded synchronously. That is to say, the multimedia segment includes audio information and video information.
  • the target video corresponding to the target music can also be generated by combining the video information in each multimedia segment. Therefore, the user can play the target music and the corresponding target video synchronously, and watch the effect of multi-person cooperative performance through the target video, thereby further increasing the fun of music performance.
  • the target video corresponding to the target music can be generated in the following manner: obtain the video template and the playing time of each multimedia segment, the video template includes a video window corresponding to each track; determine at least the video window corresponding to each video window A multimedia segment, for any video window, according to the playing time of at least one multimedia segment corresponding to the video window, the video information in the multimedia segment corresponding to the video window is concatenated to obtain the The track video corresponding to the video window is respectively added to the corresponding video window to obtain the target video corresponding to the target music.
  • FIG. 10 is a schematic diagram of a video template provided by an embodiment of the present disclosure.
  • the video template includes 3 video windows, wherein each video window is used to play a track video corresponding to a track.
  • window 1 is used to play the track video corresponding to track A
  • window 2 is used to play the track video corresponding to track B
  • window 3 is used to play the track video corresponding to track C.
  • the video information in at least one multimedia segment corresponding to track A is concatenated to obtain the track video corresponding to track A, and the track video corresponding to track A Link to window 1 for playback.
  • the video information in at least one multimedia segment corresponding to track B is concatenated to obtain the track video corresponding to track B, and the track video corresponding to track B is associated with window 2 for playback.
  • the video information in at least one multimedia segment corresponding to track C is concatenated to obtain the track video corresponding to track C, and the track video corresponding to track C is associated with window 3 for playback.
  • the video template may also include at least one of the following: title, title, transition effects, and the like.
  • title When generating the target video, you can add a title at the beginning of the target music, add a credit at the end of the target music, and add transition effects when switching between time periods, so as to improve the effect of the target video.
  • the server may send the target music and the target video to the terminal device.
  • the terminal device can play the target music and the target video synchronously, so that the user can watch the effect of multi-person cooperative performance, further increasing the fun of music performance.
  • Fig. 11 is a schematic structural diagram of a music generating device provided by an embodiment of the present disclosure.
  • the means may be in the form of software and/or hardware.
  • the apparatus may be a terminal device, or a processor, chip, chip module, module, unit, application program, etc. integrated into the terminal device.
  • the music generating device 1100 provided in this embodiment includes: a display module 1101 , a generating module 1102 and a sending module 1103 .
  • the display module 1101 is used to display a plurality of tracks corresponding to the music template, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the display module 1101 is further configured to display a multimedia recording page corresponding to the first time period in response to the selection operation of the first time period, where the first time period is any one of the time periods corresponding to the plurality of tracks;
  • a generating module 1102 configured to record multimedia information within a recording period corresponding to the multimedia recording page, and obtain a first multimedia segment corresponding to the first period;
  • the multimedia recording page includes a start control and a stop control; the generating module 1102 is specifically used for:
  • start recording multimedia information In response to the click operation on the start control, start recording multimedia information, until the click operation on the stop control is received or at the end of the recording period, stop recording multimedia information, and obtain the first multimedia information fragment.
  • the multimedia recording page further includes an information prompt area; the display module 1101 is also used for:
  • prompt information is displayed in the information prompt area, and the prompt information includes at least one of the following: lyrics and notes.
  • the music generation device provided in this embodiment can be used to implement the music generation method performed by the terminal device in any of the above method embodiments, and its implementation principle and technical effect are similar, and will not be repeated here.
  • Fig. 12 is a schematic structural diagram of another music generating device provided by an embodiment of the present disclosure.
  • the means may be in the form of software and/or hardware.
  • the device may be a server, or a processor, chip, chip module, module, unit, application program, etc. integrated into the server.
  • the music generation device 1200 provided in this embodiment includes: a determination module 1201 , an acquisition module 1202 , a processing module 1203 and a generation module 1204 .
  • the determination module 1201 is used to determine a music template, the music template includes a plurality of tracks, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the acquiring module 1202 is configured to acquire, in the database, the multimedia segment corresponding to each period of each track in the database according to the music template; wherein, the multimedia segment corresponding to each period of each track includes: The audio information obtained by playing the track melody corresponding to the time period, the database includes multiple multimedia clips played by multiple users;
  • the processing module 1203 is used to splice the audio information in the multimedia clips corresponding to each period of each track to obtain the track audio corresponding to each track;
  • the generating module 1204 is configured to merge track audio corresponding to each track to generate the target music.
  • the processing module 1203 is specifically configured to:
  • the track type is a melody track type, an accompaniment track type or a percussion track type
  • the audio information in the multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is a melody track type; the processing module 1203 is specifically configured to:
  • the first processing includes at least one of the following: treble correction processing, reverberation processing ;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is an accompaniment track type;
  • the processing module 1203 is specifically configured to:
  • the second processing includes at least one of the following: treble correction processing, adding reverberation processing , enhanced low frequency processing;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is a strike track type; the processing module 1203 is specifically configured to:
  • the third process includes at least one of the following: high-pass filter processing, adding reverberation processing ;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the generating module 1204 is specifically configured to:
  • the track audio of each track after the volume equalization process is combined and denoised to obtain the target music.
  • the multimedia segment also includes video information;
  • the acquiring module 1202 is also configured to: acquire a video template and the playing time of each multimedia segment, and the video template includes information corresponding to each track. the video window of
  • the processing module 1203 is further configured to: determine at least one multimedia segment corresponding to each video window; for any video window, according to the playing time of at least one multimedia segment corresponding to the video window, perform The video information in the segment is concatenated to obtain the track video corresponding to the video window;
  • the generating module 1204 is further configured to: respectively add the track video corresponding to each video window to the corresponding video window to obtain the target video corresponding to the target music.
  • the device further includes:
  • a sending module configured to send the target music and the target video to the terminal device.
  • the music generation device provided in this embodiment can be used to implement the music generation method performed by the server in any of the above method embodiments, and its implementation principle and technical effect are similar, and will not be repeated here.
  • the embodiments of the present disclosure further provide an electronic device.
  • the electronic device 1300 may be a terminal device or a server.
  • the terminal equipment may include but not limited to mobile phones, notebook computers, digital broadcast receivers, personal digital assistants (Personal Digital Assistant, PDA for short), tablet computers (Portable Android Device, PAD for short), portable multimedia players (Portable Media Player, referred to as PMP), mobile terminals such as vehicle-mounted terminals (such as vehicle-mounted navigation terminals), and fixed terminals such as digital TVs, desktop computers, etc.
  • PDA Personal Digital Assistant
  • PMP portable multimedia players
  • mobile terminals such as vehicle-mounted terminals (such as vehicle-mounted navigation terminals)
  • fixed terminals such as digital TVs, desktop computers, etc.
  • the electronic device shown in FIG. 13 is only an example, and should not limit the functions and scope of use of the embodiments of the present disclosure.
  • an electronic device 1300 may include a processing device (such as a central processing unit, a graphics processing unit, etc.) 1308 loads the program in the random access memory (Random Access Memory, RAM for short) 1303 to execute various appropriate actions and processes.
  • a processing device such as a central processing unit, a graphics processing unit, etc.
  • RAM Random Access Memory
  • various programs and data necessary for the operation of the electronic device 1300 are also stored.
  • the processing device 1301, ROM 1302, and RAM 1303 are connected to each other through a bus 1304.
  • An input/output (I/O) interface 1305 is also connected to the bus 1304 .
  • an input device 1306 including, for example, a touch screen, a touchpad, a keyboard, a mouse, a camera, a microphone, an accelerometer, a gyroscope, etc.; ), a speaker, a vibrator, etc.
  • a storage device 1308 including, for example, a magnetic tape, a hard disk, etc.
  • the communication means 1309 may allow the electronic device 1300 to communicate with other devices wirelessly or Wired communication to exchange data. While FIG. 13 shows electronic device 1300 having various means, it is to be understood that implementing or having all of the means shown is not a requirement. More or fewer means may alternatively be implemented or provided.
  • embodiments of the present disclosure include a computer program product, which includes a computer program carried on a computer-readable medium, where the computer program includes program codes for executing the methods shown in the flowcharts.
  • the computer program may be downloaded and installed from a network via communication means 1309, or from storage means 1308, or from ROM 1302.
  • the processing device 1301 the above-mentioned functions defined in the methods of the embodiments of the present disclosure are executed.
  • the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the two.
  • a computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination thereof.
  • Computer-readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer diskettes, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programming read-only memory (Erasable Programmable Read-Only Memory, referred to as EPROM or flash memory), optical fiber, portable compact disk read-only memory (Compact Disc Read-Only Memory, referred to as CD-ROM), optical storage device, magnetic storage device, or the above any suitable combination.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave carrying computer-readable program code therein. Such propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • the program code contained on the computer-readable medium can be transmitted by any appropriate medium, including but not limited to: electric wire, optical cable, radio frequency (Radio Frequency, RF for short), etc., or any suitable combination of the above.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device, or may exist independently without being incorporated into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device is made to execute the methods shown in the above-mentioned embodiments.
  • Computer program code for carrying out the operations of the present disclosure can be written in one or more programming languages, or combinations thereof, including object-oriented programming languages—such as Java, Smalltalk, C++, and conventional Procedural Programming Language - such as "C" or a similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or it can be connected to an external A computer (connected via the Internet, eg, using an Internet service provider).
  • LAN Local Area Network
  • WAN Wide Area Network
  • each block in a flowchart or block diagram may represent a module, program segment, or portion of code that contains one or more Executable instructions for the logical functions now specified.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by a dedicated hardware-based system that performs the specified functions or operations , or may be implemented by a combination of dedicated hardware and computer instructions.
  • the modules involved in the embodiments described in the present disclosure may be implemented by software or by hardware. Wherein, the name of the module does not constitute a limitation on the module itself under certain circumstances, for example, the determination module may also be described as "a module for determining the music template".
  • exemplary types of hardware logic components include: Field-Programmable Gate Array (Field-Programmable Gate Array, FPGA for short), Application Specific Integrated Circuit (ASIC for short), application specific standard product (Application Specific Standard Product, ASSP for short), System On a Chip (SOC for short), Complex Programmable Logic Device (CPLD for short), etc.
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, apparatus, or device.
  • a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • a machine-readable medium may include, but is not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatus, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer discs, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • magnetic storage or any suitable combination of the foregoing.
  • a music generation method including:
  • the music template includes a plurality of tracks, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the multimedia segment corresponding to each period of each track is obtained in the database; wherein, the multimedia segment corresponding to each period of each track includes: the user plays the track melody corresponding to the period in the track
  • the obtained audio information the database includes a plurality of multimedia clips performed by a plurality of users;
  • the audio information in the multimedia segment corresponding to each period of each track is spliced respectively to obtain the track audio corresponding to each track;
  • the track audio corresponding to each track is merged to generate the target music.
  • the audio information in the multimedia clip corresponding to each period of the track is spliced to obtain the track audio corresponding to each track ,include:
  • the track type is a melody track type, an accompaniment track type or a percussion track type
  • the audio information in the multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is a melody track type; according to the track type of the track, the audio information in the multimedia segment corresponding to each period of the track is spliced to obtain the Track audio corresponding to the track, including:
  • the first processing includes at least one of the following: treble correction processing, reverberation processing ;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is an accompaniment track type; according to the track type of the track, the audio information in the multimedia segment corresponding to each period of the track is spliced to obtain the Track audio corresponding to the track, including:
  • the second processing includes at least one of the following: treble correction processing, adding reverberation processing , enhanced low frequency processing;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is an attack track type; according to the track type of the track, the audio information in the multimedia segment corresponding to each period of the track is spliced to obtain the Track audio corresponding to the track, including:
  • the third process includes at least one of the following: high-pass filter processing, adding reverberation processing ;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track audio corresponding to each track is merged to generate target music, including:
  • the track audio of each track after the volume equalization process is combined and denoised to obtain the target music.
  • the multimedia segment further includes video information; the method further includes:
  • the video template includes a video window corresponding to each track
  • the video information in the multimedia segment corresponding to the video window is concatenated to obtain the track video corresponding to the video window;
  • the track video corresponding to each video window is respectively added to the corresponding video window to obtain the target video corresponding to the target music.
  • a music generation method including:
  • each track is divided into at least one period, and the plurality of tracks are arrangement tracks of music
  • the first time period In response to the selection operation of the first time period, displaying the multimedia recording page corresponding to the first time period, the first time period being any one of the time periods corresponding to the plurality of tracks;
  • Record multimedia information within the recording period corresponding to the multimedia recording page obtain a first multimedia segment corresponding to the first period, and send the first multimedia segment to the server.
  • the multimedia recording page includes a start control and a stop control; recording multimedia information within the recording time period corresponding to the multimedia recording page, and obtaining the first multi-media information corresponding to the first time period Media snippets, including:
  • start recording multimedia information In response to the click operation on the start control, start recording multimedia information, until the click operation on the stop control is received or at the end of the recording period, stop recording multimedia information, and obtain the first multimedia information fragment.
  • the multimedia recording page further includes an information prompt area; the method further includes:
  • prompt information is displayed in the information prompt area, and the prompt information includes at least one of the following: lyrics and notes.
  • a music generating device including:
  • a determining module configured to determine a music template, the music template including a plurality of tracks, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the obtaining module is used to obtain the multimedia segment corresponding to each period of each track in the database according to the music template; wherein, the multimedia segment corresponding to each period of each track includes: the user for the period in the track
  • the audio information obtained by performing the corresponding track melody, the database includes multiple multimedia clips played by multiple users;
  • the processing module is used to splicing the audio information in the multimedia clips corresponding to each time period of each track respectively to obtain the track audio corresponding to each track;
  • a generating module configured to combine track audio corresponding to each track to generate the target music.
  • the processing module is specifically configured to:
  • the track type is a melody track type, an accompaniment track type or a percussion track type
  • the audio information in the multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is a melody track type
  • the processing module is specifically used for:
  • the first processing includes at least one of the following: treble correction processing, reverberation processing ;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is an accompaniment track type
  • the processing module is specifically used for:
  • the second processing includes at least one of the following: treble correction processing, adding reverberation processing , enhanced low frequency processing;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the track type is a strike track type; the processing module is specifically used for:
  • the third process includes at least one of the following: high-pass filter processing, adding reverberation processing ;
  • the audio information in the intermediate multimedia segment corresponding to each period of the track is spliced to obtain the track audio corresponding to the track.
  • the generating module is specifically configured to:
  • the track audio of each track after the volume equalization process is combined and denoised to obtain the target music.
  • the multimedia segment also includes video information;
  • the acquiring module is also used to: acquire a video template and the playing time of each multimedia segment, and the video template includes each track The corresponding video window;
  • the processing module is also used to: determine at least one multimedia segment corresponding to each video window; for any video window, according to the playing time of at least one multimedia segment corresponding to the video window, the multimedia segment corresponding to the video window.
  • the video information in the video window is serially processed to obtain the track video corresponding to the video window;
  • the generating module is further configured to: respectively add the track video corresponding to each video window to the corresponding video window to obtain the target video corresponding to the target music.
  • the device further includes:
  • a sending module configured to send the target music and the target video to the terminal device.
  • a music generating device including:
  • the display module is used to display a plurality of tracks corresponding to the music template, each track is divided into at least one period, and the plurality of tracks are music arrangement tracks;
  • the display module is further configured to display a multimedia recording page corresponding to the first time period in response to a selection operation of the first time period, where the first time period is any one of the time periods corresponding to the plurality of tracks;
  • a generating module configured to record multimedia information within a recording period corresponding to the multimedia recording page, to obtain a first multimedia segment corresponding to the first period;
  • a sending module configured to send the first multimedia segment to a server.
  • the multimedia recording page includes a start control and a stop control; the generating module is specifically used for:
  • start recording multimedia information In response to the click operation on the start control, start recording multimedia information, until the click operation on the stop control is received or at the end of the recording period, stop recording multimedia information, and obtain the first multimedia information fragment.
  • the multimedia recording page also includes an information prompt area; the display module is also used for:
  • prompt information is displayed in the information prompt area, and the prompt information includes at least one of the following: lyrics and notes.
  • an electronic device including: at least one processor and a memory;
  • the memory stores computer-executable instructions
  • the at least one processor executes the computer-executed instructions stored in the memory, so that the at least one processor executes the music generation method described in the first aspect and various possible implementations of the first aspect, or executes the music generation method described in the first aspect above.
  • a computer-readable storage medium stores computer-executable instructions, and when a processor executes the computer-executable instructions, Realize the music generation method described in the above first aspect and various possible implementations of the first aspect, or realize the music generation method described in the above second aspect and various possible implementations of the second aspect.
  • a computer program product including a computer program.
  • the computer program When the computer program is executed by a processor, various possible functions of the first aspect and the first aspect can be realized.
  • a computer program is provided.
  • the above first aspect and various possible implementation modes of the first aspect are implemented.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • Electrophonic Musical Instruments (AREA)

Abstract

一种音乐生成方法、装置(1100、1200)、设备、存储介质及程序,方法包括:确定音乐模板,音乐模板包括多个轨道,每个轨道被划分为至少一个时段,根据音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,数据库中包括多个用户演奏的多个多媒体片段(S702),分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频(S703),对各轨道对应的轨道音频进行合并处理,生成目标音乐(S704)。增强了用户之间的合作性、互动性,增加了用户创造音乐或者演奏音乐的趣味性,提升用户体验。

Description

音乐生成方法、装置、设备、存储介质及程序
相关申请的交叉引用
本公开要求于2022年2月28日提交中国专利局、申请号为202210190764.6、申请名称为“音乐生成方法、装置、设备、存储介质及程序”的中国专利申请的优先权,其全部内容通过引用结合在本文中。
技术领域
本公开实施例涉及互联网技术领域,尤其涉及一种音乐生成方法、装置、设备、存储介质、计算机程序产品及计算机程序。
背景技术
随着终端技术的发展,用户希望能够通过终端设备来创作音乐或者演奏音乐,从而增加趣味性。
目前,终端设备可以安装K歌软件,并通过K歌软件播放音乐作品的背景音乐,用户跟随背景音乐进行K歌,以实现音乐创作或音乐演奏。
然而,上述创作/演奏音乐的方式较为单调,缺乏用户之间的合作性、互动性。
发明内容
本公开实施例提供一种音乐生成方法、装置、设备、存储介质、计算机程序产品及计算机程序,以解决音乐创作/演奏过程合作性、互动性较差问题。
第一方面,本公开实施例提供一种音乐生成方法,包括:
确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;
分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频;
对各轨道对应的轨道音频进行合并处理,生成目标音乐。
第二方面,本公开实施例提供一种音乐生成方法,包括:
显示音乐模板的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个;
在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段,并向服务器发送所述第一多媒体片段。
第三方面,本公开实施例提供一种音乐生成装置,包括:
确定模块,用于确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
获取模块,用于根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;
处理模块,用于分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频;
生成模块,用于对各轨道对应的轨道音频进行合并处理,生成目标音乐。
第四方面,本公开实施例提供一种音乐生成装置,包括:
显示模块,用于显示音乐模板对应的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
所述显示模块,还用于响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个;
生成模块,用于在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段;
发送模块,用于向服务器发送所述第一多媒体片段。
第五方面,本公开实施例提供一种电子设备,包括:处理器和存储器;
所述存储器存储计算机执行指令;
所述处理器执行所述计算机执行指令,实现如第一方面以及第一方面各种可能的实现方式中的音乐生成方法,或者,实现如第二方面以及第二方面各种可能的实现方式中的音乐生成方法。
第六方面,本公开实施例提供一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如上第一方面以及第一方面各种可能的实现方式中的音乐生成方法,或者,实现如第二方面以及第二方面各种可能的实现方式中的音乐生成方法。
第七方面,本公开实施例提供一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时,实现如第一方面以及第一方面各种可能的实现方式中的音乐生成方法,或者,实现如第二方面以及第二方面各种可能的实现方式中的音乐生成方法。
第八方面,本公开实施例提供一种计算机程序,所述计算机程序在被处理器执行时,实现如第一方面以及第一方面各种可能的实现方式中的音乐生成方法,或者,实现如第二方面以及第二方面各种可能的实现方式中的音乐生成方法。
本公开实施例提供了音乐生成方法、装置、设备、存储介质、计算机程序产品及计算机程序。该方法包括:确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段,分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频,对各轨道对应的轨道音频进行合并处理,生成所述目标音乐。上述过程中,由于数据库中存储的 各媒体片段是由多个用户演奏的,因此,服务器根据数据库中的多媒体片段生成的目标音乐,是由多个用户合作演奏生成的。通过采用多个用户合作演奏的方式来生成目标音乐,增强了用户之间的合作性、互动性,增加了用户创造音乐或者演奏音乐的趣味性,提升用户体验。
附图说明
为了更清楚地说明本公开实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本公开的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本公开实施例提供的音乐轨道和时段的示意图;
图2为本公开实施例提供的一种应用场景的示意图;
图3为本公开实施例提供的一种音乐生成方法的流程示意图;
图4为本公开实施例提供的一组显示页面的示意图;
图5为本公开实施例提供的另一组显示页面的示意图;
图6为本公开实施例提供的又一组显示页面的示意图;
图7为本公开实施例提供的另一种音乐生成方法的流程示意图;
图8为本公开实施例提供的又一种显示页面的示意图;
图9为本公开实施例提供的一种音乐生成过程的示意图;
图10为本公开实施例提供的一种视频模板的示意图;
图11为本公开实施例提供的一种音乐生成装置的结构示意图;
图12为本公开实施例提供的另一种音乐生成装置的结构示意图;
图13为本公开实施例提供的一种电子设备的结构示意图。
具体实施方式
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例中的附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本公开一部分实施例,而不是全部的实施例。基于本公开中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本公开保护的范围。
目前,用户创作音乐或者演奏音乐的方式,通常包括:听歌以及K歌。在K歌场景,终端设备安装K歌软件,并通过K歌软件播放音乐作品的背景音乐,用户跟随背景音乐进行K歌。
然而,音乐本身是一种互动性、合作性非常强的艺术形式,上述听歌以及K歌等方式较为单调,缺乏用户之间的合作性、互动性。
为了解决上述技术问题,本公开提供一种音乐生成方法、装置、设备、存储介质及程序,可以采用多个用户合作演奏的方式来生成目标音乐。其中,多个用户合作演奏的方式是指,将音乐模板划分为多个部分,每个用户演奏其中一个或者多个部分,将多个用户各自演奏的部分进行合并得到目标音乐。
本公开实施例中,音乐模板可以包括多个轨道。其中,本公开实施例中的轨道是指音乐的编奏轨道。每个轨道对应的轨道类型可以为旋律轨道类型、伴奏轨道类型、打击轨道类型中的一种。其中,旋律轨道是指为演唱歌词的声音所对应的轨道,伴奏轨道可以为吉他、钢琴、二胡、小提琴等乐器的声音所对应的轨道,打击轨道可以为架子鼓、手鼓等打击类乐器的声音所对应的轨道。
本公开实施例中对于音乐模板中所包括的轨道的数量不做限定,示例性的,有些音乐模板包括旋律轨道、钢琴伴奏轨道、二胡伴奏轨道、架子鼓打击轨道。有些音乐模板可能包括旋律轨道、吉他伴奏轨道、手鼓打击轨道。
下面结合图1对多人合作演奏的方式进行举例说明。
图1为本公开实施例提供的音乐轨道和时段的示意图。参见图1,假设音乐模板包括3个轨道,分别为轨道A、轨道B和轨道C,则可以采用如下合作演奏的方式来生成目标音乐:用户1演奏轨道A得到轨道A对应的多媒体信息、用户2演奏轨道B得到轨道B对应的多媒体信息、用户3演奏轨道C得到轨道C对应的多媒体信息。将上述轨道A、轨道B、轨道C对应的多媒体信息进行合并处理,生成目标音乐。
其中,上述的多媒体信息包括音频信息和/或视频信息。
进一步的,继续参见图1,还可以从时间维度将每个轨道划分为多个时段。每个时段对应一定时长,不同时段之间不存在重叠。例如,可以按照节拍顺序,每4-8个节拍划分为一个时段,或者,按照时间顺序,每10秒划分为一个时段。例如,图1中,轨道A划分为3个时段,分别为时段A1、时段A2、时段A3;轨道B划分为3个时段,分别为时段B1、时段B2、时段B3;轨道C划分为3个时段,分别为时段C1、时段C2、时段C3。
这样,可以采用如下合作演奏的方式来生成目标音乐:每个用户演奏上述时段A1、时段A2、时段A3、时段B1、时段B2、时段B3、时段C1、时段C2、时段C3中的部分时段。
举例而言,用户1演奏时段A1得到时段A1对应的多媒体片段,用户2演奏时段A2得到时段A2对应的多媒体片段、用户3演奏时段A3得到时段A3对应的多媒体片段,用户4演奏时段B1得到时段B1对应的多媒体片段,用户4演奏时段B2得到时段B2对应的多媒体片段、用户5演奏时段B3得到时段B3对应的多媒体片段,用户6演奏时段C1得到时段C1对应的多媒体片段,用户7演奏时段C2得到时段C2对应的多媒体片段、用户8演奏时段C3得到时段C3对应的多媒体片段。
进而,对时段A1、时段A2、时段A3对应的多媒体片段进行拼接处理,得到轨道A对应的多媒体信息;对时段B1、时段B2、时段B3对应的多媒体片段进行拼接处理,得到轨道B对应的多媒体信息;对时段C1、时段C2、时段C3对应的多媒体片段进行拼接处理,得到轨道C对应的多媒体信息。进而,将上述轨道A、轨道B、轨道C对应的多媒体信息进行合并处理,生成目标音乐。
为了便于对本公开技术方案的理解,下面结合图2对本公开实施例的应用场景进行介绍。
图2为本公开实施例提供的一种应用场景的示意图。如图2所示,该应用场景包括:终端设备、数据库和服务器。其中,终端设备可以是具有多媒体交互功能的任意电子设备,包括但不限于:智能手机、平板电脑、笔记本电脑、智能电视、智能穿戴设备、智能家居设备、智能车载设备、智能音箱等。后续附图中在示例终端设备的显示页面时,以手机页面为例进 行示例。服务器可以为云服务器,又称为云计算服务器或云主机,是云计算服务体系中的一项主机产品。服务器也可以为分布式系统的服务器,或者是结合了区块链的服务器。
本实施例提供的音乐生成方法可以包括两个过程,分别为:多媒体片段的生成过程和目标音乐的生成过程。其中,多媒体片段的生成过程可以由终端设备执行,目标音乐的生成过程可以由服务器执行。下面结合图1对上述两个过程分别进行介绍。
如图2所示,在多媒体片段的生成过程中,可以由多个用户参与。每个用户可以通过终端设备录制得到一个或者多个多媒体片段。示例性的,每个用户可以选择对某个轨道中的某个时段进行演奏,并对演奏过程进行录制得到多媒体片段。终端设备将多媒体片段上传到数据库中。数据库用于存储各个轨道中的各时段对应的多媒体片段。
继续参见图2,在目标音乐的生成过程中,服务器可以从数据库中获取各轨道中的各时段对应的多媒体片段,并对这些多媒体片段进行合并处理,从而生成目标音乐。
由图2可见,本公开应用场景中,数据库中存储的各多媒体片段是由多个用户演奏得到的,因此,服务器根据数据库中的多媒体片段生成的目标音乐,是由多个用户合作演奏生成的。
本公开技术方案中,通过采用多个用户合作演奏的方式来生成目标音乐,增强了用户之间的合作性、互动性,增加了用户创造音乐或者演奏音乐的趣味性,提升用户体验。同时,还能丰富网络用户生成内容(User Generated Content,UGC),促进音乐演奏的普及和传播。
下面以具体地实施例对本公开的技术方案进行详细说明。下面这几个具体的实施例可以相互结合,对于相同或相似的概念或过程可能在某些实施例不再赘述。
图3为本公开实施例提供的一种音乐生成方法的流程示意图。本实施例的方法可以由终端设备执行。本实施例描述一个多媒体片段的生成过程。如图3所示,本实施例的方法包括:
S301:显示音乐模板的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道。
本实施例中,音乐模板也可以称为待演奏的音乐。在终端设备执行本实施例之前,音乐模板已经被划分为多个轨道,并且,每个轨道被划分为多个时段。终端设备可以显示选择页面,以供用户选择演奏哪个轨道中的哪个时段。
一个示例中,图4为本公开实施例提供的一组显示页面的示意图。如图4所示,终端设备显示页面401,在页面401中显示多个音乐。用户可以在页面401中选择其中一个音乐进行演奏。用户选择的音乐即作为音乐模板。参见图4,响应于用户对“音乐1”的选择操作,终端设备显示“音乐1”对应的选择页面,即页面402,在页面402中显示音乐1包括的多个轨道,以及每个轨道被划分的至少一个时段。
应理解,本实施例对于音乐模板中所包括的轨道数量以及每个轨道中包括的时段数量不做限定。图4中是以音乐1包括3个轨道,每个轨道包括3个时段为例进行示意。
在页面402中,用户根据时段标识可能无法准确确定该时段对应目标音乐的哪一部分。因此,在一些可能的实现方式中,在用户选中某个时段的标识时,还可以在页面402中显示该时段对应的描述信息,例如,显示该时段对应的歌词、音符、时间区间等。本实施例对于描述信息的显示位置以及显示方式不做限定。一个示例中,当某个时段的标识被选中时,该 时段的标识被替换显示为该时段对应的描述信息。这样,用户可以根据时段的描述信息准确获知需要演奏的具体内容,便于用户进行选择。
继续参见图4,在页面402中,用户可以根据自己的意愿选择其中一个时段进行演奏,例如,假设用户想要演奏轨道A对应的时段A2,在用户选中轨道A对应的时段A2,并点击确定控件后,触发终端设备执行本实施例的S302。
本实施例中,通过将每个轨道划分为多个时段,用户可以选择一个时段进行演奏,能够降低用户演奏难度。当然,实际应用中,用户也可以选择多个时段进行演奏。
S302:响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个。
S303:在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段,并向服务器发送所述第一多媒体片段。
本实施例中,将用户选择的时段称为第一时段。第一时段可以是页面402中的任意一个时段。例如,用户在图4的页面402中选择了轨道A对应的时段A2,则轨道A对应的时段A2即为第一时段。
本实施例中,终端设备在检测到用户选择第一时段后,显示第一时段对应的多媒体录制页面,用户可以在多媒体录制页面中进行多媒体信息的录制。例如,用户可以演奏第一时段对应的音乐内容。举例而言,若第一时段对应旋律轨道,则用户演唱第一时段对应的歌词;若第一时段对应钢琴伴奏轨道,则用户演奏第一时段对应的钢琴旋律;若第一时段对应吉他伴奏轨道,则用户演奏第一时段对应的吉他旋律。进一步的,在用户演奏的同时,通过终端设备对自己的演奏过程进行录制,得到第一时段对应的第一多媒体片段。
图5为本公开实施例提供的另一组显示页面的示意图。示例性的,用户在图4所示的页面402中选中轨道A的时段A2,并点击确定控件后,终端设备可以显示如图5所示的页面501。页面501即为多媒体录制页面。用户可以在页面501中录制多媒体信息,生成多媒体片段。
在一些可能的实现方式中,多媒体录制页面中包括开始控件和停止控件,终端设备响应于对所述开始控件的点击操作,开始录制多媒体信息,直至接收到对所述停止控件的点击操作或者在所述录制时段的结束时刻,停止录制多媒体信息,得到所述第一多媒体片段。
在一些可能的实现方式中,多媒体录制页面中还可以包括信息提示区域。在多媒体信息录制过程中,终端设备获取多媒体信息的录制进度,根据所述录制进度,在信息提示区域中显示提示信息,所述提示信息包括如下至少一种:歌词、音符等。例如,图5中,页面501中的虚线框标注的即为信息提示区域。该信息提示区域中,显示有歌词(采用XXX表示),以及音符(采用黑色矩形框组成的波形表示)。
下面结合图5详细说明多媒体信息的录制过程。
参见图5,用户在页面501中点击开始控件之后,用户开始对第一时段进行演奏,并通过终端设备的音频采集装置和/或视频采集装置对其演奏过程进行录制。录制过程如页面502所示。在录制过程中,可以在信息提示区域中提示当前待演奏的歌词、音符等信息。当录制完成后,得到第一时段对应的第一多媒体片段。终端设备显示页面503,用户可以在页面503中对第一多媒体片段进行回放观看,并确定是否将第一多媒体片段上传至服务器。
可选的,终端设备录制得到第一多媒体片段之后,终端设备还可以确定第一多媒体片段与原曲之间的相似度,并根据相似度确定第一多媒体片段对应的得分,在页面503中可以显示该得分。这样,用户可以根据得分,确定是否进行重新录制。若用户点击重新录制,则终端设备显示页面501,以便用户重新开始录制。若用户点击确认,则终端设备向服务器发送第一多媒体片段。
本实施例中,服务器接收到第一多媒体片段之后,将第一多媒体片段存储至数据库中。
应理解的是,本实施例描述的是一个多媒体片段的生成过程。针对页面402中显示的每个时段,均可以采用本实施例的方式生成每个时段对应的多媒体片段。实际应用中,页面402中显示的多个时段可以由多个用户演奏完成,每个用户演奏其中的一个或者多个时段。
本实施例中,通过采用多个用户合作演奏的方式来生成目标音乐,增强了用户之间的合作性、互动性,增加了用户创造音乐或者演奏音乐的趣味性,提升用户体验。
进一步的,本实施例中多个用户采用的是“异步合作”的方式,无需多个用户同时进行演奏,降低了相互配合的难度。另外,也避免了用户找不到人一起合作的问题。
另外,考虑到用户的音乐知识可能较少,不会演奏乐器,或者对自己的演奏水平没有信心,这些因素会降低用户演奏音乐的兴趣。因此,本实施例中,还可以采用下述可能的方式来降低用户的演奏难度。
在一些可能的实现方式中,可以采用阿卡贝拉(Acappella)的合作方式。阿卡贝拉是指无伴奏合唱,以人声清唱,不使用乐器。具体而言,事先将候选歌曲(例如有改编权的歌曲,或者用户自行创造的歌曲)改编为阿卡贝拉的形式。改编好的歌曲有多个轨道,每个轨道均为阿卡贝拉的形式。用户在演奏每个轨道时,均采用人声清唱的方式进行演奏。这样,即使用户不会乐器,也可以参与多人合作,降低音乐演奏的难度。
在另一些可能的实现方式中,针对每个时段,可以事先录制教学视频。教学视频中可以由专业人员对该时段的演唱做出示范,或者,由专业人员介绍说明该时段的演唱注意事项,帮助用户掌握正确的音高和节奏。下面结合图6说明教学视频的使用过程。
图6为本公开实施例提供的又一组显示页面的示意图。示例性的,用户在图4所示的页面402中选中轨道A的时段A2,并点击确定控件后,终端设备可以先显示如图6中的页面601。页面601即为教学视频页面。在页面601中,用户可以点击开始控件,开始播放第一时段对应的教学视频。与图5所示的多媒体录制页面类似,教学视频页面中可以包括信息提示区域。在教学视频播放过程中,信息提示信息可以显示歌词、字符等提示信息。教学视频的播放过程如页面602所示,
用户可以根据教学视频学习得到第一时段对应的音高、节奏等要点。应理解,用户可以重复多次观看教学视频。如页面603所示,当用户学会之后,可以点击“学会了”控件,触发终端设备显示如图5所示的页面501,开始多媒体录制过程。
本实施例中,通过向用户提供每个时段对应的教学视频,使得用户可以先学习再演奏,进一步降低了用户的演奏难度。
上述各实施例描述了目标音乐的其中一个时段对应的多媒体片段的生成过程。应理解,当目标音乐中的所有时段对应的多媒体片段均生成后,可以对各时段对应的多媒体片段进行合并处理,生成目标音乐。下面结合几个实施例详细说明如何对多个多媒体片段进行合并处理,以生成目标音乐。
图7为本公开实施例提供的另一种音乐生成方法的流程示意图。本实施例的方法可以由服务器执行。本实施例描述目标音乐的生成过程。如图7所示,本实施例的方法,包括:
S701:确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道。
本实施例的执行时机为服务器确定需要根据音乐模板生成目标音乐时,可以有如下两种触发方式。
方式1:服务器自动触发。示例性的,服务器检测到音乐模板对应的多媒体片段已收集完成时,触发执行本实施例。例如,假设音乐模板包括3个轨道,分别为轨道A、轨道B、轨道C,轨道A被划分为3个时段,分别为时段A1、时段A2、时段A3,轨道B被划分为3个时段,分别为时段B1、时段B2、时段B3,轨道C被划分为3个时段,分别为时段C1、时段C2、时段C3,则上述时段A1、时段A2、时段A3、时段B1、时段B2、时段B3、时段C1、时段C2、时段C3对应的多媒体片段均收集完成时,可以执行本实施例的方法,以生成目标音乐。
方式2:用户手动触发。示例性的,终端设备在图4所示的页面401中,还可以展示每个时段对应的演奏信息,例如,该时段是否已被演奏。若已被演奏,说明服务器中已存在该时段对应的多媒体片段。若用户根据页面401展示的各时段的演奏信息,确定目标音乐中的各时段均已被演奏,则可以手动触发目标音乐的生成过程。示例性的,终端设备检测到用户进行音乐合成操作后,终端设备向服务器发送音乐生成请求,服务器接收到音乐生成请求后,执行本实施例的方法。
S702:根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段。
应理解,数据库中的每个多媒体片段,可以是采用上述图3所示实施例生成的。数据库中存储的各多媒体片段可以是由多个用户演奏并上传的。也就是说,数据库中包括多个用户演奏的多个多媒体片段。例如,数据库中存储有10个多媒体片段,其中,一部分多媒体片段由用户1演奏并上传,另一部分多媒体片段由用户2演奏并上传,又一部分多媒体片段由用户3演奏并上传。
可以根据音乐模板,分别获取每个轨道的每个时段对应的多媒体片段,从而得到生成目标音乐所需的多媒体片段。
其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应的轨道旋律进行演奏得到的音频信息。用户对每个轨道的每个时段对应的轨道旋律的演奏方式可以参见前述实施例的相关描述,此处不做赘述。
本实施例中,数据库在存储多媒体片段时,可以记录音乐模板、轨道、时段与多媒体片段之间的关系。一个示例中,数据库中存储的内容如表1所示。这样,服务器可以根据音乐模板中包括的多个轨道以及每个轨道对应的各时段,从数据库中获取生成目标音乐所需的多媒体片段。
表1

一些可能的场景中,针对某一个时段,可能被多个用户演奏,因此,数据库中可能存储有该时段对应的多个多媒体片段。针对该情况,可以有如下两种实现方式。
在一种可能的实现方式中,若由服务器触发目标音乐的生成过程,则服务器可以根据一个时段对应的多个多媒体片段的属性信息,在多个多媒体片段中确定一个多媒体片段用于生成目标音乐。其中,属性信息可以包括如下至少一种:演奏者信息、评分、上传时间、排名、热度值等。
示例性的,针对每个时段,若该时段对应有多个多媒体片段,则服务器可以选择排名第一的多媒体片段,或者,选择热度值最高的多媒体片段,或者,选择,最近上传的多媒体片段,或者,选择评分最高的多媒体片段,或者,选择某个指定演奏者演奏的多媒体片段。
在另一种可能的实现方式中,若由用户手动触发目标音乐的生成过程,则可以由用户在一个时段对应的多个多媒体片段中,指定使用哪个多媒体片段来生成目标音乐。
示例性的,图8为本公开实施例提供的又一种显示页面的示意图。如图8所示,示例的是音乐作品生成页面,在该页面中,显示时段A1对应的由多个用户分别演奏得到的多个多媒体片段、时段A2对应的由多个用户分别演奏得到的多个多媒体片段、以及时段A3对应的由多个用户分别演奏得到的多个多媒体片段。用户可以在图8所示页面中,选择使用哪些多媒体片段来生成目标音乐。例如,用户可以选择用户1演奏的时段A1对应的多媒体片段、用户4演奏的时段A2对应的多媒体片段、以及用户2演奏的时段A3对应的多媒体片段。这样,服务器将使用这3个多媒体片段作为轨道A对应的多媒体片段。
应理解,图8是以轨道A为例进行示意的,针对其他轨道,可以采用类似的实现方式。
S703:分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频。
S704:对各轨道对应的轨道音频进行合并处理,生成所述目标音乐。
示例性的,以轨道A为例,假设轨道A包括3个时段,分别为时段A1、时段A2、时段A3,则可以将时段A1对应的多媒体片段、时段A2对应的多媒体片段、时段A3对应的多媒体片段进行拼接处理,得到轨道A对应的轨道音频。轨道B、轨道C的处理过程与轨道A类似,此处不做赘述。
在一种可能的实现方式中,S703中,针对每个轨道,可以采用如下处理方式:确定该轨道的轨道类型,所述轨道类型为旋律轨道类型、伴奏轨道类型或者打击轨道类型。根据该轨道的轨道类型,对该轨道对应的至少一个多媒体片段中的音频信息进行处理,得到该轨道对应的轨道音频。也就是说,不同轨道类型所对应的处理方式可以不同,这样可以满足不同轨 道的音效特点,提高目标音乐的效果。针对每个轨道的具体处理方式,此处不做限定,可以参见后续实施例的详细说明。
进一步的,将轨道A对应的轨道音频、轨道B对应的轨道音频、轨道C对应的轨道音频进行合并处理,生成目标音乐。
本实施例提供的音乐生成方法,包括:确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,根据音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段,其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频,对各轨道对应的轨道音频进行合并处理,生成目标音乐。上述过程中,由于数据库中存储的各媒体片段是由多个用户演奏的,因此,服务器根据数据库中的多媒体片段生成的目标音乐,是由多个用户合作演奏生成的。通过采用多个用户合作演奏的方式来生成目标音乐,增强了用户之间的合作性、互动性,增加了用户创作音乐或者演奏音乐的趣味性,提升用户体验。
在图7所示实施例的基础上,下面结合一个具体的实施例对本公开技术方案进行更详细的说明。
图9为本公开实施例提供的一种音乐生成过程的示意图。如图9所示,假设目标音乐包括轨道A、轨道B和轨道C。轨道A的轨道类型为旋律轨道类型,轨道B的轨道类型为伴奏轨道类型,轨道C的轨道类型为打击轨道类型。轨道A被划分为3个时段,分别为时段A1、时段A2、时段A3,轨道B被划分为3个时段,分别为时段B1、时段B2、时段B3,轨道C被划分为3个时段,分别为时段C1、时段C2、时段C3。
本实施例中,服务器获取到时段A1、时段A2、时段A3、时段B1、时段B2、时段B3、时段C1、时段C2、时段C3各自对应的多媒体片段。服务器对上述多媒体片段进行如下处理,以生成目标音乐。
(1)针对轨道A,对轨道A的每个时段对应的多媒体片段中的音频信息进行第一处理,得到该时段对应的中间多媒体片段,并对轨道A的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到轨道A对应的轨道音频。
其中,所述第一处理包括如下至少一种:高音修正处理、加入混响处理。
本实施例中,音高修正处理是指,根据音乐模板对应的原曲的音高,对多媒体片段中的音频信息的音高进行修正,以使得多媒体片段中的音频信息与原曲更加相符。加入混淆处理是指,在多媒体片段中的音频信息中添加混淆效果,使得音频信号更有空间感。混响,是在一个空间内,声源停止发音后,声音继续存在、反弹的声学现象。
本实施例中,通过对旋律轨道进行音高修正处理和/或加入混响处理,能够提高生成的目标音乐的演奏效果。
(2)针对轨道B,对轨道B的每个时段对应的多媒体片段中的音频信息进行第二处理,得到该时段对应的中间多媒体片段,并对轨道B的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到轨道B对应的轨道音频。
其中,所述第二处理包括如下至少一种:高音修正处理、加入混响处理、增强低频处理。
本实施例中,针对伴奏轨道,由于伴奏轨道的有效声音主要在低频部分,因此,除了对伴奏轨道进行音高修改处理、加入混响处理之外,还可以对伴奏轨道进行增强低频处理,即,对多媒体片段中的音频信息的低频部分的响度进行增强,使得伴奏更加浑厚,能支撑起整个目标音乐作品。
(3)针对轨道C,对对轨道C的每个时段对应的多媒体片段中的音频信息进行第三处理,得到该时段对应的中间多媒体片段,并对轨道C的各时段对应的中间多媒体片段中的音频信息进行合并处理,得到轨道C对应的轨道音频。
其中,所述第三处理包括如下至少一种:高通滤波处理、加入混响处理。
本实施例中,针对打击轨道,由于打击轨道的有效声音主要在高频部分,因此,可以对打击轨道进行高频滤波处理,以消除低频噪声,使得高频部分的声音更加清晰可辨。另外,还可以对打击轨道加入混响,使得打击音效更加具有空间感。
(4)确定音乐模板对应的预设音量,根据所述预设音量,对每个轨道对应的轨道音频进行音量均衡处理,得到轨道A对应的音量均衡后的轨道音频、轨道B对应的音量均衡后的轨道音频、轨道C对应的音量均衡后的轨道音频。
实际应用中,由于各多媒体片段是由多个用户演奏得到的,不同用户在演奏时的音量可能有所不同。为了避免部分用户音量过大或过小导致整个目标音乐的不和谐,本实施例中,可以对各轨道对应的轨道音频进行音量均衡处理,以使各轨道对应的轨道音量相对均衡。
能够理解,本实施例中通过对各轨道对应的轨道音频进行音量均衡处理,一方面使得各轨道之间的音量相对均衡,另一方面也使得同一轨道中的各时段之间的音量相对均衡。
可选的,可以音乐模板对应的原曲音量,确定为所述预设音量,这样,使得本实施例生成的目标音乐的演奏效果更加趋于原曲的演奏效果。
(5)对轨道A对应的音量均衡后的轨道音频、轨道B对应的音量均衡后的轨道音频、轨道C对应的音量均衡后的轨道音频进行合并处理、以及降噪处理,得到目标音乐。
本实施例中,可以先对各轨道的轨道音频进行合并处理,得到合并音频,再对合并音频进行降噪处理,得到目标音乐。还可以先对各轨道的轨道音频分别进行降噪处理,得到各轨道对应的降噪音频,再对各轨道对应的降噪音频进行合并处理,得到目标音乐。通过进行降噪处理,能够消除目标音乐中的背景噪声,提高目标音乐的演奏效果。
上述图7和图9所示的实施例中,多媒体片段中包括音频信息,通过对多媒体片段中的音频信息进行合并处理,生成目标音乐。该目标音乐为音频形式的音乐。在一些可能的实现方式中,用户在录制多媒体片段时,除了录制音频信息,还可以同步录制视频信息。也就是说,多媒体片段中包括音频信息和视频信息。这样,还可以通过对各多媒体片段中的视频信息进行合并处理,生成目标音乐对应的目标视频。从而,用户可以同步播放目标音乐及其对应的目标视频,通过目标视频观看到多人合作演奏的效果,从而进一步增加音乐演奏的趣味性。
示例性的,可以采用如下方式生成目标音乐对应的目标视频:获取视频模板以及每个多媒体片段的播放时间,所述视频模板中包括每个轨道对应的视频窗口;确定每个视频窗口对应的至少一个多媒体片段,针对任意一个视频窗口,根据所述视频窗口对应的至少一个多媒体片段的播放时间,对所述视频窗口对应的多媒体片段中的视频信息进行串接处理,得到所 述视频窗口对应的轨道视频,分别将每个视频窗口对应的轨道视频添加至对应的视频窗口,得到所述目标音乐对应的目标视频。
一个示例中,图10为本公开实施例提供的一种视频模板的示意图。如图10所示,该视频模板中包括3个视频窗口,其中,每个视频窗口用于播放一个轨道对应的轨道视频。例如,窗口1用于播放轨道A对应的轨道视频,窗口2用于播放轨道B对应的轨道视频,窗口3用于播放轨道C对应的轨道视频。
具体的,在生成目标视频时,按照轨道A的各时段顺序,将轨道A对应的至少一个多媒体片段中的视频信息进行串接,得到轨道A对应的轨道视频,并将轨道A对应的轨道视频关联到窗口1进行播放。按照轨道B的各时段顺序,将轨道B对应的至少一个多媒体片段中的视频信息进行串接,得到轨道B对应的轨道视频,并将轨道B对应的轨道视频关联到窗口2进行播放。按照轨道C的各时段顺序,将轨道C对应的至少一个多媒体片段中的视频信息进行串接,得到轨道C对应的轨道视频,并将轨道C对应的轨道视频关联到窗口3进行播放。
可选的,视频模板中还可以包括下述中的至少一项:片头、片尾、转场特效等。在生成目标视频时,可以在目标音乐的开头添加片头、目标音乐的结尾添加片尾、在时段之间进行切换时添加转场特效,从而提升目标视频的效果。
本实施例中,在生成目标音乐以及目标音乐对应的目标视频之后,服务器可以向终端设备发送目标音乐以及目标视频。这样终端设备可以同步播放目标音乐以及目标视频,从而,用户可以观看到多人合作演奏的效果,进一步增加音乐演奏的趣味性。
图11为本公开实施例提供的一种音乐生成装置的结构示意图。该装置可以为软件和/或硬件的形式。该装置可以为终端设备,或者为集成到终端设备中的处理器、芯片、芯片模组、模块、单元、应用程序等。
如图11所示,本实施例提供的音乐生成装置1100,包括:显示模块1101、生成模块1102和发送模块1103。
其中,显示模块1101,用于显示音乐模板对应的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
所述显示模块1101,还用于响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个;
生成模块1102,用于在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段;
发送模块1103,用于向服务器发送所述第一多媒体片段。
在一种可能的实现方式中,所述多媒体录制页面中包括开始控件和停止控件;所述生成模块1102具体用于:
响应于对所述开始控件的点击操作,开始录制多媒体信息,直至接收到对所述停止控件的点击操作或者在所述录制时段的结束时刻,停止录制多媒体信息,得到所述第一多媒体片段。
在一种可能的实现方式中,所述多媒体录制页面中还包括信息提示区域;所述显示模块1101还用于:
获取多媒体信息的录制进度;
根据所述录制进度,在所述信息提示区域中显示提示信息,所述提示信息包括如下至少一种:歌词、音符。
本实施例提供的音乐生成装置,可用于执行上述任一方法实施例由终端设备执行的音乐生成方法,其实现原理和技术效果类似,此处不作赘述。
图12为本公开实施例提供的另一种音乐生成装置的结构示意图。该装置可以为软件和/或硬件的形式。该装置可以为服务器,或者为集成到服务器中的处理器、芯片、芯片模组、模块、单元、应用程序等。
如图12所示,本实施例提供的音乐生成装置1200,包括:确定模块1201、获取模块1202、处理模块1203和生成模块1204。
其中,确定模块1201,用于确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
获取模块1202,用于根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;
处理模块1203,用于分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频;
生成模块1204,用于对各轨道对应的轨道音频进行合并处理,生成所述目标音乐。
在一种可能的实现方式中,针对所述多个轨道中的任意一个轨道,所述处理模块1203具体用于:
确定所述轨道的轨道类型,所述轨道类型为旋律轨道类型、伴奏轨道类型或者打击轨道类型;
根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
在一种可能的实现方式中,所述轨道类型为旋律轨道类型;所述处理模块1203具体用于:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第一处理,得到该时段对应的中间多媒体片段,所述第一处理包括如下至少一种:高音修正处理、加入混响处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
在一种可能的实现方式中,所述轨道类型为伴奏轨道类型;所述处理模块1203具体用于:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第二处理,得到该时段对应的中间多媒体片段,所述第二处理包括如下至少一种:高音修正处理、加入混响处理、增强低频处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
在一种可能的实现方式中,所述轨道类型为打击轨道类型;所述处理模块1203具体用于:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第三处理,得到该时段对应的中间多媒体片段,所述第三处理包括如下至少一种:高通滤波处理、加入混响处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
在一种可能的实现方式中,所述生成模块1204具体用于:
确定所述音乐模板对应的预设音量;
根据所述预设音量,对每个轨道对应的轨道音频进行音量均衡处理;
对所述音量均衡处理后的各轨道的轨道音频进行合并处理、以及降噪处理,得到所述目标音乐。
在一种可能的实现方式中,所述多媒体片段中还包括视频信息;所述获取模块1202还用于:获取视频模板以及每个多媒体片段的播放时间,所述视频模板中包括每个轨道对应的视频窗口;
所述处理模块1203还用于:确定每个视频窗口对应的至少一个多媒体片段;针对任意一个视频窗口,根据所述视频窗口对应的至少一个多媒体片段的播放时间,对所述视频窗口对应的多媒体片段中的视频信息进行串接处理,得到所述视频窗口对应的轨道视频;
所述生成模块1204还用于:分别将每个视频窗口对应的轨道视频添加至对应的视频窗口,得到所述目标音乐对应的目标视频。
在一种可能的实现方式中,所述装置还包括:
发送模块,用于向终端设备发送所述目标音乐和所述目标视频。
本实施例提供的音乐生成装置,可用于执行上述任一方法实施例中由服务器执行的音乐生成方法,其实现原理和技术效果类似,此处不作赘述。
为了实现上述实施例,本公开实施例还提供了一种电子设备。
参考图13,其示出了适于用来实现本公开实施例的电子设备1300的结构示意图,该电子设备1300可以为终端设备或服务器。其中,终端设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、个人数字助理(Personal Digital Assistant,简称PDA)、平板电脑(Portable Android Device,简称PAD)、便携式多媒体播放器(Portable Media Player,简称PMP)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图13示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。
如图13所示,电子设备1300可以包括处理装置(例如中央处理器、图形处理器等)1301,其可以根据存储在只读存储器(Read Only Memory,简称ROM)1302中的程序或者从存储装置1308加载到随机访问存储器(Random Access Memory,简称RAM)1303中的程序而执行各种适当的动作和处理。在RAM 1303中,还存储有电子设备1300操作所需的各种程序和数据。处理装置1301、ROM 1302以及RAM 1303通过总线1304彼此相连。输入/输出(I/O)接口1305也连接至总线1304。
通常,以下装置可以连接至I/O接口1305:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置1306;包括例如液晶显示器(Liquid Crystal Display,简称LCD)、扬声器、振动器等的输出装置1307;包括例如磁带、硬盘等的存储装置1308;以及通信装置1309。通信装置1309可以允许电子设备1300与其他设备进行无线或 有线通信以交换数据。虽然图13示出了具有各种装置的电子设备1300,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置1309从网络上被下载和安装,或者从存储装置1308被安装,或者从ROM 1302被安装。在该计算机程序被处理装置1301执行时,执行本公开实施例的方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(Erasable Programmable Read-Only Memory,简称EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(Compact Disc Read-Only Memory,简称CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、射频(Radio Frequence,简称RF)等等,或者上述的任意合适的组合。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备执行上述实施例所示的方法。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(Local Area Network,简称LAN)或广域网(Wide Area Network,简称WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实 现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的模块可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,模块的名称在某种情况下并不构成对该模块本身的限定,例如,确定模块还可以被描述为“确定音乐模板的模块”。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(Field-Programmable Gate Array,简称FPGA)、专用集成电路(Application Specific Integrated Circuit,简称ASIC)、专用标准产品(Application Specific Standard Product,简称ASSP)、片上系统(System On a Chip,简称SOC)、复杂可编程逻辑设备(Complex Programmable Logic Device,简称CPLD)等等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。
第一方面,根据本公开的一个或多个实施例,提供了一种音乐生成方法,包括:
确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;
分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频;
对各轨道对应的轨道音频进行合并处理,生成所述目标音乐。
根据本公开的一个或多个实施例,针对所述多个轨道中的任意一个轨道,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频,包括:
确定所述轨道的轨道类型,所述轨道类型为旋律轨道类型、伴奏轨道类型或者打击轨道类型;
根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,所述轨道类型为旋律轨道类型;根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频,包括:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第一处理,得到该时段对应的中间多媒体片段,所述第一处理包括如下至少一种:高音修正处理、加入混响处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,所述轨道类型为伴奏轨道类型;根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频,包括:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第二处理,得到该时段对应的中间多媒体片段,所述第二处理包括如下至少一种:高音修正处理、加入混响处理、增强低频处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,所述轨道类型为打击轨道类型;根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频,包括:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第三处理,得到该时段对应的中间多媒体片段,所述第三处理包括如下至少一种:高通滤波处理、加入混响处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,对各轨道对应的轨道音频进行合并处理,生成目标音乐,包括:
确定所述音乐模板对应的预设音量;
根据所述预设音量,对每个轨道对应的轨道音频进行音量均衡处理;
对所述音量均衡处理后的各轨道的轨道音频进行合并处理、以及降噪处理,得到所述目标音乐。
根据本公开的一个或多个实施例,所述多媒体片段中还包括视频信息;所述方法还包括:
获取视频模板以及每个多媒体片段的播放时间,所述视频模板中包括每个轨道对应的视频窗口;
确定每个视频窗口对应的至少一个多媒体片段;
针对任意一个视频窗口,根据所述视频窗口对应的至少一个多媒体片段的播放时间,对所述视频窗口对应的多媒体片段中的视频信息进行串接处理,得到所述视频窗口对应的轨道视频;
分别将每个视频窗口对应的轨道视频添加至对应的视频窗口,得到所述目标音乐对应的目标视频。
根据本公开的一个或多个实施例,还包括:
向终端设备发送所述目标音乐和所述目标视频。
第二方面,根据本公开的一个或多个实施例,提供了一种音乐生成方法,包括:
显示音乐模板的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个;
在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段,并向服务器发送所述第一多媒体片段。
根据本公开的一个或多个实施例,所述多媒体录制页面中包括开始控件和停止控件;在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段,包括:
响应于对所述开始控件的点击操作,开始录制多媒体信息,直至接收到对所述停止控件的点击操作或者在所述录制时段的结束时刻,停止录制多媒体信息,得到所述第一多媒体片段。
根据本公开的一个或多个实施例,所述多媒体录制页面中还包括信息提示区域;所述方法还包括:
获取多媒体信息的录制进度;
根据所述录制进度,在所述信息提示区域中显示提示信息,所述提示信息包括如下至少一种:歌词、音符。
第三方面,根据本公开的一个或多个实施例,提供了一种音乐生成装置,包括:
确定模块,用于确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
获取模块,用于根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;
处理模块,用于分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频;
生成模块,用于对各轨道对应的轨道音频进行合并处理,生成所述目标音乐。
根据本公开的一个或多个实施例,针对所述多个轨道中的任意一个轨道,所述处理模块具体用于:
确定所述轨道的轨道类型,所述轨道类型为旋律轨道类型、伴奏轨道类型或者打击轨道类型;
根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,所述轨道类型为旋律轨道类型;所述处理模块具体用于:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第一处理,得到该时段对应的中间多媒体片段,所述第一处理包括如下至少一种:高音修正处理、加入混响处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,所述轨道类型为伴奏轨道类型;所述处理模块具体用于:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第二处理,得到该时段对应的中间多媒体片段,所述第二处理包括如下至少一种:高音修正处理、加入混响处理、增强低频处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,所述轨道类型为打击轨道类型;所述处理模块具体用于:
分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第三处理,得到该时段对应的中间多媒体片段,所述第三处理包括如下至少一种:高通滤波处理、加入混响处理;
对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
根据本公开的一个或多个实施例,所述生成模块具体用于:
确定所述音乐模板对应的预设音量;
根据所述预设音量,对每个轨道对应的轨道音频进行音量均衡处理;
对所述音量均衡处理后的各轨道的轨道音频进行合并处理、以及降噪处理,得到所述目标音乐。
根据本公开的一个或多个实施例,所述多媒体片段中还包括视频信息;所述获取模块还用于:获取视频模板以及每个多媒体片段的播放时间,所述视频模板中包括每个轨道对应的视频窗口;
所述处理模块还用于:确定每个视频窗口对应的至少一个多媒体片段;针对任意一个视频窗口,根据所述视频窗口对应的至少一个多媒体片段的播放时间,对所述视频窗口对应的多媒体片段中的视频信息进行串接处理,得到所述视频窗口对应的轨道视频;
所述生成模块还用于:分别将每个视频窗口对应的轨道视频添加至对应的视频窗口,得到所述目标音乐对应的目标视频。
根据本公开的一个或多个实施例,所述装置还包括:
发送模块,用于向终端设备发送所述目标音乐和所述目标视频。
第四方面,根据本公开的一个或多个实施例,提供了一种音乐生成装置,包括:
显示模块,用于显示音乐模板对应的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
所述显示模块,还用于响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个;
生成模块,用于在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段;
发送模块,用于向服务器发送所述第一多媒体片段。
根据本公开的一个或多个实施例,所述多媒体录制页面中包括开始控件和停止控件;所述生成模块具体用于:
响应于对所述开始控件的点击操作,开始录制多媒体信息,直至接收到对所述停止控件的点击操作或者在所述录制时段的结束时刻,停止录制多媒体信息,得到所述第一多媒体片段。
根据本公开的一个或多个实施例,所述多媒体录制页面中还包括信息提示区域;所述显示模块还用于:
获取多媒体信息的录制进度;
根据所述录制进度,在所述信息提示区域中显示提示信息,所述提示信息包括如下至少一种:歌词、音符。
第五方面,根据本公开的一个或多个实施例,提供了一种电子设备,包括:至少一个处理器和存储器;
所述存储器存储计算机执行指令;
所述至少一个处理器执行所述存储器存储的计算机执行指令,使得所述至少一个处理器执行如上第一方面以及第一方面各种可能的实现方式所述的音乐生成方法,或者,执行如上第二方面以及第二方面各种可能的实现方式所述的音乐生成方法。
第六方面,根据本公开的一个或多个实施例,提供了一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如上第一方面以及第一方面各种可能的实现方式所述的音乐生成方法,或者,实现如上第二方面以及第二方面各种可能的实现方式所述的音乐生成方法。
第七方面,根据本公开的一个或多个实施例,提供了一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时,实现如第一方面以及第一方面各种可能的实现方式所述的音乐生成方法,或者,实现如上第二方面以及第二方面各种可能的实现方式所述的音乐生成方法。
第八方面,根据本公开的一个或多个实施例,提供了一种计算机程序,所述计算机程序在被处理器执行时,实现如上第一方面以及第一方面各种可能的实现方式所述的音乐生成方法,或者,实现如上第二方面以及第二方面各种可能的实现方式所述的音乐生成方法。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。

Claims (17)

  1. 一种音乐生成方法,包括:
    确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
    根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;
    分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频;
    对各轨道对应的轨道音频进行合并处理,生成目标音乐。
  2. 根据权利要求1所述的方法,其中,针对所述多个轨道中的任意一个轨道,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频,包括:
    确定所述轨道的轨道类型,所述轨道类型为旋律轨道类型、伴奏轨道类型或者打击轨道类型;
    根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
  3. 根据权利要求2所述的方法,其中,所述轨道类型为旋律轨道类型;根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频,包括:
    分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第一处理,得到该时段对应的中间多媒体片段,所述第一处理包括如下至少一种:高音修正处理、加入混响处理;
    对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
  4. 根据权利要求2所述的方法,其中,所述轨道类型为伴奏轨道类型;根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频,包括:
    分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第二处理,得到该时段对应的中间多媒体片段,所述第二处理包括如下至少一种:高音修正处理、加入混响处理、增强低频处理;
    对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频。
  5. 根据权利要求2所述的方法,其中,所述轨道类型为打击轨道类型;根据所述轨道的轨道类型,对所述轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到所述轨道对应的轨道音频,包括:
    分别对所述轨道的每个时段对应的多媒体片段中的音频信息进行第三处理,得到该时段对应的中间多媒体片段,所述第三处理包括如下至少一种:高通滤波处理、加入混响处理;
    对所述轨道的各时段对应的中间多媒体片段中的音频信息进行拼接处理,得到所述轨道 对应的轨道音频。
  6. 根据权利要求1至5任一项所述的方法,其中,对各轨道对应的轨道音频进行合并处理,生成目标音乐,包括:
    确定所述音乐模板对应的预设音量;
    根据所述预设音量,对每个轨道对应的轨道音频进行音量均衡处理;
    对所述音量均衡处理后的各轨道的轨道音频进行合并处理、以及降噪处理,得到所述目标音乐。
  7. 根据权利要求1至6任一项所述的方法,其中,所述多媒体片段中还包括视频信息;所述方法还包括:
    获取视频模板以及每个多媒体片段的播放时间,所述视频模板中包括每个轨道对应的视频窗口;
    确定每个视频窗口对应的至少一个多媒体片段;
    针对任意一个视频窗口,根据所述视频窗口对应的至少一个多媒体片段的播放时间,对所述视频窗口对应的多媒体片段中的视频信息进行串接处理,得到所述视频窗口对应的轨道视频;
    分别将每个视频窗口对应的轨道视频添加至对应的视频窗口,得到所述目标音乐对应的目标视频。
  8. 根据权利要求7所述的方法,还包括:
    向终端设备发送所述目标音乐和所述目标视频。
  9. 一种音乐生成方法,包括:
    显示音乐模板的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
    响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个;
    在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段,并向服务器发送所述第一多媒体片段。
  10. 根据权利要求9所述的方法,其中,所述多媒体录制页面中包括开始控件和停止控件;在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段,包括:
    响应于对所述开始控件的点击操作,开始录制多媒体信息,直至接收到对所述停止控件的点击操作或者在所述录制时段的结束时刻,停止录制多媒体信息,得到所述第一多媒体片段。
  11. 根据权利要求9或10所述的方法,其中,所述多媒体录制页面中还包括信息提示区域;所述方法还包括:
    获取多媒体信息的录制进度;
    根据所述录制进度,在所述信息提示区域中显示提示信息,所述提示信息包括如下至少一种:歌词、音符。
  12. 一种音乐生成装置,包括:
    确定模块,用于确定音乐模板,所述音乐模板包括多个轨道,每个轨道被划分为至少一 个时段,所述多个轨道为音乐的编奏轨道;
    获取模块,用于根据所述音乐模板,在数据库中获取每个轨道的每个时段对应的多媒体片段;其中,每个轨道的每个时段对应的多媒体片段包括:用户对该轨道中的该时段对应轨道旋律进行演奏得到的音频信息,所述数据库中包括多个用户演奏的多个多媒体片段;
    处理模块,用于分别对每个轨道的各时段对应的多媒体片段中的音频信息进行拼接处理,得到每个轨道对应的轨道音频;
    生成模块,用于对各轨道对应的轨道音频进行合并处理,生成目标音乐。
  13. 一种音乐生成装置,包括:
    显示模块,用于显示音乐模板对应的多个轨道,每个轨道被划分为至少一个时段,所述多个轨道为音乐的编奏轨道;
    所述显示模块,还用于响应于对第一时段的选择操作,显示所述第一时段对应的多媒体录制页面,所述第一时段为所述多个轨道对应的时段中的任意一个;
    生成模块,用于在所述多媒体录制页面对应的录制时段内录制多媒体信息,得到所述第一时段对应的第一多媒体片段;
    发送模块,用于向服务器发送所述第一多媒体片段。
  14. 一种电子设备,包括:处理器和存储器;
    所述存储器存储计算机执行指令;
    所述处理器执行所述计算机执行指令,实现如权利要求1至8任一项所述的方法,或者,如权利要求9至11任一项所述的方法。
  15. 一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如权利要求1至8任一项所述的方法,或者,如权利要求9至11任一项所述的方法。
  16. 一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时,实现如权利要求1至8任一项所述的方法,或者,如权利要求9至11任一项所述的方法。
  17. 一种计算机程序,所述计算机程序被处理器执行时,实现如权利要求1至8任一项所述的方法,或者,如权利要求9至11任一项所述的方法。
PCT/CN2023/078545 2022-02-28 2023-02-27 音乐生成方法、装置、设备、存储介质及程序 WO2023160713A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210190764.6 2022-02-28
CN202210190764.6A CN116704978A (zh) 2022-02-28 2022-02-28 音乐生成方法、装置、设备、存储介质及程序

Publications (1)

Publication Number Publication Date
WO2023160713A1 true WO2023160713A1 (zh) 2023-08-31

Family

ID=87764892

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/078545 WO2023160713A1 (zh) 2022-02-28 2023-02-27 音乐生成方法、装置、设备、存储介质及程序

Country Status (2)

Country Link
CN (1) CN116704978A (zh)
WO (1) WO2023160713A1 (zh)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5464946A (en) * 1993-02-11 1995-11-07 Multimedia Systems Corporation System and apparatus for interactive multimedia entertainment
US20060180007A1 (en) * 2005-01-05 2006-08-17 Mcclinsey Jason Music and audio composition system
US20090164034A1 (en) * 2007-12-19 2009-06-25 Dopetracks, Llc Web-based performance collaborations based on multimedia-content sharing
US20110126103A1 (en) * 2009-11-24 2011-05-26 Tunewiki Ltd. Method and system for a "karaoke collage"
CN102568460A (zh) * 2010-12-28 2012-07-11 张泽铭 联网多人异步合唱影音作品系统及实施方法
CN103295568A (zh) * 2013-05-30 2013-09-11 北京小米科技有限责任公司 一种异步合唱方法和装置
JP2017156686A (ja) * 2016-03-04 2017-09-07 ヤマハ株式会社 録音システム、録音方法及びプログラム
CN107665703A (zh) * 2017-09-11 2018-02-06 上海与德科技有限公司 一种多用户的音频合成方法和系统以及远程服务器
CN108630240A (zh) * 2017-03-23 2018-10-09 北京小唱科技有限公司 一种合唱方法及装置
US20210407479A1 (en) * 2020-10-27 2021-12-30 Beijing Baidu Netcom Science And Technology Co., Ltd. Method for song multimedia synthesis, electronic device and storage medium

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5464946A (en) * 1993-02-11 1995-11-07 Multimedia Systems Corporation System and apparatus for interactive multimedia entertainment
US20060180007A1 (en) * 2005-01-05 2006-08-17 Mcclinsey Jason Music and audio composition system
US20090164034A1 (en) * 2007-12-19 2009-06-25 Dopetracks, Llc Web-based performance collaborations based on multimedia-content sharing
US20110126103A1 (en) * 2009-11-24 2011-05-26 Tunewiki Ltd. Method and system for a "karaoke collage"
CN102568460A (zh) * 2010-12-28 2012-07-11 张泽铭 联网多人异步合唱影音作品系统及实施方法
CN103295568A (zh) * 2013-05-30 2013-09-11 北京小米科技有限责任公司 一种异步合唱方法和装置
JP2017156686A (ja) * 2016-03-04 2017-09-07 ヤマハ株式会社 録音システム、録音方法及びプログラム
CN108630240A (zh) * 2017-03-23 2018-10-09 北京小唱科技有限公司 一种合唱方法及装置
CN107665703A (zh) * 2017-09-11 2018-02-06 上海与德科技有限公司 一种多用户的音频合成方法和系统以及远程服务器
US20210407479A1 (en) * 2020-10-27 2021-12-30 Beijing Baidu Netcom Science And Technology Co., Ltd. Method for song multimedia synthesis, electronic device and storage medium

Also Published As

Publication number Publication date
CN116704978A (zh) 2023-09-05

Similar Documents

Publication Publication Date Title
Collins et al. Electronic music
WO2020113733A1 (zh) 动画生成方法、装置、电子设备及计算机可读存储介质
US9472178B2 (en) Score-directed string retuning and gesture cueing in synthetic multi-string musical instrument
US10504498B2 (en) Real-time jamming assistance for groups of musicians
US20110023689A1 (en) Systems and methods for generating a game device music track from music
WO2020259130A1 (zh) 精选片段处理方法、装置、电子设备及可读介质
JP2019219638A (ja) 音楽合成方法及びシステム、端末並びにコンピュータ可読記憶媒体
WO2023051246A1 (zh) 视频录制方法、装置、设备及存储介质
US20120072841A1 (en) Browser-Based Song Creation
WO2022160603A1 (zh) 歌曲的推荐方法、装置、电子设备及存储介质
JP2023534975A (ja) ユーザーインタラクションに基づく音楽再生方法、装置、機器及び記憶媒体
Jackson Digital audio editing fundamentals
WO2023160713A1 (zh) 音乐生成方法、装置、设备、存储介质及程序
AU2019207800A1 (en) Musical activity system
JP5969421B2 (ja) 楽器音出力装置及び楽器音出力プログラム
Maz Music Technology Essentials: A Home Studio Guide
JP5561263B2 (ja) 楽音再生装置及びプログラム
JP5510207B2 (ja) 楽音編集装置及びプログラム
WO2024012257A1 (zh) 音频处理方法、装置及电子设备
WO2024066790A1 (zh) 音频处理方法、装置及电子设备
WO2023217003A1 (zh) 音频处理方法、装置、设备及存储介质
KR20140054810A (ko) 반주음악 제작 서비스 시스템 및 그 방법, 그리고 이에 적용되는 장치
WO2023051651A1 (zh) 音乐生成方法、装置、设备、存储介质及程序
WO2023010949A1 (zh) 一种音频数据的处理方法及装置
Heavner The Applied Music Studio: Teaching Students With Special Needs

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23759334

Country of ref document: EP

Kind code of ref document: A1