Detailed Description
Exemplary embodiments of the present disclosure are described below in conjunction with the accompanying drawings, which include various details of the embodiments of the present disclosure to facilitate understanding, and should be considered as merely exemplary. Accordingly, one of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the foregoing figures are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that the embodiments of the disclosure described herein may be capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
In accordance with the disclosed embodiments, a method of generating a work is provided, it being noted that the steps illustrated in the flowcharts of the figures may be performed in a computer system, such as a set of computer executable instructions, and that although a logical order is illustrated in the flowcharts, in some cases the steps illustrated or described may be performed in an order other than that illustrated herein.
The method embodiments provided by the embodiments of the present disclosure may be performed in a mobile terminal, a computer terminal, or similar electronic device. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular telephones, smartphones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the disclosure described and/or claimed herein. Fig. 1 shows a block diagram of a hardware configuration of a computer terminal (or mobile device) for implementing a method of generating a work.
As shown in fig. 1, the computer terminal 100 includes a computing unit 101 that can perform various appropriate actions and processes according to a computer program stored in a Read Only Memory (ROM) 102 or a computer program loaded from a storage unit 108 into a Random Access Memory (RAM) 103. In the RAM 103, various programs and data required for the operation of the computer terminal 100 can also be stored. The computing unit 101, ROM 102, and RAM 103 are connected to each other by a bus 104. An input/output (I/O) interface 105 is also connected to bus 104.
Various components in the computer terminal 100 are connected to the I/O interface 105, including an input unit 106 such as a keyboard, a mouse, etc., an output unit 107 such as various types of displays, speakers, etc., a storage unit 108 such as a magnetic disk, an optical disk, etc., and a communication unit 109 such as a network card, a modem, a wireless communication transceiver, etc. The communication unit 109 allows the computer terminal 100 to exchange information/data with other devices through a computer network such as the internet and/or various telecommunication networks.
The computing unit 101 may be a variety of general and/or special purpose processing components having processing and computing capabilities. Some examples of computing unit 101 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various specialized Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a Digital Signal Processor (DSP), and any suitable processor, controller, microcontroller, etc. The computing unit 101 performs the generation method of the work described herein. For example, in some embodiments, the method of generating a work may be implemented as a computer software program tangibly embodied on a machine-readable medium, such as the storage unit 108. In some embodiments, part or all of the computer program may be loaded and/or installed onto the computer terminal 100 via the ROM 102 and/or the communication unit 109. When the computer program is loaded into RAM 103 and executed by computing unit 101, one or more steps of the methods of generating a work described herein may be performed. Alternatively, in other embodiments, the computing unit 101 may be configured to perform the method of generating the work in any other suitable manner (e.g., by means of firmware).
Various implementations of the systems and techniques described here can be implemented in digital electronic circuitry, integrated circuitry, field Programmable Gate Arrays (FPGAs), application Specific Integrated Circuits (ASICs), application Specific Standard Products (ASSPs), systems On Chip (SOCs), load programmable logic devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include being implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be a special or general purpose programmable processor, operable to receive data and instructions from, and to transmit data and instructions to, a storage system, at least one input device, and at least one output device.
It should be noted here that, in some alternative embodiments, the electronic device shown in fig. 1 described above may include hardware elements (including circuits), software elements (including computer code stored on a computer readable medium), or a combination of both hardware and software elements. It should be noted that fig. 1 is only one example of a specific example, and is intended to illustrate the types of components that may be present in the above-described electronic devices.
In the above-described operating environment, the present disclosure provides a method of generating a work as shown in fig. 2, which may be performed by a computer terminal or similar electronic device as shown in fig. 1. Fig. 2 is a flowchart of a method of generating a work provided in accordance with an embodiment of the present disclosure. As shown in fig. 2, the method may include the steps of:
step S202, responding to an operation instruction for a first control in a first page, acquiring first audio, and jumping to a second page.
Wherein the first page includes at least one candidate audio index and the second page is used to edit text.
The first audio may be accompaniment audio.
The first page may be a page for selecting accompaniment audio, and may include a plurality of candidate audio, where each candidate audio corresponds to a first control, or an index of each candidate audio corresponds to a first control.
The first control may be a confirm button.
In an alternative embodiment, if the user wants to record a song, the user may enter a first page, where the first page includes indexes of a plurality of candidate audios, and the user may select a desired audio as an accompaniment audio according to the indexes of the plurality of candidate audios, and click a first control corresponding to the accompaniment audio index, so as to retrieve the accompaniment audio from the database according to the accompaniment audio index.
Further, after the accompaniment audio is acquired, a jump may be made to the second page to edit the lyrics of the work in the second page.
In an alternative embodiment, the second page may be used to edit lyrics of the work, and after the accompaniment audio is obtained, the page for editing the lyrics may be entered to edit the lyrics.
And step S204, acquiring a first text in the second page, and responding to an operation instruction for the second control in the second page, and jumping to a third page.
Wherein the third page is used for providing a recording audio function.
The second page may include a text box having edited text in which the user may enter the first text.
The second control may be a text input completion button.
The third page may be used to record the voice of the user.
In an alternative embodiment, the lyrics edited by the user may be obtained in the second page, and after the user finishes editing the lyrics, the user may press a button for completing text input, that is, the second control mentioned above, to jump to the page for recording the user's voice.
In another alternative embodiment, the lower word count limit and the upper word count limit of the first word may be set, and the second control may be displayed when the word count exceeds the preset word count, so that the user may click on the second control to perform the next audio recording.
In another alternative embodiment, after the user edits the text, the user may also process the edited text through a word processing control in the third page, e.g., the text that is not rhymed may be labeled, so that the user improves the edited text. After clicking the word processing control, the user can also display the word group associated with the word being edited in the word processing control for the user to select, so as to give the user the inspiration of creating the text.
Step S206, obtaining second audio in the third page, and synthesizing the first audio, the first text and the second audio in response to the operation instruction of the third control in the third page to generate a first work.
The third page may have a recording button, and the user may press the recording button to record the sound.
The third control described above may be a control for composition.
In an alternative embodiment, after the user records the audio in the third page, the user may press a control for synthesizing in the third page to synthesize the accompaniment audio, the edited lyrics, and the recorded audio, and generate the first work.
In another alternative embodiment, the user may record the sound by wearing a headset, after the recording is completed, may click a listening test button in the third page, where during listening test, the user may click to play the accompaniment audio, and if the user feels that the recorded sound and the accompaniment audio are not aligned or the rhythm is not aligned, the user may manually splice and cut the recorded audio, and generate the first work according to the spliced and cut audio.
The work generation method described above may be applied to a generation scene of a rap work.
Through the steps, the first audio can be obtained in response to the operation instruction of the first control in the first page, and the first audio is jumped to the second page, wherein the first page comprises at least one candidate audio index, the second page is used for editing texts, the first text is obtained in the second page, the third page is jumped to the third page in response to the operation instruction of the second control in the second page, the third page is used for providing the audio recording function, the second audio is obtained in the third page, the first audio, the first text and the second audio are synthesized in response to the operation instruction of the third control in the third page, the first work is generated, the purpose of creating the work is achieved, the steps of the creation process can be simplified through executing different types of creation links in different pages, the flow of the creation process is achieved, the creation efficiency is improved, and the technical problem that the creation efficiency of the complex work is too low due to the creation process in the related technology is further improved.
Optionally, after the first work is generated based on the first audio, the first text and the second audio, the method further comprises jumping to a fourth page, acquiring adjustment parameters in the fourth page, and adjusting at least one sound effect of the first work based on the adjustment parameters to obtain an adjusted first work.
The fourth page is used for adjusting the sound effect of the generated first work.
The above-mentioned sound effects may be equalization, reverberation, electrical sound, gas sound, harmony, delay, etc.
The adjustment parameter may be obtained by dragging an adjustment frame corresponding to at least one sound effect in the fourth page by the user.
The fourth page may be a post-chamber, where the post-chamber is configured to optimize the first work to obtain a more optimal first work.
In an alternative embodiment, after jumping to the fourth page with sound effects adjusted, the user may adjust the sound effects in an adjusting frame corresponding to each sound effect in the fourth page to obtain an adjusting parameter, and after obtaining the adjusting parameter, may adjust at least one sound effect of the first work according to the adjusting parameter to obtain the adjusted first work.
Optionally, after jumping to the fourth page, the method further comprises jumping to a fifth page in response to an operation instruction of the fourth page for the fourth control, wherein the fifth page comprises at least one candidate video index, acquiring a first video in response to an operation instruction of the fifth page for the fifth control, jumping back to the fourth page, and combining the adjusted first work and the first video in response to an operation instruction of the fourth page for the sixth control to generate the first work.
The fourth control may be a control to add video.
The fifth page may include a plurality of video indexes, where each video index corresponds to a fourth control.
In an alternative embodiment, after jumping to the later room, the method may respond to an operation instruction for adding the video control in the fourth page, jump to the fifth page, and the user may select the first video according to at least one video index in the fifth page, specifically, the user may click on the fifth control corresponding to the video index, so as to obtain the first video corresponding to the video index.
The sixth control may be a composite control in the fourth page.
The first video can be used for baking the atmosphere of the work, so that the work can be better presented.
In another alternative embodiment, after the first video is acquired, the fourth page may be skipped, and the sixth control is clicked on the fourth page, so that the adjusted first work and the first video are synthesized, and the first work is generated, so that the first work is more complete.
In another alternative embodiment, after the first video is acquired, the first video may be further processed, for example, special effects may be performed on the first video. Specifically, the user may select a video special effect template to add to the first video.
Optionally, after jumping back to the fourth page, the method further comprises the steps of synthesizing the first video and the first text in the fourth page to obtain a second video, and responding to an operation instruction of a sixth control in the fourth page to synthesize the adjusted first work and the second video to generate the first work.
The first text may be edited lyrics.
In an alternative embodiment, in jumping back to the fourth page, the first video and the first text may be synthesized in the fourth page to obtain the second video, where the second video is a video containing lyrics, and since the second video contains lyrics of a recorded song, the user can watch the lyrics while watching the second video, so as to improve the ornamental value of the generated first work.
In another alternative embodiment, after synthesizing the first video and the first text to obtain the second video, the user may adjust the first text in the second video so that the display of the first text may be aligned with the sound of the recorded song.
In another alternative embodiment, the lyrics fonts, colors and the like in the second video can also be set, so that the second video is more attractive.
Optionally, obtaining the second audio in the third page, synthesizing the first audio, the first text and the second audio in response to an operation instruction of the third control in the third page, and generating the first work, wherein the obtaining the clipping parameter in the third page, clipping the second audio based on the clipping parameter to obtain the third audio, and synthesizing the first audio, the first text and the third audio in response to the operation instruction of the third control in the third page, and generating the first work.
In an alternative embodiment, after the second audio is acquired in the third page, the recorded audio may be clipped in the third page according to the clipping parameters, and the unwanted sound may be removed, or the sound to be retained may be selected, so as to obtain the third audio. The clipping parameters may be obtained by manually clipping the acquired second audio by the user. After the third audio is obtained, the user may click on a synthesis control in the third page, that is, the third control, to synthesize the first audio, the first text, and the third audio, and generate the first work.
In another alternative embodiment, the user may record the sound by wearing a headset, after the recording is completed, may click a listening test button in the third page, where during listening test, the user may click to play the accompaniment audio, and if the user feels that the recorded sound and the accompaniment audio are not aligned or the rhythm is not aligned, the user may manually splice and cut the recorded audio to obtain the above-mentioned clipping parameters, and clip the second audio according to the clipping parameters to obtain the third audio. And obtaining the first work based on the generated third audio, so that the obtained first work is more complete.
Optionally, obtaining the second audio in the third page, synthesizing the first audio, the first text and the second audio in response to an operation instruction of the third control in the third page, and generating the first work, wherein the obtaining the timestamp of the second audio in the third page, marking the first text by the timestamp to obtain the second text, and synthesizing the first audio, the second text and the third audio in response to an operation instruction of the third control in the third page, and generating the first work.
In an alternative embodiment, after the second audio is obtained in the third page, a timestamp corresponding to the second audio may be obtained, where the timestamp is used to indicate a playing time of each piece of sound in the second audio, and by way of example, when there are 10 pieces of sound in the second audio, a time when the first sentence of sound starts to be recorded may be determined as a time when playing starts.
In another alternative embodiment, the audio may be segmented according to sound pauses in the second audio.
Since the second audio corresponds to the first text, the second audio may be matched to the first text using a timestamp to determine lyrics corresponding to each piece of sound in the second audio.
In another alternative embodiment, after the timestamp of the second audio is obtained, the timestamp may be used to mark the first text to obtain the second text, where the second text corresponds to the playing time of the second audio. After the second text is obtained, the user may click on a synthesis control in the third page, that is, the third control described above, and then synthesize the first audio, the second text, and the third audio based on the synthesis control, to generate the first work.
An embodiment of a method of generating a work of the present disclosure is described in detail below in conjunction with FIG. 3. The method comprises the following steps:
step S301, selecting accompaniment audio;
step S302, editing lyrics;
step S303, recording audio;
step S304, judging whether the time for recording the audio is more than 3S, if so, executing step S306, otherwise, executing step S305;
Step S305, re-clicking recording;
Step S306, editing the recorded audio to obtain edited audio;
Step S307, judging whether the voice in the clipped audio needs to be adjusted, if so, executing step S308, and if not, executing step S309;
step S308, the sliding progress bar changes the audio time to be adjusted, and adjusts the audio;
step S309, listening the audio, and determining that the audio is finished;
optionally, after the audio is produced, the produced audio may be uploaded to the local.
Step S310, labeling lyrics to the manufactured audio;
Step S311, jumping to a later room, and selecting a video corresponding to the audio;
Step S312, modifying the video, lyrics and audio;
Optionally, a video special effect template can be selected to modify the video, the font and the color of the lyrics are selected to modify the lyrics, and the audio is modified through at least one sound effect.
The above-mentioned sound effects may be equalization, reverberation, electrical sound, gas sound, harmony, delay, etc.
Step S313, synthesizing the modified video, lyrics and audio to generate a work;
step S314, the generated work is adjusted to generate a complete work.
Further, a video file corresponding to the complete work may also be generated.
Optionally, the works can be adjusted by setting filters, subtitles, width and height, an audio track set, a rap audio track, an accompaniment audio track and an empty audio track can be set for the works, so that the effects of audio noise reduction and gain of the works are ensured, and the subtitles, the mapping and the like can be synthesized for the works.
Through the steps, the complex audio production can be simplified, the music lovers can conveniently and low-threshold complete own works, good gain effects are brought to the works, more music lovers are attracted to produce, and the utilization rate of music production software is improved.
According to the embodiment of the present disclosure, a device for generating a work is further provided, where the device may execute the method for generating a work in the foregoing embodiment, and a specific implementation manner and a preferred application scenario are the same as those of the foregoing embodiment, and are not described herein.
FIG. 4 is a schematic diagram of a production device for a work, as shown in FIG. 4, according to an embodiment of the disclosure, the device comprising:
An obtaining module 402, configured to obtain a first audio in response to an operation instruction for a first control in a first page, and skip to a second page, where the first page includes at least one candidate audio index, and the second page is used for editing text;
A skip module 404, configured to obtain a first text from the second page, and skip to a third page in response to an operation instruction for the second control in the second page, where the third page is used to provide a function of recording audio;
And the synthesizing module 406 is configured to obtain the second audio in the third page, and synthesize the first audio, the first text, and the second audio in response to an operation instruction for the third control in the third page, so as to generate the first work.
Optionally, the device further comprises an adjusting module, which is used for jumping to a fourth page, acquiring adjusting parameters in the fourth page, and adjusting at least one sound effect of the first work based on the adjusting parameters to obtain an adjusted first work.
Optionally, the skip module is further configured to skip to a fifth page in response to an operation instruction for the fourth control in the fourth page, where the fifth page includes at least one candidate video index, the obtaining module is further configured to obtain a first video in response to an operation instruction for the fifth control in the fifth page, skip back to the fourth page, and the synthesizing module is further configured to synthesize the adjusted first work and the first video in response to an operation instruction for the sixth control in the fourth page, so as to generate the first work.
Optionally, the device further comprises a synthesizing module for synthesizing the first video and the first text in the fourth page to obtain the second video, and a generating module for responding to the operation instruction of the sixth control in the fourth page to synthesize the adjusted first work and the second video to generate the first work.
Optionally, the synthesizing module comprises a clipping unit for obtaining clipping parameters in the third page, clipping the second audio based on the clipping parameters to obtain third audio, and a synthesizing unit for synthesizing the first audio, the first text and the third audio to generate the first work in response to an operation instruction of the third control in the third page.
Optionally, the synthesizing unit comprises a marking subunit, a synthesizing subunit and a synthesizing subunit, wherein the marking subunit is used for obtaining the timestamp of the second audio in the third page, marking the first text by using the timestamp to obtain the second text, and the synthesizing subunit is used for responding to the operation instruction of the third control in the third page to synthesize the first audio, the second text and the third audio to generate the first work.
It should be noted that each of the above modules may be implemented by software or hardware, and the latter may be implemented by, but not limited to, the above modules all being located in the same processor, or each of the above modules being located in different processors in any combination.
There is further provided, in accordance with an embodiment of the present disclosure, an electronic device including at least one processor, and a memory communicatively coupled to the at least one processor, wherein the memory stores instructions executable by the at least one processor to enable the at least one processor to perform a method of generating any of the works described above.
There is also provided, in accordance with an embodiment of the present disclosure, a non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform a method of generating any one of the works described above.
Alternatively, in the present embodiment, the non-transitory computer readable storage medium described above may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
There is also provided, in accordance with an embodiment of the present disclosure, a computer program product comprising a computer program which, when executed by a processor, implements a method of generating any of the above works. Program code for carrying out the methods of generating a work of the present disclosure may be written in any combination of one or more programming languages. These program code may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus such that the program code, when executed by the processor or controller, causes the functions/operations specified in the flowchart and/or block diagram to be implemented. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package, partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the foregoing embodiments of the present disclosure, the descriptions of the various embodiments are emphasized, and for a portion of this disclosure that is not described in detail in this embodiment, reference is made to the related descriptions of other embodiments.
In the several embodiments provided in the present disclosure, it should be understood that the disclosed technology content may be implemented in other manners. The above-described embodiments of the apparatus are merely exemplary, and the division of the units, for example, may be a logic function division, and may be implemented in another manner, for example, a plurality of units or components may be combined or may be integrated into another system, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be through some interfaces, units or modules, or may be in electrical or other forms.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in each embodiment of the present disclosure may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated units, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present disclosure may be embodied in essence or a part contributing to the prior art or all or part of the technical solution in the form of a software product stored in a storage medium, including several instructions to cause a computer device (which may be a personal computer, a server or a network device, etc.) to perform all or part of the steps of the method described in the embodiments of the present disclosure. The storage medium includes a U disk, a read-only memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic disk, or an optical disk, etc. which can store program codes.
The foregoing is merely a preferred embodiment of the present disclosure, and it should be noted that modifications and adaptations to those skilled in the art may be made without departing from the principles of the present disclosure, which are intended to be comprehended within the scope of the present disclosure.