CN113963674B - Method, device, electronic device and storage medium for generating works - Google Patents

Method, device, electronic device and storage medium for generating works Download PDF

Info

Publication number
CN113963674B
CN113963674B CN202111161807.XA CN202111161807A CN113963674B CN 113963674 B CN113963674 B CN 113963674B CN 202111161807 A CN202111161807 A CN 202111161807A CN 113963674 B CN113963674 B CN 113963674B
Authority
CN
China
Prior art keywords
page
audio
work
control
operation instruction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111161807.XA
Other languages
Chinese (zh)
Other versions
CN113963674A (en
Inventor
郭一鸣
邹贵宝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202111161807.XA priority Critical patent/CN113963674B/en
Publication of CN113963674A publication Critical patent/CN113963674A/en
Application granted granted Critical
Publication of CN113963674B publication Critical patent/CN113963674B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0025Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/958Organisation or management of web site content, e.g. publishing, maintaining pages or automatic linking
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Management Or Editing Of Information On Record Carriers (AREA)

Abstract

本公开提供了一种作品的生成方法、装置、电子设备及存储介质,涉及音频处理技术领域,以至少解决相关技术中由于创作过程复杂导致作品生成效率较低的技术问题。具体实现方案为:响应于第一页面中对于第一控件的操作指令,获取第一音频,并跳转至第二页面,其中,第一页面包括至少一个候选的音频索引,第二页面用于编辑文本;在第二页面中获取第一文本,响应于第二页面中对于第二控件的操作指令,跳转至第三页面,其中,第三页面用于提供录制音频功能;在第三页面中获取第二音频,响应于第三页面中对于第三控件的操作指令,对所述第一音频、所述第一文本以及所述第二音频进行合成,生成第一作品。

The present disclosure provides a method, device, electronic device and storage medium for generating a work, which relates to the field of audio processing technology, in order to at least solve the technical problem in the related technology that the efficiency of generating a work is low due to the complex creation process. The specific implementation scheme is: in response to the operation instruction for the first control in the first page, the first audio is obtained, and the second page is jumped to, wherein the first page includes at least one candidate audio index, and the second page is used to edit the text; the first text is obtained in the second page, and in response to the operation instruction for the second control in the second page, the third page is jumped to, wherein the third page is used to provide a recording audio function; the second audio is obtained in the third page, and in response to the operation instruction for the third control in the third page, the first audio, the first text and the second audio are synthesized to generate the first work.

Description

Method, device, electronic equipment and storage medium for generating works
Technical Field
The present disclosure relates to the field of audio processing, and in particular, to a method and apparatus for generating a work, an electronic device, and a storage medium.
Background
At present, music creation has a certain threshold, and if a music fan wants to have a good work, the music fan needs not only good singing work, but also good equipment and the ability of proficiency in using music making software to make music. However, the operation flow of the current music production software is complex, and the music production software is usually required to be matched with professional equipment.
Disclosure of Invention
The disclosure provides a method, a device, electronic equipment and a storage medium for generating works, so as to at least solve the technical problem that the work creation efficiency is too low due to the complex creation process in the related technology.
According to one aspect of the disclosure, a work generating method is provided, which comprises the steps of responding to an operation instruction of a first control in a first page, acquiring first audio, and jumping to a second page, wherein the first page comprises at least one candidate audio index, the second page is used for editing text, acquiring the first text in the second page, responding to the operation instruction of the second control in the second page, jumping to a third page, wherein the third page is used for providing a recording audio function, acquiring second audio in the third page, responding to the operation instruction of the third control in the third page, and synthesizing the first audio, the first text and the second audio to generate the first work.
According to still another aspect of the disclosure, a generation module of a work is further provided, which comprises an acquisition module for acquiring first audio in response to an operation instruction for a first control in a first page and jumping to a second page, wherein the first page comprises at least one candidate audio index, the second page is used for editing text, a jumping module for acquiring the first text in the second page and jumping to a third page in response to the operation instruction for the second control in the second page, wherein the third page is used for providing a recording audio function, and a synthesis module for acquiring the second audio in the third page, and synthesizing the first audio, the first text and the second audio in response to the operation instruction for the third control in the third page to generate the first work.
According to yet another aspect of the present disclosure, there is provided an electronic device comprising at least one processor, and a memory communicatively coupled to the at least one processor, wherein the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of generating a work as set forth in the present disclosure.
According to yet another aspect of the present disclosure, there is provided a non-transitory computer-readable storage medium storing computer instructions for causing a computer to execute the method of generating a work set forth in the present disclosure.
According to yet another aspect of the present disclosure, there is provided a computer program product comprising a computer program which, when executed by a processor, performs a method of generating a work as set forth in the present disclosure.
In the disclosure, a first audio may be acquired in response to an operation instruction for a first control in a first page, and a jump may be made to a second page, where the first page includes at least one candidate audio index, the second page is used for editing text, the first text is acquired in the second page, the third page is used for providing a function of recording audio in response to an operation instruction for the second control in the second page, the second audio is acquired in the third page, and the first audio, the first text and the second audio are synthesized in response to an operation instruction for the third control in the third page, so as to generate a first work.
It should be understood that the description in this section is not intended to identify key or critical features of the embodiments of the disclosure, nor is it intended to be used to limit the scope of the disclosure. Other features of the present disclosure will become apparent from the following specification.
Drawings
The drawings are for a better understanding of the present solution and are not to be construed as limiting the present disclosure. Wherein:
FIG. 1 is a block diagram of a hardware architecture of a computer terminal (or mobile device) for implementing a method of generating a work according to an embodiment of the present disclosure;
FIG. 2 is a flow chart of a method of generating a work according to an embodiment of the present disclosure;
FIG. 3 is a flow chart of another method of generating a work according to an embodiment of the present disclosure;
fig. 4 is a schematic diagram of a work generating apparatus according to an embodiment of the present disclosure.
Detailed Description
Exemplary embodiments of the present disclosure are described below in conjunction with the accompanying drawings, which include various details of the embodiments of the present disclosure to facilitate understanding, and should be considered as merely exemplary. Accordingly, one of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the foregoing figures are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that the embodiments of the disclosure described herein may be capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
In accordance with the disclosed embodiments, a method of generating a work is provided, it being noted that the steps illustrated in the flowcharts of the figures may be performed in a computer system, such as a set of computer executable instructions, and that although a logical order is illustrated in the flowcharts, in some cases the steps illustrated or described may be performed in an order other than that illustrated herein.
The method embodiments provided by the embodiments of the present disclosure may be performed in a mobile terminal, a computer terminal, or similar electronic device. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular telephones, smartphones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the disclosure described and/or claimed herein. Fig. 1 shows a block diagram of a hardware configuration of a computer terminal (or mobile device) for implementing a method of generating a work.
As shown in fig. 1, the computer terminal 100 includes a computing unit 101 that can perform various appropriate actions and processes according to a computer program stored in a Read Only Memory (ROM) 102 or a computer program loaded from a storage unit 108 into a Random Access Memory (RAM) 103. In the RAM 103, various programs and data required for the operation of the computer terminal 100 can also be stored. The computing unit 101, ROM 102, and RAM 103 are connected to each other by a bus 104. An input/output (I/O) interface 105 is also connected to bus 104.
Various components in the computer terminal 100 are connected to the I/O interface 105, including an input unit 106 such as a keyboard, a mouse, etc., an output unit 107 such as various types of displays, speakers, etc., a storage unit 108 such as a magnetic disk, an optical disk, etc., and a communication unit 109 such as a network card, a modem, a wireless communication transceiver, etc. The communication unit 109 allows the computer terminal 100 to exchange information/data with other devices through a computer network such as the internet and/or various telecommunication networks.
The computing unit 101 may be a variety of general and/or special purpose processing components having processing and computing capabilities. Some examples of computing unit 101 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various specialized Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a Digital Signal Processor (DSP), and any suitable processor, controller, microcontroller, etc. The computing unit 101 performs the generation method of the work described herein. For example, in some embodiments, the method of generating a work may be implemented as a computer software program tangibly embodied on a machine-readable medium, such as the storage unit 108. In some embodiments, part or all of the computer program may be loaded and/or installed onto the computer terminal 100 via the ROM 102 and/or the communication unit 109. When the computer program is loaded into RAM 103 and executed by computing unit 101, one or more steps of the methods of generating a work described herein may be performed. Alternatively, in other embodiments, the computing unit 101 may be configured to perform the method of generating the work in any other suitable manner (e.g., by means of firmware).
Various implementations of the systems and techniques described here can be implemented in digital electronic circuitry, integrated circuitry, field Programmable Gate Arrays (FPGAs), application Specific Integrated Circuits (ASICs), application Specific Standard Products (ASSPs), systems On Chip (SOCs), load programmable logic devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include being implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be a special or general purpose programmable processor, operable to receive data and instructions from, and to transmit data and instructions to, a storage system, at least one input device, and at least one output device.
It should be noted here that, in some alternative embodiments, the electronic device shown in fig. 1 described above may include hardware elements (including circuits), software elements (including computer code stored on a computer readable medium), or a combination of both hardware and software elements. It should be noted that fig. 1 is only one example of a specific example, and is intended to illustrate the types of components that may be present in the above-described electronic devices.
In the above-described operating environment, the present disclosure provides a method of generating a work as shown in fig. 2, which may be performed by a computer terminal or similar electronic device as shown in fig. 1. Fig. 2 is a flowchart of a method of generating a work provided in accordance with an embodiment of the present disclosure. As shown in fig. 2, the method may include the steps of:
step S202, responding to an operation instruction for a first control in a first page, acquiring first audio, and jumping to a second page.
Wherein the first page includes at least one candidate audio index and the second page is used to edit text.
The first audio may be accompaniment audio.
The first page may be a page for selecting accompaniment audio, and may include a plurality of candidate audio, where each candidate audio corresponds to a first control, or an index of each candidate audio corresponds to a first control.
The first control may be a confirm button.
In an alternative embodiment, if the user wants to record a song, the user may enter a first page, where the first page includes indexes of a plurality of candidate audios, and the user may select a desired audio as an accompaniment audio according to the indexes of the plurality of candidate audios, and click a first control corresponding to the accompaniment audio index, so as to retrieve the accompaniment audio from the database according to the accompaniment audio index.
Further, after the accompaniment audio is acquired, a jump may be made to the second page to edit the lyrics of the work in the second page.
In an alternative embodiment, the second page may be used to edit lyrics of the work, and after the accompaniment audio is obtained, the page for editing the lyrics may be entered to edit the lyrics.
And step S204, acquiring a first text in the second page, and responding to an operation instruction for the second control in the second page, and jumping to a third page.
Wherein the third page is used for providing a recording audio function.
The second page may include a text box having edited text in which the user may enter the first text.
The second control may be a text input completion button.
The third page may be used to record the voice of the user.
In an alternative embodiment, the lyrics edited by the user may be obtained in the second page, and after the user finishes editing the lyrics, the user may press a button for completing text input, that is, the second control mentioned above, to jump to the page for recording the user's voice.
In another alternative embodiment, the lower word count limit and the upper word count limit of the first word may be set, and the second control may be displayed when the word count exceeds the preset word count, so that the user may click on the second control to perform the next audio recording.
In another alternative embodiment, after the user edits the text, the user may also process the edited text through a word processing control in the third page, e.g., the text that is not rhymed may be labeled, so that the user improves the edited text. After clicking the word processing control, the user can also display the word group associated with the word being edited in the word processing control for the user to select, so as to give the user the inspiration of creating the text.
Step S206, obtaining second audio in the third page, and synthesizing the first audio, the first text and the second audio in response to the operation instruction of the third control in the third page to generate a first work.
The third page may have a recording button, and the user may press the recording button to record the sound.
The third control described above may be a control for composition.
In an alternative embodiment, after the user records the audio in the third page, the user may press a control for synthesizing in the third page to synthesize the accompaniment audio, the edited lyrics, and the recorded audio, and generate the first work.
In another alternative embodiment, the user may record the sound by wearing a headset, after the recording is completed, may click a listening test button in the third page, where during listening test, the user may click to play the accompaniment audio, and if the user feels that the recorded sound and the accompaniment audio are not aligned or the rhythm is not aligned, the user may manually splice and cut the recorded audio, and generate the first work according to the spliced and cut audio.
The work generation method described above may be applied to a generation scene of a rap work.
Through the steps, the first audio can be obtained in response to the operation instruction of the first control in the first page, and the first audio is jumped to the second page, wherein the first page comprises at least one candidate audio index, the second page is used for editing texts, the first text is obtained in the second page, the third page is jumped to the third page in response to the operation instruction of the second control in the second page, the third page is used for providing the audio recording function, the second audio is obtained in the third page, the first audio, the first text and the second audio are synthesized in response to the operation instruction of the third control in the third page, the first work is generated, the purpose of creating the work is achieved, the steps of the creation process can be simplified through executing different types of creation links in different pages, the flow of the creation process is achieved, the creation efficiency is improved, and the technical problem that the creation efficiency of the complex work is too low due to the creation process in the related technology is further improved.
Optionally, after the first work is generated based on the first audio, the first text and the second audio, the method further comprises jumping to a fourth page, acquiring adjustment parameters in the fourth page, and adjusting at least one sound effect of the first work based on the adjustment parameters to obtain an adjusted first work.
The fourth page is used for adjusting the sound effect of the generated first work.
The above-mentioned sound effects may be equalization, reverberation, electrical sound, gas sound, harmony, delay, etc.
The adjustment parameter may be obtained by dragging an adjustment frame corresponding to at least one sound effect in the fourth page by the user.
The fourth page may be a post-chamber, where the post-chamber is configured to optimize the first work to obtain a more optimal first work.
In an alternative embodiment, after jumping to the fourth page with sound effects adjusted, the user may adjust the sound effects in an adjusting frame corresponding to each sound effect in the fourth page to obtain an adjusting parameter, and after obtaining the adjusting parameter, may adjust at least one sound effect of the first work according to the adjusting parameter to obtain the adjusted first work.
Optionally, after jumping to the fourth page, the method further comprises jumping to a fifth page in response to an operation instruction of the fourth page for the fourth control, wherein the fifth page comprises at least one candidate video index, acquiring a first video in response to an operation instruction of the fifth page for the fifth control, jumping back to the fourth page, and combining the adjusted first work and the first video in response to an operation instruction of the fourth page for the sixth control to generate the first work.
The fourth control may be a control to add video.
The fifth page may include a plurality of video indexes, where each video index corresponds to a fourth control.
In an alternative embodiment, after jumping to the later room, the method may respond to an operation instruction for adding the video control in the fourth page, jump to the fifth page, and the user may select the first video according to at least one video index in the fifth page, specifically, the user may click on the fifth control corresponding to the video index, so as to obtain the first video corresponding to the video index.
The sixth control may be a composite control in the fourth page.
The first video can be used for baking the atmosphere of the work, so that the work can be better presented.
In another alternative embodiment, after the first video is acquired, the fourth page may be skipped, and the sixth control is clicked on the fourth page, so that the adjusted first work and the first video are synthesized, and the first work is generated, so that the first work is more complete.
In another alternative embodiment, after the first video is acquired, the first video may be further processed, for example, special effects may be performed on the first video. Specifically, the user may select a video special effect template to add to the first video.
Optionally, after jumping back to the fourth page, the method further comprises the steps of synthesizing the first video and the first text in the fourth page to obtain a second video, and responding to an operation instruction of a sixth control in the fourth page to synthesize the adjusted first work and the second video to generate the first work.
The first text may be edited lyrics.
In an alternative embodiment, in jumping back to the fourth page, the first video and the first text may be synthesized in the fourth page to obtain the second video, where the second video is a video containing lyrics, and since the second video contains lyrics of a recorded song, the user can watch the lyrics while watching the second video, so as to improve the ornamental value of the generated first work.
In another alternative embodiment, after synthesizing the first video and the first text to obtain the second video, the user may adjust the first text in the second video so that the display of the first text may be aligned with the sound of the recorded song.
In another alternative embodiment, the lyrics fonts, colors and the like in the second video can also be set, so that the second video is more attractive.
Optionally, obtaining the second audio in the third page, synthesizing the first audio, the first text and the second audio in response to an operation instruction of the third control in the third page, and generating the first work, wherein the obtaining the clipping parameter in the third page, clipping the second audio based on the clipping parameter to obtain the third audio, and synthesizing the first audio, the first text and the third audio in response to the operation instruction of the third control in the third page, and generating the first work.
In an alternative embodiment, after the second audio is acquired in the third page, the recorded audio may be clipped in the third page according to the clipping parameters, and the unwanted sound may be removed, or the sound to be retained may be selected, so as to obtain the third audio. The clipping parameters may be obtained by manually clipping the acquired second audio by the user. After the third audio is obtained, the user may click on a synthesis control in the third page, that is, the third control, to synthesize the first audio, the first text, and the third audio, and generate the first work.
In another alternative embodiment, the user may record the sound by wearing a headset, after the recording is completed, may click a listening test button in the third page, where during listening test, the user may click to play the accompaniment audio, and if the user feels that the recorded sound and the accompaniment audio are not aligned or the rhythm is not aligned, the user may manually splice and cut the recorded audio to obtain the above-mentioned clipping parameters, and clip the second audio according to the clipping parameters to obtain the third audio. And obtaining the first work based on the generated third audio, so that the obtained first work is more complete.
Optionally, obtaining the second audio in the third page, synthesizing the first audio, the first text and the second audio in response to an operation instruction of the third control in the third page, and generating the first work, wherein the obtaining the timestamp of the second audio in the third page, marking the first text by the timestamp to obtain the second text, and synthesizing the first audio, the second text and the third audio in response to an operation instruction of the third control in the third page, and generating the first work.
In an alternative embodiment, after the second audio is obtained in the third page, a timestamp corresponding to the second audio may be obtained, where the timestamp is used to indicate a playing time of each piece of sound in the second audio, and by way of example, when there are 10 pieces of sound in the second audio, a time when the first sentence of sound starts to be recorded may be determined as a time when playing starts.
In another alternative embodiment, the audio may be segmented according to sound pauses in the second audio.
Since the second audio corresponds to the first text, the second audio may be matched to the first text using a timestamp to determine lyrics corresponding to each piece of sound in the second audio.
In another alternative embodiment, after the timestamp of the second audio is obtained, the timestamp may be used to mark the first text to obtain the second text, where the second text corresponds to the playing time of the second audio. After the second text is obtained, the user may click on a synthesis control in the third page, that is, the third control described above, and then synthesize the first audio, the second text, and the third audio based on the synthesis control, to generate the first work.
An embodiment of a method of generating a work of the present disclosure is described in detail below in conjunction with FIG. 3. The method comprises the following steps:
step S301, selecting accompaniment audio;
step S302, editing lyrics;
step S303, recording audio;
step S304, judging whether the time for recording the audio is more than 3S, if so, executing step S306, otherwise, executing step S305;
Step S305, re-clicking recording;
Step S306, editing the recorded audio to obtain edited audio;
Step S307, judging whether the voice in the clipped audio needs to be adjusted, if so, executing step S308, and if not, executing step S309;
step S308, the sliding progress bar changes the audio time to be adjusted, and adjusts the audio;
step S309, listening the audio, and determining that the audio is finished;
optionally, after the audio is produced, the produced audio may be uploaded to the local.
Step S310, labeling lyrics to the manufactured audio;
Step S311, jumping to a later room, and selecting a video corresponding to the audio;
Step S312, modifying the video, lyrics and audio;
Optionally, a video special effect template can be selected to modify the video, the font and the color of the lyrics are selected to modify the lyrics, and the audio is modified through at least one sound effect.
The above-mentioned sound effects may be equalization, reverberation, electrical sound, gas sound, harmony, delay, etc.
Step S313, synthesizing the modified video, lyrics and audio to generate a work;
step S314, the generated work is adjusted to generate a complete work.
Further, a video file corresponding to the complete work may also be generated.
Optionally, the works can be adjusted by setting filters, subtitles, width and height, an audio track set, a rap audio track, an accompaniment audio track and an empty audio track can be set for the works, so that the effects of audio noise reduction and gain of the works are ensured, and the subtitles, the mapping and the like can be synthesized for the works.
Through the steps, the complex audio production can be simplified, the music lovers can conveniently and low-threshold complete own works, good gain effects are brought to the works, more music lovers are attracted to produce, and the utilization rate of music production software is improved.
According to the embodiment of the present disclosure, a device for generating a work is further provided, where the device may execute the method for generating a work in the foregoing embodiment, and a specific implementation manner and a preferred application scenario are the same as those of the foregoing embodiment, and are not described herein.
FIG. 4 is a schematic diagram of a production device for a work, as shown in FIG. 4, according to an embodiment of the disclosure, the device comprising:
An obtaining module 402, configured to obtain a first audio in response to an operation instruction for a first control in a first page, and skip to a second page, where the first page includes at least one candidate audio index, and the second page is used for editing text;
A skip module 404, configured to obtain a first text from the second page, and skip to a third page in response to an operation instruction for the second control in the second page, where the third page is used to provide a function of recording audio;
And the synthesizing module 406 is configured to obtain the second audio in the third page, and synthesize the first audio, the first text, and the second audio in response to an operation instruction for the third control in the third page, so as to generate the first work.
Optionally, the device further comprises an adjusting module, which is used for jumping to a fourth page, acquiring adjusting parameters in the fourth page, and adjusting at least one sound effect of the first work based on the adjusting parameters to obtain an adjusted first work.
Optionally, the skip module is further configured to skip to a fifth page in response to an operation instruction for the fourth control in the fourth page, where the fifth page includes at least one candidate video index, the obtaining module is further configured to obtain a first video in response to an operation instruction for the fifth control in the fifth page, skip back to the fourth page, and the synthesizing module is further configured to synthesize the adjusted first work and the first video in response to an operation instruction for the sixth control in the fourth page, so as to generate the first work.
Optionally, the device further comprises a synthesizing module for synthesizing the first video and the first text in the fourth page to obtain the second video, and a generating module for responding to the operation instruction of the sixth control in the fourth page to synthesize the adjusted first work and the second video to generate the first work.
Optionally, the synthesizing module comprises a clipping unit for obtaining clipping parameters in the third page, clipping the second audio based on the clipping parameters to obtain third audio, and a synthesizing unit for synthesizing the first audio, the first text and the third audio to generate the first work in response to an operation instruction of the third control in the third page.
Optionally, the synthesizing unit comprises a marking subunit, a synthesizing subunit and a synthesizing subunit, wherein the marking subunit is used for obtaining the timestamp of the second audio in the third page, marking the first text by using the timestamp to obtain the second text, and the synthesizing subunit is used for responding to the operation instruction of the third control in the third page to synthesize the first audio, the second text and the third audio to generate the first work.
It should be noted that each of the above modules may be implemented by software or hardware, and the latter may be implemented by, but not limited to, the above modules all being located in the same processor, or each of the above modules being located in different processors in any combination.
There is further provided, in accordance with an embodiment of the present disclosure, an electronic device including at least one processor, and a memory communicatively coupled to the at least one processor, wherein the memory stores instructions executable by the at least one processor to enable the at least one processor to perform a method of generating any of the works described above.
There is also provided, in accordance with an embodiment of the present disclosure, a non-transitory computer-readable storage medium storing computer instructions for causing a computer to perform a method of generating any one of the works described above.
Alternatively, in the present embodiment, the non-transitory computer readable storage medium described above may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
There is also provided, in accordance with an embodiment of the present disclosure, a computer program product comprising a computer program which, when executed by a processor, implements a method of generating any of the above works. Program code for carrying out the methods of generating a work of the present disclosure may be written in any combination of one or more programming languages. These program code may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus such that the program code, when executed by the processor or controller, causes the functions/operations specified in the flowchart and/or block diagram to be implemented. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package, partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the foregoing embodiments of the present disclosure, the descriptions of the various embodiments are emphasized, and for a portion of this disclosure that is not described in detail in this embodiment, reference is made to the related descriptions of other embodiments.
In the several embodiments provided in the present disclosure, it should be understood that the disclosed technology content may be implemented in other manners. The above-described embodiments of the apparatus are merely exemplary, and the division of the units, for example, may be a logic function division, and may be implemented in another manner, for example, a plurality of units or components may be combined or may be integrated into another system, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be through some interfaces, units or modules, or may be in electrical or other forms.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in each embodiment of the present disclosure may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated units, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present disclosure may be embodied in essence or a part contributing to the prior art or all or part of the technical solution in the form of a software product stored in a storage medium, including several instructions to cause a computer device (which may be a personal computer, a server or a network device, etc.) to perform all or part of the steps of the method described in the embodiments of the present disclosure. The storage medium includes a U disk, a read-only memory (ROM), a Random Access Memory (RAM), a removable hard disk, a magnetic disk, or an optical disk, etc. which can store program codes.
The foregoing is merely a preferred embodiment of the present disclosure, and it should be noted that modifications and adaptations to those skilled in the art may be made without departing from the principles of the present disclosure, which are intended to be comprehended within the scope of the present disclosure.

Claims (7)

1.一种作品的生成方法,包括:1. A method for generating a work, comprising: 响应于第一页面中对于第一控件的操作指令,获取第一音频,并跳转至第二页面,其中,所述第一页面包括至少一个候选的音频索引,所述第二页面用于编辑文本;In response to an operation instruction for a first control on a first page, acquiring a first audio, and jumping to a second page, wherein the first page includes at least one candidate audio index, and the second page is used for editing text; 在所述第二页面中获取第一文本,响应于所述第二页面中对于第二控件的操作指令,跳转至第三页面,其中,所述第三页面用于提供录制音频功能;Acquire the first text in the second page, and jump to the third page in response to an operation instruction for the second control in the second page, wherein the third page is used to provide an audio recording function; 在所述第三页面中获取第二音频,响应于所述第三页面中对于第三控件的操作指令,对所述第一音频、所述第一文本以及所述第二音频进行合成,生成第一作品;Acquire a second audio in the third page, and synthesize the first audio, the first text, and the second audio in response to an operation instruction for a third control in the third page to generate a first work; 其中,所述方法还包括:Wherein, the method further comprises: 跳转至第四页面,在所述第四页面中获取调整参数,基于所述调整参数对所述第一作品的至少一个音效进行调整,得到调整后的第一作品;Jump to the fourth page, obtain adjustment parameters on the fourth page, and adjust at least one sound effect of the first work based on the adjustment parameters to obtain an adjusted first work; 其中,在跳转至第四页面之后,所述方法还包括:After jumping to the fourth page, the method further includes: 响应于所述第四页面中对于第四控件的操作指令,跳转至第五页面,其中,所述第五页面中包括至少一个候选的视频索引;In response to an operation instruction for a fourth control on the fourth page, jump to a fifth page, wherein the fifth page includes at least one candidate video index; 响应于所述第五页面中对于第五控件的操作指令,获取第一视频,跳转回所述第四页面;In response to an operation instruction for a fifth control on the fifth page, acquiring the first video and jumping back to the fourth page; 在跳转回所述第四页面之后,响应于所述第四页面中对于第六控件的操作指令,对所述调整后的第一作品和所述第一视频进行合成,生成所述第一作品;After jumping back to the fourth page, in response to an operation instruction for a sixth control on the fourth page, synthesizing the adjusted first work and the first video to generate the first work; 或,在跳转回所述第四页面之后,在所述第四页面中对所述第一视频和所述第一文本进行合成,得到第二视频;响应于所述第四页面中对于所述第六控件的操作指令,对所述调整后的第一作品和所述第二视频进行合成,生成所述第一作品。Or, after jumping back to the fourth page, the first video and the first text are synthesized in the fourth page to obtain the second video; in response to the operation instruction for the sixth control in the fourth page, the adjusted first work and the second video are synthesized to generate the first work. 2.根据权利要求1所述的方法,其特征在于,在所述第三页面中获取第二音频,响应于所述第三页面中对于第三控件的操作指令,对所述第一音频、所述第一文本以及所述第二音频进行合成,生成第一作品,包括:2. The method according to claim 1, characterized in that obtaining the second audio in the third page, and synthesizing the first audio, the first text and the second audio in response to an operation instruction for a third control in the third page to generate a first work, comprises: 在所述第三页面中获取剪辑参数,基于所述剪辑参数对所述第二音频进行剪辑,得到第三音频;Acquire clipping parameters in the third page, and clip the second audio based on the clipping parameters to obtain a third audio; 响应于所述第三页面中对于所述第三控件的操作指令,对所述第一音频、所述第一文本以及所述第三音频进行合成,生成所述第一作品。In response to an operation instruction for the third control in the third page, the first audio, the first text and the third audio are synthesized to generate the first work. 3.根据权利要求2所述的方法,其特征在于,在所述第三页面中获取第二音频,响应于所述第三页面中对于第三控件的操作指令,对所述第一音频、所述第一文本以及所述第二音频进行合成,生成第一作品,包括:3. The method according to claim 2, characterized in that obtaining the second audio in the third page, and synthesizing the first audio, the first text and the second audio in response to an operation instruction for a third control in the third page to generate a first work, comprises: 在所述第三页面中获取所述第二音频的时间戳,利用所述时间戳对所述第一文本进行标注,得到第二文本;Obtaining a timestamp of the second audio in the third page, and annotating the first text by using the timestamp to obtain a second text; 响应于所述第三页面中对于所述第三控件的操作指令,对所述第一音频、所述第二文本以及所述第三音频进行合成,生成所述第一作品。In response to an operation instruction for the third control in the third page, the first audio, the second text and the third audio are synthesized to generate the first work. 4.一种作品的生成装置,包括:4. A device for generating a work, comprising: 获取模块,用于响应于第一页面中对于第一控件的操作指令,获取第一音频,并跳转至第二页面,其中,所述第一页面包括至少一个候选的音频索引,所述第二页面用于编辑文本;An acquisition module, configured to acquire the first audio in response to an operation instruction for the first control in the first page, and jump to the second page, wherein the first page includes at least one candidate audio index, and the second page is used for editing text; 跳转模块,用于在所述第二页面中获取第一文本,响应于所述第二页面中对于第二控件的操作指令,跳转至第三页面,其中,所述第三页面用于提供录制音频功能;A jump module, used for acquiring the first text in the second page, and jumping to a third page in response to an operation instruction for the second control in the second page, wherein the third page is used for providing an audio recording function; 合成模块,用于在所述第三页面中获取第二音频,响应于所述第三页面中对于第三控件的操作指令,对所述第一音频、所述第一文本以及所述第二音频进行合成,生成第一作品;a synthesis module, configured to obtain the second audio in the third page, and synthesize the first audio, the first text and the second audio in response to an operation instruction for a third control in the third page to generate a first work; 其中,所述装置还用于跳转至第四页面,在所述第四页面中获取调整参数,基于所述调整参数对所述第一作品的至少一个音效进行调整,得到调整后的第一作品;The device is further configured to jump to a fourth page, obtain adjustment parameters on the fourth page, and adjust at least one sound effect of the first work based on the adjustment parameters to obtain an adjusted first work; 其中,在跳转至第四页面之后,所述装置还用于响应于所述第四页面中对于第四控件的操作指令,跳转至第五页面,其中,所述第五页面中包括至少一个候选的视频索引;响应于所述第五页面中对于第五控件的操作指令,获取第一视频,跳转回所述第四页面;在跳转回所述第四页面后,响应于所述第四页面中对于第六控件的操作指令,对所述调整后的第一作品和所述第一视频进行合成,生成所述第一作品;After jumping to the fourth page, the device is further used to respond to an operation instruction for a fourth control in the fourth page to jump to a fifth page, wherein the fifth page includes at least one candidate video index; in response to an operation instruction for a fifth control in the fifth page, obtain a first video and jump back to the fourth page; after jumping back to the fourth page, in response to an operation instruction for a sixth control in the fourth page, synthesize the adjusted first work and the first video to generate the first work; 或,在跳转回所述第四页面之后,在所述第四页面中对所述第一视频和所述第一文本进行合成,得到第二视频;响应于所述第四页面中对于所述第六控件的操作指令,对所述调整后的第一作品和所述第二视频进行合成,生成所述第一作品。Or, after jumping back to the fourth page, the first video and the first text are synthesized in the fourth page to obtain the second video; in response to the operation instruction for the sixth control in the fourth page, the adjusted first work and the second video are synthesized to generate the first work. 5.一种电子设备,包括:5. An electronic device, comprising: 至少一个处理器;以及at least one processor; and 与所述至少一个处理器通信连接的存储器;a memory communicatively coupled to the at least one processor; 其中,所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行权利要求1-3中任一项所述的方法。The memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor so that the at least one processor can execute the method according to any one of claims 1 to 3. 6.一种存储有计算机指令的非瞬时计算机可读存储介质,其中,所述计算机指令用于使所述计算机执行根据权利要求1-3中任一项所述的方法。6. A non-transitory computer-readable storage medium storing computer instructions, wherein the computer instructions are used to cause the computer to execute the method according to any one of claims 1 to 3. 7.一种计算机程序产品,包括计算机程序,所述计算机程序在被处理器执行时实现根据权利要求1-3中任一项所述的方法。7. A computer program product, comprising a computer program, which, when executed by a processor, implements the method according to any one of claims 1 to 3.
CN202111161807.XA 2021-09-30 2021-09-30 Method, device, electronic device and storage medium for generating works Active CN113963674B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111161807.XA CN113963674B (en) 2021-09-30 2021-09-30 Method, device, electronic device and storage medium for generating works

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111161807.XA CN113963674B (en) 2021-09-30 2021-09-30 Method, device, electronic device and storage medium for generating works

Publications (2)

Publication Number Publication Date
CN113963674A CN113963674A (en) 2022-01-21
CN113963674B true CN113963674B (en) 2025-07-04

Family

ID=79463351

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111161807.XA Active CN113963674B (en) 2021-09-30 2021-09-30 Method, device, electronic device and storage medium for generating works

Country Status (1)

Country Link
CN (1) CN113963674B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117059066A (en) * 2022-05-07 2023-11-14 北京字跳网络技术有限公司 Audio processing method, device, equipment and storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110189741A (en) * 2018-07-05 2019-08-30 腾讯数码(天津)有限公司 Audio synthesis method, apparatus, storage medium and computer equipment

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108024145B (en) * 2017-12-07 2020-12-11 北京百度网讯科技有限公司 Video recommendation method, apparatus, computer equipment and storage medium
CN108668164A (en) * 2018-07-12 2018-10-16 北京微播视界科技有限公司 Select method, apparatus, terminal device and the medium of background music shooting video
US20200258495A1 (en) * 2019-02-08 2020-08-13 Brett Duncan Arquette Digital audio methed for creating and sharingaudiobooks using a combination of virtual voices and recorded voices, customization based on characters, serilized content, voice emotions, and audio assembler module
CN111899706B (en) * 2020-07-30 2024-08-23 广州酷狗计算机科技有限公司 Audio production method, device, equipment and storage medium
CN112528608B (en) * 2020-12-21 2024-04-19 北京百度网讯科技有限公司 Page editing method, device, electronic device and storage medium

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110189741A (en) * 2018-07-05 2019-08-30 腾讯数码(天津)有限公司 Audio synthesis method, apparatus, storage medium and computer equipment

Also Published As

Publication number Publication date
CN113963674A (en) 2022-01-21

Similar Documents

Publication Publication Date Title
US10088976B2 (en) Systems and methods for multiple voice document narration
US8359202B2 (en) Character models for document narration
US8346557B2 (en) Systems and methods document narration
CN106716466B (en) Conference information storage device and method
US20080275700A1 (en) Method of and System for Modifying Messages
JP2021530726A (en) Methods and systems for creating object-based audio content
CN114023301A (en) Audio editing method, electronic device and storage medium
JP6641045B1 (en) Content generation system and content generation method
CN112995736A (en) Speech subtitle synthesis method, apparatus, computer device, and storage medium
CN113590247A (en) Text creation method and computer program product
TW202016922A (en) Audio playback device and audio playback method thereof
CN113676772A (en) Video generation method and device
CN113963674B (en) Method, device, electronic device and storage medium for generating works
CN112200897A (en) Intelligent animation generation method and device based on intention and equipment
JP2008217447A (en) Content generation apparatus and content generation program
KR20180078197A (en) E-voice book editor and player
JP5082971B2 (en) A speech synthesizer and a reading system using the same.
CN114550690B (en) Song synthesis method and device
CN108806732A (en) A kind of background music processing method and electronic equipment based on artificial intelligence
JP4563418B2 (en) Audio processing apparatus, audio processing method, and program
CN114242036A (en) Role dubbing method and device, storage medium and electronic equipment
JP2018072698A (en) Signal processing method and signal processing apparatus
WO2022003798A1 (en) Server, composite content data creation system, composite content data creation method, and program
CN119031176A (en) A method and device for editing narration data, electronic device, and storage medium
CN116229921A (en) Method, device, device and storage medium for making adaptive music

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant