WO2020173391A1 - 一种歌曲录制方法、修音方法及电子设备 - Google Patents

一种歌曲录制方法、修音方法及电子设备 Download PDF

Info

Publication number
WO2020173391A1
WO2020173391A1 PCT/CN2020/076099 CN2020076099W WO2020173391A1 WO 2020173391 A1 WO2020173391 A1 WO 2020173391A1 CN 2020076099 W CN2020076099 W CN 2020076099W WO 2020173391 A1 WO2020173391 A1 WO 2020173391A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
electronic device
song
original
interface
Prior art date
Application number
PCT/CN2020/076099
Other languages
English (en)
French (fr)
Inventor
陶凯
尹明婕
李延奇
徐德著
戴恒杰
缪海波
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to JP2021550197A priority Critical patent/JP7222112B2/ja
Priority to US17/434,555 priority patent/US20220130360A1/en
Priority to EP20762220.0A priority patent/EP3905246A4/en
Publication of WO2020173391A1 publication Critical patent/WO2020173391A1/zh

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • G10H1/366Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems with means for modifying or correcting the external signal, e.g. pitch correction, reverberation, changing a singer's voice
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • G10H1/0041Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • G10H1/368Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems displaying animated or moving pictures synchronized with the music or audio part
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10527Audio or video recording; Data buffering arrangements
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/005Musical accompaniment, i.e. complete instrumental rhythm synthesis added to a performed melody, e.g. as output by drum machines
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/091Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for performance evaluation, i.e. judging, grading or scoring the musical qualities or faithfulness of a performance, e.g. with respect to pitch, tempo or other timings of a reference performance
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/095Inter-note articulation aspects, e.g. legato or staccato
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/155Musical effects
    • G10H2210/195Modulation effects, i.e. smooth non-discontinuous variations over a time interval, e.g. within a note, melody or musical transition, of any sound parameter, e.g. amplitude, pitch, spectral response, playback speed
    • G10H2210/201Vibrato, i.e. rapid, repetitive and smooth variation of amplitude, pitch or timbre within a note or chord
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/005Non-interactive screen display of musical or status data
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/005Non-interactive screen display of musical or status data
    • G10H2220/011Lyrics displays, e.g. for karaoke applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/005Non-interactive screen display of musical or status data
    • G10H2220/015Musical staff, tablature or score displays, e.g. for score reading during a performance.
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/091Graphical user interface [GUI] specifically adapted for electrophonic musical instruments, e.g. interactive musical displays, musical instrument icons or menus; Details of user interactions therewith
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/171Transmission of musical instrument data, control or status information; Transmission, remote access or control of music data for electrophonic musical instruments
    • G10H2240/201Physical layer or hardware aspects of transmission to or from an electrophonic musical instrument, e.g. voltage levels, bit streams, code words or symbols over a physical link connecting network nodes or instruments
    • G10H2240/241Telephone transmission, i.e. using twisted pair telephone lines or any type of telephone network
    • G10H2240/251Mobile telephone transmission, i.e. transmitting, accessing or controlling music data wirelessly via a wireless or mobile telephone receiver, analog or digital, e.g. DECT GSM, UMTS
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10527Audio or video recording; Data buffering arrangements
    • G11B2020/10537Audio or video recording
    • G11B2020/10546Audio or video recording specifically adapted for audio data

Definitions

  • This application relates to the field of audio technology, and in particular to a song recording method, a sound repair method, and an electronic device.
  • K song applications For users who like to sing, K song applications have become popular applications installed on mobile phones, tablets and other electronic devices. Exemplarily, the user can select the song that needs to be recorded after opening the K song application, and then, the K song application can load the accompaniment of the song, and call the microphone and other pickup components on the mobile phone to record the audio of the user singing the song, thereby The user's own singing voice is integrated into the accompaniment provided by K song applications.
  • some karaoke applications provide the function of repairing the songs recorded by users.
  • a K song application can compare the song A recorded by the user with the original singing audio of the song A, and modify the pitch and rhythm of the singing voice recorded by the user to the pitch and rhythm of the original singing audio, so as to correct when the user records the song. Problems with intonation and rhythm appear.
  • This application provides a song recording method, a sound repair method, and an electronic device, which can guide the user to use corresponding singing skills when recording a song, thereby improving the recording effect and experience of the user when using K song applications.
  • this application provides a song recording method, including: an electronic device displays a first interface, and the first interface includes a recording button for recording the first song; if it is detected that the user selects the recording button, the electronic device can obtain The accompaniment of the first song and the characteristic information of the original vocals, the characteristic information of the original vocals can be used to reflect the singing characteristics of the original singer when singing the first song; further, the electronic device can start to play the accompaniment of the first song and Start to record the user's thousand voices; during the process of recording the user's thousand voices, the electronic device can display guidance information in the second interface according to the characteristic information of the original voices, and the guidance information is used to guide the user to ventilate while singing Or one or more of the vibrato.
  • the electronic device can display guidance information according to the singing characteristics of the original singer, and guide the user to one or more singing skills such as ventilation or vibrato during singing, thereby improving the recording effect of the user's recorded song. Use experience.
  • the feature information of the original vocals includes the ventilation position of the original vocals; at this time, the above-mentioned guidance information may include ventilation prompts; then, the electronic device according to the characteristics of the original vocals The information displays the guidance information in the second interface, including: the electronic device can display the ventilation prompt corresponding to the ventilation position of the original singer in the second interface according to the ventilation position of the original singer, thereby helping the user to control The breath changes during singing, thereby reducing the more obvious and disordered breathing sound in the recorded song, so as to improve the listening quality of the recorded song.
  • the above-mentioned second interface may include the lyrics of the first song; at this time, the electronic device displays the ventilation position corresponding to the ventilation position of the original song in the second interface according to the ventilation position of the original song.
  • Gas cues including: electronic equipment in the song
  • the ventilation prompt is displayed in the first position of the word, and the first position is the same as the ventilation position of the original singer.
  • the above-mentioned second interface may include a singer, which is used to display the notes of each word in the lyrics being recorded; at this time, the electronic device displays the original sound in the second interface according to the ventilation position of the original song.
  • the ventilation prompt corresponding to the ventilation position of the singing voice includes: the electronic device displays the ventilation prompt in the second position of the singer, and the second position is the same as the ventilation position of the original singing voice.
  • the feature information of the original singing voice includes the tremolo feature of the original singing voice; at this time, the corresponding guidance information includes the tremolo prompt; then, the electronic device is based on the characteristic information of the original singing voice in the first
  • the guidance information displayed in the second interface includes: the electronic device displays the vibrato prompt corresponding to the vibrato feature of the original singing thousand voices in the second interface.
  • the above-mentioned tremolo feature includes the position of the tremolo; the above-mentioned second interface includes the lyrics of the first song; at this time, the electronic device displays the tremolo feature in the second interface according to the tremolo feature of the original song.
  • the vibrato prompt corresponding to the vibrato feature includes: the electronic device displays the vibrato prompt in the third position of the lyrics, and the third position is the same as the vibrato position in the original song.
  • the vibrato feature includes the vibrato position; the second interface includes a singer, which is used to display the note of each word in the lyrics being recorded; at this time, the electronic device is based on the vibrato feature of the original singing thousand voices.
  • the vibrato prompt corresponding to the vibrato feature of the original singing voice is displayed in the second interface, including: the electronic device displays the vibrato prompt in the fourth position of the singer, and the fourth position is the same as the vibrato position in the original singing voice.
  • the feature information of the original singing voice may also include the fundamental frequency information of the original singing voice; the electronic device may generate a pitch prompt based on the fundamental frequency information of the original singing voice and display it on the second interface to guide the user when singing Intonation.
  • the above method further includes: the electronic device extracts the user's thousand voices' characteristic information from the user's thousand voices, and the user's thousand voices' characteristic information is used to reflect that the user is singing The singing characteristics of the first song; the electronic device compares the characteristic information of the user's thousand voices with the characteristic information of the original song, and displays the evaluation result in the second interface.
  • the electronic device can perform real-time evaluation of one or more singing skills such as the pitch, vibrato, and ventilation of the user based on the characteristic information of the original singing thousand voices, compared to purely based on pitch
  • the embodiment of the application also considers the user's ventilation, vibrato and other factors when evaluating the user's singing, so the obtained evaluation result is more accurate, rich and objective.
  • the electronic device can feed back the evaluation result of the user's singing in the second interface in real time to the user, so that the user can understand his current singing effect in time, and the user can timely adjust the singing method in the subsequent recording process.
  • the characteristic information of the user’s thousand voices includes the ventilation position of the user’s thousand voices
  • the characteristic information of the original voice of the thousand voices includes the ventilation position of the original voice
  • the characteristic information in the voice and the characteristic information of the original singer Qiansheng are displayed on the second interface, including:
  • the electronic device compares the ventilation position of the user Qiansheng with the ventilation position of the original singer Qiansheng in the second interface
  • the evaluation results of ventilation skills are displayed in.
  • the electronic device displays the evaluation result of the ventilation skills in the second interface by comparing the ventilation position of the user's Qiansheng with the ventilation position of the original Qiansheng, including: If the ventilating position of sing a thousand sounds is the same, the electronic device will display a prompt of accurate ventilation in the second interface; if the venting position of the user’s thousand sounds is different from that of the original singing, the electronic device is in the second interface. The prompt of ventilation error is displayed on the interface.
  • the characteristic information of the user’s thousand voices includes the vibrato characteristics of the user’s thousand voices
  • the characteristic information of the original voices of the user’s thousand voices includes the vibrato characteristics of the original voices
  • the characteristic information of the singing voice, displaying the evaluation result in the second interface includes: the electronic device displays the evaluation result of the vibrato technique in the second interface by comparing the vibrato characteristics of the user thousand voice with the vibrato characteristics of the original singing voice.
  • the electronic device displays the evaluation results of the vibrato skills in the second interface by comparing the vibrato characteristics of the user's thousand voices with the vibrato characteristics of the original singing thousand voices, including: If the vibrato feature matches, the electronic device displays a prompt that the vibrato is accurate in the second interface; if the vibrato feature of the user's thousand voices does not match the vibrato feature of the original song, the electronic device displays a prompt that the vibrato is wrong in the second interface.
  • the electronic device displaying the evaluation result in the second interface includes: the electronic device can display the evaluation result in the lyrics on the second interface; or, the electronic device can also display the evaluation result on the second interface.
  • the evaluation result is displayed in.
  • the method further includes: after the recording of the first song ends, the electronic device according to the characteristic information of the original song Fixed the singing problem in the user's Thousand Voices.
  • the singing problem includes one or more of the ventilation problem or the vibrato problem.
  • this application provides a sound repair method, including: an electronic device displays a first interface, and the first interface includes a recording button for recording the first song; if it is detected that the user selects the recording button, the electronic device can obtain The accompaniment of the first song and the characteristic information of the original singing thousand voices, the characteristic information of the original singing thousand voices is used to reflect the singing characteristics of the original singing voice when singing the first song; further, the electronic device can start playing the accompaniment of the first song and start Record the user's thousand voices sung by the user; when the first song is recorded, the electronic device can correct the singing problems in the recorded user thousand voices according to the characteristic information of the original song.
  • the singing problems include ventilation problems or vibrato problems One or more of.
  • electronic equipment can repair singing skills such as ventilation and vibrato in songs recorded by users.
  • ordinary users who have not undergone professional vocal training can also record a singing voice with a better listening quality, thereby improving the recording effect and experience of the user's recorded songs.
  • the above-mentioned singing problem is a ventilation problem; wherein the electronic device corrects the singing problem in the user's thousand voices according to the characteristic information of the original singing voice, including: the electronic device extracts the user from the user thousand voices The ventilating sound in a thousand voices; electronic equipment reduces the energy and frequency response of the user’s ventilating voice in a thousand voices, thereby weakening the ventilation problem in the user’s thousand voices, and improving the listening quality of recorded songs ⁇
  • the above-mentioned singing problem is a vibrato problem
  • the electronic device corrects the singing problem in the user's voice according to the characteristic information of the original song, including: the electronic device extracts the user's voice from the user's voice
  • the electronic device can use the vibrato feature in the original singing voice to correct the vibrato feature in the user thousand voice to correct the vibrato problem in the user thousand voice.
  • the electronic device can synthesize a new vibrato from the user's thousands of sounds by using the vibrato amplitude, the vibration frequency, and the stable trend items corresponding to the vibrato in the original singing voices, thereby improving the vibrato effect of the user's thousands of voices.
  • the above-mentioned singing problem can also be a pitch problem.
  • the electronic device can lower the pitch of the user’s thousand voices where the pitch is too high, and raise the pitch of the user’s thousand voices where the pitch is too low, so that the user’s thousand voices’ pitch It is the same or close to the tone of the original singing thousand voices.
  • the above method further includes: the electronic device displays a second interface, and the second interface includes a tone repair switch; wherein the electronic device is based on the characteristic information of the original song
  • the correction of the singing problem in the user's thousand voices includes: if it is detected that the user has turned on the above-mentioned tuning switch, the electronic device can correct the singing problem in the user's thousand voices according to the characteristic information of the original voice.
  • the above method further includes: the electronic device displays a second interface, the second interface includes a tone modification option, and the tone modification option includes an option to correct the pitch and a correction Options for ventilating sounds and options for correcting vibrato; among them, the electronic device corrects the singing problems in the user’s voice according to the characteristic information of the original song, including: in response to the user’s selected tuning option, the electronic device according to the original song The feature information of the voice corrects the singing problem corresponding to the tuning option in the user's thousand voices.
  • the above method further includes: the electronic device displays a second interface, and the second interface includes a single sentence refinement option; in response to the user selecting the single sentence refinement option, The electronic device can display a third interface, and the third interface includes N (N> 1) sentences in the first song; at this time, the electronic device corrects the singing problem in the user's song based on the characteristic information of the original song, Specifically: if the user selects the first lyrics among the above N sentences, the electronic device can correct the singing problem in the user's thousand voices corresponding to the first lyrics according to the characteristic information of the first song in the original song. In this way, the user can correct the singing problem in the user's thousand voices recorded this time sentence by sentence, thereby correcting the song recorded by the user to a more satisfactory singing effect for the user.
  • the electronic device when the electronic device displays the above third interface, it can also mark the location where the singing problem occurs in the N sentences of lyrics on the third interface. In this way, the user can intuitively see the specific problems that arise when he sings each lyrics.
  • this application provides an electronic device, including: a touch screen, one or more processors, one or more memories, one or more microphones, and one or more computer programs; where the processor and the touch screen, and the memory And the microphone are both coupled, the above-mentioned one or more computer programs are stored in the memory, and when the electronic device is running, the processor executes the one or more computer programs stored in the memory, so that the electronic device executes any of the above-mentioned songs Recording method or repairing method.
  • the present application provides a computer storage medium, including computer instructions, which when the computer instructions run on an electronic device, cause the electronic device to execute the song recording method or the sound repair method as described in any one of the first aspect.
  • this application provides a computer program product, which when the computer program product runs on an electronic device, causes the electronic device to execute the song recording method or the sound repair method as described in any one of the first aspect.
  • the present application provides a graphical user interface (GUI), the graphical user interface is stored in an electronic device, the electronic device includes a touch screen, a memory, and a processor, and the processor is configured to execute One or more computer programs in the computer program, wherein the graphical user interface includes: a first GUI displayed on the touch screen, the first GUI includes a recording button for recording a first song; in response to the recording button A second GUI is displayed on the touch screen.
  • the second GUI is a recording interface for recording the first song.
  • the second GUI includes guidance information for guiding the user to sing the first song.
  • the guidance information may include ventilation prompts and / Or vibrato prompt.
  • the above-mentioned second GUI may also include evaluation results of the user's ventilation skills and/or vibrato skills during singing.
  • the electronic device after displaying the above-mentioned second GUI on the touch screen, it further includes: after the recording of the first song ends, displaying a third GUI on the touch screen, the third GUI is the sound modification interface of the first song, and the third GUI includes Sound repair switch: When the sound repair switch is turned on by the user, the electronic device can correct the ventilating and vibrato problems in the user's voice.
  • the electronic equipment described in the third aspect, the computer storage medium described in the fourth aspect, the computer program product described in the fifth aspect, and the GUI described in the sixth aspect provided above are all used to execute the foregoing
  • the beneficial effects that can be achieved can refer to the beneficial effects in the corresponding method provided above, which will not be repeated here.
  • FIG. 1 is a first structural diagram of an electronic device according to an embodiment of the application
  • 2 is a schematic diagram of the architecture of an operating system in an electronic device provided by an embodiment of the application
  • FIG. 3 is a schematic flowchart of a song recording method and a sound repair method provided by an embodiment of the application
  • FIG. 4 is a schematic diagram 1 of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 5 is a second schematic diagram of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 6 is a third schematic diagram of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 7 is a fourth schematic diagram of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 8 is a schematic diagram five of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 9 is a sixth schematic diagram of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 10 is a seventh schematic diagram of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 11 is a schematic diagram eight of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 12 is a schematic diagram 9 of a scene of a song recording method provided by an embodiment of the application.
  • FIG. 13 is a schematic diagram 1 of a scene of a sound modification method provided by an embodiment of this application.
  • FIG. 14 is a schematic diagram 2 of a scene of a sound modification method provided by an embodiment of this application.
  • FIG. 15 is a schematic diagram three of a scene of a sound modification method provided by an embodiment of the application.
  • FIG. 16 is a fourth schematic diagram of a scene of a sound modification method provided by an embodiment of this application.
  • FIG. 17 is a second structural diagram of an electronic device provided by an embodiment of this application.
  • the song recording method and sound repair method provided in the embodiments of the present application can be applied to mobile phones, tablet computers, notebook computers, ultra-mobile personal computers (UMPC), handheld computers, netbooks, and personal computers.
  • Electronic devices such as personal digital assistants (PDAs), wearable electronic devices, virtual reality devices, etc., are not limited in this embodiment of the present application.
  • FIG. 1 shows a schematic structural diagram of an electronic device 100.
  • the electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, and an antenna 2.
  • Mobile communication module 150 wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, earphone jack 170D, sensor module 180, button 190, motor 191, indicator 192, camera 193, display 194, and Subscriber identification module (subscriber identification module, SIM) card interface 195, etc.
  • SIM Subscriber identification module
  • the sensor module 180 may include pressure sensor 180A, gyroscope sensor 180B, air pressure sensor 180C, magnetic sensor 180D, acceleration sensor 180E, distance sensor 180F, proximity light sensor 180G, fingerprint sensor 180H, temperature sensor 180J, touch sensor 180K, ambient light Sensor 180L, bone conduction sensor 180M, etc.
  • the structure illustrated in the embodiment of the present invention does not constitute a specific limitation on the electronic device 100.
  • the electronic device 100 may include more or fewer components than shown, or combine certain components, or disassemble certain components, or arrange different components.
  • the components shown in the figure can be implemented in hardware, software or a combination of software and hardware.
  • the processor 110 may include one or more processing units.
  • the processor 110 may include an application processor (AP), a modem processor, a graphics processing unit (GPU), and an image signal processor. (image signal processor, ISP), controller, video codec, digital signal processor (DSP), baseband processor, and/or neural-network processing unit (NPU), etc.
  • AP application processor
  • GPU graphics processing unit
  • ISP image signal processor
  • DSP digital signal processor
  • NPU neural-network processing unit
  • different processing units may be independent devices, or may be integrated in one or more processors.
  • the controller can generate operation control signals according to the instruction operation code and timing signals to complete the control of fetching and executing instructions.
  • a memory may also be provided in the processor 110 to store instructions and data.
  • the memory in the processor 110 is a cache memory.
  • the memory can store instructions or data that the processor 110 has just used or used cyclically. If the processor 110 needs to use the instruction or data again, it can be directly called from the memory. Repeated accesses are avoided, the waiting time of the processor 110 is reduced, and the efficiency of the system is improved.
  • the processor 110 may include one or more interfaces.
  • the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit sound (inter-integrated circuit sound, I2S) interface, a pulse code modulation (PCM) interface, and a universal asynchronous transmitter (universal asynchronous transmitter) interface.
  • I2C integrated circuit
  • I2S integrated circuit sound
  • PCM pulse code modulation
  • UART universal asynchronous transmitter
  • MIPI mobile industry processor interface
  • GPIO general-purpose input/output
  • SIM subscriber identity module
  • USB universal serial bus
  • the I2C interface is a two-way synchronous serial bus that includes a serial data line (SDA) and a derail clock line (SCL).
  • the processor 110 may include multiple sets of I2C buses.
  • the processor 110 may be coupled to the touch sensor 180K, charger, flash, camera 193, etc., through different I2C bus interfaces.
  • the processor 110 may couple the touch sensor 180K through the I2C interface, so that the processor 110 and the touch sensor 180K communicate through the I2C bus interface to realize the touch function of the electronic device 100.
  • the I2S interface can be used for audio communication.
  • the processor 110 may include multiple sets of I2S buses.
  • the processor 110 may be coupled with the audio module 170 through an I2S bus to implement communication between the processor 110 and the audio module 170.
  • the audio module 170 may transmit audio signals to the wireless communication module 160 through the I2S interface, so as to realize the function of answering calls through the Bluetooth headset.
  • the PCM interface can also be used for audio communication to sample, quantize and encode analog signals.
  • the audio module 170 and the wireless communication module 160 may be coupled through a PCM bus interface.
  • the audio module 170 may also transmit audio signals to the wireless communication module 160 through the PCM interface, so as to realize the function of answering calls through the Bluetooth headset. Both the I2S interface and the PCM interface can be used for audio communication.
  • the UART interface is a universal serial data bus used for asynchronous communication.
  • the bus can be a two-way communication bus. It converts the data to be transmitted between serial communication and parallel communication.
  • the UART interface is generally used to connect the processor 110 and the wireless communication module 160.
  • the processor 110 communicates with the Bluetooth module in the wireless communication module 160 through the UART interface to implement the Bluetooth function.
  • the audio module 170 may transmit audio signals to the wireless communication module 160 through the UART interface, so as to realize the function of playing music through the Bluetooth headset.
  • the MIPI interface can be used to connect the processor 110 with the display screen 194, the camera 193 and other peripheral devices.
  • MIPI interface includes camera serial interface (CSI), display serial interface (DSI) and so on.
  • the processor 110 and the camera 193 communicate through a CSI interface to implement the shooting function of the electronic device 100.
  • the processor 110 and the display screen 194 communicate through the DSI interface to realize the display function of the electronic device 100.
  • the GPI0 interface can be configured through software.
  • the GPI0 interface can be configured as a control signal or as a data signal.
  • the GPI0 interface may be used to connect the processor 110 with the camera 193, the display screen 194, the wireless communication module 160, the audio module 170, the sensor module 180, and so on.
  • GPI0 interface can also be configured as I2C interface, I2S interface, UART interface, MIPI interface and so on.
  • the interface connection relationship between the modules illustrated in the embodiment of the present invention is merely illustrative, and does not constitute a structural limitation of the electronic device 100.
  • the electronic device 100 may also use the above Different interface connection modes in the embodiments, or a combination of multiple interface connection modes.
  • the charging management module 140 is used to receive charging input from the charger.
  • the charger can be a wireless charger or a wired charger.
  • the charging management module 140 may receive the charging input of the wired charger through the USB interface 130.
  • the charging management module 140 may receive the wireless charging input through the wireless charging coil of the electronic device 100. While the charging management module 140 charges the battery 142, the power management module 141 can also supply power to electronic devices.
  • the power management module 141 is used to connect the battery 142, the charging management module 140 and the processor 110.
  • the power management module 141 receives input from the battery 142 and/or the charging management module 140, and supplies power to the processor 110, the internal memory 121, the display screen 194, the camera 193, and the wireless communication module 160.
  • the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle times, and battery health status (leakage, impedance).
  • the power management module 141 may also be provided in the processor 110.
  • the power management module 141 and the charging management module 140 may also be provided in the same device.
  • the wireless communication function of the electronic device 100 can be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modem processor, and the baseband processor.
  • Antenna 1 and Antenna 2 are used to transmit and receive electromagnetic wave signals.
  • Each antenna in the electronic device 100 can be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
  • Antenna 1 can be multiplexed as a diversity antenna for wireless LAN.
  • the antenna can be used in combination with a tuning switch.
  • the mobile communication module 150 can provide a wireless communication solution including 2G/3G/4G/5G and the like applied to the electronic device 100.
  • the mobile communication module 150 may include one or more filters, switches, power amplifiers, low noise amplifiers (LNA), etc.
  • the mobile communication module 150 can receive electromagnetic waves by the antenna 1, and perform processing such as filtering and amplifying the received electromagnetic waves, and transmitting them to the modem processor for demodulation.
  • the mobile communication module 150 can also amplify the signal modulated by the modem processor, and convert it into electromagnetic waves through the antenna 1 for radiation.
  • at least part of the functional modules of the mobile communication module 150 may be provided in the processor 110.
  • at least part of the functional modules of the mobile communication module 150 and at least part of the modules of the processor 110 may be provided in the same device.
  • the modem processor may include a modulator and a demodulator.
  • the modulator is used to modulate the low frequency baseband signal to be sent into a medium and high frequency signal.
  • the demodulator is used to demodulate the received electromagnetic wave signal into a low-frequency baseband signal. Then the demodulator transmits the demodulated low-frequency baseband signal to the baseband processor for processing. After the low-frequency baseband signal is processed by the baseband processor, it is passed to the application processor.
  • the application processor outputs sound signals through audio equipment (not limited to speakers 170A, receiver 170B, etc.), or displays images or videos through the display 194.
  • the modem processor may be an independent device. In other embodiments, the modem processor may be independent of the processor 110 and be provided in the same device as the mobile communication module 150 or other functional modules.
  • the wireless communication module 160 can provide applications on the electronic device 100, including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), Bluetooth (BT), and global navigation satellites. System (global navigation satellite system, GNSS), frequency modulation (FM), near field communication (NFC), infrared technology (infrared, IR) and other wireless communication solutions.
  • the wireless communication module 160 may be one or more devices integrating one or more communication processing modules.
  • the wireless communication module 160 receives electromagnetic waves via the antenna 2, frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110.
  • the wireless communication module 160 may also receive a signal to be sent from the processor 110, perform frequency modulation, amplify it, and convert it into electromagnetic waves through the antenna 2 for radiation.
  • the antenna 1 of the electronic device 100 is coupled to the mobile communication module 150, and the antenna 2 is connected to wireless communication
  • the module 160 is coupled so that the electronic device 100 can communicate with the network and other devices through wireless communication technology.
  • the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (general packet radio service, GPRS), code division multiple access (code division multiple access, CDMA), broadband Wideband code division multiple access (WCDMA), time-division code division multiple access (TD-SCDMA), long term evolution (LTE), BT, GNSS, WLAN, NFC , FM, and/or IR technology, etc.
  • the GNSS may include a global positioning system (GPS), a global navigation satellite system (GLONASS), a Beidou navigation satellite system (BDS), and a quasi-zenith satellite system (quasi). -zenith satellite system, QZSS) and/or satellite based augmentation systems (SBAS).
  • GPS global positioning system
  • GLONASS global navigation satellite system
  • BDS Beidou navigation satellite system
  • QZSS quasi-zenith satellite system
  • SBAS satellite based augmentation systems
  • the electronic device 100 implements a display function through GRJ, a display screen 194, and an application processor.
  • the GPU is a microprocessor for image processing, connected to the display 194 and the application processor.
  • the GPU is used to perform mathematical and geometric calculations, and is used for graphics rendering.
  • the processor 110 may include one or more GPUs, which execute program instructions to generate or change display information.
  • the display screen 194 is used to display images, videos, etc.
  • the display screen 194 includes a display panel.
  • the display panel can adopt liquid crystal display (LCD), organic light-emitting diode (OLED), active matrix organic light-emitting diode or active-matrix organic light-emitting diode (active-matrix organic light-emitting diode).
  • LCD liquid crystal display
  • OLED organic light-emitting diode
  • active-matrix organic light-emitting diode active-matrix organic light-emitting diode
  • AMOLED flexible light-emitting diode
  • FLED flexible light-emitting diode
  • Miniled MicroLed, Micro-oLed, quantum dot light emitting diodes (QLED), etc.
  • the electronic device 100 may include one or N display screens 194, and N is a positive integer greater than one.
  • the electronic device 100 can implement shooting functions through an ISP, a camera 193, a video codec, a GPU, a display screen 194, and an application processor.
  • the ISP is used to process the data fed back from the camera 193. For example, when taking a picture, the shutter is opened, the light is transmitted to the photosensitive element of the camera through the lens, the light signal is converted into an electrical signal, and the photosensitive element of the camera transfers the electrical signal to the ISP for processing and is converted into an image visible to the naked eye.
  • ISP can also optimize the image noise, brightness, and skin color. ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
  • the ISP may be provided in the camera 193.
  • the camera 193 is used to capture still images or videos.
  • the object generates an optical image through the lens and projects it to the photosensitive element.
  • the photosensitive element can be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the optical signal into an electrical signal, and then transfers the electrical signal to the ISP to convert it into a digital image signal.
  • ISP outputs digital image signals to DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other formats.
  • the electronic device 100 may include one or N cameras 193, and N is a positive integer greater than one.
  • the digital signal processor is used to process digital signals. In addition to processing digital image signals, it can also process other digital signals. For example, when the electronic device 100 selects the frequency point, the digital signal processor is used to perform Fourier transform on the energy of the frequency point.
  • Video codecs are used to compress or decompress digital video.
  • the electronic device 100 may support one or more video codecs. In this way, the electronic device 100 can play or record videos in multiple encoding formats, such as: moving picture experts group (MPEG)1, MPEG2, MPEG3, MPEG4, etc.
  • MPEG moving picture experts group
  • NPU is a neural-network (NN) computing processor.
  • NN neural-network
  • the NPU can realize applications such as intelligent cognition of the electronic device 100, such as image recognition, face recognition, voice recognition, text understanding, and so on.
  • the external memory interface 120 may be used to connect an external memory card, such as a Micro SD card, to expand the storage capacity of the electronic device 100.
  • the external memory card communicates with the processor 110 through the external memory interface 120 to realize the data storage function. For example, save music, video and other files in an external memory card.
  • the internal memory 121 may be used to store one or more computer programs, and the one or more computer programs include instructions.
  • the processor 110 can execute the above-mentioned instructions stored in the internal memory 121, so that the electronic device 100 executes the methods for intelligently recommending contacts provided in some embodiments of the present application, as well as various functional applications and data processing.
  • the internal memory 121 may include a program storage area and a data storage area.
  • the storage program area can store the operating system; the storage program area can also store one or more application programs (such as a gallery, contacts, etc.).
  • the data storage area can store data (such as photos, contacts, etc.) created during the use of the electronic device 101.
  • the internal memory 121 may include a high-speed random access memory, and may also include a non-volatile memory, such as one or more magnetic disk storage devices, flash memory devices, universal flash storage (UFS), and the like.
  • the processor 110 executes the instructions stored in the internal memory 121 and the instructions stored in the memory provided in the processor to cause the electronic device 100 to execute the smart recommendation provided in the embodiments of the present application. Number method, as well as various functional applications and data processing.
  • the electronic device 100 can implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, a headphone interface 170D, and an application processor. For example, music playback, recording, etc.
  • the audio module 170 is used to convert digital audio information into an analog audio signal for output, and also used to convert an analog audio input into a digital audio signal.
  • the audio module 170 can also be used to encode and decode audio signals.
  • the audio module 170 may be disposed in the processor 110, or some functional modules of the audio module 170 may be disposed in the processor 110.
  • the speaker 170A also called a “speaker” is used to convert audio electrical signals into sound signals.
  • the electronic device 100 can listen to music through the speaker 170A, or listen to a hands-free call.
  • the receiver 170B also called a "handset" is used to convert audio electrical signals into sound signals.
  • the electronic device 100 answers a call or voice message, it can receive the voice by bringing the receiver 170B close to the human ear.
  • Microphone 170C also called “microphone”, “microphone”, is used to convert sound signals into electrical signals.
  • the user can make a sound by approaching the microphone 170C through his mouth, and input the sound signal to the microphone 170C.
  • the electronic device 100 may be provided with one or more microphones 170C.
  • the electronic device 100 may be provided with two microphones 170C, which can not only collect sound signals, but also implement a noise reduction function.
  • the electronic device 100 may also be provided with three, four or more microphones 170C, which can collect sound signals, reduce noise, identify sound sources, and realize directional recording functions.
  • the microphone 170C can be used to collect the user's thousand voices when the user is singing a song, that is, a pure human voice without music.
  • the audio module 170 can use a corresponding sound modification algorithm to modify the collected voices of the user.
  • the audio module 170 may attenuate the breathing sound of the corresponding position in the user's thousand voices, thereby improving the listening quality of the singing voice .
  • the audio module 170 can detect whether the user's voice of song A also uses vibrato in the corresponding position. If the vibrato effect used by the user is not good, the audio module 170 can correct the vibrato in the corresponding position in the user's thousand voices, thereby improving the listening quality of the singing voice.
  • the audio module 170 can identify the out-of-tune position in the user's thousand voices by comparing the user's thousand voices with the original song A, and further, the audio module 170 can increase the pitch of the user's thousand voices accordingly. Or falling tone, from And fix the intonation problem of users' thousand voices.
  • the specific method for the electronic device 100 to correct the user's thousand voices collected by the microphone 170C will be described in detail in subsequent embodiments, so it will not be repeated here.
  • the electronic device can not only calibrate the intonation of the song recorded by the user, but also repair the singing technique problems such as ventilation and vibrato in the song recorded by the user.
  • ordinary users who have not undergone professional vocal training can also record songs with better listening quality, thereby improving the recording effect and experience of users when using K song applications.
  • Headphone jack 170D is used to connect wired headphones.
  • the earphone interface 170D can be a USB interface 130, or a 3.5mm open mobile terminal platform (OMTP) standard interface, and the American cellular telecommunications industry association of the USA (CTIA) standard interface .
  • OMTP open mobile terminal platform
  • CTIA American cellular telecommunications industry association of the USA
  • the sensor 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, and an ambient light sensor 180L , Bone conduction sensor 180M, etc.
  • the button 190 includes a power button, a volume button, and so on.
  • the button 190 may be a mechanical button. It can also be a touch button.
  • the electronic device 100 can receive key input, and generate key signal input related to user settings and function control of the electronic device 100.
  • the motor 191 can generate vibration prompts.
  • the motor 191 can be used for incoming call vibration notification, and can also be used for touch vibration feedback.
  • touch operations applied to different applications can correspond to different vibration feedback effects.
  • Acting on touch operations in different areas of the display screen 194, the motor 191 can also correspond to different vibration feedback effects.
  • Different application scenarios for example: time reminder, receiving message, alarm clock, game, etc.
  • the touch vibration feedback effect can also support customization.
  • the indicator 192 may be an indicator light, which may be used to indicate a charging state, a change in power, and may also be used to indicate messages, missed calls, notifications, and the like.
  • the SIM card interface 195 is used to connect to the SIM card.
  • the SIM card can be inserted into the SIM card interface 195 or pulled out from the SIM card interface 195 to achieve contact and separation with the electronic device 100.
  • the electronic device 100 may support one or N SIM card interfaces, and N is a positive integer greater than one.
  • SIM card interface 195 can support Nano SIM card, Micro SIM card, SIM card and so on.
  • the same SIM card interface 195 can insert multiple cards at the same time. The types of the multiple cards can be the same or different.
  • the SIM card interface 195 can also be compatible with different types of SIM cards.
  • the SIM card interface 195 is also compatible with external memory cards.
  • the electronic device 100 interacts with the network through the SIM card to implement functions such as call and data communication.
  • the electronic device 100 adopts an eSIM, that is, an embedded SIM card.
  • the eSIM card can be embedded in the electronic device 100 and cannot be separated from the electronic device 100.
  • the above-mentioned software system of the electronic device 100 may adopt a layered architecture, an event-driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture.
  • the embodiment of the present application takes an Android system with a layered architecture as an example to exemplarily illustrate the software structure of the electronic device 100.
  • FIG. 2 is a software structure block diagram of the electronic device 100 according to an embodiment of the present application.
  • the layered architecture divides the software into thousands of layers, and each layer has a clear role and division of labor. Communication between layers through software interface.
  • the Android system is divided into four layers, from top to bottom, the application layer, the application framework layer, the Android runtime and system libraries, and the kernel layer.
  • the application layer can include a series of application packages.
  • the application package can include apps (applications) such as call, contact, camera, gallery, calendar, map, navigation, Bluetooth, music, video, short message, etc. »
  • K song applications may also be installed in the application layer, such as Sing Bar APP, Quanmin K Song APP, etc. Users can use karaoke apps to record songs and incorporate their own singing voices into the accompaniment provided by karaoke apps.
  • the application framework layer provides an application programming interface (API) and a programming framework for applications in the application layer.
  • API application programming interface
  • the application framework layer includes some predefined functions.
  • the application framework layer may include a window manager, a content provider, a view system, a resource manager, a notification manager, and so on.
  • the window manager is used to manage window programs.
  • the window manager can obtain the size of the display screen, determine whether there is a status bar, lock the screen, capture the screen, etc.
  • the content provider is used to store and retrieve data, and make these data accessible to applications.
  • the data may include videos, images, audios, calls made and received, browsing history and bookmarks, phonebooks, etc.
  • the view system includes visual controls, such as controls that display text, controls that display pictures, and so on.
  • the view system can be used to build applications.
  • the display interface can be composed of one or more views. For example, a display interface that includes a short message notification icon may include a view that displays text and a view that displays pictures.
  • the resource manager provides various resources for the application, such as localized strings, icons, pictures, layout files, video files, and so on.
  • the notification manager enables applications to display notification information in the status bar, which can be used to convey notification-type messages, and can disappear automatically after a short stay without user interaction. For example, the notification manager is used to notify download completion, message reminders, etc.
  • the notification manager can also be a notification that appears in the status bar at the top of the system in the form of a chart or a scroll bar text, such as a notification of an application running in the background, or a notification that appears on the screen in the form of a dialog window. For example, text messages are prompted in the status bar, prompt sounds are emitted, electronic devices vibrate, and indicator lights flash.
  • the system library can include multiple functional modules. For example: surface manager (surface manager), media library (Media Libraries), 3D graphics processing library (for example: OpenGL ES), 2D graphics engine (for example: SGL), etc.
  • surface manager surface manager
  • media library Media Libraries
  • 3D graphics processing library for example: OpenGL ES
  • 2D graphics engine for example: SGL
  • the surface manager is used to manage the display subsystem and provides a combination of 2D and 3D layers for multiple applications.
  • the media library supports playback and recording of a variety of commonly used audio and video formats, as well as still image files.
  • the media library can support multiple audio and video coding formats, such as: MPEG4, H.264, MP3, AAC, AMR, JPG, PNG, etc.
  • the 3D graphics processing library is used to realize 3D graphics drawing, image slag staining, synthesis, and layer processing.
  • the 2D graphics engine is a drawing engine for 2D drawing.
  • the related sound modification algorithm provided in the embodiment of the present application can be set in the system library in the form of an algorithm dynamic link library (DLL).
  • DLL algorithm dynamic link library
  • the algorithm dynamic library of the sound modification algorithm can be stored in the system/lib directory.
  • karaoke applications can call the algorithm dynamic library of the tuning algorithm to correct the user's thousands of voices intonation, vibrato, or ventilatory sound, and realize the tuning function.
  • Android Runtime includes core libraries and virtual machines. Android runtime is responsible for the scheduling and management of the Android system.
  • the core library consists of two parts: one part is the function functions that the java language needs to call, and the other part is the core library of Android.
  • the application layer and the application framework layer run in a virtual machine.
  • the virtual machine executes the java files in the application layer and application framework layer as binary files.
  • the virtual machine is used to perform functions such as object life cycle management, stack management, thread management, security and exception management, and garbage collection.
  • the kernel layer is the layer between hardware and software.
  • the kernel layer includes at least a display driver, a camera driver, an audio driver, a sensor driver, etc., which are not limited in the embodiment of the present application.
  • the method includes the following steps S301-S306.
  • K song applications refer to application software that can provide users with the functions of song ordering, song recording, and sound repair, for example,
  • the K song application may be the Quanmin K song APP, Sing Bar APP, etc., which are not restricted in the embodiment of the application.
  • the mobile phone can start to run the K song application and display the first interface 401 of the K song application. Users can search for their favorite songs in the first interface 401 and record them. Alternatively, the user can manually search and select the song to be recorded in the search bar 402. Of course, the user can also select the song that needs to be recorded this time in other interfaces of the K song application, and the embodiment of the application does not impose any restriction on this.
  • the mobile phone can receive the user's operation to record the song "Red Bean” .
  • the mobile phone can continue to perform the following steps S302-S306 to complete the process of recording and repairing the song.
  • the mobile phone obtains the accompaniment of song A and the characteristic information of the original song.
  • the mobile phone can download the accompaniment of the song "Red Bean” from the server.
  • the mobile phone can obtain the accompaniment from the local storage.
  • the part of the song except the human voice can be called accompaniment, and some accompaniment can also include harmony and background sound, etc.
  • the mobile phone after detecting that the user clicks the record button 403 of the song "Red Bean", the mobile phone can also download the characteristic information of the original song of the song "Red Bean” from the server.
  • the characteristic information may reflect the singing characteristics of the original singer (for example, Faye Wong) when singing the song "Red Bean”.
  • the feature information may include one or more of the fundamental frequency information of the original singing voice, the ventilation position, and the vibrato feature.
  • the vibrato feature may include one or more of the vibrato position, the fundamental frequency stability information of the vibrato, the vibrato frequency, and the vibrato amplitude.
  • the fundamental frequency information can reflect the fundamental tone of the sound.
  • the server can extract the fundamental frequency-time variation curve f(t) in units of frames from the original song of the song "Red Bean",
  • the change curve f(t) is the fundamental frequency information of the song "Red Bean”.
  • the server can also detect the position of the original vocal's breathing in the original vocals, and use the time of the original vocal's breathing as a feature information.
  • the server can also detect the location of the vibrato in the original singing voice, and use the moment of the vibrato as a feature information.
  • the vibrato appears as periodic pitch changes. From the fundamental frequency curve with tremolo characteristics, the three time-varying characteristics of tremor amplitude A(t), tremor frequency 1 (1) and stable trend item V (1) can be extracted.
  • the fundamental frequency curve corresponding to vibrato 1 is f3(1) ⁇ (1)+ ⁇ (1) cos (1(1)), t 6 [8s, 10s].
  • V(1) is the stable information of the fundamental frequency of the vibrato 1
  • 1(1) is the vibrato frequency of the vibrato 1
  • A(1) is the vibrato amplitude of the vibrato 1.
  • the server can send the fundamental frequency stability information, the frequency of the vibrato and the amplitude of the vibrato of each vibrato in the song "Red Bean" as characteristic information to the mobile phone.
  • the server may extract corresponding characteristic information according to the original song, and save the extracted characteristic information in the server. Later, if the user chooses to record a certain song in the mobile phone, the mobile phone can obtain the accompaniment of the song from the server, and also obtain the characteristic information of the original song corresponding to the song from the server.
  • the acquired characteristic information of the original vocals can be used to guide users to record songs, and can also be used as a basis for tuning the songs recorded by users.
  • the mobile phone displays the recording interface of the K song application, it plays the accompaniment of song A and records the user's thousand voices sung by the user.
  • the mobile phone can start playing the accompaniment of the song “Red Bean”. At the same time, the mobile phone can turn on the microphone to start recording the user's thousand voices of the user singing the song "Red Bean”. As shown in FIG. 6, during the process of recording the song "Red Bean", the mobile phone may display the recording interface 601 of the song "Red Bean".
  • the recording interface 601 can display the lyrics 602 of the song “Red Bean", and the position of the lyrics currently being recorded can be marked in the lyrics 602 in real time, so that the user can record the song according to the prompts of the lyrics 602.
  • the mobile phone When recording song A, the mobile phone prompts the user to adjust singing methods such as pitch, ventilation, or vibrato according to the characteristic information of the original song.
  • the mobile phone may guide the user to sing the song when displaying the lyrics 602 according to the characteristic information obtained in step S302, for example, instruct the user to perform singing methods such as pitch, ventilation or vibrato.
  • the characteristic information of the original singing voice acquired by the mobile phone includes information such as the ventilation position and the vibrato position.
  • the mobile phone can mark the ventilation position 701 in the corresponding position of the lyrics 602 according to the ventilation position recorded in the characteristic information of the original singing thousand voices.
  • the user can ventilate according to the vent position 701 suggested in the lyrics 602, thereby helping the user control the breath changes during singing, thereby reducing the obvious and disorder in the recorded song To improve the listening quality of recorded songs.
  • the mobile phone can also mark the vibrato position 702 in the corresponding position of the lyrics 602 according to the vibrato position recorded in the characteristic information of the original singing thousand voices (ie Vibrato prompt).
  • the lyrics can be used in accordance with 702 602 in tips vibrato vibrato singing position, thereby enhancing the sense of depth and song singing skills when users singing, the sense of hearing to improve the quality of the recording of the song ⁇
  • the mobile phone in addition to guiding the user to sing in the lyrics displayed in real time, can also set a singer in the recording interface, and the mobile phone can guide the user in the singer to adjust the singing methods such as pitch, ventilation or vibrato.
  • the singer 801 can be set in the recording interface 601.
  • the singer 801 can display the notes of each word in the lyrics currently being recorded.
  • the mobile phone can display the pitch line 802 (that is, the pitch prompt) of each note in the singer 801 in real time according to the fundamental frequency information of the original singing thousand voices. If the pitch of the current note is higher, the position of the pitch line 802 is higher.
  • the mobile phone can display the vibrato note 803 (that is, the vibrato reminder) in the singer 801 according to the vibrato position of the original singing voice.
  • the vibrato note 803 can be represented by a wavy line or a line segment of a specific color.
  • the vibrato note 803 appears at the same position as the vibrato in the original song.
  • the mobile phone can display the ventilating symbol 804 in the singer 801 (that is, the ventilating prompt) according to the ventilation position in the original singing voice.
  • the ventilation symbol 804 can be represented by graphics of a specified shape such as a triangle or a circle. Ventilation symbol 804 appeared in the same position as the original vocal sound.
  • the user when recording a song, the user can adjust the pitch of the human voice according to the pitch line displayed in real time in the singer 801, or sing with vibrato at the corresponding position according to the real-time display of the vibrato in the singer 801. Ventilation symbols displayed in real-time in the video are ventilated at the corresponding position, so as to guide the user's singing skills such as pitch, vibrato, and ventilation during the recording process, and improve the recording quality of the song.
  • the mobile phone evaluates the user's singing by comparing the characteristic information of the original singing thousand voices with the characteristic information of the user thousand voices.
  • the mobile phone can also evaluate the user's current singing in real time based on the characteristic information of the original song.
  • the mobile phone may evaluate the user's singing with the granularity of the words, words or sentences in the lyrics. As shown in Figure 9, the mobile phone collected the user’s singing "I will believe that everything has an end.” After the user’s thousand voices of the sentence, the mobile phone can extract the characteristic information of the user’s thousand voices of the sentence.
  • the mobile phone can first extract the fundamental frequency information of the user thousand voices in this sentence, and then detect the user's ventilation position, vibrato position, and fundamental frequency stability information of the vibrato based on the fundamental frequency information of the user thousand voices , Tremolo frequency and tremolo amplitude and other one or more characteristic information.
  • the mobile phone compares the characteristic information 1 when singing the lyrics "I will believe that everything has an end” and the characteristic information 2 when the user sings the lyrics "I will believe everything has an end” to confirm that the user sings Evaluation results of singing skills such as intonation, vibrato, and ventilation of this sentence.
  • the mobile phone may display a prompt 1001 of "Pitch Master" in the recording interface 601, prompting the user to sing this sentence with the correct pitch.
  • the mobile phone can display a prompt 1002 of “Quaming Queen” in the recording interface 601.
  • the mobile phone can In the recording interface 601, a prompt 1003 of "getting a breath" is displayed.
  • the mobile phone can also display the current evaluation results to the user in real time through colors, vibrations, or special effects.
  • the vibrato used in the lyrics does not match the vibrato used in the original sing.
  • the lyric "End” 1101 corresponding to the vibrato mark the "End” 1101 as a different color from other lyrics or use a different font.
  • the mobile phone can display the lyrics corresponding to the ventilation position Mark 1102 for a ventilation error.
  • a mobile phone can also display the evaluation result to the mobile phone user in the determined recording concert interface 801 ⁇ 601 e.g., in FIG. 12 (a), the user singing "yes” when the word, the phone may not
  • the singer 801 displays the tone line 1201 of the original singing of the word "you” in real time, and can also display the tone line 1202 of the user singing the word "you” at the same time. If the user’s fundamental frequency of singing the character " ⁇ " is lower than the original singing of the character “ ⁇ ”, it means that the user’s pitch of singing the character “ ⁇ ” is lower than the original singing of the character “ ⁇ ”, then the phone will display the pitch The line 1202 is lower than the tone line 1201.
  • the pitch line 1202 displayed by the mobile phone may be higher than the pitch line 1201.
  • the pitch line 1202 in the singer 801 coincides with the pitch line 1201. In this way, the user can learn the current pitch of the singing in the singer 801 during the real-time singing.
  • the mobile phone can display the vibrato note 1203 of the original singer singing vibrato 1 in the singer 801. After the mobile phone detects the user's thousand voices when the user sings the vibrato 1, it can compare the position of the vibrato when the user sings the vibrato 1, the vibrato amplitude A(1), the vibrating frequency 1(1), and the stable trend item V(1) with the original singer's vibrato position and the vibrato amplitude when the vibrato 1 is sung Whether A(t), tremor frequency 1 (1) and stable trend item V (1) match.
  • the mobile phone can display the vibrato note 1203 as a dotted line and other display effects.
  • the mobile phone can bold or light up the vibrato note 1203 so that the user can In the process of real-time singing, the current vibrato singing effect can be obtained in the singer 801.
  • the levy information evaluates the user’s thousand voices sung in real time when the user is recording a song, and displays the evaluation results of ventilation skills, vibrato skills, or intonation to the user.
  • the mobile phone can check the user’s thousand voices based on the characteristic information of the original song. The scoring is performed and the scoring result is displayed to the user.
  • the embodiment of the present application does not have any limitation on this.
  • the embodiment of the application can perform real-time evaluation of one or more singing skills such as pitch, vibrato, and ventilation of the user according to the characteristic information of the original singing thousand voices, which is compared with The user’s singing level is evaluated purely based on the pitch, and the embodiments of the application also consider the user’s ventilation, vibrato and other factors when evaluating the user’s singing, so the obtained evaluation result is more accurate, rich and objective.
  • the mobile phone can feed back the evaluation results of the user's singing to the user in real time during the recording process, so that the user can understand his current singing effect in time, and the user can adjust the singing method in time during the subsequent recording process.
  • the mobile phone modifies the user's thousand voices according to the characteristic information of the original song A thousand voices and the characteristic information of the user thousand voices.
  • the mobile phone When the mobile phone finishes playing the accompaniment of song A, or detects the user input to end the recording operation, the mobile phone can turn off the microphone and no longer continue to collect the user's thousand voices. Moreover, after the recording of song A is finished, as shown in FIG. 13, the mobile phone may display the sound modification interface 1301 of the K song application.
  • the sound modification interface 1301 may include a sound modification switch 1302. If it is detected that the user has turned on the tuning switch 1302, the mobile phone can respond to the singing problems (such as pitch, vibrato, or tremor) that occurred in the user’s thousand voices collected when the user recorded song A according to the characteristic information of the original singing voice of song A. One or more singing problems, such as breathing sound, etc.) to be repaired.
  • the phone may display the sound modification interface 1301 as shown in FIG. If it is detected that the user has turned on the tone repair switch 1302 in the tone repair interface 1301, the mobile phone can repair the pitch, ventilatory sound and vibrato of the user's thousands of voices collected this time according to the characteristic information of the original singing voice.
  • the mobile phone can compare the fundamental frequency information of the original singing voice with the fundamental frequency information of the user's recorded song "Red Bean", so as to correct the intonation problem in the user's voice.
  • the mobile phone can increase the pitch of the thousand voices of the users
  • the processing makes the tone of the thousand voices of this part of users the same as or close to that of the original voice of thousand voices.
  • the mobile phone can respond to these users The pitch of the voice is reduced, so that the pitch of the thousand voices of this part of the user is the same or close to the pitch of the original voice.
  • the mobile phone can extract the position of the breathing sound from the user's thousand voices of the song "Red Bean" recorded this time by the user.
  • the mobile phone may use a preset ventilation detection model to detect that a ventilation sound occurs at the 14th and 78th seconds of the user's thousand voices.
  • the mobile phone can reduce the energy and frequency response of the user's thousand voices at the 14th and 78th seconds, thereby weakening the ventilation sound at the 14th and 78th seconds, so as to improve the listening quality of the recorded song.
  • the mobile phone can compare the vibrato characteristics of the original singing thousand voices with the vibrato characteristics of the user's current recorded song "Red Bean", so as to correct the vibrato of the user's thousand voices with poor vibrato effects.
  • the mobile phone may use a preset vibrato detection model to detect the location of the vibrato in the user's thousand voices. For example, if the vibrato 1 is detected at the 15th second in the user's thousand voices, if the vibration amplitude of the vibrato 1 is Al(t), the vibration frequency 1 1 (1) and the stable trend item VI (1) is the same as the original singing thousand voices at the 15th second.
  • the mobile phone can use the vibration amplitude A2(t) of the vibrato 2, the vibration frequency 1 2(1) and the stable trend item V2(1) Synthesize a new vibrato in the user's thousand voices in the 15th second, thereby improving the vibrato sound effect in the user's thousand voices.
  • the mobile phone may further display a tuning option 1401 for correcting the pitch, and a tuning option 1402 for correcting the ventilation sound. And the modification option 1403 to correct the vibrato. In this way, the user can manually select in the sound repair interface 1301 to repair the pitch, ventilation sound, or vibrato of the user's thousand voices recorded this time.
  • the mobile phone after the mobile phone collects the user’s thousand voices of the user’s recording of the song “Red Bean” this time, it can also automatically correct issues such as the pitch, ventilation sound, and vibrato of the user’s thousand voices to improve subsequent playback and recording.
  • the quality of the song's listening experience For example, as still shown in FIG. 14, a play button 1404 is provided in the sound modification interface 1301. If it is detected that the user clicks the play button 1404, the mobile phone can automatically correct the pitch, ventilatory sound, and vibrato of the user’s voice according to the characteristic information of the original song, and integrate the corrected voice into the user’s voice. Play with the accompaniment of the song "Red Bean".
  • an option 1501 for single sentence refinement may be set in the sound modification interface 1301 of the K song application. If it is detected that the user clicks on option 1501, as shown in (b) in Figure 15, the mobile phone can display a refined interface 1502. In the refined interface 1502, the mobile phone can display every lyrics sung by the user in the song "Red Bean", and each sentence of lyrics is used as an option for the user to choose. Take the lyrics 1503 that the user selects "I will believe that everything has an end" as an example. After the user selects the lyrics 1503, the mobile phone can play some of the recorded songs corresponding to the lyrics 1503.
  • the tuning interface 1301 also includes a pitch button 1504, a vibrato button 1505, and a ventilation button 1506.
  • the user can manually select in the refinement interface 1502 whether to repair one or more singing problems such as pitch, vibrato, or ventilating sound in the user's thousand voices corresponding to the lyrics 1503. If the user clicks the vibrato button 1525 and the pitch button 1504 after selecting the lyrics 1503, the mobile phone can correct the pitch and vibrato problems in the user's thousand voices corresponding to the lyrics 1503, and integrate the corrected user thousand voices into the corresponding accompaniment Play in. In this way, the user can correct the singing problem in the user's thousand voices recorded this time sentence by sentence, and the user can compare the recording effect before and after the correction sentence by sentence, thereby correcting the song recorded by the user to a more satisfactory singing effect for the user.
  • the single-sentence editing option 1601 may be set in the sound modification interface 1301 of the K song application. If it is detected that the user clicks on option 1601, as shown in (b) in Figure 16, the mobile phone can display an editing interface 1602. In the editing interface 1602, the mobile phone can display every lyric sung by the user in the song "Red Bean", and, when displaying each lyric, the mobile phone can also prompt the user to sing the lyrics in the corresponding position of the lyrics. For example, inaccurate pitch, loud breathing sound, unstable vibrato, lack of vibrato, etc. In this way, the user can intuitively see the specific problems that arise when he sings each lyrics.
  • an editing button 1603 corresponding to each sentence of lyrics can also be set in the editing interface 1602. If it is detected that the user clicks on the edit button 1603 corresponding to a certain sentence of lyrics, the mobile phone can play the accompaniment corresponding to the lyrics, and turn on the microphone to collect the user's thousand voices of the user who re-sing the lyrics. After the mobile phone collects the user's thousand voices of the user re-singing the song, the user thousand voices can be used to replace the corresponding user thousand voices in the recorded song "Red Bean". Since the mobile phone displays the lyrics on the editing interface 1602, the user has been reminded of the specific singing problems when singing the lyrics. Therefore, when the user re-sings the lyrics, the user can adjust the singing method in time to correct the singing problems that occurred during the last recording of the lyrics, thereby improving singing And recording quality.
  • an embodiment of the present application discloses an electronic device, including: a touch screen 1701, the touch screen 1701 includes a touch-sensitive surface 1706 and a display screen 1707; one or more processors 1702; a memory 1703; one Or more microphones 1708; and one or more computer programs 1704.
  • the aforementioned devices may be connected through one or more communication buses 1705.
  • the aforementioned one or more computer programs 1704 are stored in the aforementioned memory 1703 and are configured to be executed by the one or more processors 1702, and the one or more computer programs 1704 include instructions, and the aforementioned instructions can be used to execute the aforementioned Each step in the embodiment should be implemented.
  • the foregoing processor 1702 may specifically be the processor 110 shown in FIG.
  • the foregoing memory 1703 may specifically be the internal memory 121 and/or the external memory 120 shown in FIG. 1, and the foregoing display screen 1707 may specifically be FIG.
  • the microphone 1708 may specifically be the microphone 170C shown in FIG. 1
  • the touch-sensitive surface 1706 may specifically be the touch sensor 180K in the sensor module 180 shown in FIG. 1, which is not the case in this embodiment of the application. Do any restrictions.
  • the application also provides a graphical user interface (GUI), which can be stored in an electronic device.
  • GUI graphical user interface
  • the electronic device may be the electronic device shown in FIG. 1 or FIG. 17.
  • the foregoing graphical user interface includes: a first GUI displayed on the touch screen, as shown in FIG. 4, the first GUI may be an interface 401 of a K song application, and the first GUI includes a first GUI for recording the first song. Recording button 403; In response to a touch event for the recording button, the electronic device may display a second GUI on the touch screen, as shown in FIG. 7 or FIG. 8.
  • the second GUI may be a recording interface 601 for the user to record the first song.
  • the second GUI may include guidance information for instructing the user to sing the first song, and the guidance information includes ventilation prompts and/or vibrato prompts. For example, as shown in FIG.
  • the lyrics 602 displayed on the recording interface 601 include a ventilation prompt 701 and a vibrato prompt 702.
  • the guidance information in the second GUI is generated by the electronic device based on the characteristic information of the original song.
  • the aforementioned second GUI may also include an evaluation result of the user's ventilation skills and/or vibrato skills during singing.
  • the recording interface 601 includes an evaluation result 1001 of the user's ventilation skills, an evaluation result 1002 of the user's vibrato skills, and an evaluation result 1003 of the user's pitch.
  • the evaluation result of the user's pitch 1202, the evaluation result of the user's vibrato skills 1203, etc. can be displayed in the singer of the recording interface 601. The embodiment of the present application There are no restrictions on this.
  • the electronic device may display a third GUI on the touch screen, as shown in FIG. 13 or 14, the third GUI may be the sound modification interface 1301 of the first song, and the third GUI
  • the sound repair switch 1302 is included; when the user turns on the sound repair switch 1302, the electronic device can correct the singing problems such as the ventilation problem and the vibrato problem in the user's thousand voices.
  • the functional units in the various embodiments of the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated unit can be realized in the form of hardware or software functional unit.
  • the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
  • a computer readable storage medium includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute all or part of the steps of the methods described in the various embodiments of the present application.
  • the aforementioned storage media include: flash memory, mobile hard disk, read-only memory, random access memory, magnetic disk or optical disk and other media that can store program codes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Signal Processing (AREA)
  • Reverberation, Karaoke And Other Acoustics (AREA)
  • Electrophonic Musical Instruments (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

本申请的实施例提供一种歌曲录制方法、修音方法及电子设备,涉及音频技术领域,可在用户录制的歌曲中添加相应的演唱技巧,从而提高用户使用K歌类应用时的录制效果和使用体验。该方法包括:电子设备显示第一界面,第一界面中包括用于录制第一歌曲的录制按钮;响应于用户选择所述录制按钮的操作,电子设备获取所述第一歌曲的伴奏和原唱干声的特征信息,所述原唱干声的特征信息用于反映原唱在演唱所述第一歌曲时的演唱特征;电子设备开始播放所述第一歌曲的伴奏并开始录制用户演唱的用户干声;在录制用户干声的过程中,电子设备根据所述原唱干声的特征信息在第二界面中显示指导信息,所述指导信息用于指导用户演唱时的换气或颤音中的一项或多项。

Description

一种歌曲录制方法、 修音方法及电子设备 本申请要求在 2019年 2月 28 日提交中国国家知识产权局、 申请号为 201910152655. 3、 发明名称为 “一种歌曲录制方法、 修音方法及电子设备” 的中国专利申请的优先权, 其全部 内容通过引用结合在本申请中。 技术领域
本申请涉及音频技术领域, 尤其涉及一种歌曲录制方法、 修音方法及电子设备。
背景技术
对于喜爱唱歌的用户, K歌类应用已经成为手机、 平板电脑等电子设备上安装的热门应 用。 示例性的, 用户打开 K歌类应用后可选择需要录制的歌曲, 进而, K歌类应用可加载歌 曲的伴奏, 并调用手机上的麦克风等拾音组件录制用户演唱该歌曲的音频, 从而把用户自己 的歌声融入在 K歌类应用提供的伴奏中。
目前, 为了提高用户使用 K歌类应用的歌唱体验, 一些 K歌类应用对用户录制的歌曲提 供了修音功能。 例如, K歌类应用可将用户录制的歌曲 A与歌曲 A的原唱音频进行比较, 将 用户录制的歌声的音准和节奏修改为和原唱音频中的音准和节奏, 从而修正用户录制歌曲时 出现的音准和节奏问题。
但除了音准和节奏问题外, 普通用户一般没有经过专业的声乐训练, 因此, 唱功技巧不 足的用户录制歌曲时仍然无法达到很好的录制效果。
发明内容
本申请提供一种歌曲录制方法、 修音方法及电子设备, 可在用户录制歌曲时指导用户使 用相应的演唱技巧, 从而提高用户使用 K歌类应用时的录制效果和使用体验。
为达到上述目的, 本申请采用如下技术方案:
第一方面, 本申请提供一种歌曲录制方法, 包括: 电子设备显示第一界面, 第一界面中 包括用于录制第一歌曲的录制按钮; 如果检测到用户选择该录制按钮, 电子设备可获取第一 歌曲的伴奏和原唱千声的特征信息, 该原唱千声的特征信息可用于反映原唱在演唱第一歌曲 时的演唱特征; 进而, 电子设备可以开始播放第一歌曲的伴奏并开始录制用户演唱的用户千 声; 在录制用户千声的过程中, 电子设备可根据原唱千声的特征信息在第二界面中显示指导 信息, 该指导信息用于指导用户演唱时的换气或颤音中的一项或多项。 也就是说, 用户在录 制歌曲时, 电子设备可以根据原唱的演唱特征显示指导信息, 指导用户演唱时的换气或颤音 等一项或多项演唱技巧, 从而提高用户录制歌曲的录制效果和使用体验。
在一种可能的实现方式中, 上述原唱千声的特征信息包括原唱千声的换气位置; 此时, 上述指导信息可以包括换气提示; 那么, 电子设备根据原唱千声的特征信息在第二界面中显 示指导信息, 包括: 电子设备可以根据原唱千声的换气位置, 在第二界面中显示与原唱千声 的换气位置对应的换气提示, 从而帮助用户控制演唱时的气息变化, 从而减少录制的歌曲中 出现较为明显和紊乱的换气声, 以提高录制歌曲的听感质量。
示例性的, 上述第二界面中可以包括第一歌曲的歌词; 此时, 电子设备根据原唱千声的 换气位置, 在第二界面中显示与原唱千声的换气位置对应的换气提示, 包括: 电子设备在歌 词的第一位置显示该换气提示, 第一位置与原唱千声的换气位置相同。
或者, 上述第二界面中可以包括演唱器, 演唱器用于显示正在录制的歌词中每个字的音 符; 此时, 电子设备根据原唱千声的换气位置, 在第二界面中显示与原唱千声的换气位置对 应的换气提示, 包括: 电子设备在演唱器的第二位置显示该换气提示, 第二位置与原唱千声 的换气位置相同。
在一种可能的实现方式中, 上述原唱千声的特征信息包括原唱千声的颤音特征; 此时对 应的指导信息包括颤音提示; 那么, 电子设备根据原唱千声的特征信息在第二界面中显示指 导信息, 包括: 电子设备根据原唱千声的颤音特征, 在第二界面中显示与原唱千声的颤音特 征对应的颤音提示。 这样, 用户在录制歌曲时, 可以根据歌词中提示的颤音位置使用颤音演 唱, 从而提高用户演唱时的歌曲的层次感和演唱技巧, 以提高录制歌曲的听感质量。
示例性的, 上述颤音特征中包括颤音位置; 上述第二界面中包括第一歌曲的歌词; 此时, 电子设备根据原唱千声的颤音特征, 在第二界面中显示与原唱千声的颤音特征对应的颤音提 示, 包括: 电子设备在歌词的第三位置显示颤音提示, 第三位置与原唱千声中的颤音位置相 同。
或者, 上述颤音特征中包括颤音位置; 上述第二界面中包括演唱器, 演唱器用于显示正 在录制的歌词中每个字的音符; 此时, 电子设备根据原唱千声的颤音特征, 在第二界面中显 示与原唱千声的颤音特征对应的颤音提示, 包括: 电子设备在演唱器的第四位置显示颤音提 示, 第四位置与原唱千声中的颤音位置相同。
示例性的, 上述原唱千声的特征信息还可以包括原唱千声的基频信息; 电子设备可以根 据原唱千声的基频信息生成音准提示显示在第二界面中, 指导用户演唱时的音准。
在一种可能的实现方式中, 在录制用户千声的过程中, 上述方法还包括: 电子设备从用 户千声中提取用户千声的特征信息, 用户千声的特征信息用于反映用户在演唱第一歌曲时的 演唱特征; 电子设备通过对比用户千声中的特征信息与原唱千声的特征信息, 在第二界面中 显示评估结果。
可以看出, 用户录制歌曲的过程中电子设备可根据原唱千声的特征信息, 对用户演唱的 音准、 颤音以及换气等一项或多项演唱技巧进行实时评估, 相比于单纯依据音准评估用户的 演唱水平, 本申请实施例在评估用户演唱时还考虑了用户的换气、 颤音等因素, 因此得到的 评估结果更加准确、 丰富和客观。 并且, 电子设备可将对用户演唱的评估结果在第二界面中 实时反馈给用户, 便于用户及时了解自己当前的演唱效果, 使用户可以在后续录制过程中及 时调整演唱方式。
在一种可能的实现方式中, 用户千声中的特征信息包括用户千声的换气位置, 原唱千声 的特征信息包括原唱千声的换气位置; 其中, 电子设备通过对比用户千声中的特征信息与原 唱千声的特征信息, 在第二界面中显示评估结果, 包括: 电子设备通过对比用户千声的换气 位置和原唱千声的换气位置, 在第二界面中显示换气技巧的评估结果。
示例性的, 电子设备通过对比用户千声的换气位置和原唱千声的换气位置, 在第二界面 中显示换气技巧的评估结果, 包括: 若用户千声的换气位置与原唱千声的换气位置相同, 则 电子设备在第二界面中显示换气准确的提示; 若用户千声的换气位置与原唱千声的换气位置 不相同, 则电子设备在第二界面中显示换气错误的提示。
在一种可能的实现方式中, 用户千声中的特征信息包括用户千声的颤音特征, 原唱千声 的特征信息包括原唱千声的颤音特征; 其中, 电子设备通过对比用户千声中的特征信息与原 唱千声的特征信息, 在第二界面中显示评估结果, 包括: 电子设备通过对比用户千声的颤音 特征和原唱千声的颤音特征, 在第二界面中显示颤音技巧的评估结果。
示例性的, 电子设备通过对比用户千声的颤音特征和原唱千声的颤音特征, 在第二界面 中显示颤音技巧的评估结果, 包括: 若用户千声的颤音特征与原唱千声的颤音特征匹配, 则 电子设备在第二界面中显示颤音准确的提示; 若用户千声的颤音特征与原唱千声的颤音特征 不匹配, 则电子设备在第二界面中显示颤音错误的提示。
在一种可能的实现方式中, 电子设备在第二界面中显示评估结果, 包括: 电子设备可在 第二界面的歌词中显示该评估结果; 或者, 电子设备也可在第二界面的演唱器中显示该评估 结果。
在一种可能的实现方式中, 在电子设备根据原唱千声的特征信息在第二界面中显示指导 信息之后, 还包括: 第一歌曲录制结束后, 电子设备根据原唱千声的特征信息修正用户千声 中出现的演唱问题, 该演唱问题包括换气问题或颤音问题中的一项或多项。 这样, 没有经过 专业的声乐训练的普通用户也可以录制出听感质量较好的歌声, 从而提高用户录制歌曲时的 录制效果和使用体验。
第二方面, 本申请提供一种修音方法, 包括: 电子设备显示第一界面, 第一界面中包括 用于录制第一歌曲的录制按钮; 如果检测到用户选择该录制按钮, 电子设备可获取第一歌曲 的伴奏和原唱千声的特征信息, 原唱千声的特征信息用于反映原唱在演唱第一歌曲时的演唱 特征; 进而, 电子设备可开始播放第一歌曲的伴奏并开始录制用户演唱的用户千声; 当第一 歌曲录制结束后, 电子设备可根据原唱千声的特征信息修正已录制的用户千声中出现的演唱 问题, 该演唱问题包括换气问题或颤音问题中的一项或多项。 也就是说, 电子设备可以在用 户录制的歌曲中修复换气、 颤音等演唱技巧问题。 这样, 没有经过专业的声乐训练的普通用 户也可以录制出听感质量较好的歌声, 从而提高用户录制歌曲的录制效果和使用体验。
在一种可能的实现方式中, 上述演唱问题为换气问题; 其中, 电子设备根据原唱千声的 特征信息修正用户千声中出现的演唱问题, 包括: 电子设备从用户千声中提取用户千声中的 换气声; 电子设备降低用户千声中换气声的能量和频响, 从而弱化用户千声中出现的换气问 题, 以提高录制歌曲的听感质量 ^
在一种可能的实现方式中, 上述演唱问题为颤音问题, 其中, 电子设备根据原唱千声的 特征信息修正用户千声中出现的演唱问题, 包括: 电子设备从用户千声中提取用户千声的颤 音特征; 若用户千声的颤音特征与原唱千声的特征信息中的颤音特征不匹配, 例如, 用户千 声中的颤音位置、 颤动幅度、 颤动频率或平稳趋势项中的一项与原唱千声不匹配, 则电子设 备可使用原唱千声中的颤音特征修正用户千声中的颤音特征, 以修正该用户千声中出现的颤 音问题。 例如, 电子设备可使用原唱千声中对应颤音的颤动幅度、 颤动频率和平稳趋势项用 户千声中合成新的颤音, 从而提高用户千声中出现的颤音音效。
当然, 上述演唱问题也可以为音准问题, 电子设备可以对用户千声中音调过高的地方进 行降调处理, 对用户千声中音调过低的地方进行升调处理, 使得用户千声的音调与原唱千声 的音调的相同或接近。
在一种可能的实现方式中, 当第一歌曲录制结束后, 上述方法还包括: 电子设备显示第 二界面, 第二界面中包括修音开关; 其中, 电子设备根据原唱千声的特征信息修正用户千声 中出现的演唱问题, 包括: 若检测到用户打开上述修音开关, 则电子设备可根据原唱千声的 特征信息修正用户千声中出现的演唱问题。 在一种可能的实现方式中, 当第一歌曲录制结束后, 上述方法还包括: 电子设备显示第 二界面, 第二界面中包括修音选项, 该修音选项中包括修正音准的选项、 修正换气声的选项 以及修正颤音的选项; 其中, 电子设备根据原唱千声的特征信息修正用户千声中出现的演唱 问题, 包括: 响应于用户选中的修音选项, 电子设备根据原唱千声的特征信息修正用户千声 中与该修音选项对应的演唱问题。
在一种可能的实现方式中, 当第一歌曲录制结束后, 上述方法还包括: 电子设备显示第 二界面, 第二界面中包括单句精修的选项; 响应于用户选择单句精修的选项, 电子设备可显 示第三界面, 第三界面中包括第一歌曲中的 N ( N > 1 ) 句歌词; 此时, 电子设备根据原唱千 声的特征信息修正用户千声中出现的演唱问题, 具体包括: 若用户在上述 N句歌词中选中第 一歌词, 则电子设备可根据原唱千声中演唱第一歌词的特征信息, 修正与第一歌词对应的用 户千声中出现的演唱问题。这样, 用户可以逐句修正本次录制的用户千声中出现的演唱问题, 从而将用户录制的歌曲修正为用户更满意的演唱效果。
另外, 电子设备在显示上述第三界面时, 还可以在第三界面的 N句歌词中标记出现演唱 问题的位置。 这样, 用户可以直观的看到自己在演唱每一句歌词时出现的具体问题。
第三方面, 本申请提供一种电子设备, 包括: 触摸屏、 一个或多个处理器、 一个或多个 存储器、 一个或多个麦克风以及一个或多个计算机程序; 其中, 处理器与触摸屏、 存储器和 麦克风均耦合, 上述一个或多个计算机程序被存储在存储器中, 当电子设备运行时, 处理器 执行存储器存储的一个或多个计算机程序, 以使电子设备执行上述任一项所述的歌曲录制方 法或修音方法。
第四方面, 本申请提供一种计算机存储介质, 包括计算机指令, 当计算机指令在电子设 备上运行时, 使得电子设备执行如第一方面中任一项所述的歌曲录制方法或修音方法。
第五方面, 本申请提供一种计算机程序产品, 当计算机程序产品在电子设备上运行时, 使得电子设备执行如第一方面中任一项所述的歌曲录制方法或修音方法。
第六方面,本申请提供一种图形用户界面 ( GUI ), 所述图形用户界面存储在电子设备中, 所述电子设备包括触摸屏、 存储器、 处理器, 所述处理器用于执行存储在所述存储器中的一 个或多个计算机程序, 其特征在于, 所述图形用户界面包括: 显示在触摸屏上的第一 GUI, 第一 GUI中包括用于录制第一歌曲的录制按钮; 响应于针对该录制按钮的触摸事件, 在触摸 屏上显示第二 GUI, 第二 GUI为录制所述第一歌曲的录制界面, 第二 GUI中包括指导用户演 唱第一歌曲的指导信息, 该指导信息可包括换气提示和 /或颤音提示。
示例性的,上述第二 GUI中还可以包括对用户演唱时的换气技巧和 /或颤音技巧的评估结 果。
示例性的, 在触摸屏上显示上述第二 GUI之后, 还包括: 当第一歌曲录制结束后, 在触 摸屏上显示第三 GUI, 第三 GUI为第一歌曲的修音界面, 第三 GUI中包括修音开关; 当修音 开关被用户打开时, 电子设备可对用户千声中出现的换气问题和成颤音问题进行修正。
可以理解地, 上述提供的第三方面所述的电子设备、 第四方面所述的计算机存储介质、 第五方面所述的计算机程序产品以及第六方面所述的 GUI均用于执行上文所提供的对应的方 法, 因此, 其所能达到的有益效果可参考上文所提供的对应的方法中的有益效果, 此处不再 赘述。
附图说明
图 1为本申请实施例提供的一种电子设备的结构示意图一; 图 2为本申请实施例提供的一种电子设备中操作系统的架构示意图;
图 3为本申请实施例提供的一种歌曲录制方法和修音方法的流程示意图;
图 4为本申请实施例提供的一种歌曲录制方法的场景示意图一;
图 5为本申请实施例提供的一种歌曲录制方法的场景示意图二;
图 6为本申请实施例提供的一种歌曲录制方法的场景示意图三;
图 7为本申请实施例提供的一种歌曲录制方法的场景示意图四;
图 8为本申请实施例提供的一种歌曲录制方法的场景示意图五;
图 9为本申请实施例提供的一种歌曲录制方法的场景示意图六;
图 10为本申请实施例提供的一种歌曲录制方法的场景示意图七;
图 11为本申请实施例提供的一种歌曲录制方法的场景示意图八;
图 12为本申请实施例提供的一种歌曲录制方法的场景示意图九;
图 13为本申请实施例提供的一种修音方法的场景示意图一;
图 14为本申请实施例提供的一种修音方法的场景示意图二;
图 15为本申请实施例提供的一种修音方法的场景示意图三;
图 16为本申请实施例提供的一种修音方法的场景示意图四;
图 17为本申请实施例提供的一种电子设备的结构示意图二。
具体实施方式
下面将结合附图对本实施例的实施方式进行详细描述。
示例性的, 本申请实施例提供的一种歌曲录制方法和修音方法可应用于手机、平板电脑、 笔记本电脑、 超级移动个人计算机 ( ultra-mobile personal computer, UMPC )、 手持计算机、 上网本、 个人数字助理 ( personal digital assistant, PDA )、 可穿戴电子设备、 虚拟现实设备等 电子设备, 本申请实施例对此不做任何限制。
示例性的, 图 1示出了电子设备 100的结构示意图。
电子设备 100可以包括处理器 110, 外部存储器接口 120, 内部存储器 121, 通用串行总 线 (universal serial bus, USB)接口 130, 充电管理模块 140, 电源管理模块 141, 电池 142, 天 线 1, 天线 2, 移动通信模块 150, 无线通信模块 160, 音频模块 170, 扬声器 170A, 受话器 170B , 麦克风 170C, 耳机接口 170D, 传感器模块 180, 按键 190, 马达 191 , 指示器 192, 摄像头 193, 显示屏 194, 以及用户标识模块 (subscriber identification module, SIM)卡接口 195 等。其中传感器模块 180可以包括压力传感器 180A, 陀螺仪传感器 180B , 气压传感器 180C, 磁传感器 180D, 加速度传感器 180E, 距离传感器 180F, 接近光传感器 180G, 指纹传感器 180H, 温度传感器 180J, 触摸传感器 180K, 环境光传感器 180L, 骨传导传感器 180M等。
可以理解的是, 本发明实施例示意的结构并不构成对电子设备 100的具体限定。 在本申 请另一些实施例中, 电子设备 100可以包括比图示更多或更少的部件, 或者组合某些部件, 或者拆分某些部件, 或者不同的部件布置。 图示的部件可以以硬件, 软件或软件和硬件的组 合实现。
处理器 110 可以包括一个或多个处理单元, 例如: 处理器 110 可以包括应用处理器 (application processor, AP) , 调制解调处理器, 图形处理器 (graphics processing unit, GPU) , 图像信号处理器 (image signal processor, ISP) ,控制器,视频编解码器,数字信号处理器 (digital signal processor, DSP) , 基带处理器, 和 /或神经网络处理器 (neural-network processing unit, NPU)等。 其中, 不同的处理单元可以是独立的器件, 也可以集成在一个或多个处理器中。 控制器可以根据指令操作码和时序信号, 产生操作控制信号, 完成取指令和执行指令的 控制。
处理器 110中还可以设置存储器, 用于存储指令和数据。 在一些实施例中, 处理器 110 中的存储器为高速缓冲存储器。 该存储器可以保存处理器 110刚用过或循环使用的指令或数 据。 如果处理器 110需要再次使用该指令或数据, 可从所述存储器中直接调用。 避免了重复 存取, 减少了处理器 110的等待时间, 因而提高了系统的效率。
在一些实施例中, 处理器 110 可以包括一个或多个接口。 接口可以包括集成电路 (inter- integrated circuit, I2C)接口, 集成电路内置音频 (inter-integrated circuit sound, I2S)接口, 脉冲编码调制 (pulse code modulation, PCM)接口,通用异步收发传输器 (universal asynchronous receiver/transmitter , UART)接口, 移动产业处理器接口 (mobile industry processor interface, MIPI) , 通用输入输出 (general-purpose input/output , GPIO)接口, 用户标识模块 (subscriber identity module, SIM)接口, 和 /或通用串行总线 (universal serial bus, USB)接口等。
I2C接口是一种双向同步串行总线, 包括一根串行数据线 (serial data line, SDA)和一根串 行时钟线 (derail clock line, SCL)。 在一些实施例中, 处理器 110可以包含多组 I2C总线。 处 理器 110可以通过不同的 I2C总线接口分别耦合触摸传感器 180K, 充电器, 闪光灯, 摄像头 193等。 例如: 处理器 110可以通过 I2C接口耦合触摸传感器 180K, 使处理器 110与触摸传 感器 180K通过 I2C总线接口通信, 实现电子设备 100的触摸功能。
I2S接口可以用于音频通信。 在一些实施例中, 处理器 110可以包含多组 I2S总线。 处理 器 110可以通过 I2S总线与音频模块 170耦合, 实现处理器 110与音频模块 170之间的通信。 在一些实施例中, 音频模块 170可以通过 I2S接口向无线通信模块 160传递音频信号, 实现 通过蓝牙耳机接听电话的功能。
PCM接口也可以用于音频通信, 将模拟信号抽样, 量化和编码。 在一些实施例中, 音频 模块 170与无线通信模块 160可以通过 PCM总线接口耦合。 在一些实施例中, 音频模块 170 也可以通过 PCM接口向无线通信模块 160传递音频信号,实现通过蓝牙耳机接听电话的功能。 所述 I2S接口和所述 PCM接口都可以用于音频通信。
UART接口是一种通用串行数据总线, 用于异步通信。 该总线可以为双向通信总线。 它 将要传输的数据在串行通信与并行通信之间转换。 在一些实施例中, UART接口通常被用于 连接处理器 110与无线通信模块 160。例如: 处理器 110通过 UART接口与无线通信模块 160 中的蓝牙模块通信, 实现蓝牙功能。 在一些实施例中, 音频模块 170可以通过 UART接口向 无线通信模块 160传递音频信号, 实现通过蓝牙耳机播放音乐的功能。
MIPI接口可以被用于连接处理器 110与显示屏 194, 摄像头 193等外围器件。 MIPI接口 包括摄像头串行接口 (camera serial interface, CSI) , 显示屏串行接口 (display serial interface, DSI)等。 在一些实施例中, 处理器 110和摄像头 193通过 CSI接口通信, 实现电子设备 100 的拍摄功能。 处理器 110和显示屏 194通过 DSI接口通信, 实现电子设备 100的显示功能。
GPI0接口可以通过软件配置。 GPI0接口可以被配置为控制信号, 也可被配置为数据信 号。 在一些实施例中, GPI0接口可以用于连接处理器 110与摄像头 193, 显示屏 194, 无线 通信模块 160, 音频模块 170, 传感器模块 180等。 GPI0接口还可以被配置为 I2C接口, I2S 接口, UART接口, MIPI接口等。
可以理解的是, 本发明实施例示意的各模块间的接口连接关系, 只是示意性说明, 并不 构成对电子设备 100的结构限定。 在本申请另一些实施例中, 电子设备 100也可以采用上述 实施例中不同的接口连接方式, 或多种接口连接方式的组合。
充电管理模块 140用于从充电器接收充电输入。 其中, 充电器可以是无线充电器, 也可 以是有线充电器。 在一些有线充电的实施例中, 充电管理模块 140可以通过 USB接口 130接 收有线充电器的充电输入。 在一些无线充电的实施例中, 充电管理模块 140可以通过电子设 备 100的无线充电线圈接收无线充电输入。 充电管理模块 140为电池 142充电的同时, 还可 以通过电源管理模块 141为电子设备供电。
电源管理模块 141用于连接电池 142, 充电管理模块 140与处理器 110。 电源管理模块 141接收电池 142和 /或充电管理模块 140的输入, 为处理器 110, 内部存储器 121 , 显示屏 194,摄像头 193 ,和无线通信模块 160等供电。 电源管理模块 141还可以用于监测电池容量, 电池循环次数, 电池健康状态 (漏电, 阻抗)等参数。 在其他一些实施例中, 电源管理模块 141 也可以设置于处理器 110中。 在另一些实施例中, 电源管理模块 141和充电管理模块 140也 可以设置于同一个器件中。
电子设备 100的无线通信功能可以通过天线 1, 天线 2, 移动通信模块 150, 无线通信模 块 160, 调制解调处理器以及基带处理器等实现。
天线 1和天线 2用于发射和接收电磁波信号。 电子设备 100中的每个天线可用于覆盖单 个或多个通信频带。 不同的天线还可以复用, 以提高天线的利用率。 例如: 可以将天线 1复 用为无线局域网的分集天线。 在另外一些实施例中, 天线可以和调谐开关结合使用。
移动通信模块 150可以提供应用在电子设备 100上的包括 2G/3G/4G/5G等无线通信的解 决方案。 移动通信模块 150可以包括一个或多个滤波器, 开关, 功率放大器, 低噪声放大器 (low noise amplifier, LNA)等。 移动通信模块 150可以由天线 1接收电磁波, 并对接收的电磁 波进行滤波, 放大等处理, 传送至调制解调处理器进行解调。 移动通信模块 150还可以对经 调制解调处理器调制后的信号放大, 经天线 1转为电磁波辐射出去。 在一些实施例中, 移动 通信模块 150的至少部分功能模块可以被设置于处理器 110中。 在一些实施例中, 移动通信 模块 150的至少部分功能模块可以与处理器 110的至少部分模块被设置在同一个器件中。
调制解调处理器可以包括调制器和解调器。 其中, 调制器用于将待发送的低频基带信号 调制成中高频信号。 解调器用于将接收的电磁波信号解调为低频基带信号。 随后解调器将解 调得到的低频基带信号传送至基带处理器处理。 低频基带信号经基带处理器处理后, 被传递 给应用处理器。 应用处理器通过音频设备 (不限于扬声器 170A, 受话器 170B等)输出声音信 号, 或通过显示屏 194显示图像或视频。 在一些实施例中, 调制解调处理器可以是独立的器 件。 在另一些实施例中, 调制解调处理器可以独立于处理器 110, 与移动通信模块 150或其 他功能模块设置在同一个器件中。
无线通信模块 160可以提供应用在电子设备 100上的包括无线局域网 (wireless local area networks , WLAN) (如无线保真 (wireless fidelity , Wi-Fi)网络), 蓝牙 (Bluetooth, BT), 全球导 航卫星系统 (global navigation satellite system, GNSS), 调频 (frequency modulation, FM) , 近距 离无线通信技术 (near field communication, NFC) , 红外技术 (infrared, IR)等无线通信的解决 方案。 无线通信模块 160可以是集成一个或多个通信处理模块的一个或多个器件。 无线通信 模块 160经由天线 2接收电磁波, 将电磁波信号调频以及滤波处理, 将处理后的信号发送到 处理器 110。 无线通信模块 160还可以从处理器 110接收待发送的信号, 对其进行调频, 放 大, 经天线 2转为电磁波辐射出去。
在一些实施例中, 电子设备 100的天线 1和移动通信模块 150耦合, 天线 2和无线通信 模块 160耦合, 使得电子设备 100可以通过无线通信技术与网络以及其他设备通信。 所述无 线通信技术可以包括全球移动通讯系统 (global system for mobile communications , GSM) , 通 用分组无线服务 (general packet radio service, GPRS) , 码分多址接入 (code division multiple access, CDMA) , 宽带码分多址 (wideband code division multiple access, WCDMA) , 时分码分 多址 (time-division code division multiple access, TD-SCDMA) , 长期演进 (long term evolution, LTE) , BT, GNSS , WLAN, NFC, FM, 和 /或 IR技术等。 所述 GNSS可以包括全球卫星定 位系统 (global positioning system , GPS) , 全球导航卫星系统 (global navigation satellite system, GLONASS) , 北斗卫星导航系统 (beidou navigation satellite system, BDS) , 准天顶卫星系统 (quasi-zenith satellite system, QZSS)和 /或星基增强系统 (satellite based augmentation systems, SBAS)。
电子设备 100通过 GRJ, 显示屏 194, 以及应用处理器等实现显示功能。 GPU为图像处 理的微处理器, 连接显示屏 194和应用处理器。 GPU用于执行数学和几何计算, 用于图形渲 染。 处理器 110可包括一个或多个 GPU, 其执行程序指令以生成或改变显示信息。
显示屏 194用于显示图像, 视频等。 显示屏 194包括显示面板。 显示面板可以采用液晶 显示屏 (liquid crystal display, LCD) , 有机发光二极管 (organic light-emitting diode, OLED), 有 源矩阵有机发光二极体或主动矩阵有机发光二极体 (active-matrix organic light emitting diode 的, AMOLED) , 柔性发光二极管 (flex light-emitting diode , FLED) , Miniled , MicroLed , Micro-oLed, 量子点发光二极管 (quantum dot light emitting diodes, QLED)等。 在一些实施例 中, 电子设备 100可以包括 1个或 N个显示屏 194, N为大于 1的正整数。
电子设备 100可以通过 ISP, 摄像头 193 , 视频编解码器, GPU, 显示屏 194以及应用处 理器等实现拍摄功能。
ISP用于处理摄像头 193反馈的数据。 例如, 拍照时, 打开快门, 光线通过镜头被传递 到摄像头感光元件上, 光信号转换为电信号,摄像头感光元件将所述电信号传递给 ISP处理, 转化为肉眼可见的图像。 ISP还可以对图像的噪点, 亮度, 肤色进行算法优化。 ISP还可以对 拍摄场景的曝光, 色温等参数优化。 在一些实施例中, ISP可以设置在摄像头 193中。
摄像头 193用于捕获静态图像或视频。 物体通过镜头生成光学图像投射到感光元件。 感 光元件可以是电荷接合器件 (charge coupled device , CCD)或互补金属氧化物半导体 (complementary metal-oxide-semiconductor, CMOS)光电晶体管。 感光元件把光信号转换成电 信号, 之后将电信号传递给 ISP转换成数字图像信号。 ISP将数字图像信号输出到 DSP加工 处理。 DSP将数字图像信号转换成标准的 RGB , YUV等格式的图像信号。 在一些实施例中, 电子设备 100可以包括 1个或 N个摄像头 193 , N为大于 1的正整数。
数字信号处理器用于处理数字信号, 除了可以处理数字图像信号, 还可以处理其他数字 信号。 例如, 当电子设备 100在频点选择时, 数字信号处理器用于对频点能量进行傅里叶变 换等。
视频编解码器用于对数字视频压缩或解压缩。 电子设备 100可以支持一种或多种视频编 解码器。 这样, 电子设备 100可以播放或录制多种编码格式的视频, 例如: 动态图像专家组 (moving picture experts group , MPEG)1 , MPEG2 , MPEG3 , MPEG4等。
NPU为神经网络 (neural-network , NN)计算处理器, 通过借鉴生物神经网络结构, 例如 借鉴人脑神经元之间传递模式, 对输入信息快速处理, 还可以不断的自学习。 通过 NPU可以 实现电子设备 100的智能认知等应用, 例如: 图像识别, 人脸识别, 语音识别, 文本理解等。 外部存储器接口 120可以用于连接外部存储卡, 例如 Micro SD卡, 实现扩展电子设备 100的存储能力。 外部存储卡通过外部存储器接口 120与处理器 110通信, 实现数据存储功 能。 例如将音乐, 视频等文件保存在外部存储卡中。
内部存储器 121可以用于存储一个或多个计算机程序, 该一个或多个计算机程序包括指 令。 处理器 110可以通过运行存储在内部存储器 121的上述指令, 从而使得电子设备 100执 行本申请一些实施例中所提供的联系人智能推荐的方法, 以及各种功能应用和数据处理等。 内部存储器 121可以包括存储程序区和存储数据区。 其中, 存储程序区可存储操作系统; 该 存储程序区还可以存储一个或多个应用程序(比如图库、 联系人等)等。 存储数据区可存储电 子设备 101使用过程中所创建的数据(比如照片, 联系人等)等。 此外, 内部存储器 121可以 包括高速随机存取存储器, 还可以包括非易失性存储器, 例如一个或多个磁盘存储器件, 闪 存器件, 通用闪存存储器(universal flash storage, UFS)等。 在另一些实施例中, 处理器 110通 过运行存储在内部存储器 121的指令, 和成存储在设置于处理器中的存储器的指令, 来使得 电子设备 100执行本申请实施例中所提供的智能推荐号码的方法, 以及各种功能应用和数据 处理。
电子设备 100可以通过音频模块 170, 扬声器 170A, 受话器 170B , 麦克风 170C , 耳机 接口 170D, 以及应用处理器等实现音频功能。 例如音乐播放, 录音等。
音频模块 170用于将数字音频信息转换成模拟音频信号输出, 也用于将模拟音频输入转 换为数字音频信号。 音频模块 170还可以用于对音频信号编码和解码。 在一些实施例中, 音 频模块 170可以设置于处理器 110中, 或将音频模块 170的部分功能模块设置于处理器 110 中。
扬声器 170A, 也称 “喇叭”, 用于将音频电信号转换为声音信号。 电子设备 100可以通 过扬声器 170A收听音乐, 或收听免提通话。
受话器 170B , 也称 “听筒”, 用于将音频电信号转换成声音信号。 当电子设备 100接听 电话或语音信息时, 可以通过将受话器 170B靠近人耳接听语音。
麦克风 170C, 也称 “话筒”, “传声器”, 用于将声音信号转换为电信号。 当拨打电话或 发送语音信息时,用户可以通过人嘴靠近麦克风 170C发声,将声音信号输入到麦克风 170C。 电子设备 100可以设置一个或多个麦克风 170C。 在另一些实施例中, 电子设备 100可以设置 两个麦克风 170C, 除了采集声音信号, 还可以实现降噪功能。 在另一些实施例中, 电子设备 100还可以设置三个, 四个或更多麦克风 170C, 实现采集声音信号, 降噪, 还可以识别声音 来源, 实现定向录音功能等。
在本申请实施例中, 电子设备 100运行 K歌类应用时, 可使用麦克风 170C采集用户演 唱歌曲时的用户千声, 即无音乐的纯人声。 并且, 音频模块 170可使用相应的修音算法对采 集到的用户千声进行修正。
例如, 如果用户录制的歌曲 A的用户千声中出现较为明显或紊乱的换气声, 则音频模块 170可对用户千声中相应位置的换气声进行弱化处理, 从而提高歌声的听感质量。
又例如, 如果歌曲 A的原唱千声中出现了颤音, 则音频模块 170可检测用户录制歌曲 A 的用户千声中是否也在相应位置使用了颤音。如果用户使用的颤音效果不佳,则音频模块 170 可对用户千声中相应位置的颤音进行修正, 从而提高歌声的听感质量。
又例如, 音频模块 170可通过对比歌曲 A的用户千声与原唱千声识别用户千声中出现跑 调的位置, 进而, 音频模块 170可对用户千声中出现跑调的位置进行相应的升调或降调, 从 而修复用户千声的音准问题。 其中, 电子设备 100对麦克风 170C采集到的用户千声进行修 正的具体方法将在后续实施例中详细阐述, 故此处不予赘述。
可以看出, 在本申请实施例中, 电子设备除了可以对在用户录制的歌曲进行音准的校准, 还可以在用户录制的歌曲中修复换气、 颤音等演唱技巧问题。 这样, 没有经过专业的声乐训 练的普通用户也可以录制出听感质量较好的歌声, 从而提高用户使用 K歌类应用时的录制效 果和使用体验。
耳机接口 170D用于连接有线耳机。耳机接口 170D可以是 USB接口 130,也可以是 3.5mm 的开放移动电子设备平台 (open mobile terminal platform, OMTP)标准接口, 美国蜂寫电信工 业协会 (cellular telecommunications industry association of the USA, CTIA)标准接口。
传感器 180可以包括压力传感器 180A, 陀螺仪传感器 180B , 气压传感器 180C, 磁传感 器 180D, 加速度传感器 180E, 距离传感器 180F, 接近光传感器 180G, 指纹传感器 180H, 温度传感器 180J, 触摸传感器 180K, 环境光传感器 180L, 骨传导传感器 180M等。
按键 190包括开机键, 音量键等。 按键 190可以是机械按键。 也可以是触摸式按键。 电 子设备 100可以接收按键输入, 产生与电子设备 100的用户设置以及功能控制有关的键信号 输入。
马达 191可以产生振动提示。 马达 191可以用于来电振动提示, 也可以用于触摸振动反 馈。 例如, 作用于不同应用 (例如拍照, 音频播放等)的触摸操作, 可以对应不同的振动反馈 效果。 作用于显示屏 194不同区域的触摸操作, 马达 191也可对应不同的振动反馈效果。 不 同的应用场景 (例如: 时间提醒, 接收信息, 闹钟, 游戏等)也可以对应不同的振动反馈效果。 触摸振动反馈效果还可以支持自定义。
指示器 192可以是指示灯, 可以用于指示充电状态, 电量变化, 也可以用于指示消息, 未接来电, 通知等。
SIM卡接口 195用于连接 SIM卡。 SIM卡可以通过插入 SIM卡接口 195,或从 SIM卡接 口 195拔出, 实现和电子设备 100的接触和分离。 电子设备 100可以支持 1个或 N个 SIM卡 接口, N为大于 1的正整数。 SIM卡接口 195可以支持 Nano SIM卡, Micro SIM卡, SIM卡 等。 同一个 SIM卡接口 195可以同时插入多张卡。所述多张卡的类型可以相同,也可以不同。 SIM卡接口 195也可以兼容不同类型的 SIM卡。 SIM卡接口 195也可以兼容外部存储卡。 电 子设备 100通过 SIM卡和网络交互, 实现通话以及数据通信等功能。 在一些实施例中, 电子 设备 100采用 eSIM, 即: 嵌入式 SIM卡。 eSIM卡可以嵌在电子设备 100中, 不能和电子设 备 100分离。
上述电子设备 100的软件系统可以采用分层架构, 事件驱动架构, 微核架构, 微服务架 构, 或云架构。 本申请实施例以分层架构的 Android系统为例, 示例性说明电子设备 100的 软件结构。
图 2是本申请实施例的电子设备 100的软件结构框图。
分层架构将软件分成若千个层, 每一层都有清晰的角色和分工。 层与层之间通过软件接 口通信。 在一些实施例中, 将 Android 系统分为四层, 从上至下分别为应用程序层, 应用程 序框架层, 安卓运行时 (Android runtime)和系统库, 以及内核层。
应用程序层可以包括一系列应用程序包。
如图 2所示, 应用程序包可以包括通话, 联系人, 相机, 图库, 日历, 地图, 导航, 蓝 牙, 音乐, 视频, 短信息等 APP (应用, application )» 在本申请实施例中, 还可以在应用程序层中安装 K歌类应用, 例如唱吧 APP、 全民 K歌 APP等。 用户可使用 K歌类应用录制歌曲, 将自己的歌声融入在 K歌类应用提供的伴奏中。
应用程序框架层为应用程序层的应用程序提供应用编程接口 (application programming interface, API)和编程框架。 应用程序框架层包括一些预先定义的函数。
如图 2所示, 应用程序框架层可以包括窗口管理器, 内容提供器, 视图系统, 资源管理 器, 通知管理器等。
其中, 窗口管理器用于管理窗口程序。 窗口管理器可以获取显示屏大小, 判断是否有状 态栏, 锁定屏幕, 截取屏幕等。 内容提供器用来存放和获取数据, 并使这些数据可以被应用 程序访问。 所述数据可以包括视频, 图像, 音频, 拨打和接听的电话, 浏览历史和书签, 电 话簿等。 视图系统包括可视控件, 例如显示文字的控件, 显示图片的控件等。 视图系统可用 于构建应用程序。 显示界面可以由一个或多个视图组成的。 例如, 包括短信通知图标的显示 界面, 可以包括显示文字的视图以及显示图片的视图。 资源管理器为应用程序提供各种资源, 比如本地化字符串, 图标, 图片, 布局文件, 视频文件等等。 通知管理器使应用程序可以在 状态栏中显示通知信息, 可以用于传达告知类型的消息, 可以短暂停留后自动消失, 无需用 户交互。 比如通知管理器被用于告知下载完成, 消息提醒等。 通知管理器还可以是以图表或 者滚动条文本形式出现在系统顶部状态栏的通知, 例如后台运行的应用程序的通知, 还可以 是以对话窗口形式出现在屏幕上的通知。 例如在状态栏提示文本信息, 发出提示音, 电子设 备振动, 指示灯闪烁等。
系统库可以包括多个功能模块。 例如: 表面管理器 (surface manager) , 媒体库 (Media Libraries), 三维图形处理库 (例如: OpenGL ES), 2D图形引擎 (例如: SGL)等。
表面管理器用于对显示子系统进行管理,并且为多个应用程序提供了 2D和 3D图层的融 合。 媒体库支持多种常用的音频, 视频格式回放和录制, 以及静态图像文件等。 媒体库可以 支持多种音视频编码格式, 例如: MPEG4, H.264, MP3 , AAC, AMR, JPG, PNG等。 三维 图形处理库用于实现三维图形绘图, 图像渣染, 合成, 和图层处理等。 2D图形引擎是 2D绘 图的绘图引擎。
在本申请实施例中, 仍如图 2所示, 可将本申请实施例提供的相关修音算法以算法动态 库 ( dynamic link library, DLL ) 的形式设置在系统库中。 例如, 可将修音算法的算法动态库 存储在 system/lib 目录下。 在应用程序层运行 K歌类应用时, K歌类应用可调用修音算法的 算法动态库对用户千声中音准、 颤音或换气声等问题进行修正, 实现修音功能。
Android Runtime包括核心库和虚拟机。 Android runtime负责安卓系统的调度和管理。 核心库包含两部分:一部分是 java语言需要调用的功能函数,另一部分是安卓的核心库。 应用程序层和应用程序框架层运行在虚拟机中。 虚拟机将应用程序层和应用程序框架层 的 java文件执行为二进制文件。虚拟机用于执行对象生命周期的管理, 堆栈管理, 线程管理, 安全和异常的管理, 以及垃圾回收等功能。
内核层是硬件和软件之间的层。 内核层至少包含显示驱动, 摄像头驱动, 音频驱动, 传 感器驱动等, 本申请实施例对此不彳故任何限制。
以下将以手机作为电子设备举例, 结合附图详细阐述本申请实施例中提供的一种歌曲录 制方法以及修音方法。 如图 3所示, 该方法包括下述步骤 S301-S306。
S301、 在运行 K歌类应用时, 手机接收用户选择录制歌曲 A的操作。
其中, K歌类应用是指可向用户提供点歌、 歌曲录制以及修音功能的应用软件, 例如, 该 K歌类应用可以为全民 K歌 APP、 唱吧 APP等, 本申请实施例对此不做任何限制。
如果手机检测到用户打开 K歌类应用的操作, 如图 4所示, 手机可开始运行该 K歌类应 用, 并显示 K歌类应用的第一界面 401。 用户可在第一界面 401查找自己喜欢的歌曲进行录 制。 或者, 用户可在搜索栏 402中手动搜索并选择需要录制的歌曲。 当然, 用户也可以在 K 歌类应用的其他界面中选择本次需要录制的歌曲, 本申请实施例对此不做任何限制。
仍如图 4所示,如果用户想要录制歌曲《红豆》,用户可点击第一界面 401中歌曲《红豆》 的录制按钮 403, 此时, 手机可接收到用户选择录制歌曲 《红豆》 的操作。 响应于用户选择 录制歌曲 《红豆》 的操作, 手机可继续执行下述步骤 S302-S306完成歌曲的录制和修音等过 程。
5302、 手机获取歌曲 A的伴奏和原唱千声的特征信息。
仍以用户录制歌曲 《红豆》举例, 检测到用户点击歌曲 《红豆》 的录制按钮 403后, 手 机可从服务器下载歌曲 《红豆》 的伴奏。 当然, 如果手机本次存储有歌曲 《红豆》 的伴奏, 则手机可从本地的存储器中获取该伴奏。 一般, 歌曲中除人声外的部分可称为伴奏, 有的伴 奏还可包括和声以及背景音等 ^
在本申请实施例中, 检测到用户点击歌曲 《红豆》 的录制按钮 403后, 手机还可从服务 器下载歌曲 《红豆》 的原唱千声的特征信息。 其中, 该特征信息可反映出原唱 (例如王菲) 在演唱歌曲 《红豆》 时的演唱特征。 例如, 该特征信息可以包括原唱千声的基频信息、 换气 位置和颤音特征等一项或多项。 其中, 颤音特征可包括颤音位置、 颤音的基频平稳信息、 颤 音频率以及颤音幅度等一项或多项。
其中, 基频信息可反映出声音的基础音调, 如图 5所示, 服务器可预先从歌曲 《红豆》 的原唱千声中以帧为单位提取基频-时间的变化曲线 f(t) , 该变化曲线 f(t)即为歌曲《红豆》的 基频信息。 仍如图 5所示, 服务器还可在原唱千声中检测原唱换气的位置, 并将原唱换气的 时刻作为一项特征信息。 类似的, 仍如图 5所示, 服务器还可在原唱千声中检测出现颤音的 位置, 并将出现颤音的时刻作为一项特征信息。
通常, 颤音表现为周期性的音高变化。 具有颤音特征的基频曲线中可以提取到颤动幅度 A(t)、 颤动频率①⑴和平稳趋势项 V⑴这三个随时间变化的特征。 以歌曲《红豆》的第三句原 唱千声出现颤音 1为例,颤音 1对应的基频曲线 f3⑴二 ⑴+八⑴ cos (①⑴), t 6 [8s, 10s]。其中, V⑴为颤音 1的基频平稳信息, ①⑴为颤音 1的颤音频率, A⑴为颤音 1的颤音幅度。 服务器 可将歌曲 《红豆》 中每一处颤音的基频平稳信息、 颤音频率以及颤音幅度作为特征信息发送 给手机。
示例性的, 每次向服务器中的歌曲库中添加新歌时, 服务器均可根据其原唱千声提取到 对应的特征信息, 并将提取到的特征信息保存在服务器中。 后续, 如果用户在手机中选择录 制某一歌曲, 则手机除了可从服务器中获取该歌曲的伴奏外, 还可以从服务器获取与该歌曲 对应的原唱千声的特征信息。 获取到的原唱千声的特征信息可用于指导用户录制歌曲, 还可 作为修音的依据对用户录制的歌曲进行修音。
5303、手机在显示 K歌类应用的录制界面时,播放歌曲 A的伴奏并录制用户演唱的用户 千声。
仍以用户录制歌曲 《红豆》举例, 手机获取到歌曲 《红豆》 的伴奏后, 手机可以开始播 放歌曲 《红豆》 的伴奏。 同时, 手机可打开麦克风开始录制用户演唱歌曲 《红豆》 的用户千 声。 如图 6所示, 在录制歌曲 《红豆》 的过程中, 手机可显示歌曲 《红豆》 的录制界面 601。 录制界面 601可以显示歌曲 《红豆》 的歌词 602, 歌词 602中可实时标记当前正在录制的歌 词位置, 以便用户按照歌词 602的提示录制歌曲。
5304、 在录制歌曲 A时, 手机根据原唱千声的特征信息提示用户调整音准、 换气或颤音 等演唱方式。
示例性的, 用户在使用 K歌类应用录制歌曲时, 手机可根据步骤 S302获取的特征信息 在显示歌词 602时指导用户演唱歌曲, 例如指导用户的音准、 换气或颤音等演唱方式。
仍以用户录制歌曲 《红豆》举例, 手机获取的原唱千声的特征信息中包括换气位置和颤 音位置等信息。 那么, 如图 7所示, 在实时显示歌曲 《红豆》 的歌词 602时, 手机可根据原 唱千声的特征信息中记录的换气位置在歌词 602的相应位置标记换气位置 701(即换气提示 X 这样, 用户在录制歌曲 《红豆》 时, 可以根据歌词 602中提示的换气位置 701进行换气, 从 而帮助用户控制演唱时的气息变化, 从而减少录制的歌曲中出现较为明显和紊乱的换气声, 以提高录制歌曲的听感质量。
又例如, 仍如图 7所示, 在实时显示歌曲 《红豆》 的歌词 602时, 手机还可根据原唱千 声的特征信息中记录的颤音位置在歌词 602的相应位置标记颤音位置 702(即颤音提示)。 这 样, 用户在录制歌曲 《红豆》 时, 可以根据歌词 602中提示的颤音位置 702使用颤音演唱, 从而提高用户演唱时的歌曲的层次感和演唱技巧, 以提高录制歌曲的听感质量 ^
在另一些实施例中, 手机除了在实时显示的歌词中指导用户演唱外, 还可以在录制界面 中设置演唱器, 手机可在演唱器中指导用户调整音准、 换气或颤音等演唱方式。
仍以用户录制歌曲 《红豆》举例, 如图 8所示, 手机在显示录制界面 601时可在录制界 面 601中设置演唱器 801。 演唱器 801 中可显示当前正在录制的一句歌词中每个字的音符。 例如, 手机可根据原唱千声的基频信息,在演唱器 801中实时显示每个音符的音调线 802(即 音准提示)。 如果当前音符的音调越高, 则音调线 802的位置越高。 又例如, 手机可根据原唱 千声的颤音位置, 在演唱器 801中显示颤音符 803(即颤音提示)。 颤音符 803可用波浪线或 特定颜色的线段表示。 颤音符 803 出现的位置与原唱千声中出现颤音的位置相同。 又例如, 如果本句歌词对应的原唱千声中出现了换气声, 则手机可根据原唱千声中的换气位置, 在演 唱器 801中显示换气符 804(即换气提示)。 换气符 804可用三角形、 圆形等指定形状的图形 表示。 换气符 804出现的位置与原唱千声中出现换气声的位置相同。
这样, 用户在录制歌曲时, 可根据演唱器 801 中实时显示的音调线调整人声的音准, 也 可以根据演唱器 801 中实时显示的颤音符在相应位置使用颤音演唱, 还可以根据演唱器 801 中实时显示的换气符在相应位置换气, 从而在录制过程中指导用户演唱时的音准、 颤音以及 换气等演唱技巧, 提高歌曲的录制质量。
可以理解的是, 在用户录制歌曲时, 除了在录制界面显示的歌词和演唱器中指导用户演 唱外, 本领域技术人员还可以根据实际经验或实际应用场景使用原唱千声的特征信息在用户 录歌时指导用户更好的演唱歌曲, 本申请实施例对此不做任何限制。
5305、 在录制歌曲 A时, 手机通过对比原唱千声的特征信息与用户千声中的特征信息, 对用户演唱进行评估。
仍以用户录制歌曲 《红豆》举例, 在录制歌曲 《红豆》 时, 手机还可以实时的根据原唱 千声的特征信息对用户当前的演唱进行评估。 示例性的, 手机可以歌词中的字、 词或句为粒 度对用户的演唱进行评估。 如图 9所示, 手机采集到用户演唱 “我会相信一切有尽头” 这一 句的用户千声后, 手机可提取本句用户千声的特征信息。 与原唱千声的特征信息类似的, 手 机可先提取本句用户千声的基频信息, 进而基于用户千声的基频信息检测用户的换气位置、 颤音位置、 颤音的基频平稳信息、 颤音频率以及颤音幅度等一项或多项特征信息。
那么, 手机通过对比原唱演唱 “我会相信一切有尽头” 这一句歌词时的特征信息 1, 与 用户演唱 “我会相信一切有尽头” 这一句歌词时的特征信息 2, 可确定出用户演唱本句的音 准、 颤音以及换气等演唱技巧的评估结果。
例如, 如果用户千声的基频信息与原唱千声的基频信息之间的差值小于阈值, 说明用户 演唱本句的音准与原唱匹配, 如图 10中的 (a)所示, 手机可在录制界面 601中显示 “音准 达人” 的提示 1001, 提示用户演唱本句时音准无误。 如果用户千声中颤音的位置与原唱千声 中颤音的位置相同, 且用户千声中该颤音的颤音特征(即颤音频率、 幅度和基频平稳信息) 与原唱千声中对应的颤音特征之间的差值小于阈值, 说明用户演唱本句时发出的颤音与原唱 匹配, 如图 10中的 (b)所示, 手机可在录制界面 601中显示 “颤音天后” 的提示 1002。 如 果用户千声中换气声的位置与原唱千声中换气声的位置相同, 说明用户演唱本句时气息的调 整与原唱匹配, 如图 10中的 (c)所示, 手机可在录制界面 601 中显示 “换气 get” 的提示 1003。
除了通过文字实时提示用户当前的评估结果外, 手机还可以通过颜色、 震动或特效等方 式实时向用户展示当前的评估结果。
例如, 图 11 中的 (a)所示, 如果用户演唱 “我会相信一切有尽头” 这一句歌词时使用 的颤音与原唱演唱本句时使用的颤音不相符, 则手机可在显示与该颤音对应的歌词 “尽头” 1101时, 将“尽头” 1101标记为与其它歌词不同的颜色或使用不同的字体等。 又例如, 图 11 中的(b)所示, 如果用户演唱本句时的换气位置与原唱演唱本句时的换气位置不相符, 则手 机可在显示与换气位置对应的歌词时标记换气错误的标识 1102。
又或者,手机还可以在录制界面 601的演唱器 801中向用户展示手机确定出的评估结果^ 例如, 如图 12中的 (a)所示, 用户演唱 “有” 字时, 手机不仅可在演唱器 801中实时显示 原唱演唱 “有” 字的音调线 1201, 还可同时显示本次用户演唱 “有” 字的音调线 1202。 如果 用户演唱 “有” 字的基频低于原唱演唱 “有” 字的基频, 说明用户演唱 “有” 字的音调低于 原唱演唱 “有” 字的音调, 则手机显示的音调线 1202低于音调线 1201。 相应的, 如果检测 到用户演唱 “有” 字的音调高于原唱演唱 “有” 字的音调, 则手机显示的音调线 1202可高于 音调线 1201。 另外, 如果检测到用户演唱 “有” 字的音调与原唱演唱 “有” 字的音调相同, 则演唱器 801中的音调线 1202与音调线 1201重合。 这样, 用户在实时演唱的过程中可在演 唱器 801中获知当前演唱的音准情况。
又例如, 如图 12中的 (b)所示, 手机可以在演唱器 801中显示原唱演唱颤音 1的颤音 符 1203。手机检测到用户演唱颤音 1时的用户千声后,可对比用户演唱颤音 1时颤音的位置、 颤动幅度 A⑴、 颤动频率①⑴和平稳趋势项 V⑴与原唱演唱颤音 1时颤音的位置、 颤动幅度 A(t)、 颤动频率①⑴和平稳趋势项 V⑴是否匹配。 若二者不匹配, 说明用户本次演唱颤音 1时 没有达到预设的颤音效果, 则手机可将颤音符 1203显示为虚线等显示效果。 相应的, 若二者 匹配, 说明用户本次演唱颤音 1时达到了预设的颤音效果, 则图 12中的 (b)所示, 手机可 将颤音符 1203加粗或点亮,以便用户在实时演唱的过程中可在演唱器 801中获知当前颤音演 唱效果。
可以理解的是, 本领域技术人员还可以根据实际经验或实际应用场景根据原唱千声的特 征信息对用户在录制歌曲时实时演唱的用户千声进行评估, 并将换气技巧、 颤音技巧或音准 等评估结果展示给用户, 例如, 手机可以根据原唱千声的特征信息对用户千声进行打分, 并 向用户显示打分结果等, 本申请实施例对此不 i故任何限制。
可以看出, 本申请实施例在用户录制歌曲的过程中可根据原唱千声的特征信息, 对用户 演唱的音准、 颤音以及换气等一项或多项演唱技巧进行实时评估, 相比于单纯依据音准评估 用户的演唱水平, 本申请实施例在评估用户演唱时还考虑了用户的换气、 颤音等因素, 因此 得到的评估结果更加准确、 丰富和客观。 并且, 手机可将对用户演唱的评估结果在录制过程 中实时反馈给用户, 便于用户及时了解自己当前的演唱效果, 使用户可以在后续录制过程中 及时调整演唱方式。
S306、 歌曲 A录制结束后, 手机根据原唱千声的特征信息与用户千声的特征信息对用户 千声进行修音。
当手机播放完歌曲 A的伴奏, 或检测到用户输入结束录制的操作时, 手机可关闭麦克风 不再继续采集用户发出的用户千声。 并且, 歌曲 A录制结束后, 如图 13所示, 手机可显示 K 歌类应用的修音界面 1301。 修音界面 1301中可包括修音开关 1302。 如果检测到用户打开该 修音开关 1302, 手机可根据歌曲 A的原唱千声的特征信息, 对本次用户录制歌曲 A时采集 到的用户千声中出现的演唱问题 (例如音准、 颤音或换气声等一项或多项演唱问题)进行修 音。
示例性的, 仍以用户录制歌曲 《红豆》举例, 用户在录制界面录完歌曲 《红豆》后, 手 机可显示如图 13所示的修音界面 1301。 如果检测到用户打开修音界面 1301 中的修音开关 1302, 则手机可根据原唱千声的特征信息对本次采集到的用户千声中的音准、 换气声和颤音 进行修音。
例如, 手机可对比原唱千声的基频信息与用户本次录制歌曲 《红豆》 的基频信息, 从而 对本次用户千声中出现的音准问题进行修正。 示例性的, 如果用户千声中第 10秒-第 20秒的 音调低于原唱千声中第 10秒-第 20秒的音调, 则手机可对这部分用户千声的音调进行升调处 理, 使得这部分用户千声的音调与原唱千声的音调的相同或接近。 相应的, 如果用户千声中 第 1分 12秒-第 1分 27秒的音调高于原唱千声中第 1分 12秒-第 1分 27秒的音调, 则手机 可对这部分用户千声的音调进行降调处理, 使得这部分用户千声的音调与原唱千声的音调的 相同或接近。
又例如, 手机可在用户本次录制歌曲 《红豆》 的用户千声中提取换气声的位置。 示例性 的, 手机可使用预设的换气检测模型检测到在用户千声的第 14秒和第 78秒均出现换气声。 进而, 手机可降低第 14秒和第 78秒处用户千声的能量和频响, 从而弱化第 14秒和第 78秒 处出现的换气声, 以提高录制歌曲的听感质量。
又例如, 手机可对比原唱千声的颤音特征与用户本次录制歌曲 《红豆》 的颤音特征, 从 而对本次用户千声中颤音效果不好的颤音进行修正。 示例性的, 手机可使用预设的颤音检测 模型检测用户千声中出现颤音的位置。 例如, 在用户千声中第 15秒检测到颤音 1, 如果颤音 1的颤动幅度 Al(t)、颤动频率① 1⑴和平稳趋势项 VI⑴与原唱千声中第 15秒处颤音 2的颤动 幅度 A2⑴、 颤动频率① 2⑴和平稳趋势项 V2⑴匹配, 说明用户演唱的颤音 1原唱的效果基本 一致, 则手机无需再对颤音 1进行修正。 相应的, 如果颤音 1的颤动幅度 Al(t)、 颤动频率① 1⑴和平稳趋势项 VI⑴与原唱千声中第 15秒处颤音 2的颤动幅度 A2⑴、 颤动频率① 2⑴和平 稳趋势项 V2(t)不匹配, 则手机可使用颤音 2的颤动幅度 A2(t)、 颤动频率① 2⑴和平稳趋势项 V2⑴在第 15秒的用户千声中合成新的颤音, 从而提高用户千声中出现的颤音音效。 在一些实施例中, 如图 14所示, 检测到用户打开修音界面 1301中的修音开关 1302后, 手机还可进一步显示修正音准的修音选项 1401、 修正换气声的修音选项 1402以及修正颤音 的修音选项 1403。 这样, 用户可以在修音界面 1301 中手动选择修复本次录制的用户千声中 的音准、 换气声或颤音问题。
在另一些实施例中, 手机采集到本次用户录制歌曲 《红豆》 的用户千声后, 还可以自动 对用户千声中的音准、 换气声和颤音等问题进行修正, 以提高后续播放录制歌曲时歌曲的听 感质量。 例如, 仍如图 14所示, 修音界面 1301中设置有播放按钮 1404。 如果检测到用户点 击该播放按钮 1404, 则手机可根据原唱千音的特征信息自动对用户千音中的音准、 换气声和 颤音等问题进行修正, 并将修正后的用户千声融入在歌曲 《红豆》 的伴奏中进行播放。
在一些实施例中, 如图 15中的(a)所示, K歌类应用的修音界面 1301中还可以设置单 句精修的选项 1501。 如果检测到用户点击选项 1501 , 如图 15中的(b)所示, 手机可显示精 修界面 1502。 在精修界面 1502 中手机可以显示歌曲 《红豆》 中用户演唱的每一句歌词, 并 将每一句歌词作为选项供用户选择。 以用户选中 “我会相信一切有尽头”这一句歌词 1503为 例,用户选中歌词 1503后,手机可播放与歌词 1503对应的已录制的部分歌曲。修音界面 1301 中还设置音准按钮 1504、 颤音按钮 1505以及换气按钮 1506。 用户可手动在精修界面 1502中 选择是否修复歌词 1503 对应的用户千声中出现的音准、 颤音或换气声等一项或多项演唱问 题。如果用户选择歌词 1503后点击了颤音按钮 1525和音准按钮 1504,则手机可对与歌词 1503 对应的用户千声中的音准和颤音问题进行修正, 并将修正后的用户千声融入在对应的伴奏中 播放。 这样, 用户可以逐句修正本次录制的用户千声中出现的演唱问题, 并且, 用户可逐句 比较修正前后的录制效果, 从而将用户录制的歌曲修正为用户更满意的演唱效果。
在另一些实施例中, 如图 16中的(a)所示, K歌类应用的修音界面 1301中还可以设置 单句编辑的选项 1601。 如果检测到用户点击选项 1601, 如图 16中的(b)所示, 手机可显示 编辑界面 1602。 在编辑界面 1602 中手机可以显示歌曲 《红豆》 中用户演唱的每一句歌词, 并且, 在显示每一句歌词时, 手机还可在歌词的相应位置提示用户在演唱这句歌词时出现的 演唱问题, 例如, 音调不准、 换气声较大、 颤音不稳、 缺少颤音等。 这样, 用户可以直观的 看到自己在演唱每一句歌词时出现的具体问题。
另外, 仍如图 16中的(b)所示, 在编辑界面 1602中还可设置与每一句歌词对应的编辑 按钮 1603。如果检测到用户点击某一句歌词对应的编辑按钮 1603, 手机可播放与该歌词对应 的伴奏, 并打开麦克风采集用户重新演唱该歌词的用户千声。 手机采集到用户重新演唱该歌 词的用户千声后, 可使用本次重新录制的用户千声替换已录制的歌曲 《红豆》 中对应的用户 千声。 由于手机在编辑界面 1602显示歌词时已经向用户提示演唱该歌词时的具体演唱问题, 因此, 用户重新演唱该歌词时可及时调整演唱方式修正上次录制该歌词时出现的演唱问题, 从而提高演唱和录制质量。
如图 17所示,本申请实施例公开了一种电子设备, 包括: 触摸屏 1701,所述触摸屏 1701 包括触敏表面 1706和显示屏 1707; —个或多个处理器 1702; 存储器 1703; —个或多个麦克 风 1708; 以及一个或多个计算机程序 1704。 上述各器件可以通过一个或多个通信总线 1705 连接。 其中, 上述一个或多个计算机程序 1704被存储在上述存储器 1703中并被配置为被该 一个或多个处理器 1702执行, 该一个或多个计算机程序 1704包括指令, 上述指令可以用于 执行上述应实施例中的各个步骤。 示例性的, 上述处理器 1702具体可以为图 1所示的处理器 110, 上述存储器 1703具体 可以为图 1所示的内部存储器 121和 /或外部存储器 120,上述显示屏 1707具体可以为图 1所 示的显示屏 194, 上述麦克风 1708具体可以为图 1所示的麦克风 170C, 上述触敏表面 1706 具体可以为图 1所示的传感器模块 180中的触摸传感器 180K,本申请实施例对此不做任何限 制。
在一些实施例中, 本申请还提供了一种图形用户界面 (GUI), 该图形用户界面可存储在 电子设备中。 示例性的, 该电子设备可以为图 1或图 17所示的电子设备。
示例性的,上述图形用户界面包括:显示在触摸屏上的第一 GUI,如图 4所示, 第一 GUI 可以为 K歌类应用的界面 401, 第一 GUI中包括用于录制第一歌曲的录制按钮 403; 响应于 针对该录制按钮的触摸事件, 电子设备可在触摸屏上显示第二 GUI, 如图 7或图 8所示, 第 二 GUI可以为用户录制第一歌曲的录制界面 601。 第二 GUI中可以包括指导用户演唱第一歌 曲的指导信息, 该指导信息包括换气提示和 /或颤音提示。 例如, 如图 7所示, 录制界面 601 显示的歌词 602中包括换气提示 701和颤音提示 702。 又例如, 如图 8所示, 录制界面 601 显示的演唱器 801中音准提示 802、 换气提示 804和颤音提示 803。 第二 GUI中的指导信息 是电子设备根据原唱千声的特征性信息生成的。
在一些实施例中,上述第二 GUI中还可以包括对用户演唱时的换气技巧和 /或颤音技巧的 评估结果。 例如, 如图 10中的 (a) -(c)所示, 录制界面 601 中包括对用户换气技巧的评 估结果 1001, 对用户颤音技巧的评估结果 1002以及对用户音准的评估结果 1003。 又例如, 如图 12中的( a)-( b)所示,可在录制界面 601的演唱器中显示对用户音准的评估结果 1202, 对用户颤音技巧的评估结果 1203等, 本申请实施例对此不彳故任何限制。
在一些实施例中, 当第一歌曲录制结束后, 电子设备可在触摸屏上显示第三 GUI, 如图 13或 14所示, 第三 GUI可为第一歌曲的修音界面 1301 , 第三 GUI中包括修音开关 1302; 当用户打开修音开关 1302 时, 电子设备可对用户千声中出现的换气问题和成颤音问题等演 唱问题进行修正。
通过以上的实施方式的描述, 所属领域的技术人员可以清楚地了解到, 为描述的方便和 筒洁, 仅以上述各功能模块的划分进行举例说明, 实际应用中, 可以根据需要而将上述功能 分配由不同的功能模块完成, 即将装置的内部结构划分成不同的功能模块, 以完成以上描述 的全部或者部分功能。 上述描述的系统, 装置和单元的具体工作过程, 可以参考前述方法实 施例中的对应过程, 在此不再赘述。
在本申请实施例各个实施例中的各功能单元可以集成在一个处理单元中, 也可以是各个 单元单独物理存在, 也可以两个或两个以上单元集成在一个单元中。 上述集成的单元既可以 采用硬件的形式实现, 也可以采用软件功能单元的形式实现。
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时, 可以 存储在一个计算机可读取存储介质中。 基于这样的理解, 本申请实施例的技术方案本质上或 者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出 来, 该计算机软件产品存储在一个存储介质中, 包括若千指令用以使得一台计算机设备(可 以是个人计算机, 服务器, 或者网络设备等)或处理器执行本申请各个实施例所述方法的全 部或部分步骤。 而前述的存储介质包括: 快闪存储器、 移动硬盘、 只读存储器、 随机存取存 储器、 磁碟或者光盘等各种可以存储程序代码的介质。
以上所述, 仅为本申请实施例的具体实施方式, 但本申请实施例的保护范围并不局限于 此, 任何在本申请实施例揭露的技术范围内的变化或替换, 都应涵盖在本申请实施例的保护 范围之内。 因此, 本申请实施例的保护范围应以所述权利要求的保护范围为准。

Claims

权 利 要 求 书
1、 一种歌曲录制方法, 其特征在于, 包括:
电子设备显示第一界面, 所述第一界面中包括用于录制第一歌曲的录制按钮; 响应于用户选择所述录制按钮的操作, 所述电子设备获取所述第一歌曲的伴奏和原唱千 声的特征信息, 所述原唱千声的特征信息用于反映原唱在演唱所述第一歌曲时的演唱特征; 所述电子设备开始播放所述第一歌曲的伴奏并开始录制用户演唱的用户千声; 在录制用户千声的过程中, 所述电子设备根据所述原唱千声的特征信息在第二界面中显 示指导信息, 所述指导信息用于指导用户演唱时的换气或颤音中的一项或多项。
2、根据权利要求 1所述的方法, 其特征在于, 所述原唱千声的特征信息包括原唱千声的 换气位置; 所述指导信息包括换气提示;
其中, 所述电子设备根据所述原唱千声的特征信息在所述第二界面中显示指导信息, 包 括:
所述电子设备根据所述原唱千声的换气位置, 在所述第二界面中显示与所述原唱千声的 换气位置对应的换气提示。
3、根据权利要求 2所述的方法,其特征在于,所述第二界面中包括所述第一歌曲的歌词; 其中, 所述电子设备根据所述原唱千声的换气位置, 在所述第二界面中显示与所述原唱 千声的换气位置对应的换气提示, 包括:
所述电子设备在所述歌词的第一位置显示所述换气提示, 所述第一位置与所述原唱千声 的换气位置相同。
4、 根据权利要求 2所述的方法, 其特征在于, 所述第二界面中包括演唱器, 所述演唱器 用于显示正在录制的歌词中每个字的音符;
其中, 所述电子设备根据所述原唱千声的换气位置, 在所述第二界面中显示与所述原唱 千声的换气位置对应的换气提示, 包括:
所述电子设备在所述演唱器的第二位置显示所述换气提示, 所述第二位置与所述原唱千 声的换气位置相同。
5、根据权利要求 1所述的方法, 其特征在于, 所述原唱千声的特征信息包括原唱千声的 颤音特征; 所述指导信息包括颤音提示;
其中, 所述电子设备根据所述原唱千声的特征信息在所述第二界面中显示指导信息, 包 括:
所述电子设备根据所述原唱千声的颤音特征, 在所述第二界面中显示与所述原唱千声的 颤音特征对应的颤音提示。
6、 根据权利要求 5所述的方法, 其特征在于, 所述颤音特征中包括颤音位置; 所述第二 界面中包括所述第一歌曲的歌词;
其中, 所述电子设备根据所述原唱千声的颤音特征, 在所述第二界面中显示与所述原唱 千声的颤音特征对应的颤音提示, 包括:
所述电子设备在所述歌词的第三位置显示所述颤音提示, 所述第三位置与所述原唱千声 中的颤音位置相同。
7、 根据权利要求 5所述的方法, 其特征在于, 所述颤音特征中包括颤音位置; 所述第二 界面中包括演唱器, 所述演唱器用于显示正在录制的歌词中每个字的音符;
其中, 所述电子设备根据所述原唱千声的颤音特征, 在所述第二界面中显示与所述原唱 千声的颤音特征对应的颤音提示, 包括:
所述电子设备在所述演唱器的第四位置显示所述颤音提示, 所述第四位置与所述原唱千 声中的颤音位置相同。
8、 根据权利要求 1-7中任一项所述的方法, 其特征在于, 在录制用户千声的过程中, 所 述方法还包括:
所述电子设备从所述用户千声中提取所述用户千声的特征信息, 所述用户千声的特征信 息用于反映用户在演唱所述第一歌曲时的演唱特征;
所述电子设备通过对比所述用户千声中的特征信息与所述原唱千声的特征信息, 在所述 第二界面中显示评估结果。
9、根据权利要求 8所述的方法, 其特征在于, 所述用户千声中的特征信息包括用户千声 的换气位置, 所述原唱千声的特征信息包括原唱千声的换气位置;
其中, 所述电子设备通过对比所述用户千声中的特征信息与所述原唱千声的特征信息, 在所述第二界面中显示评估结果, 包括:
所述电子设备通过对比所述用户千声的换气位置和所述原唱千声的换气位置, 在所述第 二界面中显示换气技巧的评估结果。
10、 根据权利要求 9所述的方法, 其特征在于, 所述电子设备通过对比所述用户千声的 换气位置和所述原唱千声的换气位置, 在所述第二界面中显示换气技巧的评估结果, 包括: 若所述用户千声的换气位置与所述原唱千声的换气位置相同, 则所述电子设备在所述第 二界面中显示换气准确的提示;
若所述用户千声的换气位置与所述原唱千声的换气位置不相同, 则所述电子设备在所述 第二界面中显示换气错误的提示。
11、 根据权利要求 8所述的方法, 其特征在于, 所述用户千声中的特征信息包括用户千 声的颤音特征, 所述原唱千声的特征信息包括原唱千声的颤音特征;
其中, 所述电子设备通过对比所述用户千声中的特征信息与所述原唱千声的特征信息, 在所述第二界面中显示评估结果, 包括:
所述电子设备通过对比所述用户千声的颤音特征和所述原唱千声的颤音特征, 在所述第 二界面中显示颤音技巧的评估结果。
12、 根据权利要求 11所述的方法, 其特征在于, 所述电子设备通过对比所述用户千声的 颤音特征和所述原唱千声的颤音特征, 在所述第二界面中显示颤音技巧的评估结果, 包括: 若所述用户千声的颤音特征与所述原唱千声的颤音特征匹配, 则所述电子设备在所述第 二界面中显示颤音准确的提示;
若所述用户千声的颤音特征与所述原唱千声的颤音特征不匹配, 则所述电子设备在所述 第二界面中显示颤音错误的提示。
13、 根据权利要求 8-12中任一项所述的方法, 其特征在于, 所述电子设备在所述第二界 面中显示评估结果, 包括:
所述电子设备在所述第二界面的歌词中显示所述评估结果; 或者,
所述电子设备在所述第二界面的演唱器中显示所述评估结果, 所述演唱器用于显示正在 录制的歌词中每个字的音符。
14、 根据权利要求 1-13中任一项所述的方法, 其特征在于, 在所述电子设备根据所述原 唱千声的特征信息在所述第二界面中显示指导信息之后, 还包括: 所述第一歌曲录制结束后, 所述电子设备根据所述原唱千声的特征信息修正所述用户千 声中出现的演唱问题, 所述演唱问题包括换气问题或颤音问题中的一项或多项。
15、 一种修音方法, 其特征在于, 包括:
电子设备显示第一界面, 所述第一界面中包括用于录制第一歌曲的录制按钮; 响应于用户选择所述录制按钮的操作, 所述电子设备获取所述第一歌曲的伴奏和原唱千 声的特征信息, 所述原唱千声的特征信息用于反映原唱在演唱所述第一歌曲时的演唱特征; 所述电子设备开始播放所述第一歌曲的伴奏并开始录制用户演唱的用户千声; 所述第一歌曲录制结束后, 所述电子设备根据所述原唱千声的特征信息修正所述用户千 声中出现的演唱问题, 所述演唱问题包括换气问题或颤音问题中的一项或多项。
16、 根据权利要求 15所述的方法, 其特征在于, 所述演唱问题为换气问题;
其中,所述电子设备根据所述原唱千声的特征信息修正所述用户千声中出现的演唱问题, 包括:
所述电子设备从所述用户千声中提取所述用户千声中的换气声;
所述电子设备降低所述用户千声中换气声的能量和频响, 以修正所述用户千声中出现的 换气问题。
17、 根据权利要求 15所述的方法, 其特征在于, 所述演唱问题为颤音问题,
其中,所述电子设备根据所述原唱千声的特征信息修正所述用户千声中出现的演唱问题, 包括:
所述电子设备从所述用户千声中提取所述用户千声的颤音特征;
若所述用户千声的颤音特征与所述原唱千声的特征信息中的颤音特征不匹配, 则所述电 子设备使用所述原唱千声中的颤音特征修正所述用户千声中的颤音特征, 以修正所述用户千 声中出现的颤音问题。
18、 根据权利要求 15-17中任一项所述的方法, 其特征在于, 所述第一歌曲录制结束后, 所述方法还包括:
所述电子设备显示第二界面, 所述第二界面中包括修音开关;
其中,所述电子设备根据所述原唱千声的特征信息修正所述用户千声中出现的演唱问题, 包括:
若检测到用户打开所述修音开关, 则所述电子设备根据所述原唱千声的特征信息修正所 述用户千声中出现的演唱问题。
19、 根据权利要求 15-17中任一项所述的方法, 其特征在于, 所述第一歌曲录制结束后, 所述方法还包括:
所述电子设备显示第二界面, 所述第二界面中包括修音选项, 所述修音选项中包括修正 音准的选项、 修正换气声的选项以及修正颤音的选项;
其中,所述电子设备根据所述原唱千声的特征信息修正所述用户千声中出现的演唱问题, 包括:
响应于用户选中的修音选项, 所述电子设备根据所述原唱千声的特征信息修正所述用户 千声中与所述修音选项对应的演唱问题。
20、 根据权利要求 15-17中任一项所述的方法, 其特征在于, 所述第一歌曲录制结束后, 所述方法还包括:
所述电子设备显示第二界面, 所述第二界面中包括单句精修的选项; 响应于用户选择所述单句精修的选项, 所述电子设备显示第三界面, 所述第三界面中包 括所述第一歌曲中的 N句歌词, N > 1 ;
其中,所述电子设备根据所述原唱千声的特征信息修正所述用户千声中出现的演唱问题, 包括:
若用户在所述 N句歌词中选中第一歌词,则所述电子设备根据所述原唱千声的特征信息, 修正与所述第一歌词对应的用户千声中出现的演唱问题。
21、 根据权利要求 20所述的方法, 其特征在于, 所述电子设备显示第三界面, 包括: 所述电子设备在显示所述第三界面中的 N句歌词时, 标记出现所述演唱问题的位置。
22、 一种电子设备, 其特征在于, 包括:
触摸屏, 所述触摸屏包括触敏表面和显示屏;
一个或多个处理器;
一个或多个存储器;
一个或多个麦克风;
以及一个或多个计算机程序, 其中所述一个或多个计算机程序被存储在所述一个或多个 存储器中, 所述一个或多个计算机程序包括指令, 当所述指令被所述电子设备执行时, 使得 所述电子设备执行如权利要求 1-14中任一项所述的歌曲录制方法, 或权利要求 15-21中任一 项所述的修音方法。
23、 一种计算机可读存储介质, 所述计算机可读存储介质中存储有指令, 其特征在于, 当所述指令在电子设备上运行时,使得所述电子设备执行如权利要求 1-14中任一项所述的歌 曲录制方法, 或权利要求 15-21中任一项所述的修音方法。
24、 一种包含指令的计算机程序产品, 其特征在于, 当所述计算机程序产品在电子设备 上运行时, 使得所述电子设备执行如权利要求 1-14中任一项所述的歌曲录制方法, 或权利要 求 15-21中任一项所述的修音方法。
25、 一种图形用户界面 GUI, 所述图形用户界面存储在电子设备中, 所述电子设备包括 触摸屏、 存储器、 处理器, 所述处理器用于执行存储在所述存储器中的一个或多个计算机程 序, 其特征在于, 所述图形用户界面包括:
显示在所述触摸屏上的第一 GUI, 所述第一 GUI中包括用于录制第一歌曲的录制按钮; 响应于针对所述录制按钮的触摸事件, 在所述触摸屏上显示第二 GUI, 所述第二 GUI为 录制所述第一歌曲的录制界面,所述第二 GUI中包括指导用户演唱所述第一歌曲的指导信息, 所述指导信息包括换气提示和 /或颤音提示。
26、 根据权利要求 25所述的 GUI, 其特征在于, 所述第二 GUI中还包括对用户演唱时 的换气技巧和 /或颤音技巧的评估结果。
27、 根据权利要求 25或 26所述的 GUI, 其特征在于, 在所述触摸屏上显示第二 GUI之 后, 还包括:
所述第一歌曲录制结束后, 在所述触摸屏上显示第三 GUI, 所述第三 GUI为所述第一歌 曲的修音界面, 所述第三 GUI中包括修音开关; 当所述修音开关打开时, 所述电子设备对用 户千声中出现的换气问题和 /或颤音问题进行修正。
PCT/CN2020/076099 2019-02-28 2020-02-21 一种歌曲录制方法、修音方法及电子设备 WO2020173391A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2021550197A JP7222112B2 (ja) 2019-02-28 2020-02-21 歌の録音方法、音声補正方法、および電子デバイス
US17/434,555 US20220130360A1 (en) 2019-02-28 2020-02-21 Song Recording Method, Audio Correction Method, and Electronic Device
EP20762220.0A EP3905246A4 (en) 2019-02-28 2020-02-21 SONG RECORDING METHOD, SOUND CORRECTION METHOD AND ELECTRONIC DEVICE

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910152655.3A CN110010162A (zh) 2019-02-28 2019-02-28 一种歌曲录制方法、修音方法及电子设备
CN201910152655.3 2019-02-28

Publications (1)

Publication Number Publication Date
WO2020173391A1 true WO2020173391A1 (zh) 2020-09-03

Family

ID=67166258

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/076099 WO2020173391A1 (zh) 2019-02-28 2020-02-21 一种歌曲录制方法、修音方法及电子设备

Country Status (5)

Country Link
US (1) US20220130360A1 (zh)
EP (1) EP3905246A4 (zh)
JP (1) JP7222112B2 (zh)
CN (1) CN110010162A (zh)
WO (1) WO2020173391A1 (zh)

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109803523B (zh) 2019-02-23 2021-01-29 华为技术有限公司 一种封装屏蔽结构及电子设备
CN110010162A (zh) * 2019-02-28 2019-07-12 华为技术有限公司 一种歌曲录制方法、修音方法及电子设备
CN112309351A (zh) * 2019-07-31 2021-02-02 武汉Tcl集团工业研究院有限公司 一种歌曲生成方法、装置、智能终端及存储介质
CN111507254B (zh) * 2020-04-16 2023-10-24 北京雷石天地电子技术有限公司 制作逐字歌词的方法、装置、终端和非临时性计算机可读存储介质
CN111782868A (zh) * 2020-07-06 2020-10-16 腾讯音乐娱乐科技(深圳)有限公司 一种音频处理方法、装置、设备及介质
CN111968623B (zh) * 2020-08-19 2023-11-28 腾讯音乐娱乐科技(深圳)有限公司 气口位置检测方法及相关设备
CN112289300B (zh) * 2020-10-28 2024-01-09 腾讯音乐娱乐科技(深圳)有限公司 音频处理方法、装置及电子设备和计算机可读存储介质
CN112596695B (zh) * 2020-12-30 2024-03-12 北京达佳互联信息技术有限公司 歌曲导唱方法、装置、电子设备及存储介质
CN113035164A (zh) * 2021-02-24 2021-06-25 腾讯音乐娱乐科技(深圳)有限公司 歌声生成方法和装置、电子设备及存储介质
US20220293117A1 (en) * 2021-03-15 2022-09-15 Meta Platforms, Inc. Systems and methods for transforming audio in content items
US11523201B2 (en) * 2021-03-23 2022-12-06 Shanghai Greatcare Trading & Developing Co., Ltd. Multifunctional microphone
CN113791848B (zh) * 2021-08-09 2023-03-28 北京达佳互联信息技术有限公司 一种音频录制提示方法、装置、电子设备及存储介质
CN113707113B (zh) * 2021-08-24 2024-02-23 北京达佳互联信息技术有限公司 用户歌声的修音方法、装置及电子设备
CN113920786B (zh) * 2021-09-07 2024-02-23 北京小唱科技有限公司 唱歌教学方法和装置
CN114387994A (zh) * 2021-12-27 2022-04-22 北京达佳互联信息技术有限公司 音频数据获取方法及装置
CN114566191A (zh) * 2022-02-25 2022-05-31 腾讯音乐娱乐科技(深圳)有限公司 录音的修音方法及相关装置
WO2023235676A1 (en) * 2022-05-31 2023-12-07 Dts Inc. Enhanced music delivery system with metadata

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070074619A1 (en) * 2005-10-04 2007-04-05 Linda Vergo System and method for tailoring music to an activity based on an activity goal
CN103077701A (zh) * 2012-11-28 2013-05-01 福建星网视易信息系统有限公司 一种音准评定方法、装置和系统
CN103559875A (zh) * 2013-10-21 2014-02-05 福建星网视易信息系统有限公司 音准抖动修正方法、装置、系统及音视频设备和移动终端
CN104715760A (zh) * 2015-02-13 2015-06-17 朱威 一种k歌匹配分析方法及系统
CN105825844A (zh) * 2015-07-30 2016-08-03 维沃移动通信有限公司 一种修音的方法和装置
CN110010162A (zh) * 2019-02-28 2019-07-12 华为技术有限公司 一种歌曲录制方法、修音方法及电子设备

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7271329B2 (en) * 2004-05-28 2007-09-18 Electronic Learning Products, Inc. Computer-aided learning system employing a pitch tracking line
JP2007322934A (ja) * 2006-06-02 2007-12-13 Yamaha Corp 歌唱練習装置、カラオケ装置及びプログラム
CN103165119B (zh) * 2013-01-31 2015-12-09 华为技术有限公司 K歌方法及系统
JP6070652B2 (ja) * 2014-07-28 2017-02-01 ヤマハ株式会社 リファレンス表示装置およびプログラム
JP6358018B2 (ja) * 2014-09-30 2018-07-18 ブラザー工業株式会社 カラオケ装置、及びプログラム
US10043548B1 (en) * 2017-02-23 2018-08-07 International Business Machines Corporation Creating a work based on cognitive state
CN106997769B (zh) * 2017-03-25 2020-04-24 腾讯音乐娱乐(深圳)有限公司 颤音识别方法及装置
US10117021B1 (en) * 2017-03-31 2018-10-30 Polycom, Inc. Audio feedback reduction utilizing adaptive filters and nonlinear processing
CN107393519B (zh) * 2017-08-03 2020-09-15 腾讯音乐娱乐(深圳)有限公司 演唱评分的显示方法、装置及存储介质
CN108922562A (zh) * 2018-06-15 2018-11-30 广州酷狗计算机科技有限公司 演唱评价结果显示方法及装置
CN109272975B (zh) * 2018-08-14 2023-06-27 无锡冰河计算机科技发展有限公司 演唱伴奏自动调整方法、装置及ktv点唱机

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070074619A1 (en) * 2005-10-04 2007-04-05 Linda Vergo System and method for tailoring music to an activity based on an activity goal
CN103077701A (zh) * 2012-11-28 2013-05-01 福建星网视易信息系统有限公司 一种音准评定方法、装置和系统
CN103559875A (zh) * 2013-10-21 2014-02-05 福建星网视易信息系统有限公司 音准抖动修正方法、装置、系统及音视频设备和移动终端
CN104715760A (zh) * 2015-02-13 2015-06-17 朱威 一种k歌匹配分析方法及系统
CN105825844A (zh) * 2015-07-30 2016-08-03 维沃移动通信有限公司 一种修音的方法和装置
CN110010162A (zh) * 2019-02-28 2019-07-12 华为技术有限公司 一种歌曲录制方法、修音方法及电子设备

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3905246A4

Also Published As

Publication number Publication date
EP3905246A4 (en) 2022-03-23
JP7222112B2 (ja) 2023-02-14
CN110010162A (zh) 2019-07-12
JP2022522456A (ja) 2022-04-19
US20220130360A1 (en) 2022-04-28
EP3905246A1 (en) 2021-11-03

Similar Documents

Publication Publication Date Title
JP7222112B2 (ja) 歌の録音方法、音声補正方法、および電子デバイス
RU2766255C1 (ru) Способ голосового управления и электронное устройство
CN110134316B (zh) 模型训练方法、情绪识别方法及相关装置和设备
CN111345010B (zh) 一种多媒体内容同步方法、电子设备及存储介质
CN110910872B (zh) 语音交互方法及装置
WO2021147415A1 (zh) 实现立体声输出的方法及终端
CN111819533B (zh) 一种触发电子设备执行功能的方法及电子设备
CN111666119A (zh) Ui组件显示的方法及电子设备
CN111628916A (zh) 一种智能音箱与电子设备协作的方法及电子设备
WO2020239001A1 (zh) 一种哼唱识别方法及相关设备
CN113593608A (zh) 基于对象识别的美音方法、电子设备及存储介质
WO2022267468A1 (zh) 一种声音处理方法及其装置
WO2022143258A1 (zh) 一种语音交互处理方法及相关装置
WO2022127211A1 (zh) 震动方法、装置、电子设备和可读存储介质
CN111031492A (zh) 呼叫需求响应方法、装置及电子设备
CN115543145A (zh) 一种文件夹管理方法及装置
CN115359156B (zh) 音频播放方法、装置、设备和存储介质
CN114173184A (zh) 投屏方法和电子设备
EP4293664A1 (en) Voiceprint recognition method, graphical interface, and electronic device
WO2023273904A1 (zh) 音频数据的存储方法及其相关设备
WO2022007757A1 (zh) 跨设备声纹注册方法、电子设备及存储介质
WO2023000778A9 (zh) 一种音频信号的处理方法及相关电子设备
CN113407300B (zh) 应用误杀评估方法及相关设备
CN113380240B (zh) 语音交互方法和电子设备
RU2777617C1 (ru) Способ записи песни, способ коррекции звука и электронное устройство

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20762220

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020762220

Country of ref document: EP

Effective date: 20210730

ENP Entry into the national phase

Ref document number: 2021550197

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE