CN111666445A - Scene lyric display method and device and sound box equipment - Google Patents

Scene lyric display method and device and sound box equipment Download PDF

Info

Publication number
CN111666445A
CN111666445A CN201910169089.7A CN201910169089A CN111666445A CN 111666445 A CN111666445 A CN 111666445A CN 201910169089 A CN201910169089 A CN 201910169089A CN 111666445 A CN111666445 A CN 111666445A
Authority
CN
China
Prior art keywords
lyrics
song
audio data
played
scene
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910169089.7A
Other languages
Chinese (zh)
Inventor
杨廉萍
吴海全
唐大勇
张恩勤
曹磊
师瑞文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Grandsun Electronics Co Ltd
Original Assignee
Shenzhen Grandsun Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Grandsun Electronics Co Ltd filed Critical Shenzhen Grandsun Electronics Co Ltd
Priority to CN201910169089.7A priority Critical patent/CN111666445A/en
Publication of CN111666445A publication Critical patent/CN111666445A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/638Presentation of query results
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/04Time compression or expansion
    • G10L21/055Time compression or expansion for synchronising with other signals, e.g. video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/005Non-interactive screen display of musical or status data
    • G10H2220/011Lyrics displays, e.g. for karaoke applications

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Acoustics & Sound (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Signal Processing (AREA)
  • Quality & Reliability (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Reverberation, Karaoke And Other Acoustics (AREA)

Abstract

The invention is suitable for the technical field of multimedia, and provides a scene lyric display method, a device and a sound box device, wherein the method comprises the following steps: under the networking state, acquiring audio data of a song to be played from a server; acquiring lyrics matched with the song in the server according to the audio data; generating a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics; and controlling the scene mode, the lyrics and the playing progress of the song to be synchronously displayed. The invention solves the problems that the lyrics are easy to be out of synchronization with the played music or inconsistent with the content when being displayed, and the display interface is single.

Description

Scene lyric display method and device and sound box equipment
Technical Field
The invention belongs to the technical field of multimedia, and particularly relates to a scene lyric display method and device and sound box equipment.
Background
With the development of science and technology, intelligent sound boxes become an important branch of the field of artificial intelligence; as a novel intelligent household appliance, the intelligent sound box can be connected to the Internet and wirelessly exchanges data with a mobile terminal or a server or executes various commands.
At present, some wireless intelligent sound boxes with electronic screens can only display song names, singer names and photo or cover page picture information while playing music, and some sound boxes can also display lyrics, but the situation that the music is not synchronous or the content is not consistent with the played music easily exists when the lyrics are displayed, and the display interface is single and cannot meet the requirements of users.
Disclosure of Invention
In view of this, embodiments of the present invention provide a method and an apparatus for displaying contextual lyrics, and a sound box device, so as to solve the problems in the prior art that when lyrics are displayed, the lyrics are likely to be out of synchronization with played music or inconsistent in content, and a display interface is single.
The first aspect of the embodiment of the present invention provides a method for displaying scene lyrics, which is applied to a sound box device, and the method for displaying scene lyrics includes:
under the networking state, acquiring audio data of a song to be played from a server;
acquiring lyrics matched with the song in the server according to the audio data;
generating a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics;
and controlling the scene mode, the lyrics and the playing progress of the song to be synchronously displayed.
In one embodiment, before obtaining the audio data of the song to be played from the server in the networked state, the method includes:
receiving a networking trigger signal through a display screen;
and after the network connection is finished, searching audio data of the song to be played according to the received trigger signal for playing the music input by the user.
In one embodiment, the audio data includes spectral feature data, a name, and singer information for the song.
In one embodiment, the obtaining of the lyrics of the song according to the matching of the song in the audio data obtaining server comprises:
analyzing the audio data to obtain the keywords in the audio data or the frequency spectrum characteristics of the song prelude;
and acquiring and storing lyrics matched with the song according to the keywords or the frequency spectrum characteristics.
In one embodiment, generating a contextual model corresponding to the song based on the characteristics of the audio data and the lyrics comprises:
generating a driving signal of a display screen according to the characteristics of the audio data and the lyrics;
driving a display screen to display the scene mode corresponding to the song through the driving signal:
wherein the scene mode dynamically changes with the frequency spectrum of the song and the change of the lyrics.
In one embodiment, controlling the contextual model, the lyrics and the synchronous display of the playing progress of the song comprises:
and according to the playing rhythm of the lyrics, highlighting each character in the played lyrics of the current sentence one by one along with the playing rhythm under the corresponding contextual model.
In one embodiment, the contextual model includes a background color or background animation corresponding to a plurality of song genres when the song is played and lyrics are displayed.
A second aspect of an embodiment of the present invention provides a device for displaying scene lyrics, including:
the system comprises a first data acquisition unit, a second data acquisition unit and a control unit, wherein the first data acquisition unit is used for acquiring audio data of a song to be played from a server in a networking state;
the second data acquisition unit is used for acquiring lyrics matched with the song in the server according to the audio data;
the driving unit is used for generating a scene mode corresponding to the song according to the characteristics of the audio data and the lyrics;
and the control unit is used for controlling the scene mode, the lyrics and the synchronous display of the playing progress of the song.
A third aspect of the embodiments of the present invention provides a sound box apparatus, including a memory, a processor, and a computer program stored in the memory and executable on the processor, where the processor implements the steps of the method when executing the computer program.
A fourth aspect of embodiments of the present invention provides a computer-readable storage medium storing a computer program which, when executed by a processor, implements the steps of the above-described method.
Compared with the prior art, the embodiment of the invention has the following beneficial effects: according to the embodiment of the invention, under the networking state, the audio data of the song to be played is acquired from the server; acquiring lyrics matched with the song in the server according to the audio data; generating a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics; controlling the synchronous display of the contextual model, the lyrics and the playing progress of the song; the problem that the lyrics are easy to be out of synchronization with the played music or inconsistent in content when displayed is solved, and the display interface is single; through the analysis of the characteristics of the audio data and the content of the lyrics, the lyrics and the corresponding contextual model which are synchronously displayed are generated and synchronously displayed with the lyrics along with the playing progress of the song, thereby enriching the scenes of the song playing.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the embodiments or the prior art descriptions will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without inventive exercise.
Fig. 1 is a schematic flow chart illustrating an implementation of a method for displaying scene lyrics according to an embodiment of the present invention;
FIG. 2 is a diagram of a device for displaying scene lyrics according to an embodiment of the present invention;
fig. 3 is a schematic diagram of a sound box device according to an embodiment of the present invention.
Detailed Description
In the following description, for purposes of explanation and not limitation, specific details are set forth, such as particular system structures, techniques, etc. in order to provide a thorough understanding of the embodiments of the invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced in other embodiments that depart from these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present invention with unnecessary detail.
It will be understood that the terms "comprises" and/or "comprising," when used in this specification and the appended claims, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
It is also to be understood that the terminology used in the description of the invention herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used in the specification of the present invention and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise.
It should be further understood that the term "and/or" as used in this specification and the appended claims refers to and includes any and all possible combinations of one or more of the associated listed items.
In order to explain the technical means of the present invention, the following description will be given by way of specific examples.
Referring to fig. 1, a schematic diagram of an implementation flow of a method for displaying scene lyrics provided by an embodiment of the present invention is shown, where the method is applied to a sound box device, and the sound box device includes a smart sound box or a bluetooth sound box; the applicable sound box equipment further comprises a color electronic display screen, wherein the color electronic display screen can be a light emitting diode display screen (LED), an organic light emitting diode display screen (OLED), a liquid crystal display screen (LCD) and the like, and the electronic screen can display characters, color pictures, video animations and the like in different colors.
As shown, the method comprises the following steps:
step S101, under the state of networking, obtaining the audio data of the song to be played from the server.
In this embodiment, the speaker device includes a smart speaker or a bluetooth speaker; the loudspeaker box equipment can be wirelessly accessed to the Internet through Wi-Fi, or is in wireless communication connection with the mobile terminal through Bluetooth, and is in wireless communication with the Internet through the mobile terminal. The loudspeaker box equipment can also be connected to the Internet in a wired mode.
The audio data of the song to be played can be searched in a network mode according to the song name input by the user, and the audio data can be obtained from the server side or obtained by receiving song information sent by the mobile terminal. The audio data may include spectral feature data, title of the song, and singer information; the spectral feature data of a song may include the temperament, tone, and tempo information of the song, etc.
Optionally, in a networked state, before acquiring the audio data of the song to be played from the server, the method includes:
receiving a networking trigger signal through a display screen;
after the network connection is finished, searching the audio data of the song to be played according to the trigger signal for playing the music input by the user.
In this embodiment, the sound box device can receive a trigger signal set by a user through the electronic display screen, select a router device to be networked according to the trigger signal, and access the internet; or the mobile terminal establishes Bluetooth connection with the mobile terminal and selects the router device to be connected, thereby accessing the Internet.
The triggering signal for playing music input by the user comprises a signal generated by selecting a playing button on the sound box device or touching a playing control corresponding to a selected song through an electronic display screen. The trigger signal comprises a song name, and the song is searched according to the trigger signal to obtain audio data corresponding to the song name, wherein the audio data comprises frequency spectrum characteristic data, the song name and singer information of the song.
And step S102, acquiring lyrics matched with the song in the server according to the audio data.
In this embodiment, the audio data includes spectral feature data, a name, and singer information of the song; the sound box device wirelessly and quickly searches lyrics matched with the song through a specific algorithm under a networking state according to audio data, wherein the specific algorithm comprises a song name symbol (token) matching algorithm, a singer and song name text retrieval algorithm, a music melody based humming Query (QBH) retrieval algorithm and the like.
Optionally, the obtaining, according to the lyrics matched with the song in the audio data obtaining server, the method includes:
analyzing the audio data to obtain the keywords in the audio data or the frequency spectrum characteristics of the song prelude;
and acquiring and storing lyrics matched with the song according to the keywords or the frequency spectrum characteristics.
In this embodiment, the keywords include a text in a song title or a text in a singer title, and may also be searched by combining a plurality of keywords, for example, a combination of keywords such as "tomorrow" and "better" in a song title may be searched for. And the frequency spectrum characteristic of the prelude part of the song can be extracted, and the music melody of the prelude of the song can be identified and searched. After the matched lyric text is found, the lyric text is stored in a storage unit, such as a Random Access Memory (RAM) or a FLASH memory (FLASH), and the lyric text can be cached in a cache region, and corresponding cache time is set, so that the lyric can quickly respond when the same song is played next time, or lyric information is automatically deleted after the cache time is exceeded, and the memory of the memory can be reasonably and efficiently utilized.
Step S103, generating a scene mode corresponding to the song according to the characteristics of the audio data and the lyrics.
In the embodiment, the content of the temperament, tone, rhythm and lyrics in the audio data characteristics of the part to be played in the song is analyzed; determining the pitch, emotional color and frequency of the music to be played by identifying the temperament, tone and rhythm of the part of the song to be played, and setting the background colors of different contextual models corresponding to the different pitch and emotional color; for example: the emotional colors can comprise joy, worry, calm and the like, and the background with red and yellow tones is used as a joyous scene mode; blue gray background as a career profile; taking a light green background as a calm scene mode; the contextual model may also be a dynamic video model, for example, if a keyword of "river" or "sea" is recognized in the voice of the sea wave or the lyrics in the audio, a contextual model corresponding to a dynamic video of the sea may be generated; for example, if the place name exists in the lyrics, generating an image or a short video of a landmark building or a station of the corresponding place; for example, a dynamically changing ripple fluctuation video or dynamic image is generated to match the identified tempo and rhythm of the song. And performing multi-language recognition on the text in the lyrics, such as English, Japanese and the like, and generating the corresponding style contextual model.
In addition, the tone of the song can be identified, and a corresponding background color is set; for example, setting the background color of the corresponding contextual model to be pure white or milk white for the music in major key C, setting the background color of the corresponding contextual model to be pure green for the music in major key D, setting the background color of the corresponding contextual model to be light yellow or pure yellow for the music in major key E, and the like; through the recognition of the tone of the music, a corresponding contextual model is generated, and through controlling the background color or the picture of the display screen, the scenes of playing songs are enriched.
Optionally, generating a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics, including:
generating a driving signal of a display screen according to the characteristics of the audio data and the lyrics;
driving a display screen to generate a scene mode corresponding to the song through the driving signal:
wherein the scene mode dynamically changes with the frequency spectrum of the song and the change of the lyrics.
In this embodiment, the sound box device includes a driving chip for driving the color electronic display, and generates a corresponding driving signal and a corresponding contextual model by analyzing and recognizing the characteristics of the audio data and the lyrics. Wherein the state of the scene mode is dynamically changed, and a different scene mode is automatically generated with each sentence or each piece of lyrics of the song.
And step S104, controlling the scene mode, the lyrics and the playing progress of the song to be synchronously displayed.
In the embodiment, according to the playing progress of the audio frequency of the song, the color electronic display screen is controlled to display the contextual model corresponding to the currently played audio frequency part, and meanwhile, the lyrics corresponding to the currently played audio frequency part are displayed; wherein, all lyrics of the song can be displayed on the color electronic display screen, and the lyrics of the currently played part are highlighted in sequence along with the playing progress of the song; partial lyrics of the song can be displayed, the partial lyrics comprise the lyrics of the currently played part, and the displayed partial lyrics are updated along with the playing progress of the song; and only the lyrics of the current sentence can be displayed, and the size or the color of the fonts in the lyrics of the current sentence is sequentially changed according to the playing speed, so that the changing speed of the fonts is matched with the playing rhythm, beat and playing progress.
In addition, the setting of the scene mode may be set for the tone of a song, or may be set for each sentence or each paragraph of lyrics of a song and updated synchronously with the playing progress of the song.
Optionally, the controlling the scene mode, the lyrics and the playing progress of the song to be displayed synchronously includes:
and according to the playing rhythm of the lyrics, highlighting each character in the played lyrics of the current sentence one by one along with the playing rhythm under the corresponding contextual model.
In this embodiment, each lyric may correspond to a corresponding contextual model, and in the corresponding contextual model, the lyrics may be displayed as all lyrics of the song or as local lyrics, where the local lyrics include lyrics corresponding to a currently played audio portion; when the lyrics are displayed, each character in the lyrics can be highlighted one by one according to the playing rhythm of the song, and the size or the color of the font can be changed one by one; the refreshing time of the color corresponding to each word in the lyrics can be set according to the time of the audio frequency staying corresponding to each word, and the refreshing time corresponds to the beat of music playing, so that the lyrics are displayed and played synchronously.
Optionally, the contextual model includes a background color or background animation corresponding to a plurality of song genres when the song is played and the lyrics are displayed.
In this embodiment, the multiple contextual models correspond to different song styles, for example, a dynamic background with a red-yellow hue is used as a happy contextual model; blue gray background as a career profile; taking a light green background as a calm scene mode; through the combination of music and color or animation, the emotional colors such as bright, warm, tension, depression, worry and the like are displayed, the infectivity of music playing is increased, and the user experience is enriched.
Optionally, after the lyric text corresponding to the audio data is searched, the lyric text needs to be calibrated, and whether the searched lyric is the same as the currently played song is determined, so as to ensure the accuracy of the lyric.
Optionally, when the speaker device plays music, the lyrics and the contextual model are synchronously displayed, and meanwhile, the sound of the user can be received through the microphone; the tone of the user's voice and the tone of the played song can be recognized at the same time, and a graphic representation of the tone height is generated to indicate the error between the tone when the user sings and the standard tone of the played song, and meanwhile, when the user sings along with the loudspeaker box, the score or evaluation of the user's singing accuracy is given, so that the user can quickly learn to listen to the song according to the lyrics.
It should be noted that, within the technical scope of the present disclosure, other sequencing schemes that can be easily conceived by those skilled in the art should also be within the protection scope of the present disclosure, and detailed description is omitted here.
According to the embodiment, under the networking state, the audio data of the song to be played is acquired from the server; acquiring lyrics matched with the song in the server according to the audio data; generating a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics; controlling the synchronous display of the contextual model, the lyrics and the playing progress of the song; the problem that the lyrics are easy to be out of synchronization with the played music or inconsistent in content when displayed is solved, and the display interface is single; through analyzing the characteristics of the audio data and the content of the lyrics, synchronously displayed lyrics and a corresponding contextual model are generated and synchronously displayed with the lyrics along with the playing progress of the song, thereby enriching the scenes of the song playing; the user can quickly learn to listen to the song according to the lyrics, the immersion and experience of the user can be enhanced when the user enjoys the music, and the user can achieve a better experience effect.
It should be understood that, the sequence numbers of the steps in the foregoing embodiments do not imply an execution sequence, and the execution sequence of each process should be determined by its function and inherent logic, and should not constitute any limitation to the implementation process of the embodiments of the present invention.
Fig. 2 is a schematic diagram of a device for displaying scene lyrics according to an embodiment of the present invention, and for convenience of description, only the parts related to the embodiment of the present invention are shown.
The display device of the scene lyrics comprises:
a first data obtaining unit 21, configured to obtain audio data of a song to be played from a server in a networked state;
a second data obtaining unit 22, configured to obtain lyrics matched with the song in the server according to the audio data;
a driving unit 23, configured to generate a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics;
and the control unit 24 is used for controlling the scene mode, the lyrics and the synchronous display of the playing progress of the song.
According to the embodiment, under the networking state, the audio data of the song to be played is acquired from the server; acquiring lyrics matched with the song in the server according to the audio data; generating a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics; controlling the synchronous display of the contextual model, the lyrics and the playing progress of the song; the problem that the lyrics are easy to be out of synchronization with the played music or inconsistent in content when displayed is solved, and the display interface is single; through analyzing the characteristics of the audio data and the content of the lyrics, synchronously displayed lyrics and a corresponding contextual model are generated and synchronously displayed with the lyrics along with the playing progress of the song, thereby enriching the scenes of the song playing; the user can quickly learn to listen to the song according to the lyrics, the immersion and experience of the user can be enhanced when the user enjoys the music, and the user can achieve a better experience effect.
It will be apparent to those skilled in the art that, for convenience and simplicity of description, the foregoing division of the functional modules is merely illustrated, and in practical applications, the above function distribution may be performed by different functional units and modules as needed, that is, the internal structure of the mobile terminal is divided into different functional units or modules to perform all or part of the above described functions. Each functional module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional modules are only used for distinguishing one functional module from another, and are not used for limiting the protection scope of the application. The specific working process of the module in the mobile terminal may refer to the corresponding process in the foregoing method embodiment, and is not described herein again.
Fig. 3 is a schematic diagram of a sound box apparatus according to an embodiment of the present invention. As shown in fig. 3, the speaker device 3 of this embodiment includes: a processor 30, a memory 31 and a computer program 32 stored in said memory 31 and executable on said processor 30. The processor 30, when executing the computer program 32, implements the steps in the various display method embodiments described above, such as the steps 101 to 104 shown in fig. 1. Alternatively, the processor 30, when executing the computer program 32, implements the functions of the units in the above-described device embodiments, such as the functions of the modules 21 to 24 shown in fig. 2.
Illustratively, the computer program 32 may be partitioned into one or more modules/units that are stored in the memory 31 and executed by the processor 30 to implement the present invention. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, which are used to describe the execution process of the computer program 32 in the loudspeaker device 3. For example, the computer program 32 may be divided into a first data acquisition unit, a second data acquisition unit, a driving unit, and a control unit, each unit having the following specific functions:
the system comprises a first data acquisition unit, a second data acquisition unit and a control unit, wherein the first data acquisition unit is used for acquiring audio data of a song to be played from a server in a networking state;
the second data acquisition unit is used for acquiring lyrics matched with the song in the server according to the audio data;
the driving unit is used for generating a scene mode corresponding to the song according to the characteristics of the audio data and the lyrics;
and the control unit is used for controlling the scene mode, the lyrics and the synchronous display of the playing progress of the song.
The speaker device 3 may include, but is not limited to, a processor 30 and a memory 31. Those skilled in the art will appreciate that fig. 3 is merely an example of the enclosure device 3, and does not constitute a limitation of the enclosure device 3, and may include more or less components than those shown, or combine some components, or different components, for example, the enclosure device 3 may further include an input-output device, a network access device, a bus, etc.
The Processor 30 may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The memory 31 may be an internal storage unit of the sound box device 3, such as a hard disk or a memory of the sound box device 3. The memory 31 may also be an external storage device of the speaker device 3, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), and the like, provided on the speaker device 3. Further, the memory 31 may also include both an internal storage unit and an external storage device of the speaker device 3. The memory 31 is used for storing the computer program and other programs and data required by the loudspeaker device 3. The memory 31 may also be used to temporarily store data that has been output or is to be output.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-mentioned division of the functional units and modules is illustrated, and in practical applications, the above-mentioned function distribution may be performed by different functional units and modules according to needs, that is, the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-mentioned functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present invention. The specific working processes of the units and modules in the system may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus/terminal device and method may be implemented in other ways. For example, the above-described embodiments of the apparatus/terminal device are merely illustrative, and for example, the division of the modules or units is only one logical division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated modules/units, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. Based on such understanding, all or part of the flow of the method according to the embodiments of the present invention may also be implemented by a computer program, which may be stored in a computer-readable storage medium, and when the computer program is executed by a processor, the steps of the method embodiments may be implemented. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, and the like. It should be noted that the computer readable medium may contain other components which may be suitably increased or decreased as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media which may not include electrical carrier signals and telecommunications signals in accordance with legislation and patent practice.
The above-mentioned embodiments are only used for illustrating the technical solutions of the present invention, and not for limiting the same; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; such modifications and substitutions do not substantially depart from the spirit and scope of the embodiments of the present invention, and are intended to be included within the scope of the present invention.

Claims (10)

1. A method for displaying scene lyrics is applied to sound box equipment, and comprises the following steps:
under the networking state, acquiring audio data of a song to be played from a server;
acquiring lyrics matched with the song in the server according to the audio data;
generating a contextual model corresponding to the song according to the characteristics of the audio data and the lyrics;
and controlling the scene mode, the lyrics and the playing progress of the song to be synchronously displayed.
2. The method for displaying scene lyrics of claim 1, wherein before acquiring the audio data of the song to be played from the server in a networked state, the method comprises:
receiving a networking trigger signal through a display screen;
after the network connection is finished, searching the audio data of the song to be played according to the trigger signal for playing the music input by the user.
3. The method of displaying contextual lyrics of claim 1, wherein the audio data comprises spectral feature data, a name, and singer information for the song.
4. The method for displaying scene lyrics of claim 1, wherein the step of obtaining lyrics matched with the song from the audio data obtaining server comprises:
analyzing the audio data to obtain the keywords in the audio data or the frequency spectrum characteristics of the song prelude;
and acquiring and storing lyrics matched with the song according to the keywords or the frequency spectrum characteristics.
5. The method for displaying contextual lyrics according to claim 1, wherein generating a contextual model corresponding to the song based on the characteristics of the audio data and the lyrics comprises:
generating a driving signal of a display screen according to the characteristics of the audio data and the lyrics;
driving a display screen to generate a scene mode corresponding to the song through the driving signal:
wherein the scene mode dynamically changes with the frequency spectrum of the song and the change of the lyrics.
6. The method for displaying contextual lyrics according to claim 1, wherein controlling the contextual model, the lyrics and the playing progress of the song to be displayed in synchronization comprises:
and according to the playing rhythm of the lyrics, highlighting each character in the played lyrics of the current sentence one by one along with the playing rhythm under the corresponding contextual model.
7. The method of displaying contextual lyrics of claim 1, wherein the contextual model comprises a background color or background animation corresponding to a plurality of song genres when the song is played and the lyrics are displayed.
8. A display device for scene lyrics, comprising:
the system comprises a first data acquisition unit, a second data acquisition unit and a control unit, wherein the first data acquisition unit is used for acquiring audio data of a song to be played from a server in a networking state;
the second data acquisition unit is used for acquiring lyrics matched with the song in the server according to the audio data;
the driving unit is used for generating a scene mode corresponding to the song according to the characteristics of the audio data and the lyrics;
and the control unit is used for controlling the scene mode, the lyrics and the synchronous display of the playing progress of the song.
9. Loudspeaker device comprising a memory, a processor and a computer program stored in said memory and executable on said processor, characterized in that said processor implements the steps of the method according to any one of claims 1 to 7 when executing said computer program.
10. A computer-readable storage medium, in which a computer program is stored which, when being executed by a processor, carries out the steps of the method according to any one of claims 1 to 7.
CN201910169089.7A 2019-03-06 2019-03-06 Scene lyric display method and device and sound box equipment Pending CN111666445A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910169089.7A CN111666445A (en) 2019-03-06 2019-03-06 Scene lyric display method and device and sound box equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910169089.7A CN111666445A (en) 2019-03-06 2019-03-06 Scene lyric display method and device and sound box equipment

Publications (1)

Publication Number Publication Date
CN111666445A true CN111666445A (en) 2020-09-15

Family

ID=72381405

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910169089.7A Pending CN111666445A (en) 2019-03-06 2019-03-06 Scene lyric display method and device and sound box equipment

Country Status (1)

Country Link
CN (1) CN111666445A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112423107A (en) * 2020-11-18 2021-02-26 北京字跳网络技术有限公司 Lyric video display method and device, electronic equipment and computer readable medium
CN115134643A (en) * 2021-03-24 2022-09-30 腾讯科技(深圳)有限公司 Bullet screen display method and device for vehicle-mounted terminal, terminal and medium
CN117932110A (en) * 2024-03-20 2024-04-26 深圳市海勤科技有限公司 Lyric automatic processing method, computer equipment and Bluetooth sound equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080110322A1 (en) * 2006-11-13 2008-05-15 Samsung Electronics Co., Ltd. Photo recommendation method using mood of music and system thereof
CN101546587A (en) * 2008-03-25 2009-09-30 国际商业机器公司 Method and device for automatically displaying visual elements in song-playing process
US20120259634A1 (en) * 2011-04-05 2012-10-11 Sony Corporation Music playback device, music playback method, program, and data creation device
CN104882147A (en) * 2015-06-05 2015-09-02 福建星网视易信息系统有限公司 Method, device and system for displaying singing score
CN108986848A (en) * 2018-06-01 2018-12-11 四川斐讯全智信息技术有限公司 Method, system, intelligent sound box and the intelligent terminal of synchronous song information and the lyrics

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080110322A1 (en) * 2006-11-13 2008-05-15 Samsung Electronics Co., Ltd. Photo recommendation method using mood of music and system thereof
CN101546587A (en) * 2008-03-25 2009-09-30 国际商业机器公司 Method and device for automatically displaying visual elements in song-playing process
US20120259634A1 (en) * 2011-04-05 2012-10-11 Sony Corporation Music playback device, music playback method, program, and data creation device
CN104882147A (en) * 2015-06-05 2015-09-02 福建星网视易信息系统有限公司 Method, device and system for displaying singing score
CN108986848A (en) * 2018-06-01 2018-12-11 四川斐讯全智信息技术有限公司 Method, system, intelligent sound box and the intelligent terminal of synchronous song information and the lyrics

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112423107A (en) * 2020-11-18 2021-02-26 北京字跳网络技术有限公司 Lyric video display method and device, electronic equipment and computer readable medium
CN112423107B (en) * 2020-11-18 2022-05-17 北京字跳网络技术有限公司 Lyric video display method and device, electronic equipment and computer readable medium
WO2022105271A1 (en) * 2020-11-18 2022-05-27 北京字跳网络技术有限公司 Method and apparatus for presenting lyrics and video, and electronic device and computer-readable medium
CN115134643A (en) * 2021-03-24 2022-09-30 腾讯科技(深圳)有限公司 Bullet screen display method and device for vehicle-mounted terminal, terminal and medium
CN117932110A (en) * 2024-03-20 2024-04-26 深圳市海勤科技有限公司 Lyric automatic processing method, computer equipment and Bluetooth sound equipment

Similar Documents

Publication Publication Date Title
CN109979497B (en) Song generation method, device and system and data processing and song playing method
CN109543064B (en) Lyric display processing method and device, electronic equipment and computer storage medium
US9753925B2 (en) Systems, methods, and apparatus for generating an audio-visual presentation using characteristics of audio, visual and symbolic media objects
US10229669B2 (en) Apparatus, process, and program for combining speech and audio data
WO2019114516A1 (en) Media information display method and apparatus, storage medium, and electronic apparatus
CN107172485A (en) A kind of method and apparatus for being used to generate short-sighted frequency
WO2018200267A1 (en) Automatic song generation
CN111666445A (en) Scene lyric display method and device and sound box equipment
US11511200B2 (en) Game playing method and system based on a multimedia file
EP2442299B1 (en) Information processing apparatus, information processing method, and program
CN108829751A (en) Method, apparatus, electronic equipment and the storage medium for generating the lyrics, showing the lyrics
CN110782869A (en) Speech synthesis method, apparatus, system and storage medium
CN109144285A (en) A kind of input method and device
WO2023116122A1 (en) Subtitle generation method, electronic device, and computer-readable storage medium
CN108877803A (en) The method and apparatus of information for rendering
CN113538628A (en) Expression package generation method and device, electronic equipment and computer readable storage medium
CN113821189B (en) Audio playing method, device, terminal equipment and storage medium
US8681157B2 (en) Information processing apparatus, program, and information processing method
CN106055671B (en) Multimedia data processing method and equipment thereof
CN114339076A (en) Video shooting method and device, electronic equipment and storage medium
CN111125384B (en) Multimedia answer generation method and device, terminal equipment and storage medium
WO2024078293A1 (en) Image processing method and apparatus, electronic device, and storage medium
JP5344756B2 (en) Information processing apparatus, information processing method, and program
CN114974184A (en) Audio production method and device, terminal equipment and readable storage medium
CN115831080A (en) Method and device for determining audio frequency, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200915