CN117234648A - Interface display method, device, terminal equipment and storage medium - Google Patents

Interface display method, device, terminal equipment and storage medium Download PDF

Info

Publication number
CN117234648A
CN117234648A CN202311199205.2A CN202311199205A CN117234648A CN 117234648 A CN117234648 A CN 117234648A CN 202311199205 A CN202311199205 A CN 202311199205A CN 117234648 A CN117234648 A CN 117234648A
Authority
CN
China
Prior art keywords
picture
word
descriptor
sentence
reverse
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311199205.2A
Other languages
Chinese (zh)
Inventor
池振训
黄逸琛
李先剑
王志杰
胡州华
吴鸿琦
孙天塬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Kugou Computer Technology Co Ltd
Original Assignee
Guangzhou Kugou Computer Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Kugou Computer Technology Co Ltd filed Critical Guangzhou Kugou Computer Technology Co Ltd
Priority to CN202311199205.2A priority Critical patent/CN117234648A/en
Publication of CN117234648A publication Critical patent/CN117234648A/en
Pending legal-status Critical Current

Links

Landscapes

  • User Interface Of Digital Computer (AREA)

Abstract

The embodiment of the application provides an interface display method, an interface display device, terminal equipment and a storage medium, and relates to the technical field of media. The method comprises the following steps: displaying a playing interface of a song, wherein the song comprises a plurality of lyrics; displaying a first picture generated based on a first word and sentence in the process of playing song content corresponding to the first word and sentence in the plurality of words and sentences; and displaying a second picture generated based on a second word and sentence in the song content corresponding to the second word and sentence in the plurality of words, wherein the second word and sentence is different from the first word and sentence, and the second picture is different from the first picture. By adopting the technical scheme provided by the embodiment of the application, the flexibility of content display in the song playing interface can be improved.

Description

Interface display method, device, terminal equipment and storage medium
Technical Field
The embodiment of the application relates to the technical field of media, in particular to an interface display method, an interface display device, terminal equipment and a storage medium.
Background
Nowadays, more and more users enjoy songs through terminal devices such as smartphones.
In the related art, only one cover map and some play controls of a song are displayed in a play interface of the song, and the display content of the play interface is relatively dead.
Disclosure of Invention
The embodiment of the application provides an interface display method, an interface display device, terminal equipment and a storage medium, which can improve the flexibility of content display in a song playing interface. The technical scheme is as follows:
according to an aspect of an embodiment of the present application, there is provided an interface display method, including:
displaying a playing interface of a song, wherein the song comprises a plurality of lyrics;
displaying a first picture generated based on a first word and sentence in the process of playing song content corresponding to the first word and sentence in the plurality of words and sentences;
and displaying a second picture generated based on a second word and sentence in the song content corresponding to the second word and sentence in the plurality of words, wherein the second word and sentence is different from the first word and sentence, and the second picture is different from the first picture.
According to an aspect of an embodiment of the present application, there is provided an interface display apparatus including:
the interface display module is used for displaying a playing interface of songs, wherein the songs comprise a plurality of lyrics;
the picture display module is used for displaying a first picture generated based on a first word and sentence in the process of playing song content corresponding to the first word and sentence in the plurality of words and sentences;
The picture display module is further configured to display a second picture generated based on a second word and sentence in the plurality of words and sentences during playing of song content corresponding to the second word and sentence, where the second word and sentence is different from the first word and sentence, and the second picture is different from the first picture.
According to an aspect of the embodiment of the present application, there is provided a terminal device including a processor and a memory, in which a computer program is stored, the computer program being loaded and executed by the processor to implement the above-mentioned interface display method.
According to an aspect of an embodiment of the present application, there is provided a computer-readable storage medium having stored therein a computer program loaded and executed by a processor to implement the above-described interface display method.
According to an aspect of an embodiment of the present application, there is provided a computer program product loaded and executed by a processor to implement the above-described interface display method.
The technical scheme provided by the embodiment of the application can have the following beneficial effects:
the pictures displayed in the playing interface are different when different words and phrases are played in the playing interface, and the corresponding picture content is required to be displayed according to the currently played lyrics, so that the picture in the playing interface is changed along with the change of the lyrics, and the flexibility of content display in the playing interface of songs is improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the application as claimed.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings required for the description of the embodiments will be briefly described below, and it is apparent that the drawings in the following description are only some embodiments of the present application, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic illustration of an implementation environment for an embodiment of the present application;
FIG. 2 is a flow chart of an interface display method according to an embodiment of the present application;
FIG. 3 is a schematic diagram of a playback interface according to an embodiment of the present application;
FIG. 4 is a schematic diagram of a playback interface according to another embodiment of the present application;
FIG. 5 is a schematic diagram of a playback interface according to another embodiment of the present application;
FIG. 6 is a schematic diagram of a keyword display provided by one embodiment of the present application;
FIG. 7 is a schematic diagram showing keywords provided in another embodiment of the present application;
FIG. 8 is a block diagram of an interface display device provided by one embodiment of the present application;
Fig. 9 is a block diagram of a terminal device according to an embodiment of the present application.
Detailed Description
Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numbers in different drawings refer to the same or similar elements, unless otherwise indicated. The implementations described in the following exemplary examples do not represent all implementations consistent with the application. Rather, they are merely examples of methods consistent with aspects of the application as detailed in the accompanying claims.
Referring to FIG. 1, a schematic diagram of an implementation environment of an embodiment of the present application is shown, which may be implemented as an interface display system. As shown in fig. 1, the system 100 may include: a terminal device 11.
The terminal device 11 has installed and running therein a target application program, such as a client of the target application program. Optionally, the client has a user account logged in. The terminal device is an electronic device with data computing, processing and storage capabilities. The terminal device may be an electronic device such as a smart phone, a tablet computer, a PC (Personal Computer ), a wearable device, or an intelligent robot, which is not limited in the embodiment of the present application. The target application may be an audio-class application, such as a song playing application, an audio live application, a K song application, a music production application, etc., which is not particularly limited in the embodiments of the present application. The target application may also be any application with interface display functionality, such as a social application, a payment application, a video application, a shopping application, a news application, a gaming application, and the like. The method provided by the embodiment of the present application may be that the execution subject of each step is the terminal device 11, such as a client running in the terminal device 11.
In some embodiments, the system 100 further includes a server 12, where the server 12 establishes a communication connection (e.g., a network connection) with the terminal device 11, and the server 12 is configured to provide background services for the target application. The server may be an independent physical server, a server cluster or a distributed system formed by a plurality of physical servers, or a cloud server providing cloud computing service. In some embodiments, the main execution body of each step of the method provided in the embodiments of the present application may be the server 12, or may be alternatively executed by the terminal device 11 and the server 12, which is not limited in particular in the embodiments of the present application.
The technical scheme of the application is described and illustrated by the following examples.
Referring to fig. 2, a flowchart of an interface display method according to an embodiment of the application is shown. In the present embodiment, the method is mainly applied to the client described above for illustration. The method may include the following steps (210-230):
step 210, displaying a playing interface of a song, wherein the song comprises a plurality of lyrics.
In some embodiments, a playback interface for songs is displayed in the client. In some embodiments, the song may be a popular song, an ethnic song, an drama, a child song, a song in an opera, or the like, and the type of song is not particularly limited by the embodiments of the present application. In some embodiments, the playback interface may be used to control the playback progress, loudness, sound effect, playback mode (e.g., single-song loop, sequential playback, shuffle playback, etc.) of the song. In some embodiments, various controls related to playing, related information for the song, comments for the song, etc. may be displayed in the play interface for the song. The related information of the song may include, among other things, the name of the song, the producer, the singer, the time of release, the album being included, the type of song, etc.
In some embodiments, lyrics of the song may also be displayed in the play interface. In some embodiments, the playing interface may display only the words and sentences corresponding to the current playing progress each time during the song playing process, that is, only one sentence of lyrics corresponding to the current playing progress is displayed in the playing interface. In some embodiments, the playing interface may display the word corresponding to the current playing progress and the word similar to the word each time during the song playing process. For example, the playing interface may display each time a word corresponding to the current playing progress and a lyrics of a sentence after the word in the song playing process. For another example, the playing interface may display the word corresponding to the current playing progress and the lyrics of the b sentences preceding the word each time during the song playing process. For another example, the playing interface may display each time a word corresponding to the current playing progress, b words preceding the word, and a words following the word in the song playing process. Wherein, the values of a and b can be the same or different.
Wherein a may be a positive integer such as 1, 2, 3, or 4, and b may be a positive integer such as 1, 2, 3, or 4. The specific values of a and b may be set by a person skilled in the relevant art according to the actual situation, which is not particularly limited in the embodiment of the present application.
In some embodiments, as the progress of song play progresses, the lyrics displayed in the play interface change. For example, lyrics may be displayed along the playing interface in a landscape or portrait scroll. For example, the played lyrics may disappear in the playing interface through the effects of fade-out, sliding, etc., and the lyrics to be played or played may be displayed in the playing interface through the effects of fade-in, sliding, etc.
In some embodiments, the words corresponding to the current playing progress may be highlighted compared to other words displayed in the playing interface that have not yet been played. For example, the display size of the words and sentences corresponding to the current playing progress may be larger than the display size of other words and sentences; for another example, the color of the word corresponding to the current playing progress is different from the color of the other words.
Step 220, displaying the first picture generated based on the first phrase in the process of playing the song content corresponding to the first phrase in the plurality of phrases.
In some embodiments, a first picture generated based on a first phrase is displayed during playing of song content corresponding to the first phrase in the plurality of phrases of lyrics. The first picture may be a background picture of a playing interface of the song, or may be a song map corresponding to the first phrase. In some embodiments, various playback-related controls displayed in the playback interface, related information for songs, comments for songs, and the like are displayed outside or at an upper layer of the background picture.
In some embodiments, the semantic context of the first phrase is obtained by neural network model analysis. In some embodiments, the neural network model may be a large language model. In some embodiments, the picture corresponding to the first word and sentence, i.e., the first picture, is generated by using an AI (Artificial Intelligence ) model in combination with the semantic context of the first word and sentence, the mood and emotion of the whole song.
In step 230, during the process of playing the song content corresponding to the second word in the plurality of sentences, a second picture generated based on the second word, the second word being different from the first word, and the second picture being different from the first picture is displayed.
In some embodiments, a second picture generated based on a second phrase is displayed during the playing of song content corresponding to the first phrase in the plurality of phrases of lyrics. The second picture may be a background picture of the playing interface of the song, or may be a song map corresponding to the second phrase.
In some embodiments, the semantic information of the second phrase is obtained by neural network model analysis. In some embodiments, the picture corresponding to the second word and sentence, that is, the second picture, is generated by adopting the AI model in combination with the semantic information of the second word and sentence, the mood and emotion of the whole song.
In some embodiments, the first phrase and the second phrase are two adjacent phrases of lyrics in the song. In the case of switching the first expression to the second expression, switching from displaying the first picture to displaying the second picture. In some embodiments, after the playing progress of the song is switched from one lyric to the next lyric, the playing interface is also switched from displaying one picture to displaying the picture corresponding to the next lyric. That is, each time a lyric is switched, the pictures in the playing interface are also switched.
In some embodiments, the corresponding pictures (such as the first picture and the second picture) of each sentence of lyrics are pre-generated and stored, and the song is displayed on the playing interface of the song when being played. In some embodiments, the pictures may be generated by the server, or may be generated by another computer device and sent to the server, where the pictures are saved by the server. In some embodiments, the client as shown in fig. 1 may download these pictures from the server in advance and store the pictures in the terminal device, and then display the pictures on the playing interface of the song when the song is played. Or, the client shown in fig. 1 obtains the pictures from the server and displays the pictures in real time along with the playing progress of the song in the playing interface when the song is played.
In the embodiment, each time a sentence of lyrics is switched, the pictures in the playing interface are switched along with the lyrics, so that the suitability of the pictures and the song content played by the current song is ensured, the interactive content of the playing interface is enriched, and the experience of users in enjoying the song is further improved.
In summary, according to the technical scheme provided by the embodiment of the application, through different pictures displayed in the playing interface when different words and phrases are played in the playing interface, and the corresponding picture content is required to be displayed according to the currently played lyrics, the picture in the playing interface is changed along with the change of the lyrics, and the flexibility of content display in the playing interface of songs is improved.
In the following, taking the first phrase as an example, how to generate a corresponding picture based on lyrics is described. The first term may be referred to in the methods for generating the pictures corresponding to the other terms, which is not described in detail in the present application.
In some possible implementations, the method may further include the steps of:
1.1, acquiring at least one forward description word corresponding to a first word and sentence, wherein the forward description word refers to a word matched with the semantic meaning of the first word and sentence;
1.2, generating a first picture based on at least one forward descriptor.
In some embodiments, the first word and sentence is analyzed through a neural network model, and at least one forward description word corresponding to the first word and sentence is obtained. As shown in fig. 3, for the words of "the possible night sky has a meteor to go forward" and the corresponding forward descriptors may include "the night sky", "the meteor", "the go forward", then the corresponding pictures 13 of the words of the possible night sky may include elements such as the night sky, the meteor, the road, etc. As shown in fig. 4, for the words of "the solo may be the same as the solo on the western city wall", the corresponding forward descriptors may include "the solo" and "the solo", and then the picture 14 corresponding to the words of the solo may include elements such as the solo, lover, etc. As shown in fig. 5, for the words of "the possible Guilin has a fish boat to light in the confusing direction", the corresponding forward descriptors may include "Guilin", "fish boat", "confusing direction", "light", and then the corresponding pictures 15 of the words may include elements such as Gui Lin Shanshui, fish boat, light, etc.
In some embodiments, the forward descriptor may be a word that exists in the first sentence itself, or may be a word that is summarized or extended based on the first sentence. For another example, for lyrics "known to be in the sound of summer", the corresponding forward descriptors may include "known", "summer", "big tree", "sunshine". The two forward descriptors of "known" and "summer" are directly extracted from the sentence of lyrics, but the "big tree" does not appear in the sentence of lyrics, but according to the conventional knowledge, most of the forward descriptors are living on big trees with luxuriant branches, and generally only called in the noon or afternoon when sunlight is the strongest, so that the two forward descriptors of "big tree" and "sunlight" can be reasonably obtained according to the sentence of lyrics. For another example, for the lyrics "cold ice rain is clapped to the face", the corresponding forward descriptors may include "cold", "raining", "sad". The two forward descriptors of 'cold' and 'raining' are directly extracted from the lyrics, while 'sadness' does not appear in the lyrics, but the semantic analysis can understand that the surface semantics of the lyrics are 'ice-cold rainwater on the face', the cold rainwater is very uncomfortable on the face, and the emotion key of the lyrics is obviously relatively sad, so that the 'sadness' can be determined as one forward descriptor corresponding to the lyrics.
In some embodiments, at least one forward descriptor corresponding to the first term is derived based on the first term in combination with the overall lyrics content of the song. For example, if words such as "classmates", "schools" exist in a certain sentence of lyrics, but the key and the theme of the whole song are childhood hours several decades ago in mind, the forward description word corresponding to the sentence of lyrics may also include words such as "eighties", and the scene in the corresponding generated picture should also accord with the characteristics of eighties in the twentieth century.
In some embodiments, the method may further comprise the steps of:
2.1, acquiring at least one reverse descriptor corresponding to a first word and sentence, wherein the reverse descriptor refers to an element to be avoided when a picture is generated;
2.2, generating a first picture based on at least one forward descriptor and at least one reverse descriptor.
In some embodiments, the reverse descriptors corresponding to the first words and sentences are acquired and determined, and then the picture elements corresponding to the reverse descriptors should be avoided as much as possible in the correspondingly generated first picture. In some embodiments, the reverse descriptors may be some elements that easily cause drawing errors in the generated picture. For example, the reverse descriptors may include windows, because the windows reflect shadows of people or objects, but AI models are difficult to correctly show such reflections when generating pictures, and are prone to reflecting pictures that do not conform to real life and physical principles, or directly not show pictures reflected by windows, so that the appearance of picture elements such as windows should be avoided to some extent when generating pictures. For another example, the reverse descriptor may include extra limbs, and if the extra limbs appear on the body of the general class, the picture gives a feeling of tricks, incompatibility and discomfort, so that picture elements such as the extra limbs should be avoided as much as possible in the picture corresponding to the lyrics.
In some embodiments, the reverse descriptors corresponding to different words and phrases may be identical, as may the different words and phrases in a song, and the corresponding direction descriptors are all identical. In some embodiments, there is a word stock of reverse descriptors, and the reverse descriptors corresponding to the songs are all the word stock.
In some embodiments, the word stock may include the following reverse descriptors therein:
windows, canvas frames, carton, 3D (3 Dimensions, three-dimensional), disfigured, bat art, deformed, extra limbs, close up, b & w, weird colors, blurry, duplicate, morbid, and mutilated, out of frame, extra fingers, mutilated handles, poorly drawn hands, poorly drawn face, mutation, bad anatomy, bad handle, light finger light to light to light bad pro-portions, closed face, gross proportions, malformed arms, short legs, extra arms, extra legs, altered handles, fused fingers, too many fingers, long neck, photoshop, video game, tile, poorly drawn feet, bad feet, mutation, cross-eye, body out of frame (body out frame), 3D render (3D render).
In the above embodiment, by determining the forward descriptor and the reverse description corresponding to the words and sentences, it can be determined which contents need to be displayed and which contents need to be displayed as little as possible when generating the corresponding picture, so as to facilitate the generation of the picture which is more in line with the reality logic and the look and feel of the user, and reduce the logic error of the picture contents relative to the real world and the discomfort of the user watching the picture.
In some embodiments, at least one forward descriptor corresponds to a forward weight, at least one reverse descriptor corresponds to a reverse weight, the forward weight is used for indicating an importance degree of the forward descriptor when generating a picture, and the reverse weight is used for indicating a degree of avoidance of the reverse descriptor when generating the picture. In some embodiments, the first picture is generated based on the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor.
In some embodiments, each of the forward and reverse descriptors has a respective weight. The larger the weight corresponding to the forward direction descriptor is, the higher the importance of the forward direction descriptor when the first picture is generated is indicated; the smaller the weight corresponding to the forward descriptor, the lower the importance of the forward descriptor in generating the first picture. In some embodiments, the forward weights of the respective forward descriptors corresponding to the first phrase are generated and determined by the neural network model from the overall content of the first phrase and song. The larger the weight corresponding to the reverse descriptor is, the more the reverse descriptor needs to be avoided when the first picture is generated, and the lower the probability of representing the reverse descriptor in the first picture is; the smaller the weight corresponding to the reverse descriptor, the smaller the degree that the reverse descriptor needs to be avoided when generating the first picture, and the higher the probability that the reverse descriptor is represented in the first picture.
In some embodiments, generating the first picture based on the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor may include the steps of:
3.1, superposing the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor to obtain final weights respectively corresponding to the at least one forward descriptor and the at least one reverse descriptor;
and 3.2, generating a first picture based on the final weights respectively corresponding to the at least one forward descriptor and the at least one reverse descriptor.
In some embodiments, if there is a first descriptor corresponding to a first sentence, which is both a forward descriptor and a reverse descriptor, that is, the first descriptor has both a forward weight and a reverse weight, the forward weight and the reverse weight of the first descriptor are superimposed, so as to obtain a final weight corresponding to the first descriptor.
In some embodiments, if the absolute value of the forward weight corresponding to the first descriptor is greater than the absolute value of the reverse weight corresponding to the first descriptor, the first descriptor is finally determined to be the forward descriptor, the final weight corresponding to the first descriptor is the forward weight, the first descriptor still has a high probability of being displayed in the first picture, and the final weight corresponding to the first descriptor is less than the absolute value of the forward weight of the first descriptor. In some embodiments, if the absolute value of the forward weight corresponding to the first descriptor is greater than the absolute value of the reverse weight corresponding to the first descriptor, the forward weight and the reverse weight of the first descriptor are superimposed, which means that the absolute value of the forward weight is subtracted from the absolute value of the reverse weight. For example, if the forward weight corresponding to the first descriptor is 2 and the reverse weight corresponding to the first descriptor is 1, the final weight corresponding to the first descriptor (i.e., the forward weight) is 1, i.e., the final weight corresponding to the first descriptor (i.e., the forward weight) =2-1=1.
In some embodiments, if the absolute value of the forward weight corresponding to the first descriptor is smaller than the absolute value of the reverse weight corresponding to the first descriptor, the first descriptor is finally determined to be the reverse descriptor, the final weight corresponding to the first descriptor is the reverse weight, the picture element corresponding to the first descriptor needs to be prevented from appearing in the first picture as much as possible, and the final weight corresponding to the first descriptor is smaller than the absolute value of the reverse weight of the first descriptor. In some embodiments, if the absolute value of the forward weight corresponding to the first descriptor is smaller than the absolute value of the reverse weight corresponding to the first descriptor, the forward weight and the reverse weight of the first descriptor are superimposed, which means that the absolute value of the forward weight is subtracted from the absolute value of the reverse weight. For example, if the forward weight corresponding to the first descriptor is 1 and the reverse weight corresponding to the first descriptor is 3, the final weight corresponding to the first descriptor (i.e., the reverse weight) is 2, i.e., the final weight corresponding to the first descriptor (i.e., the reverse weight) =3-1=2.
In some embodiments, if the absolute value of the forward weight corresponding to the first descriptor is equal to the absolute value of the reverse weight corresponding to the first descriptor, the final weight corresponding to the first descriptor is 0. In some embodiments, if the absolute value of the forward weight corresponding to the first descriptor is equal to the absolute value of the reverse weight corresponding to the first descriptor, the forward weight and the reverse weight of the first descriptor are superimposed, which may refer to subtracting the absolute value of the forward weight from the absolute value of the reverse weight or subtracting the absolute value of the reverse weight from the absolute value of the forward weight. For example, if the forward weight corresponding to the first descriptor is 2, the reverse weight corresponding to the first descriptor is 2, so as to obtain a final weight corresponding to the first descriptor of 0, i.e. a final weight corresponding to the first descriptor=2-2=0. In this case, the first descriptor may be considered as neither a forward descriptor nor a reverse descriptor.
In some embodiments, if the forward descriptor does not belong to the reverse descriptor, the final weight of the forward descriptor is the forward weight corresponding to the forward descriptor; if the reverse descriptor does not belong to the forward descriptor, the final weight of the reverse descriptor is the reverse weight corresponding to the reverse descriptor.
In some embodiments, the first picture is generated according to the final weights corresponding to the respective descriptors. If the final weight of a certain descriptor is a forward weight, the picture element corresponding to the descriptor needs to appear in the first picture as much as possible; if the final weight of a certain descriptor is the inverse weight, the picture element corresponding to the descriptor needs to be avoided from appearing in the first picture as much as possible.
In the implementation manner, the forward description words positively related to the words and sentences are obtained by extracting the semantic information of the words and sentences, and the pictures corresponding to the words and sentences are generated according to the forward description words, so that the generated pictures are matched with the meaning of the corresponding words and sentences as much as possible, and the matching property of the lyrics and the corresponding pictures is improved. And meanwhile, the user experience when the user enjoys songs is improved.
In some possible implementations, there are at least two semantically opposite words in at least one forward description word, where semantically opposite words refer to words that cannot occur in the same scene at the same time.
In some embodiments, generating the first picture based on the at least one forward descriptor may include the steps of:
4.1, respectively generating at least two opposite pictures based on the at least two semantically opposite words, wherein each opposite picture is generated based on one semantically opposite word and other forward description words which are not opposite to the one semantically opposite word;
and 4.2, splicing at least two opposite pictures to generate a first picture.
In some embodiments, at least two opposite words exist in at least one forward descriptor corresponding to the first word and sentence, and obviously, picture elements corresponding to different opposite words cannot be simultaneously present in the same scene, so at least two opposite pictures can be generated based on each opposite word and other forward descriptors, and picture content corresponding to one opposite word exists in each opposite picture. For example, if the forward descriptors corresponding to the first phrase include "daytime", "night" and "on the road", where "daytime" and "night" are two mutually opposite words, one opposite picture is generated based on "daytime" and "on the road", another opposite picture is generated based on "night" and "on the road", and the two opposite pictures are spliced into the first picture. Therefore, not only is the situation that mutually contradictory or mutually incompatible picture elements cannot appear in the same scene of the pictures corresponding to the words and sentences ensured, but also the forward description words of the words and sentences are reflected in the pictures corresponding to the words and sentences as much as possible, and the complete reflection of key semantics in the words and sentences is ensured.
In some embodiments, the stitching at least two opposite pictures may refer to stitching at least two opposite pictures up and down or stitching left and right, so as to obtain a first picture including at least two opposite pictures, where at least two opposite pictures are displayed simultaneously.
In some embodiments, stitching at least two opposing pictures to generate a first picture may include: splicing at least two opposite pictures into a first picture in a moving picture form, wherein the first picture only displays one opposite picture in each frame; and in the process of playing the first expression, at least two opposite pictures are switched and displayed according to the song rhythm of the first expression.
In some embodiments, stitching at least two opposing pictures refers to stitching at least two opposing pictures into a motion picture, the at least two opposing pictures being grouped in different motion picture frames of the motion picture. At least two opposite pictures are displayed respectively while the first picture is displayed (played). For example, one opponent picture is displayed in the first n frames of the first picture, and another opponent picture is displayed in m frames after the n-th frame of the first picture. That is, at least two opposite pictures are alternately displayed in the first picture. Therefore, not only is the situation that mutually contradictory or mutually incompatible picture elements cannot appear in the same scene of the pictures corresponding to the words and sentences ensured, but also the forward description words of the words and sentences are reflected in the pictures corresponding to the words and sentences as much as possible, and the complete reflection of key semantics in the words and sentences is ensured.
In some embodiments, for a first picture in the form of a moving picture, when switching between displaying opposite pictures, switching is performed in accordance with the cadence of the first phrase. If the rhythm of the first word and sentence is faster, rapidly switching at least two opposite pictures; if the rhythm of the first sentence is faster, the switching process between at least two opposite pictures is slowed down.
In the implementation manner, the corresponding opposite pictures are respectively generated based on the mutually opposite words, and then the mutually opposite pictures are spliced to obtain a new picture, so that the forward description words of the words and sentences are reflected as much as possible, and the complete reflection of the semantics of the words and sentences is ensured.
In some possible implementations, the first forward descriptor of the at least one forward descriptor is a word that appears in the first sentence; and under the condition that the first forward description word is played, starting to display the picture element corresponding to the first forward description word.
In some embodiments, the first forward descriptor may be a word appearing in the first word, and then, when the first word is just played, only other picture elements in the first picture may be displayed, but the picture element corresponding to the first forward descriptor is not displayed temporarily. And when the first forward description word is played, displaying the first forward description word in the first picture. For example, if the first sentence is "you just like a meteor in the night sky," where "meteor" in the lyrics is one of its forward descriptors, then when the lyrics are played, a picture representing the night sky may be displayed first, and when the song is played to the word "meteor", the meteor element is displayed in the picture. In some embodiments, the picture element corresponding to the first forward descriptor may be gradually displayed through a fade-in special effect.
In some embodiments, a first forward descriptor of the at least one forward descriptor is a word that appears in the first sentence; and under the condition that the first forward description word is played, switching the picture element corresponding to the first forward description word from static display to dynamic display.
In some embodiments, if the first forward descriptor is a word appearing in the first word and sentence, then, when the first word and sentence has not yet been played to the first forward descriptor, the picture element corresponding to the first forward descriptor may be statically displayed; and when the first words and sentences are played to the first forward description word, changing the static display picture element corresponding to the first forward description word into dynamic display. For example, if the first phrase is "continuous tear flow", only a static face with tears is displayed on the first picture corresponding to the first phrase when the phrase lyrics are still playing the term "continuous tear flow"; the dynamic process of displaying the tear flow on the face in the first picture is only started when the song plays to the word "tear flow".
In the implementation manner, the picture element is displayed only when the lyrics corresponding to the picture element are played, or the picture element is changed from static display to dynamic display, so that important lyrics in songs are emphasized through visual enhancement, the expressive force of the pictures is improved, and the user experience when the user enjoys the songs is enhanced.
In some possible implementations, the method may further include the steps of:
5.1, acquiring the semantics of the first words and sentences and at least one keyword in the first words and sentences;
6.1, generating special effect information corresponding to at least one keyword respectively based on the semantics of the first word and sentence and the at least one keyword;
and 7.1, displaying at least one keyword according to the special effect information respectively corresponding to the at least one keyword.
In some embodiments, the keywords are words appearing in the first sentence, special effect information corresponding to the keywords is generated based on the semantics of the first sentence and the word senses of the keywords, and the keywords are displayed according to the special effect information. In some embodiments, the keywords are displayed at an upper layer of the first picture, and the keywords are displayed superimposed on the first picture. In some embodiments, each sentence of lyrics corresponds to only one keyword. As shown in fig. 3, the keyword "meteor 16" is displayed on the picture 13 corresponding to the lyric by overlapping the keyword "meteor" which is the keyword extracted by the lyric "the possible night sky has meteor to go forward" according to the fact that you can go forward ". As shown in fig. 4, the words "the solo words" are displayed superimposed on the picture 14 corresponding to the lyrics of the sentence "the solo words of a person on the wall of the city of western security are not distinguished" the corresponding extracted keywords are "the solo words". As shown in fig. 5, the keyword "the possible Guilin has the fishing boat to light you at a lost rate" corresponds to the keyword "at a lost rate", and the keyword "at a lost rate 18" is displayed on the picture 15 corresponding to the lyric in a superimposed manner.
In some embodiments, the keywords are not suddenly displayed in the play interface at one stroke, but rather are displayed by special effects fading in. As shown in fig. 6, the keyword "meteor" may be transitioned from the clutter of fragments shown in state 19 to a first word "stream" by converging a portion of fragments shown in state 20, and then to a second word "star" by converging another portion of fragments shown in state 21, and finally the two words of the keyword "meteor" are displayed completely.
In some embodiments, as shown in fig. 7, the fonts of the keywords may be multiple, and the arrangement may be multiple. In some embodiments, the keywords may be arranged laterally like the keywords 16 in fig. 7, and the keywords may be arranged longitudinally like the keywords 22, 23, 24, 25 in fig. 7, which is not particularly limited in this embodiment of the application. In some embodiments, the fonts of the keywords may be randomly selected for determination.
In some embodiments, a rate of overlap between a display region of a first keyword of the at least one keyword and a key picture element displayed in the first picture is less than or equal to a first threshold.
In some embodiments, if it is ensured that the displayed keywords do not overlap or overlap less with the key picture elements in the picture. The overlapping ratio between the display area of the first keyword and the key picture element displayed in the first picture may refer to: and overlapping the display area of the first keyword and the key picture element displayed in the first picture, wherein the ratio of the first keyword to the key picture element is calculated. In some embodiments, the first keyword is displayed in a solid color region in the first picture, the solid color region referring to a picture region having a single color ratio greater than or equal to the second threshold. Therefore, shielding of key words on key picture elements can be avoided as much as possible, and efficiency and convenience for users to understand picture content are improved.
In some embodiments, a first keyword of the at least one keyword is displayed at an upper layer of the first picture, the first keyword having transparency in an overlapping region with a key picture element in the first picture.
In some embodiments, the transparency of the first keyword in the overlapping area with the key picture element in the first picture means that the first keyword is semi-transparent in the overlapping area with the key picture element in the first picture, and the key picture element at the lower layer can be displayed in a transparent manner, so that shielding of the key word on the key picture element is reduced, and efficiency and convenience in understanding the picture content by a user are improved.
In the implementation manner, the keywords in the lyrics are displayed in a superimposed manner while the picture content corresponding to the lyrics is displayed, so that the understanding efficiency of the user on the corresponding words and sentences is improved, and the user experience of the user when enjoying songs is improved.
The following are examples of the apparatus of the present application that may be used to perform the method embodiments of the present application. For details not disclosed in the embodiments of the apparatus of the present application, please refer to the embodiments of the method of the present application.
Referring to fig. 8, a block diagram of an interface display device according to an embodiment of the application is shown. The device has the function of realizing the interface display method example, and the function can be realized by hardware or can be realized by executing corresponding software by hardware. The device may be the terminal device described above, or may be provided on the terminal device. The apparatus 800 may include: an interface display module 810 and a picture display module 820.
The interface display module 810 is configured to display a playing interface of a song, where the song includes a plurality of lyrics.
The picture display module 820 is configured to display a first picture generated based on a first phrase in the plurality of phrases during playing of song content corresponding to the first phrase.
The picture display module 820 is further configured to display a second picture generated based on a second word and sentence in the plurality of sentences during playing of song content corresponding to the second word and sentence, where the second word and sentence is different from the first word and sentence, and the second picture is different from the first picture.
In some embodiments, the apparatus 800 further comprises: the system comprises a descriptive word acquisition module and a picture generation module.
The descriptor acquisition module is used for acquiring at least one forward descriptor corresponding to the first word and sentence, wherein the forward descriptor refers to a word matched with the meaning of the first word and sentence.
The picture generation module is used for generating the first picture based on the at least one forward descriptor.
In some embodiments, the descriptor obtaining module is further configured to obtain at least one reverse descriptor corresponding to the first word and sentence, where the reverse descriptor refers to an element that needs to be avoided when generating the picture.
The picture generation module further comprises: and a picture generation sub-module.
The picture generation sub-module is used for generating the first picture based on the at least one forward descriptor and the at least one reverse descriptor.
In some embodiments, the at least one forward descriptor corresponds to a forward weight, the at least one reverse descriptor corresponds to a reverse weight, the forward weight is used for indicating the importance degree of the forward descriptor when generating a picture, and the reverse weight is used for indicating the avoidance degree of the reverse descriptor when generating a picture; the picture generation sub-module is configured to generate the first picture based on the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor.
In some embodiments, the picture generation sub-module is configured to:
superposing the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor to obtain final weights respectively corresponding to the at least one forward descriptor and the at least one reverse descriptor;
And generating the first picture based on the final weights respectively corresponding to the at least one forward descriptor and the at least one reverse descriptor.
In some embodiments, at least two semantically opposite words exist in the at least one forward description word, and the semantically opposite words refer to words which cannot simultaneously appear in the same scene; the picture generation sub-module is further used for respectively generating at least two opposite pictures based on the at least two semantically opposite words, and each opposite picture is generated based on one semantically opposite word and other forward description words which are not opposite to the one semantically opposite word.
The picture generation sub-module is further configured to splice the at least two opposite pictures to generate the first picture.
In some embodiments, the picture generation sub-module is configured to stitch the at least two opposite pictures into the first picture in a moving picture form, where the first picture displays only one opposite picture per frame; and in the process of playing the first expression, switching and displaying the at least two opposite pictures according to the song rhythm of the first expression.
In some embodiments, a first forward descriptor of the at least one forward descriptor is a word that appears in the first phrase; the apparatus 800 further comprises: and an element display module.
The element display module is used for starting to display the picture element corresponding to the first forward description word under the condition that the first forward description word is played.
The element display module is further configured to switch, when the first forward description word is played, a picture element corresponding to the first forward description word from static display to dynamic display.
In some embodiments, the apparatus 800 further comprises: the system comprises a keyword acquisition module, an information generation module and a keyword display module.
The keyword acquisition module is used for acquiring the semantics of the first words and sentences and at least one keyword in the first words and sentences.
The information generating module is used for generating special effect information corresponding to the at least one keyword respectively based on the semantics of the first word and sentence and the at least one keyword.
The keyword display module is used for displaying the at least one keyword according to the special effect information respectively corresponding to the at least one keyword.
In some embodiments, a rate of overlap between a display region of a first keyword of the at least one keyword and a key picture element displayed in the first picture is less than or equal to a first threshold; and/or, displaying a first keyword in the at least one keyword on an upper layer of the first picture, wherein the first keyword has transparency in an overlapping area with a key picture element in the first picture.
In some embodiments, the first term and the second term are two adjacent sentences of lyrics in the song; the picture display module is further configured to switch from displaying the first picture to displaying the second picture when the first sentence is switched to the second sentence.
In summary, according to the technical scheme provided by the embodiment of the application, through different pictures displayed in the playing interface when different words and phrases are played in the playing interface, and the corresponding picture content is required to be displayed according to the currently played lyrics, the picture in the playing interface is changed along with the change of the lyrics, and the flexibility of content display in the playing interface of songs is improved.
It should be noted that, in the apparatus provided in the foregoing embodiment, when implementing the functions thereof, only the division of the foregoing functional modules is used as an example, in practical application, the foregoing functional allocation may be implemented by different functional modules, that is, the internal structure of the device is divided into different functional modules, so as to implement all or part of the functions described above. In addition, the apparatus and the method embodiments provided in the foregoing embodiments belong to the same concept, and specific implementation processes of the apparatus and the method embodiments are detailed in the method embodiments and are not repeated herein.
Referring to fig. 9, a block diagram of a terminal device 900 according to an embodiment of the present application is shown. The terminal device 900 may be an electronic device such as a cell phone, tablet computer, game console, electronic book reader, multimedia playing device, wearable device, PC, etc. The terminal device is used for implementing the interface display method provided in the above embodiment. The terminal device may be the terminal device 11 in the implementation environment shown in fig. 1.
Specifically, the present application relates to a method for manufacturing a semiconductor device.
In general, the terminal device 900 includes: a processor 901 and a memory 902.
Processor 901 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and the like. The processor 901 may be implemented in at least one hardware form of DSP (Digital Signal Processing ), FPGA (Field Programmable Gate Array, field programmable gate array), PLA (Programmable Logic Array ). The processor 901 may also include a main processor and a coprocessor, the main processor being a processor for processing data in an awake state, also referred to as a CPU (Central Processing Unit ); a coprocessor is a low-power processor for processing data in a standby state. In some embodiments, the processor 901 may integrate a GPU (Graphics Processing Unit, image processor) for rendering and drawing of content required to be displayed by the display screen. In some embodiments, the processor 901 may also include an AI (Artificial Intelligence ) processor for processing computing operations related to machine learning.
The memory 902 may include one or more computer-readable storage media, which may be non-transitory. The memory 902 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 902 is used to store a computer program and is configured to be executed by one or more processors to implement the interface display methods described above.
In some embodiments, the terminal device 900 may further optionally include: a peripheral interface 903, and at least one peripheral. The processor 901, memory 902, and peripheral interface 903 may be connected by a bus or signal line. The individual peripheral devices may be connected to the peripheral device interface 903 via buses, signal lines, or circuit boards. Specifically, the peripheral device includes: at least one of radio frequency circuitry 904, a display 905, audio circuitry 906, and a power supply 907.
It will be appreciated by those skilled in the art that the structure shown in fig. 9 is not limiting and that more or fewer components than shown may be included or certain components may be combined or a different arrangement of components may be employed.
In an exemplary embodiment, there is also provided a computer readable storage medium having stored therein a computer program which, when executed by a processor, implements the above-described interface display method.
In an exemplary embodiment, a computer program product is also provided, which is loaded and executed by a processor to implement the above-described interface display method.
It should be understood that references herein to "a plurality" are to two or more. "and/or", describes an association relationship of an association object, and indicates that there may be three relationships, for example, a and/or B, and may indicate: a exists alone, A and B exist together, and B exists alone. The character "/" generally indicates that the context-dependent object is an "or" relationship.
The foregoing description of the exemplary embodiments of the application is not intended to limit the application to the particular embodiments disclosed, but on the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the application.

Claims (14)

1. An interface display method, characterized in that the method comprises:
displaying a playing interface of a song, wherein the song comprises a plurality of lyrics;
Displaying a first picture generated based on a first word and sentence in the process of playing song content corresponding to the first word and sentence in the plurality of words and sentences;
and displaying a second picture generated based on a second word and sentence in the song content corresponding to the second word and sentence in the plurality of words, wherein the second word and sentence is different from the first word and sentence, and the second picture is different from the first picture.
2. The method according to claim 1, wherein the method further comprises:
acquiring at least one forward description word corresponding to the first word and sentence, wherein the forward description word refers to a word matched with the semantic meaning of the first word and sentence;
the first picture is generated based on the at least one forward descriptor.
3. The method according to claim 2, wherein the method further comprises:
acquiring at least one reverse descriptor corresponding to the first word and sentence, wherein the reverse descriptor refers to an element to be avoided when a picture is generated;
the generating the first picture based on the at least one forward descriptor includes:
the first picture is generated based on the at least one forward descriptor and the at least one reverse descriptor.
4. A method according to claim 3, wherein the at least one forward descriptor corresponds to a forward weight, the at least one reverse descriptor corresponds to a reverse weight, the forward weight is used for indicating the importance of the forward descriptor when generating a picture, and the reverse weight is used for indicating the avoidance degree of the reverse descriptor when generating a picture;
the generating the first picture based on the at least one forward descriptor and the at least one reverse descriptor includes:
and generating the first picture based on the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor.
5. The method of claim 4, wherein generating the first picture based on the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor comprises:
superposing the forward weights respectively corresponding to the at least one forward descriptor and the reverse weights respectively corresponding to the at least one reverse descriptor to obtain final weights respectively corresponding to the at least one forward descriptor and the at least one reverse descriptor;
And generating the first picture based on the final weights respectively corresponding to the at least one forward descriptor and the at least one reverse descriptor.
6. The method of claim 2, wherein there are at least two semantically contradictory words in the at least one forward description word, the semantically contradictory words being words that cannot occur in the same scene at the same time;
the generating the first picture based on the at least one forward descriptor includes:
generating at least two opponent pictures based on the at least two semantically opponent words respectively, wherein each opponent picture is generated based on one semantically opponent word and other forward description words which are not opponent with the one semantically opponent word;
and splicing the at least two opposite pictures to generate the first picture.
7. The method of claim 6, wherein the stitching the at least two opposing pictures to generate the first picture comprises:
splicing the at least two opposite pictures into the first picture in the form of a moving picture, wherein the first picture only displays one opposite picture in each frame;
and in the process of playing the first expression, switching and displaying the at least two opposite pictures according to the song rhythm of the first expression.
8. The method of claim 2, wherein a first forward descriptor of the at least one forward descriptor is a word that appears in the first sentence;
the method further comprises the steps of:
under the condition that the first forward description word is played, starting to display a picture element corresponding to the first forward description word;
or,
and under the condition that the first forward description word is played, switching the picture element corresponding to the first forward description word from static display to dynamic display.
9. The method according to claim 1, wherein the method further comprises:
acquiring the semantics of the first words and sentences and at least one keyword in the first words and sentences;
generating special effect information corresponding to the at least one keyword respectively based on the semantics of the first word and sentence and the at least one keyword;
and displaying the at least one keyword according to the special effect information respectively corresponding to the at least one keyword.
10. The method of claim 9, wherein the step of determining the position of the substrate comprises,
the overlapping rate between the display area of the first keyword in the at least one keyword and the key picture element displayed in the first picture is smaller than or equal to a first threshold value;
And/or the number of the groups of groups,
a first keyword of the at least one keyword is displayed at an upper layer of the first picture, the first keyword having transparency in an overlapping region with a key picture element in the first picture.
11. The method of claim 1, wherein the first term and the second term are two adjacent words of lyrics in the song;
the method further comprises the steps of:
and switching from displaying the first picture to displaying the second picture when the first sentence is switched to the second sentence.
12. An interface display device, the device comprising:
the interface display module is used for displaying a playing interface of songs, wherein the songs comprise a plurality of lyrics;
the picture display module is used for displaying a first picture generated based on a first word and sentence in the process of playing song content corresponding to the first word and sentence in the plurality of words and sentences;
the picture display module is further configured to display a second picture generated based on a second word and sentence in the plurality of words and sentences during playing of song content corresponding to the second word and sentence, where the second word and sentence is different from the first word and sentence, and the second picture is different from the first picture.
13. A terminal device, characterized in that it comprises a processor and a memory, in which a computer program is stored, which computer program is loaded and executed by the processor to implement the interface display method of any of the preceding claims 1 to 11.
14. A computer readable storage medium, characterized in that the computer readable storage medium has stored therein a computer program, which is loaded and executed by a processor to implement the interface display method of any one of the preceding claims 1 to 11.
CN202311199205.2A 2023-09-15 2023-09-15 Interface display method, device, terminal equipment and storage medium Pending CN117234648A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311199205.2A CN117234648A (en) 2023-09-15 2023-09-15 Interface display method, device, terminal equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311199205.2A CN117234648A (en) 2023-09-15 2023-09-15 Interface display method, device, terminal equipment and storage medium

Publications (1)

Publication Number Publication Date
CN117234648A true CN117234648A (en) 2023-12-15

Family

ID=89094295

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311199205.2A Pending CN117234648A (en) 2023-09-15 2023-09-15 Interface display method, device, terminal equipment and storage medium

Country Status (1)

Country Link
CN (1) CN117234648A (en)

Similar Documents

Publication Publication Date Title
CN109462776B (en) Video special effect adding method and device, terminal equipment and storage medium
CN112087655B (en) Method and device for presenting virtual gift and electronic equipment
JP2021192222A (en) Video image interactive method and apparatus, electronic device, computer readable storage medium, and computer program
CN109121021A (en) A kind of generation method of Video Roundup, device, electronic equipment and storage medium
US20070238520A1 (en) Semantic annotations for virtual objects
US7019742B2 (en) Dynamic 2D imposters of 3D graphic objects
Vernallis The Oxford handbook of sound and image in digital media
CN108846886B (en) AR expression generation method, client, terminal and storage medium
WO2008151421A1 (en) User defined characteristics for inheritance based avatar generation
US20180143741A1 (en) Intelligent graphical feature generation for user content
CN113485617B (en) Animation display method and device, electronic equipment and storage medium
CN113082721B (en) Resource management method and device for application program of integrated game module, electronic equipment and storage medium
CN107358659A (en) More pictures fusion display methods and storage device based on 3D technology
CN114697703B (en) Video data generation method and device, electronic equipment and storage medium
CN112819933A (en) Data processing method and device, electronic equipment and storage medium
CN108614872A (en) Course content methods of exhibiting and device
CN113407766A (en) Visual animation display method and related equipment
CN116668733A (en) Virtual anchor live broadcast system and method and related device
CN117234648A (en) Interface display method, device, terminal equipment and storage medium
CN116017082A (en) Information processing method and electronic equipment
CN114697741B (en) Multimedia information playing control method and related equipment
US9412191B2 (en) Sex selection in inheritance based avatar generation
WO2024077492A1 (en) Playback interface displaying method and apparatus, device, medium and program product
WO2023001115A1 (en) Video generation method, electronic device and medium
Mc Glynn The “Cinematic Promise” of Video Game Music: Stylistic Convergence, Current-Generation Remakes, and the Case of Final Fantasy VII

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination