CN116975330A - Content display method and device, electronic equipment and storage medium - Google Patents

Content display method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN116975330A
CN116975330A CN202310974204.4A CN202310974204A CN116975330A CN 116975330 A CN116975330 A CN 116975330A CN 202310974204 A CN202310974204 A CN 202310974204A CN 116975330 A CN116975330 A CN 116975330A
Authority
CN
China
Prior art keywords
content
target
interaction
multimedia content
character
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310974204.4A
Other languages
Chinese (zh)
Inventor
平思嘉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202310974204.4A priority Critical patent/CN116975330A/en
Publication of CN116975330A publication Critical patent/CN116975330A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/44Browsing; Visualisation therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/435Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/438Presentation of query results

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application relates to the technical field of computers, in particular to a content display method, a content display device, electronic equipment and a storage medium, which are used for improving display diversity and search efficiency of multimedia content. The method comprises the following steps: in response to a triggering operation for the target multimedia content, dynamically displaying avatar information and character description information of at least one target character contained in the target multimedia content in a first display interface; in response to a triggering operation for a target viewing control in the at least one content viewing control, sub-multimedia content associated with the target viewing control is presented in a second presentation interface. When the object searches the target multimedia content, the virtual image information and the character description information of the target character in the target multimedia content are displayed, so that the display diversity of the content is increased, the core of the target multimedia content can be more intuitively known, whether the target multimedia content is interested or not is rapidly judged, and the content searching efficiency is improved.

Description

Content display method and device, electronic equipment and storage medium
Technical Field
The present application relates to the field of computer technologies, and in particular, to a content display method, a device, an electronic apparatus, and a storage medium.
Background
With the development of internet technology, objects can read various books at any time through comprehensive search engines on various terminal devices or special reading Application software (APP).
Before books are read, the object's core appeal is to find interesting books from a huge number of books. In general, in searching for a book, an object searches for a name of the book and determines whether to read the book by searching for basic information of the book presented on a result page.
For example, referring to fig. 1 and 2, in the display interface, basic information of a book, such as score, author, type, introduction, chapter catalog, and the like of the book, is generally presented in text form in a search result page of the book.
However, the basic information of the books is displayed singly in a text form, the form is monotonous, the object can only judge whether the books are interested or not through the text, the interested books are difficult to find quickly, furthermore, the book profile can be known only by repeatedly checking the basic information of the books, a great deal of time is consumed, and the reading desire of the object is difficult to excite.
Disclosure of Invention
The embodiment of the application provides a content display method, a content display device, electronic equipment and a storage medium, which are used for improving the display diversity and the searching efficiency of multimedia content.
The first content display method provided by the embodiment of the application comprises the following steps:
responding to triggering operation aiming at the target multimedia content, and dynamically displaying the virtual image information and the character description information of at least one target character in each candidate character contained in the target multimedia content in a first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
and responding to the triggering operation of a target view control in the at least one content view control, and presenting sub-multimedia content associated with the target view control in a second display interface.
The second content display method provided by the embodiment of the application comprises the following steps:
determining the respective avatar information and character description information of at least one target character in each candidate character contained in the target multimedia content, and feeding back to a client so that the client dynamically displays the at least one avatar information and the at least one character description information in a first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
Receiving a content viewing instruction sent by the client, wherein the content viewing instruction is as follows: the client is generated in response to triggering operation of the target viewing control in the at least one content viewing control;
and returning the sub-multimedia content associated with the target viewing control to the client so that the client presents the associated sub-multimedia content in a second display interface.
The first content display device provided by the embodiment of the application comprises:
the first response unit is used for responding to the triggering operation for the target multimedia content, and dynamically displaying the avatar information and the character description information of at least one target character in each candidate character contained in the target multimedia content in the first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
and the second response unit is used for responding to the triggering operation of the target view control in the at least one content view control, and presenting the sub-multimedia content associated with the target view control in a second display interface.
Optionally, the first display interface further includes: the corresponding background image is obtained by generating the image based on environment description information, and the environment description information is as follows: and extracting information from the sub-multimedia content to which the at least one character description information belongs.
Optionally, the first response unit is further configured to:
and sequentially playing the audio content corresponding to each of the at least one character description information according to a preset sequence.
Optionally, each avatar information is obtained by generating an image based on the avatar description information and character description information of the corresponding target character, where the avatar description information and character description information are: and extracting information from the target multimedia content.
Optionally, the content viewing control includes a presentation sub-control;
the second response unit is specifically configured to:
in response to a selection operation for a target presentation sub-control in at least one presentation sub-control, presenting summary presentation content of sub-multimedia content associated with the target presentation sub-control in the second presentation interface.
Optionally, the content viewing control includes a viewing sub-control;
The second response unit is specifically configured to:
and responding to the selection operation of the target viewing sub-control in at least one viewing sub-control, and presenting the detail display content of the sub-multimedia content associated with the target viewing sub-control in the second display interface.
Optionally, the second response unit is further configured to:
responding to content switching operation triggered based on a preset gesture, and presenting summary display content of related multimedia content corresponding to the related sub-multimedia content in the second display interface, wherein the related multimedia content is: and adjacent sub-multimedia contents with the position of the associated sub-multimedia contents in the target multimedia contents.
Optionally, the summary presentation content includes a summary presentation video of the associated sub-multimedia content;
the second response unit is further configured to:
after the playing time length of the summary display video reaches the preset time length, presenting the summary display video of the related multimedia content corresponding to the related sub-multimedia content in the second display interface in a rolling switching mode, wherein the related multimedia content is as follows: and adjacent sub-multimedia contents with the position of the associated sub-multimedia contents in the target multimedia contents.
Optionally, the arrangement order of the at least one presentation sub-control in the first presentation interface is determined based on the position order of the sub-multimedia content associated with each of the at least one content viewing control in the target multimedia content.
Optionally, the first display interface further includes: at least one interaction control of each interaction role in the candidate roles;
the device further comprises an interaction unit for:
responding to the selected operation of a target interaction control in at least one interaction control, and presenting character introduction information of a target interaction character corresponding to the target interaction control in a character interaction interface;
responding to the operation that a target object transmits the object interaction content after responding to the object interaction content input by the role interaction interface, displaying the object interaction content in the role interaction interface, and transmitting the object interaction content to a server;
and receiving the returned role interaction content of the target interaction role, and presenting the role interaction content in the role interaction interface, wherein the role interaction content is output by a target interaction model trained for the target interaction role.
Optionally, the character interaction interface further includes: the role switching control of each of the at least one interactive role;
the interaction unit is further adapted to:
responding to the target object to the selected operation of the candidate switching control in at least one role switching control, and presenting role introduction information of the interactive roles corresponding to the candidate switching control in the role interaction interface.
Optionally, the character interaction interface further includes: the associated at least one sub-multimedia content of the character interaction content is respectively checked and linked, and the associated at least one sub-multimedia content is screened from all sub-multimedia contents contained in the target multimedia content based on the character interaction content;
the interaction unit is further adapted to:
and responding to the triggering operation of the target object for the target link in at least one view link, and presenting the sub-multimedia content corresponding to the target link in a third display interface.
The second content display apparatus provided by the embodiment of the application includes:
the determining unit is used for determining the virtual image information and the character description information of each of at least one target character in candidate characters contained in the target multimedia content and feeding back the virtual image information and the character description information to the client so that the client dynamically displays the at least one virtual image information and the at least one character description information in a first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
The receiving unit is used for receiving a content viewing instruction sent by the client, wherein the content viewing instruction is as follows: the client is generated in response to triggering operation of the target viewing control in the at least one content viewing control;
and the returning unit is used for returning the sub-multimedia content associated with the target viewing control to the client so that the client presents the associated sub-multimedia content in a second display interface.
Optionally, the first display interface further includes: at least one interaction control of each interaction role in the candidate roles;
the device further comprises an interaction unit for:
receiving an interaction instruction sent by the client, wherein the interaction instruction comprises: the client responds to the selected operation of the target interaction control in at least one interaction control, and the obtained identification information of the target interaction role corresponding to the target interaction control;
screening out a target interaction model of the target interaction role from the candidate interaction models of the at least one interaction role based on the identification information, and returning role introduction information output by the target interaction model to the client;
Receiving object interaction content sent by the client, inputting the object interaction content into the target interaction model to generate interaction content, and obtaining output role interaction content;
and returning the role interaction content to the client so that the client presents the role interaction content in the role interaction interface.
Optionally, the apparatus further comprises a matching unit for:
matching the role interaction content with each sub-multimedia content in the target multimedia content respectively, and associating at least one matched sub-multimedia content with the role interaction content;
the return unit is further adapted to:
and returning the view links of the character interaction content and the matched at least one sub-multimedia content to the client so that the client presents at least one view link in the character interaction interface.
The electronic device provided by the embodiment of the application comprises a processor and a memory, wherein the memory stores a computer program, and when the computer program is executed by the processor, the processor is caused to execute the steps of any one of the content display methods.
An embodiment of the present application provides a computer-readable storage medium including a computer program for causing an electronic device to execute the steps of any one of the content presentation methods described above when the computer program is run on the electronic device.
Embodiments of the present application provide a computer program product comprising a computer program stored in a computer readable storage medium; when the processor of the electronic device reads the computer program from the computer readable storage medium, the processor executes the computer program, so that the electronic device performs the steps of any one of the content presentation methods described above.
The application has the following beneficial effects:
the embodiment of the application provides a content display method, a device, electronic equipment and a storage medium, wherein in response to triggering operation for target multimedia content, in a first display interface, the target multimedia content is dynamically displayed to contain virtual image information and character description information of a target character, so that the diversity of multimedia content display is increased and the core of target multimedia content can be displayed more vividly compared with the form of basic information of the multimedia content only displayed in a text form; further, the first display interface further comprises at least one content viewing control, and when the object wants to view the specific content of the target multimedia content, the sub multimedia content associated with the target viewing control is presented in the second display interface through triggering operation of the target viewing control. Based on the mode, the object can rapidly judge whether the object multimedia content is interested or not through the displayed virtual image information and the character description information, the basic information of the object multimedia content does not need to be checked repeatedly, the content searching efficiency is improved, and furthermore, the reading desire of the object is stimulated, and the activity of the object reading is effectively improved.
Additional features and advantages of the application will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by practice of the application. The objectives and other advantages of the application will be realized and attained by the structure particularly pointed out in the written description and claims thereof as well as the appended drawings.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this specification, illustrate embodiments of the application and together with the description serve to explain the application and do not constitute a limitation on the application. In the drawings:
FIG. 1 is a diagram of a search results page in the related art;
FIG. 2 is a schematic diagram of another search results page in the related art;
FIG. 3 is an alternative schematic diagram of an application scenario in an embodiment of the present application;
FIG. 4 is a flowchart of a content presentation method according to an embodiment of the present application;
FIG. 5A is a diagram of a search interface according to an embodiment of the present application;
FIG. 5B is a diagram of another search interface in an embodiment of the present application;
fig. 6A is a schematic diagram of avatar information in an embodiment of the present application;
Fig. 6B is a schematic diagram of another avatar information in an embodiment of the present application;
FIG. 7A is a schematic diagram of a first display interface according to an embodiment of the application;
FIG. 7B is a schematic diagram of an intermediate search interface in accordance with an embodiment of the present application;
FIG. 8A is a schematic diagram of a content viewing control in an embodiment of the application;
FIG. 8B is a schematic diagram of a second display interface according to an embodiment of the application;
FIG. 8C is a schematic diagram of another second display interface according to an embodiment of the application;
FIG. 8D is a schematic diagram of a background image in an embodiment of the application;
FIG. 9A is a schematic diagram of an image display according to an embodiment of the present application;
FIG. 9B is a diagram of a video dynamic presentation in an embodiment of the application;
FIG. 9C is a schematic diagram of AIGC generated video in an embodiment of the application;
FIG. 10A is a schematic diagram of a show child control in an embodiment of the application;
FIG. 10B is a schematic diagram of a second display interface according to an embodiment of the present application;
FIG. 11A is a schematic diagram of an interactive control in an embodiment of the application;
FIG. 11B is a schematic diagram of a character interaction interface according to an embodiment of the present application;
FIG. 11C is a schematic diagram of another character interaction interface according to an embodiment of the application;
FIG. 12A is a schematic diagram of a character interaction interface according to an embodiment of the present application;
fig. 12B is a schematic diagram of a role switching method in an embodiment of the present application;
FIG. 13A is a schematic diagram of a view link in an embodiment of the application;
FIG. 13B is a schematic diagram of a content presentation method according to an embodiment of the present application;
FIG. 14 is a flowchart illustrating a second method for displaying content according to an embodiment of the present application;
fig. 15 is a schematic structural view of a first content display apparatus according to an embodiment of the present application;
fig. 16 is a schematic structural view of a second content display apparatus according to an embodiment of the present application;
FIG. 17 is a schematic diagram of a hardware configuration of an electronic device to which embodiments of the present application are applied;
fig. 18 is a schematic diagram of a hardware configuration of another electronic device to which the embodiment of the present application is applied.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present application more apparent, the technical solutions of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application, and it is apparent that the described embodiments are some embodiments of the technical solutions of the present application, but not all embodiments. All other embodiments, based on the embodiments described in the present document, which can be obtained by a person skilled in the art without any creative effort, are within the scope of protection of the technical solutions of the present application.
Some of the concepts involved in the embodiments of the present application are described below.
In response to: for representing a condition or state upon which an operation is performed, one or more operations performed may be in real-time or with a set delay when the condition or state upon which the operation is dependent is satisfied; without being specifically described, there is no limitation in the execution sequence of the plurality of operations performed.
Multimedia content search operation: the object performs an operation in searching for the target multimedia content, and the searched keyword may be a name, an author, a type, a keyword, etc. of the target multimedia content, which can generate information directly related to the target multimedia content.
Multimedia content: in a computer system, multimedia refers to a man-machine interactive information communication and transmission medium combining two or more media. The media used include text, pictures, photographs, sounds (including music, speech sounds, special sounds), animations and movies, and interactive functions provided by programs. The target multimedia content refers to the content that the object wants to know on the display interface, and is generally static and incapable of giving visual feeling to the object, such as books, drawing sets, etc., which are not limited in detail herein. In the embodiment of the application, the target multimedia content is mainly taken as a book for illustration.
Sub-multimedia content: in order to ensure the layering and the clarity of the content, the target multimedia content is usually divided into a plurality of parts for the object to read, and the target multimedia content is taken as a book for example, and the sub multimedia content can be a chapter of the book.
Client (Client) or Client: refers to a program corresponding to a server for providing local services to clients. Except for some applications that only run locally, they are typically installed on a common client and need to run in conjunction with the server. After development of the internet, more commonly used clients include web browsers such as those used by the world wide web, email clients when receiving and sending email, and client software for instant messaging. For this type of application, there is a need for a corresponding server and service program in the network to provide corresponding services, such as database service, email service, etc., so that a specific communication connection needs to be established between the client and the server to ensure the normal operation of the application.
Embodiments of the application relate to artificial intelligence (ArtificialIntelligence, AI) and machine learning techniques, designed based on computer vision techniques and machine learning (MachineLearning, ML) in artificial intelligence.
Artificial intelligence (Artificial Intelligence, AI) is the theory, method, technique and application system that uses a digital computer or a machine controlled by a digital computer to simulate, extend and extend human intelligence, sense the environment, acquire knowledge and use the knowledge to obtain optimal results. In other words, artificial intelligence is an integrated technology of computer science that attempts to understand the essence of intelligence and to produce a new intelligent machine that can react in a similar way to human intelligence. Artificial intelligence, i.e. research on design principles and implementation methods of various intelligent machines, enables the machines to have functions of sensing, reasoning and decision.
The artificial intelligence technology is a comprehensive subject, and relates to the technology with wide fields, namely the technology with a hardware level and the technology with a software level. Artificial intelligence infrastructure technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technologies, operation/interaction systems, mechatronics, and the like. The artificial intelligence software technology mainly comprises a computer vision technology, a voice processing technology, a natural language processing technology, machine learning/deep learning and other directions.
With research and advancement of artificial intelligence technology, research and application of artificial intelligence technology is being developed in various fields, such as common smart home, smart wearable devices, virtual assistants, smart speakers, smart marketing, unmanned, automatic driving, unmanned aerial vehicles, robots, smart medical treatment, smart customer service, etc., and it is believed that with the development of technology, artificial intelligence technology will be applied in more fields and with increasing importance value.
Machine learning is a multi-domain interdisciplinary, involving multiple disciplines such as probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory, and the like. It is specially studied how a computer simulates or implements learning behavior of a human to acquire new knowledge or skills, and reorganizes existing knowledge structures to continuously improve own performance.
Machine learning is the core of artificial intelligence, a fundamental approach to letting computers have intelligence, which is applied throughout various areas of artificial intelligence. Machine learning and deep learning typically include techniques such as artificial neural networks, belief networks, reinforcement learning, transfer learning, induction learning, and the like. In the embodiment of the application, when the target multimedia content is displayed, the graphic generation model based on machine learning or deep learning is adopted, and the virtual image information, the background image and the summary display video of each sub multimedia content are generated based on the model, so that the target object can be helped to more intuitively know the core content of the target multimedia content. On the character interaction interface, an interaction model trained for the interaction character is adopted to generate character interaction content of the character, interaction with a target object is achieved, the target object is helped to better understand the multimedia content, and whether the multimedia content is interested in or not can be judged rapidly.
The following briefly describes the design concept of the embodiment of the present application:
with the development of internet technology, the object can read various books at any time through comprehensive search engines on various terminal devices or special reading APP.
Before books are read, the object's core appeal is to find interesting books from a huge number of books. In general, in searching for a book, an object searches for a name of the book and determines whether to read the book by searching for basic information of the book presented on a result page.
For example, referring to fig. 1 and 2, in the display interface, basic information of a book, such as score, author, type, introduction, chapter catalog, and the like of the book, is generally presented in text form in a search result page of the book.
However, the basic information of the books is displayed singly in a text form, the form is monotonous, the object can only judge whether the books are interested or not through the text, the interested books are difficult to find quickly, furthermore, the book profile can be known only by repeatedly checking the basic information of the books, a great deal of time is consumed, and the reading desire of the object is difficult to excite.
In view of this, the embodiments of the present application provide a content display method, apparatus, electronic device, and storage medium, when an object searches for a target multimedia content, in response to a trigger operation for the target multimedia content, in a first display interface, dynamically displaying that the target multimedia content includes avatar information and role description information of a target role, which increases diversity of multimedia content display and can display the core of the target multimedia content more vividly than a form of displaying basic information of the multimedia content only in a text form; further, the first display interface further comprises at least one content viewing control, and when the object wants to view the specific content of the target multimedia content, the sub multimedia content associated with the target viewing control is presented in the second display interface through triggering operation of the target viewing control. Based on the mode, the object can rapidly judge whether the object multimedia content is interested or not through the displayed virtual image information and the character description information, the basic information of the object multimedia content does not need to be checked repeatedly, the content searching efficiency is improved, and furthermore, the reading desire of the object is stimulated, and the activity of the object reading is effectively improved.
The preferred embodiments of the present application will be described below with reference to the accompanying drawings of the specification, it being understood that the preferred embodiments described herein are for illustration and explanation only, and not for limitation of the present application, and embodiments of the present application and features of the embodiments may be combined with each other without conflict.
Fig. 3 is a schematic diagram of an application scenario according to an embodiment of the present application. The application scenario diagram includes two terminal devices 310 and a server 320.
In the embodiment of the application, the terminal equipment comprises, but is not limited to, mobile phones, tablet computers, notebook computers, desktop computers, electronic book readers, intelligent voice interaction equipment, intelligent household appliances, vehicle-mounted terminals and other equipment; the terminal device may be provided with a client related to content presentation, where the client may be software (such as a browser, reading software, etc.), or may be a web page, an applet, etc., and the server may be a background server corresponding to the software or the web page, the applet, etc., or a server specifically used for content presentation. The server may be an independent physical server, a server cluster or a distributed system formed by a plurality of physical servers, or a cloud server providing cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, content delivery networks (Content Delivery Network, CDN), basic cloud computing services such as big data and artificial intelligent platforms, and the like.
It should be noted that, the content displaying method in the embodiment of the present application may be executed by an electronic device, and the electronic device may be a server or a terminal device, that is, the method may be executed by the server or the terminal device separately, or may be executed by both the server and the terminal device together. For example, when the server and the terminal equipment jointly execute, the server determines the virtual image information and the character description information of each candidate character contained in the target multimedia content, and feeds back the virtual image information and the character description information to the client, the client dynamically displays at least one virtual image information and the character description information in the first display interface in response to the triggering operation for the target multimedia content, the client sends a content viewing instruction to the server in response to the triggering operation for the target viewing control in the at least one content viewing control, and the server returns the sub-multimedia content associated with the target viewing control to the client so that the client displays the associated sub-multimedia content in the second display interface.
In an alternative embodiment, the communication between the terminal device and the server may be via a communication network.
In an alternative embodiment, the communication network is a wired network or a wireless network.
It should be noted that, the number of terminal devices and servers shown in fig. 3 is merely illustrative, and the number of terminal devices and servers is not limited in practice, and is not particularly limited in the embodiment of the present application.
In the embodiment of the application, when the number of the servers is multiple, the multiple servers can be formed into a blockchain, and the servers are nodes on the blockchain; the content display method disclosed by the embodiment of the application can be used for storing the related multimedia content, the virtual image information, the role description information, the background image and the like on the blockchain.
In addition, the embodiment of the application can be applied to various scenes, including not only content display scenes, but also scenes such as cloud technology, artificial intelligence, intelligent traffic, auxiliary driving and the like.
The content presentation method provided by the exemplary embodiments of the present application will be described below with reference to the accompanying drawings in conjunction with the above-described application scenarios, and it should be noted that the above-described application scenarios are only shown for the convenience of understanding the spirit and principle of the present application, and the embodiments of the present application are not limited in any way in this respect.
As shown in fig. 4, a flowchart of an implementation of a content presentation method provided by an embodiment of the present application is applied to a client installed on a terminal device, and the specific implementation of the method includes steps S41-S42 as follows:
S41: the client responds to triggering operation aiming at the target multimedia content, and in a first display interface, the virtual image information and the character description information of at least one target character in each candidate character contained in the target multimedia content are dynamically displayed; the first presentation interface further includes at least one content viewing control;
specifically, the triggering operation for the target multimedia content may be a search operation, for example, inputting related information of the target multimedia content in a search interface, where the related information may be a name, an author, a type, a character name, and the like. For example, as shown in fig. 5A, which is a schematic diagram of a search interface in the embodiment of the present application, the object inputs the name "vantage" of the target multimedia content, and clicks the control S51 to form a triggering operation for the target multimedia content.
For another example, as shown in fig. 5B, which is a schematic diagram of another search interface in the embodiment of the present application, the object inputs the role name "Zhang Chen" contained in the target multimedia content, a plurality of choices are presented below the search box, and the object clicks "Zhang Chen vanity" to form a triggering operation for the target multimedia content.
The avatar information of the target character may be a cartoon image or a face modeling image, which is not particularly limited herein, for example, referring to fig. 6A and 6B, respectively, which are schematic diagrams of the avatar information in the embodiment of the present application.
The character description information of the target character may be a word of the target character in the target multimedia content, for example, "first repair, i only smelt the efficacy of two Qi tonifying pellets, which has not been done before, and have long been futured for a long time", or may be an introduction of the target character, for example, "Zhang Chen, which is the only person repairing to the highest level in one year, and certainly becomes the most attention person nowadays.
The avatar information and the character description information may be dynamically displayed in the first display interface in the form of an image or a video, and referring to fig. 7A, which is a schematic diagram of the first display interface in the embodiment of the present application, each candidate character included in the target multimedia content "vanity" is "dust", "Chen Mo", "moon glass", "baili" and "water-dependent", where the target character is "dust", and the corresponding avatar information 71 and the character description information 72 are displayed on the display interface.
In addition, an AI tool can be used for generating the 3D inching effect by using the image containing the avatar information and the character description information, and the effect is dynamically displayed on the first display interface.
The client may directly present the first presentation interface in response to a trigger operation for the target multimedia content, for example, the user inputs "vanity in the search interface", and the client presents the first presentation interface as shown in fig. 7A. Further, since many multimedia contents with duplicate names may be included in the network, the user needs to further select the target multimedia contents after inputting the search keyword. Specifically, the user inputs "vancomic' in the client search interface, the client displays the intermediate search interface shown in fig. 7B, the user clicks S71 to form a trigger operation for the target multimedia content, and the client displays the first display interface shown in fig. 7A.
Optionally, each avatar information is obtained by generating an image based on the avatar description information and character description information of the corresponding target character, and the avatar description information and character description information are obtained by extracting information from the target multimedia content.
Specifically, first, through data processing of the target multimedia content, character descriptors (character description information) and character descriptors (character description information) of the target character are extracted, and then the extracted character descriptors and character descriptors are input as keywords (prompt) into graphic generation software such as an AI drawing tool (midjourn ey) to generate a character figure (avatar information) of the target character.
Further, in order to make a video of a target character, after face images of the target character in different pictures are controlled to be consistent through a seed (seed) command, generating a plurality of different pictures, using an AI tool for generating the video, synthesizing the generated plurality of pictures into a video, and using a plurality of character videos of each target character as character animation materials according to chapters.
The character description information of the target character may be obtained by respectively extracting the speech and the bystander description of the target character in each sub-multimedia content by performing data processing on the target multimedia content.
In order to make the avatar information and the character description information more flexible on the first display interface, a plurality of images containing the avatar information and the character description information can be combined to generate a video, and the video corresponding to the target character is displayed on the first display interface.
In the above, in order to further enhance the user experience of the user on the content, the audio-visual display device may further use a dubbing tool to dub the audio description information according to the gender, age, character and other basic information of the target character, and sequentially play the audio content corresponding to at least one character description information in the first display interface according to the preset sequence. Therefore, the virtual image information and the character description information of the target character are presented, and the sound of the target character can be heard, so that the sound, the picture and the character are synchronous, and a more natural and smooth video effect is displayed.
Each content viewing control in the first display interface is associated with one sub-multimedia content in the target multimedia content, wherein the sub-multimedia content can be in various forms such as characters, images, videos, music and the like, the target multimedia content is taken as a novel example, the sub-multimedia content can be a chapter of the novel, the target multimedia content is taken as a television play example, and the sub-multimedia content can be a set of television plays.
S42: and the client responds to the triggering operation of the target view control in the at least one content view control, and the sub-multimedia content associated with the target view control is presented in the second display interface.
Specifically, the triggering operation of the target view control may be clicking, long pressing, double clicking, and the like, which is not limited herein. Specific content of the sub-multimedia content, such as a chapter content of a novel, an episode of a television series, a minute episode of a movie, and brief introduction content of the sub-multimedia content, for introducing a core episode of the sub-multimedia content, may be presented in the second presentation interface, so that a user may quickly determine whether the user is interested.
Optionally, the content viewing control comprises a showing sub-control and a viewing sub-control, and the client responds to the selected operation of the target viewing sub-control in at least one viewing sub-control, and in the second showing interface, details showing content of sub-multimedia content associated with the target viewing sub-control is shown; and the client responds to the selected operation of the target showing sub-control in the at least one showing sub-control, and presents the summary showing content of the sub-multimedia content associated with the target showing sub-control in the second showing interface.
The detail display content is specific content of the sub-multimedia content, namely, when a user determines that the target multimedia content is interested in through the virtual image information and the role description information displayed on the first display interface, the user can check the specific content of the corresponding sub-multimedia content by clicking the viewing sub-control.
Fig. 8A is a schematic diagram of a viewing sub-control according to an embodiment of the present application, where S81-S84 are all viewing sub-controls, S81 and S82 are both associated with a first chapter of a novel, and S83 and S84 are associated with a second chapter and a third chapter of the novel, respectively. In fig. 8A, taking the target viewing sub-control as an example of S83, clicking S83 to present the content of the second chapter, the object can view the content of the second chapter through the second presentation interface as shown in fig. 8B.
The summary presentation content is used for introducing a core scenario of the sub-multimedia content, and a user can further know the sub-multimedia content through the summary presentation content and judge whether the sub-multimedia content is interested or not, and the summary presentation content can be in the form of characters, images, videos and music, and is described in detail below.
It should be noted that, the first display interface may include only a display sub-control, or may include only a viewing sub-control, or may include both a display sub-control and a viewing sub-control, and fig. 8A illustrates that only a viewing sub-control is included.
In fig. 8B, the second display interface is taken as a separate interface for illustration, and in fact, the second display interface may be the same interface as the first display interface, and the sub-multimedia content is displayed on the first display interface, see fig. 8C in particular.
In the embodiment of the application, when an object searches for the target multimedia content, the trigger operation for the target multimedia content is responded, and the target multimedia content is dynamically displayed in the first display interface, wherein the virtual image information and the character description information of the target character are contained in the target multimedia content, so that the diversity of multimedia content display is increased and the core of the target multimedia content can be displayed more vividly compared with the form of only displaying the basic information of the multimedia content in a text form; further, the first display interface further comprises at least one content viewing control, and when the object wants to view the specific content of the target multimedia content, the sub multimedia content associated with the target viewing control is presented in the second display interface through triggering operation of the target viewing control. Based on the mode, the object can rapidly judge whether the object multimedia content is interested or not through the displayed virtual image information and the character description information, the basic information of the object multimedia content does not need to be checked repeatedly, the content searching efficiency is improved, and furthermore, the reading desire of the object is stimulated, and the activity of the object reading is effectively improved.
Optionally, the first display interface further includes a corresponding background image, where the background image is obtained by generating an image based on environment description information, and the environment description information is obtained by extracting information from sub-multimedia content to which at least one role description information belongs.
When the first display interface presents the virtual image information and the role description information of the target object, corresponding background images can be presented, information extraction is carried out on each sub-multimedia content respectively, the contained description words of the environment are obtained, a picture is generated based on the environment description information based on an AI tool, and each sub-multimedia content corresponds to at least one background image. Fig. 8D is a schematic diagram of a background image according to an embodiment of the present application.
When matching the background image with the avatar information and the character description information, the background image, the avatar image (information) and the character description image (information) corresponding to the same sub-multimedia content may be matched into a group and displayed together on the first display interface, for example, the background image 1, the avatar image 1 and the character description image 1 are combined into a group, the environment description information of the background image 1 is extracted from chapter 3 of the novel, the avatar descriptor and the character descriptor of the avatar image 1 are extracted from chapter 3 of the novel, and the speech included in the character description image 1 is also extracted from chapter 3 of the novel.
In order to increase the diversity of the presentation, it is also possible to match the background image, the avatar image, and the character description image corresponding to different sub-multimedia contents into one group, for example, the background image 2, the avatar image 2, and the character description image 2 are one group, the environment description information of the background image 2 is extracted from chapter 4 of the novel, the avatar descriptor and the character descriptor of the avatar image 2 are extracted from chapter 5 of the novel, and the station word included in the character description image 3 is also extracted from chapter 6 of the novel.
For the above-mentioned background image, avatar image, and character description image, they may be combined together as an image for static presentation when presented in the first presentation interface, see fig. 9A. The multiple images can be combined together to form a dynamic video, the video is displayed in the first display interface, the change of the video is shown in fig. 9B, taking a key picture of a segment as an example, a picture 1, a picture 2, a picture 3 and a picture 4, and a picture core element consists of characters, the lines of the characters, background images, audio content corresponding to the lines and background sound effects. The video picture is the dialogue with different roles as the main form, and the picture switches the roles of the dialogue in turn. In the dialogue process, characters and background pictures generate micro-animation based on original pictures through AI, the characters of the lines of the dialogue gradually type and the animation appears, and sound is synthesized by the sound effect of the corresponding characters.
The video displayed in the first display interface may be a core scenario of the target multimedia content or a latest scenario, so that the target object can intuitively understand the core scenario, and for each sub multimedia content included in the target multimedia content, a corresponding summary display video may also be generated for reference by the target object.
Taking a target multimedia content as a novel, the generation process of the displayed video is introduced:
generating images of all roles of the novel, generating corresponding lines and corresponding dubbing of the roles, generating background pictures through generating artificial intelligence (Artificial Intelligence Generated Content, AIGC), and generating videos through the elements. When the video content is applied to the result header card in the first display interface, the content of the video can be generated by selecting the content of the latest chapter. When applied to chapter videos, a corresponding video is generated for each chapter. As shown in fig. 9C, a schematic diagram of generating video by the AIGC according to an embodiment of the present application may be divided into the following four processes:
A. character image animation generation flow
1. Extracting image descriptors and character descriptors of each role through data processing of the novel text content;
2. the descriptive words are used as the campt and are input into graphic generation software such as midjourn ey and the like;
3. generating a shape image of the character, and generating a plurality of different images after the face images of the character are controlled to be consistent through a seed command;
4. using an AI tool for generating videos, and generating videos from a plurality of pictures of the roles;
5. generating a plurality of character videos for each character according to the chapters and taking the video as character animation materials;
B. The background picture animation generation flow is as follows:
1. extracting the description words of the environment corresponding to each chapter through the data processing of the novel text content;
2. the description words of the scene are used as the prompt and are input into graphic generation software such as midjourn ey and the like;
3. generating scene pictures of each chapter and marking corresponding chapters;
4. using an AI tool for video generation, generating a video from the background map;
5. generating a plurality of background animation materials according to the chapters;
C. the text animation generation flow is as follows:
1. through the data processing of the original text content of the novel, the lines of each role are extracted in sections;
2. matching the lines with the chapters and roles;
3. matching different sounds according to character gender using dubbing tool
D. The flow of synthesizing video:
1. importing the materials generated by the ABC process into a video generation AI tool;
2. the background picture in the B is used as the lowest layer picture layer, and the characters are used as the uppermost layer picture layer;
3. and extracting a background image and a corresponding line corresponding to each chapter according to the chapters. Matching corresponding character images according to the lines;
4. different background animation materials and character animation materials can be combined;
5. the animation effect of the character of the line by character is increased, and the dubbing is matched with the line and displayed synchronously;
6. Videos are generated, with each video marker corresponding to a chapter.
Based on the mode, the novel role image and the scene graph are generated through the AIGC technology, the novel with pure characters can be generated to be more fit with the novel role through the role and the background graph AI tool to generate the animation, and the cost of manually manufacturing the animation is greatly saved; the speech and text of the character are combined through appearing character by character and synchronizing AI dubbing, so that the characters of the sound picture are synchronized, and the effect of more natural and smooth video is displayed; through carrying out layering treatment on the background images, the personas and the line texts, and then carrying out respective animation as original materials and then combining, different background image animations and different persona animations can be combined and matched in a crossing way, more video effects are generated, the types of generated videos are more abundant, and the cost is lower.
Referring to fig. 10A, three presentation sub-controls S1001-S1003 are included on a first presentation interface, a user clicks S1001, and a client presents summary presentation content of chapter 001 in a second presentation interface shown in fig. 10B.
Optionally, the arrangement order of the at least one presentation sub-control in the first presentation interface is determined based on the position order of the sub-multimedia content associated with each of the at least one presentation sub-control in the target multimedia content.
As shown in fig. 10A, the display sub-controls S1001 to S1003 respectively correspond to chapter 001 to 003, so that the display sub-controls are ordered in the first interface in the order from S1001 to S1003 from top to bottom, and the order of the display sub-controls may be from left to right or from bottom to top, which is not particularly limited herein, as long as the positional order of the sub-multimedia content is satisfied.
Optionally, when the second display interface displays the summary display content, the second display interface may switch to other summary display content in the following two manners:
mode 1: the client responds to video switching operation triggered based on a preset gesture, and in a second display interface, summary display content of related multimedia content corresponding to the related sub-multimedia content is presented, wherein the related multimedia content is as follows: adjacent sub-multimedia content to the location of the associated sub-multimedia content in the target multimedia content;
specifically, the preset gesture may be sliding up or sliding down, taking the summary display content as an example of the video, and currently playing the summary display video of chapter 5 as an example, when the user slides on the screen of the terminal device, the client presents the summary display video of chapter 4, and when the user slides down the screen of the terminal device, the client presents the summary display video of chapter 6.
Mode 2: if the summary display content comprises the summary display video of the associated sub-multimedia content, after the playing time length of the summary display video reaches the preset time length, the client displays the summary display video of the related multimedia content corresponding to the associated sub-multimedia content in the second display interface in a rolling switching mode, wherein the related multimedia content is as follows: adjacent sub-multimedia content is adjacent to the location of the associated sub-multimedia content in the target multimedia content.
Specifically, the preset duration may be determined based on a duration of the summary display video, for example, the summary display video is 2 minutes and 25 seconds, then the preset duration may be set to 2 minutes and 26 seconds, after the playing duration reaches the preset duration, the content introduction duration may be considered to be played completely, then the next summary display video may be switched to, taking the summary display video currently playing chapter 5 as an example, at this time, the related multimedia content is chapter 6, and the related multimedia content is automatically switched to the summary display video of chapter 6.
In the first display interface, basic information of characters contained in the target multimedia content, such as character profiles, is often displayed, but the objects cannot directly know the character information based on the mode, and optionally, the first display interface further comprises respective interaction controls of at least one interaction character in each candidate character; the client responds to the selected operation of the target interaction control in at least one interaction control, and role introduction information of a target interaction role corresponding to the target interaction control is presented in a role interaction interface; responding to the object interaction content input by the target object in the role interaction interface, responding to the operation of the target object for transmitting the object interaction content, displaying the object interaction content in the role interaction interface, and transmitting the object interaction content to the server; and receiving the character interaction content of the returned target interaction character, and presenting the character interaction content in a character interaction interface, wherein the character interaction content is output by a target interaction model trained for the target interaction character.
Specifically, the interactive control of each interactive role is displayed in the first display interface, and the user can select one interactive role and start a dialogue on the role interactive interface with the interactive role. For example, referring to fig. 11A, the interactive roles displayed on the first display interface are "Zhang Chen", "moon glass" and "water depending", respectively, and the head of each role is used as its interactive control, and the user clicks on the head of the target interactive role, and enters the role interaction interface to start a dialogue with the target interactive role.
Firstly, character introduction information of the character is presented on a character interaction interface, wherein the character introduction information can be self introduction of a target interaction character, classical speech or evaluation of readers and authors on the character. As shown in fig. 11B, in an embodiment of the present application, a client displays object interactive contents input by a user on a character interactive interface, and sends the object interactive contents to a server, and a target interactive model trained for a target interactive character exists on the server side. The client is presented on a character interaction interface as shown in fig. 11C.
Optionally, the character interaction interface further includes: at least one role switching control of each interactive role; and the client responds to the selected operation of the target object aiming at the candidate switching control in at least one role switching control, and presents role introduction information of the interactive role corresponding to the candidate switching control in the role interaction interface.
Specifically, when a user wants to switch to a conversation with other roles in the process of interacting with a target interactive role, the user can switch without exiting the role interaction interface and directly selecting the role which wants to be interacted, at this time, the role presentation interface presents role introduction information of the interactive role corresponding to the candidate switching control, and when the user inputs object interaction content, the server outputs the role interaction content by using an interaction model trained for the interactive role corresponding to the candidate switching control, thereby achieving the purpose of the conversation between the user and different roles.
Fig. 12A is a schematic diagram of another character interaction interface according to an embodiment of the present application, in which a user clicks a character switching control S1201 to switch to a dialogue with "moon glass", and a client presents the interface shown in fig. 12B and presents character introduction information of "moon glass".
Optionally, the character interaction interface further includes: the associated at least one sub-multimedia content is selected from all sub-multimedia contents contained in the target multimedia content based on the role interaction content; and the client responds to the triggering operation of the target object for the target link in the at least one view link, and the sub-multimedia content corresponding to the target link is presented in the third display interface.
Specifically, the character interaction content may further include some content in the target multimedia content, the character interaction content is matched with each sub-multimedia content, the matched sub-multimedia content is associated with the character interaction content, a viewing link of the associated sub-multimedia content is provided, the user can click on the viewing link, and the jump refers to a viewing interface of the corresponding content.
As shown in fig. 13A, in an embodiment of the present application, when the content of the character answer corresponds to the chapter of the novel, shortcut links S1301 (view link) and S1302 are provided below the dialogue content, and after clicking S1301, the corresponding chapter may be opened directly, as shown in fig. 8B.
The third display interface and the second display interface may be the same interface or different interfaces, and are not particularly limited herein.
As shown in fig. 13B, taking a target multimedia content as a novel example, when a target object wants to search a novel, the novel name "puzzle of universe" is first input in the search interface, and then the generated novel moving picture video is displayed in the head (area 1) of the first display interface to introduce the core scenario of the novel, and the video picture content mainly comprises three parts: (1) an avatar of a principal angle in a novel; (2) a line of a main corner; (3) background map of scene in novel. The target object can primarily learn about the novice through the above content, further, by clicking the display sub-control in the first display interface, the summary display video of the corresponding chapter can be checked, for example, after clicking S1303, the summary display video of chapter 001 is checked in the second display interface, and the object can also click the interactive control in the first display interface to talk with the role in the novice, for example, after clicking S1304, the object can talk with the "green doctor" in the role interactive interface. Based on the above manner, the capability of AI to generate text and pictures is used to innovatively present key content of novels in various forms of text, video and virtual character conversations in the novel search results. Therefore, the method helps the user to quickly know the contents such as core points, character characters and the like of the novels in a more vivid mode, helps the user to judge whether the user is interested in the current novels, improves the efficiency of finding the novels by the user, and provides brand new experience of novels searching for the user.
As shown in fig. 14, a flowchart of implementation of a second content presentation method according to an embodiment of the present application is applied to a server, and the specific implementation of the method includes steps S1401-S1403 as follows:
s1401: the server determines the virtual image information and the character description information of each of the candidate characters contained in the target multimedia content, and feeds back the virtual image information and the character description information of each of the at least one target character to the client so that the client dynamically displays the at least one virtual image information and the at least one character description information in a first display interface;
the server pre-determines the avatar information and the character description information of each character aiming at each character contained in the target multimedia content and feeds back the avatar information and the character description information to the client, and the client can display the avatar information and the character description information of the character when a user searches the target multimedia content.
S1402: the method comprises the steps that a server receives a content viewing instruction sent by a client;
s1403: and the server returns the sub-multimedia content associated with the target viewing control to the client so that the client presents the associated sub-multimedia content in the second display interface.
The first display interface further comprises at least one content viewing control, each content viewing control is associated with one sub-multimedia content in the target multimedia content, and the client responds to a content viewing instruction generated for triggering operation of the target viewing control in the at least one content viewing control and sends the content viewing instruction to the server. The server sends the sub-multimedia content associated with the target viewing control pair to the client for viewing by the user, and the detailed description is omitted herein.
In the embodiment of the application, when an object searches for the target multimedia content, the trigger operation for the target multimedia content is responded, and the target multimedia content is dynamically displayed in the first display interface, wherein the virtual image information and the character description information of the target character are contained in the target multimedia content, so that the diversity of multimedia content display is increased and the core of the target multimedia content can be displayed more vividly compared with the form of only displaying the basic information of the multimedia content in a text form; further, the first display interface further comprises at least one content viewing control, and when the object wants to view the specific content of the target multimedia content, the sub multimedia content associated with the target viewing control is presented in the second display interface through triggering operation of the target viewing control. Based on the mode, the object can rapidly judge whether the object multimedia content is interested or not through the displayed virtual image information and the character description information, the basic information of the object multimedia content does not need to be checked repeatedly, the content searching efficiency is improved, and furthermore, the reading desire of the object is stimulated, and the activity of the object reading is effectively improved.
Optionally, the first display interface further includes respective interaction controls of at least one interaction role in the candidate roles; the server receives an interaction instruction sent by the client, wherein the interaction instruction comprises the following steps: the client responds to the selected operation of the target interaction control in at least one interaction control, and the obtained identification information of the target interaction role corresponding to the target interaction control; screening out target interaction models of target interaction roles from candidate interaction models of at least one interaction role based on the identification information, and returning role introduction information output by the target interaction models to the client; receiving object interaction content sent by a client, inputting the object interaction content into a target interaction model to generate interaction content, and obtaining output role interaction content; and returning the character interaction content to the client so that the client presents the character interaction content in the character interaction interface.
Specifically, the server trains a corresponding candidate interaction model for each interaction role, is used for talking with a user, and after the user selects one target role to talk, the client sends an interaction instruction to the server, and the server determines which interaction model to use for talking based on the interaction instruction. For example, there are an interaction model 1 corresponding to "Zhang Chen", an interaction model 2 corresponding to "Chen Mo", an interaction model 3 corresponding to "moon glass", an interaction model 4 corresponding to "baili", and an interaction model 5 corresponding to "water-dependent", and "Zhang Chen" selected by the user, the interaction instruction includes identification information 001 of "Zhang Chen", and the server determines the target interaction model as the interaction model 1 based on the identification information, and outputs character interaction content by adopting the interaction model 1.
Taking a target multimedia content as a novel example, the following describes a training process of a candidate interaction model corresponding to an interaction role, wherein the training process comprises the following two steps:
method 1:
(1) The number and name of the novel main angles are obtained from the existing fields of the novel database. Three different models of ABC are trained independently, provided that ABC three principal angles are provided. The next step takes training model a as an example;
(2) Determining a data set: the composition of the data set is composed of the relevant parts (including sentences, behavioral and psychological activities and the like which are said by the role A) in the novel text A, the content discussion about the role A in the network and comments which are published by novel authors about the role A;
(3) Creating a character feature tag comprising: in order for the model to simulate the kissing, thinking and knowledge of character a when replying, define a set of labels and attributes for the character to evoke its specific personality and characteristics in the dataset;
(4) Marking according to knowledge quantity of the roles: due to the difference of knowledge reserves of different roles, the data needs to be marked according to the background of the roles and the knowledge reserve degree. It is possible to leave the data intact to record the knowledge and insight of the character into a particular topic and to indicate this directly using markers. For example, for a character of a doctor, if the content related to the disease needs to be marked, the marks such as 'patient symptom', 'disease detection', and the like can be used for describing the range of knowledge reserves, which helps the model to better know the speaking characteristics of the character and improve the authenticity of the dialogue;
(5) Cleaning and pretreatment: to ensure the effectiveness and quality of the training data, the data was cleaned and pre-processed using a natural language processing tool library similar to NLTK or SpaCy. For example, invalid information, duplicate data, and unnecessary data are removed in the dataset, and the data is subjected to preprocessing such as normalization, word segmentation, and the like. This will help to guarantee the quality of the data set and improve the performance of the model;
(6) Model training: the sequence-to-sequence model (Sequence to Sequence, seq2 Seq) is implemented using a deep learning framework, such as a machine learning platform (TensorFlow or pyrerch). An appropriate loss function and optimizer is selected to train the model. Common loss functions include cross entropy and mean square error. The optimizer may use Adam, random gradient descent (Stochastic Gradient Descent, SGD), or adaptive gradient (Adaptive Gradient, adagard), etc.;
(7) Generating self-introduction of the character: self-introduction (character introduction information) is added in the training data set of step (2), for example, sentences such as "i am character a, very happy to answer your question", etc. Thus, when the model generates an answer, it will automatically add these self-introductory sentences;
(8) Adding self-introduced sentences when generating answers: when using the trained Seq2Seq model, a specific self-introduced sentence can be added when generating the answer, so that the model is self-called role a.
Method 2:
(1) The content such as the novel and the view of the role A is arranged into a training data set. Next, the original text needs to be preprocessed by using the GPT model for training;
(2) A method of transfer learning (Transfer Learning) is required to train the Seq2Seq model by fine tuning a Generative Pre-trained Transformer (GPT) model. The capability of extracting text features by using the GPT model can be utilized, so that the performance of the Seq2Seq model is further improved;
(3) The trained Seq2Seq model needs to be used to generate the answer for role a. The trained model is integrated into the backend so that the user can talk to the character.
In order to fine tune the GPT model, a Decoder (Decoder) is connected to the GPT Transformer network to form a Seq2Seq model. The weight of the GPT model can be used as an initial value, and the model can be fine-tuned by using contents such as the plain text and the perspective of the character a. During the fine tuning process, some skill may be used to improve model convergence speed and performance, such as learning rate decay, dropouts (dropouts), sequence sampling, etc.
The method 1 can better control the training process of the model and meet the requirements. But require greater computing resources. The method 2 is based on GPT for fine tuning, and can train a model faster. Meanwhile, the method ensures that the performance and the reliability are more advantageous, and can be selected according to requirements.
The following describes the flow of a user's dialogue with a character:
(1) Marking the role selection of the user: when a user clicks on the avatar of a character, a conversation is conducted with the character. The user's input is marked as belonging to that character. For example, if the user clicks on the avatar of character a, the user's input is marked as belonging to character a;
(2) Using the role-specific model: the marked user input is processed using the model associated with the particular character. This model is a trained, deep learning model based on the specific manner of speaking that recognizes the character, emotional preferences, and knowledge background. Always calling a model of the role A to answer the user's question until the user switches roles;
(3) Generating answers and using role-specific tokenization techniques: when the model generates an answer, it should answer the question using the character's mood, thought pattern and knowledge background. The role-specific tagging technique is used to ensure that the generated answer is characteristic of a particular role. The technique includes defining specific tags and attributes for the character to further describe the character's personality and characteristics, ensuring that the generated answer is consistent with the character's personality;
(4) Outputting the answers of the roles: the generated answer is output and returned to the user.
Optionally, the server matches the character interaction content with each sub-multimedia content in the target multimedia content, and associates at least one matched sub-multimedia content with the character interaction content; and returning the view links of the character interaction content and the matched at least one sub-multimedia content to the client so that the client presents at least one view link in the character interaction interface.
Taking the target multimedia content as a novel example, when the role interaction content of the target role is mentioned to a certain chapter in the novel, searching the novel text according to the content; when the matched content is retrieved, the chapter position corresponding to the content and the uniform resource locator (uniform resource locator, url) of the chapter in the novel are obtained, namely, the link is checked, the url is displayed in a dialog box of the role interaction interface by the terminal equipment, and the user clicks and then enters the corresponding chapter position in the novel for reading.
Based on the same inventive concept, the embodiment of the application provides a first content display device. As shown in fig. 15, which is a schematic structural diagram of the content display apparatus 1500, may include:
A first response unit 1501 for dynamically displaying avatar information and character description information of each of at least one target character among candidate characters included in the target multimedia content in a first display interface in response to a trigger operation for the target multimedia content; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content in the target multimedia content;
a second response unit 1502 is configured to, in response to a triggering operation for a target view control in the at least one content view control, present, in a second presentation interface, sub-multimedia content associated with the target view control.
Optionally, the first display interface further includes: corresponding background images are obtained by generating images based on environment description information, wherein the environment description information is as follows: and extracting information from the sub-multimedia content to which the at least one character description information belongs.
Optionally, the first response unit 1501 is further configured to:
and sequentially playing the audio content corresponding to each of the at least one character description information according to a preset sequence.
Optionally, each avatar information is obtained by generating an image based on the avatar description information and character description information of the corresponding target character, where the avatar description information and character description information are: and extracting information from the target multimedia content.
Optionally, the content viewing control includes a presentation sub-control;
the second response unit 1502 is specifically configured to:
in response to a selection operation for a target presentation sub-control in the at least one presentation sub-control, presenting summary presentation content of sub-multimedia content associated with the target presentation sub-control in a second presentation interface.
Optionally, the content viewing control includes a viewing sub-control;
the second response unit 1502 is specifically configured to:
and responding to the selected operation of the target viewing sub control in the at least one viewing sub control, and presenting the detail display content of the sub multimedia content associated with the target viewing sub control in the second display interface.
Optionally, the second response unit 1502 is further configured to:
responding to content switching operation triggered based on a preset gesture, and presenting summary display content of related multimedia content corresponding to the related sub-multimedia content in a second display interface, wherein the related multimedia content is as follows: adjacent sub-multimedia content is adjacent to the location of the associated sub-multimedia content in the target multimedia content.
Optionally, the summary presentation content comprises a summary presentation video of the associated sub-multimedia content;
the second response unit 1502 is further configured to:
After the playing time length of the summary display video reaches the preset time length, presenting the summary display video of the related multimedia content corresponding to the related sub-multimedia content in a second display interface in a rolling switching mode, wherein the related multimedia content is as follows: adjacent sub-multimedia content is adjacent to the location of the associated sub-multimedia content in the target multimedia content.
Optionally, the arrangement order of the at least one presentation sub-control in the first presentation interface is determined based on the positional order of the sub-multimedia content associated with each of the at least one content viewing control in the target multimedia content.
Optionally, the first display interface further includes: at least one interaction role in each candidate role is provided with respective interaction controls;
the apparatus further comprises an interaction unit 1503 for:
responding to the selected operation of the target interaction control in at least one interaction control, and presenting the character introduction information of the target interaction character corresponding to the target interaction control in a character interaction interface;
responding to the object interaction content input by the target object in the role interaction interface, responding to the operation of the target object for transmitting the object interaction content, displaying the object interaction content in the role interaction interface, and transmitting the object interaction content to the server;
And receiving the character interaction content of the returned target interaction character, and presenting the character interaction content in a character interaction interface, wherein the character interaction content is output by a target interaction model trained for the target interaction character.
Optionally, the character interaction interface further includes: at least one role switching control of each interactive role;
the interaction unit 1503 is also for:
responding to the selection operation of the target object aiming at least one role switching control, and presenting role introduction information of the interaction roles corresponding to the candidate switching control in a role interaction interface.
Optionally, the character interaction interface further includes: the associated at least one sub-multimedia content is selected from all sub-multimedia contents contained in the target multimedia content based on the role interaction content;
the interaction unit 1503 is also for:
and responding to the triggering operation of the target object for the target link in the at least one view link, and presenting the sub-multimedia content corresponding to the target link in a third display interface.
Based on the same inventive concept, the embodiment of the application provides a second content display device. As shown in fig. 16, which is a schematic structural diagram of the content display apparatus 1600, may include:
A determining unit 1601, configured to determine, from among the candidate characters included in the target multimedia content, avatar information and character description information of each of at least one target character, and feed back the avatar information and the character description information to the client, so that the client dynamically displays the at least one avatar information and the at least one character description information in the first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content in the target multimedia content;
the receiving unit 1602 is configured to receive a content viewing instruction sent by a client, where the content viewing instruction is: the client responds to the triggering operation generation of the target viewing control in the at least one content viewing control;
and the returning unit 1603 is configured to return the sub-multimedia content associated with the target viewing control to the client, so that the client presents the associated sub-multimedia content in the second presentation interface.
Optionally, the first display interface further includes: at least one interaction role in each candidate role is provided with respective interaction controls;
the apparatus further comprises an interaction unit 1604 for:
receiving an interaction instruction sent by a client, wherein the interaction instruction comprises: the client responds to the selected operation of the target interaction control in at least one interaction control, and the obtained identification information of the target interaction role corresponding to the target interaction control;
Screening out target interaction models of target interaction roles from candidate interaction models of at least one interaction role based on the identification information, and returning role introduction information output by the target interaction models to the client;
receiving object interaction content sent by a client, inputting the object interaction content into a target interaction model to generate interaction content, and obtaining output role interaction content;
and returning the character interaction content to the client so that the client presents the character interaction content in the character interaction interface.
Optionally, the apparatus further comprises a matching unit 1605 for:
matching the role interaction content with each sub-multimedia content in the target multimedia content respectively, and associating at least one matched sub-multimedia content with the role interaction content;
the return unit 1603 is also for:
and returning the view links of the character interaction content and the matched at least one sub-multimedia content to the client so that the client presents at least one view link in the character interaction interface.
For convenience of description, the above parts are described as being functionally divided into modules (or units) respectively. Of course, the functions of each module (or unit) may be implemented in the same piece or pieces of software or hardware when implementing the present application.
Those skilled in the art will appreciate that the various aspects of the application may be implemented as a system, method, or program product. Accordingly, aspects of the application may be embodied in the following forms, namely: an entirely hardware embodiment, an entirely software embodiment (including firmware, micro-code, etc.) or an embodiment combining hardware and software aspects may be referred to herein as a "circuit," module "or" system.
The embodiment of the application also provides electronic equipment based on the same conception as the embodiment of the method. In one embodiment, the electronic device may be a server, such as the server shown in FIG. 1. In this embodiment, the electronic device may be configured as shown in fig. 17, including a memory 1701, a communication module 1703, and one or more processors 1702.
A memory 1701 for storing computer programs for execution by the processor 1702. The memory 1701 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, a program required for running an instant messaging function, and the like; the storage data area can store various instant messaging information, operation instruction sets and the like.
The memory 1701 may be a volatile memory (RAM) such as a random-access memory (RAM); the memory 1701 may also be a nonvolatile memory (non-volatile memory), such as a read-only memory, a flash memory (flash memory), a hard disk (HDD) or a Solid State Drive (SSD); or any other medium that can be used to carry or store a desired computer program in the form of instructions or data structures and that can be accessed by a computer, without limitation. The memory 1701 may be a combination of the above.
The processor 1702 may include one or more central processing units (central processing unit, CPU) or digital processing units, or the like. Processor 1702 is configured to implement the content presentation method described above when calling the computer program stored in memory 1701.
The communication module 1703 is used for communicating with a terminal device and other servers.
The specific connection medium between the memory 1701, the communication module 1703 and the processor 1702 is not limited to the above embodiments of the present application. The embodiment of the present application is illustrated in fig. 17 by a bus 1704 between the memory 1701 and the processor 1702, and the bus 1704 is illustrated in fig. 17 by a bold line, and the connection between other components is merely illustrative and not limiting. The bus 1704 may be classified as an address bus, a data bus, a control bus, or the like. For ease of description, only one thick line is depicted in fig. 17, but only one bus or one type of bus is not depicted.
The memory 1701 stores therein a computer storage medium having stored therein computer executable instructions for implementing the content presentation method of the embodiment of the present application. The processor 1702 is configured to perform the content presentation method described above, as shown in fig. 4 or 14.
In another embodiment, the electronic device may also be other electronic devices, such as the terminal device shown in fig. 1. In this embodiment, the structure of the electronic device may include, as shown in fig. 18: communication component 1810, memory 1820, display unit 1830, camera 1840, sensor 1850, audio circuitry 1860, bluetooth module 1870, processor 1880, and the like.
The communication component 1810 is for communicating with a server. In some embodiments, a circuit wireless fidelity (Wireless Fidelity, wiFi) module may be included, where the WiFi module belongs to a short-range wireless transmission technology, and the electronic device may help the user to send and receive information through the WiFi module.
Memory 1820 may be used for storing software programs and data. The processor 1880 performs various functions of the terminal device and data processing by executing software programs or data stored in the memory 1820. Memory 1820 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device. The memory 1820 stores an operating system that enables the terminal device to operate. The memory 1820 may store an operating system and various application programs, and may also store a computer program for executing the content presentation method according to the embodiment of the present application.
The display unit 1830 may also be used to display information input by a user or information provided to the user and a graphical user interface (graphical user interface, GUI) of various menus of the terminal device. In particular, the display unit 1830 may include a display screen 1832 disposed on a front surface of the terminal device. The display 1832 may be configured in the form of a liquid crystal display, light emitting diodes, or the like. The display unit 1830 may be used to display a first presentation interface or the like in the embodiment of the present application.
The display unit 1830 may also be used to receive input numeric or character information, generate signal inputs related to user settings and function control of the terminal device, and in particular, the display unit 1830 may include a touch screen 1831 provided on the front of the terminal device, and may collect touch operations on or near the user, such as clicking buttons, dragging scroll boxes, and the like.
The touch screen 1831 may be covered on the display screen 1832, or the touch screen 1831 may be integrated with the display screen 1832 to implement input and output functions of the terminal device, and the integrated touch screen may be abbreviated as a touch screen. The display unit 1830 may display an application program and corresponding operation steps in the present application.
The camera 1840 may be used to capture still images and the user may comment the image captured by the camera 1840 through the application. The camera 1840 may be one or more. The object generates an optical image through the lens and projects the optical image onto the photosensitive element. The photosensitive element may be a charge coupled device (charge coupled device, CCD) or a Complementary Metal Oxide Semiconductor (CMOS) phototransistor. The photosensitive elements convert the optical signals to electrical signals, which are then passed to a processor 1880 for conversion to digital image signals.
The terminal device may further comprise at least one sensor 1850, such as an acceleration sensor 1851, a distance sensor 1852, a fingerprint sensor 1853, a temperature sensor 1854. The terminal device may also be configured with other sensors such as gyroscopes, barometers, hygrometers, thermometers, infrared sensors, light sensors, motion sensors, and the like.
Audio circuitry 1860, speaker 1861, microphone 1862 may provide an audio interface between a user and the terminal device. The audio circuit 1860 may transmit the received electrical signal converted from audio data to the speaker 1861, and may be converted into a sound signal by the speaker 1861 for output. The terminal device may also be configured with a volume button for adjusting the volume of the sound signal. On the other hand, microphone 1862 converts the collected sound signals into electrical signals, which are received by audio circuit 1860 and converted into audio data, which are output to communication component 1810 for transmission to, for example, another terminal device, or to memory 1820 for further processing.
The bluetooth module 1870 is used for exchanging information with other bluetooth devices having a bluetooth module through a bluetooth protocol. For example, the terminal device may establish a bluetooth connection with a wearable electronic device (e.g., a smart watch) that also has a bluetooth module through the bluetooth module 1870, thereby performing data interaction.
The processor 1880 is a control center of the terminal device, connects various parts of the entire terminal using various interfaces and lines, and performs various functions of the terminal device and processes data by running or executing software programs stored in the memory 1820, and calling data stored in the memory 1820. In some embodiments, the processor 1880 may include one or more processing units; the processor 1880 may also integrate an application processor that primarily processes operating systems, user interfaces, applications, etc., and a baseband processor that primarily processes wireless communications. It will be appreciated that the baseband processor described above may not be integrated into the processor 1880. The processor 1880 of the present application may run an operating system, application programs, user interface displays and touch responses, as well as content presentation methods of embodiments of the present application. In addition, the processor 1880 is coupled to a display unit 1830.
In some possible embodiments, aspects of the content presentation method provided by the present application may also be implemented in the form of a program product comprising a computer program for causing an electronic device to perform the steps in the content presentation method according to the various exemplary embodiments of the application described herein above when the program product is run on the electronic device, e.g. the electronic device may perform the steps as shown in fig. 4 or 14.
The program product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium can be, for example, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium would include the following: an electrical connection having one or more wires, a portable disk, a hard disk, random Access Memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The program product of embodiments of the present application may take the form of a portable compact disc read only memory (CD-ROM) and comprise a computer program and may be run on an electronic device. However, the program product of the present application is not limited thereto, and in this document, a readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with a command execution system, apparatus, or device.
The readable signal medium may comprise a data signal propagated in baseband or as part of a carrier wave in which a readable computer program is embodied. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A readable signal medium may also be any readable medium that is not a readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with a command execution system, apparatus, or device.
A computer program embodied on a readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer programs for performing the operations of the present application may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C++ or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The computer program may execute entirely on the consumer electronic device, partly on the consumer electronic device, as a stand-alone software package, partly on the consumer electronic device and partly on a remote electronic device or entirely on the remote electronic device or server. In the case of remote electronic devices, the remote electronic device may be connected to the consumer electronic device through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external electronic device (e.g., connected through the internet using an internet service provider).
It should be noted that although several units or sub-units of the apparatus are mentioned in the above detailed description, such a division is merely exemplary and not mandatory. Indeed, the features and functions of two or more of the elements described above may be embodied in one element in accordance with embodiments of the present application. Conversely, the features and functions of one unit described above may be further divided into a plurality of units to be embodied.
Furthermore, although the operations of the methods of the present application are depicted in the drawings in a particular order, this is not required to either imply that the operations must be performed in that particular order or that all of the illustrated operations be performed to achieve desirable results. Additionally or alternatively, certain steps may be omitted, multiple steps combined into one step to perform, and/or one step decomposed into multiple steps to perform.
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having a computer-usable computer program embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program commands may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the commands executed by the processor of the computer or other programmable data processing apparatus produce means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program commands may also be stored in a computer readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the commands stored in the computer readable memory produce an article of manufacture including command means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While preferred embodiments of the present application have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. It is therefore intended that the following claims be interpreted as including the preferred embodiments and all such alterations and modifications as fall within the scope of the application.
It will be apparent to those skilled in the art that various modifications and variations can be made to the present application without departing from the spirit or scope of the application. Thus, it is intended that the present application also include such modifications and alterations insofar as they come within the scope of the appended claims or the equivalents thereof.

Claims (20)

1. A method of content presentation, the method comprising:
responding to triggering operation aiming at target multimedia content, and dynamically displaying the virtual image information and the character description information of at least one target character in each candidate character contained in the target multimedia content in a first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
and responding to the triggering operation of a target view control in the at least one content view control, and presenting sub-multimedia content associated with the target view control in a second display interface.
2. The method of claim 1, wherein the first presentation interface further comprises: the corresponding background image is obtained by generating the image based on environment description information, and the environment description information is as follows: and extracting information from the sub-multimedia content to which the at least one character description information belongs.
3. The method of claim 1, wherein dynamically displaying, in the first display interface, avatar information and character description information of each of at least one target character among candidate characters included in the target multimedia content, further comprises:
and sequentially playing the audio content corresponding to each of the at least one character description information according to a preset sequence.
4. The method of claim 1, wherein each avatar information is obtained by performing image generation based on avatar description information and character description information of a corresponding target character, the avatar description information and the character description information being: and extracting information from the target multimedia content.
5. The method of claim 1, wherein the content viewing control comprises a show child control.
The responding to the triggering operation of the target view control in the at least one content view control presents the sub-multimedia content associated with the target view control in a second display interface, and the method comprises the following steps:
in response to a selection operation for a target presentation sub-control in at least one presentation sub-control, presenting summary presentation content of sub-multimedia content associated with the target presentation sub-control in the second presentation interface.
6. The method of claim 1, wherein the content viewing control comprises a view child control.
The responding to the triggering operation of the target view control in the at least one content view control presents the sub-multimedia content associated with the target view control in a second display interface, and the method comprises the following steps:
and responding to the selection operation of the target viewing sub-control in at least one viewing sub-control, and presenting the detail display content of the sub-multimedia content associated with the target viewing sub-control in the second display interface.
7. The method of claim 1, wherein the order of arrangement of the at least one content viewing control in the first presentation interface is determined based on the respective associated sub-multimedia content of the at least one content viewing control, a positional order in the target multimedia content.
8. The method of claim 5, wherein the method further comprises:
responding to content switching operation triggered based on a preset gesture, and presenting summary display content of related multimedia content corresponding to the related sub-multimedia content in the second display interface, wherein the related multimedia content is: and adjacent sub-multimedia contents with the position of the associated sub-multimedia contents in the target multimedia contents.
9. The method of claim 5, wherein the summary presentation content comprises a summary presentation video of the associated sub-multimedia content;
the method further comprises:
after the playing time length of the summary display video reaches the preset time length, presenting the summary display video of the related multimedia content corresponding to the related sub-multimedia content in the second display interface in a rolling switching mode, wherein the related multimedia content is as follows: and adjacent sub-multimedia contents with the position of the associated sub-multimedia contents in the target multimedia contents.
10. The method of any one of claims 1-7, wherein the first presentation interface further comprises: at least one interaction control of each interaction role in the candidate roles;
the method further comprises:
responding to the selected operation of a target interaction control in at least one interaction control, and presenting character introduction information of a target interaction character corresponding to the target interaction control in a character interaction interface;
responding to the operation that a target object transmits the object interaction content after responding to the object interaction content input by the role interaction interface, displaying the object interaction content in the role interaction interface, and transmitting the object interaction content to a server;
And receiving the returned role interaction content of the target interaction role, and presenting the role interaction content in the role interaction interface, wherein the role interaction content is output by a target interaction model trained for the target interaction role.
11. The method of claim 10, wherein the character interaction interface further comprises: the role switching control of each of the at least one interactive role;
the method further comprises:
responding to the target object to the selected operation of the candidate switching control in at least one role switching control, and presenting role introduction information of the interactive roles corresponding to the candidate switching control in the role interaction interface.
12. The method of claim 10, wherein the character interaction interface further comprises: the associated at least one sub-multimedia content of the character interaction content is respectively checked and linked, and the associated at least one sub-multimedia content is screened from all sub-multimedia contents contained in the target multimedia content based on the character interaction content;
the method further comprises:
and responding to the triggering operation of the target object for the target link in at least one view link, and presenting the sub-multimedia content corresponding to the target link in a third display interface.
13. A method of content presentation, the method comprising:
determining the respective avatar information and character description information of at least one target character in each candidate character contained in the target multimedia content, and feeding back to a client so that the client dynamically displays the at least one avatar information and the at least one character description information in a first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
receiving a content viewing instruction sent by the client, wherein the content viewing instruction is as follows: the client is generated in response to triggering operation of the target viewing control in the at least one content viewing control;
and returning the sub-multimedia content associated with the target viewing control to the client so that the client presents the associated sub-multimedia content in a second display interface.
14. The method of claim 13, wherein the first presentation interface further comprises: at least one interaction control of each interaction role in the candidate roles;
The method further comprises:
receiving an interaction instruction sent by the client, wherein the interaction instruction comprises: the client responds to the selected operation of the target interaction control in at least one interaction control, and the obtained identification information of the target interaction role corresponding to the target interaction control;
screening out a target interaction model of the target interaction role from the candidate interaction models of the at least one interaction role based on the identification information, and returning role introduction information output by the target interaction model to the client;
receiving object interaction content sent by the client, inputting the object interaction content into the target interaction model to generate interaction content, and obtaining output role interaction content;
and returning the role interaction content to the client so that the client presents the role interaction content in the role interaction interface.
15. The method of claim 14, further comprising, prior to said returning said character interaction content to said client:
matching the role interaction content with each sub-multimedia content in the target multimedia content respectively, and associating at least one matched sub-multimedia content with the role interaction content;
The step of returning the character interaction content to the client side further comprises the following steps:
and returning the view links of the character interaction content and the matched at least one sub-multimedia content to the client so that the client presents at least one view link in the character interaction interface.
16. A content presentation device, comprising:
the first response unit is used for responding to the triggering operation for the target multimedia content, and dynamically displaying the avatar information and the character description information of at least one target character in each candidate character contained in the target multimedia content in a first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
and the second response unit is used for responding to the triggering operation of the target view control in the at least one content view control, and presenting the sub-multimedia content associated with the target view control in a second display interface.
17. A content presentation device, comprising:
The determining unit is used for determining the virtual image information and the character description information of each of at least one target character in candidate characters contained in the target multimedia content and feeding back the virtual image information and the character description information to the client so that the client dynamically displays the at least one virtual image information and the at least one character description information in a first display interface; the first presentation interface further includes at least one content viewing control; each content viewing control is associated with one sub-multimedia content of the target multimedia content;
the receiving unit is used for receiving a content viewing instruction sent by the client, wherein the content viewing instruction is as follows: the client is generated in response to triggering operation of the target viewing control in the at least one content viewing control;
and the returning unit is used for returning the sub-multimedia content associated with the target viewing control to the client so that the client presents the associated sub-multimedia content in a second display interface.
18. An electronic device comprising a processor and a memory, wherein the memory stores a computer program which, when executed by the processor, causes the processor to perform the steps of the method of any of claims 1 to 15.
19. A computer readable storage medium, characterized in that it comprises a computer program for causing an electronic device to perform the steps of the method of any one of claims 1-15 when said computer program is run on the electronic device.
20. A computer program product comprising a computer program, the computer program being stored on a computer readable storage medium; when the computer program is read from the computer readable storage medium by a processor of an electronic device, the processor executes the computer program, causing the electronic device to perform the steps of the method of any one of claims 1-15.
CN202310974204.4A 2023-08-02 2023-08-02 Content display method and device, electronic equipment and storage medium Pending CN116975330A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310974204.4A CN116975330A (en) 2023-08-02 2023-08-02 Content display method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310974204.4A CN116975330A (en) 2023-08-02 2023-08-02 Content display method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN116975330A true CN116975330A (en) 2023-10-31

Family

ID=88474737

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310974204.4A Pending CN116975330A (en) 2023-08-02 2023-08-02 Content display method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN116975330A (en)

Similar Documents

Publication Publication Date Title
CN111143610B (en) Content recommendation method and device, electronic equipment and storage medium
Park et al. A metaverse: Taxonomy, components, applications, and open challenges
CN109952572B (en) Suggested response based on message decal
JP6967059B2 (en) Methods, devices, servers, computer-readable storage media and computer programs for producing video
US20190303768A1 (en) Community Question Answering-Based Article Recommendation Method, System, and User Device
CN111611436B (en) Label data processing method and device and computer readable storage medium
CN104735468B (en) A kind of method and system that image is synthesized to new video based on semantic analysis
CN107294837A (en) Engaged in the dialogue interactive method and system using virtual robot
CN105204886B (en) A kind of method, user terminal and server activating application program
CN113766299B (en) Video data playing method, device, equipment and medium
JP7240505B2 (en) Voice packet recommendation method, device, electronic device and program
CN115114395B (en) Content retrieval and model training method and device, electronic equipment and storage medium
US20230214423A1 (en) Video generation
CN113569037A (en) Message processing method and device and readable storage medium
CN110019934A (en) Identify the correlation of video
JP7337172B2 (en) Voice packet recommendation method, device, electronic device and program
EP4075411A1 (en) Device and method for providing interactive audience simulation
CN113746875A (en) Voice packet recommendation method, device, equipment and storage medium
CN116977701A (en) Video classification model training method, video classification method and device
CN116955591A (en) Recommendation language generation method, related device and medium for content recommendation
CN115238126A (en) Method, device and equipment for reordering search results and computer storage medium
CN112533050B (en) Video processing method, device, equipment and medium
CN114816180A (en) Content browsing guiding method and device, electronic equipment and storage medium
CN116980665A (en) Video processing method, device, computer equipment, medium and product
CN116975330A (en) Content display method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication