CN109036397B - Method and apparatus for presenting content - Google Patents

Method and apparatus for presenting content Download PDF

Info

Publication number
CN109036397B
CN109036397B CN201810718896.5A CN201810718896A CN109036397B CN 109036397 B CN109036397 B CN 109036397B CN 201810718896 A CN201810718896 A CN 201810718896A CN 109036397 B CN109036397 B CN 109036397B
Authority
CN
China
Prior art keywords
application
presentation mode
content
user
mode template
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810718896.5A
Other languages
Chinese (zh)
Other versions
CN109036397A (en
Inventor
唐欢
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Shanghai Xiaodu Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN201810718896.5A priority Critical patent/CN109036397B/en
Publication of CN109036397A publication Critical patent/CN109036397A/en
Application granted granted Critical
Publication of CN109036397B publication Critical patent/CN109036397B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/133Protocols for remote procedure calls [RPC]
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The embodiment of the application discloses a method and a device for presenting content. One embodiment of the method comprises: in response to receiving a request for creating a content presentation mode for a target application accessing a voice service, searching a candidate presentation mode template corresponding to a data type of content to be presented, which is indicated by the request for creating the content presentation mode, from a preset presentation mode template library, wherein the presentation mode template in the preset presentation mode template library is configured with the corresponding data type; and receiving an instruction of selecting at least one candidate presentation mode template as a content presentation mode template of the target application, and associating the content presentation mode template of the target application with the acquired application identification of the target application. The implementation mode reduces the cost of accessing the voice service by the target application and improves the flexibility of accessing the voice service by the target application.

Description

Method and apparatus for presenting content
Technical Field
The embodiment of the application relates to the technical field of computers, in particular to a method and a device for presenting content.
Background
With the development of information technology, artificial intelligence technology has been widely used in various fields. The intelligent voice service is an important branch of artificial intelligence, and can provide more convenient and more intelligent life service for users by identifying voice signals sent by users, generating response information based on identification results, and converting the response information into voice signals for output through a voice synthesis technology.
In the current voice service technology, an intelligent voice device can access services provided by a voice service provider to realize a voice service function with a personalized function or mode. After accessing the voice service, the device developer can configure the related setting data of each function item individually according to the function requirement of the device developer through the voice service platform.
Disclosure of Invention
The embodiment of the application provides a method and a device for presenting content.
In a first aspect, an embodiment of the present application provides a method for presenting content, including: in response to receiving a request for creating a content presentation mode for a target application accessing a voice service, searching a candidate presentation mode template corresponding to a data type of content to be presented, which is indicated by the request for creating the content presentation mode, from a preset presentation mode template library, wherein the presentation mode template in the preset presentation mode template library is configured with the corresponding data type; and receiving an instruction of selecting at least one candidate presentation mode template as a content presentation mode template of the target application, and associating the content presentation mode template of the target application with the acquired application identification of the target application.
In some embodiments, the method further comprises: analyzing the received voice request of the user, determining application information of the called application and the intention of the user, and generating response information meeting the intention of the user, wherein the application information of the called application comprises an application identifier of the called application; and presenting the user intention response information by adopting the content presentation mode template associated with the application identification of the called application.
In some embodiments, the application information of the invoked application further comprises keywords for characterizing the invoked application; analyzing the received voice request of the user, determining application information and user intention of the called application, and generating response information meeting the user intention, wherein the response information comprises: performing semantic recognition on the voice request, and determining whether the recognized semantic information contains keywords for representing the called application and keywords for representing the user intention; in response to determining that the identified semantic information contains keywords for characterizing the invoked application and keywords for characterizing the user intent, sending the keywords for characterizing the user intent to a server for supporting the invoked application; and receiving application service data returned by the server for supporting the called application, and generating response information meeting the user intention according to the application service data.
In some embodiments, before receiving a result of processing application service data returned by a server for supporting the invoked application, and taking the processing result as response information meeting the user intention according to the application server interaction data, the method further includes: and in response to determining that the identified semantic information contains the keywords for characterizing the called application and does not contain the keywords for characterizing the user intention, sending the voice request of the user and the preset default intention to a server for supporting the called application.
In some embodiments, prior to associating the content presentation style template of the target application with the obtained application identification of the target application, the method further comprises: acquiring to-be-presented contents of preset filling items to be filled into the selected candidate presentation mode template; and filling the content to be presented to the selected candidate presentation mode template to generate a content presentation mode template of the target application.
In a second aspect, an embodiment of the present application provides an apparatus for presenting content, including: the searching unit is configured to respond to a received request for creating a content presentation mode for a target application accessing the voice service, and search a candidate presentation mode template corresponding to a data type of content to be presented, which is indicated by the request for creating the content presentation mode, from a preset presentation mode template library, wherein the presentation mode template in the preset presentation mode template library is configured with the corresponding data type; and the association unit is configured to receive an instruction of selecting at least one candidate presentation mode template as the content presentation mode template of the target application, and associate the content presentation mode template of the target application with the acquired application identifier of the target application.
In some embodiments, the apparatus further comprises: the analysis unit is configured to analyze the received voice request of the user, determine application information of the called application and the intention of the user, and generate response information meeting the intention of the user, wherein the application information of the called application comprises an application identifier of the called application; and the presentation unit is configured to present the response information of the voice request by adopting the content presentation mode template associated with the application identification of the called application.
In some embodiments, the application information of the invoked application further comprises keywords for characterizing the invoked application; the analysis unit comprises a determining subunit, a processing unit and a processing unit, wherein the determining subunit is configured to perform semantic recognition on the voice request, and determine whether the recognized semantic information contains a keyword for representing the called application and a keyword for representing the user intention; a first sending subunit configured to send, in response to determining that the identified semantic information includes a keyword for characterizing the invoked application and a keyword for characterizing a user intention, the keyword for characterizing the user intention to a server for supporting the invoked application; and the receiving subunit is configured to receive application service data returned by the server for supporting the called application, and generate response information meeting the user intention according to the application service data.
In some embodiments, the parsing unit further comprises: and the second sending subunit is used for responding to the fact that the identified semantic information contains the keywords for representing the called application and does not contain the keywords for representing the user intention, and sending the voice request of the user and the preset default intention to the server for supporting the called application.
In some embodiments, the association unit is further configured to: acquiring to-be-presented contents of preset filling items to be filled into the selected candidate presentation mode template; and filling the content to be presented to the selected candidate presentation mode template to generate a content presentation mode template of the target application.
In a third aspect, an embodiment of the present application provides an electronic device, including: one or more processors; a storage device having one or more programs stored thereon, which when executed by one or more processors, cause the one or more processors to implement a method as in any embodiment of the first aspect.
In a fourth aspect, the present application provides a computer-readable storage medium, on which a computer program is stored, which when executed by a processor implements the method as in any embodiment of the first aspect.
According to the method and the device for presenting the content, when a request for creating a content presentation mode for a target application accessing a voice service is received, candidate presentation mode templates corresponding to the data type of the content to be presented, indicated by the request for creating the content presentation mode, are searched from a preset presentation mode template library, then an instruction for selecting at least one candidate presentation mode template as the content presentation mode template of the target application is received, and the content presentation mode template of the target application is associated with the acquired application identifier of the target application, so that the target application accessing the voice service does not need to recreate the content presentation mode, the cost for accessing the voice service by the target application is reduced, and the flexibility for accessing the voice service by the target application is improved.
Drawings
Other features, objects and advantages of the present application will become more apparent upon reading of the following detailed description of non-limiting embodiments thereof, made with reference to the accompanying drawings in which:
FIG. 1 is an exemplary system architecture diagram that may be employed therein according to an embodiment of the present application;
FIG. 2 is a flow diagram of one embodiment of a method for presenting content, according to the present application;
FIG. 3 is a schematic illustration of an application scenario of a method for presenting content according to the present application;
FIG. 4 is a flow diagram of another embodiment of a method for presenting content according to the present application;
FIG. 5 is a schematic diagram of an embodiment of an apparatus for presenting content in accordance with the present application;
FIG. 6 is a schematic block diagram of a computer system suitable for use in implementing a server according to embodiments of the present application.
Detailed Description
The present application will be described in further detail with reference to the following drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the relevant invention and not restrictive of the invention. It should be noted that, for convenience of description, only the portions related to the related invention are shown in the drawings.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
Fig. 1 illustrates an exemplary system architecture 100 to which a method for presenting content or an apparatus for presenting content of embodiments of the present application may be applied.
As shown in fig. 1, the system architecture 100 may include terminal devices 101, 102, a network 103, and a server 104. The network 103 serves as a medium for providing communication links between the terminal devices 101, 102 and the server 104. Network 103 may include various connection types, such as wired, wireless communication links, or fiber optic cables, to name a few.
Terminal device 101 may interact with server 104 via network 103 for accessing terminal device 102 to server 104. Here, the terminal device 101 may be installed with an application for supporting the access of the terminal device 102 to the server 104. The terminal device 101 may be hardware having a file editing function and a display function, including but not limited to a mobile phone, a computer, and the like.
The terminal device 102 may interact with the server 104 through the network 103, and is configured to send the received voice information to the server 104 through the network 103, so that the server 104 performs semantic analysis. Here, the terminal device 102 may be hardware or software. When the terminal device 102 is hardware, it may be various electronic devices having a display screen and supporting interactive interaction and video playing, including but not limited to a smart phone, a tablet computer, a smart air conditioner, a smart refrigerator, a smart television, and the like. When the terminal device 102 is software, it can be installed in the electronic devices listed above. It may be implemented as multiple pieces of software or software modules (e.g., to provide distributed services) or as a single piece of software or software module. And is not particularly limited herein.
The server 104 may be a server that provides various services (e.g., voice services), for example, the server may receive a request sent by the terminal device 101 to access an application in the terminal device 102 to the voice service, analyze and process the request, and return a processing result to the terminal device 101; after the application in the terminal device 102 accesses the voice service, the voice information sent by the terminal device 102 may be processed, and the processing result may be presented through the terminal device 102.
It is noted that the content-based methods provided by the embodiments of the present application are generally performed by the server 104. Accordingly, the means for presenting content is generally disposed in the server 104.
The server may be hardware or software. When the server is hardware, it may be implemented as a distributed server cluster formed by multiple servers, or may be implemented as a single server. When the server is software, it may be implemented as multiple pieces of software or software modules (e.g., to provide distributed services), or as a single piece of software or software module. And is not particularly limited herein.
It should be understood that the number of terminal devices, networks, and servers in fig. 1 is merely illustrative. There may be any number of terminal devices, networks, and servers, as desired for implementation.
With continued reference to FIG. 2, a flow 200 of one embodiment of a method for presenting content in accordance with the present application is shown. The method for presenting content comprises the following steps:
step 201, in response to receiving a request for creating a content presentation mode for a target application accessing a voice service, finding out a candidate presentation mode template corresponding to a data type of content to be presented, indicated by the request for creating the content presentation mode, from a preset presentation mode template library.
In the present embodiment, an execution subject (e.g., a server shown in fig. 1) of the method for presenting content may provide support for various voice services. Here, the voice service may include, for example, a service in which the user may perform voice interaction with the terminal device performing the main body communication connection to acquire the intention of the user. Here, the execution body may provide voice service support to a terminal device provided with a display screen.
In this embodiment, the execution main body may provide a voice service for various terminal devices that have access to the voice service provided by the execution main body, or may provide a voice service for various applications provided in the terminal devices. Here, the terminal device may include, for example, various smart devices such as a refrigerator, a stereo, an air conditioner, a television, a smart watch, a smart learning machine, and the like; the applications may include, for example, query (e.g., weather query, ticket query, etc.) applications, shopping applications, search applications, web browsing applications, etc., which are typically provided in terminal devices such as mobile phones, computers, etc.
In this embodiment, for a terminal device capable of displaying content (i.e., "a screen-side device"), a presentation mode template library for presenting content may be preset in the execution main body so that the terminal device presents content. Here, the terminal device capable of displaying the content may be, for example, a mobile phone with a screen, a television with a screen, a projector, and the like, and is not limited herein. The presentation mode template in the preset presentation mode template library may be configured with a corresponding data type. The data type may include, for example, a text data type, a picture data type, a video data type, and the like. When a terminal device with a display function requests to access to the execution main body to acquire a voice service, a request for creating a content presentation mode can be sent to the execution main body, so that interactive content can be displayed on the terminal device when a user interacts with the terminal device. After receiving the request, the execution main body may search, from a preset presentation mode template library, a candidate presentation mode template corresponding to the data type of the content to be presented, which is indicated by the request for creating the content presentation mode. Here, the presentation template in the presentation template library may include, for example, a presentation template that presents only text data, a presentation template that presents only picture data, a presentation template that presents only video data, a presentation template that presents both picture data and text data, and the like. Here, in the presentation template in which only text data or picture data is presented, the text data or the picture data may be presented in a horizontal list or a vertical list. The presentation mode template for simultaneously presenting the picture data and the text data may further include a presentation mode template in which the image data and the text data have different relative positions in the display screen. For example, a presentation mode template in which pictures are presented above the screen and text is presented below the screen may be included; and the method can also comprise a presentation mode template in which pictures are presented on the left side of the screen and characters are presented on the right side of the screen.
Taking a presentation mode template for presenting the picture data and the text data at the same time as an example, the presentation mode template may include a title content, a picture content, and a text content for representing the intention information. The title content may be generally positioned in a central location on top of the display screen when the display screen is viewed from the user's perspective. The title content may be represented using one or more of chinese, numeric, alphabetic, and underlined. When the user performs voice interaction with the terminal device supported by the execution main body, the picture content presented by the terminal device may be set on the left side of the display screen, and the text content presented may be set on the right side of the display screen. Here, the terminal device provided with the target application accessing the voice service may custom set whether the background picture is displayed in the display screen when the user performs voice interaction with the terminal device. When displaying a background picture, the background picture may support both JPEG and PNG formats.
As an example, when the data type presented by the target application accessing the voice service is only a text type, the executing entity may find a content presentation mode template presenting only text content; when the data type presented by the target application accessing the voice service is a text type and a picture type, the execution main body can find out a content presentation mode template presenting the text content and the picture content at the same time.
Step 202, receiving an instruction for selecting at least one candidate presentation mode template as a content presentation mode template of the target application, and associating the content presentation mode template of the target application with the acquired application identifier of the target application.
In this embodiment, after finding out the candidate presentation mode templates in step 201, the execution main body may receive an instruction to select at least one candidate presentation mode template as the content presentation mode template of the target application. Here, the instruction may be transmitted for a terminal provided with a target application accessing the voice service. Here, the target application accessing the voice service further includes an application identifier for uniquely identifying the target application, and the application identifier may be, for example, a text or an icon, etc. When receiving that the target application accesses the voice service, the application identifier of the target application can be acquired. After receiving the instruction, the execution subject may further associate the content presentation mode template of the target application with the acquired application identifier of the target application. Here, associating may include, for example, storing the content presentation style template of the target application in correspondence with the acquired application identification of the target application.
In some optional implementations of this embodiment, before associating the content presentation mode template of the target application with the obtained application identifier of the target application, the method may further include: and acquiring the content to be presented of a preset filling item to be filled into the selected candidate presentation mode template, and then filling the content to be presented into the selected candidate presentation mode template to be used as the content presentation mode template of the target application.
Specifically, the preset padding items may include, for example, an application identifier of the target application, a welcome phrase when the target application is started, an end phrase when the target application is closed, and the like. The executing body may obtain a resource storage address (e.g., a storage address of a picture) in which the content to be presented is stored, and then obtain the content to be presented in the preset padding from the resource storage address. The contents of the application identifier, the welcome language when the target application is started, the closing language when the target application is closed and the like are preset by the terminal. After the content to be presented is acquired, the execution main body may fill the content to be presented into the selected candidate presentation mode template, so as to serve as the content presentation mode template of the target application.
With continued reference to fig. 3, fig. 3 is a schematic diagram of an application scenario for presenting content according to the present embodiment. In the application scenario of fig. 3, a user sends a request to the server 302 via the computer 301 to create a presentation of content for a "weather query" application accessing a voice service. In the request, the data type of the content to be presented indicated by the request is also carried, and the data type is text data and picture data. Then, after the server 302 responds to the request, candidate presentation mode templates c and d capable of presenting the text content and the picture content at the same time are found from the preset presentation mode template library 303. Next, the user selects the presentation mode template d as a content presentation mode template of the "weather query" application, and after receiving the instruction, the server 302 associates the application identifier of the "weather query" application that has been acquired with the presentation mode template d.
According to the method for presenting the content, when a request for creating a content presentation mode for a target application accessing the voice service is received, candidate presentation mode templates corresponding to the data type of the content to be presented and indicated by the request for creating the content presentation mode are found from a preset presentation mode template library, then an instruction for selecting at least one candidate presentation mode template as the content presentation mode template of the target application is received, and the content presentation mode template of the target application is associated with the obtained application identifier of the target application, so that the target application accessing the voice service does not need to recreate the content presentation mode, the cost for accessing the voice service by the target application is reduced, and the flexibility for accessing the voice service by the target application is improved.
With further reference to FIG. 4, a flow 400 of another embodiment of a method for presenting content in accordance with the present application is shown. The method for presenting content comprises the following steps:
step 401, in response to receiving a request for creating a content presentation manner for a target application accessing a voice service, finding a candidate presentation manner template corresponding to a data type of content to be presented, indicated by the request for creating the content presentation manner, from a preset presentation manner template library.
In this embodiment, for a terminal device provided with a display screen, a presentation mode template library for content presentation is generally preset in the execution body so that the terminal device can present content. The presentation mode template in the preset presentation mode template library may be configured with a corresponding data type. The data type may include, for example, a text data type, a picture data type, a voice data type, and the like. The execution main body can search out a candidate presentation mode template corresponding to the data type of the content to be presented, which is indicated by the request for creating the content presentation mode, from a preset presentation mode template library in response to receiving the request for creating the content presentation mode for the target application accessing the voice service.
Step 402, receiving an instruction for selecting at least one candidate presentation mode template as a content presentation mode template of the target application, and associating the content presentation mode template of the target application with the acquired application identifier of the target application.
In this embodiment, after finding out the candidate presentation mode templates in step 201, the execution main body may receive an instruction to select at least one candidate presentation mode template as the content presentation mode template of the target application. Here, the instruction may be transmitted for a terminal where a target application for accessing the voice service is located. After receiving the instruction, the execution subject may further associate the content presentation mode template of the target application with the acquired application identifier of the target application.
Here, the specific processing of step 401 and step 402 and the beneficial effects thereof may refer to the related description of the implementation manners of step 201 and step 202 in the corresponding embodiment of fig. 2, and are not described herein again.
Step 403, analyzing the received voice request of the user, determining the application information of the called application and the intention of the user, and generating response information meeting the intention of the user.
In this embodiment, the execution subject may receive a voice request of a user, where the voice request may be a request for a specific application to provide data services, such as a request for a weather application to provide weather data, or a request for a music playing application to play music, and so on. The execution agent may parse the received voice request of the user, so as to determine the application information of the application requested to be invoked (i.e., determine the application information of the invoked application) and the intention of the user. Here, the application information of the invoked application may include an application identification of the invoked application. Specifically, the executing agent may recognize the received voice of the user by using an existing voice recognition technology, and convert the voice into a text. Then, the text obtained by conversion is subjected to semantic analysis and other processing by using language processing technology such as natural language processing, so that the application identifier and the user intention of the application to be called are determined. Here, the user intention may be, for example, an intention to inquire about weather, an intention to set an alarm, or the like. Finally, the execution subject may query the analyzed user intention, and generate response information satisfying the user intention. As an example, when the execution subject parses out the keyword of "a city" and "weather" contained in the voice request of the user, the function of querying weather, which is preset based on the keyword, "a city weather", may be queried, so as to generate response information about "a city weather".
In some optional implementation manners of this embodiment, the application information of the invoked application further includes a keyword for characterizing the invoked application, and the executing body may further perform semantic recognition on the voice request, and determine whether the recognized semantic information includes the keyword for characterizing the invoked application and the keyword for characterizing the user intention; in response to determining that the identified semantic information contains keywords for characterizing the invoked application and keywords for characterizing the user intent, sending the keywords for characterizing the user intent to a server for supporting the invoked application; and receiving application data returned by the server for supporting the called application, and generating response information meeting the user intention according to the application service data.
In particular, the invoked application may be supported by a backend server of the application. Here, the background server of the application is different from the execution subject. After receiving a voice request from a user, the execution body may convert the voice request into a corresponding text by using a voice recognition technology. Then, keywords in the voice request converted into text are extracted using keyword extraction, for example, a keyword extraction method based on word frequency, a keyword extraction method based on a neural network, or the like. And then comparing the keywords with keywords in a preset word library so as to determine whether the identified semantic information contains the keywords for representing the called application and the keywords for representing the user intention. Here, the keyword for characterizing the invoked application may include, for example, an application name of the invoked application. The execution main body may transmit the extracted keyword representing the user intention as the user intention to a server for supporting the invoked application in response to determining that the identified semantic information includes the keyword representing the invoked application and the keyword representing the user intention. After the server supporting the called application completes the information analysis processing, the processing result may be returned to the execution main body. Therefore, the execution main body can receive the application service data returned by the server supporting the called application, and take the application service data as the response information meeting the user intention.
In some embodiments, before receiving a processing result returned by the server and using the processing result as response information meeting the user intention, the method may further include: and in response to determining that the identified semantic information contains keywords for characterizing the called application and does not contain keywords for characterizing the user intention, sending the voice request of the user and preset default intention to a server for supporting the called application. Specifically, the execution body converts the voice request into a corresponding text by using a voice recognition technology, and extracts keywords in the voice request converted into the text by using a keyword extraction method, such as a keyword extraction method based on a word frequency, a keyword extraction method based on a neural network, and the like. And then comparing the keywords with keywords in a preset word library so as to determine whether the identified semantic information contains the keywords for representing the called application and the keywords for representing the user intention. When the execution subject determines that the recognized semantic information contains a keyword for characterizing the called application and does not contain the keyword for characterizing the user intention, the text converted from the voice request of the user and the preset default intention can be sent to the server for supporting the called application. Here, the preset default intention may be used, for example, to characterize information that cannot recognize the intention of the user. Then, the execution main body may receive a processing result returned by the server supporting the invoked application, so that the execution main body may use the processing result as response information satisfying the user's intention.
Step 404, presenting the response information of the voice request by using the content presentation mode template associated with the application identifier of the called application.
In this embodiment, after associating the content presentation mode template of the target application with the acquired application identifier of the target application in step 402, according to the application identifier of the called application determined in step 403, the execution main body may present response information corresponding to the voice request received in step 403 by using the content presentation mode template associated with the application identifier of the called application.
As can be seen from fig. 4, unlike the embodiment shown in fig. 2, the present embodiment highlights the step of presenting the response information of the user's voice request using the created content presentation style template. By presenting the response information using the created content presentation mode template, it is possible to quickly and clearly present the content requested by the user in the created template.
With further reference to fig. 5, as an implementation of the method shown in the above figures, the present application provides an embodiment of an apparatus for presenting content, which corresponds to the method embodiment shown in fig. 2, and which is particularly applicable in various electronic devices.
As shown in fig. 5, the apparatus 500 for presenting content of the present embodiment includes: a lookup unit 501 and an association unit 502. The searching unit 501 is configured to, in response to receiving a request for creating a content presentation manner for a target application accessing a voice service, search, from a preset presentation manner template library, a candidate presentation manner template corresponding to a data type of content to be presented, which is indicated by the request for creating the content presentation manner, where the presentation manner template in the preset presentation manner template library is configured with the corresponding data type; the associating unit 502 is configured to receive an instruction to select at least one candidate presentation mode template as a content presentation mode template of the target application, associate the content presentation mode template of the target application with the acquired application identification of the target application.
In this embodiment, for specific processing of the searching unit 501 and the associating unit 502 in the device 500 for presenting content and beneficial effects brought by the specific processing, reference may be made to related descriptions of implementation manners of step 201 and step 202 in the corresponding embodiment of fig. 2, and details are not described herein again.
In some optional implementations of this embodiment, the apparatus 500 for presenting content further includes: a parsing unit (not shown) configured to parse the received voice request of the user, determine application information of the invoked application and an intention of the user, and generate response information satisfying the intention of the user, wherein the application information of the invoked application includes an application identifier of the invoked application; a presentation unit (not shown) configured to present the response information of the voice request using the content presentation style template associated with the application identification of the invoked application.
In some optional implementations of this embodiment, the application information of the invoked application further includes a keyword for characterizing the invoked application; the analysis unit (not shown) comprises a determining subunit (not shown) and a processing unit, wherein the determining subunit is configured to semantically identify the voice request, and determine whether the identified semantic information contains a keyword for representing the called application and a keyword for representing the user intention; a first transmitting subunit (not shown) configured to transmit, in response to a determination that the recognized voice information contains a keyword for characterizing the invoked application and a keyword for characterizing a user intention, the keyword for characterizing the user intention to a server for supporting the invoked application; a receiving subunit (not shown) configured to receive application service data returned by the server for supporting the invoked application, and generate response information satisfying the user's intention according to the application service data
In some optional implementations of this embodiment, the parsing unit (not shown) further includes: and a second transmitting subunit (not shown) for transmitting the voice request of the user and a preset default intention to the server for supporting the invoked application in response to determining that the recognized semantic information contains a keyword for characterizing the invoked application and does not contain a keyword for characterizing the user intention.
In some optional implementations of this embodiment, the associating unit 502 is further configured to: acquiring to-be-presented contents of preset filling items to be filled into the selected candidate presentation mode template; and filling the content to be presented to the selected candidate presentation mode template to generate a content presentation mode template of the target application.
Referring now to FIG. 6, shown is a block diagram of a computer system 600 suitable for use in implementing the electronic device of an embodiment of the present application. The electronic device shown in fig. 6 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present application.
As shown in fig. 6, the computer system 600 includes a Central Processing Unit (CPU)601 that can perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM)602 or a program loaded from a storage section 608 into a Random Access Memory (RAM) 603. In the RAM 603, various programs and data necessary for the operation of the system 600 are also stored. The CPU 601, ROM 602, and RAM 603 are connected to each other via a bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
To the I/O interface 605, AN input section 606 including a keyboard, a mouse, and the like, AN output section 607 including a network interface card such as a Cathode Ray Tube (CRT), a liquid crystal display (L CD), and the like, a speaker, and the like, a storage section 608 including a hard disk, and the like, and a communication section 609 including a network interface card such as a L AN card, a modem, and the like, the communication section 609 performs communication processing via a network such as the internet, a drive 610 is also connected to the I/O interface 605 as necessary, a removable medium 611 such as a magnetic disk, AN optical disk, a magneto-optical disk, a semiconductor memory, and the like is mounted on the drive 610 as necessary, so that a computer program read out therefrom is mounted into the storage section 608 as necessary.
In particular, according to an embodiment of the present disclosure, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network through the communication section 609, and/or installed from the removable medium 611. The computer program performs the above-described functions defined in the method of the present application when executed by a Central Processing Unit (CPU) 601. It should be noted that the computer readable medium described herein can be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present application, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In this application, however, a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wire, fiber optic cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations of the present application may be written in any combination of one or more programming languages, including AN object oriented programming language such as Java, Smalltalk, C + +, and conventional procedural programming languages, such as the "C" programming language or similar programming languages.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present application may be implemented by software or hardware. The described units may also be provided in a processor, and may be described as: a processor includes a lookup unit and an association unit. The names of the units do not form a limitation on the units themselves in some cases, for example, the search unit may also be described as a unit that searches a candidate presentation mode template corresponding to the data type of the content to be presented indicated by the request for creating the content presentation mode from a preset presentation mode template library.
As another aspect, the present application also provides a computer-readable medium, which may be contained in the electronic device described in the above embodiments; or may exist separately without being assembled into the electronic device. The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: in response to receiving a request for creating a content presentation mode for a target application accessing a voice service, searching a candidate presentation mode template corresponding to a data type of content to be presented, which is indicated by the request for creating the content presentation mode, from a preset presentation mode template library, wherein the presentation mode template in the preset presentation mode template library is configured with the corresponding data type; and receiving an instruction of selecting at least one candidate presentation mode template as a content presentation mode template of the target application, and associating the content presentation mode template of the target application with the acquired application identification of the target application.
The above description is only a preferred embodiment of the application and is illustrative of the principles of the technology employed. It will be appreciated by those skilled in the art that the scope of the invention herein disclosed is not limited to the particular combination of features described above, but also encompasses other arrangements formed by any combination of the above features or their equivalents without departing from the spirit of the invention. For example, the above features may be replaced with (but not limited to) features having similar functions disclosed in the present application.

Claims (12)

1. A method for presenting content, comprising:
in response to receiving a request for creating a content presentation mode for a target application accessing a voice service, searching a candidate presentation mode template corresponding to a data type of content to be presented, which is indicated by the request for creating the content presentation mode, from a preset presentation mode template library, wherein the presentation mode template in the preset presentation mode template library is configured with the corresponding data type; the data type comprises one or more of a text data type, a picture data type or a video data type;
receiving an instruction of selecting at least one candidate presentation mode template as a content presentation mode template of a target application, and associating the content presentation mode template of the target application with the acquired application identification of the target application.
2. The method of claim 1, wherein the method further comprises:
analyzing a received voice request of a user, determining application information of a called application and the intention of the user, and generating response information meeting the intention of the user, wherein the application information of the called application comprises an application identifier of the called application;
and presenting the response information of the user intention by adopting the content presentation mode template associated with the application identification of the called application.
3. The method of claim 2, wherein the application information of the invoked application further comprises keywords for characterizing the invoked application; and
the analyzing the received voice request of the user, determining the application information of the called application and the user intention, and generating response information meeting the user intention, comprises:
performing semantic recognition on the voice request, and determining whether the recognized semantic information contains keywords for representing the called application and keywords for representing the user intention;
in response to determining that the identified semantic information includes keywords characterizing the invoked application and keywords characterizing the user intent, sending the keywords characterizing the user intent to a server supporting the invoked application;
receiving application service data returned by the server for supporting the called application, and generating the response information meeting the user intention according to the application service data.
4. The method of claim 3, wherein before receiving application service data returned by the server for supporting the invoked application, generating the response information satisfying the user intention according to the application service data, further comprising:
and in response to determining that the identified semantic information contains keywords for characterizing the called application and does not contain keywords for characterizing the user intention, sending the voice request of the user and preset default intention to a server for supporting the called application.
5. The method of claim 1, wherein prior to associating a content presentation template of a target application with the obtained application identification of the target application, the method further comprises:
acquiring to-be-presented contents of preset filling items to be filled into the selected candidate presentation mode template;
and filling the content to be presented to the selected candidate presentation mode template to generate the content presentation mode template of the target application.
6. An apparatus for presenting content, comprising:
the search unit is configured to search a candidate presentation mode template corresponding to the data type of the content to be presented, which is indicated by a request for creating the content presentation mode, from a preset presentation mode template library in response to receiving the request for creating the content presentation mode for the target application accessing the voice service, wherein the presentation mode template in the preset presentation mode template library is configured with the corresponding data type; the data type comprises one or more of a text data type, a picture data type or a video data type;
the association unit is configured to receive an instruction of selecting at least one candidate presentation mode template as a content presentation mode template of a target application, and associate the content presentation mode template of the target application with the acquired application identifier of the target application.
7. The apparatus of claim 6, wherein the apparatus further comprises:
the system comprises an analysis unit, a processing unit and a processing unit, wherein the analysis unit is configured to analyze a received voice request of a user, determine application information of an invoked application and an intention of the user, and generate response information meeting the intention of the user, wherein the application information of the invoked application comprises an application identifier of the invoked application;
a presentation unit configured to present response information of the user intention using a content presentation mode template associated with the application identification of the invoked application.
8. The apparatus of claim 7, wherein the application information of the invoked application further comprises keywords for characterizing the invoked application; and
the analysis unit includes:
a determining subunit, configured to perform semantic recognition on the voice request, and determine whether the recognized semantic information includes a keyword for characterizing the invoked application and a keyword for characterizing the user intention;
a first transmitting subunit configured to transmit, in response to determining that the identified semantic information includes a keyword for characterizing an invoked application and a keyword for characterizing the user intention, the keyword for characterizing the user intention to a server for supporting the invoked application;
and the receiving subunit is configured to receive the application service data returned by the server for supporting the called application, and generate response information meeting the user intention according to the application service data.
9. The apparatus of claim 8, wherein the parsing unit further comprises:
and the second sending subunit is used for responding to the fact that the identified semantic information contains the keywords for representing the called application and does not contain the keywords for representing the user intention, and sending the voice request of the user and the preset default intention to a server for supporting the called application.
10. The apparatus of claim 6, wherein the association unit is further configured to:
acquiring to-be-presented contents of preset filling items to be filled into the selected candidate presentation mode template;
and filling the content to be presented to the selected candidate presentation mode template to generate the content presentation mode template of the target application.
11. An electronic device, comprising:
one or more processors;
a storage device having one or more programs stored thereon,
when executed by the one or more processors, cause the one or more processors to implement the method of any one of claims 1-5.
12. A computer-readable storage medium, on which a computer program is stored, which program, when being executed by a processor, carries out the method according to any one of claims 1-5.
CN201810718896.5A 2018-06-29 2018-06-29 Method and apparatus for presenting content Active CN109036397B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810718896.5A CN109036397B (en) 2018-06-29 2018-06-29 Method and apparatus for presenting content

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810718896.5A CN109036397B (en) 2018-06-29 2018-06-29 Method and apparatus for presenting content

Publications (2)

Publication Number Publication Date
CN109036397A CN109036397A (en) 2018-12-18
CN109036397B true CN109036397B (en) 2020-08-04

Family

ID=65522223

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810718896.5A Active CN109036397B (en) 2018-06-29 2018-06-29 Method and apparatus for presenting content

Country Status (1)

Country Link
CN (1) CN109036397B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109901899A (en) * 2019-01-28 2019-06-18 百度在线网络技术(北京)有限公司 Video speech technical ability processing method, device, equipment and readable storage medium storing program for executing
CN109976829A (en) * 2019-03-18 2019-07-05 百度在线网络技术(北京)有限公司 A kind of voice technical ability signal language configuration method and device
CN110162776A (en) 2019-03-26 2019-08-23 腾讯科技(深圳)有限公司 Interaction message processing method, device, computer equipment and storage medium
CN110727435B (en) * 2019-10-21 2023-09-05 百度在线网络技术(北京)有限公司 Rendering method, rendering device, electronic equipment and storage medium
CN111309857A (en) * 2020-01-20 2020-06-19 联想(北京)有限公司 Processing method and processing device
CN111273964B (en) * 2020-02-14 2024-03-08 抖音视界有限公司 Data loading method and device
CN113051477A (en) * 2021-03-26 2021-06-29 北京百度网讯科技有限公司 Content recommendation method and device, electronic equipment and medium

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103425741A (en) * 2013-07-16 2013-12-04 北京中科汇联信息技术有限公司 Information exhibiting method and device
US20150169529A1 (en) * 2013-12-16 2015-06-18 Sap Ag Mobile device data rendering
CN104090904B (en) * 2014-05-16 2018-03-23 百度在线网络技术(北京)有限公司 A kind of method and apparatus for being used to provide target search result
CN104916287A (en) * 2015-06-10 2015-09-16 青岛海信移动通信技术股份有限公司 Voice control method and device and mobile device
CN105574138A (en) * 2015-12-10 2016-05-11 商丘师范学院 Information retrieval system
CN106096064A (en) * 2016-07-19 2016-11-09 百度在线网络技术(北京)有限公司 For the method and apparatus generating the page
CN107256266A (en) * 2017-06-16 2017-10-17 珠海格力电器股份有限公司 One kind inquiry content displaying method and its system

Also Published As

Publication number Publication date
CN109036397A (en) 2018-12-18

Similar Documents

Publication Publication Date Title
CN109036397B (en) Method and apparatus for presenting content
CN107844586B (en) News recommendation method and device
US10824874B2 (en) Method and apparatus for processing video
JP7335062B2 (en) Voice service providing method and apparatus
US20190147864A1 (en) Voice interaction based method and apparatus for generating multimedia playlist
US10878044B2 (en) System and method for providing content recommendation service
US10936645B2 (en) Method and apparatus for generating to-be-played multimedia content
US10789474B2 (en) System, method and apparatus for displaying information
US20200012675A1 (en) Method and apparatus for processing voice request
CN109165302A (en) Multimedia file recommendation method and device
CN109271556B (en) Method and apparatus for outputting information
CN109359194B (en) Method and apparatus for predicting information categories
US20170132267A1 (en) Pushing system and method based on natural information recognition, and a client end
CN111522909B (en) Voice interaction method and server
CN109582825B (en) Method and apparatus for generating information
US20200409998A1 (en) Method and device for outputting information
CN110737824B (en) Content query method and device
CN108924604A (en) Method and apparatus for playing video
CN111324700A (en) Resource recall method and device, electronic equipment and computer-readable storage medium
US10872108B2 (en) Method and apparatus for updating multimedia playlist
WO2020199659A1 (en) Method and apparatus for determining push priority information
CN111274819A (en) Resource acquisition method and device
CN108573053B (en) Information pushing method, device and system
CN112148962B (en) Method and device for pushing information
CN110489326B (en) IDS-based HTTPAPI debugging method device, medium and equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20210510

Address after: 100085 Baidu Building, 10 Shangdi Tenth Street, Haidian District, Beijing

Patentee after: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd.

Patentee after: Shanghai Xiaodu Technology Co.,Ltd.

Address before: 100085 Baidu Building, 10 Shangdi Tenth Street, Haidian District, Beijing

Patentee before: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd.

TR01 Transfer of patent right