CN110020411B - Image-text content generation method and equipment - Google Patents

Image-text content generation method and equipment Download PDF

Info

Publication number
CN110020411B
CN110020411B CN201910248465.1A CN201910248465A CN110020411B CN 110020411 B CN110020411 B CN 110020411B CN 201910248465 A CN201910248465 A CN 201910248465A CN 110020411 B CN110020411 B CN 110020411B
Authority
CN
China
Prior art keywords
text
user
picture
keywords
keyword
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910248465.1A
Other languages
Chinese (zh)
Other versions
CN110020411A (en
Inventor
陈大年
杨凌枫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Zhangmen Science and Technology Co Ltd
Original Assignee
Shanghai Zhangmen Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Zhangmen Science and Technology Co Ltd filed Critical Shanghai Zhangmen Science and Technology Co Ltd
Priority to CN201910248465.1A priority Critical patent/CN110020411B/en
Publication of CN110020411A publication Critical patent/CN110020411A/en
Application granted granted Critical
Publication of CN110020411B publication Critical patent/CN110020411B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/103Formatting, i.e. changing of presentation of documents
    • G06F40/106Display of layout of documents; Previewing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Abstract

The scheme can respond to the text input operation of a user, determine a text input by the user, display the text in a display interface, determine a keyword according to the text, acquire a picture matched with the keyword, and generate and display the image-text content comprising the text and the picture in the display interface. In the whole process, a user does not need to manually search and select matching, the picture matched with the text content can be directly and automatically found by the input text, and then the image-text content containing the picture and the characters is generated, so that the user experience can be improved, the user can effectively express the emotion in the social contact process, and a large amount of complicated operations of the user are not needed to be added.

Description

Image-text content generation method and equipment
Technical Field
The present application relates to the field of information technologies, and in particular, to a method and a corresponding device for generating image-text content.
Background
With the continuous development of internet technology, various social contact modes based on the internet are continuously appeared. For example, a user can use the social platform to publish his own dynamic and original content in a publishing interface for other users to view, comment and the like, or chat with other users through instant messaging software.
In the social process, simple character interaction is more and more difficult to meet the expression requirements of the user, so that the user often uses characters in cooperation with pictures, and the current emotion is expressed more accurately. However, in an actual scene, a user often cannot quickly find a picture suitable for the current expression requirement each time. The user may need to switch from the social application to the search application, perform complex picture search using a search engine of the search application, and may use the pictures in the social application only after the pictures are found and downloaded and stored, which is a complicated operation in the whole process and poor user experience.
Content of application
An object of the present application is to provide a method and apparatus for generating image-text content.
To achieve the above object, some embodiments of the present application provide a method for generating teletext content, the method comprising:
in response to a text input operation of a user, determining a text input by the user, and displaying the text in a display interface;
determining keywords according to the text, and acquiring pictures matched with the keywords;
and generating and displaying the image-text content containing the text and the picture in the display interface.
According to another aspect of the present application, there is also provided a teletext content generation arrangement comprising a memory for storing computer program instructions and a processor for executing the computer program instructions, wherein the computer program instructions, when executed by the processor, trigger the arrangement to perform the teletext content generation method.
In addition, the embodiment of the application also provides a computer readable medium, on which computer program instructions are stored, and the computer program instructions can be executed by a processor to realize the image-text content generation method.
In the scheme provided by some embodiments of the present application, in response to a text input operation of a user, a text input by the user can be determined, the text is displayed in a display interface, a keyword is determined according to the text, an image matched with the keyword is obtained, and image-text content including the text and the image is generated and displayed in the display interface. In the whole process, a user does not need to manually search and select matching, the picture matched with the text content can be directly and automatically found by the input text, and then the image-text content containing the picture and the characters is generated, so that the user experience can be improved, the user can effectively express the emotion in the social contact process, and a large amount of complicated operations of the user are not needed to be added.
Drawings
Other features, objects and advantages of the present application will become more apparent upon reading of the following detailed description of non-limiting embodiments thereof, made with reference to the accompanying drawings in which:
fig. 1 is a processing flow chart of a method for generating image-text content according to an embodiment of the present application;
FIG. 2 is a flowchart illustrating a process for determining keywords according to the text in an embodiment of the present application;
FIG. 3 is a diagram illustrating the display effect of candidate keywords in a keyword display area according to an example of the present application;
FIG. 4 is a schematic diagram of a user interaction interface in a process of obtaining a picture according to an embodiment of the present application;
FIG. 5 is a flowchart of a process for implementing intelligent matching of text images by using the solution provided by the embodiment of the present application;
FIG. 6 is a diagram illustrating pop-up window options in an implementation process of intelligent matching of a text in an embodiment of the present application;
FIG. 7 is a schematic diagram of an apparatus suitable for implementing the methods and/or aspects of embodiments of the present disclosure;
the same or similar reference numbers in the drawings identify the same or similar elements.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
In a typical configuration of the present application, the terminal, the devices serving the network each include one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include forms of volatile memory in a computer readable medium, Random Access Memory (RAM) and/or non-volatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). Memory is an example of a computer-readable medium.
Computer-readable media, which include both non-transitory and non-transitory, removable and non-removable media, may implement the information storage by any method or technology. The information may be computer program instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device.
The embodiment of the application provides a method for generating image-text contents, which can generate the image-text contents containing characters and pictures, can directly and automatically find the pictures matched with the text contents from the input texts without manually searching and selecting matching pictures by a user in the whole process, and then generates and displays the image-text contents containing the pictures and the characters, so that the user experience can be improved, the user can effectively express emotion in the social process, and a large amount of complicated operations of the user are not required to be added.
In a practical scenario, the execution subject of the method may be a user equipment, or a device formed by integrating the user equipment and a network device through a network. The user equipment comprises an input and output device for realizing interaction with a user, and the data processing process can be realized locally in the user equipment or in network equipment and provides a processing result to the user equipment through a network. The user equipment comprises but is not limited to various terminal equipment such as a computer, a mobile phone and a tablet computer; including but not limited to implementations such as a network host, a single network server, multiple sets of network servers, or a cloud-computing-based collection of computers. Here, the cloud is made up of a large number of hosts or network servers based on cloud computing (CloudComputing), which is a type of distributed computing, one virtual computer consisting of a collection of loosely coupled computers.
Fig. 1 shows a processing flow of a method for generating teletext content according to an embodiment of the present application, where the method includes the following processing steps:
step S101, responding to the text input operation of the user, determining the text input by the user, and displaying the text in the display interface. The text input operation may be an input operation implemented by an input device in the user equipment, for example, a corresponding text input operation implemented by an input device such as a keyboard, a tablet, a touch screen, etc., so that the user equipment may determine the text input by the user according to the operations.
In an actual scenario, the text input operation of the user may be input based on an input interface provided by an application program running in the user equipment, for example, the text input operation may be implemented based on an input interface such as a chat interface in an instant messaging application, an editing interface of a mail application, and a publishing interface of user original content in a social platform application. Taking the example that the user chats with the friend by using the instant messaging application of the mobile phone, the virtual keyboard can be provided on the chatting interface for the user to operate, and the user clicks to complete the input of the text. According to the text input operation of the user, specific text content can be determined, and then the input text is displayed in an input box of a chat interface of the instant messaging application. Therefore, the display interface for generating the image-text content in the embodiment of the application may be a User Generated Content (UGC) publishing interface of a social platform such as a microblog, a blog, a friend group, and the like, or may also be a communication interface of a communication tool such as instant messaging, a mail, and the like.
And S102, determining keywords according to the text, and acquiring pictures matched with the keywords. After the user inputs the text, keywords related to the text can be determined through the text, for example, the text input by the user is "good hot today, and wants to eat a frozen sucker", the keywords can be extracted as "good hot" and "frozen sucker" after the text is processed, and then pictures matched with "good hot" and "frozen sucker" are automatically searched.
In some embodiments of the present application, after performing step S101, step S102 may be triggered by a specific condition. The trigger condition may be the detection of a specific user operation or the detection of specific input content. The specific user operation may include, but is not limited to, an operation of clicking a trigger button, a preset gesture operation on a touch screen, an operation of driving the user equipment to implement a preset movement trend, and the like.
The operation of clicking the trigger button may refer to an operation of clicking a specific button displayed on the interactive interface by the user equipment, for example, the chat interface may further include a button for intelligent matching, and after the user inputs the text, the button may be clicked, so that the user equipment may be triggered to perform the subsequent steps. The operation of the click operation may be a trigger operation performed by a user controlling a cursor using an external input device such as a mouse or a keyboard, or may be a touch operation performed by a user using a touch input device such as a finger or an input pen, which is not particularly limited in this embodiment.
The preset gesture operation on the touch screen may refer to a contact sliding track of a user on a touch device of the user equipment. The contact sliding trajectory may include, but is not limited to, a straight line or a curve with an arbitrary shape, which is composed of a plurality of touch points corresponding to a plurality of consecutive touch events, and this embodiment is not particularly limited thereto. For example, a long press operation gesture, a swipe gesture, etc. on the interactive interface may be possible.
The operation of driving the user equipment to implement the preset motion trend may refer to that the user equipment is held by a hand of a user, and when the user equipment is in a specific state (for example, in a state of inputting a text), a motion track of the user equipment is driven to move, for example, shake, turn over, and the like.
In the implementation process, the operation gesture of the user on the page where the shortcut entrance is located may be detected specifically by using a sensor device. Specifically, the sensor device may include, but is not limited to, at least one of a gravity sensor, an acceleration sensor, a pressure sensor, an infrared sensor, a distance sensor, and an image sensor, and this embodiment is not particularly limited thereto.
The distance sensor may be an ultrasonic distance sensor, or may also be an infrared distance sensor, or may also be a laser distance sensor, or may also be a microwave distance sensor, which is not particularly limited in this embodiment. These distance sensors are well known in the art, and the detailed description can be referred to the relevant contents in the prior art, and will not be described herein.
The image sensor may be a Charge Coupled Device (CCD) sensor, or may also be a metal-Oxide Semiconductor (CMOS) sensor, which is not particularly limited in this embodiment.
In addition, the detection of the specific input content may be content detection based on the input text, for example, a character may be preset as a trigger character, and when the trigger character is detected at any position or a preset position in the text, the step S102 is triggered to be executed. For example, in some embodiments of the present application, the trigger character may be # and the trigger condition is satisfied when the end of the text is detected as the character "# #". Taking the text "good hot today, want to eat a frozen sucker" as an example, if the subsequent matching step is to be triggered, the user may actually input "good hot today, want to eat a frozen sucker # #", and thus after step S101 is completed, the subsequent step is triggered to be executed.
In some embodiments of the present application, a processing flow as shown in fig. 2 may be adopted when determining keywords from the text, including the following processing steps:
step S201, firstly, performing word segmentation processing on the text to obtain word-segmented phrases contained in the text. For example, for the text "it is good today and wants to eat a frozen sucker", after performing word segmentation processing on the text, the processing result includes the following word segmentation phrases: today/hot/o/want to eat/ice-lolly. When performing word segmentation, any mature word segmentation algorithm may be used, for example, a dictionary-based word segmentation algorithm such as a forward maximum matching method, a reverse maximum matching method, a bidirectional matching word segmentation method, or a word segmentation algorithm based on machine learning such as HMM, CRF, SVM, and deep learning.
The text processed in this step may be all the texts input in step S101, or may be a partial text selected from all the texts input in step S101, for example, some sentence in a sentence, or some part of a sentence may be selected. Taking the text "today is good at home, wants to eat a frozen sucker" as an example, the selection operation performed by the user on the text may be obtained, for example, the user may mark the text that needs to be selected by long pressing, sliding, and the like, the device may select a part of the text to be processed according to the selection operation of the user, if the user selects "wants to eat a frozen sucker", the word segmentation processing may be performed on the part of the text in this step, and the following word segmentation phrases are obtained: want to eat/ice-lolly.
Step S202, determining keywords according to the word segmentation phrases, wherein one way may be to directly select one or more word segmentation phrases from the word segmentation phrases as the keywords, for example, three of "good hot", "want to eat", and "frozen sucker" may be selected as the keywords to search for the picture.
In addition, another way for determining keywords from a text is provided in the embodiments of the present application, where the way first performs semantic analysis on the text to determine semantic information corresponding to the text, and then determines keywords according to the semantic information. The semantic information indicates the meaning of the text to be expressed, and may specifically be a plurality of information such as the emotion expressed by the text, the described behavior, or the contained things, and the keyword determined by the semantic information is different from the manner shown in fig. 2.
In an actual scene, a deep learning model can be constructed in advance by using a deep learning mode, and the deep learning model is trained by using a sample set labeled with semantic information. After the model training is finished, the deep learning model can be used for finishing semantic analysis and identifying semantic information corresponding to the text. For example, for a text "do not want to move at all but want to lie today", the text is input into a deep learning model for semantic analysis, and if recognized semantic information can include moods such as sluggish expressed by the text, keywords related to the moods can be determined.
Because the keywords are used for matching the pictures, and the pictures are used for more accurately expressing the emotion of the user in the social contact process, the accuracy of the keywords is closely related to the emotion required to be expressed by the pictures and the characters. Therefore, in some embodiments of the present application, instead of automatically determining the keyword for matching, a candidate keyword may be determined according to a text, the candidate keyword is presented to a user, then an operation of the user is detected, and the candidate keyword is adjusted in response to a keyword adjustment operation of the user to determine the keyword.
For example, for the text "today is good hot and wants to eat a frozen sucker", all word segmentation phrases of the word segmentation result may be used as candidate keywords, that is, six candidate keywords including "today", "good hot", "a", "want to eat", "individual" and "frozen sucker" may be included. These keywords will be presented to the user so that the user can adjust according to the actual situation.
In some embodiments of the present application, after determining the candidate keyword, the user equipment may set a keyword display area on the display interface, where the keyword display area may be presented in the display interface in the form of a pop-up window, a floating layer, or the like, and the candidate keyword may be displayed in the form of a tag in the keyword display area. Fig. 3 shows the display effect of the above six candidate keywords in the keyword display area.
The keyword adjustment operation is to adjust the candidate keyword, and may include, but is not limited to, selection, modification, addition, deletion, etc. of the keyword. For example, for the aforementioned six candidate keywords "today", "nice hot", "o", "want to eat", "one", "frozen sucker", 4 of them may be deleted, and only "nice hot" and "frozen sucker" are taken as the keywords finally used for picture matching. In an actual scenario, a user may determine keywords from candidate keywords by clicking on the keyword tags, for example, if it is required to use "nice hot" and "frozen sucker" as the keywords finally used for picture matching, the tags corresponding to "nice hot" and "frozen sucker" may be clicked among six candidate keywords. The device, upon detecting a click operation, will highlight the two tags to distinguish them from the other tags that were not selected. In addition, when determining the keyword, other phrases may be additionally added as the keyword according to the requirement of the user, for example, a new keyword "summer" is added, or a candidate keyword therein may also be modified, for example, "popsicle" is modified to "popsicle" or the like. By adjusting the candidate keywords, keywords which meet the requirements of users better can be determined, so that the matched pictures are more accurate.
When there are a plurality of keywords for matching pictures, picture matching is performed based on the plurality of keywords, for example, matching may be performed with tag information of the pictures themselves, and if the tags are completely matched or partially matched, the pictures may be taken as pictures matched with the keywords. For example, when 4 keywords A, B, C, D are included, matching may be performed based on all keywords and tag information of the picture itself, and if the pictures in the preset picture data set include pictures PIC 1-4, where the tag information of the picture PIC1 is A, B, C, E, the tag information of the picture PIC2 is B, C, E, F, the tag information of the picture PIC3 is C, G, and the tag information of the picture PIC4 is A, C, D, E, G. If at least two pieces of tag information included in the pictures are the same as the keywords, the pictures PIC1, PIC2 and PIC4 may be determined as the pictures matching the keywords. It should be understood by those skilled in the art that the matching of the tag information of the above-mentioned pictures and the keywords is only an example, and other existing or future modifications or extensions based on similar principles should be included in the protection scope of the present application if they can be applied to the present application, and are included by reference.
In an actual scenario, the preset picture data set may be a preset picture database, or may be a result set obtained by searching the internet using a search engine and using a keyword as a search condition. When a plurality of matched pictures are obtained, for example, three pictures PIC1, PIC2 and PIC4 are obtained in the foregoing scene, these pictures may be used as candidate pictures, and the candidate pictures are presented to the user, fig. 4 shows a schematic diagram of a user interaction interface in an embodiment of the present application, where area1 is a display area of a keyword determined based on text, and area2 is a display area of a candidate picture obtained based on the keyword. The user can perform a selection operation in the display area, the user equipment, in response to the selection operation of the user, will determine a target picture in the candidate pictures, for example, three pictures PIC1, PIC2 and PIC4 are displayed in the display area, and the user can select at least one of the pictures as the target picture by a click operation, so that text and the target picture are included in the teletext content subsequently generated and displayed in the display interface.
In other embodiments of the present application, when there are multiple keywords used for matching pictures, a core keyword may be determined in the keywords, and taking the aforementioned 4 keywords A, B, C, D as an example, at least one of the keywords may be determined to be the core keyword, if the determined core keyword in this embodiment is a. And when picture matching is carried out in a preset picture data set according to the keywords and pictures matched with the keywords are obtained, the weight of the core keywords is greater than that of non-core keywords when the pictures are matched.
According to different practical application scenarios, the weighting may be embodied differently, for example, only pictures in which the tag information includes the core keyword are determined as matching results, so that the keyword A, B, C, D is searched for in the pictures PIC 1-4, and since B, C two tag information in the picture PIC2 are the same as the keyword, but the matching results are the pictures PIC1 and PIC4 without including the picture PIC2 because the core keyword a is not included. Or the weighting may be embodied in the order of displaying the candidate pictures, for example, preferentially displaying the pictures including the core keyword in the tag information.
Step S103, after the picture is determined, generating and displaying the image-text content containing the text and the picture in the display interface. For example, taking an instant messaging scenario as an example, after a user inputs a text "good hot today and wants to eat a frozen sucker" in an input frame of a chat interface, the user can automatically find a picture PIC1 matched with keywords such as "good hot today" and "frozen sucker" and generate image-text content including the text "good hot today and wants to eat a frozen sucker" and the picture PIC1, and the image-text content can be displayed in the chat interface.
For example, taking the social platform to publish the message, if a friend ring is published, a text "good hot today and want to eat a frozen sucker" is first input into an input box of the publishing interface, and a picture PIC1 matched with keywords such as "good hot today" and "frozen sucker" can be automatically found based on the text, so that image-text content including the text "good hot today and want to eat a frozen sucker" and the picture PIC1 can be generated, and the image-text content is displayed in the publishing interface.
Therefore, in the whole process, a user does not need to manually search and select matching pictures, pictures matched with the text contents can be directly and automatically found through the input text, and then the picture-text contents containing the pictures and the characters are generated and displayed, so that the user experience can be improved, the user can effectively express the emotion in the social contact process, and meanwhile, a large number of complicated operations of the user are not needed to be added.
In some embodiments of the present application, when generating and displaying the teletext content including the text and the picture in the display interface, the user equipment may insert the picture in a preset area of the display interface on which the text is displayed, so that the display interface displays the teletext content including the text and the picture. The preset area may be a preset position around the text, such as the end of the text, the lower part of the text, and the like, and forms an integral text content with the text. In addition, in a specific scene, for example, in a chat interface of an instant messaging application, the picture can be used as a background picture in the chat interface as a scheme for automatically changing the background picture of the chat interface, so that the interestingness of the instant messaging application is improved, and the user experience is better.
Fig. 5 shows a process flow for implementing intelligent matching of texts by using the solution provided by the embodiment of the present application, which includes the following process steps:
the method comprises the following steps:
1. after the text box enters the text, click the add picture button. For example, the text input in this embodiment is "you are hot today and want to eat a frozen sucker".
2. In the pop-up option that appears, the intelligent match is selected, as shown in FIG. 6. The intelligent image matching function is a function corresponding to the image-text content generation scheme provided by the embodiment of the application.
Step two:
1. after the intelligent matching is selected, the input text is subjected to word segmentation, and the interactive interface is as shown in fig. 4. The text box area1 above the interactive interface will show the word segmentation result of the text, and each word segmentation phrase will be shown in a separate tag. For example, the word-segmentation phrases in this embodiment are six word-segmentation phrases, i.e., "today", "nice and hot", "o", "want to eat", "personal", and "frozen sucker", and these six word-segmentation phrases are candidate keywords.
2. Clicking on the tag needed for picture matching selects one or more of the tags, and the selected tags are displayed in a highlighted form. In addition, the word-separating phrases can be directly modified, or more keywords can be input.
3. After searching for pictures according to the keywords, the matching candidate pictures appear in the lower candidate picture area 2.
4. And selecting a satisfactory picture, namely inserting the picture into the text to generate the image-text content. If the candidate pictures do not have satisfactory pictures, the user can continue to adjust the keywords and re-match the pictures.
Based on the same inventive concept, the embodiment of the application also provides a device for generating the image-text content, the corresponding method of the device is the method for generating the image-text content in the previous embodiment, and the principle of solving the problem is similar to the method. The apparatus comprises a memory for storing computer program instructions and a processor for executing the computer program instructions, wherein the computer program instructions, when executed by the processor, trigger the apparatus to perform the aforementioned teletext content generation method.
The specific implementation of the device may be a user device, or a device formed by integrating a user device and a network device through a network. The user equipment comprises an input and output device for realizing interaction with a user, and the data processing process can be realized locally in the user equipment or in network equipment and provides a processing result to the user equipment through a network. The user equipment comprises but is not limited to various terminal equipment such as a computer, a mobile phone and a tablet computer; including but not limited to implementations such as a network host, a single network server, multiple sets of network servers, or a cloud-computing-based collection of computers. Here, the Cloud is made up of a large number of hosts or web servers based on Cloud Computing (Cloud Computing), which is a type of distributed Computing, one virtual computer consisting of a collection of loosely coupled computers.
Fig. 7 shows a structure of a device suitable for implementing the method and/or technical solution in the embodiment of the present application, and the device 700 includes a Central Processing Unit (CPU)701, which can execute various suitable actions and processes according to a program stored in a Read Only Memory (ROM) 702 or a program loaded from a storage portion 708 into a Random Access Memory (RAM) 703. In the RAM703, various programs and data necessary for system operation are also stored. The CPU 701, the ROM 702, and the RAM703 are connected to each other via a bus 704. An Input/Output (I/O) interface 705 is also connected to the bus 704.
The following components are connected to the I/O interface 705: an input portion 706 including a keyboard, a mouse, a touch screen, a microphone, an infrared sensor, and the like; an output section 707 including a Display panel such as a Cathode Ray Tube (CRT), a Liquid Crystal Display (LCD), an LED Display, an OLED Display, and the like, and a speaker; a storage portion 708 comprising one or more computer-readable media such as a hard disk, optical disk, magnetic disk, semiconductor memory, or the like; and a communication section 709 including a Network interface card such as a LAN (Local Area Network) card, a modem, or the like. The communication section 709 performs communication processing via a network such as the internet.
In particular, the methods and/or embodiments in the embodiments of the present application may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. The computer program, when executed by a Central Processing Unit (CPU)701, performs the above-described functions defined in the method of the present application.
It should be noted that the computer readable medium described herein can be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present application, a computer readable medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
In this application, however, a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wire, fiber optic cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present application may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
The flowchart or block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
As another aspect, the present application also provides a computer-readable medium, which may be contained in the apparatus described in the above embodiments; or may be separate and not incorporated into the device. The computer-readable medium carries one or more computer program instructions that are executable by a processor to implement the methods and/or aspects of the embodiments of the present application as described above.
In summary, in the solutions provided in some embodiments of the present application, a text input by a user can be determined in response to a text input operation of the user, the text is displayed in a display interface, a keyword is determined according to the text, a picture matched with the keyword is obtained, and image-text content including the text and the picture is generated and displayed in the display interface. In the whole process, a user does not need to manually search and select matching, the picture matched with the text content can be directly and automatically found by the input text, and then the image-text content containing the picture and the characters is generated, so that the user experience can be improved, the user can effectively express the emotion in the social contact process, and a large amount of complicated operations of the user are not needed to be added.
It should be noted that the present application may be implemented in software and/or a combination of software and hardware, for example, implemented using Application Specific Integrated Circuits (ASICs), general purpose computers or any other similar hardware devices. In some embodiments, the software programs of the present application may be executed by a processor to implement the above steps or functions. Likewise, the software programs (including associated data structures) of the present application may be stored in a computer readable recording medium, such as RAM memory, magnetic or optical drive or diskette and the like. Additionally, some of the steps or functions of the present application may be implemented in hardware, for example, as circuitry that cooperates with the processor to perform various steps or functions.
It will be evident to those skilled in the art that the present application is not limited to the details of the foregoing illustrative embodiments, and that the present application may be embodied in other specific forms without departing from the spirit or essential attributes thereof. The present embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the application being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein. Any reference sign in a claim should not be construed as limiting the claim concerned. Furthermore, it is obvious that the word "comprising" does not exclude other elements or steps, and the singular does not exclude the plural. A plurality of units or means recited in the apparatus claims may also be implemented by one unit or means in software or hardware. The terms first, second, etc. are used to denote names, but not any particular order.

Claims (8)

1. A method for generating teletext content, wherein the method comprises:
in response to a text input operation of a user, determining a text input by the user, and displaying the text in a display interface;
when a trigger character is detected at any position or a preset position in the text, performing word segmentation processing on the text to obtain word segmentation phrases contained in the text, performing semantic analysis on the text, and determining semantic information corresponding to the text;
determining candidate keywords according to the word segmentation phrases and semantic information, and displaying the candidate keywords to a user;
responding to the keyword adjusting operation of the user, adjusting the candidate keywords and determining the keywords;
obtaining a picture matched with the keyword and used for realizing social contact;
and generating and displaying the image-text content containing the text and the picture in the display interface.
2. The method of claim 1, wherein presenting the candidate keywords to a user comprises:
after determining the candidate keywords, setting a keyword display area on the display interface;
and displaying the candidate keywords in a label form in the keyword display area.
3. The method of claim 1 or 2, wherein obtaining a picture matching the keyword comprises:
determining a core keyword among the keywords;
and performing picture matching in a preset picture data set according to the keywords to obtain pictures matched with the keywords, wherein the weight of the core keywords is greater than that of the non-core keywords when the pictures are matched.
4. The method of claim 1, wherein obtaining a picture matching the keyword comprises:
acquiring candidate pictures matched with the keywords, and displaying the candidate pictures to a user;
in response to a selection operation of a user, determining a target picture in the candidate pictures;
generating and displaying the image-text content containing the text and the picture in the display interface, wherein the image-text content comprises:
generating and displaying the image-text content containing the text and the target picture in the display interface.
5. The method of claim 1, wherein generating and displaying in the display interface teletext content comprising the text and the picture comprises:
and inserting the picture into a preset area of a display interface on which the text is displayed, so that the display interface displays the image-text content comprising the text and the picture.
6. The method of claim 5, wherein the display interface comprises at least any one of: UGC publication interface or communication interface.
7. Teletext content generation device comprising a memory for storing computer program instructions and a processor for executing the computer program instructions, wherein the computer program instructions, when executed by the processor, trigger the device to perform the method of any one of claims 1 to 6.
8. A computer readable medium having stored thereon computer program instructions executable by a processor to implement the method of any one of claims 1 to 6.
CN201910248465.1A 2019-03-29 2019-03-29 Image-text content generation method and equipment Active CN110020411B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910248465.1A CN110020411B (en) 2019-03-29 2019-03-29 Image-text content generation method and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910248465.1A CN110020411B (en) 2019-03-29 2019-03-29 Image-text content generation method and equipment

Publications (2)

Publication Number Publication Date
CN110020411A CN110020411A (en) 2019-07-16
CN110020411B true CN110020411B (en) 2020-10-09

Family

ID=67190215

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910248465.1A Active CN110020411B (en) 2019-03-29 2019-03-29 Image-text content generation method and equipment

Country Status (1)

Country Link
CN (1) CN110020411B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110674634A (en) * 2019-09-10 2020-01-10 吴杰 Character interaction method and server equipment
CN110865833A (en) * 2019-09-29 2020-03-06 上海萌家网络科技有限公司 Application client, server and method for updating user state
CN110879840A (en) * 2019-11-19 2020-03-13 珠海格力电器股份有限公司 Information feedback method, device and storage medium
CN111158817A (en) * 2019-12-24 2020-05-15 维沃移动通信有限公司 Information processing method and electronic equipment
WO2021160100A1 (en) * 2020-02-12 2021-08-19 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Methods for searching images and for indexing images, and electronic device
CN113448461A (en) * 2020-06-24 2021-09-28 北京新氧科技有限公司 Information processing method, device and equipment
CN112036128A (en) * 2020-08-21 2020-12-04 百度在线网络技术(北京)有限公司 Text content processing method, device, equipment and storage medium
CN114338572B (en) * 2020-09-28 2023-07-18 荣耀终端有限公司 Information processing method, related device and storage medium
CN112364610A (en) * 2020-12-01 2021-02-12 深圳市房多多网络科技有限公司 Method and device for inserting building card in house source article and computing equipment
CN112613270A (en) * 2020-12-22 2021-04-06 百色学院 Method, system, equipment and storage medium for recommending target text based on pre-trained drawing creation model
CN113536006B (en) * 2021-06-25 2023-06-13 北京百度网讯科技有限公司 Method, apparatus, device, storage medium and computer product for generating picture
CN115546355B (en) * 2022-11-28 2023-03-31 北京红棉小冰科技有限公司 Text matching method and device

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107015979B (en) * 2016-01-27 2021-04-06 斑马智行网络(香港)有限公司 Data processing method and device and intelligent terminal
CN107305566B (en) * 2016-04-21 2019-10-18 北京搜狗科技发展有限公司 A kind of method and device to search for information matches picture
CN106528588A (en) * 2016-09-14 2017-03-22 厦门幻世网络科技有限公司 Method and apparatus for matching resources for text information
KR20190002073A (en) * 2017-06-29 2019-01-08 네이버 주식회사 Method and system for automatic activation of machine
CN108052563A (en) * 2017-12-05 2018-05-18 广东小天才科技有限公司 A kind of picture searching method and mobile terminal
CN108733779B (en) * 2018-05-04 2022-10-04 百度在线网络技术(北京)有限公司 Text matching method and device

Also Published As

Publication number Publication date
CN110020411A (en) 2019-07-16

Similar Documents

Publication Publication Date Title
CN110020411B (en) Image-text content generation method and equipment
US10645142B2 (en) Video keyframes display on online social networks
RU2632144C1 (en) Computer method for creating content recommendation interface
US9977835B2 (en) Queryless search based on context
US10140314B2 (en) Previews for contextual searches
CN111095215B (en) Inter-application delivery format specific data objects
US10289727B2 (en) Incorporation of semantic attributes within social media
WO2017124116A1 (en) Searching, supplementing and navigating media
US20170357521A1 (en) Virtual keyboard with intent-based, dynamically generated task icons
CN114375435A (en) Enhancing tangible content on a physical activity surface
JP6361351B2 (en) Method, program and computing system for ranking spoken words
WO2017076073A1 (en) Method and apparatus for search and recommendation
US20230130438A1 (en) Method and apparatus for providing multimedia content, and device
KR101911999B1 (en) Feature-based candidate selection
US9818080B2 (en) Categorizing a use scenario of a product
US10169374B2 (en) Image searches using image frame context
CN109726712A (en) Character recognition method, device and storage medium, server
CN110851712B (en) Method, device and computer readable medium for recommending book information
CN116821475B (en) Video recommendation method and device based on client data and computer equipment
CN112667118A (en) Method, apparatus and computer readable medium for displaying historical chat messages
CN110827058A (en) Multimedia promotion resource insertion method, equipment and computer readable medium
US9619519B1 (en) Determining user interest from non-explicit cues
US20190227634A1 (en) Contextual gesture-based image searching
WO2023124793A1 (en) Image pushing method and device
CN113157170B (en) Data labeling method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant