WO2021042234A1 - Application introduction method, mobile terminal, and server - Google Patents

Application introduction method, mobile terminal, and server Download PDF

Info

Publication number
WO2021042234A1
WO2021042234A1 PCT/CN2019/104000 CN2019104000W WO2021042234A1 WO 2021042234 A1 WO2021042234 A1 WO 2021042234A1 CN 2019104000 W CN2019104000 W CN 2019104000W WO 2021042234 A1 WO2021042234 A1 WO 2021042234A1
Authority
WO
WIPO (PCT)
Prior art keywords
keywords
application
information
introduction
mobile terminal
Prior art date
Application number
PCT/CN2019/104000
Other languages
French (fr)
Chinese (zh)
Inventor
艾静雅
柳彤
朱大卫
汤慧秀
Original Assignee
深圳海付移通科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳海付移通科技有限公司 filed Critical 深圳海付移通科技有限公司
Priority to CN201980010315.5A priority Critical patent/CN111801673A/en
Priority to PCT/CN2019/104000 priority patent/WO2021042234A1/en
Publication of WO2021042234A1 publication Critical patent/WO2021042234A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • G10L21/10Transforming into visible information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/60Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals

Definitions

  • This application relates to the technical field of application programs, in particular to an application program introduction method, mobile terminal and server.
  • this application provides an application introduction method, mobile terminal and server. On the one hand, it can adapt to different user groups so that the application can meet the needs of more user groups. On the other hand, it adopts the form of animation.
  • the introduction of the application can increase the personalization of the introduction of the application, increase the interest, and improve the user experience.
  • the first technical solution adopted by this application is to provide an application introduction method, including: obtaining introduction requirement information about the application; wherein the introduction requirement information is used to indicate the requirement for introducing the application; and the introduction requirement information is extracted Keywords; obtain related images and voices based on keywords; process images and voices to form a video for introducing the application.
  • the introduction demand information is audio information
  • extracting keywords in the introduction demand information includes: performing voice recognition on the audio information to obtain text information; performing keyword extraction on the text information to obtain keywords.
  • keyword extraction is performed on text information to obtain keywords, including: semantic segmentation of text information; keywords are obtained based on the result of semantic segmentation.
  • performing semantic segmentation on text information includes: inputting the text information into a convolutional neural network for deep learning, so as to perform semantic segmentation on the text information to obtain keywords.
  • the introduction demand information is text information
  • the extraction of keywords in the introduction demand information includes: semantic segmentation of the text information; keywords are obtained based on the result of the semantic segmentation.
  • obtaining the associated image and voice based on the keyword includes: sending the keyword to the server so that the server generates the associated image and voice based on the keyword; obtaining the image and voice sent by the server.
  • the image and voice are processed to form a video for introducing the application, including: image segmentation of multiple corresponding images, extraction of feature information in the image; combination of feature information to generate multiple images Frame; multiple image frames are formed into animation; animation and voice are merged to form a video used to introduce the application.
  • the method further includes: obtaining background music sent by the server; wherein, the background music is music generated by the server based on keywords; and adding the background music to the video.
  • the second technical solution adopted in this application is to provide an application introduction method, including: acquiring keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application.
  • the introduction demand information is used to express the demand for introducing the application; generate related images and voice based on keywords; send the image and voice to the mobile terminal, so that the mobile terminal can process the image and voice to form the application program Introductory video.
  • generating related images and voices based on keywords includes: deep learning the keywords to obtain related images from a preset image library.
  • generating related images and voices based on keywords includes: applying keywords through deep learning to generate text information that meets the keyword scene; and converting text information into voice.
  • a mobile terminal which includes a processor and a memory connected to the processor; the memory is used to store program data, and the processor is used to execute the program data, so as to implement the first solution described above.
  • the server includes a processor and a memory connected to the processor; the memory is used to store program data, and the processor is used to execute the program data, so as to implement the above-mentioned second solution.
  • Another technical solution adopted by the present application is to provide a computer storage medium, which is used to store program data, and when the program data is executed by a processor, it is used to implement any of the methods provided in the above-mentioned solutions.
  • the mobile terminal includes: an acquisition module for acquiring introduction requirement information about an application program; wherein the introduction requirement information is used to indicate a requirement for introducing an application program; extraction The module is used to extract keywords in the introduction demand information; the acquisition module is also used to obtain related images and voices based on the keywords; the processing module is used to process the images and voices to form an introduction to the application Video.
  • the server includes: an acquisition module for acquiring keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application , The introduction demand information is used to express the demand for introducing the application; the processing module is used to generate associated images and voices based on keywords; the sending module is used to send images and voices to the mobile terminal so that the mobile terminal can respond to the image And voice processing to form a video for introducing the application.
  • an application introduction method of this application includes: obtaining introduction requirement information about the application; wherein, the introduction requirement information is used to indicate information about the introduction of the application. Demand; extract the keywords in the introduction demand information; obtain related images and voices based on the keywords; process the images and voices to form a video for introducing the application program.
  • FIG. 1 is a schematic flowchart of a first embodiment of an application program introduction method provided by the present application
  • FIG. 2 is a schematic flowchart of a second embodiment of an application program introduction method provided by the present application
  • FIG. 3 is a schematic flowchart of a third embodiment of an application program introduction method provided by the present application.
  • FIG. 4 is a schematic flowchart of a fourth embodiment of an application program introduction method provided by the present application.
  • FIG. 5 is a schematic flowchart of a fifth embodiment of an application program introduction method provided by the present application.
  • FIG. 6 is a schematic flowchart of a sixth embodiment of an application program introduction method provided by the present application.
  • FIG. 7 is a schematic flowchart of a seventh embodiment of an application program introduction method provided by the present application.
  • FIG. 8 is a schematic structural diagram of a first embodiment of a mobile terminal provided by the present application.
  • FIG. 9 is a schematic structural diagram of a first embodiment of a server provided by the present application.
  • FIG. 10 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application.
  • FIG. 11 is a schematic structural diagram of a second embodiment of a mobile terminal provided by the present application.
  • Fig. 12 is a schematic structural diagram of a second embodiment of a server provided by the present application.
  • Fig. 1 is a schematic flowchart of a first embodiment of the application introduction method provided by the present application. The method is implemented based on a mobile terminal, and the method includes:
  • Step 11 Obtain the introduction requirement information about the application; among them, the introduction requirement information is used to indicate the requirement for the introduction of the application.
  • the mobile terminal after the mobile terminal responds to the user to download the application and the installation is complete, it obtains the introduction requirement information about the application.
  • the introduction requirement information may be audio information or text information.
  • the audio information is collected by the microphone of the mobile terminal, and the text information can be manually input, or keywords prompted by the application can be selected as the text information.
  • the introduction requirement information is used to indicate the requirement for the introduction application. For example, when users need to know the income of financial management applications, they can use “income” as the introduction demand information.
  • Step 12 Extract the keywords in the introduction requirement information.
  • the mobile terminal performs keyword extraction on the content of the introduction requirement information.
  • the obtained introduction requirement information is audio information
  • the text information parsed from the audio information is "How is this application safe?" "Payment", the extracted keyword is "secure payment”.
  • the keyword extraction method can be based on a keyword extraction algorithm based on statistical features.
  • the keyword extraction algorithm uses the statistical information of the words in the document to extract the keywords of the document.
  • the text is preprocessed to obtain a set of candidate words, and then keywords are obtained from the candidate set by means of feature value quantification.
  • the feature value quantization methods include feature quantification based on word weight, feature quantification based on word document location, and feature quantification based on word-related information.
  • Feature quantification based on word weight mainly includes part-of-speech, word frequency, inverse document frequency, relative word frequency, word length, etc.; word-based feature quantification of document position is based on the assumption that sentences at different positions of the article have different importance to the document.
  • words in the first N words, last N words, beginning of paragraph, end of paragraph, title, introduction, etc. of the article are representative.
  • Word related information refers to the degree of relevance information between words and words, words and documents, including mutual information, hits value, contribution degree, dependency degree, TF-IDF value, etc.
  • Keyword extraction methods can also be based on deep learning methods.
  • Step 13 Obtain related images and voices based on keywords.
  • acquiring the associated image may be that the mobile terminal sends the keyword to the server, and the server performs image retrieval in a preset image library to obtain multiple images.
  • obtaining the associated voice may be that the mobile terminal sends the keywords to the server, and the server generates multiple texts conforming to the application scenarios through the keywords and application scenarios, and then sends the multiple texts to the mobile terminal, and then the mobile terminal sends the multiple texts to the mobile terminal. Convert text messages into voice messages.
  • acquiring the associated image may be that the mobile terminal performs image retrieval in a local preset image library to obtain multiple images.
  • acquiring the associated voice may be that the mobile terminal generates multiple paragraphs of text that conform to the application scenario through keywords and application scenarios, and then converts the text information into voice information.
  • Step 14 Process the image and voice to form a video for introducing the application.
  • the keywords are "birds" and "trees”; one image has the characteristic information of a tree, and another image has the characteristic information of a bird. These two pieces of information can be extracted to form a bird stop according to the scene. Image on the tree. After composing a series of complete images, the image is smoothed and other enhanced processing, the purpose is to make the image content more natural.
  • the voice information and the image information are merged to form a video for introducing the application program.
  • the application reminds the user to say what he wants to know.
  • the audio information collected by the mobile terminal at this time is "I open an account for the first time, how can I invest in order to achieve high returns and The risk is small, and there is also how to pay.
  • the keywords extracted by the mobile terminal are "open account for the first time”, “investment”, “high return”, “small risk”, and “how to pay”. Then, based on these keywords, search for the corresponding images in the preset image library. For example, “open an account for the first time” will search for the account opening screen and images of animated characters, and “high return” and “low risk” will search for warnings and recommendations.
  • the image of the product is then formed into a section with an animated character explaining how to invest in the best way to ensure low risk, and then there is another picture where payment safety is also very important.
  • text information that meets the scene is generated, the text information is converted into voice, and the voice information and image information are merged to form a video that introduces user needs.
  • background music can also be added to the video.
  • the mobile terminal can generate an application introduction corresponding to the sound and image according to a piece of voice, or tell a story to the child through voice.
  • the child can describe the type of story he likes to listen to, and generate a short story with pictures and texts through machine learning. , So that children are more interested, and it can also make some children who are not literate can acquire corresponding knowledge through animation.
  • the mobile terminal According to different application programs and different user requirements, the mobile terminal generates a video with pictures and texts corresponding to the application program for the user to watch.
  • an application introduction method of this application includes: obtaining introduction requirement information about the application; wherein the introduction requirement information is used to indicate the requirement for introducing the application; and the introduction requirement information is extracted Keywords; obtain related images and voices based on keywords; process images and voices to form a video for introducing the application.
  • Figure 2 is a schematic flowchart of a second embodiment of the application introduction method provided by the present application. The method is implemented based on a mobile terminal, and the method includes:
  • Step 21 Acquire audio information about the application program; among them, the audio information is used to indicate the demand for introducing the application program.
  • the user's audio information is collected to indicate the demand for introducing the application program.
  • the audio information may be audio information related to the application that the user wants to learn about the application.
  • the audio information may be text information displayed to the user after the application is started to prompt the user to learn about the application, so that the user can quickly speak the corresponding keyword information.
  • Step 22 Perform voice recognition on the audio information to obtain text information.
  • speech recognition is to convert a piece of audio information into corresponding text information.
  • the system mainly includes four parts: feature extraction, acoustic model, language model, dictionary and decoding.
  • the collected Perform audio data preprocessing work such as filtering and framing of the audio information to properly extract the audio information that needs to be analyzed from the original signal;
  • feature extraction converts the audio information from the time domain to the frequency domain to provide a suitable acoustic model Feature vector:
  • the acoustic model calculates the score of each feature vector on the acoustic feature according to the acoustic characteristics; while the language model calculates the probability of the sound signal corresponding to the possible sequence of phrases according to the linguistic theory; finally according to the existing dictionary , Decode the phrase sequence to get the final possible text representation.
  • Step 23 Input the text information into the convolutional neural network for deep learning, so as to perform semantic segmentation on the text information to obtain keywords.
  • a large amount of information is trained in advance through a convolutional neural network for deep learning to generate a corresponding semantic segmentation model.
  • the semantic segmentation model gets the text information, it can get the keywords.
  • Step 24 Obtain related images and voices based on the keywords.
  • Step 25 Process the image and voice to form a video for introducing the application.
  • Steps 24-25 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
  • FIG. 3 is a schematic flowchart of a third embodiment of the application introduction method provided by the present application.
  • the method is implemented based on a mobile terminal, and the method includes:
  • Step 31 Acquire text information about the application program; where the text information is used to indicate a requirement for introducing the application program.
  • the text information may be manually input by the user, or may be generated by the user's selection by the application prompting multiple paragraphs of text.
  • Step 32 Perform semantic segmentation on the text information.
  • Step 33 Obtain keywords based on the result of semantic segmentation.
  • Steps 32-33 can be specifically:
  • Adopt TF-IDF term frequency inverse document frequency, a common weighting technique for information retrieval data mining
  • TextRank a general graph-based sorting algorithm for natural language processing
  • Rake Rapid Automatic Keyword Extraction, fast automatic keyword extraction
  • Topic -Model theme model
  • TF-IDF TF*IDF
  • a semantic segmentation model may be established in advance through deep learning of a neural network to achieve rapid extraction of keywords.
  • Step 34 Obtain related images and voices based on the keywords.
  • Step 35 Process the image and voice to form a video for introducing the application program.
  • Steps 34-35 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
  • FIG. 4 is a schematic flowchart of a fourth embodiment of the application introduction method provided by the present application.
  • the method is implemented based on a mobile terminal, and the method includes:
  • Step 41 Obtain the introduction requirement information about the application program; wherein the introduction requirement information is used to indicate the requirement for the introduction application program.
  • Step 42 Extract keywords in the introduction requirement information.
  • Steps 41-42 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
  • Step 43 Send the keywords to the server, so that the server generates associated images and voices based on the keywords.
  • deep learning based on the convolutional neural network can obtain images and voices associated with the keywords.
  • the image may be obtained by the server, and the voice may be recognized by the mobile terminal on the keywords, so as to generate multiple paragraphs of text that match the scene, and convert them into voice.
  • Step 44 Obtain the image and voice sent by the server.
  • Step 45 Process the image and voice to form a video for introducing the application program.
  • Steps 44-45 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
  • FIG. 5 is a schematic flowchart of a fifth embodiment of the application introduction method provided by the present application, and the method includes:
  • Step 51 Obtain introduction requirement information about the application; where the introduction requirement information is used to indicate the requirement for the introduction of the application.
  • Step 52 Extract keywords in the introduction requirement information.
  • Step 53 Send the keywords to the server, so that the server generates associated images and voices based on the keywords.
  • Step 54 Obtain the image and voice sent by the server.
  • Steps 51-54 have the same or similar technical solutions as the above-mentioned embodiment, and will not be repeated here.
  • Step 55 Perform image segmentation on multiple corresponding images, and extract feature information from the images.
  • Image segmentation is the technique and process of dividing an image into a number of specific areas with unique properties and proposing objects of interest. It is a key step from image processing to image analysis.
  • the existing image segmentation methods are mainly divided into the following categories: threshold-based segmentation methods, region-based segmentation methods, edge-based segmentation methods, and segmentation methods based on specific theories.
  • image segmentation is the process of dividing a digital image into disjoint areas.
  • the process of image segmentation is also a marking process, that is, the pixels belonging to the same area are assigned the same number.
  • the threshold-based segmentation method is a region-based image segmentation technology, the principle is to divide the image pixels into several categories.
  • Image thresholding segmentation is one of the most commonly used traditional image segmentation methods. It has become the most basic and most widely used segmentation technique in image segmentation due to its simple implementation, small calculation amount, and stable performance. It is especially suitable for images where the target and background occupy different gray scale ranges. It can not only greatly compress the amount of data, but also greatly simplifies the analysis and processing steps. Therefore, in many cases, it is a necessary image preprocessing process before image analysis, feature extraction and pattern recognition.
  • the purpose of image thresholding is to divide the pixel set according to the gray level, and each of the obtained subsets forms an area corresponding to the real scene. Each area has the same attributes, while the adjacent areas do not have this Consistent attributes. Such division can be achieved by selecting one or more thresholds starting from the gray level.
  • the region-based segmentation method is a segmentation technique based on directly finding the region.
  • the specific algorithms include region growth and region separation and merging algorithms.
  • region growth which starts from a single pixel and gradually merges to form the required segmentation area; the other is to start from the overall situation and gradually cut to the required segmentation area.
  • edge-based segmentation mainly includes point-based detection, line-based detection, and edge-based detection.
  • segmentation methods based on specific theories can be divided into cluster analysis, fuzzy set theory, gene coding, wavelet transform and other methods.
  • feature extraction is performed based on the keywords and the scene, to perform step 56.
  • Step 56 Combine the feature information to generate multiple image frames.
  • Step 57 Form multiple image frames into animation.
  • steps 55-57 are specifically:
  • Deep learning is performed in advance through the convolutional neural network to establish an image model so that the corresponding feature information generates multiple image frames, and then the multiple image frames are formed into an animation.
  • Step 58 The animation and voice are merged to form a video for introducing the application.
  • FIG. 6 is a schematic flowchart of a sixth embodiment of the application introduction method provided by the present application.
  • the method is implemented based on a server, and the method includes:
  • Step 61 Acquire keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction requirement information about the application, and the introduction requirement information is used to indicate the requirement for introducing the application.
  • the mobile terminal After the mobile terminal obtains the introduction requirement information about the application, it extracts the keywords and sends the keywords to the server.
  • Step 62 Generate associated images and voices based on the keywords.
  • the server performs model training through the relevant content of the application in advance, so that when the keyword of the mobile terminal is obtained, it responds quickly and obtains the image and voice associated with the keyword.
  • Step 63 Send the image and voice to the mobile terminal, so that the mobile terminal processes the image and voice to form a video for introducing the application program.
  • the generated image and voice are sent to the mobile terminal, so that the mobile terminal extracts feature information of the image, and then combines the feature information to generate multiple image frames and combine to generate multiple image frames.
  • the mobile terminal merges multiple image frames into animation and voice to form a video for introducing the application.
  • an application introduction method of this application includes: acquiring keywords sent by a mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction requirement information about the application.
  • the introduction demand information is used to express the demand for introducing the application; generate related images and voice based on keywords; send the image and voice to the mobile terminal, so that the mobile terminal can process the image and voice to form the application program Introductory video.
  • FIG. 7 is a schematic flowchart of a seventh embodiment of the application introduction method provided by the present application, and the method includes:
  • Step 71 Acquire keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction requirement information about the application, and the introduction requirement information is used to indicate the requirement for introducing the application.
  • Step 72 Pass the keywords through deep learning to obtain associated images from the preset image library.
  • Deep learning models include convolutional neural network (convolutional neural network), DBN (Deep Belief Network, deep trust network model) and stacked auto-encoder network models.
  • Convolutional neural networks are inspired by the structure of the visual system.
  • the first convolutional neural network calculation model was proposed in the neurocognitive machine. Based on the local connection between neurons and the layered organization image conversion, the neurons with the same parameters are applied to the difference of the previous layer of neural network. Position, get a translation-invariant neural network structure. Later, on the basis of this idea, a convolutional neural network was designed and trained with error gradients to obtain superior performance on some pattern recognition tasks.
  • DBN can be interpreted as a Bayesian probability generation model, which is composed of multiple layers of random latent variables.
  • the upper two layers have undirected symmetrical connections, and the lower layer gets top-down directed connections from the upper layer, and the lowest layer unit
  • the state of is the visible input data vector.
  • the DBN is composed of a stack of 2F structural units, and the structural unit is usually RBM (RestIlcted Boltzmann Machine, Restricted Boltzmann Machine).
  • RBM RasterIlcted Boltzmann Machine, Restricted Boltzmann Machine
  • the input samples are used to train the first-layer RBM units, and their output is used to train the second-layer RBM models, and the RBM models are stacked to improve the model performance by adding layers.
  • the unsupervised pre-training process after the DBN code is input to the top RBM, the state of the top layer is decoded to the unit of the bottom layer to realize the reconstruction of the input.
  • RBM shares parameters with each layer of DBN.
  • the structure of the stacked self-encoding network is similar to that of the DBN, consisting of a stack of several structural units. The difference is that the structural unit is an auto-en-coder instead of RBM.
  • the self-encoding model is a two-layer neural network, the first layer is called the coding layer, and the second layer is called the decoding layer.
  • the server needs to generate a corresponding scene prediction according to the characteristics of the application and the characteristics of the keywords, and search for the corresponding image according to the scene.
  • the server will search for images in the Internet.
  • Step 73 Pass the keywords through deep learning to generate text information that meets the keyword scene.
  • Step 74 Convert the text information into voice.
  • Step 75 Send the image and voice to the mobile terminal, so that the mobile terminal processes the image and voice to form a video for introducing the application program.
  • the server first retrieves a large number of images, sends the images to the mobile terminal, and the mobile terminal divides the images according to keywords, and then combines them according to the scene to form an animation. It is voice fusion to form a video for introducing the application.
  • FIG. 8 is a schematic structural diagram of a first embodiment of a mobile terminal provided by the present application.
  • the mobile terminal 80 includes a processor 81 and a memory 82 connected to the processor 81; the memory 82 is used to store program data, and the processor 81 Used to execute program data to implement the following methods:
  • the introduction demand information is used to express the demand for introducing the application; extract the keywords in the introduction demand information; obtain the associated images and voices based on the keywords; process the images and voices To form a video for introducing the application.
  • the processor 81 for executing the program data is also used to implement the following methods: perform voice recognition on audio information to obtain text information; and perform keyword extraction on text information to obtain keywords.
  • the processor 81 is used to execute the program data to implement the following method: perform semantic segmentation on the text information; obtain keywords based on the result of the semantic segmentation.
  • the processor 81 used to execute the program data is also used to implement the following method: input text information into a convolutional neural network for deep learning, so as to perform semantic segmentation on the text information to obtain keywords.
  • the processor 81 is used to execute the program data to implement the following method: perform semantic segmentation on the text information; obtain keywords based on the result of the semantic segmentation.
  • the processor 81 is used to execute the program data to implement the following method: sending keywords to the server so that the server generates associated images and voices based on the keywords; acquiring images and voices sent by the server.
  • the processor 81 is configured to execute the program data to implement the following method: image segmentation of multiple corresponding images, extraction of feature information in the image; combination of feature information to generate multiple image frames; Multiple image frames are formed into animation; animation and voice are merged to form a video for introducing the application.
  • the processor 81 is used to execute the program data to implement the following method: acquiring background music sent by the server; wherein the background music is music generated by the server based on keywords; adding the background music to the video.
  • FIG. 9 is a schematic structural diagram of a first embodiment of a server provided by the present application.
  • the server 90 includes a processor 91 and a memory 92 connected to the processor 91; the memory 92 is used to store program data, and the processor 91 is used to store program data.
  • the program data is executed to achieve the following methods:
  • Acquire keywords sent by the mobile terminal among them, the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application, and the introduction demand information is used to indicate the demand for introducing the application; the associated image is generated based on the keywords And voice; send images and voices to the mobile terminal so that the mobile terminal can process the images and voices to form a video for introducing the application.
  • the processor 91 used to execute the program data is also used to implement the following method: pass keywords through deep learning to obtain associated images from a preset image library.
  • the processor 91 used to execute the program data is also used to implement the following method: pass keywords through deep learning to generate text information that meets the keyword scene; convert the text information into voice
  • FIG. 10 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application.
  • the computer storage medium 100 is used to store program data 101.
  • the program data 101 is executed by a processor, it is used to implement the following methods:
  • the introduction demand information is used to express the demand for introducing the application; extract the keywords in the introduction demand information; obtain the associated images and voices based on the keywords; process the images and voices , To form a video for introducing the application;
  • the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application, and the introduction demand information is used to indicate the demand for introducing the application; and the correlation is generated based on the keywords
  • the image and voice of the mobile terminal send the image and voice to the mobile terminal so that the mobile terminal can process the image and voice to form a video for introducing the application.
  • the computer storage medium can be applied to the above-mentioned mobile terminal or the above-mentioned server to implement the method of any one of the above-mentioned embodiments.
  • FIG. 11 is a schematic structural diagram of a second embodiment of a mobile terminal provided by the present application.
  • the mobile terminal 110 includes: an acquisition module 111, an extraction module 112, and a processing module 113.
  • the obtaining module 111 is used for obtaining introduction requirement information about the application program; wherein, the introduction requirement information is used to indicate the requirement for introducing the application program;
  • the extraction module 112 is used to extract keywords in the introduction demand information
  • the obtaining module 111 is also used to obtain related images and voices based on keywords;
  • the processing module 113 is used to process images and voices to form a video for introducing the application program.
  • FIG. 12 is a schematic structural diagram of a second embodiment of a server provided by the present application.
  • the server 120 includes: an obtaining module 121, a processing module 122, and a sending module 123.
  • the obtaining module 121 is used to obtain keywords sent by the mobile terminal; wherein, the keywords are extracted by the mobile terminal based on the obtained introduction requirement information about the application, and the introduction requirement information is used to indicate the requirement for introducing the application;
  • the processing module 122 is configured to generate associated images and voices based on keywords
  • the sending module 123 is configured to send images and voices to the mobile terminal, so that the mobile terminal processes the images and voices to form a video for introducing the application program.
  • the disclosed method and device may be implemented in other ways.
  • the device implementation described above is only illustrative.
  • the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, for example, multiple units or components may be Combined or can be integrated into another system, or some features can be ignored or not implemented.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of this embodiment.
  • the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
  • the integrated unit in the other embodiments described above is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
  • the technical solution of the present application essentially or the part that contributes to the existing technology or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , Including several instructions to enable a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute all or part of the steps of the methods described in the various embodiments of the present application.
  • the aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks and other media that can store program codes. .

Abstract

An application introduction method, a mobile terminal, and a server. The method comprises: obtaining introduction requirement information about an application, wherein the introduction requirement information is used for indicating a requirement for introducing the application (11); extracting a keyword in the introduction requirement information (12); obtaining associated image and voice on the basis of the keyword (13); and processing the image and the voice to form a video for introducing the application (14). On the one hand, the method can adapt to different user groups, so that the application satisfies the requirements of more user groups. On the other hand, the application is introduced in the form of animation, making the introduction of the application more personalized and more interesting, thereby improving user experience.

Description

应用程序的介绍方法、移动终端及服务器Application introduction method, mobile terminal and server 【技术领域】【Technical Field】
本申请涉及应用程序技术领域,具体涉及一种应用程序的介绍方法、移动终端及服务器。This application relates to the technical field of application programs, in particular to an application program introduction method, mobile terminal and server.
【背景技术】【Background technique】
随着移动终端的普及,在移动终端上使用的应用程序也越来越多。用户在移动终端下载应用程序后,通常希望能在短时间内了解该应用程序的使用方法,使用场景,以及需要关注的部分,与自身更相关的部分等。如支付与金融理财等应用程序,主要的应用程序介绍,均是一些常规图片文字,这些比较固定,没有吸引性,会显得死板没有个性以及无趣。With the popularity of mobile terminals, more and more applications are used on mobile terminals. After downloading an application on a mobile terminal, a user usually hopes to understand the usage method, usage scenario, and parts that need attention, and more relevant parts of the application within a short period of time. For applications such as payment and financial management, the main application introductions are all regular pictures and texts. These are relatively fixed, unattractive, and will appear rigid, non-personal and boring.
【发明内容】[Summary of the invention]
为了解决上述问题,本申请提供一种应用程序的介绍方法、移动终端及服务器,一方面能够适应于不同的用户群体,使应用程序满足更多用户群体的需求,另一方面采用动画的形式进行应用程序的介绍能够增加应用程序介绍的个性化,增加趣味性,提高用户体验。In order to solve the above-mentioned problems, this application provides an application introduction method, mobile terminal and server. On the one hand, it can adapt to different user groups so that the application can meet the needs of more user groups. On the other hand, it adopts the form of animation. The introduction of the application can increase the personalization of the introduction of the application, increase the interest, and improve the user experience.
本申请采用的第一种技术方案是提供一种应用程序的介绍方法,包括:获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求;提取介绍需求信息中的关键词;基于关键词获取相关联的图像和语音;对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。The first technical solution adopted by this application is to provide an application introduction method, including: obtaining introduction requirement information about the application; wherein the introduction requirement information is used to indicate the requirement for introducing the application; and the introduction requirement information is extracted Keywords; obtain related images and voices based on keywords; process images and voices to form a video for introducing the application.
其中,介绍需求信息为音频信息;提取介绍需求信息中的关键词,包括:对音频信息进行语音识别,以得到文本信息;对文本信息进行关键词提取,以得到关键词。Among them, the introduction demand information is audio information; extracting keywords in the introduction demand information includes: performing voice recognition on the audio information to obtain text information; performing keyword extraction on the text information to obtain keywords.
其中,对文本信息进行关键词提取,以得到关键词,包括:对文本信息进行语义分割;基于语义分割的结果得到关键词。Wherein, keyword extraction is performed on text information to obtain keywords, including: semantic segmentation of text information; keywords are obtained based on the result of semantic segmentation.
其中,对文本信息进行语义分割,包括:将文本信息输入至卷积神经网络进行深度学习,以将文本信息进行语义分割,以得到关键词。Wherein, performing semantic segmentation on text information includes: inputting the text information into a convolutional neural network for deep learning, so as to perform semantic segmentation on the text information to obtain keywords.
其中,介绍需求信息为文本信息;提取介绍需求信息中的关键词,包括:对文本信息进行语义分割;基于语义分割的结果得到关键词。Among them, the introduction demand information is text information; the extraction of keywords in the introduction demand information includes: semantic segmentation of the text information; keywords are obtained based on the result of the semantic segmentation.
其中,基于关键词获取相关联的图像和语音,包括:将关键词发送给服务器,以使服务器基于关键词生成相关联的图像和语音;获取服务 器发送的图像和语音。Among them, obtaining the associated image and voice based on the keyword includes: sending the keyword to the server so that the server generates the associated image and voice based on the keyword; obtaining the image and voice sent by the server.
其中,对图像和语音进行处理,以形成用于对应用程序进行介绍的视频,包括:对多个对应的图像进行图像分割,提取图像中特征信息;将特征信息进行组合,以生成多个图像帧;将多个图像帧形成动画;将动画与语音进行融合,以形成用于对应用程序进行介绍的视频。Among them, the image and voice are processed to form a video for introducing the application, including: image segmentation of multiple corresponding images, extraction of feature information in the image; combination of feature information to generate multiple images Frame; multiple image frames are formed into animation; animation and voice are merged to form a video used to introduce the application.
其中,该方法还包括:获取服务器发送的背景音乐;其中,背景音乐是服务器基于关键词生成的音乐;将背景音乐添加至视频。Wherein, the method further includes: obtaining background music sent by the server; wherein, the background music is music generated by the server based on keywords; and adding the background music to the video.
本申请采用的第二种技术方案是提供一种应用程序的介绍方法,包括:获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于表示对于介绍应用程序的需求;基于关键词生成相关联的图像和语音;向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。The second technical solution adopted in this application is to provide an application introduction method, including: acquiring keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application. The introduction demand information is used to express the demand for introducing the application; generate related images and voice based on keywords; send the image and voice to the mobile terminal, so that the mobile terminal can process the image and voice to form the application program Introductory video.
其中,基于关键词生成相关联的图像和语音,包括:将关键词通过深度学习,以从预设图像库得到相关联的图像。Among them, generating related images and voices based on keywords includes: deep learning the keywords to obtain related images from a preset image library.
其中,基于关键词生成相关联的图像和语音,包括:将关键词通过深度学习,以生成符合关键词场景的文字信息;将文字信息转换为语音。Among them, generating related images and voices based on keywords includes: applying keywords through deep learning to generate text information that meets the keyword scene; and converting text information into voice.
本申请采用的另一种技术方案是提供一种移动终端,移动终端包括处理器以及与处理器连接的存储器;存储器用于存储程序数据,处理器用于执行程序数据,以实现上述第一种方案中提供的方法。Another technical solution adopted in this application is to provide a mobile terminal, which includes a processor and a memory connected to the processor; the memory is used to store program data, and the processor is used to execute the program data, so as to implement the first solution described above. Method provided in.
本申请采用的另一种技术方案是提供一种服务器,服务器包括处理器以及与处理器连接的存储器;存储器用于存储程序数据,处理器用于执行程序数据,以实现上述第二种方案中提供的方法。Another technical solution adopted by this application is to provide a server, the server includes a processor and a memory connected to the processor; the memory is used to store program data, and the processor is used to execute the program data, so as to implement the above-mentioned second solution. Methods.
本申请采用的另一种技术方案是提供一种计算机存储介质,计算机存储介质用于存储程序数据,程序数据在被处理器执行时,用于实现上述方案中提供的任一方法。Another technical solution adopted by the present application is to provide a computer storage medium, which is used to store program data, and when the program data is executed by a processor, it is used to implement any of the methods provided in the above-mentioned solutions.
本申请采用的另一种技术方案是提供一种移动终端,移动终端包括:获取模块,用于获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求;提取模块,用于提取介绍需求信息中的关键词;获取模块还用于基于关键词获取相关联的图像和语音;处理模块,用于对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Another technical solution adopted in this application is to provide a mobile terminal. The mobile terminal includes: an acquisition module for acquiring introduction requirement information about an application program; wherein the introduction requirement information is used to indicate a requirement for introducing an application program; extraction The module is used to extract keywords in the introduction demand information; the acquisition module is also used to obtain related images and voices based on the keywords; the processing module is used to process the images and voices to form an introduction to the application Video.
本申请采用的另一种技术方案是提供一种服务器,服务器包括:获取模块,用于获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于 表示对于介绍应用程序的需求;处理模块,用于基于关键词生成相关联的图像和语音;发送模块,用于向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Another technical solution adopted in this application is to provide a server. The server includes: an acquisition module for acquiring keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application , The introduction demand information is used to express the demand for introducing the application; the processing module is used to generate associated images and voices based on keywords; the sending module is used to send images and voices to the mobile terminal so that the mobile terminal can respond to the image And voice processing to form a video for introducing the application.
本申请的有益效果是:区别于现有技术的情况,本申请的一种应用程序的介绍方法,包括:获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求;提取介绍需求信息中的关键词;基于关键词获取相关联的图像和语音;对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。通过上述方式,能够便捷的获取到用户的需求,并根据用户的不同需求进行不同的应用程序介绍,一方面能够适应于不同的用户群体,使应用程序满足更多用户群体的需求,另一方面采用动画的形式进行应用程序的介绍能够增加应用程序介绍的个性化,增加趣味性,提高用户体验。The beneficial effect of this application is that, different from the situation in the prior art, an application introduction method of this application includes: obtaining introduction requirement information about the application; wherein, the introduction requirement information is used to indicate information about the introduction of the application. Demand; extract the keywords in the introduction demand information; obtain related images and voices based on the keywords; process the images and voices to form a video for introducing the application program. Through the above methods, users’ needs can be easily obtained, and different application programs can be introduced according to the different needs of users. On the one hand, it can adapt to different user groups and make the application meet the needs of more user groups. On the other hand, The introduction of the application in the form of animation can increase the personalization of the application introduction, increase the interest, and improve the user experience.
【附图说明】【Explanation of the drawings】
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。其中:In order to more clearly describe the technical solutions in the embodiments of the present application, the following will briefly introduce the drawings that need to be used in the description of the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application. For those of ordinary skill in the art, other drawings can be obtained from these drawings without creative work. among them:
图1是本申请提供的应用程序的介绍方法第一实施例的流程示意图;FIG. 1 is a schematic flowchart of a first embodiment of an application program introduction method provided by the present application;
图2是本申请提供的应用程序的介绍方法第二实施例的流程示意图;FIG. 2 is a schematic flowchart of a second embodiment of an application program introduction method provided by the present application;
图3是本申请提供的应用程序的介绍方法第三实施例的流程示意图;FIG. 3 is a schematic flowchart of a third embodiment of an application program introduction method provided by the present application;
图4是本申请提供的应用程序的介绍方法第四实施例的流程示意图;FIG. 4 is a schematic flowchart of a fourth embodiment of an application program introduction method provided by the present application;
图5是本申请提供的应用程序的介绍方法第五实施例的流程示意图;FIG. 5 is a schematic flowchart of a fifth embodiment of an application program introduction method provided by the present application;
图6是本申请提供的应用程序的介绍方法第六实施例的流程示意图;FIG. 6 is a schematic flowchart of a sixth embodiment of an application program introduction method provided by the present application;
图7是本申请提供的应用程序的介绍方法第七实施例的流程示意图;FIG. 7 is a schematic flowchart of a seventh embodiment of an application program introduction method provided by the present application;
图8是本申请提供的移动终端第一实施例的结构示意图;FIG. 8 is a schematic structural diagram of a first embodiment of a mobile terminal provided by the present application;
图9是本申请提供的服务器第一实施例的结构示意图;FIG. 9 is a schematic structural diagram of a first embodiment of a server provided by the present application;
图10是本申请提供的计算机存储介质一实施例的结构示意图;FIG. 10 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application;
图11是本申请提供的移动终端第二实施例的结构示意图;FIG. 11 is a schematic structural diagram of a second embodiment of a mobile terminal provided by the present application;
图12是本申请提供的服务器第二实施例的结构示意图。Fig. 12 is a schematic structural diagram of a second embodiment of a server provided by the present application.
【具体实施方式】【detailed description】
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述。可以理解的是,此处所描述的具体实施例仅用于解释本申请,而非对本申请的限定。另外还需要说明的是,为了便于描述,附图中仅示出了与本申请相关的部分而非全部结构。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly and completely described below in conjunction with the drawings in the embodiments of the present application. It can be understood that the specific embodiments described here are only used to explain the application, but not to limit the application. In addition, it should be noted that, for ease of description, the drawings only show a part of the structure related to the present application instead of all of the structure. Based on the embodiments in this application, all other embodiments obtained by a person of ordinary skill in the art without creative work shall fall within the protection scope of this application.
本申请中的术语“第一”、“第二”等是用于区别不同对象,而不是用于描述特定顺序。此外,术语“包括”和“具有”以及它们任何变形,意图在于覆盖不排他的包含。例如包含了一系列步骤或单元的过程、方法、系统、产品或设备没有限定于已列出的步骤或单元,而是可选地还包括没有列出的步骤或单元,或可选地还包括对于这些过程、方法、产品或设备固有的其它步骤或单元。The terms "first", "second", etc. in this application are used to distinguish different objects, rather than to describe a specific sequence. In addition, the terms "including" and "having" and any variations thereof are intended to cover non-exclusive inclusions. For example, a process, method, system, product, or device that includes a series of steps or units is not limited to the listed steps or units, but optionally includes unlisted steps or units, or optionally also includes Other steps or units inherent in these processes, methods, products or equipment.
在本文中提及“实施例”意味着,结合实施例描述的特定特征、结构或特性可以包含在本申请的至少一个实施例中。在说明书中的各个位置出现该短语并不一定均是指相同的实施例,也不是与其它实施例互斥的独立的或备选的实施例。本领域技术人员显式地和隐式地理解的是,本文所描述的实施例可以与其它实施例相结合。Reference to "embodiments" herein means that a specific feature, structure, or characteristic described in conjunction with the embodiments may be included in at least one embodiment of the present application. The appearance of the phrase in various places in the specification does not necessarily refer to the same embodiment, nor is it an independent or alternative embodiment mutually exclusive with other embodiments. Those skilled in the art clearly and implicitly understand that the embodiments described herein can be combined with other embodiments.
参阅图1,图1是本申请提供的应用程序的介绍方法第一实施例的流程示意图,该方法基于移动终端进行实施,该方法包括:Referring to Fig. 1, Fig. 1 is a schematic flowchart of a first embodiment of the application introduction method provided by the present application. The method is implemented based on a mobile terminal, and the method includes:
步骤11:获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求。Step 11: Obtain the introduction requirement information about the application; among them, the introduction requirement information is used to indicate the requirement for the introduction of the application.
可选的,移动终端响应用户下载应用程序并安装完成后,获取到关于应用程序的介绍需求信息。Optionally, after the mobile terminal responds to the user to download the application and the installation is complete, it obtains the introduction requirement information about the application.
可选的,介绍需求信息可以是音频信息,也可以是文本信息。音频信息通过移动终端的麦克风采集,文本信息可以通过手动输入,或者选择应用程序提示的关键词作为文本信息。Optionally, the introduction requirement information may be audio information or text information. The audio information is collected by the microphone of the mobile terminal, and the text information can be manually input, or keywords prompted by the application can be selected as the text information.
可选的,介绍需求信息用于表示对于介绍应用程序的需求。例如:用户需要了解理财类应用程序的收益时,即可将“收益”作为介绍需求信息。Optionally, the introduction requirement information is used to indicate the requirement for the introduction application. For example, when users need to know the income of financial management applications, they can use “income” as the introduction demand information.
步骤12:提取介绍需求信息中的关键词。Step 12: Extract the keywords in the introduction requirement information.
可选的,当获取到介绍需求信息后,移动终端对介绍需求信息的内 容进行关键词提取,如:获取到的介绍需求信息为音频信息,音频信息解析出的文本信息为“此应用如何安全支付”,则提取到的关键词为“安全支付”。Optionally, when the introduction requirement information is obtained, the mobile terminal performs keyword extraction on the content of the introduction requirement information. For example, the obtained introduction requirement information is audio information, and the text information parsed from the audio information is "How is this application safe?" "Payment", the extracted keyword is "secure payment".
关键词提取方法可以基于统计特征的关键词提取算法。The keyword extraction method can be based on a keyword extraction algorithm based on statistical features.
基于统计特征的关键词提取算法是利用文档中词语的统计信息抽取文档的关键词。通常将文本经过预处理得到候选词语的集合,然后采用特征值量化的方式从候选集合中得到关键词。The keyword extraction algorithm based on statistical features uses the statistical information of the words in the document to extract the keywords of the document. Usually, the text is preprocessed to obtain a set of candidate words, and then keywords are obtained from the candidate set by means of feature value quantification.
其中,特征值量化的方式有基于词权重的特征量化、基于词的文档位置的特征量化、基于词的关联信息的特征量化。基于词权重的特征量化主要包括词性、词频、逆向文档频率、相对词频、词长等;基于词的文档位置的特征量化是根据文章不同位置的句子对文档的重要性不同的假设来进行的。通常,文章的前N个词、后N个词、段首、段尾、标题、引言等位置的词具有代表性,这些词作为关键词可以表达整个的主题;基于词的关联信息的特征量化:词的关联信息是指词与词、词与文档的关联程度信息,包括互信息、hits值、贡献度、依存度、TF-IDF值等。Among them, the feature value quantization methods include feature quantification based on word weight, feature quantification based on word document location, and feature quantification based on word-related information. Feature quantification based on word weight mainly includes part-of-speech, word frequency, inverse document frequency, relative word frequency, word length, etc.; word-based feature quantification of document position is based on the assumption that sentences at different positions of the article have different importance to the document. Generally, words in the first N words, last N words, beginning of paragraph, end of paragraph, title, introduction, etc. of the article are representative. These words can express the entire topic as keywords; quantify the characteristics based on the related information of the words : Word related information refers to the degree of relevance information between words and words, words and documents, including mutual information, hits value, contribution degree, dependency degree, TF-IDF value, etc.
关键词的提取方法还可以基于深度学习的方法进行提取。Keyword extraction methods can also be based on deep learning methods.
可以理解,提取关键词的方法有多种,这里不一一列举。It is understandable that there are many ways to extract keywords, which are not listed here.
步骤13:基于关键词获取相关联的图像和语音。Step 13: Obtain related images and voices based on keywords.
可选的,获取相关联的图像可以是移动终端将关键词发送至服务器,由服务器在预设图像库中进行图像检索,以得到多个图像。Optionally, acquiring the associated image may be that the mobile terminal sends the keyword to the server, and the server performs image retrieval in a preset image library to obtain multiple images.
可选的,获取相关联的语音可以是移动终端将关键词发送至服务器,由服务器在通过关键词及应用场景生成符合应用场景的多段文字,在将多段文字发送给移动终端,再由移动终端将文字信息转换成语音信息。Optionally, obtaining the associated voice may be that the mobile terminal sends the keywords to the server, and the server generates multiple texts conforming to the application scenarios through the keywords and application scenarios, and then sends the multiple texts to the mobile terminal, and then the mobile terminal sends the multiple texts to the mobile terminal. Convert text messages into voice messages.
可选的,获取相关联的图像可以是移动终端在本地预设图像库中进行图像检索,以得到多个图像。Optionally, acquiring the associated image may be that the mobile terminal performs image retrieval in a local preset image library to obtain multiple images.
可选的,获取相关联的语音可以是移动终端通过关键词及应用场景生成符合应用场景的多段文字,然后将文字信息转换成语音信息。Optionally, acquiring the associated voice may be that the mobile terminal generates multiple paragraphs of text that conform to the application scenario through keywords and application scenarios, and then converts the text information into voice information.
步骤14:对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Step 14: Process the image and voice to form a video for introducing the application.
可选的,对图像进行图像分割,提取符合关键词的特征信息,将特征信息组成新的图像。如:关键词是“鸟”、“树”;某图像存在一棵树的特征信息,另一张图像存在一只鸟的特征信息,将这两个信息提取出来,可以根据场景,组成鸟停在树上的图像。在组成一系列完整图像 后,对图像进行平滑等增强处理,目的是使图像内容更自然。Optionally, perform image segmentation on the image, extract feature information that matches the keyword, and compose the feature information into a new image. For example, the keywords are "birds" and "trees"; one image has the characteristic information of a tree, and another image has the characteristic information of a bird. These two pieces of information can be extracted to form a bird stop according to the scene. Image on the tree. After composing a series of complete images, the image is smoothed and other enhanced processing, the purpose is to make the image content more natural.
可选的,将语音信息与图像信息进行融合,形成用于对应用程序进行介绍的视频。Optionally, the voice information and the image information are merged to form a video for introducing the application program.
举例说明:for example:
用户下载了一个理财类应用程序,在启动应用程序时,应用程序提醒用户请说出想要了解的内容,此时移动终端采集的音频信息为“我第一次开户,如何投资才能收益高且风险小,还有就是如何支付”,移动终端提取的关键词是“第一次开户”、“投资”、“收益高”、“风险小”、“如何支付”。那么根据这些关键词,在预设图像库中搜索对应的图像,如“第一次开户”就搜索到开户画面及动画人物的图像,“收益高”、“风险小”就搜索警示和推荐相关产品的图像,然后组成一段有个动画人物介绍怎么在保证风险低的情况下最好投资方法,然后还要一个画面是支付安全也很重要的画面。同时根据这些关键词和场景,生成符合场景的文字信息,将文字信息转换为语音,将语音信息和图像信息融合,形成了对用户需求介绍的视频。The user downloads a financial management application. When starting the application, the application reminds the user to say what he wants to know. The audio information collected by the mobile terminal at this time is "I open an account for the first time, how can I invest in order to achieve high returns and The risk is small, and there is also how to pay. The keywords extracted by the mobile terminal are "open account for the first time", "investment", "high return", "small risk", and "how to pay". Then, based on these keywords, search for the corresponding images in the preset image library. For example, “open an account for the first time” will search for the account opening screen and images of animated characters, and “high return” and “low risk” will search for warnings and recommendations. The image of the product is then formed into a section with an animated character explaining how to invest in the best way to ensure low risk, and then there is another picture where payment safety is also very important. At the same time, based on these keywords and scenes, text information that meets the scene is generated, the text information is converted into voice, and the voice information and image information are merged to form a video that introduces user needs.
在一些实施例中,还可以给视频加入背景音乐。In some embodiments, background music can also be added to the video.
在其他实施例中,移动终端可以根据一段语音,生成声音和图像对应的应用程序介绍,或者通过语音给小朋友讲故事,小朋友可以描述自己喜欢听的故事类型,通过机器学习,生成图文并茂的小故事,让小朋友更感兴趣,也可以使得一些不太识字的小朋友可以通过动画获取相应的知识。In other embodiments, the mobile terminal can generate an application introduction corresponding to the sound and image according to a piece of voice, or tell a story to the child through voice. The child can describe the type of story he likes to listen to, and generate a short story with pictures and texts through machine learning. , So that children are more interested, and it can also make some children who are not literate can acquire corresponding knowledge through animation.
在其他实施例中,根据不同的应用程序及不同的用户需求,移动终端生成对应需求的应用程序的图文并茂的视频,以供用户观看。In other embodiments, according to different application programs and different user requirements, the mobile terminal generates a video with pictures and texts corresponding to the application program for the user to watch.
区别于现有技术的情况,本申请的一种应用程序的介绍方法,包括:获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求;提取介绍需求信息中的关键词;基于关键词获取相关联的图像和语音;对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。通过上述方式,能够便捷的获取到用户的需求,并根据用户的不同需求进行不同的应用程序介绍,一方面能够适应于不同的用户群体,使应用程序满足更多用户群体的需求,另一方面采用动画的形式进行应用程序的介绍能够增加应用程序介绍的个性化,增加趣味性,提高用户体验。Different from the situation in the prior art, an application introduction method of this application includes: obtaining introduction requirement information about the application; wherein the introduction requirement information is used to indicate the requirement for introducing the application; and the introduction requirement information is extracted Keywords; obtain related images and voices based on keywords; process images and voices to form a video for introducing the application. Through the above methods, users’ needs can be easily obtained, and different application programs can be introduced according to the different needs of users. On the one hand, it can adapt to different user groups and make the application meet the needs of more user groups. On the other hand, The introduction of the application in the form of animation can increase the personalization of the application introduction, increase the interest, and improve the user experience.
参阅图2,图2是本申请提供的应用程序的介绍方法第二实施例的流程示意图,该方法基于移动终端进行实施,该方法包括:Referring to Figure 2, Figure 2 is a schematic flowchart of a second embodiment of the application introduction method provided by the present application. The method is implemented based on a mobile terminal, and the method includes:
步骤21:获取关于应用程序的音频信息;其中,音频信息用于表示 对于介绍应用程序的需求。Step 21: Acquire audio information about the application program; among them, the audio information is used to indicate the demand for introducing the application program.
在本实施例中,收集用户的音频信息来表示对于介绍应用程序的需求。In this embodiment, the user's audio information is collected to indicate the demand for introducing the application program.
可选的,音频信息可以是用户想要了解应用程序是说出的与应用程序相关的音频信息。Optionally, the audio information may be audio information related to the application that the user wants to learn about the application.
可选的,音频信息可以是在应用程序启动后显示给用户的文字信息,以提示用户对应用程序进行哪方面的了解,方便用户快速的说出相应的关键词信息。Optionally, the audio information may be text information displayed to the user after the application is started to prompt the user to learn about the application, so that the user can quickly speak the corresponding keyword information.
步骤22:对音频信息进行语音识别,以得到文本信息。Step 22: Perform voice recognition on the audio information to obtain text information.
可选的,语音识别是将一段音频信息转换成相对应的文本信息,系统主要包含特征提取、声学模型,语言模型以及字典与解码四大部分,为了更有效地提取特征还需要对所采集到的音频信息进行滤波、分帧等音频数据预处理工作,将需要分析的音频信息从原始信号中合适地提取出来;特征提取工作将音频信息从时域转换到频域,为声学模型提供合适的特征向量;声学模型中再根据声学特性计算每一个特征向量在声学特征上的得分;而语言模型则根据语言学相关的理论,计算该声音信号对应可能词组序列的概率;最后根据已有的字典,对词组序列进行解码,得到最后可能的文本表示。Optionally, speech recognition is to convert a piece of audio information into corresponding text information. The system mainly includes four parts: feature extraction, acoustic model, language model, dictionary and decoding. In order to extract features more effectively, the collected Perform audio data preprocessing work such as filtering and framing of the audio information to properly extract the audio information that needs to be analyzed from the original signal; feature extraction converts the audio information from the time domain to the frequency domain to provide a suitable acoustic model Feature vector: The acoustic model calculates the score of each feature vector on the acoustic feature according to the acoustic characteristics; while the language model calculates the probability of the sound signal corresponding to the possible sequence of phrases according to the linguistic theory; finally according to the existing dictionary , Decode the phrase sequence to get the final possible text representation.
步骤23:将文本信息输入至卷积神经网络进行深度学习,以将文本信息进行语义分割,以得到关键词。Step 23: Input the text information into the convolutional neural network for deep learning, so as to perform semantic segmentation on the text information to obtain keywords.
可选的,根据应用程序的特点,预先把大量的信息通过卷积神经网络进行深度学习进行训练,以生成对应的语义分割模型。当语义分割模型在得到文本信息时,即可得到关键词。Optionally, according to the characteristics of the application program, a large amount of information is trained in advance through a convolutional neural network for deep learning to generate a corresponding semantic segmentation model. When the semantic segmentation model gets the text information, it can get the keywords.
步骤24:基于关键词获取相关联的图像和语音。Step 24: Obtain related images and voices based on the keywords.
步骤25:对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Step 25: Process the image and voice to form a video for introducing the application.
步骤24-25与上述实施例具有相同或相似的技术方案,这里不做赘述。Steps 24-25 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
参阅图3,图3是本申请提供的应用程序的介绍方法第三实施例的流程示意图,该方法基于移动终端进行实施,该方法包括:Referring to FIG. 3, FIG. 3 is a schematic flowchart of a third embodiment of the application introduction method provided by the present application. The method is implemented based on a mobile terminal, and the method includes:
步骤31:获取关于应用程序的文本信息;其中,文本信息用于表示对于介绍应用程序的需求。Step 31: Acquire text information about the application program; where the text information is used to indicate a requirement for introducing the application program.
可选的,文本信息可以是用户手动输入,也可以是应用程序提示多段文字,由用户进行选择而生成的。Optionally, the text information may be manually input by the user, or may be generated by the user's selection by the application prompting multiple paragraphs of text.
步骤32:对文本信息进行语义分割。Step 32: Perform semantic segmentation on the text information.
步骤33:基于语义分割的结果得到关键词。Step 33: Obtain keywords based on the result of semantic segmentation.
步骤32-33可以具体是:Steps 32-33 can be specifically:
采用TF-IDF(term frequency inverse document frequency,信息检索数据挖掘的常用加权技术)、TextRank(自然语言处理的通用基于图的排序算法)、Rake(Rapid Automatic Keyword Extraction,快速自动关键字提取)、Topic-Model(主题模型)等方法,可以得到关键词。Adopt TF-IDF (term frequency inverse document frequency, a common weighting technique for information retrieval data mining), TextRank (a general graph-based sorting algorithm for natural language processing), Rake (Rapid Automatic Keyword Extraction, fast automatic keyword extraction), Topic -Model (theme model) and other methods, you can get keywords.
TF-IDF:TF衡量了一个词在文本信息中出现的频率,一个文本信息中多次出现的词总是有一定的特殊意义,但是并不是所有多次出现的词就都是有意义的,如果一个词在所有的文档中都多次出现,那么这个词就没有什么价值了。TF-IDF就很好地衡量了这些因素:TF=(词在文本信息中出现的次数)/(文章总词数),IDF=log(语料库中文本信息综述/(包含该词的文本信息数+1));TF-IDF: TF measures the frequency of a word in text information. Words that appear multiple times in a text message always have a certain special meaning, but not all words that appear multiple times are meaningful. If a word appears multiple times in all documents, then the word has no value. TF-IDF measures these factors well: TF=(the number of times the word appears in the text information)/(the total number of words in the article), IDF=log(summary of the text information in the corpus/(the number of text information containing the word) +1));
TF-IDF=TF*IDF;TF-IDF=TF*IDF;
TF-IDF值越大,则这个词成为一个关键词的概率就越大。The greater the TF-IDF value, the greater the probability of the word becoming a keyword.
Rake算法的流程为分词,如以标点符号及停用词作为分词标准;然后构建共现矩阵;特征提取。包含词频freq、度deg以及度与频率之比deg/freq三个特征;定义score。score=deg/freq;降序输出。按score大小降序输出1/3文档词汇量的关键词。The process of the Rake algorithm is word segmentation, such as punctuation and stop words as the word segmentation standard; then construct a co-occurrence matrix; feature extraction. Contains three features: word frequency freq, degree deg, and the ratio of degree to frequency deg/freq; defines score. score=deg/freq; output in descending order. Output keywords of 1/3 document vocabulary in descending order of score size.
其中,提取特征后有个特殊处理,对于相邻的关键词,如果满足同一文档和相同顺序中至少两次相邻,则进行合并,成为新的候选关键词后,score定义为合并前的候选关键词score之和。这样操作的原因是,这些相邻候选关键词相对较少,简单对score相加,增加了它们的重要性。Among them, there is a special process after extracting features. For adjacent keywords, if at least two adjacent keywords in the same document and in the same order are met, they will be merged and become a new candidate keyword. Score is defined as the candidate before merging. The sum of the keyword score. The reason for this operation is that these adjacent candidate keywords are relatively few, and simply adding the score increases their importance.
在其他实施例中,还可以通过神经网络的深度学习,预先建立语义分割模型,以实现快速提取关键词。In other embodiments, a semantic segmentation model may be established in advance through deep learning of a neural network to achieve rapid extraction of keywords.
步骤34:基于关键词获取相关联的图像和语音。Step 34: Obtain related images and voices based on the keywords.
步骤35:对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Step 35: Process the image and voice to form a video for introducing the application program.
步骤34-35与上述实施例具有相同或相似的技术方案,这里不做赘述。Steps 34-35 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
参阅图4,图4是本申请提供的应用程序的介绍方法第四实施例的流程示意图,该方法基于移动终端进行实施,该方法包括:Referring to FIG. 4, FIG. 4 is a schematic flowchart of a fourth embodiment of the application introduction method provided by the present application. The method is implemented based on a mobile terminal, and the method includes:
步骤41:获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求。Step 41: Obtain the introduction requirement information about the application program; wherein the introduction requirement information is used to indicate the requirement for the introduction application program.
步骤42:提取介绍需求信息中的关键词。Step 42: Extract keywords in the introduction requirement information.
步骤41-42与上述实施例具有相同或相似的技术方案,这里不做赘述。Steps 41-42 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
步骤43:将关键词发送给服务器,以使服务器基于关键词生成相关联的图像和语音。Step 43: Send the keywords to the server, so that the server generates associated images and voices based on the keywords.
可选的,当服务器接收到关键词后,基于卷积神经网络的深度学习,得到与关键词相关联的图像和语音。Optionally, after the server receives the keywords, deep learning based on the convolutional neural network can obtain images and voices associated with the keywords.
在其他实施例中,图像可以由服务器得到,而语音可以由移动终端自行对关键词进行识别,以生成多段符合场景的多段文字,并转换为语音。In other embodiments, the image may be obtained by the server, and the voice may be recognized by the mobile terminal on the keywords, so as to generate multiple paragraphs of text that match the scene, and convert them into voice.
步骤44:获取服务器发送的图像和语音。Step 44: Obtain the image and voice sent by the server.
步骤45:对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Step 45: Process the image and voice to form a video for introducing the application program.
步骤44-45与上述实施例具有相同或相似的技术方案,这里不做赘述。Steps 44-45 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
参阅图5,图5是本申请提供的应用程序的介绍方法第五实施例的流程示意图,该方法包括:Referring to FIG. 5, FIG. 5 is a schematic flowchart of a fifth embodiment of the application introduction method provided by the present application, and the method includes:
步骤51:获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求。Step 51: Obtain introduction requirement information about the application; where the introduction requirement information is used to indicate the requirement for the introduction of the application.
步骤52:提取介绍需求信息中的关键词。Step 52: Extract keywords in the introduction requirement information.
步骤53:将关键词发送给服务器,以使服务器基于关键词生成相关联的图像和语音。Step 53: Send the keywords to the server, so that the server generates associated images and voices based on the keywords.
步骤54:获取服务器发送的图像和语音。Step 54: Obtain the image and voice sent by the server.
步骤51-54与上述实施例具有相同或相似的技术方案,这里不做赘述。Steps 51-54 have the same or similar technical solutions as the above-mentioned embodiment, and will not be repeated here.
步骤55:对多个对应的图像进行图像分割,提取图像中特征信息。Step 55: Perform image segmentation on multiple corresponding images, and extract feature information from the images.
图像分割就是把图像分成若干个特定的、具有独特性质的区域并提出感兴趣目标的技术和过程。它是由图像处理到图像分析的关键步骤。现有的图像分割方法主要分以下几类:基于阈值的分割方法、基于区域的分割方法、基于边缘的分割方法以及基于特定理论的分割方法等。从数学角度来看,图像分割是将数字图像划分成互不相交的区域的过程。图像分割的过程也是一个标记过程,即把属于同一区域的像素赋予相同的编号。Image segmentation is the technique and process of dividing an image into a number of specific areas with unique properties and proposing objects of interest. It is a key step from image processing to image analysis. The existing image segmentation methods are mainly divided into the following categories: threshold-based segmentation methods, region-based segmentation methods, edge-based segmentation methods, and segmentation methods based on specific theories. From a mathematical point of view, image segmentation is the process of dividing a digital image into disjoint areas. The process of image segmentation is also a marking process, that is, the pixels belonging to the same area are assigned the same number.
其中,基于阈值的分割方法是一种基于区域的图像分割技术,原理是把图像象素点分为若干类。图像阈值化分割是一种传统的最常用的图像分割方法,因其实现简单、计算量小、性能较稳定而成为图像分割中 最基本和应用最广泛的分割技术。它特别适用于目标和背景占据不同灰度级范围的图像。它不仅可以极大的压缩数据量,而且也大大简化了分析和处理步骤,因此在很多情况下,是进行图像分析、特征提取与模式识别之前的必要的图像预处理过程。图像阈值化的目的是要按照灰度级,对像素集合进行一个划分,得到的每个子集形成一个与现实景物相对应的区域,各个区域内部具有一致的属性,而相邻区域不具有这种一致属性。这样的划分可以通过从灰度级出发选取一个或多个阈值来实现。Among them, the threshold-based segmentation method is a region-based image segmentation technology, the principle is to divide the image pixels into several categories. Image thresholding segmentation is one of the most commonly used traditional image segmentation methods. It has become the most basic and most widely used segmentation technique in image segmentation due to its simple implementation, small calculation amount, and stable performance. It is especially suitable for images where the target and background occupy different gray scale ranges. It can not only greatly compress the amount of data, but also greatly simplifies the analysis and processing steps. Therefore, in many cases, it is a necessary image preprocessing process before image analysis, feature extraction and pattern recognition. The purpose of image thresholding is to divide the pixel set according to the gray level, and each of the obtained subsets forms an area corresponding to the real scene. Each area has the same attributes, while the adjacent areas do not have this Consistent attributes. Such division can be achieved by selecting one or more thresholds starting from the gray level.
其中,基于区域的分割方法是以直接寻找区域为基础的分割技术,具体算法有区域生长和区域分离与合并算法。基于区域提取方法有两种基本形式:一种是区域生长,从单个像素出发,逐步合并以形成所需要的分割区域;另一种是从全局出发,逐步切割至所需的分割区域。Among them, the region-based segmentation method is a segmentation technique based on directly finding the region. The specific algorithms include region growth and region separation and merging algorithms. There are two basic forms of region-based extraction methods: one is region growth, which starts from a single pixel and gradually merges to form the required segmentation area; the other is to start from the overall situation and gradually cut to the required segmentation area.
其中,基于边缘的分割则主要有基于点的检测、基于线的检测以及基于边缘检测等几种方法。Among them, edge-based segmentation mainly includes point-based detection, line-based detection, and edge-based detection.
其中,基于特定理论的分割方法可以分为聚类分析、模糊集理论、基因编码、小波变换等方法。Among them, segmentation methods based on specific theories can be divided into cluster analysis, fuzzy set theory, gene coding, wavelet transform and other methods.
可选的,在图像分割后,基于关键词及场景,进行特征提取,以执行步骤56。Optionally, after the image is segmented, feature extraction is performed based on the keywords and the scene, to perform step 56.
步骤56:将特征信息进行组合,以生成多个图像帧。Step 56: Combine the feature information to generate multiple image frames.
步骤57:将多个图像帧形成动画。Step 57: Form multiple image frames into animation.
可选的,步骤55-57具体是:Optionally, steps 55-57 are specifically:
通过卷积神经网络预先进行深度学习,建立图像模型,以使对应的特征信息生成多个图像帧,然后将多个图像帧形成动画。Deep learning is performed in advance through the convolutional neural network to establish an image model so that the corresponding feature information generates multiple image frames, and then the multiple image frames are formed into an animation.
步骤58:将动画与语音进行融合,以形成用于对应用程序进行介绍的视频。Step 58: The animation and voice are merged to form a video for introducing the application.
参阅图6,图6是本申请提供的应用程序的介绍方法第六实施例的流程示意图,该方法基于服务器进行实施,该方法包括:Referring to FIG. 6, FIG. 6 is a schematic flowchart of a sixth embodiment of the application introduction method provided by the present application. The method is implemented based on a server, and the method includes:
步骤61:获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于表示对于介绍应用程序的需求。Step 61: Acquire keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction requirement information about the application, and the introduction requirement information is used to indicate the requirement for introducing the application.
可选的,当移动终端获取到关于应用程序的介绍需求信息后,提取出关键词,将关键词发送给服务器。Optionally, after the mobile terminal obtains the introduction requirement information about the application, it extracts the keywords and sends the keywords to the server.
步骤62:基于关键词生成相关联的图像和语音。Step 62: Generate associated images and voices based on the keywords.
可选的,服务器预先通过应用程序的相关内容进行模型训练,以使在获得移动终端的关键词时,快速进行响应,得到与关键词相关联的图 像和语音。Optionally, the server performs model training through the relevant content of the application in advance, so that when the keyword of the mobile terminal is obtained, it responds quickly and obtains the image and voice associated with the keyword.
步骤63:向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Step 63: Send the image and voice to the mobile terminal, so that the mobile terminal processes the image and voice to form a video for introducing the application program.
可选的,向移动终端发送生成的图像和语音,以使移动终端对图像进行特征信息提取,再将特征信息进行组合,以生成多个图像帧进行组合,以生成多个图像帧。移动终端将多个图像帧形成动画与语音融合,以形成用于对应用程序进行介绍的视频。Optionally, the generated image and voice are sent to the mobile terminal, so that the mobile terminal extracts feature information of the image, and then combines the feature information to generate multiple image frames and combine to generate multiple image frames. The mobile terminal merges multiple image frames into animation and voice to form a video for introducing the application.
区别于现有技术的情况,本申请的一种应用程序的介绍方法,包括:获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于表示对于介绍应用程序的需求;基于关键词生成相关联的图像和语音;向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。通过上述方式,能够便捷的获取到用户的需求,并根据用户的不同需求进行不同的应用程序介绍,一方面能够适应于不同的用户群体,使应用程序满足更多用户群体的需求,另一方面采用动画的形式进行应用程序的介绍能够增加应用程序介绍的个性化,增加趣味性,提高用户体验Different from the situation in the prior art, an application introduction method of this application includes: acquiring keywords sent by a mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction requirement information about the application. The introduction demand information is used to express the demand for introducing the application; generate related images and voice based on keywords; send the image and voice to the mobile terminal, so that the mobile terminal can process the image and voice to form the application program Introductory video. Through the above methods, the needs of users can be easily obtained, and different application programs can be introduced according to the different needs of users. On the one hand, it can adapt to different user groups and make the application meet the needs of more user groups. The introduction of the application in the form of animation can increase the personalization of the application introduction, increase the interest, and improve the user experience
参阅图7,图7是本申请提供的应用程序的介绍方法第七实施例的流程示意图,该方法包括:Referring to FIG. 7, FIG. 7 is a schematic flowchart of a seventh embodiment of the application introduction method provided by the present application, and the method includes:
步骤71:获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于表示对于介绍应用程序的需求。Step 71: Acquire keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction requirement information about the application, and the introduction requirement information is used to indicate the requirement for introducing the application.
步骤72:将关键词通过深度学习,以从预设图像库得到相关联的图像。Step 72: Pass the keywords through deep learning to obtain associated images from the preset image library.
深度学习模型有卷积神经网络(convolutional neural network)、DBN(Deep Belief Network,深度信任网络模型)和堆栈自编码网络(stacked auto-encoder network)模型。Deep learning models include convolutional neural network (convolutional neural network), DBN (Deep Belief Network, deep trust network model) and stacked auto-encoder network models.
卷积神经网络受视觉系统的结构启发而产生。第一个卷积神经网络计算模型是在神经认知机中提出的,基于神经元之间的局部连接和分层组织图像转换,将有相同参数的神经元应用于前一层神经网络的不同位置,得到一种平移不变神经网络结构形式。后来,在该思想的基础上,用误差梯度设计并训练卷积神经网络,在一些模式识别任务上得到优越的性能。Convolutional neural networks are inspired by the structure of the visual system. The first convolutional neural network calculation model was proposed in the neurocognitive machine. Based on the local connection between neurons and the layered organization image conversion, the neurons with the same parameters are applied to the difference of the previous layer of neural network. Position, get a translation-invariant neural network structure. Later, on the basis of this idea, a convolutional neural network was designed and trained with error gradients to obtain superior performance on some pattern recognition tasks.
DBN可以解释为贝叶斯概率生成模型,由多层随机隐变量组成,上面的两层具有无向对称连接,下面的层得到来自上一层的自顶向下的有 向连接,最底层单元的状态为可见输入数据向量。DBN由若2F结构单元堆栈组成,结构单元通常为RBM(RestIlcted Boltzmann Machine,受限玻尔兹曼机)。堆栈中每个RBM单元的可视层神经元数量等于前一RBM单元的隐层神经元数量。根据深度学习机制,采用输入样例训练第一层RBM单元,并利用其输出训练第二层RBM模型,将RBM模型进行堆栈通过增加层来改善模型性能。在无监督预训练过程中,DBN编码输入到顶层RBM后,解码顶层的状态到最底层的单元,实现输入的重构。RBM作为DBN的结构单元,与每一层DBN共享参数。DBN can be interpreted as a Bayesian probability generation model, which is composed of multiple layers of random latent variables. The upper two layers have undirected symmetrical connections, and the lower layer gets top-down directed connections from the upper layer, and the lowest layer unit The state of is the visible input data vector. The DBN is composed of a stack of 2F structural units, and the structural unit is usually RBM (RestIlcted Boltzmann Machine, Restricted Boltzmann Machine). The number of neurons in the visible layer of each RBM unit in the stack is equal to the number of neurons in the hidden layer of the previous RBM unit. According to the deep learning mechanism, the input samples are used to train the first-layer RBM units, and their output is used to train the second-layer RBM models, and the RBM models are stacked to improve the model performance by adding layers. In the unsupervised pre-training process, after the DBN code is input to the top RBM, the state of the top layer is decoded to the unit of the bottom layer to realize the reconstruction of the input. As the structural unit of DBN, RBM shares parameters with each layer of DBN.
堆栈自编码网络的结构与DBN类似,由若干结构单元堆栈组成,不同之处在于其结构单元为自编码模型(auto-en-coder)而不是RBM。自编码模型是一个两层的神经网络,第一层称为编码层,第二层称为解码层。The structure of the stacked self-encoding network is similar to that of the DBN, consisting of a stack of several structural units. The difference is that the structural unit is an auto-en-coder instead of RBM. The self-encoding model is a two-layer neural network, the first layer is called the coding layer, and the second layer is called the decoding layer.
可选的,服务器需要根据应用程序的特点及关键词的特点,产生对应的场景预判,根据场景搜索对应的图像。Optionally, the server needs to generate a corresponding scene prediction according to the characteristics of the application and the characteristics of the keywords, and search for the corresponding image according to the scene.
可选的,当预设图像库不满足搜索需求时,服务器将在互联中搜索图像。Optionally, when the preset image library does not meet the search requirements, the server will search for images in the Internet.
步骤73:将关键词通过深度学习,以生成符合关键词场景的文字信息。Step 73: Pass the keywords through deep learning to generate text information that meets the keyword scene.
步骤74:将文字信息转换为语音。Step 74: Convert the text information into voice.
步骤75:向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Step 75: Send the image and voice to the mobile terminal, so that the mobile terminal processes the image and voice to form a video for introducing the application program.
可选的,还可以是服务器先检索大量图像,将图像发送给移动终端,由移动终端根据关键词进行图像分割,再根据场景进行组合,形成动画。在于语音融合,以形成用于对应用程序进行介绍的视频。Optionally, the server first retrieves a large number of images, sends the images to the mobile terminal, and the mobile terminal divides the images according to keywords, and then combines them according to the scene to form an animation. It is voice fusion to form a video for introducing the application.
参阅图8,图8是本申请提供的移动终端第一实施例的结构示意图,该移动终端80包括处理器81以及与处理器81连接的存储器82;存储器82用于存储程序数据,处理器81用于执行程序数据,以实现以下方法:Referring to FIG. 8, FIG. 8 is a schematic structural diagram of a first embodiment of a mobile terminal provided by the present application. The mobile terminal 80 includes a processor 81 and a memory 82 connected to the processor 81; the memory 82 is used to store program data, and the processor 81 Used to execute program data to implement the following methods:
获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求;提取介绍需求信息中的关键词;基于关键词获取相关联的图像和语音;对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Obtain the introduction demand information about the application; among them, the introduction demand information is used to express the demand for introducing the application; extract the keywords in the introduction demand information; obtain the associated images and voices based on the keywords; process the images and voices To form a video for introducing the application.
可选地,处理器81用于执行该程序数据还用以实现以下的方法:对音频信息进行语音识别,以得到文本信息;对文本信息进行关键词提取,以得到关键词。Optionally, the processor 81 for executing the program data is also used to implement the following methods: perform voice recognition on audio information to obtain text information; and perform keyword extraction on text information to obtain keywords.
可选地,处理器81用于执行该程序数据还用以实现以下的方法:对文本信息进行语义分割;基于语义分割的结果得到关键词。Optionally, the processor 81 is used to execute the program data to implement the following method: perform semantic segmentation on the text information; obtain keywords based on the result of the semantic segmentation.
可选地,处理器81用于执行该程序数据还用以实现以下的方法:将文本信息输入至卷积神经网络进行深度学习,以将文本信息进行语义分割,以得到关键词。Optionally, the processor 81 used to execute the program data is also used to implement the following method: input text information into a convolutional neural network for deep learning, so as to perform semantic segmentation on the text information to obtain keywords.
可选地,处理器81用于执行该程序数据还用以实现以下的方法:对文本信息进行语义分割;基于语义分割的结果得到关键词。Optionally, the processor 81 is used to execute the program data to implement the following method: perform semantic segmentation on the text information; obtain keywords based on the result of the semantic segmentation.
可选地,处理器81用于执行该程序数据还用以实现以下的方法:将关键词发送给服务器,以使服务器基于关键词生成相关联的图像和语音;获取服务器发送的图像和语音。Optionally, the processor 81 is used to execute the program data to implement the following method: sending keywords to the server so that the server generates associated images and voices based on the keywords; acquiring images and voices sent by the server.
可选地,处理器81用于执行该程序数据还用以实现以下的方法:对多个对应的图像进行图像分割,提取图像中特征信息;将特征信息进行组合,以生成多个图像帧;将多个图像帧形成动画;将动画与语音进行融合,以形成用于对应用程序进行介绍的视频。Optionally, the processor 81 is configured to execute the program data to implement the following method: image segmentation of multiple corresponding images, extraction of feature information in the image; combination of feature information to generate multiple image frames; Multiple image frames are formed into animation; animation and voice are merged to form a video for introducing the application.
可选地,处理器81用于执行该程序数据还用以实现以下的方法:获取服务器发送的背景音乐;其中,背景音乐是服务器基于关键词生成的音乐;将背景音乐添加至视频。Optionally, the processor 81 is used to execute the program data to implement the following method: acquiring background music sent by the server; wherein the background music is music generated by the server based on keywords; adding the background music to the video.
参阅图9,图9是本申请提供的服务器第一实施例的结构示意图,该服务器90包括处理器91以及与处理器91连接的存储器92;存储器92用于存储程序数据,处理器91用于执行程序数据,以实现以下方法:Referring to FIG. 9, FIG. 9 is a schematic structural diagram of a first embodiment of a server provided by the present application. The server 90 includes a processor 91 and a memory 92 connected to the processor 91; the memory 92 is used to store program data, and the processor 91 is used to store program data. The program data is executed to achieve the following methods:
获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于表示对于介绍应用程序的需求;基于关键词生成相关联的图像和语音;向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Acquire keywords sent by the mobile terminal; among them, the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application, and the introduction demand information is used to indicate the demand for introducing the application; the associated image is generated based on the keywords And voice; send images and voices to the mobile terminal so that the mobile terminal can process the images and voices to form a video for introducing the application.
可选地,处理器91用于执行该程序数据还用以实现以下的方法:将关键词通过深度学习,以从预设图像库得到相关联的图像。Optionally, the processor 91 used to execute the program data is also used to implement the following method: pass keywords through deep learning to obtain associated images from a preset image library.
可选地,处理器91用于执行该程序数据还用以实现以下的方法:将关键词通过深度学习,以生成符合关键词场景的文字信息;将文字信息转换为语音Optionally, the processor 91 used to execute the program data is also used to implement the following method: pass keywords through deep learning to generate text information that meets the keyword scene; convert the text information into voice
参阅图10,图10是本申请提供的计算机存储介质一实施例的结构示意图,该计算机存储介质100用于存储程序数据101,程序数据101在被处理器执行时,用于实现以下方法:Referring to FIG. 10, FIG. 10 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application. The computer storage medium 100 is used to store program data 101. When the program data 101 is executed by a processor, it is used to implement the following methods:
获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求;提取介绍需求信息中的关键词;基于关键词 获取相关联的图像和语音;对图像和语音进行处理,以形成用于对应用程序进行介绍的视频;Obtain the introduction demand information about the application; among them, the introduction demand information is used to express the demand for introducing the application; extract the keywords in the introduction demand information; obtain the associated images and voices based on the keywords; process the images and voices , To form a video for introducing the application;
或者,获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于表示对于介绍应用程序的需求;基于关键词生成相关联的图像和语音;向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。Or, acquire keywords sent by the mobile terminal; wherein, the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application, and the introduction demand information is used to indicate the demand for introducing the application; and the correlation is generated based on the keywords The image and voice of the mobile terminal; send the image and voice to the mobile terminal so that the mobile terminal can process the image and voice to form a video for introducing the application.
可以理解,计算机存储介质既可以应用于上述的移动终端,也可以应用于上述的服务器,实现上述任一实施例的方法。It can be understood that the computer storage medium can be applied to the above-mentioned mobile terminal or the above-mentioned server to implement the method of any one of the above-mentioned embodiments.
参阅图11,图11是本申请提供的移动终端第二实施例的结构示意图,移动终端110包括:获取模块111、提取模块112、处理模块113。Referring to FIG. 11, FIG. 11 is a schematic structural diagram of a second embodiment of a mobile terminal provided by the present application. The mobile terminal 110 includes: an acquisition module 111, an extraction module 112, and a processing module 113.
获取模块111用于获取关于应用程序的介绍需求信息;其中,介绍需求信息用于表示对于介绍应用程序的需求;The obtaining module 111 is used for obtaining introduction requirement information about the application program; wherein, the introduction requirement information is used to indicate the requirement for introducing the application program;
提取模块112用于提取介绍需求信息中的关键词;The extraction module 112 is used to extract keywords in the introduction demand information;
获取模块111还用于基于关键词获取相关联的图像和语音;The obtaining module 111 is also used to obtain related images and voices based on keywords;
处理模块113用于对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。The processing module 113 is used to process images and voices to form a video for introducing the application program.
参阅图12,图12是本申请提供的服务器第二实施例的结构示意图,服务器120包括:获取模块121、处理模块122、发送模块123。Referring to FIG. 12, FIG. 12 is a schematic structural diagram of a second embodiment of a server provided by the present application. The server 120 includes: an obtaining module 121, a processing module 122, and a sending module 123.
获取模块121用于获取移动终端发送的关键词;其中,关键词是移动终端基于获取的关于应用程序的介绍需求信息提取得到的,介绍需求信息用于表示对于介绍应用程序的需求;The obtaining module 121 is used to obtain keywords sent by the mobile terminal; wherein, the keywords are extracted by the mobile terminal based on the obtained introduction requirement information about the application, and the introduction requirement information is used to indicate the requirement for introducing the application;
处理模块122用于基于关键词生成相关联的图像和语音;The processing module 122 is configured to generate associated images and voices based on keywords;
发送模块123用于向移动终端发送图像和语音,以使移动终端对图像和语音进行处理,以形成用于对应用程序进行介绍的视频。The sending module 123 is configured to send images and voices to the mobile terminal, so that the mobile terminal processes the images and voices to form a video for introducing the application program.
在本申请所提供的几个实施方式中,应该理解到,所揭露的方法以及设备,可以通过其它的方式实现。例如,以上所描述的设备实施方式仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。In the several implementation manners provided in this application, it should be understood that the disclosed method and device may be implemented in other ways. For example, the device implementation described above is only illustrative. For example, the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, for example, multiple units or components may be Combined or can be integrated into another system, or some features can be ignored or not implemented.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施方式方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of this embodiment.
另外,在本申请各个实施方式中的各功能单元可以集成在一个处理 单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
上述其他实施方式中的集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)或处理器(processor)执行本申请各个实施方式所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。If the integrated unit in the other embodiments described above is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium. Based on this understanding, the technical solution of the present application essentially or the part that contributes to the existing technology or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , Including several instructions to enable a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute all or part of the steps of the methods described in the various embodiments of the present application. The aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks and other media that can store program codes. .
以上仅为本申请的实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。The above are only examples of this application, and do not limit the scope of this application. Any equivalent structure or equivalent process transformation made using the content of the description and drawings of this application, or directly or indirectly applied to other related technical fields, The same reasoning is included in the scope of patent protection of this application.

Claims (16)

  1. 一种应用程序的介绍方法,其特征在于,包括:A method for introducing an application program, which is characterized in that it includes:
    获取关于应用程序的介绍需求信息;其中,所述介绍需求信息用于表示对于介绍所述应用程序的需求;Acquire introduction requirement information about the application program; wherein, the introduction requirement information is used to indicate a requirement for introducing the application program;
    提取所述介绍需求信息中的关键词;Extract the keywords in the introduction demand information;
    基于所述关键词获取相关联的图像和语音;Acquiring associated images and voices based on the keywords;
    对所述图像和所述语音进行处理,以形成用于对所述应用程序进行介绍的视频。The image and the voice are processed to form a video for introducing the application program.
  2. 根据权利要求1所述的方法,其特征在于,The method of claim 1, wherein:
    所述介绍需求信息为音频信息;The introduction requirement information is audio information;
    所述提取所述介绍需求信息中的关键词,包括:The extraction of keywords in the introduction requirement information includes:
    对所述音频信息进行语音识别,以得到文本信息;Performing voice recognition on the audio information to obtain text information;
    对所述文本信息进行关键词提取,以得到关键词。Keyword extraction is performed on the text information to obtain keywords.
  3. 根据权利要求2所述的方法,其特征在于,The method of claim 2, wherein:
    所述对所述文本信息进行关键词提取,以得到关键词,包括:The keyword extraction on the text information to obtain keywords includes:
    对所述文本信息进行语义分割;Perform semantic segmentation on the text information;
    基于所述语义分割的结果得到关键词。A keyword is obtained based on the result of the semantic segmentation.
  4. 根据权利要求3所述的方法,其特征在于,The method of claim 3, wherein:
    所述对所述文本信息进行语义分割,包括:The semantic segmentation of the text information includes:
    将所述文本信息输入至卷积神经网络进行深度学习,以将所述文本信息进行语义分割,以得到关键词。The text information is input to a convolutional neural network for deep learning, so as to perform semantic segmentation on the text information to obtain keywords.
  5. 根据权利要求1所述的方法,其特征在于,The method of claim 1, wherein:
    所述介绍需求信息为文本信息;The introduction requirement information is text information;
    所述提取所述介绍需求信息中的关键词,包括:The extraction of keywords in the introduction requirement information includes:
    对所述文本信息进行语义分割;Perform semantic segmentation on the text information;
    基于所述语义分割的结果得到关键词。A keyword is obtained based on the result of the semantic segmentation.
  6. 根据权利要求1所述的方法,其特征在于,The method of claim 1, wherein:
    所述基于所述关键词获取相关联的图像和语音,包括:The obtaining the associated image and voice based on the keyword includes:
    将所述关键词发送给服务器,以使所述服务器基于所述关键词生成相关联的图像和语音;Sending the keywords to a server, so that the server generates associated images and voices based on the keywords;
    获取所述服务器发送的所述图像和所述语音。Acquiring the image and the voice sent by the server.
  7. 根据权利要求1所述的方法,其特征在于,The method of claim 1, wherein:
    所述对所述图像和所述语音进行处理,以形成用于对所述应用程序进行介绍的视频,包括:The processing the image and the voice to form a video for introducing the application program includes:
    对多个所述对应的图像进行图像分割,提取所述图像中特征信息;Performing image segmentation on a plurality of the corresponding images, and extracting feature information in the images;
    将所述特征信息进行组合,以生成多个图像帧;Combining the feature information to generate multiple image frames;
    将所述多个图像帧形成动画;Forming the plurality of image frames into an animation;
    将所述动画与所述语音进行融合,以形成用于对所述应用程序进行介绍的视频。The animation and the voice are merged to form a video for introducing the application.
  8. 根据权利要求7所述的方法,其特征在于,The method according to claim 7, wherein:
    所述方法还包括:The method also includes:
    获取所述服务器发送的背景音乐;其中,所述背景音乐是所述服务器基于所述关键词生成的音乐;Acquiring background music sent by the server; wherein the background music is music generated by the server based on the keywords;
    将所述背景音乐添加至所述视频。Add the background music to the video.
  9. 一种应用程序的介绍方法,其特征在于,包括:A method for introducing an application program, which is characterized in that it includes:
    获取移动终端发送的关键词;其中,所述关键词是所述移动终端基于获取的关于应用程序的介绍需求信息提取得到的,所述介绍需求信息用于表示对于介绍所述应用程序的需求;Acquiring keywords sent by the mobile terminal; wherein the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application, and the introduction demand information is used to indicate the demand for introducing the application;
    基于所述关键词生成相关联的图像和语音;Generate related images and voices based on the keywords;
    向所述移动终端发送所述图像和所述语音,以使所述移动终端对所述图像和所述语音进行处理,以形成用于对所述应用程序进行介绍的视频。The image and the voice are sent to the mobile terminal, so that the mobile terminal processes the image and the voice to form a video for introducing the application program.
  10. 根据权利要求9所述的方法,其特征在于,The method of claim 9, wherein:
    所述基于所述关键词生成相关联的图像和语音,包括:The generating associated images and voices based on the keywords includes:
    将所述关键词通过深度学习,以从预设图像库得到相关联的图像。The keywords are subjected to deep learning to obtain associated images from a preset image library.
  11. 根据权利要求10所述的方法,其特征在于,The method of claim 10, wherein:
    所述基于所述关键词生成相关联的图像和语音,包括:The generating associated images and voices based on the keywords includes:
    将所述关键词通过深度学习,以生成符合所述关键词场景的文字信息;Pass the keyword through deep learning to generate text information that meets the keyword scene;
    将所述文字信息转换为所述语音。Convert the text information into the voice.
  12. 一种移动终端,其特征在于,所述移动终端包括处理器以及与所述处理器连接的存储器;A mobile terminal, characterized in that the mobile terminal includes a processor and a memory connected to the processor;
    所述存储器用于存储程序数据,所述处理器用于执行所述程序数据,以实现如权利要求1-8任一项所述的方法。The memory is used to store program data, and the processor is used to execute the program data to implement the method according to any one of claims 1-8.
  13. 一种服务器,其特征在于,所述服务器包括处理器以及与所述处理器连接的存储器;A server, characterized in that the server includes a processor and a memory connected to the processor;
    所述存储器用于存储程序数据,所述处理器用于执行所述程序数据,以实现如权利要求9-11任一项所述的方法。The memory is used to store program data, and the processor is used to execute the program data to implement the method according to any one of claims 9-11.
  14. 一种计算机存储介质,其特征在于,所述计算机存储介质用于存 储程序数据,所述程序数据在被处理器执行时,用于实现如权利要求1-11任一项所述的方法。A computer storage medium, wherein the computer storage medium is used to store program data, and when the program data is executed by a processor, it is used to implement the method according to any one of claims 1-11.
  15. 一种移动终端,其特征在于,所述移动终端包括:A mobile terminal, characterized in that, the mobile terminal includes:
    获取模块,用于获取关于应用程序的介绍需求信息;其中,所述介绍需求信息用于表示对于介绍所述应用程序的需求;The obtaining module is used to obtain introduction requirement information about the application program; wherein, the introduction requirement information is used to indicate the requirement for introducing the application program;
    提取模块,用于提取所述介绍需求信息中的关键词;The extraction module is used to extract keywords in the introduction demand information;
    所述获取模块还用于基于所述关键词获取相关联的图像和语音;The acquisition module is also used to acquire associated images and voices based on the keywords;
    处理模块,用于对所述图像和所述语音进行处理,以形成用于对所述应用程序进行介绍的视频。The processing module is configured to process the image and the voice to form a video for introducing the application program.
  16. 一种服务器,其特征在于,所述服务器包括:A server, characterized in that the server includes:
    获取模块,用于获取移动终端发送的关键词;其中,所述关键词是所述移动终端基于获取的关于应用程序的介绍需求信息提取得到的,所述介绍需求信息用于表示对于介绍所述应用程序的需求;The acquiring module is used to acquire keywords sent by the mobile terminal; wherein, the keywords are extracted by the mobile terminal based on the acquired introduction demand information about the application, and the introduction demand information is used to indicate that the introduction Application requirements;
    处理模块,用于基于所述关键词生成相关联的图像和语音;A processing module for generating associated images and voices based on the keywords;
    发送模块,用于向所述移动终端发送所述图像和所述语音,以使所述移动终端对所述图像和所述语音进行处理,以形成用于对所述应用程序进行介绍的视频。The sending module is configured to send the image and the voice to the mobile terminal, so that the mobile terminal processes the image and the voice to form a video for introducing the application program.
PCT/CN2019/104000 2019-09-02 2019-09-02 Application introduction method, mobile terminal, and server WO2021042234A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201980010315.5A CN111801673A (en) 2019-09-02 2019-09-02 Application program introduction method, mobile terminal and server
PCT/CN2019/104000 WO2021042234A1 (en) 2019-09-02 2019-09-02 Application introduction method, mobile terminal, and server

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2019/104000 WO2021042234A1 (en) 2019-09-02 2019-09-02 Application introduction method, mobile terminal, and server

Publications (1)

Publication Number Publication Date
WO2021042234A1 true WO2021042234A1 (en) 2021-03-11

Family

ID=72805590

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/104000 WO2021042234A1 (en) 2019-09-02 2019-09-02 Application introduction method, mobile terminal, and server

Country Status (2)

Country Link
CN (1) CN111801673A (en)
WO (1) WO2021042234A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117041627B (en) * 2023-09-25 2024-03-19 宁波均联智行科技股份有限公司 Vlog video generation method and electronic equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104731959A (en) * 2015-04-03 2015-06-24 北京威扬科技有限公司 Video abstraction generating method, device and system based on text webpage content
CN106547748A (en) * 2015-09-16 2017-03-29 中国移动通信集团公司 The creation method and device of a kind of APP index databases, the method and device of search APP
CN108965737A (en) * 2017-05-22 2018-12-07 腾讯科技(深圳)有限公司 media data processing method, device and storage medium
CN109145152A (en) * 2018-06-28 2019-01-04 中山大学 A kind of self-adapting intelligent generation image-text video breviary drawing method based on query word

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104820546B (en) * 2014-12-30 2018-02-13 广州酷狗计算机科技有限公司 Function information methods of exhibiting and device
CN106648675A (en) * 2016-12-28 2017-05-10 乐蜜科技有限公司 Method and device for displaying application program using information and electronic equipment
CN106919317A (en) * 2017-02-27 2017-07-04 珠海市魅族科技有限公司 A kind of information displaying method and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104731959A (en) * 2015-04-03 2015-06-24 北京威扬科技有限公司 Video abstraction generating method, device and system based on text webpage content
CN106547748A (en) * 2015-09-16 2017-03-29 中国移动通信集团公司 The creation method and device of a kind of APP index databases, the method and device of search APP
CN108965737A (en) * 2017-05-22 2018-12-07 腾讯科技(深圳)有限公司 media data processing method, device and storage medium
CN109145152A (en) * 2018-06-28 2019-01-04 中山大学 A kind of self-adapting intelligent generation image-text video breviary drawing method based on query word

Also Published As

Publication number Publication date
CN111801673A (en) 2020-10-20

Similar Documents

Publication Publication Date Title
CN110427617B (en) Push information generation method and device
CN108197111B (en) Text automatic summarization method based on fusion semantic clustering
US9754585B2 (en) Crowdsourced, grounded language for intent modeling in conversational interfaces
CN109582952B (en) Poetry generation method, poetry generation device, computer equipment and medium
CN108538286A (en) A kind of method and computer of speech recognition
CN111160452A (en) Multi-modal network rumor detection method based on pre-training language model
CN110347790B (en) Text duplicate checking method, device and equipment based on attention mechanism and storage medium
CN110287314B (en) Long text reliability assessment method and system based on unsupervised clustering
CN110472043B (en) Clustering method and device for comment text
CN110807324A (en) Video entity identification method based on IDCNN-crf and knowledge graph
CN111966826A (en) Method, system, medium and electronic device for constructing text classification system
CN111506794A (en) Rumor management method and device based on machine learning
CN110493612B (en) Barrage information processing method, server and computer readable storage medium
EP4060548A1 (en) Method and device for presenting prompt information and storage medium
CN109635303B (en) Method for recognizing meaning-changing words in specific field
CN115470344A (en) Video barrage and comment theme fusion method based on text clustering
CN113407842B (en) Model training method, theme recommendation reason acquisition method and system and electronic equipment
WO2021042234A1 (en) Application introduction method, mobile terminal, and server
CN110781327B (en) Image searching method and device, terminal equipment and storage medium
CN115188376A (en) Personalized voice interaction method and system
CN114428852A (en) Chinese text abstract extraction method and device based on BERT pre-training model
CN114328910A (en) Text clustering method and related device
KR20220143229A (en) Apparatus and method extracting core sentence based on korean language model
CN114722267A (en) Information pushing method and device and server
CN112016003A (en) Social crowd user tag mining and similar user recommending method based on CNN

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19943978

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19943978

Country of ref document: EP

Kind code of ref document: A1