CN113689633A - Scenic spot human-computer interaction method, device and system - Google Patents

Scenic spot human-computer interaction method, device and system Download PDF

Info

Publication number
CN113689633A
CN113689633A CN202110986296.9A CN202110986296A CN113689633A CN 113689633 A CN113689633 A CN 113689633A CN 202110986296 A CN202110986296 A CN 202110986296A CN 113689633 A CN113689633 A CN 113689633A
Authority
CN
China
Prior art keywords
index
knowledge
video
tourist
ticket
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110986296.9A
Other languages
Chinese (zh)
Other versions
CN113689633B (en
Inventor
郑勇
陈海江
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Lishi Technology Co Ltd
Original Assignee
Zhejiang Lishi Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Lishi Technology Co Ltd filed Critical Zhejiang Lishi Technology Co Ltd
Priority to CN202110986296.9A priority Critical patent/CN113689633B/en
Publication of CN113689633A publication Critical patent/CN113689633A/en
Application granted granted Critical
Publication of CN113689633B publication Critical patent/CN113689633B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/75Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0241Advertisements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/14Travel agencies
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Strategic Management (AREA)
  • Data Mining & Analysis (AREA)
  • Tourism & Hospitality (AREA)
  • Economics (AREA)
  • Accounting & Taxation (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Business, Economics & Management (AREA)
  • Development Economics (AREA)
  • Multimedia (AREA)
  • Finance (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Marketing (AREA)
  • Game Theory and Decision Science (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Probability & Statistics with Applications (AREA)
  • Human Resources & Organizations (AREA)
  • Primary Health Care (AREA)
  • Databases & Information Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a scenic spot human-computer interaction method, a device and a system, wherein the method comprises the following steps: pre-establishing a knowledge base aiming at a target scenic spot, wherein the knowledge base comprises: one or a combination of the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge in the scenic spot; the method comprises the steps of obtaining a voice ticket-taking instruction sent by a tourist, and analyzing the voice ticket-taking instruction of the tourist by using an NLP (line segment processing) module, wherein the voice ticket-taking instruction comprises the following steps: a ticket fetching instruction; carrying out authentication processing on the tourist and drawing a ticket; and acquiring a voice consultation instruction sent by a user, matching the voice consultation instruction based on a pre-established knowledge base, and outputting a matching result. The invention utilizes voice recognition to intelligently ask and answer, can also realize explanation on basic knowledge, basic environment, basic service and the like of scenic spots, and enhances the participation sense of tourists. The temperature is provided in the ticketing link, and the utilization rate of the self-service device is improved.

Description

Scenic spot human-computer interaction method, device and system
Technical Field
The invention relates to a man-machine interaction method, in particular to a scenic spot man-machine interaction method, a device and a system.
Background
With the development of tourism and the progress of modern science and technology, most scenic spots are provided with scenic spot automatic ticketing and marketing distribution systems, which play a role in relieving the pressure of scenic spot ticketing, reducing the time of tourists queuing for ticketing and improving the experience of tourists, wherein the self-service ticketing terminal mainly uses an automatic ticketing machine to realize the functions of inquiry, ticketing, payment and the like. The background system of the automatic ticket selling and taking machine comprises a main control module, a union pay card payment module and a ticket drawing module. The master control module is used for controlling the operation of the union pay card payment module and the ticket drawing module, the union pay card payment module is used for completing the fee deduction operation of the union pay card according to the fee deduction instruction of the master control module, and the ticket drawing module is used for completing the ticket drawing operation according to the ticket drawing instruction of the master control module.
However, there are two main problems in the prior art: 1. the self-service ticket selling and taking machine is not practical due to more carrying functions in the whole view, and has negative effects brought by complicated operation; 2. neglecting the emotional feeling of the tourists, having no interaction with the tourists, having low utilization rate, people still tend to manually ticket selling windows. And further, the input-output ratio of the self-service ticket selling and taking machine is not high, the investment is lost, and the comprehensive benefit of the scenic spot is low.
Disclosure of Invention
The invention aims to provide a scenic spot human-computer interaction method, device and system, and aims to solve the problem of low comprehensive benefit caused by low equipment utilization rate due to weak participation of tourists in the prior art.
The invention solves the technical problems through the following technical scheme:
the invention provides a scenic spot human-computer interaction method, which comprises the following steps:
pre-establishing a knowledge base aiming at a target scenic spot, wherein the knowledge base comprises: one or a combination of the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge in the scenic spot;
the method comprises the steps of obtaining a voice ticket-taking instruction sent by a tourist, and analyzing the voice ticket-taking instruction of the tourist by using an NLP (line segment processing) module, wherein the voice ticket-taking instruction comprises the following steps: a ticket fetching instruction; carrying out authentication processing on the tourist and drawing a ticket;
and acquiring a voice consultation instruction sent by a user, matching the voice consultation instruction based on a pre-established knowledge base, and outputting a matching result.
Optionally, the knowledge base of the target scenic spot further includes a customized knowledge base, and the establishing process of the customized knowledge base includes:
acquiring related knowledge, and classifying the related knowledge by using a clustering algorithm to obtain a plurality of knowledge classifications;
acquiring a scene label corresponding to each knowledge classification, and marking the corresponding knowledge classification by using the scene label, wherein the scene label comprises: meeting scene labels, research scene labels, hotel scene labels, natural knowledge scene labels, historical culture knowledge scene labels, or a combination thereof.
Optionally, the method further includes:
for each tourist, collecting video data of the tourist by using a preset camera, and cutting the video data into a plurality of video sections;
for each video segment, extracting user emotional characteristics, and outputting the user emotional characteristics to a pre-trained first neural network model to obtain an emotional index corresponding to the video segment, wherein the emotional characteristics include:
sound, motion, expression or a combination thereof;
when the emotion index is higher than a first preset threshold value, taking the video segment as a target video segment;
and adding the target video segment into the VLOG of the tourist.
Optionally, when the emotional features of the user are extracted, the method further includes:
evaluating the aesthetic feeling index of each video frame of each video segment by utilizing a pre-constructed second neural network;
when the aesthetic feeling index is higher than a second preset threshold value, taking the video frame as a target video frame;
adding the target video frame to the guest's VLOG.
Optionally, when the emotional features of the user are extracted, the method further includes:
for each video segment, extracting images of tourists contained in the video segment, and evaluating the color index, the stature index and the lap index of the tourists by utilizing a pre-constructed third neural network;
taking an index with a value larger than a third preset threshold value from the color index, the stature index and the lap index as a target index;
when the target index is a color value index, using a video segment of a close shot and/or a video frame of the close shot contained in the video segment;
when the target index is a figure index or a cross-over index, adding a video segment of a distant view and/or a video frame of the distant view included in the video segment into the VLOG of the tourist;
when the target index is a color value index, adding a video segment of a close shot and/or a video frame of the close shot included in the video segment of the close shot into the VLOG of the tourist;
and when the target index is a color value index, a stature index and a cross-matching index, adding a video frame of a close scene in the video segment of the close scene and/or a video frame of a close scene in the video segment of the close scene, and a video frame of a far scene in the video segment of the far scene and/or a video frame of the far scene in the VLOG of the tourist.
Optionally, when the target index does not include the color value index, the stature index and the lap index, extracting a video segment and/or a video frame of the visitor, and adding the extracted content into the VLOG of the visitor.
Optionally, the extracting the video segments and/or the video frames of the guest includes:
evaluating a second aesthetic index of each video frame by utilizing a pre-constructed fourth neural network aiming at each video frame of each video segment; calculating a third aesthetic index of the video segment based on the aesthetic index of the video frame; when the second aesthetic index is larger than a fourth preset threshold value, adding the video frame into the VLOG of the tourist; and when the third aesthetic index is larger than a fifth preset threshold value, adding the video segment into the VLOG of the tourist. Optionally, after extracting the video segments and/or video frames of the guest, before adding the extracted content to the guest's VLOG, the method further comprises:
and performing beautifying processing on the images of the tourists in the extracted content by using a beautifying algorithm.
The invention also provides a scenic spot human-computer interaction device, which comprises:
the system comprises an establishing module, a searching module and a display module, wherein the establishing module is used for establishing a knowledge base aiming at a target scenic spot in advance, and the knowledge base comprises: one or a combination of the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge in the scenic spot; the acquisition module is used for acquiring a voice ticket-taking instruction sent by the tourist and analyzing the voice ticket-taking instruction of the tourist by using the NLP module, wherein the voice ticket-taking instruction comprises: a ticket fetching instruction; carrying out authentication processing on the tourist and drawing a ticket;
and acquiring a voice consultation instruction sent by a user, matching the voice consultation instruction based on a pre-established knowledge base, and outputting a matching result.
The invention also provides a scenic spot human-computer interaction system, which comprises: the system comprises an input/output controller, an in-place sensor, a human body detection sensor, a help calling button, a ticket gate indicator lamp and a buzzer, wherein the in-place sensor, the human body detection sensor, the help calling button, the ticket gate indicator lamp and the buzzer are connected with the input/output controller;
the intelligent bank card comprises a main controller connected with an input/output controller, a Unionpay card module, a two-dimensional code reader, an identity card reader, a camera, a receipt printer, an entrance ticket printer, a touch display screen, an LED rolling screen and a voice module, wherein the main controller is used for executing the method, and the Unionpay card module, the two-dimensional code reader, the identity card reader, the camera, the receipt printer, the entrance ticket printer, the touch display screen, the LED rolling screen and the voice module are connected with the main controller.
Compared with the prior art, the invention has the following advantages:
according to the embodiment of the invention, through the NLP intelligent answering module, the man-machine interaction of the ticket selling and taking link is realized, the intelligent question answering is realized from voice recognition, the explanation on basic knowledge, basic environment, basic service and the like of scenic spots can be realized, and the participation sense of tourists is enhanced. The ticket selling link has temperature, so that tourists are willing to use the family interaction equipment, the utilization rate of the self-service device is improved, and further the comprehensive benefits of scenic spots are improved.
Drawings
Fig. 1 is a schematic flow chart of a scenic spot human-computer interaction method according to an embodiment of the present invention;
FIG. 2 is a schematic diagram illustrating a method for man-machine interaction in a scenic spot according to an embodiment of the present invention;
fig. 3 is a schematic diagram of an algorithm architecture in a scenic spot human-computer interaction method according to an embodiment of the present invention;
fig. 4 is a schematic diagram of a VLOG generation algorithm in a scenic spot human-computer interaction method according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of a human-computer interaction system in a scenic spot according to an embodiment of the present invention.
Detailed Description
The following examples are given for the detailed implementation and specific operation of the present invention, but the scope of the present invention is not limited to the following examples.
Example 1
The embodiment of the invention can be deployed on a self-traveling assistant platform.
Fig. 1 is a schematic flow chart of a scenic spot human-computer interaction method according to an embodiment of the present invention; FIG. 2 is a schematic diagram illustrating a method for man-machine interaction in a scenic spot according to an embodiment of the present invention; fig. 3 is a schematic diagram of an algorithm architecture in a scenic spot human-computer interaction method provided by an embodiment of the present invention, as shown in fig. 1-3, the method includes: step S1: pre-establishing a knowledge base aiming at a target scenic spot, wherein the knowledge base comprises: and one or a combination of the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge in the scenic spot.
Illustratively, the knowledge base constructed for the tourism industry comprises the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge for scenic spots, and fully covers the information consultation requirements of tourists before, during and after the tourism.
Further, a customized knowledge base can be constructed for different scenes, including:
and (3) conference knowledge: aiming at a knowledge base framework of the exhibition, the common relevant knowledge of the exhibition is included; the scene label is a conference scene label;
research knowledge: a knowledge framework aiming at research scenes comprises research culture and culture content related knowledge; the scene label is a research scene label;
hotel knowledge: the knowledge base framework aiming at hotel and accommodation scenes comprises hotel services, catering services, invoices, check-in and other related knowledge; the scene label is a hotel scene label;
natural encyclopedia knowledge: a knowledge framework for museums and science popularization property scenes, including natural science related common knowledge; the scene label is a natural knowledge scene label;
history culture knowledge: the knowledge frame customized for the historical culture venue comprises local culture and historical humanity related knowledge, and the scene label is a historical culture knowledge scene label.
S2: the method comprises the steps of obtaining a voice ticket-taking instruction sent by a tourist, and analyzing the voice ticket-taking instruction of the tourist by using an NLP (line segment processing) module, wherein the voice ticket-taking instruction comprises the following steps: a ticket fetching instruction; and (4) carrying out authentication processing on the tourists and drawing a ticket. The multifunctional integrated scenic spot marketing device automatically acquires ticket business information of the tourists by means of face recognition, identity card induction or mobile phone numbers of the tourists and the like. The printing module in the multifunctional integrated scenic spot marketing device is started, the ticket of the tourist is made and printed, and then the printing function of the ticket is realized, and the tourist is prompted to carry the ticket to check the ticket. Or, the tourist sends a ticket buying instruction to the multi-function integrated scenic spot marketing device, for example, the device receives the instruction and then starts the NLP module and the self-traveling assistant platform; the natural language is analyzed through the NLP module, and the multifunctional integrated scenic spot marketing device can meet mobile payment modes such as code scanning payment and face brushing payment according to ticket purchasing requirements of tourists; through interacting with the tourists, the printing of paper tickets and the generation of electronic tickets are supported, and the tourists can conveniently select the paper tickets according to own preferences. In addition, voice operation is simpler than clicking operation which requires searching for a specific single page or key.
Specifically, NLP (Natural Language Processing) module research is a comprehensive study, which is far from machine learning algorithms. Text changes are more complex than images or speech, which is a cross domain of computer science, artificial intelligence, and linguistics. The target is to let the computer process or "understand" natural language, and divide task NLP basic technology such as language translation and question answering into three types of lexical analysis, syntactic analysis and semantic analysis. The lexical analysis comprises word segmentation and part-of-speech tagging, wherein the word segmentation method is realized based on a word list matching and statistical model, namely, character strings are scanned word by word, and when a substring of the character strings is found to be the same as words in a word list, the character strings are matched; and modeling the Chinese according to the part of speech and the statistical characteristics of the manual labeling, calculating the probability of the occurrence of various word segmentations through the model, and taking the word segmentation result with the maximum probability as a final result. The purpose of part-of-speech tagging is to assign a specific category to each word in the sentence, i.e. to tag part-of-speech for each word in the word segmentation result. A method based on rules and a method of statistical models, a part-of-speech disambiguation rule is built by a part-of-speech collocation relationship and a context; and calculating the occurrence probability of various parts of speech through the model, and taking the part of speech with the maximum probability as a final result.
Syntactic parsing is the main task of parsing a syntactical parsing, which is to parse an input text sentence (string) to obtain a sentence syntactical structure. On the one hand the nlu (natural language understanding) task's own requirements, and on the other hand may provide support for other nlu tasks. Tasks are classified into 3 types according to different expression forms of syntax structures: dependency syntax analysis dependency syntax parsing, mainly aiming at identifying the interdependence relationship among the vocabularies in the sentence; phrase structure syntax analysis phrase-structure syntax parsing, also called syntactic analysis syntax parsing, the main task is to identify the hierarchical syntax relationship between phrase structures and phrases in sentences; the deep grammar syntactic analysis is mainly used for carrying out deep syntactic and semantic analysis on a sentence by utilizing deep grammars, wherein the deep grammars comprise adjacent lexical trees, functional lexical grammars, combined category grammars and the like.
Semantic analysis refers to meaning included in natural language, and in the field of computer science, semantics can be understood as meaning of a concept represented by a real-world object corresponding to data. Semantic analysis refers to applying various machine learning methods to make a machine learn and understand semantic contents represented by a text. Any understanding of a language can be classified as semantic analysis, involving linguistics, computational linguistics, artificial intelligence, machine learning, and even cognitive language. The ultimate goal of semantic analysis is to understand the true meaning of a sentence expression.
S201 (not shown in the figure): the method comprises the steps of preprocessing an original voice signal input by a tourist and carrying out A/D conversion. Converting the sound continuous waveform into discrete data points by sampling and quantizing with a certain sampling rate and sampling bit number; the MFCC feature extraction is used, and a first-order high-pass filter is used, so that the acoustic model can better utilize a high-frequency formant, and the identification accuracy is improved; endpoint detection and frame windowing, extracting the significant portion of the speech input data. In this MFCC, windowing uses a Hamming window where the edges are smoothly dropped to 0.
S202 (not shown in the figure): and (4) feature processing, namely using Mel cepstrum coefficient feature processing. And extracting MFCC parameters and PLCC parameters, performing maximum likelihood value estimation on a training process of the four parameters of a state transition matrix A, a weight matrix C with mixed Gaussian distribution, a mean vector II and a covariance matrix U, and processing the speech input characteristics of the tourists.
S203 (not shown in the figure): and generating a voice training template, and identifying according to distortion judgment criteria such as Euclidean distance, covariance matrix, Bayesian distance and the like.
S204 (not shown): and performing pattern matching, namely using hidden Markov pattern matching of the HMM to determine an optimal state transition sequence, and calculating the output probability of the observation sequence to the HMM model, or performing judgment according to the speech training template of S103. And judging the recognition result of the voice command of the tourist according to the judgment result.
S205 (not shown): obtaining an identification result, using OpenMP programming to realize multi-core tourist voice parallel operation, using OpenMP to improve two-dimensional discrete fast Fourier transform, using a voice matching result as a multi-operation result, finally programming through an FFT algorithm to output a natural sequence, and outputting a tourist query result.
S3: and acquiring a voice consultation instruction sent by a user, matching the voice consultation instruction based on a pre-established knowledge base, and outputting a matching result.
The tourists can send out consultation instructions to the multi-function integrated scenic spot marketing device according to the demands of the tourists, and the device starts the NLP module to communicate with the tourists after receiving the instructions; the method comprises the steps of analyzing natural languages through an NLP (neural-linear Programming) module, supporting multiple languages such as Mandarin, northeast, Cantonese and English to communicate, matching each scene label according to analyzed semantics to obtain a scene label with the highest matching degree, matching the analyzed semantics with each knowledge in the scene label according to the scene label to obtain a matching result, and displaying the matching result to a user.
For example: the tourists need to consult nearby food, and can say that the device says 'what food is recommended nearby', and the multifunctional integrated scenic spot marketing device lists a food ranking list; when the tourist needs to consult a nearby toilet, the tourist can say that the nearby toilet is located with the device, the device can display a map, show the current position of the tourist, mark the position of the nearby toilet, and meanwhile, the device can lead the tourist to go to the nearby toilet. Furthermore, a user track database, a behavior database, a question and answer database, a basic travel database and the like can be preset in the multifunctional integrated scenic spot marketing device, and the data are input into the tourist question prediction model, the consultation recommendation algorithm model and the guidance recommendation algorithm model, so that the behavior prediction of the tourists is realized, and the experience of the tourists is improved.
According to the embodiment of the invention, through the NLP intelligent answering module, the man-machine interaction of the ticket selling and taking link is realized, the intelligent question answering is realized from voice recognition, the explanation on basic knowledge, basic environment, basic service and the like of scenic spots can be realized, and the participation sense of tourists is enhanced. The temperature is kept in the ticketing link, the utilization rate of the self-service device is improved, and the man-machine interaction in the ticketing link is realized;
moreover, a powerful knowledge base is created, and a closed loop of data acquisition, analysis and feedback is constructed, so that the quality and efficiency of customer service can be continuously and iteratively improved; the data can provide support for management, marketing of the destination platform, and can further be used to enable ecological merchants.
In addition, the cognitive ability is continuously improved by means of a deep learning technology. And providing multiple intelligent text processing capabilities including intelligent word segmentation, entity recognition, keyword extraction, intelligent chatting, text and travel knowledge query and the like. The method can be widely applied to scenes such as tourist answering, function retrieval in the journey, complaint analysis in the journey process and the like, and the text intelligent requirements in the literature and travel industry are met.
Example 2
Fig. 4 is a schematic view of a VLOG generation algorithm in a scenic spot human-computer interaction method provided by an embodiment of the present invention, as shown in fig. 4, based on embodiment 1 of the present invention, embodiment 2 adds the following steps on the basis of embodiment 1:
s4 (not shown): and aiming at each tourist, the video data of the tourist is collected by utilizing a preset camera, and the video data is cut into a plurality of video sections.
The cameras are configured according to needs to carry out point location deployment; the intelligent edge all-in-one machine is locally deployed, the number of the intelligent edge all-in-one machines is configured according to the number of the cameras, the intelligent edge all-in-one machines are generally deployed in a central machine room, are in data transmission with a camera local area network, and are in butt joint with a VLOG operation cloud platform through a public network fixed IP. The VLOG operation platform is deployed in the Alice cloud, a scenic spot administrator can log in and manage through a web terminal, and a consumer can log in and select to manufacture and obtain the VLOG through a small program and an APP. The system automatically cuts the guest's video into video segments, typically 3-5 seconds in length.
S5 (not shown): for each video segment, extracting user emotional characteristics, and outputting the user emotional characteristics to a pre-trained first neural network model to obtain an emotional index corresponding to the video segment, wherein the emotional characteristics include: sound, motion, expression or a combination thereof. The user expression may include: fear, surprise, feeling, embarrassment, shyness, happiness, etc.
It can be understood that the training method of the first neural network model includes: manually identifying volume data, language data and the like in sound generated by sample tourists in the scenic spot in the playing process to obtain an emotion index, and then taking the emotion index as a tag of the sound data to obtain a sound sample; manually identifying according to action amplitude, action acceleration and the like in the actions of the sample tourists, identifying the exciting degree of the sample tourists to further obtain an emotion index, and taking the emotion index as a label of the action data of the sample tourists to further obtain an action sample of the sample tourists; similarly, quantification is carried out according to the degree of the expression, such as the level of surprise of the sample tourist and the level of embarrassment of the sample tourist, so as to obtain a corresponding emotion index and further obtain an expression sample; and training the first neural network model by using the sound sample, the action sample and the expression sample to further obtain a trained model.
It can be understood that when the game of the tourist is finished and the exclusive tourist plays the photos and videos, the privacy and the uniqueness of the photos need to be ensured.
S6 (not shown): and when the emotion index is higher than a first preset threshold value, taking the video segment as a target video segment.
It is understood that the deeper the emotional depth, the greater the value corresponding to the emotional index. For example, a larger voice indicates a higher emotion index, and a faster pace indicates a higher emotion index. The more embarrassing the expression of the tourist, the higher the emotion index is.
Therefore, in the embodiment of the invention, the video segment with the emotion index ranking of the top 10% can be selected as the target video segment; and taking a boundary of the top 10% of the ranking as a first preset threshold.
S7 (not shown): and adding the target video segment into the VLOG of the tourist.
Furthermore, when the photos of the tourists are synthesized into the exclusive and private customized VLOG, the images, the color distribution of the videos, the texture characteristics and the shapes of main objects can be identified by using an image identification algorithm, spots are identified by using a Gaussian Laplacian operator detection method (LOG), local characteristics of corner points are identified by using a Harris algorithm and a FAST algorithm, and different characteristics of the images are classified according to different identification algorithms. The guest may select a favorite category or categories based on the classification results. Then, the existing VLOG generation algorithm is used to perform overlay synthesis and arrangement processing on the contents selected by the guest, so as to obtain an arrangement picture, and the arrangement picture processing is called as VLOG.
After the visitor went into the garden, VLOG terminal equipment can snatch the wonderful of visitor in the scenic spot in the twinkling of an eye to upload data information to back-end system platform, when the visitor needed the exclusive video of VLOG oneself, can consult multi-functional integrated scenic spot marketing device. For example, "help me generate my exclusive micro video", the device can find the picture or video information of the corresponding guest by face recognition of the guest, automatically clip and synthesize the picture or video information, and automatically generate the exclusive micro video of the corresponding guest through the VLOG platform.
By means of the VLOG technology, wonderful images and videos of tourists in scenic spots are captured, permanent memory of the tourists in the scenic spots is synthesized, the drips played by the tourists in the scenic spots are recorded, the memory feeling of the tourists to the scenic spots is enhanced, the scenic spots can be circularly played by utilizing the synthesized videos to conduct scenic spot propaganda, and a scenic spot propaganda content library is formed.
The ticket selling mode with single function of the traditional self-service ticket selling and taking machine is enriched, the pressure of queuing for buying tickets in an artificial ticket selling window is relieved, and the working efficiency of selling and taking tickets is improved; the function of interactive communication with the tourists is added, more ways for the tourists to consult problems are provided, the tourists can speak the needs of the tourists, and the operation is more convenient and fast; the wonderful moment of the tourists in the scenic spot is recorded, and a special micro video is generated, so that the satisfaction degree of the tourists is improved; through the circulation of the scenic spot micro-video, on the other hand, propaganda ways of the scenic spot are increased, and the brand awareness of the scenic spot is improved.
Example 3
The embodiment 3 of the invention is added with the following steps on the basis of the embodiment 2:
evaluating the aesthetic feeling index of each video frame of each video segment by utilizing a pre-constructed second neural network; when the aesthetic feeling index is higher than a second preset threshold value, taking the video frame as a target video frame; adding the target video frame to the guest's VLOG.
A large number of video segments or video frames are stored in the system, and the aesthetic feeling of the video segments or the video frames is subjective manually from the aspects of composition, light and shadow cooperation, coordination between people and nature and the like; and (3) calculating an average value by adopting multiple scoring times by multiple people, and taking the average value as the aesthetic indexes of the video frames or video segments, wherein the value range of each index is 1-10, and the aesthetic indexes are no exception. And then training a second neural network by using the scored video frames and video segments, and then performing aesthetic evaluation on the video frames or the video segments of the tourists by using the trained second neural network to obtain a corresponding aesthetic index.
By applying the embodiment of the invention, pictures with high aesthetic feeling can be screened out and added into the VLOG, and because the pictures are shot by the camera, on one hand, the pictures can give surprise to tourists in unconscious times, and on the other hand, the invention uses the second neural network to identify the aesthetic feeling pictures, so that the pictures are more professional, the effect of the obtained pictures is better, and the user experience is further improved.
Or, when the emotional features of the user are extracted, the method further includes:
for each video segment, extracting images of tourists contained in the video segment, and evaluating the color index, the stature index and the lap index of the tourists by utilizing a pre-constructed third neural network;
taking an index with a value larger than a third preset threshold value from the color index, the stature index and the lap index as a target index; when the target index is a color value index, using a video segment of a close shot and/or a video frame of the close shot contained in the video segment; when the target index is a figure index or a cross-over index, adding a video segment of a distant view and/or a video frame of the distant view included in the video segment into the VLOG of the tourist; when the target index is a color value index, adding a video segment of a close shot and/or a video frame of the close shot included in the video segment of the close shot into the VLOG of the tourist; and when the target index is a color value index, a stature index and a cross-matching index, adding a video frame of a close scene in the video segment of the close scene and/or a video frame of a close scene in the video segment of the close scene, and a video frame of a far scene in the video segment of the far scene and/or a video frame of the far scene in the VLOG of the tourist.
For example, the system makes intelligent recommendations based on the guest's color value, stature, fit, make-up.
If the visitor has a high color value and a good stature and wins the lottery, the VLOG is generated by fusing a far scene and a near scene and a video and a photo.
The VLOG is generated in an intelligent mode by mainly selecting distant view or image conversion processing due to high color value and common stature of tourists. The tourist has common color value and good stature. The lens special effect and the beauty special effect are started, and the visitor color value is improved comprehensively. Intelligent video compositing uses mainly perspective to generate VLOG.
And when the visitor image is not good in the color value, the figure and the putting on and putting off, performing the beautifying processing on the visitor image in the extracted content by using the existing beautifying algorithm.
Or when the visitor is not good in color value, stature and putting on, the standard is properly reduced, such as:
regarding each video frame of each video segment, taking the video frame with the aesthetic feeling index ranking 10 percent as the video frame of the tourist;
and taking the video segment with the aesthetic index ranking 10 percent of the video segments as the video segment for the tourist.
Therefore, by applying the embodiment of the invention, the condition that the video frames or video segments of the tourists do not meet the common aesthetic standard can be avoided, a certain number of video frames and video segments can still be screened out, and the experience of the tourists is improved.
Example 4
Based on any one of embodiments 1 to 3, embodiment 4 of the present invention provides a scenic spot human-computer interaction device, including:
the system comprises an establishing module, a searching module and a display module, wherein the establishing module is used for establishing a knowledge base aiming at a target scenic spot in advance, and the knowledge base comprises: one or a combination of the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge in the scenic spot; the acquisition module is used for acquiring a voice ticket-taking instruction sent by the tourist and analyzing the voice ticket-taking instruction of the tourist by using the NLP module, wherein the voice ticket-taking instruction comprises: a ticket fetching instruction; carrying out authentication processing on the tourist and drawing a ticket;
and acquiring a voice consultation instruction sent by a user, matching the voice consultation instruction based on a pre-established knowledge base, and outputting a matching result.
Example 5
Fig. 5 is a schematic structural diagram of a human-computer interaction system in a scenic spot according to an embodiment of the present invention, and as shown in fig. 5, based on any one of embodiments 1 to 4, embodiment 5 of the present invention provides a human-computer interaction system in a scenic spot, where the system includes: the system comprises an input/output controller, an in-place sensor, a human body detection sensor, a help calling button, a ticket gate indicator lamp and a buzzer, wherein the in-place sensor, the human body detection sensor, the help calling button, the ticket gate indicator lamp and the buzzer are connected with the input/output controller;
a main controller connected with the input and output controller, the main controller is used for executing the method as claimed in any one of claims 1 to 8, and a Unionpay card module, a two-dimensional code reader, an identity card reader, a camera, a receipt printer, an entrance ticket printer, a touch display screen, an LED scroll screen and a voice module which are connected with the main controller. The device adopts an I/O controller designed by an embedded industrial control computer and a microprocessor, has definite division of labor, strong function, high reliability and convenient function expansion; the damp-proof, mould-proof and salt fog-proof design of the electronic circuit board ensures the long-term stable operation of the I/O controller of the self-service machine; strict grounding measures and lightning protection designs are adopted, and the safety is good; standard components and a universal interface are adopted, so that upgrading and maintenance are convenient;
TABLE 1 multifunctional integrated scenic spot marketing device VLOG module function list
Figure BDA0003230795740000141
Figure BDA0003230795740000151
TABLE 2 parameters of multifunctional integrated scenic spot marketing device
Figure BDA0003230795740000152
Figure BDA0003230795740000161
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents and improvements made within the spirit and principle of the present invention are intended to be included within the scope of the present invention.

Claims (10)

1. A scenic spot human-computer interaction method is characterized by comprising the following steps:
pre-establishing a knowledge base aiming at a target scenic spot, wherein the knowledge base comprises: one or a combination of the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge in the scenic spot;
the method comprises the steps of obtaining a voice ticket-taking instruction sent by a tourist, and analyzing the voice ticket-taking instruction of the tourist by using an NLP (line segment processing) module, wherein the voice ticket-taking instruction comprises the following steps: a ticket fetching instruction; carrying out authentication processing on the tourist and drawing a ticket;
and acquiring a voice consultation instruction sent by a user, matching the voice consultation instruction based on a pre-established knowledge base, and outputting a matching result.
2. The human-computer interaction method for scenic spots as claimed in claim 1, wherein the knowledge base of the target scenic spot further comprises a customized knowledge base, and the establishment process of the customized knowledge base comprises:
acquiring related knowledge, and classifying the related knowledge by using a clustering algorithm to obtain a plurality of knowledge classifications;
acquiring a scene label corresponding to each knowledge classification, and marking the corresponding knowledge classification by using the scene label, wherein the scene label comprises: meeting scene labels, research scene labels, hotel scene labels, natural knowledge scene labels, historical culture knowledge scene labels, or a combination thereof.
3. A scenic spot human-computer interaction method as claimed in claim 1, further comprising:
for each tourist, collecting video data of the tourist by using a preset camera, and cutting the video data into a plurality of video sections;
for each video segment, extracting user emotional characteristics, and outputting the user emotional characteristics to a pre-trained first neural network model to obtain an emotional index corresponding to the video segment, wherein the emotional characteristics include: sound, motion, expression or a combination thereof;
when the emotion index is higher than a first preset threshold value, taking the video segment as a target video segment;
and adding the target video segment into the VLOG of the tourist.
4. A scenic spot human-computer interaction method according to claim 3, wherein when the emotional features of the user are extracted, the method further comprises:
evaluating the aesthetic feeling index of each video frame of each video segment by utilizing a pre-constructed second neural network;
when the aesthetic feeling index is higher than a second preset threshold value, taking the video frame as a target video frame;
adding the target video frame to the guest's VLOG.
5. The human-computer interaction method for scenic spots as claimed in claim 1, wherein when the emotional features of the user are extracted, the method further comprises:
for each video segment, extracting images of tourists contained in the video segment, and evaluating the color index, the stature index and the lap index of the tourists by utilizing a pre-constructed third neural network;
taking an index with a value larger than a third preset threshold value from the color index, the stature index and the lap index as a target index;
when the target index is a color value index, using a video segment of a close shot and/or a video frame of the close shot contained in the video segment;
when the target index is a figure index or a cross-over index, adding a video segment of a distant view and/or a video frame of the distant view included in the video segment into the VLOG of the tourist;
when the target index is a color value index, adding a video segment of a close shot and/or a video frame of the close shot included in the video segment of the close shot into the VLOG of the tourist;
and when the target index is a color value index, a stature index and a cross-matching index, adding a video frame of a close scene in the video segment of the close scene and/or a video frame of a close scene in the video segment of the close scene, and a video frame of a far scene in the video segment of the far scene and/or a video frame of the far scene in the VLOG of the tourist.
6. The human-computer interaction method for scenic spots of claim 5, wherein when none of the target indexes comprises a color value index, a stature index and a cross-matching index, video segments and/or video frames of the tourists are extracted, and the extracted contents are added to VLOGs of the tourists.
7. A scenic spot human-computer interaction method according to claim 6, wherein the extracting of video segments and/or video frames of tourists comprises:
regarding each video frame of each video segment, taking the video frame with the aesthetic feeling index ranking 10 percent as the video frame of the tourist;
and taking the video segment with the aesthetic index ranking 10 percent of the video segments as the video segment for the tourist.
8. A scenic spot human-computer interaction method as claimed in claim 6, wherein after extracting a guest's video segment and/or video frame, before adding the extracted content to the guest's VLOG, the method further comprises:
and performing beautifying processing on the images of the tourists in the extracted content by using a beautifying algorithm.
9. A scenic spot human-computer interaction device, the device comprising:
the system comprises an establishing module, a searching module and a display module, wherein the establishing module is used for establishing a knowledge base aiming at a target scenic spot in advance, and the knowledge base comprises: one or a combination of the playing knowledge, the traffic knowledge, the food culture knowledge and the popular characteristic knowledge in the scenic spot;
the acquisition module is used for acquiring a voice ticket-taking instruction sent by the tourist and analyzing the voice ticket-taking instruction of the tourist by using the NLP module, wherein the voice ticket-taking instruction comprises: a ticket fetching instruction; carrying out authentication processing on the tourist and drawing a ticket;
and acquiring a voice consultation instruction sent by a user, matching the voice consultation instruction based on a pre-established knowledge base, and outputting a matching result.
10. A scenic spot human-computer interaction system, the system comprising: the system comprises an input/output controller, an in-place sensor, a human body detection sensor, a help calling button, a ticket gate indicator lamp and a buzzer, wherein the in-place sensor, the human body detection sensor, the help calling button, the ticket gate indicator lamp and the buzzer are connected with the input/output controller;
a main controller connected with the input and output controller, the main controller is used for executing the method as claimed in any one of claims 1 to 8, and a Unionpay card module, a two-dimensional code reader, an identity card reader, a camera, a receipt printer, an entrance ticket printer, a touch display screen, an LED scroll screen and a voice module which are connected with the main controller.
CN202110986296.9A 2021-08-26 2021-08-26 Scenic spot human-computer interaction method, device and system Active CN113689633B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110986296.9A CN113689633B (en) 2021-08-26 2021-08-26 Scenic spot human-computer interaction method, device and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110986296.9A CN113689633B (en) 2021-08-26 2021-08-26 Scenic spot human-computer interaction method, device and system

Publications (2)

Publication Number Publication Date
CN113689633A true CN113689633A (en) 2021-11-23
CN113689633B CN113689633B (en) 2023-03-17

Family

ID=78582766

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110986296.9A Active CN113689633B (en) 2021-08-26 2021-08-26 Scenic spot human-computer interaction method, device and system

Country Status (1)

Country Link
CN (1) CN113689633B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115374268A (en) * 2022-10-25 2022-11-22 广州市明道文化产业发展有限公司 Multi-role decentralized collaborative interaction method and system

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020198706A1 (en) * 2001-05-07 2002-12-26 Yu-Hung Kao Implementing a high accuracy continuous speech recognizer on a fixed-point processor
JP2008216461A (en) * 2007-03-01 2008-09-18 Nec Corp Speech recognition, keyword extraction, and knowledge base retrieval coordinating device
US20090248671A1 (en) * 2008-03-28 2009-10-01 Daisuke Maruyama Information classification system, information processing apparatus, information classification method and program
CN104078039A (en) * 2013-03-27 2014-10-01 广东工业大学 Voice recognition system of domestic service robot on basis of hidden Markov model
CN206282340U (en) * 2016-12-23 2017-06-27 上海木爷机器人技术有限公司 Electronic bill processing system
CN107451276A (en) * 2017-08-05 2017-12-08 龙飞 A kind of intelligent self-service guide system and its method based on deep learning
US10026394B1 (en) * 2012-08-31 2018-07-17 Amazon Technologies, Inc. Managing dialogs on a speech recognition platform
CN108766419A (en) * 2018-05-04 2018-11-06 华南理工大学 A kind of abnormal speech detection method based on deep learning
CN208314860U (en) * 2018-05-10 2019-01-01 科大讯飞股份有限公司 A kind of intelligent sound ticket-selling device and system
CN110263220A (en) * 2019-06-28 2019-09-20 北京奇艺世纪科技有限公司 A kind of video highlight segment recognition methods and device
CN111651554A (en) * 2020-04-17 2020-09-11 世纪保众(北京)网络科技有限公司 Insurance question-answer method and device based on natural language understanding and processing
CN112182299A (en) * 2020-09-25 2021-01-05 北京字节跳动网络技术有限公司 Method, device, equipment and medium for acquiring highlight segments in video
CN113225586A (en) * 2020-02-05 2021-08-06 字节跳动有限公司 Video processing method and device, electronic equipment and storage medium

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020198706A1 (en) * 2001-05-07 2002-12-26 Yu-Hung Kao Implementing a high accuracy continuous speech recognizer on a fixed-point processor
JP2008216461A (en) * 2007-03-01 2008-09-18 Nec Corp Speech recognition, keyword extraction, and knowledge base retrieval coordinating device
US20090248671A1 (en) * 2008-03-28 2009-10-01 Daisuke Maruyama Information classification system, information processing apparatus, information classification method and program
US10026394B1 (en) * 2012-08-31 2018-07-17 Amazon Technologies, Inc. Managing dialogs on a speech recognition platform
CN104078039A (en) * 2013-03-27 2014-10-01 广东工业大学 Voice recognition system of domestic service robot on basis of hidden Markov model
CN206282340U (en) * 2016-12-23 2017-06-27 上海木爷机器人技术有限公司 Electronic bill processing system
CN107451276A (en) * 2017-08-05 2017-12-08 龙飞 A kind of intelligent self-service guide system and its method based on deep learning
CN108766419A (en) * 2018-05-04 2018-11-06 华南理工大学 A kind of abnormal speech detection method based on deep learning
CN208314860U (en) * 2018-05-10 2019-01-01 科大讯飞股份有限公司 A kind of intelligent sound ticket-selling device and system
CN110263220A (en) * 2019-06-28 2019-09-20 北京奇艺世纪科技有限公司 A kind of video highlight segment recognition methods and device
CN113225586A (en) * 2020-02-05 2021-08-06 字节跳动有限公司 Video processing method and device, electronic equipment and storage medium
CN111651554A (en) * 2020-04-17 2020-09-11 世纪保众(北京)网络科技有限公司 Insurance question-answer method and device based on natural language understanding and processing
CN112182299A (en) * 2020-09-25 2021-01-05 北京字节跳动网络技术有限公司 Method, device, equipment and medium for acquiring highlight segments in video

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
许喜占: "《人像摄影》", 30 April 2011 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115374268A (en) * 2022-10-25 2022-11-22 广州市明道文化产业发展有限公司 Multi-role decentralized collaborative interaction method and system

Also Published As

Publication number Publication date
CN113689633B (en) 2023-03-17

Similar Documents

Publication Publication Date Title
US10977452B2 (en) Multi-lingual virtual personal assistant
CN107492379B (en) Voiceprint creating and registering method and device
JP3848319B2 (en) Information processing method and information processing apparatus
US7225128B2 (en) System and method for providing information using spoken dialogue interface
CN104050160B (en) Interpreter's method and apparatus that a kind of machine is blended with human translation
CN107657017A (en) Method and apparatus for providing voice service
CN109658271A (en) A kind of intelligent customer service system and method based on the professional scene of insurance
CN110349564A (en) Across the language voice recognition methods of one kind and device
WO2007043679A1 (en) Information processing device, and program
CN108986790A (en) The method and apparatus of voice recognition of contact
CN113380271B (en) Emotion recognition method, system, device and medium
CN109325124A (en) A kind of sensibility classification method, device, server and storage medium
KR101988165B1 (en) Method and system for improving the accuracy of speech recognition technology based on text data analysis for deaf students
CN113689633B (en) Scenic spot human-computer interaction method, device and system
CN116343747A (en) Speech synthesis method, speech synthesis device, electronic device, and storage medium
CN114996506A (en) Corpus generation method and device, electronic equipment and computer-readable storage medium
CN110516083A (en) Photograph album management method, storage medium and electronic equipment
CN113850083A (en) Method, device and equipment for determining broadcast style and computer storage medium
CN113763925B (en) Speech recognition method, device, computer equipment and storage medium
US11681364B1 (en) Gaze prediction
CN114492382A (en) Character extraction method, text reading method, dialog text generation method, device, equipment and storage medium
CN115705705A (en) Video identification method, device, server and storage medium based on machine learning
JP6087704B2 (en) Communication service providing apparatus, communication service providing method, and program
CN111554269A (en) Voice number taking method, system and storage medium
US11792365B1 (en) Message data analysis for response recommendations

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant