CN112132095B - Dangerous state identification method and device, electronic equipment and storage medium - Google Patents

Dangerous state identification method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN112132095B
CN112132095B CN202011062013.3A CN202011062013A CN112132095B CN 112132095 B CN112132095 B CN 112132095B CN 202011062013 A CN202011062013 A CN 202011062013A CN 112132095 B CN112132095 B CN 112132095B
Authority
CN
China
Prior art keywords
information
target user
lip
preset
dangerous state
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011062013.3A
Other languages
Chinese (zh)
Other versions
CN112132095A (en
Inventor
谭皓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority to CN202011062013.3A priority Critical patent/CN112132095B/en
Publication of CN112132095A publication Critical patent/CN112132095A/en
Application granted granted Critical
Publication of CN112132095B publication Critical patent/CN112132095B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/237Lexical tools
    • G06F40/242Dictionaries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/26Government or public services
    • G06Q50/265Personal security, identity or safety
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Business, Economics & Management (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Tourism & Hospitality (AREA)
  • Psychiatry (AREA)
  • Computer Security & Cryptography (AREA)
  • Development Economics (AREA)
  • Educational Administration (AREA)
  • Social Psychology (AREA)
  • Economics (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • General Business, Economics & Management (AREA)
  • Image Analysis (AREA)

Abstract

The application discloses a dangerous state identification method, a dangerous state identification device, electronic equipment and a storage medium, and relates to the technical field of safety. Acquiring a current face image of a target user, identifying lip features of the target user based on the face image, performing lip recognition based on the lip features to obtain a recognition result, and matching the recognition result with preset lip information to obtain a matching result, wherein the preset lip information is preset according to the lip information representing a dangerous state, and determining whether the target user is in the dangerous state based on the matching result. In this way, a determination of whether the user is in a dangerous state based on the lip motion of the user may be achieved.

Description

Dangerous state identification method and device, electronic equipment and storage medium
Technical Field
The present disclosure relates to the field of security technologies, and in particular, to a method and apparatus for identifying a dangerous state, an electronic device, and a storage medium.
Background
With rapid progress in living standard and technological level, people have a higher demand for safety protection, and more safety protection technologies are also in the life of people. In the related art, a user may actively report a dangerous state and further obtain help, however, the convenience and safety of such a manner are to be improved.
Disclosure of Invention
In view of the foregoing, the present application proposes a dangerous state identification method, device, electronic apparatus and storage medium, so as to improve the above-mentioned problems.
In a first aspect, an embodiment of the present application provides a method for identifying a dangerous state, where the method includes: acquiring a current face image of a target user; identifying lip features of the target user based on the face image; performing lip language identification based on the lip features to obtain an identification result; matching the identification result with preset lip language information to obtain a matching result, wherein the preset lip language information is preset according to the lip language information representing the dangerous state; and determining whether the target user is in a dangerous state or not based on the matching result.
In a second aspect, an embodiment of the present application provides an apparatus for identifying a dangerous state, the apparatus including: the system comprises an image acquisition module, a lip recognition module, a lip matching module and a state determination module. The image acquisition module is used for acquiring the current face image of the target user; the lip recognition module is used for recognizing the lip characteristics of the target user based on the face image; the lip language identification module is used for carrying out lip language identification based on the lip features to obtain an identification result; the lip language matching module is used for matching the identification result with preset lip language information to obtain a matching result, wherein the preset lip language information is preset according to the lip language information representing the dangerous state; and the state determining module is used for determining whether the target user is in a dangerous state or not based on the matching result.
In a third aspect, an embodiment of the present application provides an electronic device, including: one or more processors; a memory; one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the one or more programs configured to perform the method of identifying a hazard condition provided in the first aspect.
In a fourth aspect, embodiments of the present application provide a computer readable storage medium having stored therein program code that is callable by a processor to perform the method of identifying a dangerous state provided in the first aspect.
Compared with the prior art, in the scheme provided by the application, the lip characteristics of the target user are identified based on the face image by acquiring the current face image of the target user, the lip recognition is performed based on the lip characteristics, the recognition result is obtained, the recognition result is matched with the preset lip information, and the matching result is obtained, wherein the preset lip information is preset according to the lip information representing the dangerous state, and whether the target user is in the dangerous state is determined based on the matching result. Therefore, the user can judge whether the user is in a dangerous state or not according to the lip language identification result by carrying out lip language identification on the target user, so that the user can trigger and identify the dangerous state through lip actions, the user can report the dangerous state conveniently, and the safety can be improved due to the secrecy of the lip actions.
These and other aspects of the present application will be more readily apparent from the following description of the embodiments.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are needed in the description of the embodiments will be briefly introduced below, it being obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a flow chart of a dangerous state identification method according to an embodiment of the present application.
Fig. 2 is a flow chart illustrating a method for identifying a dangerous state according to another embodiment of the present application.
Fig. 3 shows a schematic flow chart of the substeps of step S240 shown in fig. 2 in one embodiment.
Fig. 4 is a flow chart of a method for identifying a dangerous state according to another embodiment of the present application.
Fig. 5 shows a schematic flow chart of the substeps of step S460 shown in fig. 4 in an embodiment.
Fig. 6 is a block diagram of a dangerous state identification device according to an embodiment of the present application.
Fig. 7 is a block diagram of an electronic device for performing the method of identifying a dangerous state according to an embodiment of the present application.
Fig. 8 is a storage unit for storing or carrying program code for implementing the path generating method according to the embodiment of the present application.
Detailed Description
In order to enable those skilled in the art to better understand the present application, the following description will make clear and complete descriptions of the technical solutions in the embodiments of the present application with reference to the accompanying drawings in the embodiments of the present application.
In practical application, network communication is increasingly developed, people rely on and use networks increasingly, business such as takeaway, windward, network friend making and the like are also spreading and coming in the world, intersections among strangers are also increasing, but after the intersections with the strangers are generated, the strangers may expose themselves to danger. In daily news, people often hear the news that the people catch a windward, take out from a spot or meet with an internet friend and meet the injury, and when the people meet the danger, the people can choose to call for an alarm or send help seeking information to families or friends through social software.
The inventor discovers through long-term research that in the processing mode when the dangerous situation is encountered, the user is required to report the dangerous state through voice input or manual operation, which is quite inconvenient. And the arrival of danger is often foreseeable, and under the condition that criminals are fully prepared and lose intelligence, trade and alarm or ask for help probably irritate the criminals, so that the parties are in a more dangerous state.
In order to solve the above problems, the inventor proposes a dangerous state identification method, a dangerous state identification device and an electronic device, and can judge whether a user is in a dangerous state according to an identification result by carrying out lip language identification on a target user, thereby judging whether alarm information needs to be sent according to the state of the user. This will be described in detail below.
Referring to fig. 1, fig. 1 is a flow chart of a dangerous state identification method according to an embodiment of the present application. The method for identifying a dangerous state according to the embodiment of the present application will be described in detail with reference to fig. 1. The dangerous state identification method may include the steps of:
step S110: and acquiring the current face image of the target user.
In the embodiment of the application, the lip action of the target user can be identified to identify whether the target user is in a dangerous state. When the electronic device recognizes the lip motion of the target user, the electronic device may acquire the current face image thereof to recognize the lip motion thereof through the face image thereof. The target user can refer to any user, and also can refer to a user of the electronic device; the current face image of the target user refers to a face image obtained by collecting the face image of the target user at the current moment.
In some embodiments, the electronic device may perform acquiring the current face image of the target user at preset time intervals, and perform subsequent steps to determine whether the target user is in a dangerous state. The preset time may be preset, for example, according to a user operation, and the specific size of the preset time of the electronic device may not be limited, for example, the preset time may be 5 minutes, or may be 10 minutes.
In other embodiments, the electronic device may also perform the acquiring the current face image of the target user when detecting an instruction for triggering recognition of the dangerous state. It will be appreciated that the user may instruct the electronic device to identify whether it is in a dangerous state, i.e. perform the process of the dangerous state identification method, by actively inputting a corresponding instruction to the electronic device.
In this embodiment, the instructions for triggering the identification of the dangerous state may include: one or more of a preset gesture input by a target user and detected by the electronic device, an operation on a specified key and a detected key combination operation and a detected specified expression of the target user can be preset by the electronic device, and specific content can be not limited.
In some implementations, the electronic device may include an image capture device, such as a color camera, a depth camera, or the like. When the electronic equipment needs to identify whether the target user is in a dangerous state or not, the face image of the target user can be acquired through the image acquisition device. As an implementation manner, the electronic device may collect multiple frames of face images of the target user at present, for example, collect multiple frames of face images continuously, that is, collect a video including a face of the target user; as another implementation manner, the electronic device may also collect a current frame of face image of the target user.
Step S120: and identifying lip features of the target user based on the face image.
In the embodiment of the application, the electronic device can identify the lip language characteristics of the target user based on the acquired face image of the target user so as to identify the lip action of the target user, identify the lip language of the target user and further judge whether the target user is in a dangerous state or not.
In some embodiments, a lip image of the lip region may be extracted from the face image, and then lip features may be extracted from the lip image to obtain the lip features of the target user. The lip region can be segmented from the face image in a threshold segmentation mode, so that a lip image of the lip region is obtained; the face image may be input into the region extraction model through a region extraction model of the lip region trained in advance, so as to obtain the lip image output by the region extraction model, and the specific manner of obtaining the lip image may not be limited. When the lip features are extracted according to the lip images, the lip images can be input into the feature extraction model through a feature extraction model trained in advance, so that the lip features output by the feature extraction model are obtained; lip feature points can also be extracted from the lip image by singular value decomposition, discrete cosine transform, discrete wavelet transform, and the like. The above region extraction model may be a deep neural network, and the like, and is not limited herein.
In this embodiment, when the electronic device obtains that the face image of the target user is a continuous multi-frame face image, that is, a video including the face of the target user, the lip image in each frame of image may be extracted, the lip image in each frame of image may be adjusted to the same size, and the extracted lip image sample data may be spliced and stored according to the chronological order, so as to generate the lip feature dataset. The video containing the face of the target user needs to be post-processed to make the frame rates equal, for example, the frame rates are all 30f/s.
In the above embodiment, when the lip feature is extracted, the lip image may be obtained first, and then the corresponding lip feature may be obtained from the lip image, based on which, in the process of performing lip language recognition, interference of other irrelevant facial information may be eliminated, and the recognition accuracy may be improved.
Step S130: and carrying out lip language identification based on the lip characteristics to obtain an identification result.
In the embodiment of the application, the electronic device can identify the identified lip features so as to acquire lip information corresponding to the lip features, wherein the lip information is a final identification result, and further whether the target user is in a dangerous state or not can be judged based on the identification result.
In some embodiments, the lip language recognition may be understood as obtaining the content of the target user expressed by the lip language by analyzing the lip movement characteristics of the target user, and the analysis process may include two steps, namely, pinyin sequence recognition and Chinese character sequence recognition, wherein the pinyin sequence recognition maps continuous lip feature images into pinyin sentences, and the Chinese character sequence recognition is to translate the pinyin sequence into corresponding Chinese character sentences. As an implementation mode, the pinyin sequence recognition can be performed by a convolutional neural network model, a pinyin sequence recognition network framework and other methods, further, the Chinese character sequence recognition can be performed based on an Encoder-Decoder model and a Chinese character sequence recognition network, so that lip language information corresponding to the lip feature is obtained, and the embodiment of the application is not limited to the lip language recognition method.
Step S140: and matching the identification result with preset lip language information to obtain a matching result, wherein the preset lip language information is preset according to the lip language information representing the dangerous state.
In the embodiment of the present application, since the preset lip information is preset according to the lip information representing the dangerous state, after the electronic device obtains the recognition result of the target user's lip, the recognition result may be matched with the preset lip information to determine whether the target user is in the dangerous state.
In some embodiments, the preset lip information may be lip information that is identified by the electronic device according to a lip image input by a user in advance. The electronic device may display a setting interface of the lip information, prompt the user to perform lip input, collect face images of the user after detecting that the user starts the operation of lip input, and finally perform lip recognition according to the collected face images of the user, so as to obtain the lip information set by the user, that is, the preset lip information.
Step S150: and determining whether the target user is in a dangerous state or not based on the matching result.
In the embodiment of the application, after the electronic device obtains the matching result of the identification result and the preset lip language information, whether the target user is in a dangerous state or not can be determined according to the matching result. When the identification result is matched with the preset lip language information, the target user can be determined to be in a dangerous state; when the identification result is not matched with the preset lip language information, the target user can be determined not to be in a dangerous state.
In this embodiment, after the lip information of the target user is obtained by the above-mentioned method for identifying the lip, the content of the lip information may be matched with the content of the preset lip information, where the content of the preset lip information may be the lip information characterized as being in a dangerous state, and by determining whether the content of the lip information is matched with the content of the preset lip information, it may be further determined whether the target user is in a dangerous state based on the matching result.
Referring to fig. 2, fig. 2 is a flow chart illustrating a method for identifying a dangerous state according to another embodiment of the present application. As will be explained in detail below with respect to the flow shown in fig. 2, the identification of the dangerous state may specifically include the following steps:
step S210: and acquiring the current face image of the target user.
Step S220: and identifying lip features of the target user based on the face image.
Step S230: and carrying out lip language identification based on the lip characteristics to obtain an identification result.
In the embodiment of the present application, the contents of step S210 to step S230 may refer to the contents of step S110 to step S130 in the foregoing embodiment, and are not described herein.
Step S240: and matching the identification result with preset lip language information to obtain a matching result, wherein the preset lip language information is preset according to the lip language information representing the dangerous state.
In some embodiments, referring to fig. 3, step S240 may include:
step S241: and obtaining the similarity between the identification result and the preset lip language information.
In the embodiment of the present application, the similarity between the recognition result and the preset lip information may be understood as the similarity between the text information corresponding to the lip feature of the target user and the text information representing the dangerous state.
In one possible example, a dictionary library for representing dangerous states is established, where the dictionary library includes all words in a training corpus, for example, "rescue me", "alarm", "rescue" or "help me", and the words in the dictionary library can represent that the user is in a dangerous state, that is, the word in the dictionary library is in the recognition result for the lip feature of the target user, and then it can be determined that the recognition result for the lip feature of the user matches with the preset lip information. In some cases, the word information expressed by the target user may not be completely consistent with the words in the dictionary library, and then the similarity between the word information expressed by the target user through the lip language and the words in the dictionary library can be judged through some algorithms, and whether the recognition result of the lip feature of the user is matched with the preset lip language information or not is judged through the similarity.
Step S242: and if the similarity is larger than a preset similarity threshold, determining that the identification result is matched with the preset lip language information.
The preset similarity threshold may be a fixed value preset in advance, for example, 0.4, and whether the identification result for the target user is matched with the preset lip language information is determined by determining the size between the acquired similarity and the preset similarity threshold. In practical application, the main application scenario of the scheme is that the user is in a dangerous state, if the preset similarity threshold value set for the target user is higher, under the condition that the user is stressed, the characters expressed by the lips may be intermittent, for example, "quick … rescue … rescue me", the similarity between the character information and the preset lip information may be lower, for example 0.5, in this case, the calculated similarity may be smaller than the similarity threshold value, correspondingly, the recognition result is not matched with the preset lip information, and further, whether the target user is in a dangerous state may be misjudged. Therefore, in the embodiment of the present application, the preset similarity threshold is generally set to be smaller, for example, 0.3 or 0.4, and under the condition that the preset similarity threshold is set to be lower, even if the text information expressed by the target user due to factors such as tension is incomplete or intermittent, the matching result between the recognition result of the target user and the preset lip language will not be misjudged.
For example, if the recognition result of the lip feature of the target user is "rescue me, rescue me", the similarity of the word information "rescue me, rescue me" and the word in the dictionary library representing the dangerous state may be calculated, for example, 0.7, the preset similarity threshold may be 0.4, correspondingly, it may be determined that the similarity is greater than the preset similarity threshold, and further, it may be determined that the recognition result of the target user matches with the preset lip language information.
Step S243: and if the similarity is smaller than or equal to a preset similarity threshold, determining that the identification result is not matched with the lip language information.
For example, if the recognition result of the lip feature of the target user is "misplaced", the similarity between the word information "misplaced" and the word in the dictionary library representing the dangerous state may be calculated, for example, 0.1, and the preset similarity threshold may be 0.4, and correspondingly, it may be determined that the similarity is smaller than the preset similarity threshold, and further, it may be determined that the recognition result of the target user is not matched with the preset lip information.
In the embodiment of the application, whether the identification result aiming at the target user is matched with the preset lip language information can also be judged by carrying out semantic analysis and emotion classification on the text information expressed by the target user. Correspondingly, each word can be mapped into a vector with a fixed dimension based on a large number of text corpuses through similar neural network model training, the dimension is between tens and hundreds of dimensions, each vector represents the word, and the semantic and grammatical similarity of the word is judged through the similarity among the vectors. The method can segment the text information expressed by the target user through the lip language, extract nouns, adjectives, adverbs, connecting words and the like in the text information, sequentially compare the similarity of the segmented text information with word vectors in a text corpus, further analyze emotion of the text information, and judge whether emotion corresponding to the text information is negative emotion, wherein the negative emotion can comprise tension, anxiety, difficulty, fear and the like.
Based on the above, the text information expressed by the target user is subjected to emotion classification, word segmentation and stop word processing can be performed on the text information, and character string matching is performed on the text information by using a member-good emotion dictionary, so that positive and negative information is mined. The emotion dictionary comprises four parts including a positive word dictionary, a negative word dictionary and a degree adverb dictionary, and each dictionary generally comprises two parts, words and weights. And performing text matching by using the emotion dictionary, namely traversing words summarized by sentences after word segmentation of the word information of the target user, if the words hit the corresponding emotion dictionary, performing corresponding weight processing, wherein the positive word weight is addition, the negative word weight is subtraction, the negative word weight takes the opposite number, the degree adverb weight is multiplied by the word weight modified by the negative word weight, and judging the emotion classification of the word information according to the finally output weight value. It will be appreciated that in the case where the target user is in a dangerous state, the emotion of the target user is generally tension, fear, anxiety, etc., and thus the corresponding weight is negative. When the semantic meaning of the text information expressed by the target user expresses the dangerous state of the user, the emotion classification corresponding to the text information is detected, if the weight corresponding to the text information is calculated to be negative, the text information can be judged to belong to negative emotion, further, the recognition result aiming at the lip characteristics of the target user is analyzed by combining semantic meaning analysis and emotion classification, and when the semantic meaning of the text information expresses the dangerous state of the user and the emotion classification corresponding to the text information is negative emotion, whether the target user is in the dangerous state can be judged further.
Step S250: and when the matching result characterizes that the identification result is matched with the preset lip language information, expression information corresponding to the face image is obtained.
In this embodiment, when the recognition result for the target user matches with the preset lip information, expression information corresponding to the face image of the target user may also be obtained.
It can be understood that when the recognition result of the target user is matched with the preset lip language information, the probability that the user is in a dangerous state can be judged to be relatively high, but in some cases, for example, a command for a child to trigger the recognition of the dangerous state by mistake or a command for triggering the recognition of the dangerous state by miscreant when the child uses the mobile phone, if the user is judged only according to the result of the lip language recognition, the situation of misjudgment can possibly occur, so that whether the target user is in the dangerous state can be further determined through facial expression recognition, and the judgment accuracy can be improved.
Step S260: judging whether the expression information meets preset expression conditions or not, wherein the preset expression conditions are preset according to the expression information representing dangerous states.
Based on the method, the obtained expression information can be judged, and whether the expression information meets the preset expression condition for representing the dangerous state or not is judged.
In some embodiments, the facial contours, feature points of the five sense organs and the emotional states of the five sense organs can be marked manually by using a convolutional neural network, using the position coordinates of the feature points of the images in the training set as input, using the emotional states corresponding to the images in the training set as output, and training the convolutional neural network to obtain the expression recognition model for expression recognition. When the expression recognition model is used for carrying out expression recognition, the facial recognition model can be used for obtaining the emotional state output by the expression recognition model based on the facial image of the target user obtained through a camera, determining the position and the shape (such as facial contours and five sense organs) of the face and main components of the face, extracting characteristic points of the face, inputting the relation (such as distance and angle) among the characteristic points into the expression recognition model as characteristic vectors, and obtaining the emotional state output by the expression recognition model as recognized expression information, so that the expression recognition is realized.
Specifically, the relation between the feature points is input into the expression model as a feature vector to obtain an emotional state output by the expression recognition model, which can be understood as that 68 feature points are marked on the face of the target user, the data coordinates obtained by the 68 feature points are obtained through calculation by a predictor, the eyebrow picking degree and the eyebrow tattooing degree of the target user can be analyzed through 10 feature points on two eyebrows, the eyebrow picking degree can correspond to happiness or surprise, the eyebrow tattooing degree can correspond to confusion, worry or fear, and the like, the opening degree of the eyes of the target user can be obtained through analyzing the feature points around the eyes, further, the opening degree of the mouth is analyzed through the feature points around the mouth, the mouth opening degree can correspond to happiness, surprise or fear, and the like, the mouth opening is not opened, namely, the 68 feature points of the face of the target user can be analyzed to realize the recognition of the target user.
Step S270: and if the expression information meets the preset expression condition, determining that the target user is in a dangerous state.
It can be understood that on the premise that the recognition result of the target user is matched with the preset lip language information, if the expression information of the target user is judged to be a fear, fear or anxiety and other negative emotion in the mode, that is, the preset expression condition is met, the target user can be determined to be in a dangerous state.
Based on this, after determining that the target user is in a dangerous state, alarm information may be transmitted through the steps shown in fig. 4.
Step S280: and if the expression information does not meet the preset expression condition, determining that the target user is not in a dangerous state.
Optionally, on the premise that the recognition result of the target user is matched with the preset lip language information, if the expression information of the target user is judged to be positive emotion such as happiness, happiness and the like, the preset condition is not met, miscreants of the user of the mobile phone or intentional testing of the function can be considered, and when the situation that the target user is not in a dangerous state is determined, the electronic equipment correspondingly does not acquire real-time position information of the target user and generates alarm information according to the position information.
Step S290: and when the matching result characterizes that the identification result is not matched with the preset lip information, determining that the target user is not in a dangerous state.
It can be understood that when the lip language information obtained by the lip language identification technology is not matched with the preset lip language information, it indicates that the target user is not in a dangerous state, and correspondingly, the electronic device cannot obtain the real-time position information of the target user and generates alarm information according to the position information.
In this embodiment, the lip feature of the target user is identified by the lip language identification technology, the identification result is analyzed, the judgment of the expression information of the target user is combined, and whether the target user is in a dangerous state or not is determined by combining the analysis result and the judgment result, so that the user can report the dangerous state of the target user conveniently.
Referring to fig. 4, fig. 4 is a flow chart illustrating a method for identifying a dangerous state according to another embodiment of the present application. As will be explained in detail below with respect to the flow shown in fig. 4, the identification of the dangerous state may specifically include the following steps:
step S410: and acquiring the current face image of the target user.
Step S420: and identifying lip features of the target user based on the face image.
Step S430: and carrying out lip language identification based on the lip characteristics to obtain an identification result.
Step S440: and matching the identification result with preset lip language information to obtain a matching result, wherein the preset lip language information is preset according to the lip language information representing the dangerous state.
Step S450: and determining whether the target user is in a dangerous state or not based on the matching result.
In the embodiment of the present application, the contents of step S410 to step S450 may refer to the contents of step S110 to step S150 in the foregoing embodiment, and are not described herein.
Step S460: and when the target user is in a dangerous state, generating alarm information, wherein the alarm information is used for indicating that the target user is in the dangerous state.
In some embodiments, the alert information may be generated through the steps described in fig. 5, i.e., step S460 may include the steps shown in fig. 5.
Step S461: and acquiring the position information of the position of the target user, wherein the position information comprises real-time positioning information and/or background environment information.
In practical application, when determining that a target user is in a dangerous state, the real-time positioning information of the position of the target user can be obtained through a positioning system carried by a mobile phone of the target user, wherein the positioning system can be a local satellite positioning system such as a GPS (global positioning system ), a BDS (Beidou satellite navigation system, beiDou Navigation Satellite System), a GLONASS (global satellite navigation system, global Navigation Satellite System) and the like, or network positioning can be used, the network positioning can be in two modes, one mode can be Wi-Fi (Wireless Fidelity) small-range positioning, positioning is performed according to the position of a Wi-Fi router, the positioning accuracy is high but unreliable through the Wi-Fi router, because google has no method for recording the position of each router on the earth, the phenomenon that the target user is positioned to other places or even other provinces is sometimes caused, and the target user is not connected to a certain Wi-Fi router in the dangerous state, and the Wi-Fi router positioning is not used in practical application; another way in network positioning may be base station positioning, which is reliable but with large errors, because it depends on the base station distribution density. The urban positioning accuracy in developed areas can be higher, and the highest positioning accuracy can be within tens of meters to hundreds of meters at present. However, when the distribution distance of the base stations in the remote areas is relatively large, the error is large, and sometimes can reach more than a few kilometers. The network positioning has the advantages that the positioning speed is high, so long as the mobile phone of the target user is in a networking state, the instant positioning can be realized, and the real-time position information of the target user in a dangerous state can be positioned in the shortest time, which is important; the satellite positioning system has the advantages that the satellite positioning system is accurate in positioning, is not limited by a network, can be positioned on the sea in the deserts of the barren cigarettes, is high in precision, and can accurately position real-time position information of a target user in a dangerous state, but the satellite positioning system depends on satellites in space, and is characterized in that the first positioning reaction is slow, and the first positioning of the mobile phone in the market at present needs 10 seconds or even more than 20 seconds. Based on the method, the network positioning and the satellite positioning can be combined for use, and the positioning is assisted, so that the positioning accuracy can be ensured, the positioning timeliness can also be ensured, and the real-time position information of the target user in a dangerous state can be accurately and quickly positioned.
In one embodiment, only the real-time positioning information of the target user in the dangerous state can be obtained, the real-time positioning information is used as the position information of the position of the target user, the position information can represent the accurate geographic position of the target user, and the position information can be the longitude and latitude of the position of the target user.
Meanwhile, background environment information of the specific position of the target user can be obtained through a camera of the mobile phone of the target user. The method includes the steps that a picture or a video containing a target user is obtained through a camera of a mobile phone, and a specific background environment where the target user is located is determined through algorithm analysis, for example, when the target user is on a taxi, the specific background environment where the target user is located can be analyzed to be in the taxi according to the collected picture or video containing the target user; when the target user is in the house, the specific background where the target user is located can be analyzed to be indoor according to the collected picture or video containing the target user; when the target user is outdoors, the specific background where the target user is located can be analyzed to be outdoors according to the collected picture or video containing the target user.
In still another embodiment, the background environment information of the target user acquired by the image acquisition device may be used as location information of the location where the target user is located, that is, the location information represents a specific scene where the user is located, where the location information may be an indoor scene, an outdoor scene, or an in-vehicle scene.
In another embodiment, it may be appreciated that combining the obtained real-time location information of the target user with the specific context information of the location of the target user may help to determine the location information of the target user more accurately, i.e., the obtained location information of the location of the target user may include the real-time location information and the specific context information.
Alternatively, there are a number of ways to determine the real-time location information of where the target user is located. In one embodiment, when it is determined that the target user is in a dangerous state, the real-time positioning information of the position where the target user is located at the current moment is obtained, that is, the real-time positioning information of the target user is obtained only once, where the obtained real-time positioning information may include longitude and latitude of the position where the target user is located and time information of the current moment, for example: (118.5, 31.5, 200927.0915), wherein 118.5 represents the longitude of the location of the target user, 31.5 represents the latitude of the location of the target user, and 200927.0915 represents the time of 9 months of 2020 at 9 am at 15 minutes.
In another embodiment, the longitude and latitude of the position of the target user can be obtained through a cyclic program, that is, the longitude and latitude of the position of the target user can be obtained every other preset time period, for example, the longitude and latitude of the position of the target user can be obtained through the positioning system every 1 second, based on the longitude and latitude, the real-time positioning information of the target user can be continuously obtained, and compared with the mode of obtaining the real-time positioning information of the target user only once, the real-time positioning information of the target user can be obtained when the target user moves, so that the user receiving the real-time positioning information can observe the position information of the target user in real time, for example: when the target user encounters danger on the vehicle during driving, the real-time position information of the target user is always changed, and the user receiving the real-time position information of the target user can display the real-time position information of the target user by means of an API (application program interface, application Programming Interface) of the map application, so that the real-time position information of the target user can be observed more intuitively.
Step S462: and generating alarm information carrying the position information based on the position information.
Optionally, the acquired real-time positioning information and/or background environment information of the target user are integrated to generate alarm information for the target user, wherein the alarm information can indicate that the user is in a dangerous state and also contains real-time position information of the target user.
Step S470: and sending the alarm information to a designated device.
The designated device may be a mobile phone of a family or/and a friend of the target user, an electronic device for receiving an alarm call by a police department, a server of an alarm platform, or the like, which is not limited in the embodiment of the present application.
In practical application, after generating corresponding alarm information, the mobile phone of the target user can automatically send the alarm information to the appointed equipment without the operation of the target user. Correspondingly, after the relevant personnel managing the designated equipment receive the alarm information, the target user can be rescued according to the position information of the target user carried by the alarm information. For example, if the family or friend of the target user receives the help seeking information, the family or friend of the target user can arrive at the corresponding position according to the position information carried in the help seeking information and help the corresponding position, and the family or friend of the target user can also alarm according to the position information carried in the help seeking information and find the target user through the help of police; if police receives the help seeking information, the police can rescue the target user directly according to the position information carried in the help seeking information.
In this embodiment, the lip feature of the target user is identified by the lip recognition technology, the identification result is analyzed, whether the target user is in a dangerous state is determined according to the analysis result, real-time position information of the target user in the dangerous state is obtained by combining the positioning technology, and alarm information about the target user is generated based on the position information, so that related rescue workers can rescue the target user according to the alarm information. The application of the lip language recognition technology in specific life is well realized, the risk that a target user is possibly injured when encountering danger is reduced, the success rate of alarming help seeking is improved, the target user is helped to send effective help seeking information in time, and the purpose of protecting personal and property safety of the user is achieved.
Referring to fig. 6, a block diagram of a dangerous state recognition device 600 according to an embodiment of the present application is shown. The apparatus 600 may include: an image acquisition module 610, a lip recognition module 620, a lip recognition module 630, a lip matching module 640, and a status determination module 650.
The image acquisition module 610 is configured to acquire a current face image of a target user.
The lip recognition module 620 is configured to recognize lip features of the target user based on the face image.
The lip recognition module 630 is configured to perform lip recognition based on the lip feature, and obtain a recognition result.
The lip matching module 630 is configured to match the identification result with preset lip information, so as to obtain a matching result, where the preset lip information is preset according to the lip information representing the dangerous state.
The state determining module 650 is configured to determine whether the target user is in a dangerous state based on the matching result.
Optionally, the state determination module 650 may include: a first determination unit and a second determination unit. The first determining unit is used for determining that the target user is in a dangerous state when the matching result characterizes that the identification result is matched with the preset lip language information; and the second determining unit is used for determining that the target user is not in a dangerous state when the matching result characterizes that the identification result is not matched with the lip language information.
Optionally, the lip matching module 630 may be further configured to obtain a similarity between the identification result and the preset lip information, determine that the identification result is matched with the preset lip information if the similarity is greater than a preset similarity threshold, and determine that the identification result is not matched with the lip information if the similarity is less than or equal to a preset similarity threshold.
Alternatively, the state determination module 650 may be specifically configured to: and when the matching result represents that the identification result is matched with the preset lip language information, acquiring expression information corresponding to the face image, judging whether the expression information meets preset expression conditions or not, wherein the preset expression conditions are preset according to the expression information representing the dangerous state, and determining that the target user is in the dangerous state if the expression information meets the preset expression conditions.
Optionally, the state determining module 650 may be further configured to determine that the target user is not in a dangerous state if the expression information does not satisfy the preset expression condition.
The dangerous state recognition device 600 may further include: and the alarm information generation module and the information sending module. The alarm information generation module is used for generating alarm information when the target user is determined to be in a dangerous state after the target user is determined to be in the dangerous state based on the matching result, wherein the alarm information is used for indicating that the target user is in the dangerous state; the alarm information generation module may be specifically configured to obtain location information of a location where the target user is located, where the location information includes real-time positioning information and/or background environment information, and generate alarm information carrying the location information based on the location information; and the information sending module is used for sending the alarm information to the appointed equipment.
The image acquisition module 610 may be specifically configured to: and executing the current face image of the target user when the instruction for triggering the identification of the dangerous state is detected.
It will be clearly understood by those skilled in the art that, for convenience and brevity of description, the specific working process of the apparatus and modules described above may refer to the corresponding process in the foregoing method embodiment, which is not repeated herein.
In several embodiments provided herein, the coupling of the modules to each other may be electrical, mechanical, or other.
In addition, each functional module in each embodiment of the present application may be integrated into one processing module, or each module may exist alone physically, or two or more modules may be integrated into one module. The integrated modules may be implemented in hardware or in software functional modules.
In summary, in the solution provided in the embodiment of the present application, by acquiring a current face image of a target user, identifying lip features of the target user based on the face image, performing lip recognition based on the lip features, obtaining a recognition result, and matching the recognition result with preset lip information to obtain a matching result, where the preset lip information is preset according to lip information representing a dangerous state, and determining whether the target user is in the dangerous state based on the matching result. Therefore, the user can judge whether the user is in a dangerous state or not according to the lip language identification result by carrying out lip language identification on the target user, so that the user can trigger and identify the dangerous state through lip actions, the user can report the dangerous state conveniently, and the safety can be improved due to the secrecy of the lip actions.
Referring to fig. 7, a block diagram of an electronic device 700 according to an embodiment of the present application is shown, and a method for identifying a dangerous state according to an embodiment of the present application may be performed by the electronic device 700.
The electronic device 700 in embodiments of the present application may include one or more of the following components: a processor 701, a memory 702, and one or more application programs, wherein the one or more application programs may be stored in the memory 702 and configured to be executed by the one or more processors 701, the one or more program configured to perform the method as described in the foregoing method embodiments.
The processor 701 may include one or more processing cores. The processor 701 utilizes various interfaces and lines to connect various portions of the overall electronic device 700, perform various functions of the electronic device 700, and process data by executing or executing instructions, programs, code sets, or instruction sets stored in the memory 702, and invoking data stored in the memory 702. Alternatively, the processor 701 may be implemented in hardware in at least one of digital signal processing (Digital Signal Processing, DSP), field programmable gate array (Field-Programmable Gate Array, FPGA), programmable logic array (Programmable Logic Array, PLA). The processor 701 may integrate one or a combination of several of a central processing unit (Central Processing Unit, CPU), an image processor (Graphics Processing Unit, GPU), and a modem, etc. The CPU mainly processes an operating system, a user interface, an application program and the like; the GPU is used for being responsible for rendering and drawing of display content; the modem is used to handle wireless communications. It will be appreciated that the modem may not be integrated into the processor 701 and may be implemented solely by a single communication chip.
The Memory 702 may include random access Memory (Random Access Memory, RAM) or Read-Only Memory (RAM). Memory 702 may be used to store instructions, programs, code, sets of codes, or instruction sets. The memory 702 may include a stored program area and a stored data area, wherein the stored program area may store instructions for implementing an operating system, instructions for implementing at least one function (e.g., a touch function, a sound playing function, an image playing function, etc.), instructions for implementing the various method embodiments described below, and the like. The storage data area may also store data created by the electronic device 700 in use (such as the various correspondences described above), and so forth.
It will be clearly understood by those skilled in the art that, for convenience and brevity of description, the specific working process of the apparatus and modules described above may refer to the corresponding process in the foregoing method embodiment, which is not repeated herein.
In the several embodiments provided herein, the illustrated or discussed coupling or direct coupling or communication connection of the modules to each other may be through some interfaces, indirect coupling or communication connection of devices or modules, electrical, mechanical, or other forms.
In addition, each functional module in each embodiment of the present application may be integrated into one processing module, or each module may exist alone physically, or two or more modules may be integrated into one module. The integrated modules may be implemented in hardware or in software functional modules.
Referring to fig. 8, a block diagram of a computer readable storage medium according to an embodiment of the present application is shown. The computer readable medium 800 has stored therein program code which can be invoked by a processor to perform the methods described in the method embodiments described above.
The computer readable storage medium 800 may be an electronic memory such as a flash memory, an EEPROM (electrically erasable programmable read only memory), an EPROM, a hard disk, or a ROM. Optionally, the computer readable storage medium 800 comprises a non-transitory computer readable medium (non-transitory computer-readable storage medium). The computer readable storage medium 800 has storage space for program code 810 that performs any of the method steps described above. The program code can be read from or written to one or more computer program products. Program code 810 may be compressed, for example, in a suitable form.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present application, and are not limiting thereof; although the present application has been described in detail with reference to the foregoing embodiments, one of ordinary skill in the art will appreciate that: the technical scheme described in the foregoing embodiments can be modified or some technical features thereof can be replaced by equivalents; such modifications and substitutions do not drive the essence of the corresponding technical solutions to depart from the spirit and scope of the technical solutions of the embodiments of the present application.

Claims (9)

1. A method for identifying a dangerous state, applied to an electronic device, the method comprising:
when an instruction for triggering the recognition of a dangerous state is detected, acquiring a current face image of a target user, wherein the instruction comprises one or more of a preset gesture, a specified expression, an operation on a specified key and a key combination operation which are input by the target user, and the target user is a current user of the electronic equipment;
identifying lip features of the target user based on the face image;
performing lip language identification based on the lip features to obtain an identification result, wherein the identification result comprises text information and emotion classification corresponding to the text information;
Matching the text information with preset lip language information to obtain a matching result, wherein the preset lip language information is preset according to the lip language information representing a dangerous state;
if the matching result represents that the text information is matched with preset lip language information and the emotion is classified as negative emotion, determining that the target user is in a dangerous state;
acquiring real-time position information of the target user, wherein the real-time position information comprises real-time positioning information and background environment information, the real-time positioning information is determined based on satellite positioning and network positioning, and the background environment information is determined based on acquired pictures containing the target user;
generating alarm information carrying the real-time position information based on the real-time position information, wherein the alarm information is used for indicating that the target user is in a dangerous state;
and sending the alarm information to a designated device.
2. The method of claim 1, wherein the determining whether the target user is in a dangerous state based on the matching result comprises:
when the matching result characterizes that the identification result is matched with the preset lip information, determining that the target user is in a dangerous state;
And when the matching result characterizes that the identification result is not matched with the preset lip information, determining that the target user is not in a dangerous state.
3. The method according to claim 2, wherein matching the identification result with preset lip language information to obtain a matching result comprises:
obtaining the similarity between the identification result and the preset lip language information;
if the similarity is larger than a preset similarity threshold, determining that the identification result is matched with the preset lip language information;
and if the similarity is smaller than or equal to a preset similarity threshold, determining that the identification result is not matched with the lip language information.
4. The method according to claim 2, wherein determining that the target user is in a dangerous state when the matching result characterizes that the recognition result matches the preset lip information comprises:
when the matching result characterizes that the identification result is matched with the preset lip information, expression information corresponding to the face image is obtained;
judging whether the expression information meets preset expression conditions or not, wherein the preset expression conditions are preset according to expression information representing dangerous states;
And if the expression information meets the preset expression condition, determining that the target user is in a dangerous state.
5. The method of claim 4, wherein the determining whether the target user is in a dangerous state based on the matching result further comprises:
and if the expression information does not meet the preset expression condition, determining that the target user is not in a dangerous state.
6. The method according to any one of claims 1-5, wherein the acquiring the current face image of the target user includes:
and executing the current face image of the target user when the instruction for triggering the identification of the dangerous state is detected.
7. A dangerous state identification device, characterized by being applied to an electronic apparatus, the device comprising:
the image acquisition module is used for acquiring a current face image of a target user when an instruction for triggering the recognition of a dangerous state is detected, wherein the instruction comprises one or more of a preset gesture, a specified expression, an operation on a specified key and a key combination operation which are input by the target user, and the target user is a current user of the electronic equipment;
The lip recognition module is used for recognizing the lip characteristics of the target user based on the face image;
the lip recognition module is used for carrying out lip recognition based on the lip characteristics to obtain a recognition result, wherein the recognition result comprises character information and emotion classification corresponding to the character information;
the lip language matching module is used for matching the identification result with preset lip language information to obtain a matching result, wherein the preset lip language information is preset according to the lip language information representing the dangerous state;
the state determining module is used for determining whether the target user is in a dangerous state or not based on the matching result;
the alarm information generation module is used for determining that the target user is in a dangerous state if the matching result represents that the text information is matched with the preset lip language information and the emotion is classified as a negative emotion; acquiring real-time position information of the target user, wherein the real-time position information comprises real-time positioning information and background environment information, the real-time positioning information is determined based on satellite positioning and network positioning, and the background environment information is determined based on acquired pictures containing the target user; generating alarm information carrying the real-time position information based on the real-time position information, wherein the alarm information is used for indicating that the target user is in a dangerous state;
And the information sending module is used for sending the alarm information to the appointed equipment.
8. An electronic device, comprising:
one or more processors;
a memory;
one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the one or more programs configured to perform the method of any of claims 1-6.
9. A computer readable storage medium, characterized in that the computer readable storage medium has stored therein a program code, which is callable by a processor for performing the method according to any one of claims 1-6.
CN202011062013.3A 2020-09-30 2020-09-30 Dangerous state identification method and device, electronic equipment and storage medium Active CN112132095B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011062013.3A CN112132095B (en) 2020-09-30 2020-09-30 Dangerous state identification method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011062013.3A CN112132095B (en) 2020-09-30 2020-09-30 Dangerous state identification method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112132095A CN112132095A (en) 2020-12-25
CN112132095B true CN112132095B (en) 2024-02-09

Family

ID=73844975

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011062013.3A Active CN112132095B (en) 2020-09-30 2020-09-30 Dangerous state identification method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112132095B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112926407A (en) * 2021-02-02 2021-06-08 华南师范大学 Distress signal detection method, device and system based on campus deception
CN115691498A (en) * 2021-07-29 2023-02-03 华为技术有限公司 Voice interaction method, electronic device and medium
CN114612977A (en) * 2022-03-10 2022-06-10 苏州维科苏源新能源科技有限公司 Big data based acquisition and analysis method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110113319A (en) * 2019-04-16 2019-08-09 深圳壹账通智能科技有限公司 Identity identifying method, device, computer equipment and storage medium
CN110427809A (en) * 2019-06-21 2019-11-08 平安科技(深圳)有限公司 Lip reading recognition methods, device, electronic equipment and medium based on deep learning
CN111045639A (en) * 2019-12-11 2020-04-21 深圳追一科技有限公司 Voice input method, device, electronic equipment and storage medium

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2368214A2 (en) * 2008-12-24 2011-09-28 David Matsumoto Systems and methods for analyzing facial expressions, identifying intent and transforming images trhough review of facial expressions
KR101840644B1 (en) * 2011-05-31 2018-03-22 한국전자통신연구원 System of body gard emotion cognitive-based, emotion cognitive device, image and sensor controlling appararus, self protection management appararus and method for controlling the same
CN110377761A (en) * 2019-07-12 2019-10-25 深圳传音控股股份有限公司 A kind of method and device enhancing video tastes

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110113319A (en) * 2019-04-16 2019-08-09 深圳壹账通智能科技有限公司 Identity identifying method, device, computer equipment and storage medium
CN110427809A (en) * 2019-06-21 2019-11-08 平安科技(深圳)有限公司 Lip reading recognition methods, device, electronic equipment and medium based on deep learning
CN111045639A (en) * 2019-12-11 2020-04-21 深圳追一科技有限公司 Voice input method, device, electronic equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
一种基于long short-term memory的唇语识别方法;马宁;田国栋;周曦;;中国科学院大学学报(第01期);全文 *

Also Published As

Publication number Publication date
CN112132095A (en) 2020-12-25

Similar Documents

Publication Publication Date Title
CN112132095B (en) Dangerous state identification method and device, electronic equipment and storage medium
CN109145680B (en) Method, device and equipment for acquiring obstacle information and computer storage medium
CN107391505B (en) Image processing method and system
US20210365343A1 (en) Artificial Intelligence (AI)-Based Cloud Computing Safety Monitoring System
CN109492221B (en) Information reply method based on semantic analysis and wearable equipment
US11521378B2 (en) Refined searching based on detected object configurations
EP3872652A2 (en) Method and apparatus for processing video, electronic device, medium and product
CN109471919B (en) Zero pronoun resolution method and device
CN111144360A (en) Multimode information identification method and device, storage medium and electronic equipment
CN112150457A (en) Video detection method, device and computer readable storage medium
US20230410221A1 (en) Information processing apparatus, control method, and program
CN112052911A (en) Method and device for identifying riot and terrorist content in image, electronic equipment and storage medium
CN113850109A (en) Video image alarm method based on attention mechanism and natural language processing
Presa-Reyes et al. Deep learning with weak supervision for disaster scene description in low-altitude imagery
CN115966061B (en) Disaster early warning processing method, system and device based on 5G message
CN106791010B (en) Information processing method and device and mobile terminal
CN108241678A (en) The method for digging and device of interest point data
CN112633244B (en) Social relationship identification method and device, electronic equipment and storage medium
CN115331097A (en) Image detection model training method and device and image detection method
CN113573009A (en) Video processing method, video processing device, computer equipment and storage medium
CN113342978A (en) City event processing method and device
CN111291597A (en) Image-based crowd situation analysis method, device, equipment and system
CN114302171B (en) Video auditing method, device and storage medium
CN116702094B (en) Group application preference feature representation method
US11423760B2 (en) Device for detecting drowning individuals or individuals in a situation presenting a risk of drowning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant