WO2021190086A1 - Face-to-face examination risk control method and apparatus, computer device, and storage medium - Google Patents
Face-to-face examination risk control method and apparatus, computer device, and storage medium Download PDFInfo
- Publication number
- WO2021190086A1 WO2021190086A1 PCT/CN2021/070931 CN2021070931W WO2021190086A1 WO 2021190086 A1 WO2021190086 A1 WO 2021190086A1 CN 2021070931 W CN2021070931 W CN 2021070931W WO 2021190086 A1 WO2021190086 A1 WO 2021190086A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- stress
- information
- target object
- face
- video
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 51
- 238000012954 risk control Methods 0.000 title claims abstract description 21
- 230000004044 response Effects 0.000 claims abstract description 68
- 238000012552 review Methods 0.000 claims description 40
- 230000008859 change Effects 0.000 claims description 29
- 230000008921 facial expression Effects 0.000 claims description 21
- 230000014509 gene expression Effects 0.000 claims description 15
- 238000000605 extraction Methods 0.000 claims description 13
- 230000001815 facial effect Effects 0.000 claims description 13
- 238000012550 audit Methods 0.000 claims description 12
- 230000033001 locomotion Effects 0.000 claims description 11
- 230000009471 action Effects 0.000 claims description 3
- 238000000354 decomposition reaction Methods 0.000 claims description 3
- 230000008569 process Effects 0.000 abstract description 14
- 238000011156 evaluation Methods 0.000 abstract description 7
- 238000004891 communication Methods 0.000 description 8
- 238000001514 detection method Methods 0.000 description 6
- 238000013507 mapping Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 4
- 238000003062 neural network model Methods 0.000 description 4
- 230000006399 behavior Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 230000001413 cellular effect Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000000513 principal component analysis Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 238000012797 qualification Methods 0.000 description 2
- 208000003028 Stuttering Diseases 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000013077 scoring method Methods 0.000 description 1
- 230000004936 stimulating effect Effects 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q40/00—Finance; Insurance; Tax strategies; Processing of corporate or income taxes
- G06Q40/03—Credit; Loans; Processing thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/06—Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
- G06Q10/063—Operations research, analysis or management
- G06Q10/0639—Performance analysis of employees; Performance analysis of enterprise or organisation operations
- G06Q10/06393—Score-carding, benchmarking or key performance indicator [KPI] analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8456—Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
Definitions
- This application relates to the field of intelligent identification technology. Specifically, this application relates to a face-to-face risk control method, device, computer equipment, and storage medium.
- the problems during the interviews are usually extracted randomly from the database.
- the stress question is manually formulated, the question is output by voice through the robot, and the whole process of answering the lender’s question is recorded in video. Later, the above-mentioned recorded video file is passed Processing, identifying the micro-expression of the lender when answering questions, in order to objectively score the performance of the lender’s answer.
- the stress question to the lender cannot be automatically generated by the robot according to the individual's own situation, and the evaluation of the micro-expression of the lender answering the question is evaluated by the overall performance of the lender in the entire video, not by the computer The evaluation is performed automatically according to the specific problems, and the accuracy of the evaluation results is not high.
- a face-to-face risk control method including:
- Identify the time node raised by each stress question in the video information and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding
- the image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
- the face-to-face review score value of the target object is generated according to each response information.
- a face-to-face audit risk control device including:
- Obtaining module configured to perform acquisition of video information of a target object responding to a stress question, wherein the stress question is a set of questions generated based on the collected personal information of the target object;
- Decomposition module configured to execute the time node proposed according to the identification of each stress problem in the video information, and divide the video information into several video segments according to the time node, where the video segment includes Image information that characterizes the change in expression of the target object when answering the corresponding stress question and audio information that characterizes the change in voice of the target object and the reply answer when answering the corresponding stress question;
- Extraction module configured to perform extraction of response information when the target object in each video segment replies to a corresponding stress question based on the image information and audio information;
- the scoring module is configured to execute the generation of the score value of the face-to-face review of the target object according to each response information.
- a computer device includes a memory and a processor, and computer-readable instructions are stored in the memory, and when the computer-readable instructions are executed by the processor, the processor executes the following steps:
- Identify the time node raised by each stress question in the video information and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding
- the image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
- the face-to-face review score value of the target object is generated according to each response information.
- a storage medium storing computer-readable instructions.
- the one or more processors execute the following steps:
- Identify the time node raised by each stress question in the video information and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding
- the image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
- the face-to-face review score value of the target object is generated according to each response information.
- This application eliminates the phenomenon of unfair scoring due to manual subjectivity, and the entire process is completed by a computer without manual operation, which saves labor costs.
- Figure 1 is the flow chart of the risk control method for face-to-face examination of the application
- Figure 2 is a flow chart of the scoring method for face-to-face review of the application
- Figure 3 is a flowchart of a method for generating a micro-expression feature set in this application
- Figure 4 is a flow chart of a method for generating face-to-face review scores based on multiple videos in this application;
- Figure 5 is a flow chart of the method for generating stress problems in this application.
- Figure 6 is a flowchart of a method for listing stress data to generate stress problems in this application
- FIG. 7 is a flow chart of the method for obtaining the video information of the target object responding to the stress problem in this application.
- Figure 8 is a schematic diagram of the application face-to-face risk control device module
- FIG. 9 is a block diagram of the basic structure of the computer equipment of this application.
- terminal and terminal equipment used herein include both wireless signal receiver equipment, which only has equipment with wireless signal receivers without transmitting capability, and also includes receiving and transmitting hardware equipment.
- Such equipment may include: cellular or other communication equipment, which has a single-line display or multi-line display or cellular or other communication equipment without a multi-line display; PCS (Personal Communications Service, personal communication system), which can combine voice and data Processing, fax and/or data communication capabilities; PDA (Personal Digital Assistant), which can include radio frequency receivers, pagers, Internet/Intranet access, web browsers, notebooks, calendars, and/or GPS (Global Positioning System (Global Positioning System) receiver; a conventional laptop and/or palmtop computer or other device that has and/or includes a radio frequency receiver.
- PCS Personal Communications Service, personal communication system
- PDA Personal Digital Assistant
- terminal and terminal equipment used here may be portable, transportable, installed in vehicles (aviation, sea and/or land), or suitable and/or configured to operate locally, and/or In a distributed form, it runs on the earth and/or any other location in space.
- the "terminal” and “terminal device” used here can also be communication terminals, Internet terminals, music/video playback terminals, such as PDA, MID (Mobile Internet Device, mobile Internet device) and/or music/video playback Functional mobile phones can also be devices such as smart TVs and set-top boxes.
- This application discloses a face-to-face risk control method, including:
- the target object is a person who needs to conduct a face-to-face review.
- the application scenario of this application is to use a computer to conduct a full face-to-face review of the target object.
- the stress question in the face-to-face interview process is generated by the personal information of the target object, and then interacts with the target object in the way of voice playback. While the target object responds to the stress question, the process of the target object answering the stress question is recorded. To collect real-time video information.
- the computer can identify each stress problem raised in the video information Time node, the time period between the time node of the previous stress problem and the time node of the next stress problem is the time period for the target object to reply to the previous stress problem, and the video of this time period can be intercepted.
- there are multiple stress questions so multiple video clips can be segmented, and each video clip includes image information that characterizes the change in expression of the target object in answering the stress question, and image information that characterizes the target object’s answer to the stress question. The voice change when stimulating the question and the audio message for replying to the answer.
- S3000 Based on the image information and audio information, extract response information when the target object in each video clip replies to a corresponding stress question;
- the above-mentioned segmented video segment includes image information that characterizes the change in expression of the target object and audio information that characterizes the change in voice of the target object and a reply answer.
- Different stress problems involve different aspects, the target object will have different reactions, and the target object’s response to different stress problems is of different importance to the face-to-face review, for example, the stress problem during the face-to-face loan review of the target object
- the target object’s actual age is 35 but deliberately said to be 32 years old, and the stress problem is the annual income
- the target object’s actual annual income is 20W, but it is said to be 250,000, if all are identified by micro-expression The target person lied.
- the video information is divided into stress problems, and scores are based on the performance of the target object under a single stress problem.
- the performance of the target user is the facial expressions and body expressions recognized by the image information and the audio information.
- the voice intonation and question answer are comprehensively reflected by multiple factors. Based on these factors, a comprehensive face-to-face review score value is calculated to make the face-to-face review score more objective.
- An application of the technical solution of this application is to conduct AI interviews with target objects, where the stress question is raised and the video recording is done by the computer itself, and the stress question of the target object is based on the collected target object’s individual
- the computer outputs the stress problem in the form of voice broadcast, and records the video information of the target responding to the stress problem.
- the computer decomposes the collected video information into several video clips according to the time node of the stress problem.
- the performance of the target object and the response information in the video clip are scored, and then the total face-to-face review score is calculated.
- the performance of the target object is scored according to different questions, which can evaluate the target object objectively and eliminate the artificial
- the subjectivity of scoring leads to the phenomenon of unfair scoring, and the whole process is completed by computer without manual operation, which saves labor costs.
- the response information includes a micro-expression feature set, a voice feature set, and a reply answer
- the extracted response information when the target object responds to a corresponding stress question in each video clip includes:
- S3100 Generate a micro-expression feature set that characterizes the video clip from the facial and body actions of the target object in the image information in the video clip;
- the response information is extracted by the target object's micro-expression and audio information to evaluate the question.
- the micro expressions are the facial expressions and body movements of the target object.
- the micro expressions that characterize the video clip are generated from the facial and body movements of the target object in the image information in the video clip.
- the feature set includes:
- S3120 Mapping the recognized micro-expression and the time point in the corresponding video segment to form a micro-expression feature set.
- the image information of the video clip is first input into the micro-expression recognition model.
- the micro-expression recognition model is a neural network model that is pre-trained to a convergent state and can recognize various expressions of the target object. Summarize the facial expressions of the human face, and map the changes in facial features when the human body produces the facial expressions to perform facial expression recognition.
- the change value of the facial features is the movement feature of the facial expression.
- the facial expression recognition is performed through the video image through the movement feature between the contour of the face and the facial features in each frame of the video connected first. The change is recognized. Specifically, the specific performance of the motion characteristics of the micro-expressions can be found in Table 1:
- Facial expression recognition based on neural network is mainly divided into three steps: face detection, expression feature extraction and expression classification.
- face detection is equivalent to detecting the face in the video or image to be processed and segmenting it from the image, thereby effectively reducing facial expression recognition
- Interference information that may appear in the process.
- Face detection is mainly based on the unique characteristics of the face, detecting whether there is a face in the image area to be tested, and comparing the detection possibility with a threshold, so as to locate the coordinate information of the face and segment the face. Location.
- the face area When the face area is determined, it is equivalent to reducing the detection area for facial expression recognition, and feature extraction of the facial expression information is performed on the facial area.
- the more popular expression feature extraction methods include principal component analysis (PCA), local binary mode (LBP), and some other feature extraction methods based on motion and deformation.
- the time axis information corresponding to the micro-expression is obtained, and the time-axis information is matched with the corresponding micro-expression to form a micro-expression feature set.
- the method of recognizing the specific content of voice information can also be processed by neural network model, for example, using Kaldi network model to extract features, output voice content in text, and recognize the voice, intonation, and speed of the target object when responding And amplitude information, etc., these voice information constitute a set of sound features of the target object in the video segment.
- the combination of voice feature set and micro-expression feature set can be used as a factor to evaluate the credit score of the video clip. For example, when the micro-expression is recognized as the face turns red and the eyes turn left and right unconsciously, and the voice feature at the same micro-expression time point is stuttering, it can be recognized that the current state of the target object is nervous and may be suspected of lying. Inputting the recognized micro-expression feature set and voice feature set into the scoring model, the corresponding credit score score can be obtained.
- the scoring model is a pre-trained model that scores based on stress questions and possible responses and expressions.
- S4000 Generate a face-to-face review score value of the target object according to each response information.
- step S3000 After each video clip passes step S3000, a corresponding response message is generated, and each response message is processed, and finally the face-to-face review score value of the target object can be generated.
- generating the face-to-face review score value of the target object according to each response information includes:
- the performance of the target object in the video segment is scored through a neural network model to obtain a credit score score.
- the credit score is the value of the probability that the target object may lie when answering the question.
- generating the stress question according to the collected personal information of the target object includes:
- S1200 Matching stress data in a preset stress database according to the personal information
- One method of generating stress problems is to generate corresponding stress problems according to preset rules through the collected personal information of the target object.
- Personal information includes, but is not limited to, personal identity information, historical behavior data, and historical credit data, etc.
- personal identity information includes age, marital status, occupation, income, etc.
- historical behavior data includes asset purchases, loan records, and historical credit data including Credit rating, etc.
- Different personal information corresponds to different stress data, and different stress data includes one or more different questions.
- Methods of listing to generate stress problems include:
- the stress data is matched from the stress database according to personal information, it can be responded to according to a preset rule.
- the questions in the stress data are listed.
- the preset rule is to list the questions according to the priority level of the stress data to generate stress questions. Therefore, in addition to mapping the association relationship between personal information and stress data in the preset stress database, the priority level of the stress data is also mapped.
- Table 2 As an example, please refer to Table 2:
- Table 2 List of the relationship between personal information and stress data
- the above multiple databases are combined to form a preset stress database.
- the corresponding databases are selected based on the personal identity information of the target object. Since there are multiple databases, each database contains multiple stress issues. Therefore, the lender can be asked in turn for each matching database.
- the stress data is a combination of multiple questions listed according to a preset association relationship, and the combination relationship is arranged according to the logical relationship between the questions. For example, in a married database, the first The question is whether there are children. If not, ask when you plan to have children; if so, ask the number of children, then ask how old each child is, and then ask the child’s monthly expenditure percentage, etc., in the database of the number of fixed assets , You can ask about the type of fixed assets, and then according to the identified type, look up the next question in the corresponding type database. For example, if the lender answers that the fixed asset is a real estate, then ask about the number of real estate, then ask about the location, and then ask about the real estate.
- Ask about the total area of the property ask about the purchase period of the real estate, ask whether the nature of the real estate is shared or exclusive, whether it is an apartment or a residence, and ask about the mortgage of the real estate.
- the questions in the same stress database are asked in the order in which they are arranged.
- the stress data obtained is a combination of multiple questions listed according to a preset association relationship
- the obtaining video information of the target object responding to the stress question includes:
- S1400 Acquire image information and audio information of the target object's response to the stress problem, and associate the image information and audio information in a time axis manner;
- S1500 Identify the reply answer corresponding to the target object in the audio information
- each question is arranged in accordance with the logical relationship between each other. Therefore, the method of obtaining the video information of the target object's response to the stress question is to first obtain the stress with the highest priority. Ask the first question in the data. At the same time, the image information and audio information of the target object's reply are obtained, and the audio information is recognized. After the reply answer is identified, the follow-up question combination in the stress data is selected according to the reply answer. The question is played by voice, and the video information of the target object's reply is obtained synchronously. When all the stressful questions are answered, the video shooting ends.
- This application discloses a face-to-face risk control device, including:
- Obtaining module 1000 configured to perform acquisition of video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
- Decomposition module 2000 configured to execute the time node proposed according to the identification of each stress problem in the video information, and divide the video information into several video segments according to the time node, wherein, in the video segment Including image information that characterizes the change in expression of the target object when answering the corresponding stress question and audio information that characterizes the change in voice of the target object and the answer to the answer when answering the corresponding stress question;
- Extraction module 3000 configured to perform, based on the image information and audio information, extract response information when the target object in each video clip responds to a corresponding stress question
- the scoring module 4000 is configured to execute the generation of the score value of the face-to-face review of the target object according to each response information.
- the response information includes a micro-expression feature set, a voice feature set, and a reply answer
- the extraction module includes:
- the first feature generation module configured to execute the generation of a micro-expression feature set based on the facial and body movements of the target object in the image information;
- the second feature generation module configured to perform extraction of a voice feature set characterizing the target object of the video clip and a reply answer from the audio information in the video clip, wherein the voice feature set includes voice, intonation, and speech rate And amplitude information;
- Response information generation module configured to perform mapping of the micro-expression feature set, the sound feature set and the response answer to each other in a time axis manner to obtain response information corresponding to the stress question.
- the scoring module includes:
- Credit score output module configured to input each of the response information into the scoring model in turn to output the credit score scores of the corresponding video clips in turn, where the scoring model is pre-trained to a convergent state, according to the application Models for scoring exciting questions and possible response answers, micro-expression characteristics and voice characteristics;
- Weight acquisition module configured to execute and acquire the scoring weight value of the stress problem corresponding to each video clip
- the weighted calculation module is configured to perform calculation and generate the face-to-face review score value in a weighted average manner according to the score weight value and the corresponding credit score score value.
- the first feature set generating module includes:
- Image input module configured to perform input of the image information into the micro-expression recognition model to output the micro-expression in the image information
- Mapping module configured to perform mutual mapping of the recognized micro-expression and the time point that appears in the video to form a micro-expression feature set.
- it also includes:
- Personal information acquisition module configured to perform acquisition of the personal information of the target object
- Matching module configured to perform matching of stress data in a preset stress database according to the personal information
- the first generating module is configured to perform listing of the stress data according to preset rules to generate stress questions.
- the first generating module includes:
- Priority acquisition module configured to execute the priority acquisition of the stress data
- the second generating module is configured to perform listing of the stress data according to the priority level to generate stress questions.
- the stress data is a combination of multiple questions listed according to a preset association relationship
- the acquisition module includes:
- Associating module configured to perform acquisition of image information and audio information of the target object's response to the stress problem, and to associate the image information and audio information in a time axis manner;
- Recognition module configured to perform recognition of the reply answer of the target object in the audio information
- Selection module configured to perform selection of a combination of questions in the stress data according to the reply answer to raise subsequent stress questions, and obtain video information synchronously.
- FIG. 9 Please refer to FIG. 9 for the basic structure block diagram of the computer equipment provided in the embodiment of the present application.
- the computer equipment includes a processor, a non-volatile storage medium, a memory, and a network interface connected through a system bus.
- the non-volatile storage medium of the computer device stores an operating system, a database, and computer-readable instructions.
- the database may store control information sequences.
- the processor can realize a A kind of face-to-face audit risk control method.
- the processor of the computer equipment is used to provide calculation and control capabilities, and supports the operation of the entire computer equipment.
- the computer readable instructions may be stored in the memory of the computer device, and when the computer readable instructions are executed by the processor, the processor may execute the following steps:
- Identify the time node raised by each stress question in the video information and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding
- the image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
- the face-to-face review score value of the target object is generated according to each response information.
- the network interface of the computer device is used to connect and communicate with the terminal.
- FIG. 9 is only a block diagram of a part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied.
- the specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
- the computer device receives the status information of the prompt behavior sent by the associated client, that is, whether the associated terminal opens the prompt and whether the user closes the prompt task.
- the corresponding preset instruction is sent to the associated terminal, so that the associated terminal can perform corresponding operations according to the preset instruction, thereby realizing effective supervision of the associated terminal.
- the server side controls the associated terminal to continue ringing, so as to prevent the prompt task of the associated terminal from automatically terminating after a period of execution.
- the present application also provides a storage medium storing computer-readable instructions.
- the computer-readable instructions When executed by one or more processors, the one or more processors perform the following steps:
- Identify the time node raised by each stress question in the video information and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding
- the image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
- the face-to-face review score value of the target object is generated according to each response information.
- the computer program can be stored in a computer readable storage medium. When executed, it may include the procedures of the above-mentioned method embodiments.
- the aforementioned storage medium can be a non-volatile storage medium such as a magnetic disk, an optical disc, a read-only memory (Read-Only Memory, ROM), or a volatile storage medium, or a random access memory (Random Access Memory). , RAM) etc.
- the face-to-face audit risk control method provided in this application further ensures the privacy and security of all the above-mentioned data
- all the above-mentioned data can also be stored in a node of a blockchain.
- video information and audio information, etc. these data can be stored in the blockchain node.
- the blockchain referred to in this application is a new application mode of computer technology such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm.
- Blockchain essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information for verification. The validity of the information (anti-counterfeiting) and the generation of the next block.
- the blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
Landscapes
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Human Resources & Organizations (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Development Economics (AREA)
- Economics (AREA)
- Health & Medical Sciences (AREA)
- Strategic Management (AREA)
- Theoretical Computer Science (AREA)
- General Business, Economics & Management (AREA)
- Entrepreneurship & Innovation (AREA)
- Marketing (AREA)
- General Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Finance (AREA)
- Accounting & Taxation (AREA)
- Educational Administration (AREA)
- Technology Law (AREA)
- Game Theory and Decision Science (AREA)
- Operations Research (AREA)
- Quality & Reliability (AREA)
- Tourism & Hospitality (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Child & Adolescent Psychology (AREA)
- Hospice & Palliative Care (AREA)
- Psychiatry (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Acoustics & Sound (AREA)
- Electrically Operated Instructional Devices (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
A face-to-face examination risk control method and apparatus, a computer device, and a storage medium. Said method comprises: acquiring video information of a target object replying to stress questions, the stress questions being a question set generated according to collected personal information of the target object (S1000); identifying, in the video information, a time node where each stress question is proposed, and segmenting the video information into several video segments according to the time nodes (S2000); on the basis of image information and audio information in the video segments, extracting from each video segment response information of the target object when replying to a corresponding stress question (S3000); and generating a face-to-face examination score value of the target object according to the response information (S4000). The performance of a target object is scored according to different questions, providing objective evaluation of a target object; the whole process needs no manual operation, reducing manpower costs.
Description
本申请要求于2020年3月26日提交中国专利局、申请号为CN202010225123.0,发明名称为“面审风险控制方法、装置、计算机设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed with the Chinese Patent Office on March 26, 2020, the application number is CN202010225123.0, and the invention title is "face-to-face risk control methods, devices, computer equipment and storage media", all of which The content is incorporated in this application by reference.
本申请涉及智能识别技术领域,具体而言,本申请涉及一种面审风险控制方法、装置、计算机设备及存储介质。This application relates to the field of intelligent identification technology. Specifically, this application relates to a face-to-face risk control method, device, computer equipment, and storage medium.
随着金融业的发展以及经济的发展,各项新兴业务和项目不断展开,很多企业或者个人在执行这些项目或业务,且需要用到资金时,会在银行等金融机构申请贷款以解决当前的公司或个人用钱的困境。银行等金融机构为了确保贷款被用在该用的合适的地方,且确保带出去的款项能够被收回来,会在贷款前对贷款人进行资格审查,资格审查包括历史征信信息审查,个人身份审查以及现场面审表现等。With the development of the financial industry and economic development, various emerging businesses and projects continue to unfold. When many companies or individuals implement these projects or businesses and need funds, they will apply for loans from banks and other financial institutions to solve the current problems. The dilemma of companies or individuals using money. In order to ensure that the loan is used in the appropriate place for the purpose, and to ensure that the money taken out can be recovered, the bank and other financial institutions will conduct a qualification review of the lender before the loan. The qualification review includes a review of historical credit information and personal identity. Review and on-site interview performance, etc.
发明人意识到在现有技术中,现场面审一般通过人工进行审核,面审员通过自己的经验对贷款者提交的审核资料进行提问,观察贷款人的表情和回答,来判断贷款人是否真实地回答所提出的问题,评价贷款人的诚实度。但是由于贷款人员众多,不用的面审员经验水平不一样,判断人的标准也不一样,且人工审核主观因素过多,不能保证贷款人评价标准的一致性,且人工审核,费时费力,占用过多的人力成本。为了减小贷款人评价的主观因素,通过对人工面审过程中进行全程视频录制,或者直接采用机器人进行面审,现有的机器面审中,对于面审时的问题,通常采用数据库随机抽取的方式,或者通过人工对贷款人进行背景调查后,人工制定应激问题,通过机器人将问题以语音播放的方式输出,对贷款人问题回答过程进行全程视频录制,后期将上述录制的视频文件通过处理,识别贷款人在回答问题时的微表情,以对贷款人回答的表现进行客观评分。但是采用这种方式,对贷款人的应激问题不能由机器人针对个人自身情况自动生成,且对贷款人回答问题的微表情的评价是通过整段视频贷款人的整体表现来评价,不能由计算机自动根据具体的问题分别进行评价,评价结果精确度不高。The inventor realizes that in the prior art, on-site interviews are generally conducted manually, and the interviewers use their own experience to ask questions about the review materials submitted by the lender, observe the expressions and answers of the lender to determine whether the lender is authentic Answer the questions raised and evaluate the honesty of the lender. However, due to the large number of loan personnel, different face-to-face reviewers have different experience levels, different criteria for judging people, and there are too many subjective factors in manual review, which cannot guarantee the consistency of the lender’s evaluation standards, and manual review is time-consuming and labor-intensive. Excessive labor costs. In order to reduce the subjective factors of the lender’s evaluation, the entire process of manual interviews is recorded by video recording, or the robots are used directly for the interviews. In the existing machine interviews, the problems during the interviews are usually extracted randomly from the database. After conducting a background check on the lender manually, the stress question is manually formulated, the question is output by voice through the robot, and the whole process of answering the lender’s question is recorded in video. Later, the above-mentioned recorded video file is passed Processing, identifying the micro-expression of the lender when answering questions, in order to objectively score the performance of the lender’s answer. However, in this way, the stress question to the lender cannot be automatically generated by the robot according to the individual's own situation, and the evaluation of the micro-expression of the lender answering the question is evaluated by the overall performance of the lender in the entire video, not by the computer The evaluation is performed automatically according to the specific problems, and the accuracy of the evaluation results is not high.
一种面审风险控制方法,包括:A face-to-face risk control method, including:
获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;
根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
一种面审风险控制装置,包括:A face-to-face audit risk control device, including:
获取模块:被配置为执行获取目标对象回复应激问题的视频信息,其中,所述应 激问题为根据收集的所述目标对象的个人信息生成的问题集合;Obtaining module: configured to perform acquisition of video information of a target object responding to a stress question, wherein the stress question is a set of questions generated based on the collected personal information of the target object;
分解模块:被配置为执行根据识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Decomposition module: configured to execute the time node proposed according to the identification of each stress problem in the video information, and divide the video information into several video segments according to the time node, where the video segment includes Image information that characterizes the change in expression of the target object when answering the corresponding stress question and audio information that characterizes the change in voice of the target object and the reply answer when answering the corresponding stress question;
提取模块:被配置为执行基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Extraction module: configured to perform extraction of response information when the target object in each video segment replies to a corresponding stress question based on the image information and audio information;
评分模块:被配置为执行根据各回应信息生成所述目标对象的面审评分值。The scoring module is configured to execute the generation of the score value of the face-to-face review of the target object according to each response information.
一种计算机设备,包括存储器和处理器,所述存储器中存储有计算机可读指令,所述计算机可读指令被所述处理器执行时,使得所述处理器执行以下步骤:A computer device includes a memory and a processor, and computer-readable instructions are stored in the memory, and when the computer-readable instructions are executed by the processor, the processor executes the following steps:
获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;
根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
一种存储有计算机可读指令的存储介质,所述计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行以下步骤:A storage medium storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors execute the following steps:
获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;
根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
本申请消除了因人工的主观性导致评分不公平的现象,且整个过程通过计算机完成,无需人工操作,节省了人力成本。This application eliminates the phenomenon of unfair scoring due to manual subjectivity, and the entire process is completed by a computer without manual operation, which saves labor costs.
本申请上述的和/或附加的方面和优点从下面结合附图对实施例的描述中将变得明显和容易理解,其中:The above and/or additional aspects and advantages of the present application will become obvious and easy to understand from the following description of the embodiments in conjunction with the accompanying drawings, in which:
图1为本申请面审风险控制方法流程图;Figure 1 is the flow chart of the risk control method for face-to-face examination of the application;
图2为本申请面审评分方法流程图;Figure 2 is a flow chart of the scoring method for face-to-face review of the application;
图3为本申请生成微表情特征集合的方法流程图;Figure 3 is a flowchart of a method for generating a micro-expression feature set in this application;
图4为本申请根据多个视频生成面审评分值的方法流程图;Figure 4 is a flow chart of a method for generating face-to-face review scores based on multiple videos in this application;
图5为本申请应激问题生成方法流程图;Figure 5 is a flow chart of the method for generating stress problems in this application;
图6为本申请将应激数据罗列生成应激问题的方法流程图;Figure 6 is a flowchart of a method for listing stress data to generate stress problems in this application;
图7为本申请获取目标对象回复应激问题的视频信息的方法流程图;FIG. 7 is a flow chart of the method for obtaining the video information of the target object responding to the stress problem in this application;
图8为本申请面审风险控制装置模块示意图;Figure 8 is a schematic diagram of the application face-to-face risk control device module;
图9为本申请计算机设备基本结构框图。Figure 9 is a block diagram of the basic structure of the computer equipment of this application.
下面详细描述本申请的实施例,所述实施例的示例在附图中示出,其中自始至终相同或类似的标号表示相同或类似的元件或具有相同或类似功能的元件。下面通过参考附图描述的实施例是示例性的,仅用于解释本申请,而不能解释为对本申请的限制。The embodiments of the present application are described in detail below. Examples of the embodiments are shown in the accompanying drawings, wherein the same or similar reference numerals indicate the same or similar elements or elements with the same or similar functions. The embodiments described below with reference to the drawings are exemplary, and are only used to explain the present application, and cannot be construed as a limitation to the present application.
本技术领域技术人员可以理解,除非特意声明,这里使用的单数形式“一”、“一个”、“所述”和“该”也可包括复数形式。应该进一步理解的是,本申请的说明书中使用的措辞“包括”是指存在所述特征、整数、步骤、操作、元件和/或组件,但是并不排除存在或添加一个或多个其他特征、整数、步骤、操作、元件、组件和/或它们的组。应该理解,当我们称元件被“连接”或“耦接”到另一元件时,它可以直接连接或耦接到其他元件,或者也可以存在中间元件。此外,这里使用的“连接”或“耦接”可以包括无线连接或无线耦接。这里使用的措辞“和/或”包括一个或更多个相关联的列出项的全部或任一单元和全部组合。Those skilled in the art can understand that, unless specifically stated, the singular forms "a", "an", "said" and "the" used herein may also include plural forms. It should be further understood that the term "comprising" used in the specification of this application refers to the presence of the described features, integers, steps, operations, elements and/or components, but does not exclude the presence or addition of one or more other features, Integers, steps, operations, elements, components, and/or groups of them. It should be understood that when we refer to an element as being "connected" or "coupled" to another element, it can be directly connected or coupled to the other element, or intervening elements may also be present. In addition, “connected” or “coupled” used herein may include wireless connection or wireless coupling. The term "and/or" as used herein includes all or any unit and all combinations of one or more associated listed items.
本技术领域技术人员可以理解,除非另外定义,这里使用的所有术语(包括技术术语和科学术语),具有与本申请所属领域中的普通技术人员的一般理解相同的意义。还应该理解的是,诸如通用字典中定义的那些术语,应该被理解为具有与现有技术的上下文中的意义一致的意义,并且除非像这里一样被特定定义,否则不会用理想化或过于正式的含义来解释。Those skilled in the art can understand that, unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meanings as those commonly understood by those of ordinary skill in the art to which this application belongs. It should also be understood that terms such as those defined in general dictionaries should be understood to have a meaning consistent with the meaning in the context of the prior art, and unless specifically defined as here, they will not be idealized or overly Explain the formal meaning.
本技术领域技术人员可以理解,这里所使用的“终端”、“终端设备”既包括无线信号接收器的设备,其仅具备无发射能力的无线信号接收器的设备,又包括接收和发射硬件的设备,其具有能够在双向通信链路上,执行双向通信的接收和发射硬件的设备。这种设备可以包括:蜂窝或其他通信设备,其具有单线路显示器或多线路显示器或没有多线路显示器的蜂窝或其他通信设备;PCS(Personal Communications Service,个人通信系统),其可以组合语音、数据处理、传真和/或数据通信能力;PDA(Personal Digital Assistant,个人数字助理),其可以包括射频接收器、寻呼机、互联网/内联网访问、网络浏览器、记事本、日历和/或GPS(Global Positioning System,全球定位系统)接收器;常规膝上型和/或掌上型计算机或其他设备,其具有和/或包括射频接收器的常规膝上型和/或掌上型计算机或其他设备。这里所使用的“终端”、“终端设备”可以是便携式、可运输、安装在交通工具(航空、海运和/或陆地)中的,或者适合于和/或配置为在本地运行,和/或以分布形式,运行在地球和/或空间的任何其他位置运行。这里所使用的“终端”、“终端设备”还可以是通信终端、上网终端、音乐/视频播放终端,例如可以是PDA、MID(Mobile Internet Device,移动互联网设备)和/或具有音乐/视频播放功能的移动电话,也可以是智能电视、机顶盒等设备。Those skilled in the art can understand that the "terminal" and "terminal equipment" used herein include both wireless signal receiver equipment, which only has equipment with wireless signal receivers without transmitting capability, and also includes receiving and transmitting hardware equipment. A device that has a device capable of performing two-way communication receiving and transmitting hardware on a two-way communication link. Such equipment may include: cellular or other communication equipment, which has a single-line display or multi-line display or cellular or other communication equipment without a multi-line display; PCS (Personal Communications Service, personal communication system), which can combine voice and data Processing, fax and/or data communication capabilities; PDA (Personal Digital Assistant), which can include radio frequency receivers, pagers, Internet/Intranet access, web browsers, notebooks, calendars, and/or GPS (Global Positioning System (Global Positioning System) receiver; a conventional laptop and/or palmtop computer or other device that has and/or includes a radio frequency receiver. The "terminal" and "terminal equipment" used here may be portable, transportable, installed in vehicles (aviation, sea and/or land), or suitable and/or configured to operate locally, and/or In a distributed form, it runs on the earth and/or any other location in space. The "terminal" and "terminal device" used here can also be communication terminals, Internet terminals, music/video playback terminals, such as PDA, MID (Mobile Internet Device, mobile Internet device) and/or music/video playback Functional mobile phones can also be devices such as smart TVs and set-top boxes.
具体的,请参阅图1,本申请公开了一种面审风险控制方法,包括:Specifically, please refer to Figure 1. This application discloses a face-to-face risk control method, including:
S1000、获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;S1000. Obtain video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
目标对象为需要进行面审的人员,本申请的应用场景为采用计算机对目标对象进行全程面审。面审过程中的应激问题通过目标对象的个人信息生成,然后以语音播放的方式与目标对象进行交互,在目标对象回复应激问题的同时,对目标对象回答应激问题的过程进行录制,以采集实时的视频信息。The target object is a person who needs to conduct a face-to-face review. The application scenario of this application is to use a computer to conduct a full face-to-face review of the target object. The stress question in the face-to-face interview process is generated by the personal information of the target object, and then interacts with the target object in the way of voice playback. While the target object responds to the stress question, the process of the target object answering the stress question is recorded. To collect real-time video information.
S2000、识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;S2000. Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes a character that the target object is Image information of facial expression changes when answering the corresponding stress question and audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
由于应激问题是通过计算机以语音播放方式播放的,且在整个面审过程中都进行 有视频录制,视频录制以时间轴方式采集,因此计算机可以在视频信息识别出每个应激问题提出的时间节点,前一个应激问题的时间节点与后一个应激问题的时间节点之间的时间段则为目标对象回复该前一应激问题的时间段,将该时间段的视频进行截取则可获得针对该应激问题的视频片段。在一实施例中,有多个应激问题,因此则可分割出多个视频片段,每个视频片段中包括表征目标对象回答该应激问题的表情变化的图像信息和表征目标对象回答该应激问题时的声音变化及回复答案的音频信息。Since the stress problem is played through the computer by voice playback, and there is video recording during the entire interview process, and the video recording is collected in a timeline manner, the computer can identify each stress problem raised in the video information Time node, the time period between the time node of the previous stress problem and the time node of the next stress problem is the time period for the target object to reply to the previous stress problem, and the video of this time period can be intercepted. Obtain a video clip for the stress problem. In one embodiment, there are multiple stress questions, so multiple video clips can be segmented, and each video clip includes image information that characterizes the change in expression of the target object in answering the stress question, and image information that characterizes the target object’s answer to the stress question. The voice change when stimulating the question and the audio message for replying to the answer.
S3000、基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;S3000: Based on the image information and audio information, extract response information when the target object in each video clip replies to a corresponding stress question;
上述分割的视频片段为包括表征所述目标对象表情变化的图像信息和表征所述目标对象声音变化及回复答案的音频信息。不同的应激问题涉及不同的方面,目标对象会有不同的反应,且目标对象对不同应激问题的反应对面审的重要程度不同,例如在对目标对象进行贷款面审过程中,应激问题为年龄多大,若目标对象实际年龄为35岁,而却故意说成32岁,以及应激问题为年收入多少,目标对象实际年收入为20W,却说成25万,若都通过微表情识别出目标对象说谎,在对整个面审过程进行评分时,年收入问题说谎的严重性要高于年龄说谎的严重性,因此在进行最终面审评分时比重越大。基于这一原因,将视频信息按照应激问题进行拆分,根据目标对象在单个应激问题下的表现进行评分,而目标用户的表现是通过图像信息识别的面部表情和肢体表情以及音频信息中的语音语调和问题答复多个因素综合体现出来的,根据这些因素再计算出综合的面审评分值,使面审评分值更客观。The above-mentioned segmented video segment includes image information that characterizes the change in expression of the target object and audio information that characterizes the change in voice of the target object and a reply answer. Different stress problems involve different aspects, the target object will have different reactions, and the target object’s response to different stress problems is of different importance to the face-to-face review, for example, the stress problem during the face-to-face loan review of the target object For the age, if the target object’s actual age is 35 but deliberately said to be 32 years old, and the stress problem is the annual income, the target object’s actual annual income is 20W, but it is said to be 250,000, if all are identified by micro-expression The target person lied. When scoring the entire interview process, the severity of annual income lie is higher than the severity of age lie, so the greater the proportion in the final face-to-face scoring. For this reason, the video information is divided into stress problems, and scores are based on the performance of the target object under a single stress problem. The performance of the target user is the facial expressions and body expressions recognized by the image information and the audio information. The voice intonation and question answer are comprehensively reflected by multiple factors. Based on these factors, a comprehensive face-to-face review score value is calculated to make the face-to-face review score more objective.
本申请的技术方案的一种应用在于对目标对象进行AI面试,其中应激问题的提出和视频的录制都是计算机自己完成,所述目标对象的应激问题为根据所收集的目标对象的个人信息而生成,计算机以语音播报的方式输出应激问题,并记录目标对象回复应激问题的视频信息,计算机根据应激问题的时间节点,将采集的视频信息分解成若干个视频片段,对单个视频片段中目标对象的表现和回复信息进行评分,然后再计算出总的面审评分,根据不同的问题对目标对象的表现进行打分,更能客观地对目标对象进行评价,消除了因人工的主观性导致评分不公平的现象,且整个过程通过计算机完成,无需人工操作,节省了人力成本。An application of the technical solution of this application is to conduct AI interviews with target objects, where the stress question is raised and the video recording is done by the computer itself, and the stress question of the target object is based on the collected target object’s individual The computer outputs the stress problem in the form of voice broadcast, and records the video information of the target responding to the stress problem. The computer decomposes the collected video information into several video clips according to the time node of the stress problem. The performance of the target object and the response information in the video clip are scored, and then the total face-to-face review score is calculated. The performance of the target object is scored according to different questions, which can evaluate the target object objectively and eliminate the artificial The subjectivity of scoring leads to the phenomenon of unfair scoring, and the whole process is completed by computer without manual operation, which saves labor costs.
在一实施例中,请参阅图2,所述回应信息包括微表情特征集合、声音特征集合和回复答案,所述提取各视频片段中所述目标对象回复对应应激问题时的回应信息包括:In one embodiment, referring to FIG. 2, the response information includes a micro-expression feature set, a voice feature set, and a reply answer, and the extracted response information when the target object responds to a corresponding stress question in each video clip includes:
S3100、通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合;S3100: Generate a micro-expression feature set that characterizes the video clip from the facial and body actions of the target object in the image information in the video clip;
S3200、通过视频片段中的所述音频信息提取表征该视频片段所述目标对象的声音特征集合和回复答案,其中,所述声音特征集合包括语音、语调、语速和振幅信息;S3200. Extract a voice feature set that characterizes the target object of the video clip and a reply answer from the audio information in the video clip, where the voice feature set includes voice, intonation, speech rate, and amplitude information;
S3300、将所述微表情特征集合与所述声音特征集合和回复答案以时间轴的方式相互映射以得到对应应激问题的回应信息。S3300. Mapping the micro-expression feature set, the sound feature set and the reply answer in a time axis manner to obtain response information corresponding to the stress question.
在一实施例中,通过目标对象回答问题的微表情以及音频信息提取回复信息进行评价。微表情为目标对象的面部表情和肢体动作,在一实施例中,请参阅图3,所述通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合包括:In one embodiment, the response information is extracted by the target object's micro-expression and audio information to evaluate the question. The micro expressions are the facial expressions and body movements of the target object. In one embodiment, please refer to FIG. 3. The micro expressions that characterize the video clip are generated from the facial and body movements of the target object in the image information in the video clip. The feature set includes:
S3110、将所述图像信息输入至微表情识别模型中以输出所述图像信息中的微表情;S3110. Input the image information into a micro-expression recognition model to output the micro-expression in the image information;
S3120、将所识别的微表情与对应视频片段中出现的时间点相互映射形成微表情特征集合。S3120: Mapping the recognized micro-expression and the time point in the corresponding video segment to form a micro-expression feature set.
在单独进行微表情识别过程中,先将视频片段的图像信息输入至微表情识别模型中,微表情识别模型为预先训练至收敛状态,可以识别出目标对象各种表情的神经网 络模型,其通过对人脸的表情进行归纳,并映射人体产生该表情时面部五官的变化以进行表情识别。面部五官的变化值则为该表情的运动特征,在本实施例中,通过视频图像进行表情识别,是通过视频中先连的每一帧画面中,人脸的轮廓与五官之间的运动特征变化识别的。具体的,微表情的运动特征的具体表现可参阅表1:In the process of individual micro-expression recognition, the image information of the video clip is first input into the micro-expression recognition model. The micro-expression recognition model is a neural network model that is pre-trained to a convergent state and can recognize various expressions of the target object. Summarize the facial expressions of the human face, and map the changes in facial features when the human body produces the facial expressions to perform facial expression recognition. The change value of the facial features is the movement feature of the facial expression. In this embodiment, the facial expression recognition is performed through the video image through the movement feature between the contour of the face and the facial features in each frame of the video connected first. The change is recognized. Specifically, the specific performance of the motion characteristics of the micro-expressions can be found in Table 1:
表1:微表情的运动特征具体表现Table 1: Specific performance of micro-expressions
基于神经网络的人脸表情识别主要分为三步:即人脸检测、表情特征提取以及表情分类。通常情况下,进行人脸表情识别的图像中存在其他环境的干扰信息,因此人脸检测相当于在待处理的视频或图像中将人脸进行检测并从图像中进行分割,从而有效降低表情识别过程中可能出现的干扰信息。人脸检测主要是根据人脸的特有特征,在待测的图像区域中检测是否存在人脸,并将检测的可能性与阈值进行比较,从而定位人脸的坐标信息并分割出人脸所在的位置。Facial expression recognition based on neural network is mainly divided into three steps: face detection, expression feature extraction and expression classification. Normally, there is interference information from other environments in the image for facial expression recognition, so face detection is equivalent to detecting the face in the video or image to be processed and segmenting it from the image, thereby effectively reducing facial expression recognition Interference information that may appear in the process. Face detection is mainly based on the unique characteristics of the face, detecting whether there is a face in the image area to be tested, and comparing the detection possibility with a threshold, so as to locate the coordinate information of the face and segment the face. Location.
当确定人脸区域后,相当于缩小了表情识别的检测区域,在人脸区域上对表情信息进行特征提取。目前较为流行的表情特征提取法由主成分分析法(PCA)、局部二值模式(LBP)以及一些其他基于运动和形变的特征提取方法。When the face area is determined, it is equivalent to reducing the detection area for facial expression recognition, and feature extraction of the facial expression information is performed on the facial area. At present, the more popular expression feature extraction methods include principal component analysis (PCA), local binary mode (LBP), and some other feature extraction methods based on motion and deformation.
在该阶段需要针对待求问题选择合适的分类器训练得到分类准确、泛化能力强的分类器,并将需要检测的图像作为输入,经过人脸检测、提取表情特征以及分类器分 类得到人脸表情的所属类别。At this stage, it is necessary to select a suitable classifier for the problem to be asked for training to obtain a classifier with accurate classification and strong generalization ability, and use the image to be detected as input, and then face detection, expression feature extraction and classifier classification to obtain the face The category of the emoticon.
当通过上述神经网络模型识别出视频中相关联的帧画面中的微表情后,获取产生该微表情对应的时间轴信息,将时间轴信息与对应的微表情匹配形成微表情特征集合。After the micro-expression in the associated frame of the video is identified through the neural network model, the time axis information corresponding to the micro-expression is obtained, and the time-axis information is matched with the corresponding micro-expression to form a micro-expression feature set.
识别语音信息的具体内容的方法也可采用神经网络模型进行处理,例如,使用kaldi(卡尔迪)网络模型提取特征,以文字方式输出语音内容,同时识别目标对象回复时的语音、语调、语速和振幅信息等,这些语音信息构成了目标对象在该视频片段中的声音特征集合。声音特征集合与微表情特征集合结合起来可作为评价该段视频片段的信用评分分值的因素。例如,当微表情识别为脸变红,眼睛不自觉左右转动,而同一微表情时间点上的声音特征为结巴,则可识别出该目标对象当前的状态为紧张,可能有说谎的嫌疑,通过将识别的微表情特征集合与声音特征集合输入评分模型中,则可得到对应的信用评分分值。在一实施例中,所述评分模型为预先训练好的,根据应激问题及其可能的回复和表情进行评分的模型。The method of recognizing the specific content of voice information can also be processed by neural network model, for example, using Kaldi network model to extract features, output voice content in text, and recognize the voice, intonation, and speed of the target object when responding And amplitude information, etc., these voice information constitute a set of sound features of the target object in the video segment. The combination of voice feature set and micro-expression feature set can be used as a factor to evaluate the credit score of the video clip. For example, when the micro-expression is recognized as the face turns red and the eyes turn left and right unconsciously, and the voice feature at the same micro-expression time point is stuttering, it can be recognized that the current state of the target object is nervous and may be suspected of lying. Inputting the recognized micro-expression feature set and voice feature set into the scoring model, the corresponding credit score score can be obtained. In one embodiment, the scoring model is a pre-trained model that scores based on stress questions and possible responses and expressions.
S4000、根据各回应信息生成所述目标对象的面审评分值。S4000: Generate a face-to-face review score value of the target object according to each response information.
每一个视频片段都通过步骤S3000后会生成一个对应的回应信息,对每个回应信息进行处理,最后即可生成目标对象的面审评分值。After each video clip passes step S3000, a corresponding response message is generated, and each response message is processed, and finally the face-to-face review score value of the target object can be generated.
在一实施例中,请参阅图4,所述根据各回应信息生成所述目标对象的面审评分值包括:In an embodiment, referring to FIG. 4, generating the face-to-face review score value of the target object according to each response information includes:
S4100、将各个所述回应信息依次输入评分模型中以依次输出对应的视频片段的信用评分分值,其中,所述评分模型为预先训练至收敛状态,根据应激问题及其可能的回复答案、微表情特征和声音特征进行评分的模型;S4100. Input each of the response information into a scoring model in turn to output the credit scores of the corresponding video clips in turn, where the scoring model is pre-trained to a convergent state, and according to the stress question and its possible response answers, A model for scoring micro-expression features and voice features;
S4200、获取各个视频片段对应的应激问题的评分权重值;S4200: Obtain the scoring weight value of the stress problem corresponding to each video clip;
S4300、根据所述评分权重值与对应的所述信用评分分值采用加权平均的方式计算生成面审评分值。S4300: According to the scoring weight value and the corresponding credit scoring value, a weighted average method is used to calculate and generate a face-to-face review score value.
在一实施例中,将视频信息拆解成单个的视频片段后,通过神经网络模型来对该视频片段中的目标对象的表现进行打分,得到信用评分分值。信用评分分值为表针目标对象在回答该问题时可能说谎的概率值。In one embodiment, after the video information is disassembled into a single video segment, the performance of the target object in the video segment is scored through a neural network model to obtain a credit score score. The credit score is the value of the probability that the target object may lie when answering the question.
由于上述公开了不同的应激问题对面审结果的重要程度不同,通过对不同的应激问题制定不同的评分权重值以凸显在某些问题上进行欺诈的严重性。计算机获取了对应视频片段的信用评分分值后,再获取该应激问题的评分权重值,将所有的信用评分分值与对应的评分权重值相乘,再加权平均则可得到最后的面审评分值。例如应激问题包括A、B、C三个,对应的评那份权重值分别为0.3,0.6和0.1,应激问题A、B、C对应的信用评分分值分别为80,60,50,则面审评分值=(0.3*80+0.6*60+0.1*50)/(0.3+0.6+0.1)=65,最终的面审评分值为65分。Since the above-mentioned disclosures of different stress issues have different importance to the face-to-face audit results, different scoring weight values are formulated for different stress issues to highlight the seriousness of fraud on certain issues. After the computer obtains the credit scoring value of the corresponding video clip, it obtains the scoring weight value of the stress problem, multiplying all the credit scoring scores by the corresponding scoring weight value, and then weighting the average to get the final face-to-face review The score value. For example, there are three stress problems including A, B, and C, and the corresponding evaluation weights are 0.3, 0.6, and 0.1 respectively. The corresponding credit scores of stress problems A, B, and C are 80, 60, and 50, respectively. Then the score value of face-to-face review=(0.3*80+0.6*60+0.1*50)/(0.3+0.6+0.1)=65, and the final score value of face-to-face review is 65 points.
在一实施例中,请参阅图5,根据收集的所述目标对象的个人信息生成所述应激问题包括:In one embodiment, referring to FIG. 5, generating the stress question according to the collected personal information of the target object includes:
S1100、获取所述目标对象的个人信息;S1100. Acquire personal information of the target object;
S1200、根据所述个人信息在预设应激数据库中匹配应激数据;S1200: Matching stress data in a preset stress database according to the personal information;
S1300、根据预设规则将所述应激数据进行罗列以生成应激问题。S1300. List the stress data according to a preset rule to generate a stress problem.
在一实施例中,在面审过程中,不同的目标对象具有不同的应激问题,以提高面审的可参考性。一种生成应激问题的方法为通过所收集的目标对象的个人信息,按照预设的规则生成对应的应激问题。个人信息包括但是不局限于个人身份信息、历史行为数据和历史征信数据等,个人身份信息包括年龄、婚姻状态、职业、收入等,历史行为数据包括资产购买、贷款记录、历史征信数据包括征信级别等。不同的个人信息对应有不同的应激数据,不同的应激数据中包括一个或多个不同的问题,在一实施例中,请参阅图6,所述根据预设规则将所述应激数据进行罗列以生成应激问题的方法 包括:In one embodiment, during the face-to-face interview process, different target objects have different stress problems, so as to improve the referenceability of the face-to-face interview. One method of generating stress problems is to generate corresponding stress problems according to preset rules through the collected personal information of the target object. Personal information includes, but is not limited to, personal identity information, historical behavior data, and historical credit data, etc. Personal identity information includes age, marital status, occupation, income, etc., historical behavior data includes asset purchases, loan records, and historical credit data including Credit rating, etc. Different personal information corresponds to different stress data, and different stress data includes one or more different questions. In one embodiment, please refer to FIG. Methods of listing to generate stress problems include:
S1310、获取所述应激数据的优先级别;S1310. Obtain the priority level of the stress data;
S1320、根据所述优先级别对所述应激数据进行罗列以生成应激问题。S1320. List the stress data according to the priority level to generate a stress problem.
为了有次序地进行应激问题提问,需要对获取的应激数据进行排列,由于应激数据是根据个人信息从应激数据库中匹配的数据,因此,可根据一种预设的规则对该应激数据中的问题进行罗列,在一实施例中,这种预设规则为根据应激数据的优先级别对问题进行罗列生成应激问题。因此在预设的应激数据库中除了映射个人信息与应激数据之间的关联关系外,还映射该应激数据的优先级别,作为一示例,请参阅表2:In order to ask questions about stress in an orderly manner, it is necessary to arrange the obtained stress data. Since the stress data is matched from the stress database according to personal information, it can be responded to according to a preset rule. The questions in the stress data are listed. In one embodiment, the preset rule is to list the questions according to the priority level of the stress data to generate stress questions. Therefore, in addition to mapping the association relationship between personal information and stress data in the preset stress database, the priority level of the stress data is also mapped. As an example, please refer to Table 2:
表2:个人信息与应激数据关系列表Table 2: List of the relationship between personal information and stress data
上述多个数据库组合起来构成预设应激数据库,在一实施例中,分别根据目标对象的个人身份信息选取出对应的数据库,由于有多个数据库,每个数据库中都包含多个应激问题,因此,可对贷款人针对每一个匹配的数据库进行依次提问。The above multiple databases are combined to form a preset stress database. In one embodiment, the corresponding databases are selected based on the personal identity information of the target object. Since there are multiple databases, each database contains multiple stress issues. Therefore, the lender can be asked in turn for each matching database.
在一实施例中,所述应激数据为按照预设的关联关系罗列的多个问题的组合,组合的关系为根据问题之间的逻辑关系进行排列,例如,在已婚数据库中,第一个问题为是否有小孩,若无,则问打算什么时候要小孩,若有,则问孩子数量,再问每个孩子多大,接着问孩子每月开销占比等,在固定资产数量的数据库中,可想问固定资产的类型,再根据所识别的类型,在对应的类型数据库中查找下一步的问题,例如贷款人回答固定资产为房产,则问房产数量,再问的位置,再问房产的总面积,再问房产的购买年限,再问房产的性质是共有还是独占,是公寓还是住宅,再问房产的抵押情况等。在进行提问过程中,同一应激数据库中的问题按照排列的先后顺序进行提问。In one embodiment, the stress data is a combination of multiple questions listed according to a preset association relationship, and the combination relationship is arranged according to the logical relationship between the questions. For example, in a married database, the first The question is whether there are children. If not, ask when you plan to have children; if so, ask the number of children, then ask how old each child is, and then ask the child’s monthly expenditure percentage, etc., in the database of the number of fixed assets , You can ask about the type of fixed assets, and then according to the identified type, look up the next question in the corresponding type database. For example, if the lender answers that the fixed asset is a real estate, then ask about the number of real estate, then ask about the location, and then ask about the real estate. Ask about the total area of the property, ask about the purchase period of the real estate, ask whether the nature of the real estate is shared or exclusive, whether it is an apartment or a residence, and ask about the mortgage of the real estate. In the process of questioning, the questions in the same stress database are asked in the order in which they are arranged.
在一实施例中,请参阅图7,获所述应激数据为按照预设的关联关系罗列的多个问题的组合,所述获取目标对象回复应激问题的视频信息包括:In one embodiment, referring to FIG. 7, the stress data obtained is a combination of multiple questions listed according to a preset association relationship, and the obtaining video information of the target object responding to the stress question includes:
S1400、获取所述目标对象回复所述应激问题的图像信息和音频信息,并将所述图像信息和音频信息以时间轴方式进行关联;S1400: Acquire image information and audio information of the target object's response to the stress problem, and associate the image information and audio information in a time axis manner;
S1500、识别所述音频信息中所述目标对象对应的回复答案;S1500: Identify the reply answer corresponding to the target object in the audio information;
S1600、根据所述回复答案选择所述应激数据中的问题组合以进行后续应激的问题提出,并同步获取视频信息。S1600. Select a combination of questions in the stress data according to the reply answer to raise subsequent stress questions, and obtain video information synchronously.
由于上述公开的应激数据中包括多个问题,各个问题按照相互之间的逻辑关系进行排列,因此,在获取目标对象回复应激问题的视频信息的方法为,先获取优先级别最高的应激数据中的第一个问题进行提问,同时获取目标对象回复时的图像信息和音频信息,对音频信息进行识别,识别了回复答案后,根据回复答案在应激数据中选择其问题组合中的后续问题进行语音播放,并同步获取目标对象回复的视频信息,当所有的应激问题都回复完毕,视频拍摄结束。Since the above-disclosed stress data includes multiple questions, each question is arranged in accordance with the logical relationship between each other. Therefore, the method of obtaining the video information of the target object's response to the stress question is to first obtain the stress with the highest priority. Ask the first question in the data. At the same time, the image information and audio information of the target object's reply are obtained, and the audio information is recognized. After the reply answer is identified, the follow-up question combination in the stress data is selected according to the reply answer. The question is played by voice, and the video information of the target object's reply is obtained synchronously. When all the stressful questions are answered, the video shooting ends.
另一方面,请参阅图8,本申请公开一种面审风险控制装置,包括:On the other hand, please refer to Figure 8. This application discloses a face-to-face risk control device, including:
获取模块1000:被配置为执行获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Obtaining module 1000: configured to perform acquisition of video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
分解模块2000:被配置为执行根据识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Decomposition module 2000: configured to execute the time node proposed according to the identification of each stress problem in the video information, and divide the video information into several video segments according to the time node, wherein, in the video segment Including image information that characterizes the change in expression of the target object when answering the corresponding stress question and audio information that characterizes the change in voice of the target object and the answer to the answer when answering the corresponding stress question;
提取模块3000:被配置为执行基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息Extraction module 3000: configured to perform, based on the image information and audio information, extract response information when the target object in each video clip responds to a corresponding stress question
评分模块4000:被配置为执行根据各回应信息生成所述目标对象的面审评分值。The scoring module 4000 is configured to execute the generation of the score value of the face-to-face review of the target object according to each response information.
可选的,所述回应信息包括微表情特征集合、声音特征集合和回复答案,所述提取模块包括:Optionally, the response information includes a micro-expression feature set, a voice feature set, and a reply answer, and the extraction module includes:
第一特征生成模块:被配置为执行通过所述图像信息中目标对象的面部和肢体动作生成微表情特征集合;The first feature generation module: configured to execute the generation of a micro-expression feature set based on the facial and body movements of the target object in the image information;
第二特征生成模块:被配置为执行通过视频片段中的所述音频信息提取表征该视频片段所述目标对象的声音特征集合和回复答案,其中,所述声音特征集合包括语音、语调、语速和振幅信息;The second feature generation module: configured to perform extraction of a voice feature set characterizing the target object of the video clip and a reply answer from the audio information in the video clip, wherein the voice feature set includes voice, intonation, and speech rate And amplitude information;
回应信息生成模块:被配置为执行将所述微表情特征集合与所述声音特征集合和回复答案以时间轴的方式相互映射以得到对应应激问题的回应信息。Response information generation module: configured to perform mapping of the micro-expression feature set, the sound feature set and the response answer to each other in a time axis manner to obtain response information corresponding to the stress question.
可选的,所述评分模块包括:Optionally, the scoring module includes:
信用评分分值输出模块:被配置为执行将各个所述回应信息依次输入评分模型中以依次输出对应的视频片段的信用评分分值,其中,所述评分模型为预先训练至收敛状态,根据应激问题及其可能的回复答案、微表情特征和声音特征进行评分的模型;Credit score output module: configured to input each of the response information into the scoring model in turn to output the credit score scores of the corresponding video clips in turn, where the scoring model is pre-trained to a convergent state, according to the application Models for scoring exciting questions and possible response answers, micro-expression characteristics and voice characteristics;
权重获取模块:被配置为执行获取各个视频片段对应的应激问题的评分权重值;Weight acquisition module: configured to execute and acquire the scoring weight value of the stress problem corresponding to each video clip;
加权计算模块:被配置为执行根据所述评分权重值与对应的所述信用评分分值采用加权平均的方式计算生成面审评分值。The weighted calculation module is configured to perform calculation and generate the face-to-face review score value in a weighted average manner according to the score weight value and the corresponding credit score score value.
可选的,所述第一特征集生成模块包括:Optionally, the first feature set generating module includes:
图像输入模块:被配置为执行将所述图像信息输入至微表情识别模型中以输出所述图像信息中的微表情;Image input module: configured to perform input of the image information into the micro-expression recognition model to output the micro-expression in the image information;
映射模块:被配置为执行将所识别的微表情与在视频中出现的时间点相互映射形成微表情特征集合。Mapping module: configured to perform mutual mapping of the recognized micro-expression and the time point that appears in the video to form a micro-expression feature set.
可选的,还包括:Optionally, it also includes:
个人信息获取模块:被配置为执行获取所述目标对象的个人信息;Personal information acquisition module: configured to perform acquisition of the personal information of the target object;
匹配模块:被配置为执行根据所述个人信息在预设应激数据库中匹配应激数据;Matching module: configured to perform matching of stress data in a preset stress database according to the personal information;
第一生成模块:被配置为执行根据预设规则将所述应激数据进行罗列以生成应激问题。The first generating module is configured to perform listing of the stress data according to preset rules to generate stress questions.
可选的,所述第一生成模块包括:Optionally, the first generating module includes:
优先级获取模块:被配置为执行获取所述应激数据的优先级别;Priority acquisition module: configured to execute the priority acquisition of the stress data;
第二生成模块:被配置为执行根据所述优先级别对所述应激数据进行罗列以生成应激问题。The second generating module is configured to perform listing of the stress data according to the priority level to generate stress questions.
可选的,所述应激数据为按照预设的关联关系罗列的多个问题的组合,所述获取模块包括:Optionally, the stress data is a combination of multiple questions listed according to a preset association relationship, and the acquisition module includes:
关联模块:被配置为执行获取所述目标对象回复所述应激问题的图像信息和音频信息,并将所述图像信息和音频信息以时间轴方式进行关联;Associating module: configured to perform acquisition of image information and audio information of the target object's response to the stress problem, and to associate the image information and audio information in a time axis manner;
识别模块:被配置为执行识别所述音频信息中所述目标对象的回复答案;Recognition module: configured to perform recognition of the reply answer of the target object in the audio information;
选择模块:被配置为执行根据所述回复答案选择所述应激数据中的问题组合以进行后续应激问题的提出,并同步获取视频信息。Selection module: configured to perform selection of a combination of questions in the stress data according to the reply answer to raise subsequent stress questions, and obtain video information synchronously.
本申请实施例提供计算机设备基本结构框图请参阅图9。Please refer to FIG. 9 for the basic structure block diagram of the computer equipment provided in the embodiment of the present application.
该计算机设备包括通过系统总线连接的处理器、非易失性存储介质、存储器和网络接口。其中,该计算机设备的非易失性存储介质存储有操作系统、数据库和计算机可读指令,数据库中可存储有控件信息序列,该计算机可读指令被处理器执行时,可使得处理器实现一种面审风险控制方法。该计算机设备的处理器用于提供计算和控制能力,支撑整个计算机设备的运行。该计算机设备的存储器中可存储有计算机可读指令,该计算机可读指令被处理器执行时,可使得处理器执行以下步骤:The computer equipment includes a processor, a non-volatile storage medium, a memory, and a network interface connected through a system bus. Wherein, the non-volatile storage medium of the computer device stores an operating system, a database, and computer-readable instructions. The database may store control information sequences. When the computer-readable instructions are executed by the processor, the processor can realize a A kind of face-to-face audit risk control method. The processor of the computer equipment is used to provide calculation and control capabilities, and supports the operation of the entire computer equipment. The computer readable instructions may be stored in the memory of the computer device, and when the computer readable instructions are executed by the processor, the processor may execute the following steps:
获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;
根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
本申请计算机设备的具体实施例与上述面审风险控制方法各实施例基本相同,在此不作赘述。The specific embodiments of the computer equipment of this application are basically the same as the embodiments of the above-mentioned face-to-face risk control method, and will not be repeated here.
该计算机设备的网络接口用于与终端连接通信。本领域技术人员可以理解,图9中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。The network interface of the computer device is used to connect and communicate with the terminal. Those skilled in the art can understand that the structure shown in FIG. 9 is only a block diagram of a part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied. The specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
计算机设备通过接收关联的客户端发送的提示行为的状态信息,即关联终端是否开启提示以及用户是否关闭该提示任务。通过验证上述任务条件是否达成,进而向关联终端发送对应的预设指令,以使关联终端能够根据该预设指令执行相应的操作,从而实现了对关联终端的有效监管。同时,在提示信息状态与预设的状态指令不相同时,服务器端控制关联终端持续进行响铃,以防止关联终端的提示任务在执行一段时间后自动终止的问题。The computer device receives the status information of the prompt behavior sent by the associated client, that is, whether the associated terminal opens the prompt and whether the user closes the prompt task. By verifying whether the above-mentioned task conditions are fulfilled, the corresponding preset instruction is sent to the associated terminal, so that the associated terminal can perform corresponding operations according to the preset instruction, thereby realizing effective supervision of the associated terminal. At the same time, when the prompt information state is different from the preset state instruction, the server side controls the associated terminal to continue ringing, so as to prevent the prompt task of the associated terminal from automatically terminating after a period of execution.
本申请还提供一种存储有计算机可读指令的存储介质,所述计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行以下步骤:The present application also provides a storage medium storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors perform the following steps:
获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;
识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回 答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;
基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;
根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
本申请计算机可读指令的存储介质的具体实施例与上述面审风险控制方法各实施例基本相同,在此不作赘述。The specific embodiments of the storage medium of the computer-readable instructions of the present application are basically the same as the embodiments of the above-mentioned face-to-face risk control method, and will not be repeated here.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,该计算机程序可存储于一计算机可读取存储介质中,该程序在执行时,可包括如上述各方法的实施例的流程。其中,前述的存储介质可为磁碟、光盘、只读存储记忆体(Read-Only Memory,ROM)等非易失性存储介质,或易失性存储介质,或随机存储记忆体(Random Access Memory,RAM)等。A person of ordinary skill in the art can understand that all or part of the processes in the above-mentioned embodiment methods can be implemented by instructing relevant hardware through a computer program. The computer program can be stored in a computer readable storage medium. When executed, it may include the procedures of the above-mentioned method embodiments. Among them, the aforementioned storage medium can be a non-volatile storage medium such as a magnetic disk, an optical disc, a read-only memory (Read-Only Memory, ROM), or a volatile storage medium, or a random access memory (Random Access Memory). , RAM) etc.
在另一个实施例中,本申请所提供的面审风险控制方法,为进一步保证上述所有出现的数据的私密和安全性,上述所有数据还可以存储于一区块链的节点中。例如视频信息及音频信息等,这些数据均可存储在区块链节点中。In another embodiment, the face-to-face audit risk control method provided in this application further ensures the privacy and security of all the above-mentioned data, all the above-mentioned data can also be stored in a node of a blockchain. For example, video information and audio information, etc., these data can be stored in the blockchain node.
本申请所指区块链是分布式数据存储、点对点传输、共识机制、加密算法等计算机技术的新型应用模式。区块链(Blockchain),本质上是一个去中心化的数据库,是一串使用密码学方法相关联产生的数据块,每一个数据块中包含了一批次网络交易的信息,用于验证其信息的有效性(防伪)和生成下一个区块。区块链可以包括区块链底层平台、平台产品服务层以及应用服务层等。The blockchain referred to in this application is a new application mode of computer technology such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm. Blockchain, essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information for verification. The validity of the information (anti-counterfeiting) and the generation of the next block. The blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
应该理解的是,虽然附图的流程图中的各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,其可以以其他的顺序执行。而且,附图的流程图中的至少一部分步骤可以包括多个子步骤或者多个阶段,这些子步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,其执行顺序也不必然是依次进行,而是可以与其他步骤或者其他步骤的子步骤或者阶段的至少一部分轮流或者交替地执行。It should be understood that although the various steps in the flowchart of the drawings are displayed in sequence as indicated by the arrows, these steps are not necessarily performed in sequence in the order indicated by the arrows. Unless explicitly stated in this article, the execution of these steps is not strictly limited in order, and they can be executed in other orders. Moreover, at least part of the steps in the flowchart of the drawings may include multiple sub-steps or multiple stages. These sub-steps or stages are not necessarily executed at the same time, but can be executed at different times, and the order of execution is also It is not necessarily performed sequentially, but may be performed alternately or alternately with at least a part of other steps or sub-steps or stages of other steps.
以上所述仅是本申请的部分实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本申请原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本申请的保护范围。The above are only part of the implementation of this application. It should be pointed out that for those of ordinary skill in the art, without departing from the principles of this application, several improvements and modifications can be made, and these improvements and modifications are also Should be regarded as the scope of protection of this application.
Claims (20)
- 一种面审风险控制方法,其中,包括:A face-to-face risk control method, which includes:获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
- 根据权利要求1所述的面审风险控制方法,其中,所述回应信息包括微表情特征集合、声音特征集合和回复答案,所述提取各视频片段中所述目标对象回复对应应激问题时的回应信息包括:The face-to-face audit risk control method according to claim 1, wherein the response information includes a set of micro-expression features, a set of voice features, and a response answer, and the extraction of each video clip when the target object responds to the corresponding stress question The response information includes:通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合;Generating a set of micro-expression features that characterize the video segment from the facial and body movements of the target object in the image information in the video segment;通过视频片段中的所述音频信息提取表征该视频片段所述目标对象的声音特征集合和回复答案,其中,所述声音特征集合包括语音、语调、语速和振幅信息;Extracting a voice feature set that characterizes the target object of the video clip and a reply answer from the audio information in the video clip, where the voice feature set includes voice, intonation, speech rate, and amplitude information;将所述微表情特征集合与所述声音特征集合和回复答案以时间轴的方式相互映射以得到对应应激问题的回应信息。The micro-expression feature set, the voice feature set and the reply answer are mapped to each other in a time axis manner to obtain response information corresponding to the stress question.
- 根据权利要求2所述的面审风险控制方法,其中,所述根据各回应信息生成所述目标对象的面审评分值包括:The method for controlling face-to-face audit risk according to claim 2, wherein said generating the face-to-face audit score value of the target object according to each response information comprises:将各个所述回应信息依次输入评分模型中以依次输出对应的视频片段的信用评分分值,其中,所述评分模型为预先训练至收敛状态,根据应激问题及其可能的回复答案、微表情特征和声音特征进行评分的模型;Each of the response information is sequentially input into the scoring model to sequentially output the credit scoring value of the corresponding video clip, wherein the scoring model is pre-trained to a convergent state, based on the stress question and its possible response answers, micro-expressions A model for scoring features and voice features;获取各个视频片段对应的应激问题的评分权重值;Obtain the scoring weight value of the stress problem corresponding to each video clip;根据所述评分权重值与对应的所述信用评分分值采用加权平均的方式计算生成面审评分值。According to the scoring weight value and the corresponding credit scoring value, a weighted average method is used to calculate and generate a face-to-face review score value.
- 根据权利要求2所述的面审风险控制方法,其中,所述通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合包括:The face-to-face audit risk control method according to claim 2, wherein said generating a micro-expression feature set characterizing the video clip from the facial and body actions of the target object in the image information in the video clip comprises:将所述图像信息输入至微表情识别模型中以输出所述图像信息中的微表情;Inputting the image information into a micro-expression recognition model to output the micro-expression in the image information;将所识别的微表情与对应视频片段中出现的时间点相互映射形成微表情特征集合。The identified micro-expression and the time point in the corresponding video segment are mapped to each other to form a micro-expression feature set.
- 根据权利要求1所述的面审风险控制方法,其中,根据收集的所述目标对象的个人信息生成所述应激问题包括:The face-to-face audit risk control method according to claim 1, wherein generating the stress question according to the collected personal information of the target object comprises:获取所述目标对象的个人信息;Obtaining personal information of the target object;根据所述个人信息在预设应激数据库中匹配应激数据;Matching stress data in a preset stress database according to the personal information;根据预设规则将所述应激数据进行罗列以生成应激问题。List the stress data according to preset rules to generate stress problems.
- 根据权利要求5所述的面审风险控制方法,其中,所述根据预设规则将所述应激数据进行罗列以生成应激问题包括:The face-to-face audit risk control method according to claim 5, wherein the listing the stress data according to a preset rule to generate a stress problem comprises:获取所述应激数据的优先级别;Acquiring the priority level of the stress data;根据所述优先级别对所述应激数据进行罗列以生成应激问题。The stress data is listed according to the priority level to generate stress problems.
- 根据权利要求6所述的面审风险控制方法,其中,所述应激数据为按照预设的关联关系罗列的多个问题的组合,所述获取目标对象回复应激问题的视频信息包括:The face-to-face audit risk control method according to claim 6, wherein the stress data is a combination of a plurality of questions listed according to a preset association relationship, and the obtaining the video information of the target object responding to the stress question comprises:获取所述目标对象回复所述应激问题的图像信息和音频信息,并将所述图像信息 和音频信息以时间轴方式进行关联;Acquiring image information and audio information of the target object in reply to the stress question, and associating the image information and audio information in a time axis manner;识别所述音频信息中所述目标对象的回复答案;Identifying the reply answer of the target object in the audio information;根据所述回复答案选择所述应激数据中的问题组合以进行后续应激问题的提出,并同步获取视频信息。According to the reply answer, a combination of questions in the stress data is selected to raise subsequent stress questions, and video information is obtained synchronously.
- 一种面审风险控制装置,其中,包括:A face-to-face audit risk control device, which includes:获取模块:被配置为执行获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;An acquisition module: configured to perform acquisition of video information of a target object responding to a stress question, wherein the stress question is a question set generated according to the collected personal information of the target object;分解模块:被配置为执行根据识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Decomposition module: configured to execute the time node proposed according to the identification of each stress problem in the video information, and divide the video information into several video segments according to the time node, where the video segment includes Image information that characterizes the change in expression of the target object when answering the corresponding stress question and audio information that characterizes the change in voice of the target object and the reply answer when answering the corresponding stress question;提取模块:被配置为执行基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Extraction module: configured to perform extraction of response information when the target object in each video segment replies to a corresponding stress question based on the image information and audio information;评分模块:被配置为执行根据各回应信息生成所述目标对象的面审评分值。The scoring module is configured to execute the generation of the score value of the face-to-face review of the target object according to each response information.
- 一种计算机设备,包括存储器和处理器,所述存储器中存储有计算机可读指令,所述计算机可读指令被所述处理器执行时,使得所述处理器执行如下步骤:A computer device includes a memory and a processor. The memory stores computer-readable instructions, and when the computer-readable instructions are executed by the processor, the processor executes the following steps:获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
- 根据权利要求9所述的计算机设备,其中,所述回应信息包括微表情特征集合、声音特征集合和回复答案,所述提取各视频片段中所述目标对象回复对应应激问题时的回应信息包括:The computer device according to claim 9, wherein the response information includes a micro-expression feature set, a sound feature set, and a response answer, and the extracted response information when the target object in each video clip responds to a corresponding stress question includes :通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合;Generating a set of micro-expression features that characterize the video segment from the facial and body movements of the target object in the image information in the video segment;通过视频片段中的所述音频信息提取表征该视频片段所述目标对象的声音特征集合和回复答案,其中,所述声音特征集合包括语音、语调、语速和振幅信息;Extracting a voice feature set that characterizes the target object of the video clip and a reply answer from the audio information in the video clip, where the voice feature set includes voice, intonation, speech rate, and amplitude information;将所述微表情特征集合与所述声音特征集合和回复答案以时间轴的方式相互映射以得到对应应激问题的回应信息。The micro-expression feature set, the voice feature set and the reply answer are mapped to each other in a time axis manner to obtain response information corresponding to the stress question.
- 根据权利要求10所述的计算机设备,其中,所述根据各回应信息生成所述目标对象的面审评分值包括:11. The computer device according to claim 10, wherein said generating the face-to-face review score value of the target object according to each response information comprises:将各个所述回应信息依次输入评分模型中以依次输出对应的视频片段的信用评分分值,其中,所述评分模型为预先训练至收敛状态,根据应激问题及其可能的回复答案、微表情特征和声音特征进行评分的模型;Each of the response information is sequentially input into the scoring model to sequentially output the credit scoring value of the corresponding video clip, wherein the scoring model is pre-trained to a convergent state, based on the stress question and its possible response answers, micro-expressions A model for scoring features and voice features;获取各个视频片段对应的应激问题的评分权重值;Obtain the scoring weight value of the stress problem corresponding to each video clip;根据所述评分权重值与对应的所述信用评分分值采用加权平均的方式计算生成面审评分值。According to the scoring weight value and the corresponding credit scoring value, a weighted average method is used to calculate and generate a face-to-face review score value.
- 根据权利要求10所述的计算机设备,其中,所述通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合包括:10. The computer device according to claim 10, wherein said generating a set of micro-expression features characterizing the video clip from the facial and body actions of the target object in the image information in the video clip comprises:将所述图像信息输入至微表情识别模型中以输出所述图像信息中的微表情;Inputting the image information into a micro-expression recognition model to output the micro-expression in the image information;将所识别的微表情与对应视频片段中出现的时间点相互映射形成微表情特征集合。The identified micro-expression and the time point in the corresponding video segment are mapped to each other to form a micro-expression feature set.
- 根据权利要求9所述的计算机设备,其中,根据收集的所述目标对象的个人信息生成所述应激问题包括:The computer device according to claim 9, wherein generating the stress question based on the collected personal information of the target object comprises:获取所述目标对象的个人信息;Obtaining personal information of the target object;根据所述个人信息在预设应激数据库中匹配应激数据;Matching stress data in a preset stress database according to the personal information;根据预设规则将所述应激数据进行罗列以生成应激问题。List the stress data according to preset rules to generate stress problems.
- 根据权利要求13所述的计算机设备,其中,所述根据预设规则将所述应激数据进行罗列以生成应激问题包括:The computer device according to claim 13, wherein the listing the stress data according to a preset rule to generate a stress question comprises:获取所述应激数据的优先级别;Acquiring the priority level of the stress data;根据所述优先级别对所述应激数据进行罗列以生成应激问题。The stress data is listed according to the priority level to generate stress problems.
- 根据权利要求14所述的计算机设备,其中,所述应激数据为按照预设的关联关系罗列的多个问题的组合,所述获取目标对象回复应激问题的视频信息包括:The computer device according to claim 14, wherein the stress data is a combination of a plurality of questions listed according to a preset association relationship, and the obtaining the video information of the target object responding to the stress question comprises:获取所述目标对象回复所述应激问题的图像信息和音频信息,并将所述图像信息和音频信息以时间轴方式进行关联;Acquiring image information and audio information of the target object in reply to the stress question, and associating the image information and audio information in a time axis manner;识别所述音频信息中所述目标对象的回复答案;Identifying the reply answer of the target object in the audio information;根据所述回复答案选择所述应激数据中的问题组合以进行后续应激问题的提出,并同步获取视频信息。According to the reply answer, a combination of questions in the stress data is selected to raise subsequent stress questions, and video information is obtained synchronously.
- 一种存储有计算机可读指令的存储介质,所述计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行如下步骤:A storage medium storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors execute the following steps:获取目标对象回复应激问题的视频信息,其中,所述应激问题为根据收集的所述目标对象的个人信息生成的问题集合;Acquiring video information of a target object responding to a stress question, where the stress question is a set of questions generated based on the collected personal information of the target object;识别所述视频信息中每个应激问题提出的时间节点,并根据所述时间节点将所述视频信息分割成若干个视频片段,其中,所述视频片段中包括表征所述目标对象在回答对应的应激问题时表情变化的图像信息和回答对应的应激问题时表征所述目标对象声音变化及回复答案的音频信息;Identify the time node raised by each stress question in the video information, and divide the video information into a number of video segments according to the time node, where the video segment includes the character that the target object is answering the corresponding The image information of the facial expression change during the stress question and the audio information representing the voice change of the target object and the reply answer when answering the corresponding stress question;基于所述图像信息和音频信息,提取各视频片段中所述目标对象回复对应应激问题时的回应信息;Based on the image information and audio information, extract the response information when the target object in each video clip responds to the corresponding stress question;根据各回应信息生成所述目标对象的面审评分值。The face-to-face review score value of the target object is generated according to each response information.
- 根据权利要求16所述的存储介质,其中,所述回应信息包括微表情特征集合、声音特征集合和回复答案,所述提取各视频片段中所述目标对象回复对应应激问题时的回应信息包括:The storage medium according to claim 16, wherein the response information includes a micro-expression feature set, a sound feature set, and a response answer, and the extracted response information when the target object in each video clip responds to a corresponding stress question includes :通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合;Generating a set of micro-expression features that characterize the video segment from the facial and body movements of the target object in the image information in the video segment;通过视频片段中的所述音频信息提取表征该视频片段所述目标对象的声音特征集合和回复答案,其中,所述声音特征集合包括语音、语调、语速和振幅信息;Extracting a voice feature set that characterizes the target object of the video clip and a reply answer from the audio information in the video clip, where the voice feature set includes voice, intonation, speech rate, and amplitude information;将所述微表情特征集合与所述声音特征集合和回复答案以时间轴的方式相互映射以得到对应应激问题的回应信息。The micro-expression feature set, the voice feature set and the reply answer are mapped to each other in a time axis manner to obtain response information corresponding to the stress question.
- 根据权利要求17所述的存储介质,其中,所述根据各回应信息生成所述目标对象的面审评分值包括:18. The storage medium according to claim 17, wherein said generating the face-to-face review score value of the target object according to each response information comprises:将各个所述回应信息依次输入评分模型中以依次输出对应的视频片段的信用评分分值,其中,所述评分模型为预先训练至收敛状态,根据应激问题及其可能的回复答案、微表情特征和声音特征进行评分的模型;Each of the response information is sequentially input into the scoring model to sequentially output the credit scoring value of the corresponding video clip, wherein the scoring model is pre-trained to a convergent state, based on the stress question and its possible response answers, micro-expressions A model for scoring features and voice features;获取各个视频片段对应的应激问题的评分权重值;Obtain the scoring weight value of the stress problem corresponding to each video clip;根据所述评分权重值与对应的所述信用评分分值采用加权平均的方式计算生成面审评分值。According to the scoring weight value and the corresponding credit scoring value, a weighted average method is used to calculate and generate a face-to-face review score value.
- 根据权利要求17所述的存储介质,其中,所述通过视频片段中的所述图像信息中目标对象的面部和肢体动作生成表征该视频片段的微表情特征集合包括:18. The storage medium according to claim 17, wherein said generating a set of micro-expression features characterizing the video clip from the facial and body movements of the target object in the image information in the video clip comprises:将所述图像信息输入至微表情识别模型中以输出所述图像信息中的微表情;Inputting the image information into a micro-expression recognition model to output the micro-expression in the image information;将所识别的微表情与对应视频片段中出现的时间点相互映射形成微表情特征集合。The identified micro-expression and the time point in the corresponding video segment are mapped to each other to form a micro-expression feature set.
- 根据权利要求16所述的存储介质,其中,根据收集的所述目标对象的个人信息生成所述应激问题包括:The storage medium according to claim 16, wherein generating the stress question according to the collected personal information of the target object comprises:获取所述目标对象的个人信息;Obtaining personal information of the target object;根据所述个人信息在预设应激数据库中匹配应激数据;Matching stress data in a preset stress database according to the personal information;根据预设规则将所述应激数据进行罗列以生成应激问题。List the stress data according to preset rules to generate stress problems.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010225123.0 | 2020-03-26 | ||
CN202010225123.0A CN111429267A (en) | 2020-03-26 | 2020-03-26 | Face examination risk control method and device, computer equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021190086A1 true WO2021190086A1 (en) | 2021-09-30 |
Family
ID=71548955
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/070931 WO2021190086A1 (en) | 2020-03-26 | 2021-01-08 | Face-to-face examination risk control method and apparatus, computer device, and storage medium |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111429267A (en) |
WO (1) | WO2021190086A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118536879A (en) * | 2024-07-24 | 2024-08-23 | 杭州视洞科技有限公司 | Intelligent quality inspection system platform based on audio and video |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111429267A (en) * | 2020-03-26 | 2020-07-17 | 深圳壹账通智能科技有限公司 | Face examination risk control method and device, computer equipment and storage medium |
CN112151027B (en) * | 2020-08-21 | 2024-05-03 | 深圳追一科技有限公司 | Method, device and storage medium for querying specific person based on digital person |
CN112182537A (en) * | 2020-09-28 | 2021-01-05 | 深圳前海微众银行股份有限公司 | Monitoring method, device, server, system and storage medium |
CN112381513A (en) * | 2020-11-13 | 2021-02-19 | 平安普惠企业管理有限公司 | Information approval method and device, electronic equipment and storage medium |
CN113080969B (en) * | 2021-03-29 | 2022-06-21 | 济南大学 | Multi-mode feature-based lie detection data processing method and system |
CN113065449B (en) * | 2021-03-29 | 2022-08-19 | 济南大学 | Face image acquisition method and device, computer equipment and storage medium |
CN113487400B (en) * | 2021-06-04 | 2022-10-11 | 长春工业大学 | Financial credit consensus method based on honesty bidirectional selection |
CN116109989B (en) * | 2023-04-14 | 2023-06-27 | 中关村科学城城市大脑股份有限公司 | Evaluation information generation method, apparatus, electronic device, and computer-readable medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104644189A (en) * | 2015-03-04 | 2015-05-27 | 刘镇江 | Analysis method for psychological activities |
CN105069318A (en) * | 2015-09-12 | 2015-11-18 | 宁波江东泓源工业设计有限公司 | Emotion analysis method |
US20180225307A1 (en) * | 2016-08-31 | 2018-08-09 | Robert William Kocher | Two-stage, facial recognition and identification system (two-stage facial R & I system) |
CN110135800A (en) * | 2019-04-23 | 2019-08-16 | 南京葡萄诚信息科技有限公司 | A kind of artificial intelligence video interview method and system |
CN111429267A (en) * | 2020-03-26 | 2020-07-17 | 深圳壹账通智能科技有限公司 | Face examination risk control method and device, computer equipment and storage medium |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107704834B (en) * | 2017-10-13 | 2021-03-30 | 深圳壹账通智能科技有限公司 | Micro-surface examination assisting method, device and storage medium |
CN109829358A (en) * | 2018-12-14 | 2019-05-31 | 深圳壹账通智能科技有限公司 | Micro- expression loan control method, device, computer equipment and storage medium |
CN109815803B (en) * | 2018-12-18 | 2023-04-18 | 平安科技(深圳)有限公司 | Face examination risk control method and device, computer equipment and storage medium |
CN110148406B (en) * | 2019-04-12 | 2022-03-04 | 北京搜狗科技发展有限公司 | Data processing method and device for data processing |
CN110378228A (en) * | 2019-06-17 | 2019-10-25 | 深圳壹账通智能科技有限公司 | Video data handling procedure, device, computer equipment and storage medium are examined in face |
-
2020
- 2020-03-26 CN CN202010225123.0A patent/CN111429267A/en active Pending
-
2021
- 2021-01-08 WO PCT/CN2021/070931 patent/WO2021190086A1/en active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104644189A (en) * | 2015-03-04 | 2015-05-27 | 刘镇江 | Analysis method for psychological activities |
CN105069318A (en) * | 2015-09-12 | 2015-11-18 | 宁波江东泓源工业设计有限公司 | Emotion analysis method |
US20180225307A1 (en) * | 2016-08-31 | 2018-08-09 | Robert William Kocher | Two-stage, facial recognition and identification system (two-stage facial R & I system) |
CN110135800A (en) * | 2019-04-23 | 2019-08-16 | 南京葡萄诚信息科技有限公司 | A kind of artificial intelligence video interview method and system |
CN111429267A (en) * | 2020-03-26 | 2020-07-17 | 深圳壹账通智能科技有限公司 | Face examination risk control method and device, computer equipment and storage medium |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118536879A (en) * | 2024-07-24 | 2024-08-23 | 杭州视洞科技有限公司 | Intelligent quality inspection system platform based on audio and video |
Also Published As
Publication number | Publication date |
---|---|
CN111429267A (en) | 2020-07-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021190086A1 (en) | Face-to-face examination risk control method and apparatus, computer device, and storage medium | |
Buolamwini | Gender shades: intersectional phenotypic and demographic evaluation of face datasets and gender classifiers | |
CN107680019B (en) | Examination scheme implementation method, device, equipment and storage medium | |
WO2020119272A1 (en) | Risk identification model training method and apparatus, and server | |
US10685329B2 (en) | Model-driven evaluator bias detection | |
US11800014B2 (en) | Method and system for proactive fraudster exposure in a customer service channel | |
US11252279B2 (en) | Method and system for fraud clustering by content and biometrics analysis | |
CN109767321A (en) | Question answering process optimization method, device, computer equipment and storage medium | |
US20200294130A1 (en) | Loan matching system and method | |
Liu et al. | Can listing information indicate borrower credit risk in online peer-to-peer lending? | |
Chariri et al. | Individual characteristics, financial literacy and ability in detecting investment scams | |
KR102648772B1 (en) | Online examination platform system of preventing and coping with cheating using metaverse | |
KR102353529B1 (en) | Method and server for collecting job duty information of company based on blockchain | |
Qi et al. | Do facial images matter? Understanding the role of private information disclosure in crowdfunding markets | |
Preu et al. | Perception vs. Reality: understanding and evaluating the impact of synthetic image deepfakes over college students | |
Moreno Moreno et al. | A model proposal to determine a crowd-credit-scoring | |
Karim et al. | Financial Literacy among University Students and its Implications towards Financial Scams | |
CN113723774A (en) | Answer scoring method and device, computer equipment and storage medium | |
US11508377B2 (en) | System and method for detecting fraud rings | |
TWM574723U (en) | Capital loan intelligent audit system | |
Kugler et al. | Surveillance Duration Doesn't Affect Privacy Expectations: An Empirical Test of the Mosaic Theory | |
TWI827910B (en) | Credit evaluation method and system | |
Pawde et al. | Synthesizing data for collusion-based malpractice of shell companies in money laundering | |
CN116401685A (en) | Data full-flow risk internal examination method, device and platform | |
Mulgund et al. | Deepfake: The Lay of the Land |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21775863 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 17.01.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21775863 Country of ref document: EP Kind code of ref document: A1 |