WO2023231211A1 - Voice recognition method and apparatus, electronic device, storage medium, and product - Google Patents

Voice recognition method and apparatus, electronic device, storage medium, and product Download PDF

Info

Publication number
WO2023231211A1
WO2023231211A1 PCT/CN2022/117333 CN2022117333W WO2023231211A1 WO 2023231211 A1 WO2023231211 A1 WO 2023231211A1 CN 2022117333 W CN2022117333 W CN 2022117333W WO 2023231211 A1 WO2023231211 A1 WO 2023231211A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
vehicle
recognition
voice information
state
Prior art date
Application number
PCT/CN2022/117333
Other languages
French (fr)
Chinese (zh)
Inventor
蒋磊
蔡勇
Original Assignee
合众新能源汽车股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 合众新能源汽车股份有限公司 filed Critical 合众新能源汽车股份有限公司
Publication of WO2023231211A1 publication Critical patent/WO2023231211A1/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Definitions

  • the present application relates to the field of speech understanding technology, and in particular to a speech recognition method, device, electronic equipment, computer-readable storage medium and computer program product.
  • voice function is an important function of smart cars.
  • the user needs to use the wake-up word every time he communicates with the car, for example, "The user says: Hello Nezha" to wake up the voice function of the car. Since the wake-up word has to be used every time, it will be more troublesome.
  • This application provides a speech recognition method, device, electronic equipment, computer-readable storage medium and computer program product, to at least solve the problem in related technologies that the vehicle-machine instructions cannot be accurately recognized due to the inability to accurately recognize the voice in the car, resulting in vehicle-machine execution errors. instructions, technical problems that increase the rate of misoperation.
  • the technical solution of this application is as follows:
  • a speech recognition method including:
  • the recognition result is a vehicle-machine instruction
  • the corresponding operation is performed according to the vehicle-machine instruction.
  • the method also includes:
  • determining the current status of the user based on facial features on the facial image includes at least one of the following:
  • the voice information is recognized and the recognition result is obtained, including:
  • the current state of the user is at least one of: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state, it is determined that the user satisfies the set condition;
  • the voice information is recognized and a recognition result is obtained.
  • the recognition of the voice information to obtain the recognition result includes:
  • performing corresponding operations according to the vehicle-machine instruction includes:
  • the recognition result is judged by a trained vehicle-machine command recognition model, and the recognition result is a vehicle-machine command; wherein the trained vehicle-machine command recognition model is based on multiple histories of human-vehicle interaction A model obtained by learning and training audio pairs, text pairs, scenes and keywords;
  • a speech recognition device including:
  • the acquisition module is used to respond to the voice information of the user in the car and obtain the facial image of the user;
  • a determining module configured to determine the current status of the user based on facial features on the facial image
  • a recognition module used to recognize the voice information and obtain a recognition result when the current status of the user meets the set conditions
  • An execution module when the recognition result is a vehicle-machine instruction, executes the corresponding operation according to the vehicle-machine instruction.
  • the device also includes:
  • a recognition rejection module is configured to reject recognition of the voice information when the user's current status does not meet the set conditions.
  • the determination module includes at least one of the following modules:
  • a first determination module configured to determine that the user is in a non-phone state when determining that the vehicle-mounted Bluetooth phone is not turned on based on the obtained information status of the vehicle and the facial features of the facial image;
  • a second determination module configured to determine that the user is in a state of facing forward when it is determined based on the facial features of the facial image that the user's front face is looking in the direction of vehicle travel;
  • the third determination module is configured to determine that the user is in a speaking state when it is determined that the user's mouth is in an opening and closing state based on the facial features of the facial image.
  • the identification module includes:
  • the first judgment module is used to determine that the set conditions are met when the current state of the user is at least one of: the user is not on the phone, the user is in a state of facing forward, and the user is in a speaking state. ;
  • a speech recognition module is used to recognize the speech information and obtain a recognition result.
  • the speech recognition module includes: a speech conversion module; and/or a sending module and a receiving module, wherein,
  • the voice conversion module is used to perform local voice-to-text conversion processing on the voice information to obtain converted text information
  • the sending module is used to send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information;
  • the receiving module is used to receive the converted text information sent by the cloud.
  • the execution module includes:
  • the second judgment module is used to judge the recognition result through a trained vehicle-machine command recognition model, and obtain that the recognition result is a vehicle-machine command; wherein the trained vehicle-machine command recognition model is based on human and computer commands.
  • a model obtained by learning and training multiple historical audio pairs, text pairs, scenes and keywords of vehicle-computer interaction;
  • An instruction execution module is used to execute corresponding operations according to the vehicle-machine instructions obtained by the second judgment module.
  • an electronic device including:
  • memory for storing instructions executable by the processor
  • the processor is configured to execute the instructions to implement the speech recognition method as described above.
  • a computer-readable storage medium which when instructions in the computer-readable storage medium are executed by a processor of an electronic device, enables the electronic device to perform speech recognition as described above. method.
  • a computer program product including a computer program or instructions that implement the speech recognition method as described above when executed by a processor.
  • the user's facial image in response to the voice information of the user in the car, the user's facial image is obtained; the current state of the user is determined based on the facial features on the facial image; and the user's current state satisfies the set conditions
  • the voice information is recognized to obtain a recognition result; when the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction. That is to say, in the embodiment of the present application, the current status of the user is determined based on the facial features on the facial image, and the voice information is recognized based on the user's current status, so that it can be accurately determined which voice information is from the vehicle.
  • Commands, which voice messages are not vehicle-machine instructions improve the efficiency of the vehicle-machine accurately executing vehicle-machine instructions, reduce the rate of vehicle-machine misoperation, and also improve the user experience.
  • Figure 1 is a flow chart of a speech recognition method provided by an embodiment of the present application.
  • Figure 2 is a flow chart of an application example of a speech recognition method provided by an embodiment of the present application.
  • Figure 3 is a block diagram of a speech recognition device provided by an embodiment of the present application.
  • Figure 4 is another block diagram of a speech recognition device provided by an embodiment of the present application.
  • Figure 5 is a block diagram of a determination module provided by an embodiment of the present application.
  • FIG. 6 is a block diagram of an identification module provided by an embodiment of the present application.
  • FIG. 7 is a block diagram of an execution module provided by an embodiment of the present application.
  • Figure 8 is a block diagram of an electronic device provided by an embodiment of the present application.
  • Figure 9 is a block diagram of a device for speech recognition provided by an embodiment of the present application.
  • Artificial Intelligence is an emerging science and technology that studies and develops theories, methods, technologies and application systems for simulating and extending human intelligence.
  • the subject of artificial intelligence is a comprehensive subject, involving many types of technologies such as chips, big data, cloud computing, Internet of Things, distributed storage, deep learning, machine learning, neural networks, etc.
  • Computer vision as an important branch of artificial intelligence, specifically allows machines to recognize the world.
  • Computer vision technology usually includes face recognition, live body detection, fingerprint recognition and anti-counterfeiting verification, biometric recognition, face detection, pedestrian detection, target detection, pedestrian Recognition, image processing, image recognition, image semantic understanding, image retrieval, text recognition, video processing, video content recognition, behavior recognition, 3D reconstruction, virtual reality, augmented reality, simultaneous localization and mapping (SLAM), computational photography, robotics Navigation and positioning technologies.
  • FIG. 1 is a flow chart of a speech recognition method provided by an embodiment of the present application. As shown in Figure 1, the speech recognition method includes the following steps:
  • Step 101 Respond to the voice message of the user in the car and obtain the user's facial image.
  • Step 102 Determine the current status of the user based on facial features on the facial image.
  • Step 103 When the user's current status meets the set conditions, recognize the voice information and obtain a recognition result.
  • Step 104 When the recognition result is a vehicle-machine instruction, perform the corresponding operation according to the vehicle-machine instruction.
  • the speech recognition method described in this application can be applied to vehicle-machine terminals, etc., and is not limited here.
  • vehicle-machine terminal implementation equipment can be electronic equipment such as smart car-machine, vehicle-machine platform, etc., which is not limited here.
  • step 101 in response to the voice information of the user in the car, the facial image of the user is obtained.
  • the car terminal can detect the user's voice information through the microphone on the vehicle.
  • the face of the user in the car can be obtained through the image collection device (such as a camera, etc.) on the vehicle.
  • Image the facial image can be a single frame image or multiple frame images.
  • the image acquisition device can be set at a position aimed at the driver, so that the image acquisition device can clearly capture the driver's facial image.
  • step 102 the current status of the user is determined based on facial features on the facial image.
  • facial image recognition uses computer image processing technology to extract facial feature points from facial images, such as whether the eyes are open, whether the mouth is open, etc.
  • determining the current user's facial state based on facial features may include at least one of the following, but is not limited to this:
  • the vehicle first obtain the information status of the vehicle, such as whether the Bluetooth phone in the car (i.e., the car Bluetooth phone) is turned on, etc., and then combine the facial features in the facial image (such as whether the mouth is open and closed, etc.) to determine the user Are you on the phone? For example, if the car Bluetooth phone is in the open state and the user's mouth is opening and closing, it is determined that the user is making a call at this time; otherwise, it is determined that the user is in a non-call state, for example, the car Bluetooth phone is in a non-open state.
  • the information status of the vehicle such as whether the Bluetooth phone in the car (i.e., the car Bluetooth phone) is turned on, etc.
  • the facial features in the facial image such as whether the mouth is open and closed, etc.
  • the user's mouth is in an open and closed state, then it is determined that the user is in a speaking state, not that the user is in a non-calling state; of course, if the car Bluetooth phone is in a non-open state and the user's mouth is in a closed state, then Make sure the user is not speaking, is in a quiet state, etc.
  • multi-angle face recognition technology can be used to determine whether the user's front face is looking in the direction of the vehicle. If so, it is determined that the user is facing forward. Otherwise, it is determined that the user is not facing forward.
  • the state of looking forward That is to say, it is determined whether the user's front face is looking within 90 degrees of the vehicle's driving direction. If so, it is determined that the front face is facing forward.
  • the multi-angle face recognition technology is a branch of the multi-pose face recognition technology.
  • a deep learning multi-angle face recognition algorithm includes: first, constructing a deep learning training data set; second, training a deep face classifier; finally, applying the classifier for face detection.
  • the specific implementation process is a well-known technology in this field and will not be described again.
  • this algorithm takes the side image of the face as input and the corresponding frontal image of the face as the output.
  • the supervised model learns the mapping from the side image of the face in different poses to the frontal image, thus increasing the effectiveness in recognition. facial information.
  • the trained face angle classification model can also be used to determine the frontal angle of whether the user is looking forward. If it is judged that the user's front face is within the range of 90 degrees forward, all Make sure the user is facing forward.
  • this step it is determined according to the facial features of the facial image whether the user's mouth is in an open and closed state. If so, it is determined that the user is in a speaking state; otherwise, it is determined that the user is in a non-speaking state.
  • the lip movement feature extraction algorithm can be used to determine whether the user opens his mouth, thereby determining whether the user has lip movement.
  • speaking user identification technology based on lip movement can also be used to extract information that reflects both the physiological characteristics of the speaker's mouth and the behavioral characteristics of the speaker's lip movements from the image sequence of the speaking user through discrete cosine transformation.
  • Visual features Based on these visual features, a static and dynamic hybrid model is established for the speaking user to determine whether the user has lip movements. The specific process is a familiar technique to those skilled in the art and will not be described in detail here.
  • step 103 when the current status of the user meets the set conditions, the voice information is recognized and a recognition result is obtained.
  • the setting conditions include at least one of the following: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state.
  • the recognition of the voice information can be performed. The best way for this embodiment is to satisfy all the above setting conditions.
  • the voice information is recognized and the recognition results are obtained, including:
  • the voice information is subjected to local voice-to-text conversion processing to obtain converted text information.
  • Another situation is to send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information; and receives the converted text information sent by the cloud.
  • step 104 when the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction.
  • the trained vehicle-to-machine command model is a model obtained by learning and training based on multiple historical audio pairs, text pairs, scenes and keywords of human-vehicle interaction.
  • the vehicle-machine command recognition model is trained in advance, wherein the input for training the vehicle-machine command recognition model usually selects historical dialogue audio of multiple human-vehicle-machine (referred to as human-machine) interactions, etc.
  • the output results include: 1 means speaking to the vehicle and the computer, that is, the vehicle and computer commands; 0 does not speak to the car, that is, it is not a vehicle and computer command; of course, it can also be set to: 0 means speaking to the vehicle and the computer, 1 does not speak to the vehicle and the computer, etc. , there is no limitation in this embodiment.
  • the training of the vehicle-machine command recognition model is to allow the vehicle-machine command model to learn more vehicle-machine commands from it, thereby improving the accuracy of training the vehicle-machine command recognition model.
  • this embodiment selects a large number of data groups for learning.
  • Each group of data includes: historical audio and current audio to learn which type of audio speaks to the vehicle, that is, the vehicle instructions issued to the vehicle. .
  • this embodiment can also learn which texts are command words from the text. If the command words are not rich enough, use the historical results as the input this time, thereby improving the accuracy of the vehicle-machine command recognition model training. .
  • the user's facial image in response to the voice information of the user in the car, the user's facial image is obtained; the current state of the user is determined based on the facial features on the facial image; and the user's current state satisfies the set conditions
  • the voice information is recognized to obtain a recognition result; when the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction. That is to say, in the embodiment of the present application, the current status of the user is determined based on the facial features on the facial image, and the voice information is recognized based on the user's current status, so that it can be accurately determined which voice information is from the vehicle. Instructions, which voice messages are not vehicle-machine instructions, improve the efficiency of the vehicle-machine executing accurate vehicle-machine instructions, reduce the rate of vehicle-machine misoperation, and also improve the user experience.
  • Figure 2 is an application example diagram of a speech recognition method provided by an embodiment of the present application.
  • the method is applied to a vehicle-machine terminal.
  • the method includes:
  • Step 201 Detect the voice information of the user in the car
  • the car terminal detects the user's voice information.
  • Step 202 Obtain the user's facial image
  • Step 203 Determine the current status of the user based on the facial features on the facial image
  • the current state of the user includes: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state, but in practical applications, it is not limited to this.
  • Step 204 Determine whether the user is currently on the phone. If not, perform step 205; otherwise, perform step 210:
  • Step 205 Determine whether the current state of the user is facing the direction of the vehicle. If so, perform step 206; otherwise, perform step 210:
  • Step 206 Determine whether the current state of the user is in an open and closed state. If so, perform step 207; otherwise, perform step 210:
  • Step 207 Recognize the voice information and obtain the recognition result
  • Step 208 Determine whether the recognition result is a vehicle-machine command. If so, execute step 209; otherwise, execute step 211;
  • Step 209 Perform corresponding operations according to the vehicle and machine instructions
  • Step 210 Refuse to recognize the voice information, that is, reject recognition.
  • Step 211 Refuse to execute the identification result.
  • the current status of the user is determined based on the facial features on the facial image, and the voice information is recognized based on the user's current status, so that it can be accurately determined which voice information is a vehicle-machine command and which voice information is
  • the information is not a vehicle-machine instruction, that is, multi-mode (such as visual and audio, etc.) is used to determine whether the voice message is a vehicle-machine instruction, which improves the efficiency of the vehicle-machine executing accurate vehicle-machine instructions and reduces the probability of "false recall" of the vehicle-machine. It also improves user experience. That is to say, the embodiment of the present application uses the visual system and voice system in the car to reduce the "false recall" rate of the car and improve the user experience.
  • Figure 3 is a block diagram of a speech recognition device provided by an embodiment of the present application.
  • the device includes: an acquisition module 301, a determination module 302, an identification module 303 and an execution module 304, where,
  • the acquisition module 301 is used to respond to the voice information of the user in the car and acquire the facial image of the user;
  • the determination module 302 is used to determine the current status of the user according to the facial features on the facial image
  • the recognition module 303 is used to recognize the voice information and obtain a recognition result when the user's current status meets the set conditions;
  • the execution module 304 when the recognition result is a vehicle-machine instruction, performs corresponding operations according to the vehicle-machine instruction.
  • the device further includes: a rejection identification module 401, the structural block diagram of which is shown in Figure 4, wherein,
  • the recognition rejection module 401 is used to reject recognition of the voice information when the user's current status does not meet the set conditions.
  • the determination module 302 includes at least one of the following modules: a first determination module 501, a second determination module 502 and a third determination module. 503, the structural block diagram of which is shown in Figure 5, in which this embodiment includes all modules at the same time as an example:
  • the first determination module 501 is used to determine that the user is in a non-phone state when determining that the vehicle-mounted Bluetooth phone is not turned on based on the information status of the vehicle and the facial features of the facial image;
  • the second determination module 502 is configured to determine that the user is in a state of facing forward when it is determined based on the facial features of the facial image that the user's front face is looking in the direction of vehicle travel;
  • the third determination module 503 is configured to determine that the user is in a speaking state when it is determined that the user's mouth is in an opening and closing state based on the facial features of the facial image.
  • the recognition module 303 includes: a first judgment module 601 and a speech recognition module 602, whose structural block diagram is shown in Figure 6, where ,
  • the first judgment module 601 is used to judge that when the current state of the user is at least one of: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state, the condition is satisfied. set conditions;
  • the speech recognition module 602 is used to recognize the speech information and obtain a recognition result.
  • the speech recognition module includes: a speech conversion module; and/or a sending module and a receiving module, wherein,
  • the voice conversion module is used to perform local voice-to-text conversion processing on the voice information to obtain converted text information
  • the sending module is used to send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information;
  • the receiving module is used to receive the converted text information sent by the cloud.
  • the execution module 304 includes: a second judgment module 701 and an instruction execution module 702, whose structural block diagram is shown in Figure 7, where ,
  • the second judgment module 701 is used to judge the recognition result through a trained vehicle-machine command recognition model, and obtain that the recognition result is a vehicle-machine command; wherein the trained vehicle-machine command recognition model is based on A model obtained by learning and training multiple historical audio pairs, text pairs, scenes and keywords of human-vehicle interaction;
  • the instruction execution module 702 is used to execute corresponding operations according to the vehicle-machine instructions obtained by the second judgment module 701 .
  • the device embodiments described above are only illustrative.
  • the modules described as separate components may or may not be physically separated.
  • the components shown as modules may or may not be physical modules, that is, they may be located in One place, or it can be distributed across multiple networks. Some or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment. Persons of ordinary skill in the art can understand and implement the method without any creative effort.
  • this embodiment of the present application also provides an electronic device, including:
  • memory for storing instructions executable by the processor
  • the processor is configured to execute the instructions to implement the speech recognition method as described above.
  • an embodiment of the present application also provides a computer-readable storage medium.
  • the computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
  • this embodiment of the present application further provides a computer program product, including a computer program or instructions, which implement the speech recognition method as described above when executed by a processor.
  • FIG. 8 is a block diagram of an electronic device 800 provided by an embodiment of the present application.
  • the electronic device 800 may be a mobile terminal or a server.
  • the electronic device 800 is a mobile terminal as an example for description.
  • the electronic device 800 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, etc.
  • the electronic device 800 may include one or more of the following components: a processing component 802 , a memory 804 , a power component 806 , a multimedia component 808 , an audio component 810 , an input/output (I/O) interface 812 , and a sensor component 814 , and communication component 816.
  • a processing component 802 a memory 804 , a power component 806 , a multimedia component 808 , an audio component 810 , an input/output (I/O) interface 812 , and a sensor component 814 , and communication component 816.
  • Processing component 802 generally controls the overall operations of electronic device 800, such as operations associated with display, phone calls, data communications, camera operations, and recording operations.
  • the processing component 802 may include one or more processors 820 to execute instructions to complete all or part of the steps of the above method.
  • processing component 802 may include one or more modules that facilitate interaction between processing component 802 and other components.
  • processing component 802 may include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
  • Memory 804 is configured to store various types of data to support operations at device 800 . Examples of such data include instructions for any application or method operating on the electronic device 800, contact data, phonebook data, messages, pictures, videos, etc.
  • Memory 804 may be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EEPROM), Programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read-only memory
  • EEPROM erasable programmable read-only memory
  • EPROM Programmable read-only memory
  • PROM programmable read-only memory
  • ROM read-only memory
  • magnetic memory flash memory, magnetic or optical disk.
  • Power supply component 806 provides power to various components of electronic device 800 .
  • Power supply components 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power to electronic device 800 .
  • Multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user.
  • the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touches, swipes, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide action.
  • multimedia component 808 includes a front-facing camera and/or a rear-facing camera.
  • the front camera and/or the rear camera may receive external multimedia data.
  • Each front-facing camera and rear-facing camera can be a fixed optical lens system or have a focal length and optical zoom capabilities.
  • Audio component 810 is configured to output and/or input audio signals.
  • audio component 810 includes a microphone (MIC) configured to receive external audio signals when electronic device 800 is in operating modes, such as call mode, recording mode, and voice recognition mode. The received audio signal may be further stored in memory 804 or sent via communication component 816 .
  • audio component 810 also includes a speaker for outputting audio signals.
  • the I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module, which may be a keyboard, a click wheel, a button, etc. These buttons may include, but are not limited to: Home button, Volume buttons, Start button, and Lock button.
  • Sensor component 814 includes one or more sensors for providing various aspects of status assessment for electronic device 800 .
  • the sensor component 814 can detect the open/closed state of the device 800, the relative positioning of components, such as the display and keypad of the electronic device 800.
  • the sensor component 814 can also detect the electronic device 800 or a component of the electronic device 800. changes in position, the presence or absence of user contact with the electronic device 800 , the orientation or acceleration/deceleration of the electronic device 800 and changes in the temperature of the electronic device 800 .
  • Sensor assembly 814 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • Sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • Communication component 816 is configured to facilitate wired or wireless communication between electronic device 800 and other devices.
  • the electronic device 800 can access a wireless network based on a communication standard, such as WiFi, a carrier network (such as 2G, 3G, 4G or 5G), or a combination thereof.
  • the communication component 816 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel.
  • the communications component 816 also includes a near field communications (NFC) module to facilitate short-range communications.
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • electronic device 800 may be configured by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gates Array (FPGA), controller, microcontroller, microprocessor or other electronic components are implemented for executing the speech recognition method shown above.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field programmable gates Array
  • controller microcontroller, microprocessor or other electronic components are implemented for executing the speech recognition method shown above.
  • a computer-readable storage medium such as a memory 804 including instructions, and the instructions can be executed by the processor 820 of the electronic device 800 to complete the speech recognition method shown above.
  • the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
  • a computer program product is also provided.
  • the instructions in the computer program product are executed by the processor 820 of the electronic device 800, the electronic device 800 performs the speech recognition method shown above.
  • FIG. 9 is a block diagram of a device 900 for speech recognition provided by an embodiment of the present application.
  • device 900 may be provided as a server.
  • apparatus 900 includes a processing component 922, which further includes one or more processors, and memory resources represented by memory 932 for storing instructions, such as application programs, executable by processing component 922.
  • the application program stored in memory 932 may include one or more modules, each corresponding to a set of instructions.
  • the processing component 922 is configured to execute instructions to perform the above-described method.
  • Device 900 may also include a power supply component 926 configured to perform power management of device 900, a wired or wireless network interface 950 configured to connect device 900 to a network, and an input-output (I/O) interface 958.
  • Device 900 may operate based on an operating system stored in memory 932, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or the like.

Abstract

A voice recognition method and apparatus, an electronic device, a storage medium, and a product. The method comprises: in response to voice information of a user in a vehicle, obtaining a facial image of the user; determining a current state of the user according to a facial feature on the facial image; when the current state of the user satisfies a set condition, recognizing the voice information to obtain a recognition result; and when the recognition result is an in-vehicle infotainment instruction, executing a corresponding operation according to the in-vehicle infotainment instruction.

Description

语音识别方法、装置、电子设备、存储介质及产品Speech recognition methods, devices, electronic equipment, storage media and products
本申请要求在2022年6月1日提交中国专利局、申请号为202210617455.2、发明名称为“语音识别方法、装置、电子设备、存储介质及产品”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application requires the priority of the Chinese patent application submitted to the China Patent Office on June 1, 2022, with the application number 202210617455.2 and the invention title "Speech Recognition Methods, Devices, Electronic Equipment, Storage Media and Products", and its entire content has been approved. This reference is incorporated into this application.
技术领域Technical field
本申请涉及语音理解技术领域,尤其涉及一种语音识别方法、装置、电子设备、计算机可读存储介质及计算机程序产品。The present application relates to the field of speech understanding technology, and in particular to a speech recognition method, device, electronic equipment, computer-readable storage medium and computer program product.
背景技术Background technique
随着智能汽车的快速发展,语音功能是智能汽车的一个重要功能。用户每次和车机交流的时候都需要使用唤醒词,比如,“用户说:你好哪吒”,从而唤醒车机的语音功能,由于每次都要使用唤醒词,所以会比较麻烦。With the rapid development of smart cars, voice function is an important function of smart cars. The user needs to use the wake-up word every time he communicates with the car, for example, "The user says: Hello Nezha" to wake up the voice function of the car. Since the wake-up word has to be used every time, it will be more troublesome.
基于此,相关技术中提出了“免唤醒”方案,但是,在“免唤醒”方案中,对于用户在车中说话时,车机并不能准确的判断出哪些话是对“车机的指令”,哪些话“不是对车机的指令”。从而就会造成“误召回”,导致车机错误的执行指令,从而影响用户体验。Based on this, a "wake-free" solution has been proposed in related technologies. However, in the "wake-free" solution, when the user speaks in the car, the car computer cannot accurately determine which words are "instructions to the car computer" , which words "are not instructions to the vehicle." This will cause a "false recall", causing the vehicle to incorrectly execute instructions, thus affecting the user experience.
因此,在检测到车内用户的语音时,如何准确的识别出哪些语音是车机的指令,降低车机的误操作率是目前有待解决的技术问题。Therefore, when detecting the voice of a user in the car, how to accurately identify which voices are instructions for the car and reduce the misoperation rate of the car is a technical problem that needs to be solved.
概述Overview
本申请提供一种语音识别方法、装置、电子设备、计算机可读存储介质及计算机程序产品,以至少解决相关技术中由于对车内语音不能准确的识别出车机指令,导致车机执行错误的指令,增加误操作率的的技术问题。本申请的技术方案如下:This application provides a speech recognition method, device, electronic equipment, computer-readable storage medium and computer program product, to at least solve the problem in related technologies that the vehicle-machine instructions cannot be accurately recognized due to the inability to accurately recognize the voice in the car, resulting in vehicle-machine execution errors. instructions, technical problems that increase the rate of misoperation. The technical solution of this application is as follows:
根据本申请实施例的第一方面,提供一种语音识别方法,包括:According to a first aspect of the embodiment of the present application, a speech recognition method is provided, including:
响应车内用户的语音信息,获取所述用户的面部图像;Respond to the voice information of the user in the car and obtain the facial image of the user;
根据所述面部图像上的面部特征确定所述用户的当前状态;Determine the current status of the user based on facial features on the facial image;
在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果;When the user's current status meets the set conditions, recognize the voice information and obtain a recognition result;
在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。When the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction.
可选的,所述方法还包括:Optionally, the method also includes:
在所述用户的当前状态不满足设定条件时,拒绝对所述语音信息进行识别。When the current status of the user does not meet the set conditions, the recognition of the voice information is refused.
可选的,所述根据所述面部图像上的面部特征确定所述用户的当前状态,至少包括下述一种:Optionally, determining the current status of the user based on facial features on the facial image includes at least one of the following:
获取车辆的信息状态,基于所述信息状态和所述面部图像的面部特征判定车载蓝牙 电话没有开启时,确定所述用户处于非打电话状态;Obtain the information status of the vehicle, and determine that the vehicle-mounted Bluetooth phone is not turned on based on the information status and the facial features of the facial image, and determine that the user is in a non-calling state;
在根据所述面部图像的面部特征判定所述用户的正脸看向车辆行驶方向时,确定所述用户处于正脸看向前的状态;When it is determined based on the facial features of the facial image that the user's front face is looking in the direction of vehicle travel, it is determined that the user is in a state of facing forward;
在根据所述面部图像的面部特征判定所述用户的嘴巴处于张合状态时,确定所述用户处于说话状态。When it is determined that the user's mouth is in an opening and closing state based on the facial features of the facial image, it is determined that the user is in a speaking state.
可选的,所述在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果,包括:Optionally, when the current status of the user meets the set conditions, the voice information is recognized and the recognition result is obtained, including:
在所述用户的当前状态为:所述用户处于非打电话状态、用户处于正脸向前看的状态和用户处于说话状态的至少一种时,确定所述用户满足设定条件;When the current state of the user is at least one of: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state, it is determined that the user satisfies the set condition;
对所述语音信息进行识别,得到识别结果。The voice information is recognized and a recognition result is obtained.
可选的,所述对所述语音信息进行识别,得到识别结果,包括:Optionally, the recognition of the voice information to obtain the recognition result includes:
将所述语音信息进行本地语音文字转换处理,得到转换后的文本信息;或者Perform local speech-to-text conversion processing on the voice information to obtain converted text information; or
将所述语音信息发送给云端,由所述云端进行语音文字转换处理后得到文本信息;Send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information;
接收所述云端发送的转换后的文本信息。Receive the converted text information sent by the cloud.
可选的,所述在所述识别结果为车机指令时,按照所述车机指令执行对应的操作,包括:Optionally, when the recognition result is a vehicle-machine instruction, performing corresponding operations according to the vehicle-machine instruction includes:
将所述识别结果通过训练好的车机指令识别模型进行判断,得到所述识别结果是车机指令;其中,所述训练好的车机指令识别模型是基于人与车机交互的多个历史音频对,文本对,以及场景和关键词进行学习训练得到的模型;The recognition result is judged by a trained vehicle-machine command recognition model, and the recognition result is a vehicle-machine command; wherein the trained vehicle-machine command recognition model is based on multiple histories of human-vehicle interaction A model obtained by learning and training audio pairs, text pairs, scenes and keywords;
按照得到的所述车机指令执行对应的操作。Perform corresponding operations according to the obtained vehicle and machine instructions.
根据本申请实施例的第二方面,提供一种语音识别装置,包括:According to a second aspect of the embodiment of the present application, a speech recognition device is provided, including:
获取模块,用于响应车内用户的语音信息,获取所述用户的面部图像;The acquisition module is used to respond to the voice information of the user in the car and obtain the facial image of the user;
确定模块,用于根据所述面部图像上的面部特征确定所述用户的当前状态;a determining module configured to determine the current status of the user based on facial features on the facial image;
识别模块,用于在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果;A recognition module, used to recognize the voice information and obtain a recognition result when the current status of the user meets the set conditions;
执行模块,在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。An execution module, when the recognition result is a vehicle-machine instruction, executes the corresponding operation according to the vehicle-machine instruction.
可选的,所述装置还包括:Optionally, the device also includes:
拒识别模块,用于在所述用户的当前状态不满足设定条件时,拒绝对所述语音信息进行识别。A recognition rejection module is configured to reject recognition of the voice information when the user's current status does not meet the set conditions.
可选的,所述确定模块至少包括下述一个模块:Optionally, the determination module includes at least one of the following modules:
第一确定模块,用于基于获取车辆的信息状态和所述面部图像的面部特征判定车载蓝牙电话没有开启时,确定所述用户处于非打电话状态;A first determination module configured to determine that the user is in a non-phone state when determining that the vehicle-mounted Bluetooth phone is not turned on based on the obtained information status of the vehicle and the facial features of the facial image;
第二确定模块,用于在根据所述面部图像的面部特征判定所述用户的正脸看向车辆行驶方向时,确定所述用户处于正脸看向前的状态;a second determination module, configured to determine that the user is in a state of facing forward when it is determined based on the facial features of the facial image that the user's front face is looking in the direction of vehicle travel;
第三确定模块,用于在根据所述面部图像的面部特征判定所述用户的嘴巴处于张合状态时,确定所述用户处于说话状态。The third determination module is configured to determine that the user is in a speaking state when it is determined that the user's mouth is in an opening and closing state based on the facial features of the facial image.
可选的,所述识别模块包括:Optionally, the identification module includes:
第一判断模块,用于在所述用户的当前状态为:所述用户处于非打电话状态、用户处于正脸向前看的状态和用户处于说话状态的至少一种时,判定满足设定条件;The first judgment module is used to determine that the set conditions are met when the current state of the user is at least one of: the user is not on the phone, the user is in a state of facing forward, and the user is in a speaking state. ;
语音识别模块,用于对所述语音信息进行识别,得到识别结果。A speech recognition module is used to recognize the speech information and obtain a recognition result.
可选的,所述语音识别模块包括:语音转换模块;和/或发送模块和接收模块,其中,Optionally, the speech recognition module includes: a speech conversion module; and/or a sending module and a receiving module, wherein,
所述语音转换模块,用于将所述语音信息进行本地语音文字转换处理,得到转换后的文本信息;The voice conversion module is used to perform local voice-to-text conversion processing on the voice information to obtain converted text information;
所述发送模块,用于将所述语音信息发送给云端,由所述云端进行语音文字转换处理后得到文本信息;The sending module is used to send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information;
所述接收模块,用于接收所述云端发送的转换后的文本信息。The receiving module is used to receive the converted text information sent by the cloud.
可选的,所述执行模块包括:Optionally, the execution module includes:
第二判断模块,用于将所述识别结果通过训练好的车机指令识别模型进行判断,得到所述识别结果是车机指令;其中,所述训练好的车机指令识别模型是基于人与车机交互的多个历史音频对,文本对,以及场景和关键词进行学习训练得到的模型;The second judgment module is used to judge the recognition result through a trained vehicle-machine command recognition model, and obtain that the recognition result is a vehicle-machine command; wherein the trained vehicle-machine command recognition model is based on human and computer commands. A model obtained by learning and training multiple historical audio pairs, text pairs, scenes and keywords of vehicle-computer interaction;
指令执行模块,用于按照第二判断模块得到的所述车机指令执行对应的操作。An instruction execution module is used to execute corresponding operations according to the vehicle-machine instructions obtained by the second judgment module.
根据本申请实施例的第三方面,提供一种电子设备,包括:According to a third aspect of the embodiment of the present application, an electronic device is provided, including:
处理器;processor;
用于存储所述处理器可执行指令的存储器;memory for storing instructions executable by the processor;
其中,所述处理器被配置为执行所述指令,以实现如上所述的语音识别方法。Wherein, the processor is configured to execute the instructions to implement the speech recognition method as described above.
根据本申请实施例的第四方面,提供一种计算机可读存储介质,当所述计算机可读存储介质中的指令由电子设备的处理器执行时,使得电子设备能够执行如上所述的语音识别方法。According to a fourth aspect of an embodiment of the present application, a computer-readable storage medium is provided, which when instructions in the computer-readable storage medium are executed by a processor of an electronic device, enables the electronic device to perform speech recognition as described above. method.
根据本申请实施例的第五方面,提供一种计算机程序产品,包括计算机程序或指令,所述计算机程序或指令被处理器执行时实现如上所述的语音识别方法。According to a fifth aspect of the embodiments of the present application, a computer program product is provided, including a computer program or instructions that implement the speech recognition method as described above when executed by a processor.
本申请的实施例提供的技术方案至少带来以下有益效果:The technical solutions provided by the embodiments of the present application at least bring the following beneficial effects:
本申请实施例中,响应车内用户的语音信息,获取所述用户的面部图像;根据所述面部图像上的面部特征确定所述用户的当前状态;在所述用户的当前状态满足设定条件 时,对所述语音信息进行识别,得到识别结果;在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。也就是说,本申请实施例中,根据面部图像上的面部特征来确定所述用户的当前状态,基于的用户的当前对语音信息进行识别,进而可以准确的判断出哪那些语音信息是车机指令,哪些语音信息不是车机指令,提高了车机准确执行车机指令的效率,降低车机误操作率,也提升了用户体验。In the embodiment of the present application, in response to the voice information of the user in the car, the user's facial image is obtained; the current state of the user is determined based on the facial features on the facial image; and the user's current state satisfies the set conditions When, the voice information is recognized to obtain a recognition result; when the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction. That is to say, in the embodiment of the present application, the current status of the user is determined based on the facial features on the facial image, and the voice information is recognized based on the user's current status, so that it can be accurately determined which voice information is from the vehicle. Commands, which voice messages are not vehicle-machine instructions, improve the efficiency of the vehicle-machine accurately executing vehicle-machine instructions, reduce the rate of vehicle-machine misoperation, and also improve the user experience.
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,并不能限制本申请。It should be understood that the above general description and the following detailed description are only exemplary and explanatory, and do not limit the present application.
上述说明仅是本申请技术方案的概述,为了能够更清楚了解本申请的技术手段,而可依照说明书的内容予以实施,并且为了让本申请的上述和其它目的、特征和优点能够更明显易懂,以下特举本申请的具体实施方式。The above description is only an overview of the technical solutions of the present application. In order to have a clearer understanding of the technical means of the present application, they can be implemented according to the content of the description, and in order to make the above and other purposes, features and advantages of the present application more obvious and understandable. , the specific implementation methods of the present application are specifically listed below.
附图说明Description of the drawings
此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本申请的实施例,并与说明书一起用于解释本申请的原理,并不构成对本申请的不当限定。为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。The drawings herein are incorporated into the specification and constitute a part of the specification, illustrate embodiments consistent with the present application, and are used together with the description to explain the principles of the present application, and do not constitute undue limitations on the present application. In order to more clearly explain the embodiments of the present application or the technical solutions in the prior art, the following will briefly introduce the drawings that need to be used in the description of the embodiments or the prior art. Obviously, the drawings in the following description These are some embodiments of the present application. For those of ordinary skill in the art, other drawings can be obtained based on these drawings without exerting creative efforts.
图1是本申请实施例提供的一种语音识别方法的流程图。Figure 1 is a flow chart of a speech recognition method provided by an embodiment of the present application.
图2是本申请实施例提供的一种语音识别方法的应用实例的流程图。Figure 2 is a flow chart of an application example of a speech recognition method provided by an embodiment of the present application.
图3是本申请实施例提供的一种语音识别装置的框图。Figure 3 is a block diagram of a speech recognition device provided by an embodiment of the present application.
图4是本申请实施例提供的一种语音识别装置的另一框图。Figure 4 is another block diagram of a speech recognition device provided by an embodiment of the present application.
图5是本申请实施例提供的一种确定模块的框图。Figure 5 is a block diagram of a determination module provided by an embodiment of the present application.
图6是本申请实施例提供的一种识别模块的框图。Figure 6 is a block diagram of an identification module provided by an embodiment of the present application.
图7是本申请实施例提供的一种执行模块的框图。Figure 7 is a block diagram of an execution module provided by an embodiment of the present application.
图8是本申请实施例提供的一种电子设备的框图。Figure 8 is a block diagram of an electronic device provided by an embodiment of the present application.
图9是本申请实施例提供的一种用于语音识别的装置的框图。Figure 9 is a block diagram of a device for speech recognition provided by an embodiment of the present application.
详细描述A detailed description
为了使本领域普通人员更好地理解本申请的技术方案,下面将结合附图,对本申请实施例中的技术方案进行清楚、完整地描述。In order to enable ordinary people in the art to better understand the technical solutions of the present application, the technical solutions in the embodiments of the present application will be clearly and completely described below in conjunction with the accompanying drawings.
需要说明的是,本申请的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的本申请的实施例能够以除了在这里 图示或描述的那些以外的顺序实施。以下示例性实施例中所描述的实施方式并不代表与本申请相一致的所有实施方式。相反,它们仅是与如所附权利要求书中所详述的、本申请的一些方面相一致的装置和方法的例子。It should be noted that the terms "first", "second", etc. in the description and claims of this application and the above-mentioned drawings are used to distinguish similar objects and are not necessarily used to describe a specific order or sequence. It is to be understood that the data so used are interchangeable under appropriate circumstances so that the embodiments of the application described herein can be practiced in sequences other than those illustrated or described herein. The implementations described in the following exemplary embodiments do not represent all implementations consistent with this application. Rather, they are merely examples of apparatus and methods consistent with aspects of the application as detailed in the appended claims.
近年来,基于人工智能的计算机视觉、深度学习、机器学习、图像处理、图像识别等技术研究取得了重要进展。人工智能(Artificial Intelligence,AI)是研究、开发用于模拟、延伸人的智能的理论、方法、技术及应用系统的新兴科学技术。人工智能学科是一门综合性学科,涉及芯片、大数据、云计算、物联网、分布式存储、深度学习、机器学习、神经网络等诸多技术种类。计算机视觉作为人工智能的一个重要分支,具体是让机器识别世界,计算机视觉技术通常包括人脸识别、活体检测、指纹识别与防伪验证、生物特征识别、人脸检测、行人检测、目标检测、行人识别、图像处理、图像识别、图像语义理解、图像检索、文字识别、视频处理、视频内容识别、行为识别、三维重建、虚拟现实、增强现实、同步定位与地图构建(SLAM)、计算摄影、机器人导航与定位等技术。随着人工智能技术的研究和进步,该项技术在众多领域展开了应用,例如安防、城市管理、交通管理、楼宇管理、园区管理、人脸通行、人脸考勤、物流管理、仓储管理、机器人、智能营销、计算摄影、手机影像、云服务、智能家居、穿戴设备、无人驾驶、自动驾驶、智能医疗、人脸支付、人脸解锁、指纹解锁、人证核验、智慧屏、智能电视、摄像机、移动互联网、网络直播、美颜、美妆、医疗美容、智能测温等领域。In recent years, research on computer vision, deep learning, machine learning, image processing, image recognition and other technologies based on artificial intelligence has made important progress. Artificial Intelligence (AI) is an emerging science and technology that studies and develops theories, methods, technologies and application systems for simulating and extending human intelligence. The subject of artificial intelligence is a comprehensive subject, involving many types of technologies such as chips, big data, cloud computing, Internet of Things, distributed storage, deep learning, machine learning, neural networks, etc. Computer vision, as an important branch of artificial intelligence, specifically allows machines to recognize the world. Computer vision technology usually includes face recognition, live body detection, fingerprint recognition and anti-counterfeiting verification, biometric recognition, face detection, pedestrian detection, target detection, pedestrian Recognition, image processing, image recognition, image semantic understanding, image retrieval, text recognition, video processing, video content recognition, behavior recognition, 3D reconstruction, virtual reality, augmented reality, simultaneous localization and mapping (SLAM), computational photography, robotics Navigation and positioning technologies. With the research and progress of artificial intelligence technology, this technology has been applied in many fields, such as security, urban management, traffic management, building management, park management, face traffic, face attendance, logistics management, warehousing management, robots , intelligent marketing, computational photography, mobile imaging, cloud services, smart home, wearable devices, driverless driving, autonomous driving, smart medical care, face payment, face unlocking, fingerprint unlocking, ID verification, smart screen, smart TV, Cameras, mobile Internet, webcasting, beauty, cosmetics, medical cosmetology, intelligent temperature measurement and other fields.
图1是本申请实施例提供的一种语音识别方法的流程图,如图1所示,该语音识别方法包括以下步骤:Figure 1 is a flow chart of a speech recognition method provided by an embodiment of the present application. As shown in Figure 1, the speech recognition method includes the following steps:
步骤101:响应车内用户的语音信息,获取所述用户的面部图像。Step 101: Respond to the voice message of the user in the car and obtain the user's facial image.
步骤102:根据所述面部图像上的面部特征确定所述用户的当前状态。Step 102: Determine the current status of the user based on facial features on the facial image.
步骤103:在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果。Step 103: When the user's current status meets the set conditions, recognize the voice information and obtain a recognition result.
步骤104:在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。Step 104: When the recognition result is a vehicle-machine instruction, perform the corresponding operation according to the vehicle-machine instruction.
本申请所述的语音识别方法可以应用于车机终端等,在此不作限制,其车机终端实施设备可以是智能车机,车机平台等等电子设备,在此不作限制。The speech recognition method described in this application can be applied to vehicle-machine terminals, etc., and is not limited here. The vehicle-machine terminal implementation equipment can be electronic equipment such as smart car-machine, vehicle-machine platform, etc., which is not limited here.
下面结合图1,对本申请实施例提供的一种语音识别方法的具体实施步骤进行详细说明。The specific implementation steps of a speech recognition method provided by an embodiment of the present application will be described in detail below with reference to Figure 1 .
在步骤101中,响应车内用户的语音信息,获取所述用户的面部图像。In step 101, in response to the voice information of the user in the car, the facial image of the user is obtained.
该步骤中,在车内的用户说话时,车机终端可以通过车辆上的麦克风检测到用户的语音信息,此时,可以通过车辆上的图像采集设备(比如摄像头等)获取车内用户的面部图像,该面部图像可以是一帧图像,也可以是多帧图像。其中,该图像采集设备可以 设置在对准驾驶员的位置,以便于图像采集设备能清晰的采集到驾驶员的面部图像。In this step, when the user in the car speaks, the car terminal can detect the user's voice information through the microphone on the vehicle. At this time, the face of the user in the car can be obtained through the image collection device (such as a camera, etc.) on the vehicle. Image, the facial image can be a single frame image or multiple frame images. Wherein, the image acquisition device can be set at a position aimed at the driver, so that the image acquisition device can clearly capture the driver's facial image.
在步骤102中,根据所述面部图像上的面部特征确定所述用户的当前状态。In step 102, the current status of the user is determined based on facial features on the facial image.
该步骤中,对获取到的面部图像进行识别,得到该面部图像上的面部特征点,根据该面部特征点确定该用户面部的当前状态。其中,对面部图像的识别,是利用计算机图像处理技术从面部图像中提取人像面部的特征点,比如,提取眼睛是否睁开,嘴巴是否张开等等。In this step, the acquired facial image is recognized, facial feature points on the facial image are obtained, and the current state of the user's face is determined based on the facial feature points. Among them, facial image recognition uses computer image processing technology to extract facial feature points from facial images, such as whether the eyes are open, whether the mouth is open, etc.
之后,根据面部特征确定当前用户的面部状态可以包括下述至少一种,但并不限于此:Afterwards, determining the current user's facial state based on facial features may include at least one of the following, but is not limited to this:
1)获取车辆的信息状态,基于所述信息状态和所述面部图像的面部特征判定车载蓝牙电话没有开启时,确定所述用户处于非打电话状态。1) Obtain the information status of the vehicle, and determine that the user is in a non-phone state when it is determined that the vehicle-mounted Bluetooth phone is not turned on based on the information status and the facial features of the facial image.
也就是说,先获取车辆的信息状态,比如,车内的蓝牙电话(即车载蓝牙电话)是否开启等,然后,再结合面部图像中的面部特征(比如嘴巴是否张合等)来判断该用户是否在打电话。比如,如果车载蓝牙电话处于打开状态,且该用户的嘴巴处于张合状态,则确定该用户此时正在打电话状态;否则,判断用户处于非打电话状态,比如车载蓝牙电话处于非打开状态,且该用户的嘴巴处于张合状态,则确定该用户处于说话状态,而并非是该用户处于非打电话状态;当然,如果车载蓝牙电话处于非打开状态,且该用户的嘴巴处于闭合状态,则确定该用户没有说话,处于安静状态等。That is to say, first obtain the information status of the vehicle, such as whether the Bluetooth phone in the car (i.e., the car Bluetooth phone) is turned on, etc., and then combine the facial features in the facial image (such as whether the mouth is open and closed, etc.) to determine the user Are you on the phone? For example, if the car Bluetooth phone is in the open state and the user's mouth is opening and closing, it is determined that the user is making a call at this time; otherwise, it is determined that the user is in a non-call state, for example, the car Bluetooth phone is in a non-open state. And the user's mouth is in an open and closed state, then it is determined that the user is in a speaking state, not that the user is in a non-calling state; of course, if the car Bluetooth phone is in a non-open state and the user's mouth is in a closed state, then Make sure the user is not speaking, is in a quiet state, etc.
2)在根据所述面部图像的面部特征判定所述用户的双目看向车辆行驶方向时,确定所述用户处于正面看向前的状态。2) When it is determined that the user's eyes are looking in the vehicle traveling direction based on the facial features of the facial image, it is determined that the user is in a state of looking forward.
该步骤中,可以通过多角度人脸识别技术来判断用户的正脸是否看向车辆行驶方向,如果是,则确定该用户处于正脸向前看的状态,否则,判断该用户处于非正脸向前看的状态。也就是判断用户的正脸是否向车辆行驶方向的九十度范围内看,如果是,则确定正脸处于向前的状态。In this step, multi-angle face recognition technology can be used to determine whether the user's front face is looking in the direction of the vehicle. If so, it is determined that the user is facing forward. Otherwise, it is determined that the user is not facing forward. The state of looking forward. That is to say, it is determined whether the user's front face is looking within 90 degrees of the vehicle's driving direction. If so, it is determined that the front face is facing forward.
该实施例中,多角度人脸识别技术是多姿态人脸识别技术的分支。其中,一种深度学习的多角度人脸识别算法包括:首先,构建深度学习训练数据集,其次,训练一个深度人脸分类器;最后,应用分类器进行人脸检测。其具体的的实现过程,对于本领域已是熟知技术,在不再赘述。In this embodiment, the multi-angle face recognition technology is a branch of the multi-pose face recognition technology. Among them, a deep learning multi-angle face recognition algorithm includes: first, constructing a deep learning training data set; second, training a deep face classifier; finally, applying the classifier for face detection. The specific implementation process is a well-known technology in this field and will not be described again.
也就是说,这种算法就是将人脸侧面图像作为输入,相应的人脸正面图像作为输出,监督模型学习出从不同姿态的人脸侧面图像到正面图像的映射,从而增加了识别中的有效面部信息。当然,在实际应用中,并不限于此,比如,还可以通过训练好的人脸角度分类模型判断用户是否向前看的正面角度,如果判断该用户正脸向前九十度范围内,都确定该用户正脸向前看。In other words, this algorithm takes the side image of the face as input and the corresponding frontal image of the face as the output. The supervised model learns the mapping from the side image of the face in different poses to the frontal image, thus increasing the effectiveness in recognition. facial information. Of course, in practical applications, it is not limited to this. For example, the trained face angle classification model can also be used to determine the frontal angle of whether the user is looking forward. If it is judged that the user's front face is within the range of 90 degrees forward, all Make sure the user is facing forward.
3)在根据所述面部图像判定所述用户的嘴巴处于张合状态时,确定所述用户处于说话状态。3) When it is determined that the user's mouth is in an opening and closing state based on the facial image, it is determined that the user is in a speaking state.
该步骤中,根据面部图像的面部特征判定所述用户的嘴巴是否处于张合状态,如果是,确定所述用户处于说话状态,否则确定该用户处于非说话状态。In this step, it is determined according to the facial features of the facial image whether the user's mouth is in an open and closed state. If so, it is determined that the user is in a speaking state; otherwise, it is determined that the user is in a non-speaking state.
具体的,可以通过唇动特征提取算法(或者是唇动模型)来判断用户是否开口,从而确定该用户是否唇动。当然,也可以基于唇动的说话用户识别技术,通过离散余弦变换,从说话用户讲话时的图像序列,提取既能反映说话人嘴部生理特性,又能反映了说话人唇动的行为特性的视觉特征,基于这些视觉特征,为说话用户建立静态与动态混合模型,用来判断用户是否发生唇动。其具体的过程,对于本领域技术人员来说,已是熟知技术,在此不再赘述。Specifically, the lip movement feature extraction algorithm (or lip movement model) can be used to determine whether the user opens his mouth, thereby determining whether the user has lip movement. Of course, speaking user identification technology based on lip movement can also be used to extract information that reflects both the physiological characteristics of the speaker's mouth and the behavioral characteristics of the speaker's lip movements from the image sequence of the speaking user through discrete cosine transformation. Visual features. Based on these visual features, a static and dynamic hybrid model is established for the speaking user to determine whether the user has lip movements. The specific process is a familiar technique to those skilled in the art and will not be described in detail here.
在步骤103中,在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果。In step 103, when the current status of the user meets the set conditions, the voice information is recognized and a recognition result is obtained.
该步骤中,在确定用户的当前状态后,需要判断该用户的当前状态是否处于满足设定条件,如果满足,则执行所述对所述语音信息进行识别的步骤,否则,则拒绝对所述语音信息进行识别,即拒识别。其中,设定条件至少包括下述一种:用户处于非打电话状态,用户处于正脸向前看的状态,以及用户处于说话状态等。在判断用户的当前状态满足上述至少一个设定条件时,就可以执行对该语音信息的识别。本实施例的最佳方式就是全部满足上述所有设定条件。In this step, after determining the current status of the user, it is necessary to determine whether the current status of the user satisfies the set conditions. If satisfied, the step of identifying the voice information is performed. Otherwise, the step of identifying the voice information is rejected. The voice information is recognized, that is, the recognition is refused. Among them, the setting conditions include at least one of the following: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state. When it is determined that the user's current status satisfies at least one of the above set conditions, the recognition of the voice information can be performed. The best way for this embodiment is to satisfy all the above setting conditions.
其中,再另一实施例中,对所述语音信息进行识别,得到识别结果,包括:In another embodiment, the voice information is recognized and the recognition results are obtained, including:
一种情况是,将所述语音信息进行本地语音文字转换处理,得到转换后的文本信息。In one case, the voice information is subjected to local voice-to-text conversion processing to obtain converted text information.
另一种情况是,将所述语音信息发送给云端,由所述云端进行语音文字转换处理后得到文本信息;接收所述云端发送的转换后的文本信息。Another situation is to send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information; and receives the converted text information sent by the cloud.
其具体的语音文字转换处理过程,对于本领域技术人员来说已是熟知技术,在此不再赘述。The specific speech-to-text conversion process is a familiar technology to those skilled in the art and will not be described in detail here.
在步骤104中,在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。In step 104, when the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction.
在对该语音信息进行识别后,将得到的识别结果输入到训练好的车机指令模型进行判断,该所述识别结果是否是车机指令。其中,所述训练好的车机指令识别模型是基于人与车机交互的多个历史音频对,文本对,以及场景和关键词进行学习训练得到的模型。After the voice information is recognized, the obtained recognition result is input into the trained vehicle-to-machine command model to determine whether the recognition result is a vehicle-to-machine command. Among them, the trained vehicle-machine command recognition model is a model obtained by learning and training based on multiple historical audio pairs, text pairs, scenes and keywords of human-vehicle interaction.
可选的,一种实施例中,预先对车机指令识别模型进行训练,其中,训练该车机指令识别模型的输入通常选取多次人与车机(简称人机)交互的历史对话音频等,比如选取10轮人机交互对话的音频、文本,并确认该用户是否对车机进行说话(或对车机说的车机指令)记录每一轮识别的结果;其车机指令识别模型的输出结果包括:1是对车机说 话,即是车机指令;0不是对车说话,即非车机指令;当然,也可以设置为:0是对车机说话,1不是对车机说话等,本实施例不做限制。Optionally, in one embodiment, the vehicle-machine command recognition model is trained in advance, wherein the input for training the vehicle-machine command recognition model usually selects historical dialogue audio of multiple human-vehicle-machine (referred to as human-machine) interactions, etc. , for example, select the audio and text of 10 rounds of human-computer interaction dialogues, and confirm whether the user speaks to the vehicle and the computer (or gives the vehicle and computer commands to the vehicle and the computer) and record the results of each round of recognition; its vehicle-to-machine command recognition model The output results include: 1 means speaking to the vehicle and the computer, that is, the vehicle and computer commands; 0 does not speak to the car, that is, it is not a vehicle and computer command; of course, it can also be set to: 0 means speaking to the vehicle and the computer, 1 does not speak to the vehicle and the computer, etc. , there is no limitation in this embodiment.
该实施例中,对车机指令识别模型的训练,就是让车机指令模型从中学习到更多的车机指令,来提高车机指令识别模型训练的精度。In this embodiment, the training of the vehicle-machine command recognition model is to allow the vehicle-machine command model to learn more vehicle-machine commands from it, thereby improving the accuracy of training the vehicle-machine command recognition model.
一种情况是,本实施例选取大量的数据组进行学习,每组数据包含:历史音频和当前音频中学习出哪种类型的音频是对车机说话的,即对车机发出的车机指令。In one case, this embodiment selects a large number of data groups for learning. Each group of data includes: historical audio and current audio to learn which type of audio speaks to the vehicle, that is, the vehicle instructions issued to the vehicle. .
另一种情况,本实施例还可以从文本中,学习出哪些文本是命令词,在命令词不够丰富的情况下,用历史结果作为本次的输入,从而提高车机指令识别模型训练的精度。In another case, this embodiment can also learn which texts are command words from the text. If the command words are not rich enough, use the historical results as the input this time, thereby improving the accuracy of the vehicle-machine command recognition model training. .
本申请实施例中,响应车内用户的语音信息,获取所述用户的面部图像;根据所述面部图像上的面部特征确定所述用户的当前状态;在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果;在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。也就是说,本申请实施例中,根据面部图像上的面部特征来确定所述用户的当前状态,基于的用户的当前对语音信息进行识别,进而可以准确的判断出哪那些语音信息是车机指令,哪些语音信息不是车机指令,提高了车机执行准确车机指令的效率,降低车机误操作率,也提升了用户体验。In the embodiment of the present application, in response to the voice information of the user in the car, the user's facial image is obtained; the current state of the user is determined based on the facial features on the facial image; and the user's current state satisfies the set conditions When, the voice information is recognized to obtain a recognition result; when the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction. That is to say, in the embodiment of the present application, the current status of the user is determined based on the facial features on the facial image, and the voice information is recognized based on the user's current status, so that it can be accurately determined which voice information is from the vehicle. Instructions, which voice messages are not vehicle-machine instructions, improve the efficiency of the vehicle-machine executing accurate vehicle-machine instructions, reduce the rate of vehicle-machine misoperation, and also improve the user experience.
还请参阅图2,为本申请实施例提供的一种语音识别方法的应用实例图,所述方法应用于车机终端,所述方法包括:Please also refer to Figure 2, which is an application example diagram of a speech recognition method provided by an embodiment of the present application. The method is applied to a vehicle-machine terminal. The method includes:
步骤201:检测到车内用户的语音信息;Step 201: Detect the voice information of the user in the car;
该步骤中,当车内有用户说话时,车机终端检测到该用户的语音信息。In this step, when a user speaks in the car, the car terminal detects the user's voice information.
步骤202:获取所述用户的面部图像;Step 202: Obtain the user's facial image;
步骤203:根据所述面部图像上的面部特征确定所述用户的当前状态;Step 203: Determine the current status of the user based on the facial features on the facial image;
其中,所述用户的当前状态以包括:用户处于非打电话状态,用户处于正脸看向前方的状态和所述用户处于说话状态为例,但在实际应用中,并不限于此。For example, the current state of the user includes: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state, but in practical applications, it is not limited to this.
步骤204:判断所述用户的当前状态是否在打电话,如果否,执行步骤205;否则,执行步骤210:Step 204: Determine whether the user is currently on the phone. If not, perform step 205; otherwise, perform step 210:
步骤205:判断所述用户的当前状态是否处于正脸看向车辆行驶方向,如果是,执行步骤206;否则,执行步骤210:Step 205: Determine whether the current state of the user is facing the direction of the vehicle. If so, perform step 206; otherwise, perform step 210:
步骤206:判断所述用户的当前状态是否嘴巴处于张合状态,如果是,执行步骤207;否则,执行步骤210:Step 206: Determine whether the current state of the user is in an open and closed state. If so, perform step 207; otherwise, perform step 210:
步骤207:对所述语音信息进行识别,得到识别结果;Step 207: Recognize the voice information and obtain the recognition result;
步骤208:判断所述识别结果是否为车机指令,如果是,执行步骤209;否则,执行步骤211;Step 208: Determine whether the recognition result is a vehicle-machine command. If so, execute step 209; otherwise, execute step 211;
步骤209:按照所述车机指令执行对应的操作;Step 209: Perform corresponding operations according to the vehicle and machine instructions;
步骤210:拒绝对所述语音信息进行识别,即拒识别。Step 210: Refuse to recognize the voice information, that is, reject recognition.
步骤211:拒绝执行所述识别结果。Step 211: Refuse to execute the identification result.
当然也可以删除或忽略该识别结果。Of course, the recognition result can also be deleted or ignored.
该实施例中,各个步骤的实现过程详见上述对应实施例的实现过程,在此不再赘述。In this embodiment, the details of the implementation process of each step can be found in the implementation process of the above corresponding embodiment, which will not be described again here.
本申请实施例中,根据面部图像上的面部特征来确定所述用户的当前状态,基于的用户的当前对语音信息进行识别,进而可以准确的判断出哪那些语音信息是车机指令,哪些语音信息不是车机指令,即通过多模(比如视觉和音频等)去判断该语音信息是不是车机指令,提高了车机执行准确车机指令的效率,降低车机“误召回”的几率,也提升了用户体验。也即是说,本申请实施例利用车内的视觉系统和语音系统,降低车机“误召回”率,提升了用户体验。In the embodiment of the present application, the current status of the user is determined based on the facial features on the facial image, and the voice information is recognized based on the user's current status, so that it can be accurately determined which voice information is a vehicle-machine command and which voice information is The information is not a vehicle-machine instruction, that is, multi-mode (such as visual and audio, etc.) is used to determine whether the voice message is a vehicle-machine instruction, which improves the efficiency of the vehicle-machine executing accurate vehicle-machine instructions and reduces the probability of "false recall" of the vehicle-machine. It also improves user experience. That is to say, the embodiment of the present application uses the visual system and voice system in the car to reduce the "false recall" rate of the car and improve the user experience.
需要说明的是,对于方法实施例,为了简单描述,故将其都表述为一系列的动作组合,但是本领域技术人员应该知悉,本实施公开并不受所描述的动作顺序的限制,因为依据本申请,某些步骤可以采用其他顺序或者同时进行。其次,本领域技术人员也应该知悉,说明书中所描述的实施例均属于优选实施例,所涉及的动作并不一定是本申请所必须的。It should be noted that for the sake of simple description, method embodiments are expressed as a series of action combinations. However, those skilled in the art should know that this implementation disclosure is not limited by the described action sequence, because according to In this application, certain steps may be performed in other orders or simultaneously. Secondly, those skilled in the art should also know that the embodiments described in the specification are preferred embodiments, and the actions involved are not necessarily necessary for this application.
图3是本申请实施例提供的一种语音识别装置的框图。参照图3,该装置包括:获取模块301,确定模块302,识别模块303和执行模块304,其中,Figure 3 is a block diagram of a speech recognition device provided by an embodiment of the present application. Referring to Figure 3, the device includes: an acquisition module 301, a determination module 302, an identification module 303 and an execution module 304, where,
该获取模块301,用于响应车内用户的语音信息,获取所述用户的面部图像;The acquisition module 301 is used to respond to the voice information of the user in the car and acquire the facial image of the user;
该确定模块302,用于根据所述面部图像上的面部特征确定所述用户的当前状态;The determination module 302 is used to determine the current status of the user according to the facial features on the facial image;
该识别模块303,用于在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果;The recognition module 303 is used to recognize the voice information and obtain a recognition result when the user's current status meets the set conditions;
该执行模块304,在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。The execution module 304, when the recognition result is a vehicle-machine instruction, performs corresponding operations according to the vehicle-machine instruction.
可选的,在另一实施例中,该实施例在上述实施例的基础上,所述装置还包括:拒识别模块401,其结构框图如图4所示,其中,Optionally, in another embodiment, based on the above embodiment, the device further includes: a rejection identification module 401, the structural block diagram of which is shown in Figure 4, wherein,
该拒识别模块401,用于在所述用户的当前状态不满足设定条件时,拒绝对所述语音信息进行识别。The recognition rejection module 401 is used to reject recognition of the voice information when the user's current status does not meet the set conditions.
可选的,在另一实施例中,该实施例在上述实施例的基础上,所述确定模块302至少包括下述一个模块:第一确定模块501,第二确定模块502和第三确定模块503,其结构框图如图5所示,其中,本实施例以同时包括所有模块为例:Optionally, in another embodiment, based on the above embodiment, the determination module 302 includes at least one of the following modules: a first determination module 501, a second determination module 502 and a third determination module. 503, the structural block diagram of which is shown in Figure 5, in which this embodiment includes all modules at the same time as an example:
该第一确定模块501,用于基于获取车辆的信息状态和所述面部图像的面部特征判 定车载蓝牙电话没有开启时,确定所述用户处于非打电话状态;The first determination module 501 is used to determine that the user is in a non-phone state when determining that the vehicle-mounted Bluetooth phone is not turned on based on the information status of the vehicle and the facial features of the facial image;
该第二确定模块502,用于在根据所述面部图像的面部特征判定所述用户的正脸看向车辆行驶方向时,确定所述用户处于正脸看向前的状态;The second determination module 502 is configured to determine that the user is in a state of facing forward when it is determined based on the facial features of the facial image that the user's front face is looking in the direction of vehicle travel;
该第三确定模块503,用于在根据所述面部图像的面部特征判定所述用户的嘴巴处于张合状态时,确定所述用户处于说话状态。The third determination module 503 is configured to determine that the user is in a speaking state when it is determined that the user's mouth is in an opening and closing state based on the facial features of the facial image.
可选的,在另一实施例中,该实施例在上述实施例的基础上,所述识别模块303包括:第一判断模块601和语音识别模块602,其结构框图如图6所示,其中,Optionally, in another embodiment, based on the above embodiment, the recognition module 303 includes: a first judgment module 601 and a speech recognition module 602, whose structural block diagram is shown in Figure 6, where ,
该第一判断模块601,用于在所述用户的当前状态为:所述用户处于非打电话状态、用户处于正脸向前看的状态和用户处于说话状态的至少一种时,判定满足设定条件;The first judgment module 601 is used to judge that when the current state of the user is at least one of: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state, the condition is satisfied. set conditions;
该语音识别模块602,用于对所述语音信息进行识别,得到识别结果。The speech recognition module 602 is used to recognize the speech information and obtain a recognition result.
可选的,在另一实施例中,该实施例在上述实施例的基础上,所述语音识别模块包括:语音转换模块;和/或发送模块和接收模块,其中,Optionally, in another embodiment, based on the above embodiment, the speech recognition module includes: a speech conversion module; and/or a sending module and a receiving module, wherein,
该语音转换模块,用于将所述语音信息进行本地语音文字转换处理,得到转换后的文本信息;The voice conversion module is used to perform local voice-to-text conversion processing on the voice information to obtain converted text information;
该发送模块,用于将所述语音信息发送给云端,由所述云端进行语音文字转换处理后得到文本信息;The sending module is used to send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information;
该接收模块,用于接收所述云端发送的转换后的文本信息。The receiving module is used to receive the converted text information sent by the cloud.
可选的,在另一实施例中,该实施例在上述实施例的基础上,所述执行模块304包括:第二判断模块701和指令执行模块702,其结构框图如图7所示,其中,Optionally, in another embodiment, based on the above embodiment, the execution module 304 includes: a second judgment module 701 and an instruction execution module 702, whose structural block diagram is shown in Figure 7, where ,
该第二判断模块701,用于将所述识别结果通过训练好的车机指令识别模型进行判断,得到所述识别结果是车机指令;其中,所述训练好的车机指令识别模型是基于人与车机交互的多个历史音频对,文本对,以及场景和关键词进行学习训练得到的模型;The second judgment module 701 is used to judge the recognition result through a trained vehicle-machine command recognition model, and obtain that the recognition result is a vehicle-machine command; wherein the trained vehicle-machine command recognition model is based on A model obtained by learning and training multiple historical audio pairs, text pairs, scenes and keywords of human-vehicle interaction;
该指令执行模块702,用于按照第二判断模块701得到的所述车机指令执行对应的操作。The instruction execution module 702 is used to execute corresponding operations according to the vehicle-machine instructions obtained by the second judgment module 701 .
关于上述实施例中的装置,其中各个模块执行操作的具体方式已经在有关该方法的实施例中进行了详细描述,此处将不做详细阐述说明。Regarding the devices in the above embodiments, the specific manner in which each module performs operations has been described in detail in the embodiments related to the method, and will not be described in detail here.
以上所描述的装置实施例仅仅是示意性的,其中所述作为分离部件说明的模块可以是或者也可以不是物理上分开的,作为模块显示的部件可以是或者也可以不是物理模块,即可以位于一个地方,或者也可以分布到多个网络上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。本领域普通技术人员在不付出创造性的劳动的情况下,即可以理解并实施。The device embodiments described above are only illustrative. The modules described as separate components may or may not be physically separated. The components shown as modules may or may not be physical modules, that is, they may be located in One place, or it can be distributed across multiple networks. Some or all of the modules can be selected according to actual needs to achieve the purpose of the solution of this embodiment. Persons of ordinary skill in the art can understand and implement the method without any creative effort.
可选的,本申请实施例还提供一种电子设备,包括:Optionally, this embodiment of the present application also provides an electronic device, including:
处理器;processor;
用于存储所述处理器可执行指令的存储器;memory for storing instructions executable by the processor;
其中,所述处理器被配置为执行所述指令,以实现如上所述的语音识别方法。Wherein, the processor is configured to execute the instructions to implement the speech recognition method as described above.
可选的,本申请实施例还一种计算机可读存储介质,当所述计算机可读存储介质中的指令由电子设备的处理器执行时,使得电子设备能够执行如上所述的语音识别方法。可选地,计算机可读存储介质可以是ROM、随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。Optionally, an embodiment of the present application also provides a computer-readable storage medium. When instructions in the computer-readable storage medium are executed by a processor of an electronic device, the electronic device can perform the speech recognition method as described above. Alternatively, the computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
可选的,本申请实施例还一种计算机程序产品,包括计算机程序或指令,所述计算机程序或指令被处理器执行时实现如上所述的语音识别方法。Optionally, this embodiment of the present application further provides a computer program product, including a computer program or instructions, which implement the speech recognition method as described above when executed by a processor.
图8是本申请实施例提供的一种电子设备800的框图。例如,电子设备800可以为移动终端也可以为服务器,本申请实施例中以电子设备为移动终端为例进行说明。例如,电子设备800可以是移动电话,计算机,数字广播终端,消息收发设备,游戏控制台,平板设备,医疗设备,健身设备,个人数字助理等。FIG. 8 is a block diagram of an electronic device 800 provided by an embodiment of the present application. For example, the electronic device 800 may be a mobile terminal or a server. In the embodiment of this application, the electronic device 800 is a mobile terminal as an example for description. For example, the electronic device 800 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, etc.
参照图8,电子设备800可以包括以下一个或多个组件:处理组件802,存储器804,电力组件806,多媒体组件808,音频组件810,输入/输出(I/O)的接口812,传感器组件814,以及通信组件816。Referring to FIG. 8 , the electronic device 800 may include one or more of the following components: a processing component 802 , a memory 804 , a power component 806 , a multimedia component 808 , an audio component 810 , an input/output (I/O) interface 812 , and a sensor component 814 , and communication component 816.
处理组件802通常控制电子设备800的整体操作,诸如与显示,电话呼叫,数据通信,相机操作和记录操作相关联的操作。处理组件802可以包括一个或多个处理器820来执行指令,以完成上述的方法的全部或部分步骤。此外,处理组件802可以包括一个或多个模块,便于处理组件802和其他组件之间的交互。例如,处理组件802可以包括多媒体模块,以方便多媒体组件808和处理组件802之间的交互。 Processing component 802 generally controls the overall operations of electronic device 800, such as operations associated with display, phone calls, data communications, camera operations, and recording operations. The processing component 802 may include one or more processors 820 to execute instructions to complete all or part of the steps of the above method. Additionally, processing component 802 may include one or more modules that facilitate interaction between processing component 802 and other components. For example, processing component 802 may include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
存储器804被配置为存储各种类型的数据以支持在设备800的操作。这些数据的示例包括用于在电子设备800上操作的任何应用程序或方法的指令,联系人数据,电话簿数据,消息,图片,视频等。存储器804可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。 Memory 804 is configured to store various types of data to support operations at device 800 . Examples of such data include instructions for any application or method operating on the electronic device 800, contact data, phonebook data, messages, pictures, videos, etc. Memory 804 may be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EEPROM), Programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disk.
电源组件806为电子设备800的各种组件提供电力。电源组件806可以包括电源管理系统,一个或多个电源,及其他与为电子设备800生成、管理和分配电力相关联的组件。 Power supply component 806 provides power to various components of electronic device 800 . Power supply components 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power to electronic device 800 .
多媒体组件808包括在所述电子设备800和用户之间的提供一个输出接口的屏幕。在一些实施例中,屏幕可以包括液晶显示器(LCD)和触摸面板(TP)。如果屏幕包括 触摸面板,屏幕可以被实现为触摸屏,以接收来自用户的输入信号。触摸面板包括一个或多个触摸传感器以感测触摸、滑动和触摸面板上的手势。所述触摸传感器可以不仅感测触摸或滑动动作的边界,而且还检测与所述触摸或滑动操作相关的持续时间和压力。在一些实施例中,多媒体组件808包括一个前置摄像头和/或后置摄像头。当设备800处于操作模式,如拍摄模式或视频模式时,前置摄像头和/或后置摄像头可以接收外部的多媒体数据。每个前置摄像头和后置摄像头可以是一个固定的光学透镜系统或具有焦距和光学变焦能力。 Multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user. In some embodiments, the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user. The touch panel includes one or more touch sensors to sense touches, swipes, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide action. In some embodiments, multimedia component 808 includes a front-facing camera and/or a rear-facing camera. When the device 800 is in an operating mode, such as a shooting mode or a video mode, the front camera and/or the rear camera may receive external multimedia data. Each front-facing camera and rear-facing camera can be a fixed optical lens system or have a focal length and optical zoom capabilities.
音频组件810被配置为输出和/或输入音频信号。例如,音频组件810包括一个麦克风(MIC),当电子设备800处于操作模式,如呼叫模式、记录模式和语音识别模式时,麦克风被配置为接收外部音频信号。所接收的音频信号可以被进一步存储在存储器804或经由通信组件816发送。在一些实施例中,音频组件810还包括一个扬声器,用于输出音频信号。 Audio component 810 is configured to output and/or input audio signals. For example, audio component 810 includes a microphone (MIC) configured to receive external audio signals when electronic device 800 is in operating modes, such as call mode, recording mode, and voice recognition mode. The received audio signal may be further stored in memory 804 or sent via communication component 816 . In some embodiments, audio component 810 also includes a speaker for outputting audio signals.
I/O接口812为处理组件802和外围接口模块之间提供接口,上述外围接口模块可以是键盘,点击轮,按钮等。这些按钮可包括但不限于:主页按钮、音量按钮、启动按钮和锁定按钮。The I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module, which may be a keyboard, a click wheel, a button, etc. These buttons may include, but are not limited to: Home button, Volume buttons, Start button, and Lock button.
传感器组件814包括一个或多个传感器,用于为电子设备800提供各个方面的状态评估。例如,传感器组件814可以检测到设备800的打开/关闭状态,组件的相对定位,例如所述组件为电子设备800的显示器和小键盘,传感器组件814还可以检测电子设备800或电子设备800一个组件的位置改变,用户与电子设备800接触的存在或不存在,电子设备800方位或加速/减速和电子设备800的温度变化。传感器组件814可以包括接近传感器,被配置用来在没有任何的物理接触时检测附近物体的存在。传感器组件814还可以包括光传感器,如CMOS或CCD图像传感器,用于在成像应用中使用。在一些实施例中,该传感器组件814还可以包括加速度传感器,陀螺仪传感器,磁传感器,压力传感器或温度传感器。 Sensor component 814 includes one or more sensors for providing various aspects of status assessment for electronic device 800 . For example, the sensor component 814 can detect the open/closed state of the device 800, the relative positioning of components, such as the display and keypad of the electronic device 800. The sensor component 814 can also detect the electronic device 800 or a component of the electronic device 800. changes in position, the presence or absence of user contact with the electronic device 800 , the orientation or acceleration/deceleration of the electronic device 800 and changes in the temperature of the electronic device 800 . Sensor assembly 814 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact. Sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor component 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
通信组件816被配置为便于电子设备800和其他设备之间有线或无线方式的通信。电子设备800可以接入基于通信标准的无线网络,如WiFi,运营商网络(如2G、3G、4G或5G),或它们的组合。在一个示例性实施例中,通信组件816经由广播信道接收来自外部广播管理系统的广播信号或广播相关信息。在一个示例性实施例中,所述通信组件816还包括近场通信(NFC)模块,以促进短程通信。例如,在NFC模块可基于射频识别(RFID)技术,红外数据协会(IrDA)技术,超宽带(UWB)技术,蓝牙(BT)技术和其他技术来实现。 Communication component 816 is configured to facilitate wired or wireless communication between electronic device 800 and other devices. The electronic device 800 can access a wireless network based on a communication standard, such as WiFi, a carrier network (such as 2G, 3G, 4G or 5G), or a combination thereof. In one exemplary embodiment, the communication component 816 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel. In one exemplary embodiment, the communications component 816 also includes a near field communications (NFC) module to facilitate short-range communications. For example, the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
在实施例中,电子设备800可以被一个或多个应用专用集成电路(ASIC)、数字 信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,用于执行上述所示的语音识别方法。In embodiments, electronic device 800 may be configured by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gates Array (FPGA), controller, microcontroller, microprocessor or other electronic components are implemented for executing the speech recognition method shown above.
在实施例中,还提供了一种计算机可读存储介质,例如包括指令的存储器804,上述指令可由电子设备800的处理器820执行以完成上述所示的语音识别方法。例如,所述非临时性计算机可读存储介质可以是ROM、随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。In an embodiment, a computer-readable storage medium is also provided, such as a memory 804 including instructions, and the instructions can be executed by the processor 820 of the electronic device 800 to complete the speech recognition method shown above. For example, the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
在实施例中,还提供了一种计算机程序产品,当计算机程序产品中的指令由电子设备800的处理器820执行时,使得电子设备800执行上述所示的语音识别方法。In an embodiment, a computer program product is also provided. When the instructions in the computer program product are executed by the processor 820 of the electronic device 800, the electronic device 800 performs the speech recognition method shown above.
图9是本申请实施例提供的一种用于语音识别的装置900的框图。例如,装置900可以被提供为一服务器。参照图9,装置900包括处理组件922,其进一步包括一个或多个处理器,以及由存储器932所代表的存储器资源,用于存储可由处理组件922的执行的指令,例如应用程序。存储器932中存储的应用程序可以包括一个或一个以上的每一个对应于一组指令的模块。此外,处理组件922被配置为执行指令,以执行上述方法。Figure 9 is a block diagram of a device 900 for speech recognition provided by an embodiment of the present application. For example, device 900 may be provided as a server. Referring to Figure 9, apparatus 900 includes a processing component 922, which further includes one or more processors, and memory resources represented by memory 932 for storing instructions, such as application programs, executable by processing component 922. The application program stored in memory 932 may include one or more modules, each corresponding to a set of instructions. Furthermore, the processing component 922 is configured to execute instructions to perform the above-described method.
装置900还可以包括一个电源组件926被配置为执行装置900的电源管理,一个有线或无线网络接口950被配置为将装置900连接到网络,和一个输入输出(I/O)接口958。装置900可以操作基于存储在存储器932的操作系统,例如Windows ServerTM,Mac OS XTM,UnixTM,LinuxTM,FreeBSDTM或类似。 Device 900 may also include a power supply component 926 configured to perform power management of device 900, a wired or wireless network interface 950 configured to connect device 900 to a network, and an input-output (I/O) interface 958. Device 900 may operate based on an operating system stored in memory 932, such as Windows Server™, Mac OS X™, Unix™, Linux™, FreeBSD™ or the like.
本领域技术人员在考虑说明书及实践这里公开的发明后,将容易想到本申请的其它实施方案。本申请旨在涵盖本申请的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本申请的一般性原理并包括本申请未公开的本技术领域中的公知常识或惯用技术手段。说明书和实施例仅被视为示例性的,本申请的真正范围和精神由下面的权利要求指出。Other embodiments of the present application will be readily apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. This application is intended to cover any variations, uses, or adaptations of this application that follow the general principles of this application and include common knowledge or customary technical means in the technical field that are not disclosed in this application. . It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the application being indicated by the following claims.
应当理解的是,本申请并不局限于上面已经描述并在附图中示出的精确结构,并且可以在不脱离其范围进行各种修改和改变。本申请的范围仅由所附的权利要求来限制。It is to be understood that the present application is not limited to the precise structures described above and illustrated in the accompanying drawings, and that various modifications and changes may be made without departing from the scope thereof. The scope of the application is limited only by the appended claims.

Claims (10)

  1. 一种语音识别方法,其特征在于,包括:A speech recognition method, characterized by including:
    响应车内用户的语音信息,获取所述用户的面部图像;Respond to the voice information of the user in the car and obtain the facial image of the user;
    根据所述面部图像上的面部特征确定所述用户的当前状态;Determine the current status of the user based on facial features on the facial image;
    在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果;When the user's current status meets the set conditions, recognize the voice information and obtain a recognition result;
    在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。When the recognition result is a vehicle-machine instruction, the corresponding operation is performed according to the vehicle-machine instruction.
  2. 根据权利要求1所述的语音识别方法,其特征在于,所述方法还包括:The speech recognition method according to claim 1, characterized in that the method further includes:
    在所述用户的当前状态不满足设定条件时,拒绝对所述语音信息进行识别。When the current status of the user does not meet the set conditions, the recognition of the voice information is refused.
  3. 根据权利要求1或2所述的语音识别方法,其特征在于,所述根据所述面部图像上的面部特征确定所述用户的当前状态,至少包括下述一种:The speech recognition method according to claim 1 or 2, characterized in that determining the current state of the user based on facial features on the facial image includes at least one of the following:
    获取车辆的信息状态,基于所述信息状态和所述面部图像的面部特征判定车载蓝牙电话没有开启时,确定所述用户处于非打电话状态;Obtain the information status of the vehicle, and when it is determined that the vehicle-mounted Bluetooth phone is not turned on based on the information status and the facial features of the facial image, it is determined that the user is in a non-phone state;
    在根据所述面部图像的面部特征判定所述用户的正脸看向车辆行驶方向时,确定所述用户处于正脸看向前的状态;When it is determined based on the facial features of the facial image that the user's front face is looking in the direction of vehicle travel, it is determined that the user is in a state of facing forward;
    在根据所述面部图像的面部特征判定所述用户的嘴巴处于张合状态时,确定所述用户处于说话状态。When it is determined that the user's mouth is in an opening and closing state based on the facial features of the facial image, it is determined that the user is in a speaking state.
  4. 根据权利要求3所述的语音识别方法,其特征在于,所述在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果,包括:The speech recognition method according to claim 3, characterized in that, when the current state of the user meets the set conditions, the speech information is recognized and the recognition result is obtained, including:
    在所述用户的当前状态为:所述用户处于非打电话状态、用户处于正脸向前看的状态和用户处于说话状态的至少一种时,确定所述用户满足设定条件;When the current state of the user is at least one of: the user is in a non-phone state, the user is in a state of facing forward, and the user is in a speaking state, it is determined that the user satisfies the set condition;
    对所述语音信息进行识别,得到识别结果。The voice information is recognized and a recognition result is obtained.
  5. 根据权利要求4所述的语音识别方法,其特征在于,所述对所述语音信息进行识别,得到识别结果,包括:The speech recognition method according to claim 4, wherein the step of identifying the speech information to obtain a recognition result includes:
    将所述语音信息进行本地语音文字转换处理,得到转换后的文本信息;或者Perform local speech-to-text conversion processing on the voice information to obtain converted text information; or
    将所述语音信息发送给云端,由所述云端进行语音文字转换处理后得到文本信息;Send the voice information to the cloud, and the cloud performs voice-to-text conversion processing to obtain text information;
    接收所述云端发送的转换后的文本信息。Receive the converted text information sent by the cloud.
  6. 根据权利要求4所述的语音识别方法,其特征在于,所述在所述识别结果为车机指令时,按照所述车机指令执行对应的操作,包括:The speech recognition method according to claim 4, characterized in that when the recognition result is a vehicle-machine instruction, performing corresponding operations according to the vehicle-machine instruction includes:
    将所述识别结果通过训练好的车机指令识别模型进行判断,得到所述识别结果是车机指令;其中,所述训练好的车机指令识别模型是基于人与车机交互的多个历史音频对,文本对,以及场景和关键词进行学习训练得到的模型;The recognition result is judged by a trained vehicle-machine command recognition model, and the recognition result is a vehicle-machine command; wherein the trained vehicle-machine command recognition model is based on multiple histories of human-vehicle interaction A model obtained by learning and training audio pairs, text pairs, scenes and keywords;
    按照得到的所述车机指令执行对应的操作。Perform corresponding operations according to the obtained vehicle and machine instructions.
  7. 一种语音识别装置,其特征在于,包括:A speech recognition device, characterized by including:
    获取模块,用于响应车内用户的语音信息,获取所述用户的面部图像;The acquisition module is used to respond to the voice information of the user in the car and obtain the facial image of the user;
    确定模块,用于根据所述面部图像上的面部特征确定所述用户的当前状态;a determining module configured to determine the current status of the user based on facial features on the facial image;
    识别模块,用于在所述用户的当前状态满足设定条件时,对所述语音信息进行识别,得到识别结果;A recognition module, used to recognize the voice information and obtain a recognition result when the current status of the user meets the set conditions;
    执行模块,在所述识别结果为车机指令时,按照所述车机指令执行对应的操作。An execution module, when the recognition result is a vehicle-machine instruction, executes the corresponding operation according to the vehicle-machine instruction.
  8. 一种电子设备,其特征在于,包括:An electronic device, characterized by including:
    处理器;processor;
    用于存储所述处理器可执行指令的存储器;memory for storing instructions executable by the processor;
    其中,所述处理器被配置为执行所述指令,以实现如权利要求1至6中任一项所述的语音识别方法。Wherein, the processor is configured to execute the instructions to implement the speech recognition method according to any one of claims 1 to 6.
  9. 一种计算机可读存储介质,其特征在于,当所述计算机可读存储介质中的指令由电子设备的处理器执行时,使得电子设备能够执行如权利要求1至6中任一项所述的语音识别方法。A computer-readable storage medium, characterized in that, when the instructions in the computer-readable storage medium are executed by a processor of an electronic device, the electronic device is capable of executing the method described in any one of claims 1 to 6 Speech recognition methods.
  10. 一种计算机程序产品,包括计算机程序或指令,其特征在于,所述计算机程序或指令被处理器执行时实现权利要求1至6任一项所述的语音识别方法。A computer program product, including a computer program or instructions, characterized in that when the computer program or instructions are executed by a processor, the speech recognition method according to any one of claims 1 to 6 is implemented.
PCT/CN2022/117333 2022-06-01 2022-09-06 Voice recognition method and apparatus, electronic device, storage medium, and product WO2023231211A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210617455.2A CN115171678A (en) 2022-06-01 2022-06-01 Voice recognition method, device, electronic equipment, storage medium and product
CN202210617455.2 2022-06-01

Publications (1)

Publication Number Publication Date
WO2023231211A1 true WO2023231211A1 (en) 2023-12-07

Family

ID=83483240

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/117333 WO2023231211A1 (en) 2022-06-01 2022-09-06 Voice recognition method and apparatus, electronic device, storage medium, and product

Country Status (2)

Country Link
CN (1) CN115171678A (en)
WO (1) WO2023231211A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117238291A (en) * 2023-11-14 2023-12-15 暗物智能科技(广州)有限公司 Multi-mode voice refusing identification method

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017035768A1 (en) * 2015-09-01 2017-03-09 涂悦 Voice control method based on visual wake-up
CN112567457A (en) * 2019-12-13 2021-03-26 华为技术有限公司 Voice detection method, prediction model training method, device, equipment and medium
CN113539265A (en) * 2021-07-13 2021-10-22 中国第一汽车股份有限公司 Control method, device, equipment and storage medium
CN113674746A (en) * 2021-08-18 2021-11-19 北京百度网讯科技有限公司 Man-machine interaction method, device, equipment and storage medium
CN114187637A (en) * 2021-12-13 2022-03-15 中国第一汽车股份有限公司 Vehicle control method, device, electronic device and storage medium
CN114333017A (en) * 2021-12-29 2022-04-12 阿波罗智联(北京)科技有限公司 Dynamic pickup method and device, electronic equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017035768A1 (en) * 2015-09-01 2017-03-09 涂悦 Voice control method based on visual wake-up
CN112567457A (en) * 2019-12-13 2021-03-26 华为技术有限公司 Voice detection method, prediction model training method, device, equipment and medium
CN113539265A (en) * 2021-07-13 2021-10-22 中国第一汽车股份有限公司 Control method, device, equipment and storage medium
CN113674746A (en) * 2021-08-18 2021-11-19 北京百度网讯科技有限公司 Man-machine interaction method, device, equipment and storage medium
CN114187637A (en) * 2021-12-13 2022-03-15 中国第一汽车股份有限公司 Vehicle control method, device, electronic device and storage medium
CN114333017A (en) * 2021-12-29 2022-04-12 阿波罗智联(北京)科技有限公司 Dynamic pickup method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN115171678A (en) 2022-10-11

Similar Documents

Publication Publication Date Title
WO2020134858A1 (en) Facial attribute recognition method and apparatus, electronic device, and storage medium
CN108874967B (en) Dialogue state determining method and device, dialogue system, terminal and storage medium
TWI759647B (en) Image processing method, electronic device, and computer-readable storage medium
WO2021135685A1 (en) Identity authentication method and device
EP4064284A1 (en) Voice detection method, prediction model training method, apparatus, device, and medium
CN108363706A (en) The method and apparatus of human-computer dialogue interaction, the device interacted for human-computer dialogue
KR102420567B1 (en) Method and device for voice recognition
KR20190106921A (en) Communication robot and method for operating the same
US20230129816A1 (en) Speech instruction control method in vehicle cabin and related device
CN112860169B (en) Interaction method and device, computer readable medium and electronic equipment
WO2023273064A1 (en) Object speaking detection method and apparatus, electronic device, and storage medium
US20210192192A1 (en) Method and apparatus for recognizing facial expression
WO2023231211A1 (en) Voice recognition method and apparatus, electronic device, storage medium, and product
WO2023173660A1 (en) User recognition method and apparatus, storage medium, electronic device, computer program product and computer program
CN111583923A (en) Information control method and device, and storage medium
CN114333774B (en) Speech recognition method, device, computer equipment and storage medium
CN112286364A (en) Man-machine interaction method and device
WO2023273063A1 (en) Passenger speaking detection method and apparatus, and electronic device and storage medium
CN115291724A (en) Man-machine interaction method and device, storage medium and electronic equipment
CN111435422B (en) Action recognition method, control method and device, electronic equipment and storage medium
CN108388399B (en) Virtual idol state management method and system
CN112639964A (en) Method, system and computer readable medium for recognizing speech using depth information
CN114187637A (en) Vehicle control method, device, electronic device and storage medium
CN113611318A (en) Audio data enhancement method and related equipment
KR20190048630A (en) Electric terminal and method for controlling the same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22944521

Country of ref document: EP

Kind code of ref document: A1