CN110444211A - A kind of audio recognition method and equipment - Google Patents

A kind of audio recognition method and equipment Download PDF

Info

Publication number
CN110444211A
CN110444211A CN201910785213.2A CN201910785213A CN110444211A CN 110444211 A CN110444211 A CN 110444211A CN 201910785213 A CN201910785213 A CN 201910785213A CN 110444211 A CN110444211 A CN 110444211A
Authority
CN
China
Prior art keywords
electronic device
target electronic
wearable device
speech
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910785213.2A
Other languages
Chinese (zh)
Inventor
王康
曹建伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qingdao Hisense Electronics Co Ltd
Original Assignee
Qingdao Hisense Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qingdao Hisense Electronics Co Ltd filed Critical Qingdao Hisense Electronics Co Ltd
Priority to CN201910785213.2A priority Critical patent/CN110444211A/en
Publication of CN110444211A publication Critical patent/CN110444211A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

This application discloses a kind of audio recognition method and equipment, it is related to field of speech recognition, for using target electronic device as the speech recognition module of wearable device, wearable device to be without increasing additional module, the cost and weight for reducing wearable device, improve the usage experience of user.This method comprises: sending speech recognition request to target electronic device by wearable device, after target electronic device receives speech recognition request, the voice input of user's sending is included.Target electronic device identify to user speech and recognition result is sent to wearable device.Wearable device receives recognition result and executes corresponding function in recognition result.The application is applied to speech recognition.

Description

A kind of audio recognition method and equipment
Technical field
The present invention relates to field of speech recognition more particularly to a kind of audio recognition method and equipment.
Background technique
Most of wearable device haves the defects that very big that man-machine interaction mode is cumbersome in man-machine interaction mode at present It is complicated.Especially when user needs to carry out text input or assigns instruction, user needs positioning cursor repeatedly To the key of dummy keyboard, such input mode extremely very complicated greatly reduces the usage experience of user.Existing rank Although section has had manufacturer to propose the technical solution of wearable device integrating speech sound identifying system, this technical solution is needed Dedicated speech recognition system is configured for wearable device, comprising: integrated identification chip, audio signal reception device etc..It in this way can be into one The cost and weight of the increase wearable device of step, also will increase the calculation amount of wearable device, reduce calculated performance.
Summary of the invention
Embodiments herein provides a kind of audio recognition method and equipment, can for solving that user in the prior art wears When wearable device, text inputs or assigns the problem of instruction process very complicated inconvenience.
In order to achieve the above objectives, embodiments herein adopts the following technical scheme that
In a first aspect, this method is applied to wearable device this application provides a kind of audio recognition method and equipment, it should Method includes: to send speech recognition request to target electronic device, and speech recognition request is for making target electronic device include use Family voice simultaneously generates recognition result according to user speech, and recognition result is sent to wearable device.Target electronic is received to set The recognition result that preparation is sent, and execute function corresponding with recognition result.
Second aspect, this application provides a kind of audio recognition method and equipment, this method is applied to target electronic device, This method comprises: receiving the speech recognition request that wearable device is sent includes user speech in response to speech recognition request. Recognition result is generated according to user speech.Recognition result is sent to wearable device, recognition result is for holding wearable device Row function corresponding with recognition result.
The third aspect, this application provides a kind of wearable device, the wearable device include: processor, memory and Communication interface.Wherein, communication interface is for wearable device and other equipment or network communication.Memory for store one or Multiple programs, which includes computer executed instructions, and when wearable device operation, processor executes storage Computer executed instructions of device storage, so that wearable device executes the audio recognition method of above-mentioned first aspect.
Fourth aspect, this application provides a kind of electronic equipment, which includes: processor, memory and communication Interface.Wherein, communication interface is for electronic equipment and other equipment or network communication.Memory is for storing one or more journeys Sequence, which includes computer executed instructions, and when electronic equipment operation, processor executes memory storage Computer executed instructions, so that electronic equipment executes the audio recognition method of above-mentioned second aspect.
5th aspect, this application provides a kind of corresponding computers for storing one or more programs of wearable device can Storage medium is read, one or more programs include instruction, and instruction executes wearable device when being executed by wearable device State the audio recognition method of first aspect.
6th aspect, this application provides the computer-readable of the one or more programs of the corresponding storage of a kind of electronic equipment Storage medium, one or more programs include instruction, and instruction makes electronic equipment execute above-mentioned second when being executed by electronic equipment The audio recognition method of aspect.
A kind of audio recognition method of the target electronic device that embodiments herein provides as wearable device.Pass through Wearable device sends speech recognition request to target electronic device, after target electronic device receives speech recognition request, receives Employ the voice input of family sending.Target electronic device identify to user speech and recognition result is sent to wearable set It is standby.Wearable device receives recognition result and executes corresponding function in recognition result.Present invention greatly reduces user with The interaction of wearable device, especially in terms of text inputs or assigns instruction, using electronic equipment as wearable device Speech recognition module does not increase the cost and weight of wearable device, greatly improves user without increasing additional module Usage experience.
Detailed description of the invention
Fig. 1 is a kind of flow diagram for audio recognition method that embodiments herein provides;
Fig. 2 is the flow diagram for another audio recognition method that embodiments herein provides;
Fig. 3 is the interface display after a kind of target electronic device that embodiments herein provides is connect with wearable device Screen locking picture drawing;
Fig. 4 is a kind of structural schematic diagram for wearable device that embodiments herein provides;
Fig. 5 is the structural schematic diagram for another wearable device that embodiments herein provides;
Fig. 6 is the structural schematic diagram for a kind of electronic equipment that embodiments herein provides;
Fig. 7 is the structural schematic diagram for another electronic equipment that embodiments herein provides.
Specific embodiment
Below first to the invention relates to some concepts simply introduced.
AR (AugmentedReality augmented reality), also referred to as mixed reality.It, will be virtual by computer technology To real world, true environment and virtual object have been added in real time the same picture or space while to be deposited Information application In.
VR (Virtual Reality virtual reality), specific intension are comprehensive utilization computer graphics systems and various The interface equipments such as reality and control, the technology for immersing feeling is provided in three-dimensional environment generating on computers, can interacting.
By the way that various living scenes can be experienced in the case where staying indoors using VR product and AR product.Such as: In When carrying out shopping online, clothing can be tried on by VR product and AR product;It is seen by VR product and AR product Concert or sports tournament, can be to the feeling as a kind of scene of people;It is each with AR product to can see the world by VR product The tourist attractions on ground, the feeling as on the spot in person.But existing VR product and AR product in man-machine interaction mode also Very big defect, man-machine interaction mode are more complicated.It thus will affect the usage experience of user.Existing VR product and AR are produced Product are human-computer interaction to be realized by virtual screen key, or pre-set voice on VR product and AR product mostly Module or an external voice module realize human-computer interaction.Thus will increase the cost of VR product and AR product with Weight can also reduce the performance of VR product Yu AR product.Wherein, meaning VR product in the application, is specifically as follows VR glasses.AR Product then specifically can be the equipment such as AR glasses.
The embodiment of the present application provides a kind of audio recognition method and equipment, to solve the deficiencies in the prior art. The application is not necessarily to pre-set voice module on VR product and AR product, does not also need external voice module.The application is by target Speech recognition equipment of the electronic equipment as VR product and AR product, VR product and AR product are by way of wireless connection and mesh Mark electronic equipment interacts.
Below in conjunction with the attached drawing in the embodiment of the present application, technical solutions in the embodiments of the present application is described.
In the description of the present application, unless otherwise indicated, "/" indicates the meaning of "or", for example, A/B can indicate A or B. "and/or" herein is only a kind of incidence relation for describing affiliated partner, indicates may exist three kinds of relationships, for example, A And/or B, can indicate: individualism A exists simultaneously A and B, these three situations of individualism B.In addition, "at least one" is Refer to one or more, " multiple " refer to two or more.The printed words such as " first ", " second " are not to quantity and execution order It is defined, and the printed words such as " first ", " second " also do not limit certain difference.
Technical solution provided by the embodiments of the present application can be applied to speech recognition of the electronic equipment as wearable device Module.In the embodiment of the present application, in the specific implementation, it is each specifically to can be mobile phone, tablet computer or MP4 etc. for electronic equipment Class has the function of that voice is included and the equipment of wireless connecting function.Which kind of equipment specifically to realize the embodiment of the present application is provided by Audio recognition method in electronic equipment function, then can be selected according to concrete scene, can be with no restrictions to this application.
Embodiment one
When the wearable device in the embodiment of the present application is AR glasses, as shown in Figure 1, being provided by the embodiments of the present application A kind of flow diagram of audio recognition method, this method comprises:
S101, AR glasses and target electronic device establish connection, and target electronic device enters particular state.
Specifically, AR glasses and target electronic device open the wireless connecting functions such as Bluetooth function, the direct-connected function of WIFI, or Person accesses in same local area network.Target electronic device scans peripheral equipment by wireless connecting function and transmission is established connection and asked Ask, AR glasses receive establish connection request after to electronic equipment returning equipment information.Target electronic device is sent out according to AR glasses The facility information and AR glasses sent, which is established, to be matched.So far, successful connection has been established in AR glasses and target electronic device.
Wherein, it is connect to prompt ownership goal electronic equipment to establish with AR glasses, while alsoing for avoiding Connection carries out maloperation to target electronic device after establishing, after AR glasses connect foundation with target electronic device, target electronic Equipment enters particular state.Wherein, particular state is as shown in figure 3, the display interface of target electronic device shows screen locking picture simultaneously And the brightness of display interface is in minimum state.Also, in a particular state, target electronic device is able to carry out following operation: The speech recognition request of AR glasses transmission is received, user speech is included, recognition result is generated according to user speech, sends identification knot Fruit operates to AR glasses etc., and the specific implementation content of aforesaid operations can refer to corresponding contents hereinafter.
Before the operation that AR glasses receive user, target electronic device receives user to the 4th behaviour of target electronic device Make, sends request voice recognition instruction and give AR glasses.After AR glasses receive request voice recognition instruction, speech recognition request is sent To target electronic device.
S102, AR glasses receive the first operation that user treats input area.
Wherein, region to be entered can specifically include: need to input the region of text on the display interface of AR glasses.So It afterwards, can be by the behavior act of acquisition user, to determine that user treats the operation of input area.For example, user can pass through The touch tablet in AR eyeglasses temple arm is arranged in touch-control, thus the cursor on mobile display interface, and be moved to by cursor to defeated After entering region, input area can be treated and carry out such as other operations of input text.For another example detection user's eye can also be passed through The focus of eyeball, to determine eyes of user position of interest.When determining that eyes of user watches attentively at region to be entered, then can treat Input area carries out other subsequent operations.
Illustratively, the first operation that user treats input area may include: that user positions cursor to area to be entered Domain.Alternatively, the first operation can also include: user's long-pressing region to be entered.
S103, AR glasses are shown in the display interface in response to the first operation for prompting user to pass through target electronic The prompt information of equipment inputting word information.
For example, cursor is positioned to region to be entered, will pop up one on display interface when user is by the first operation Indicate the pop-up of " whether needing to input text by voice ".
The second operation is carried out (for example, user clicks the determination in pop-up to prompt information shown on interface in user Key) after, method provided by the embodiment of the present application is specific further comprising the steps of:
S104, AR glasses send speech recognition request to target electronic device, and target electronic device receives speech recognition and asks It asks.
Specifically, step S104 includes: of AR glasses in response to user to prompt information after executing step S103 Two operations send speech recognition request to target electronic device.
Wherein, speech recognition request is identified for making target electronic device include user speech and be generated according to user speech As a result, and recognition result is sent to wearable device, i.e., the AR glasses in the present embodiment.
Specifically, in the present embodiment, speech recognition request, specifically for making target electronic device include user speech simultaneously It generates corresponding text according to user speech, and text is sent to wearable device, i.e., the AR glasses in the present embodiment.In After AR glasses send speech recognition request to target electronic device, display reminding window in display interface.Include in prompt window Cancel key, is instructed after cancelling key when the user clicks to target electronic device transmission cancellation, so that target electronic device is stopped Only include user speech.
Illustratively, the prompt information in step S103 can be with are as follows: display interface shows that one indicates and " whether needs to lead to Cross voice input text " or " i.e. will by voice input text " pop-up.Alternatively, showing input method in the display interface The key of speech voice input function is triggered at interface including one.
In turn, the second operation can be with are as follows: user clicks the determination key in pop-up, and AR glasses are clicked in user and determine key Afterwards, speech recognition request is sent to target electronic device.
It should be noted that be in the embodiment of the present application by user first operation, second operation come trigger voice know It does not invite and asks, in some implementations, other triggering modes can also be set according to specific implementation situation.When utilize other When triggering mode triggers, then the content of S102 and S103 can not be executed.
S105, target electronic device include user speech in response to speech recognition request.
Specifically, target electronic device in the present embodiment can have for mobile phone, tablet computer etc. include voice and Each class of electronic devices of speech identifying function.
In one implementation, it is contemplated that after user carries out the above-mentioned second operation to AR glasses, target electronic device It is different surely to call sound-recording function in a short time, if therefore user carry out second operation after directly loquitur, may Appearance can not be indexed to user speak audio the case where.Therefore, language is included in order to allow users to perceive target electronic device The time of sound, in the embodiment of the present application, S105 is specifically included:
S1051, target electronic device issue standby signal after receiving the speech recognition request.
Specifically, standby signal at least may include vibration or playing alert tones.
S1052, the user speech after the standby signal issues is included.
Specifically, generating the recognition result containing text according to user speech after target electronic device includes user speech.
S106, target electronic device send recognition result and give AR glasses, and AR glasses receive recognition result and execute and identify As a result corresponding function.
Wherein, recognition result is for making AR glasses execute function corresponding with recognition result.
Specifically, AR glasses receive the recognition result that target electronic device is sent, recognition result includes user speech input Corresponding text.Text in recognition result is input in region to be entered by AR glasses.AR glasses are receiving identification knot After fruit, recognition result is presented on prompt window, executes function corresponding with recognition result after user's confirmation.
Below in conjunction with example, audio recognition method provided by the embodiment of the present application is introduced: using mobile phone as AR The target electronic device of glasses, AR glasses are matched with mobile phone by Bluetooth function.Mobile phone opens client application and passes through nothing Line functionality scan peripheral equipment sends discovering device and requests to give AR glasses.The service of AR glasses running background, which receives discovery, asks After asking, a facility information, the equipment that mobile phone is matched according to facility information selection are returned to cell phone client.AR glasses and mobile phone After successful matching, when user needs inputting word information, user can show that one passes through by clicking input frame, AR glasses The prompt information of mobile phone input text.Then AR glasses send speech recognition request to mobile phone, and mobile phone is known in response to this voice It does not invite to seek and open voice and include function and starts typing voice.The speech recognition that cell phone client will be included after voice is included It generates corresponding Text region result and recognition result is sent to AR glasses.AR glasses tie identification after receiving recognition result Text corresponding to fruit is input in input frame.
It should be noted that in some implementations, other electronic equipment conducts can also be used in those skilled in the art The language energy identification module of wearable device.For example, using tablet computer as VR glasses or the speech recognition mould of AR glasses Block, can be with no restrictions to this application.
Embodiment two
When the wearable device in the embodiment of the present application is AR glasses, as shown in Fig. 2, being provided by the embodiments of the present application The flow diagram of another audio recognition method, which comprises
S201, AR glasses and target electronic device establish connection, and target electronic device enters particular state.
Specifically, the particular content of the step can refer to step S101 content, do not illustrate herein.
S202, AR glasses receive the third operation of user.
Specifically, third operates are as follows: user nods, user shakes the head or presses specified key, after waking up AR glasses Platform service.Third operation is to need operation when assigning control instruction to AR glasses in user.For example, when user is using AR For mirror when watching music video, user can activate speech recognition request by third operation, to be set by target electronic The standby control instruction assigned such as " switching next track " or " closing music video ".AR glasses can pass through the behaviour of acquisition user Make, to determine whether user needs to assign control instruction to AR glasses by voice.Illustratively, user can be by by dividing into The key in AR eyeglasses temple arm is set, to activate the speech recognition request of AR glasses, " is cut to be assigned by target electronic device Change next track " control instruction.
S203, AR glasses are operated in response to third, are shown in the display interface for prompting user to pass through target electronic Equipment assigns the prompt information of control instruction.
" finger will be assigned by voice for example, will pop up one after user finishes third operation, on display interface and indicate Enable " pop-up.AR glasses send speech recognition request to target electronic device simultaneously.
S204, AR glasses send speech recognition request to target electronic device, and target electronic device receives speech recognition and asks It asks.
Wherein, speech recognition request is controlled for making target electronic device include user speech and be generated according to user speech Instruction, and control instruction is sent to wearable device, i.e., the AR glasses in the present embodiment.
Specifically, in the present embodiment, speech recognition request, specifically for making target electronic device include user speech simultaneously Control instruction is generated according to user speech, and control instruction is sent to AR glasses.AR glasses send speech recognition request to After target electronic device, display reminding window in display interface.Comprising cancelling key in prompt window, cancels pressing when the user clicks It is sent after key to target electronic device and cancels instruction, so that user speech is included in target electronic device stopping.
It should be noted that be that speech recognition request is triggered by the third operation of user in the embodiment of the present application, In some implementations, other triggering modes can also be set according to specific implementation situation.It is touched when using other triggering modes When hair, then the content of S202 and S203 can not be executed.
S205, target electronic device include user speech in response to speech recognition request.
Specifically, target electronic device in the present embodiment can have for mobile phone, tablet computer etc. include voice and Each class of electronic devices of speech identifying function.
In one implementation, it is contemplated that after user carries out above-mentioned third operation to AR glasses, target electronic device It is different surely to call sound-recording function in a short time, if therefore user directly loquitur after carrying out third operation, may Appearance can not be indexed to user speak audio the case where.Therefore, language is included in order to allow users to perceive target electronic device The time of sound, in the embodiment of the present application, S205 is specifically included:
S2051, target electronic device issue standby signal after receiving speech recognition request.
Specifically, standby signal at least may include vibration or playing alert tones.
S2052, the user speech after standby signal issues is included.
Specifically, generating the identification containing control instruction according to user speech after target electronic device includes user speech As a result.
S206, target electronic device send recognition result and give AR glasses, and AR glasses receive recognition result and execute identification knot Control instruction in fruit.
Wherein, the recognition result is for making the wearable device execute function corresponding with the recognition result.
Specifically, AR glasses receive the recognition result that target electronic device is sent, recognition result includes user speech input Corresponding control instruction.AR glasses execute the control instruction in recognition result.AR glasses will be known after receiving recognition result Other result is presented on prompt window, executes function corresponding with recognition result after user's confirmation.
Below in conjunction with example, audio recognition method provided by the embodiment of the present application is introduced: using mobile phone as AR The target electronic device of glasses, AR glasses are matched with mobile phone by Bluetooth function.Mobile phone opens client application and passes through nothing Line functionality scan peripheral equipment sends discovering device and requests to give AR glasses.The service of AR glasses running background, which receives discovery, asks After asking, a facility information, the equipment that mobile phone is matched according to facility information selection are returned to cell phone client.AR glasses and mobile phone After successful matching, when user needs to assign instruction to AR glasses, for example music player is opened.User is by specifically grasping Make, for example nod, AR glasses can send speech recognition request to mobile phone, and mobile phone is in response to this speech recognition request and opens language Sound includes function and starts typing voice.The speech recognition included is generated corresponding control by cell phone client after voice is included System instructs and control instruction is sent to AR glasses.AR glasses execute control instruction after receiving control instruction.
It should be noted that in some implementations, other electronic equipment conducts can also be used in those skilled in the art The language energy identification module of wearable device.For example, using tablet computer as VR glasses or the speech recognition mould of AR glasses Block, can be with no restrictions to this application.
The embodiment of the present application also provides a kind of equipment of speech recognition, knows for realizing voice provided by above-described embodiment The correlation function of wearable device in other method.Specifically, as shown in figure 4, wearable for one kind provided by the embodiments of the present application The structural schematic diagram of equipment.Wherein, which includes the first transmission unit 401, the first receiving unit 402, first Processing unit 403.Wherein:
First transmission unit 401, for sending speech recognition request to target electronic device, speech recognition request is for making Target electronic device includes user speech and generates recognition result according to user speech.
First receiving unit 402, for receiving recognition result transmitted by target electronic device.
First processing units 403, for after the first receiving unit 402 receives recognition result, handling recognition result, and Execute function corresponding with recognition result.
First receiving unit 402 is also used to receive the first operation, the second operation and third operation of user.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, as shown in figure 5, being the structure of another wearable device provided by the embodiments of the present application Schematic diagram.Wherein, wearable device 50 includes: processor 501, memory 502, communication interface 503 and bus 504.Communication connects Mouth 503 is for wearable device 50 and other equipment or network communication, and memory 502 is for storing one or more programs.It should One or more programs include computer executed instructions.Processor 501 executes the computer execution that memory 502 stores and refers to It enables, so that wearable device 50 executes the audio recognition method in above-described embodiment.
Wherein, above-mentioned processor 501 may be implemented or execute various exemplary in conjunction with described in present disclosure Logic block, unit and circuit.The processor 501 can be central processing unit, general processor, digital signal processor, Specific integrated circuit, field programmable gate array or other programmable logic device, transistor logic, hardware component or Person's any combination thereof.It, which may be implemented or executes, combines various illustrative logic blocks described in present disclosure, Unit and circuit.The processor 501 is also possible to realize the combination of computing function, such as includes one or more microprocessors Combination, DSP and the combination of microprocessor etc..
Memory 502 may include volatile memory, such as random access memory.The memory 502 also can wrap Include nonvolatile memory, such as read-only memory, flash memory, hard disk or solid state hard disk.Memory 502 includes above-mentioned kind The combination of the memory of class.
Bus 504 can be expanding the industrial standard structure (Extended Industry Standard Architecture, EISA) bus etc..Bus 504 can be divided into address bus, data/address bus, control bus etc..For convenient for table Show, Fig. 5 is only indicated with a thick line, it is not intended that an only bus or a type of bus.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, the application also provides a kind of computer readable storage medium, computer-readable storage medium Instruction is stored in matter, when computer executes the instruction, which executes method flow shown in above method embodiment Each step performed by middle wearable device.
Wherein, computer readable storage medium, such as electricity, magnetic, optical, electromagnetic, infrared ray can be but not limited to or partly led System, device or the device of body, or any above combination.The more specific example of computer readable storage medium is (non-poor The list of act) it include: the electrical connection with one or more conducting wires, portable computer diskette, hard disk.Random access memory (Random Access Memory, RAM), read-only memory (Read-Only Memory, ROM), erasable type may be programmed read-only It is memory (Erasable Programmable Read Only Memory, EPROM), register, hard disk, optical fiber, portable Compact disc read-only memory (Compact Disc Read-Only Memory, CD-ROM), light storage device, magnetic memory Part or above-mentioned people are in appropriate combination or the computer readable storage medium of any other form of this field numerical value. A kind of illustrative storage medium is coupled to processor, to enable a processor to from the read information, and can be to Information is written in the storage medium.Certainly, storage medium is also possible to the component part of processor.Pocessor and storage media can be with In application-specific IC (Application Specific Integrated Circuit, ASIC).In the application In embodiment, computer readable storage medium can be any tangible medium for including or store program, which can be referred to Enable execution system, device or device use or in connection.
In another embodiment, the application also provides a kind of computer program product, when instruction is on wearable device When operation, so that wearable device executes step performed by wearable device in the method for speech recognition as shown in Figure 1.
The embodiment of the present application also provides a kind of equipment of speech recognition, knows for realizing voice provided by above-described embodiment The correlation function of electronic equipment in other method.Specifically, as shown in fig. 6, being a kind of electronic equipment provided by the embodiments of the present application Structural schematic diagram.Wherein, which includes the second transmission unit 601, the second receiving unit 602, second processing list Member 603.
Wherein:
Second receiving unit 602, for receiving the speech recognition request of wearable device transmission, speech recognition request is used for So that target electronic device is included user speech and recognition result is generated according to user speech.
Second transmission unit 601, for sending recognition result to user equipment.
The second processing unit 603, for including user's language after the second receiving unit 602 receives speech recognition request Sound simultaneously handles user speech and is converted into corresponding text or control instruction.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, as shown in fig. 7, the structure for another electronic equipment provided by the embodiments of the present application is shown It is intended to.Wherein, electronic equipment 70 includes: processor 701, memory 702 and communication interface 703.Wherein, communication interface 703 is used In electronic equipment 70 and other equipment or network communication, memory is for storing one or more programs.The one or more journey Sequence includes computer executed instructions, and when electronic equipment 70 is run, the computer execution that processor executes memory storage refers to It enables, so that electronic equipment 70 executes the audio recognition method in above-described embodiment.
Wherein, above-mentioned processor 701 may be implemented or execute various exemplary in conjunction with described in present disclosure Logic block, unit and circuit.The processor 701 can be central processing unit, general processor, digital signal processor, Specific integrated circuit, field programmable gate array or other programmable logic device, transistor logic, hardware component or Person's any combination thereof.It, which may be implemented or executes, combines various illustrative logic blocks described in present disclosure, Unit and circuit.The processor 701 is also possible to realize the combination of computing function, such as includes one or more microprocessors Combination, DSP and the combination of microprocessor etc..
Memory 702 may include volatile memory, such as random access memory.The memory 702 also can wrap Include nonvolatile memory, such as read-only memory, flash memory, hard disk or solid state hard disk.Memory 702 includes above-mentioned kind The combination of the memory of class.
Bus 704 can be expanding the industrial standard structure (Extended Industry Standard Architecture, EISA) bus etc..Bus 704 can be divided into address bus, data/address bus, control bus etc..For convenient for table Show, Fig. 7 is only indicated with a thick line, it is not intended that an only bus or a type of bus.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, the application also provides a kind of computer readable storage medium, computer-readable storage medium Instruction is stored in matter, when computer executes the instruction, which executes method flow shown in above method embodiment Each step performed by middle electronic equipment.
Wherein, computer readable storage medium, such as electricity, magnetic, optical, electromagnetic, infrared ray can be but not limited to or partly led System, device or the device of body, or any above combination.The more specific example of computer readable storage medium is (non-poor The list of act) it include: the electrical connection with one or more conducting wires, portable computer diskette, hard disk.Random access memory (Random Access Memory, RAM), read-only memory (Read-Only Memory, ROM), erasable type may be programmed read-only It is memory (Erasable Programmable Read Only Memory, EPROM), register, hard disk, optical fiber, portable Compact disc read-only memory (Compact Disc Read-Only Memory, CD-ROM), light storage device, magnetic memory Part or above-mentioned people are in appropriate combination or the computer readable storage medium of any other form of this field numerical value. A kind of illustrative storage medium is coupled to processor, to enable a processor to from the read information, and can be to Information is written in the storage medium.Certainly, storage medium is also possible to the component part of processor.Pocessor and storage media can be with In application-specific IC (Application Specific Integrated Circuit, ASIC).In the application In embodiment, computer readable storage medium can be any tangible medium for including or store program, which can be referred to Enable execution system, device or device use or in connection.
In another embodiment, the application also provides a kind of computer program product, when instruction is transported on an electronic device When row, so that network side equipment executes step performed by electronic equipment in audio recognition method as shown in Figure 2.
By in an embodiment of the present invention audio recognition method and equipment, computer readable storage medium, computer journey Sequence product can be applied to the above method, therefore, can be obtained technical effect see also above method embodiment, this hair Details are not described herein for bright embodiment.
The above, the only specific embodiment of the application, but the protection scope of the application is not limited thereto, it is any Change or replacement within the technical scope of the present application should all be covered within the scope of protection of this application.
In several embodiments provided herein, it should be understood that disclosed method and apparatus can pass through it Its mode is realized.For example, apparatus embodiments described above are merely indicative, for example, the division of the unit, only Only a kind of logical function partition, there may be another division manner in actual implementation, such as multiple units or components can be tied Another system is closed or is desirably integrated into, or some features can be ignored or not executed.Another point, it is shown or discussed Mutual coupling, direct-coupling or communication connection can be through some interfaces, the INDIRECT COUPLING or logical of equipment or unit Letter connection can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme 's.
It, can also be in addition, each functional unit in each embodiment of the application can integrate in one processing unit It is that each unit physically exists alone, can also be integrated in one unit with two or more units.
In the above-described embodiments, can come wholly or partly by software, hardware, firmware or any combination thereof real It is existing.When being realized using software program, can entirely or partly realize in the form of a computer program product.The computer Program product includes one or more computer instructions.On computers load and execute computer program instructions when, all or It partly generates according to process or function described in the embodiment of the present application.The computer can be general purpose computer, dedicated meter Calculation machine, computer network or other programmable devices.The computer instruction can store in computer readable storage medium In, or from a computer readable storage medium to the transmission of another computer readable storage medium, for example, the computer Instruction can pass through wired (such as coaxial cable, optical fiber, number from a web-site, computer, server or data center Word user line (Digital Subscriber Line, DSL)) or wireless (such as infrared, wireless, microwave etc.) mode to another A web-site, computer, server or data center are transmitted.The computer readable storage medium can be computer Any usable medium that can be accessed either includes the numbers such as one or more server, data centers that medium can be used to integrate According to storage equipment.The usable medium can be magnetic medium (for example, floppy disk, hard disk, tape), optical medium (for example, DVD), Or semiconductor medium (such as solid state hard disk (Solid State Disk, SSD)) etc..
The above, the only specific embodiment of the application, but the protection scope of the application is not limited thereto, it is any Those familiar with the art within the technical scope of the present application, can easily think of the change or the replacement, and should all contain Lid is within the scope of protection of this application.Therefore, the protection scope of the application should be based on the protection scope of the described claims.

Claims (10)

1. a kind of audio recognition method is applied to wearable device, the wearable device includes Virtual Reality product or increasing One of strong reality AR product, which is characterized in that the described method includes:
Speech recognition request is sent to target electronic device;The speech recognition request is for including the target electronic device User speech simultaneously generates recognition result according to the user speech, and the recognition result is sent to the wearable device;
The recognition result that the target electronic device is sent is received, and executes function corresponding with the recognition result.
2. audio recognition method according to claim 1, which is characterized in that the method also includes:
Receive the first operation that user treats input area;The region to be entered includes: display circle of the wearable device Need to input the region of text on face;
In response to first operation, shown on the display interface for prompting user to pass through the target electronic device The prompt information of inputting word information;
The transmission speech recognition request is specifically included to target electronic device: in response to user to the of the prompt information Two operations, send the speech recognition request;The speech recognition request, specifically for making the target electronic device include use Family voice simultaneously generates corresponding text according to the user speech, and the text is sent to the wearable device;
The recognition result for receiving the target electronic device and sending, and execute function corresponding with the recognition result Can, it specifically includes: receiving the text, and the text is inputted in the region to be entered.
3. audio recognition method according to claim 1, which is characterized in that the method also includes:
User is received to operate the third of the wearable device;
The transmission speech recognition request is specifically included to target electronic device: being operated in response to the third, is sent institute's predicate Sound identification request;The speech recognition request, specifically for making the target electronic device include user speech and according to described User speech generates corresponding control instruction, and the control instruction is sent to the wearable device;
The recognition result for receiving the target electronic device and sending, and execute function corresponding with the recognition result Can, it specifically includes: receiving the control instruction, and execute the control instruction.
4. audio recognition method described in -3 any bars according to claim 1, which is characterized in that
The transmission speech recognition request is to target electronic device, comprising: the wearable device sends the speech recognition and asks It asks to after the target electronic device, display reminding window in display interface;In the prompt window comprising cancel key, when with Family instructs after clicking the cancellation key to target electronic device transmission cancellation, so that the target electronic device stops Include user speech;
The recognition result for receiving the target electronic device and sending, and execute function corresponding with the recognition result It can, comprising: recognition result is presented on the prompt window by the wearable device after receiving the recognition result, Function corresponding with the recognition result is executed after user's confirmation.
5. a kind of audio recognition method is applied to target electronic device, which is characterized in that the described method includes:
Receive the speech recognition request that wearable device is sent;
In response to the speech recognition request, user speech is included;
Recognition result is generated according to the user speech;
The recognition result is sent to the wearable device, the recognition result is for making the wearable device execution and institute State the corresponding function of recognition result.
6. audio recognition method according to claim 5, which is characterized in that
In response to the speech recognition request, user speech is included, is specifically included:
The target electronic device issues standby signal after receiving the speech recognition request;The standby signal is at least Including vibration or playing alert tones;Include the user speech after the standby signal issues.
7. audio recognition method according to claim 5, which is characterized in that the method also includes:
When the wearable device and after the target electronic device successful connection, control the target electronic device enter it is specific State;Wherein, under the particular state, the display interface of the target electronic device shows screen locking picture and the display The brightness at interface is in minimum state;Also, under the particular state, the target electronic device is able to carry out the reception Speech recognition request that the wearable device is sent, it is described include user speech, described generated according to the user speech is known Other result, the transmission recognition result are to the wearable device.
8. according to audio recognition method described in claim 5-7 any bar, which is characterized in that receiving the wearable device Before the speech recognition request of transmission, the method also includes:
User is received to the 4th of target electronic device operation, request voice recognition instruction is sent and wearable is set to described It is standby;
The speech recognition request for receiving the wearable device and sending, comprising: request voice recognition instruction in the transmission Afterwards, the speech recognition request that the wearable device is sent is received.
9. a kind of wearable device characterized by comprising processor, memory and communication interface;Wherein, communication interface is used In the wearable device and other equipment or network communication;The memory is for storing one or more programs, and this Or multiple programs include computer executed instructions, when wearable device operation, the processor executes the memory Computer executed instructions of storage, so that the wearable device perform claim requires voice described in any one of 1-4 to know Other method.
10. a kind of electronic equipment characterized by comprising processor, memory and communication interface;Wherein, communication interface is used for The electronic equipment and other equipment or network communication;The memory is for storing one or more programs, and this or more A program includes computer executed instructions, and when electronic equipment operation, the processor executes the memory storage The computer executed instructions, so that the electronic equipment perform claim requires audio recognition method described in any one of 5-8.
CN201910785213.2A 2019-08-23 2019-08-23 A kind of audio recognition method and equipment Pending CN110444211A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910785213.2A CN110444211A (en) 2019-08-23 2019-08-23 A kind of audio recognition method and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910785213.2A CN110444211A (en) 2019-08-23 2019-08-23 A kind of audio recognition method and equipment

Publications (1)

Publication Number Publication Date
CN110444211A true CN110444211A (en) 2019-11-12

Family

ID=68437396

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910785213.2A Pending CN110444211A (en) 2019-08-23 2019-08-23 A kind of audio recognition method and equipment

Country Status (1)

Country Link
CN (1) CN110444211A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111694605A (en) * 2020-05-18 2020-09-22 Oppo广东移动通信有限公司 Voice information processing method and device, storage medium and electronic equipment
CN111782053A (en) * 2020-08-10 2020-10-16 Oppo广东移动通信有限公司 Model editing method, device, equipment and storage medium
CN112885356A (en) * 2021-01-29 2021-06-01 焦作大学 Voice recognition method based on voiceprint
CN113986011A (en) * 2021-10-27 2022-01-28 北京乐驾科技有限公司 Input method and device for AR glasses, storage medium and electronic device

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111694605A (en) * 2020-05-18 2020-09-22 Oppo广东移动通信有限公司 Voice information processing method and device, storage medium and electronic equipment
CN111782053A (en) * 2020-08-10 2020-10-16 Oppo广东移动通信有限公司 Model editing method, device, equipment and storage medium
CN112885356A (en) * 2021-01-29 2021-06-01 焦作大学 Voice recognition method based on voiceprint
CN112885356B (en) * 2021-01-29 2021-09-24 焦作大学 Voice recognition method based on voiceprint
CN113986011A (en) * 2021-10-27 2022-01-28 北京乐驾科技有限公司 Input method and device for AR glasses, storage medium and electronic device

Similar Documents

Publication Publication Date Title
CN110444211A (en) A kind of audio recognition method and equipment
CN107851144B (en) Request the user interface of the equipment of remote authorization
CN109656512A (en) Exchange method, device, storage medium and terminal based on voice assistant
EP3584787A1 (en) Headless task completion within digital personal assistants
CN106131057B (en) Certification based on virtual reality scenario and device
CN107430858A (en) The metadata of transmission mark current speaker
CN106575243A (en) Hypervisor-hosted virtual machine forensics
CN109994108A (en) Full-duplex communication technology for the session talk between chat robots and people
CN108133707A (en) A kind of content share method and system
CN107967055A (en) A kind of man-machine interaction method, terminal and computer-readable medium
CN106462713B (en) The interface display method and terminal of terminal
CN107209905A (en) For personalized and task completion service, correspondence spends theme and sorted out
CN107870711A (en) Page navigation method, the method and client that user interface is provided
CN103902056B (en) Dummy keyboard input method, equipment and system
CN108446320A (en) A kind of data processing method, device and the device for data processing
CN108229535A (en) Relate to yellow image audit method, apparatus, computer equipment and storage medium
CN109474658A (en) Electronic equipment, server and the recording medium of task run are supported with external equipment
US11233897B1 (en) Secure call center communications
CN108345667A (en) A kind of searching method and relevant apparatus
CN110047484A (en) A kind of speech recognition exchange method, system, equipment and storage medium
US20130120249A1 (en) Electronic device
CN109144458A (en) For executing the electronic equipment for inputting corresponding operation with voice
CN109448727A (en) Voice interactive method and device
US20230131975A1 (en) Music playing method and apparatus based on user interaction, and device and storage medium
CN109547836A (en) Exchange method and device, electronic equipment, storage medium is broadcast live

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 266555 Qingdao economic and Technological Development Zone, Shandong, Hong Kong Road, No. 218

Applicant after: Hisense Video Technology Co.,Ltd.

Address before: 266555 Qingdao economic and Technological Development Zone, Shandong, Hong Kong Road, No. 218

Applicant before: QINGDAO HISENSE ELECTRONICS Co.,Ltd.

SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination