CN110444211A - A kind of audio recognition method and equipment - Google Patents
A kind of audio recognition method and equipment Download PDFInfo
- Publication number
- CN110444211A CN110444211A CN201910785213.2A CN201910785213A CN110444211A CN 110444211 A CN110444211 A CN 110444211A CN 201910785213 A CN201910785213 A CN 201910785213A CN 110444211 A CN110444211 A CN 110444211A
- Authority
- CN
- China
- Prior art keywords
- electronic device
- target electronic
- wearable device
- speech
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Abstract
This application discloses a kind of audio recognition method and equipment, it is related to field of speech recognition, for using target electronic device as the speech recognition module of wearable device, wearable device to be without increasing additional module, the cost and weight for reducing wearable device, improve the usage experience of user.This method comprises: sending speech recognition request to target electronic device by wearable device, after target electronic device receives speech recognition request, the voice input of user's sending is included.Target electronic device identify to user speech and recognition result is sent to wearable device.Wearable device receives recognition result and executes corresponding function in recognition result.The application is applied to speech recognition.
Description
Technical field
The present invention relates to field of speech recognition more particularly to a kind of audio recognition method and equipment.
Background technique
Most of wearable device haves the defects that very big that man-machine interaction mode is cumbersome in man-machine interaction mode at present
It is complicated.Especially when user needs to carry out text input or assigns instruction, user needs positioning cursor repeatedly
To the key of dummy keyboard, such input mode extremely very complicated greatly reduces the usage experience of user.Existing rank
Although section has had manufacturer to propose the technical solution of wearable device integrating speech sound identifying system, this technical solution is needed
Dedicated speech recognition system is configured for wearable device, comprising: integrated identification chip, audio signal reception device etc..It in this way can be into one
The cost and weight of the increase wearable device of step, also will increase the calculation amount of wearable device, reduce calculated performance.
Summary of the invention
Embodiments herein provides a kind of audio recognition method and equipment, can for solving that user in the prior art wears
When wearable device, text inputs or assigns the problem of instruction process very complicated inconvenience.
In order to achieve the above objectives, embodiments herein adopts the following technical scheme that
In a first aspect, this method is applied to wearable device this application provides a kind of audio recognition method and equipment, it should
Method includes: to send speech recognition request to target electronic device, and speech recognition request is for making target electronic device include use
Family voice simultaneously generates recognition result according to user speech, and recognition result is sent to wearable device.Target electronic is received to set
The recognition result that preparation is sent, and execute function corresponding with recognition result.
Second aspect, this application provides a kind of audio recognition method and equipment, this method is applied to target electronic device,
This method comprises: receiving the speech recognition request that wearable device is sent includes user speech in response to speech recognition request.
Recognition result is generated according to user speech.Recognition result is sent to wearable device, recognition result is for holding wearable device
Row function corresponding with recognition result.
The third aspect, this application provides a kind of wearable device, the wearable device include: processor, memory and
Communication interface.Wherein, communication interface is for wearable device and other equipment or network communication.Memory for store one or
Multiple programs, which includes computer executed instructions, and when wearable device operation, processor executes storage
Computer executed instructions of device storage, so that wearable device executes the audio recognition method of above-mentioned first aspect.
Fourth aspect, this application provides a kind of electronic equipment, which includes: processor, memory and communication
Interface.Wherein, communication interface is for electronic equipment and other equipment or network communication.Memory is for storing one or more journeys
Sequence, which includes computer executed instructions, and when electronic equipment operation, processor executes memory storage
Computer executed instructions, so that electronic equipment executes the audio recognition method of above-mentioned second aspect.
5th aspect, this application provides a kind of corresponding computers for storing one or more programs of wearable device can
Storage medium is read, one or more programs include instruction, and instruction executes wearable device when being executed by wearable device
State the audio recognition method of first aspect.
6th aspect, this application provides the computer-readable of the one or more programs of the corresponding storage of a kind of electronic equipment
Storage medium, one or more programs include instruction, and instruction makes electronic equipment execute above-mentioned second when being executed by electronic equipment
The audio recognition method of aspect.
A kind of audio recognition method of the target electronic device that embodiments herein provides as wearable device.Pass through
Wearable device sends speech recognition request to target electronic device, after target electronic device receives speech recognition request, receives
Employ the voice input of family sending.Target electronic device identify to user speech and recognition result is sent to wearable set
It is standby.Wearable device receives recognition result and executes corresponding function in recognition result.Present invention greatly reduces user with
The interaction of wearable device, especially in terms of text inputs or assigns instruction, using electronic equipment as wearable device
Speech recognition module does not increase the cost and weight of wearable device, greatly improves user without increasing additional module
Usage experience.
Detailed description of the invention
Fig. 1 is a kind of flow diagram for audio recognition method that embodiments herein provides;
Fig. 2 is the flow diagram for another audio recognition method that embodiments herein provides;
Fig. 3 is the interface display after a kind of target electronic device that embodiments herein provides is connect with wearable device
Screen locking picture drawing;
Fig. 4 is a kind of structural schematic diagram for wearable device that embodiments herein provides;
Fig. 5 is the structural schematic diagram for another wearable device that embodiments herein provides;
Fig. 6 is the structural schematic diagram for a kind of electronic equipment that embodiments herein provides;
Fig. 7 is the structural schematic diagram for another electronic equipment that embodiments herein provides.
Specific embodiment
Below first to the invention relates to some concepts simply introduced.
AR (AugmentedReality augmented reality), also referred to as mixed reality.It, will be virtual by computer technology
To real world, true environment and virtual object have been added in real time the same picture or space while to be deposited Information application
In.
VR (Virtual Reality virtual reality), specific intension are comprehensive utilization computer graphics systems and various
The interface equipments such as reality and control, the technology for immersing feeling is provided in three-dimensional environment generating on computers, can interacting.
By the way that various living scenes can be experienced in the case where staying indoors using VR product and AR product.Such as: In
When carrying out shopping online, clothing can be tried on by VR product and AR product;It is seen by VR product and AR product
Concert or sports tournament, can be to the feeling as a kind of scene of people;It is each with AR product to can see the world by VR product
The tourist attractions on ground, the feeling as on the spot in person.But existing VR product and AR product in man-machine interaction mode also
Very big defect, man-machine interaction mode are more complicated.It thus will affect the usage experience of user.Existing VR product and AR are produced
Product are human-computer interaction to be realized by virtual screen key, or pre-set voice on VR product and AR product mostly
Module or an external voice module realize human-computer interaction.Thus will increase the cost of VR product and AR product with
Weight can also reduce the performance of VR product Yu AR product.Wherein, meaning VR product in the application, is specifically as follows VR glasses.AR
Product then specifically can be the equipment such as AR glasses.
The embodiment of the present application provides a kind of audio recognition method and equipment, to solve the deficiencies in the prior art.
The application is not necessarily to pre-set voice module on VR product and AR product, does not also need external voice module.The application is by target
Speech recognition equipment of the electronic equipment as VR product and AR product, VR product and AR product are by way of wireless connection and mesh
Mark electronic equipment interacts.
Below in conjunction with the attached drawing in the embodiment of the present application, technical solutions in the embodiments of the present application is described.
In the description of the present application, unless otherwise indicated, "/" indicates the meaning of "or", for example, A/B can indicate A or B.
"and/or" herein is only a kind of incidence relation for describing affiliated partner, indicates may exist three kinds of relationships, for example, A
And/or B, can indicate: individualism A exists simultaneously A and B, these three situations of individualism B.In addition, "at least one" is
Refer to one or more, " multiple " refer to two or more.The printed words such as " first ", " second " are not to quantity and execution order
It is defined, and the printed words such as " first ", " second " also do not limit certain difference.
Technical solution provided by the embodiments of the present application can be applied to speech recognition of the electronic equipment as wearable device
Module.In the embodiment of the present application, in the specific implementation, it is each specifically to can be mobile phone, tablet computer or MP4 etc. for electronic equipment
Class has the function of that voice is included and the equipment of wireless connecting function.Which kind of equipment specifically to realize the embodiment of the present application is provided by
Audio recognition method in electronic equipment function, then can be selected according to concrete scene, can be with no restrictions to this application.
Embodiment one
When the wearable device in the embodiment of the present application is AR glasses, as shown in Figure 1, being provided by the embodiments of the present application
A kind of flow diagram of audio recognition method, this method comprises:
S101, AR glasses and target electronic device establish connection, and target electronic device enters particular state.
Specifically, AR glasses and target electronic device open the wireless connecting functions such as Bluetooth function, the direct-connected function of WIFI, or
Person accesses in same local area network.Target electronic device scans peripheral equipment by wireless connecting function and transmission is established connection and asked
Ask, AR glasses receive establish connection request after to electronic equipment returning equipment information.Target electronic device is sent out according to AR glasses
The facility information and AR glasses sent, which is established, to be matched.So far, successful connection has been established in AR glasses and target electronic device.
Wherein, it is connect to prompt ownership goal electronic equipment to establish with AR glasses, while alsoing for avoiding
Connection carries out maloperation to target electronic device after establishing, after AR glasses connect foundation with target electronic device, target electronic
Equipment enters particular state.Wherein, particular state is as shown in figure 3, the display interface of target electronic device shows screen locking picture simultaneously
And the brightness of display interface is in minimum state.Also, in a particular state, target electronic device is able to carry out following operation:
The speech recognition request of AR glasses transmission is received, user speech is included, recognition result is generated according to user speech, sends identification knot
Fruit operates to AR glasses etc., and the specific implementation content of aforesaid operations can refer to corresponding contents hereinafter.
Before the operation that AR glasses receive user, target electronic device receives user to the 4th behaviour of target electronic device
Make, sends request voice recognition instruction and give AR glasses.After AR glasses receive request voice recognition instruction, speech recognition request is sent
To target electronic device.
S102, AR glasses receive the first operation that user treats input area.
Wherein, region to be entered can specifically include: need to input the region of text on the display interface of AR glasses.So
It afterwards, can be by the behavior act of acquisition user, to determine that user treats the operation of input area.For example, user can pass through
The touch tablet in AR eyeglasses temple arm is arranged in touch-control, thus the cursor on mobile display interface, and be moved to by cursor to defeated
After entering region, input area can be treated and carry out such as other operations of input text.For another example detection user's eye can also be passed through
The focus of eyeball, to determine eyes of user position of interest.When determining that eyes of user watches attentively at region to be entered, then can treat
Input area carries out other subsequent operations.
Illustratively, the first operation that user treats input area may include: that user positions cursor to area to be entered
Domain.Alternatively, the first operation can also include: user's long-pressing region to be entered.
S103, AR glasses are shown in the display interface in response to the first operation for prompting user to pass through target electronic
The prompt information of equipment inputting word information.
For example, cursor is positioned to region to be entered, will pop up one on display interface when user is by the first operation
Indicate the pop-up of " whether needing to input text by voice ".
The second operation is carried out (for example, user clicks the determination in pop-up to prompt information shown on interface in user
Key) after, method provided by the embodiment of the present application is specific further comprising the steps of:
S104, AR glasses send speech recognition request to target electronic device, and target electronic device receives speech recognition and asks
It asks.
Specifically, step S104 includes: of AR glasses in response to user to prompt information after executing step S103
Two operations send speech recognition request to target electronic device.
Wherein, speech recognition request is identified for making target electronic device include user speech and be generated according to user speech
As a result, and recognition result is sent to wearable device, i.e., the AR glasses in the present embodiment.
Specifically, in the present embodiment, speech recognition request, specifically for making target electronic device include user speech simultaneously
It generates corresponding text according to user speech, and text is sent to wearable device, i.e., the AR glasses in the present embodiment.In
After AR glasses send speech recognition request to target electronic device, display reminding window in display interface.Include in prompt window
Cancel key, is instructed after cancelling key when the user clicks to target electronic device transmission cancellation, so that target electronic device is stopped
Only include user speech.
Illustratively, the prompt information in step S103 can be with are as follows: display interface shows that one indicates and " whether needs to lead to
Cross voice input text " or " i.e. will by voice input text " pop-up.Alternatively, showing input method in the display interface
The key of speech voice input function is triggered at interface including one.
In turn, the second operation can be with are as follows: user clicks the determination key in pop-up, and AR glasses are clicked in user and determine key
Afterwards, speech recognition request is sent to target electronic device.
It should be noted that be in the embodiment of the present application by user first operation, second operation come trigger voice know
It does not invite and asks, in some implementations, other triggering modes can also be set according to specific implementation situation.When utilize other
When triggering mode triggers, then the content of S102 and S103 can not be executed.
S105, target electronic device include user speech in response to speech recognition request.
Specifically, target electronic device in the present embodiment can have for mobile phone, tablet computer etc. include voice and
Each class of electronic devices of speech identifying function.
In one implementation, it is contemplated that after user carries out the above-mentioned second operation to AR glasses, target electronic device
It is different surely to call sound-recording function in a short time, if therefore user carry out second operation after directly loquitur, may
Appearance can not be indexed to user speak audio the case where.Therefore, language is included in order to allow users to perceive target electronic device
The time of sound, in the embodiment of the present application, S105 is specifically included:
S1051, target electronic device issue standby signal after receiving the speech recognition request.
Specifically, standby signal at least may include vibration or playing alert tones.
S1052, the user speech after the standby signal issues is included.
Specifically, generating the recognition result containing text according to user speech after target electronic device includes user speech.
S106, target electronic device send recognition result and give AR glasses, and AR glasses receive recognition result and execute and identify
As a result corresponding function.
Wherein, recognition result is for making AR glasses execute function corresponding with recognition result.
Specifically, AR glasses receive the recognition result that target electronic device is sent, recognition result includes user speech input
Corresponding text.Text in recognition result is input in region to be entered by AR glasses.AR glasses are receiving identification knot
After fruit, recognition result is presented on prompt window, executes function corresponding with recognition result after user's confirmation.
Below in conjunction with example, audio recognition method provided by the embodiment of the present application is introduced: using mobile phone as AR
The target electronic device of glasses, AR glasses are matched with mobile phone by Bluetooth function.Mobile phone opens client application and passes through nothing
Line functionality scan peripheral equipment sends discovering device and requests to give AR glasses.The service of AR glasses running background, which receives discovery, asks
After asking, a facility information, the equipment that mobile phone is matched according to facility information selection are returned to cell phone client.AR glasses and mobile phone
After successful matching, when user needs inputting word information, user can show that one passes through by clicking input frame, AR glasses
The prompt information of mobile phone input text.Then AR glasses send speech recognition request to mobile phone, and mobile phone is known in response to this voice
It does not invite to seek and open voice and include function and starts typing voice.The speech recognition that cell phone client will be included after voice is included
It generates corresponding Text region result and recognition result is sent to AR glasses.AR glasses tie identification after receiving recognition result
Text corresponding to fruit is input in input frame.
It should be noted that in some implementations, other electronic equipment conducts can also be used in those skilled in the art
The language energy identification module of wearable device.For example, using tablet computer as VR glasses or the speech recognition mould of AR glasses
Block, can be with no restrictions to this application.
Embodiment two
When the wearable device in the embodiment of the present application is AR glasses, as shown in Fig. 2, being provided by the embodiments of the present application
The flow diagram of another audio recognition method, which comprises
S201, AR glasses and target electronic device establish connection, and target electronic device enters particular state.
Specifically, the particular content of the step can refer to step S101 content, do not illustrate herein.
S202, AR glasses receive the third operation of user.
Specifically, third operates are as follows: user nods, user shakes the head or presses specified key, after waking up AR glasses
Platform service.Third operation is to need operation when assigning control instruction to AR glasses in user.For example, when user is using AR
For mirror when watching music video, user can activate speech recognition request by third operation, to be set by target electronic
The standby control instruction assigned such as " switching next track " or " closing music video ".AR glasses can pass through the behaviour of acquisition user
Make, to determine whether user needs to assign control instruction to AR glasses by voice.Illustratively, user can be by by dividing into
The key in AR eyeglasses temple arm is set, to activate the speech recognition request of AR glasses, " is cut to be assigned by target electronic device
Change next track " control instruction.
S203, AR glasses are operated in response to third, are shown in the display interface for prompting user to pass through target electronic
Equipment assigns the prompt information of control instruction.
" finger will be assigned by voice for example, will pop up one after user finishes third operation, on display interface and indicate
Enable " pop-up.AR glasses send speech recognition request to target electronic device simultaneously.
S204, AR glasses send speech recognition request to target electronic device, and target electronic device receives speech recognition and asks
It asks.
Wherein, speech recognition request is controlled for making target electronic device include user speech and be generated according to user speech
Instruction, and control instruction is sent to wearable device, i.e., the AR glasses in the present embodiment.
Specifically, in the present embodiment, speech recognition request, specifically for making target electronic device include user speech simultaneously
Control instruction is generated according to user speech, and control instruction is sent to AR glasses.AR glasses send speech recognition request to
After target electronic device, display reminding window in display interface.Comprising cancelling key in prompt window, cancels pressing when the user clicks
It is sent after key to target electronic device and cancels instruction, so that user speech is included in target electronic device stopping.
It should be noted that be that speech recognition request is triggered by the third operation of user in the embodiment of the present application,
In some implementations, other triggering modes can also be set according to specific implementation situation.It is touched when using other triggering modes
When hair, then the content of S202 and S203 can not be executed.
S205, target electronic device include user speech in response to speech recognition request.
Specifically, target electronic device in the present embodiment can have for mobile phone, tablet computer etc. include voice and
Each class of electronic devices of speech identifying function.
In one implementation, it is contemplated that after user carries out above-mentioned third operation to AR glasses, target electronic device
It is different surely to call sound-recording function in a short time, if therefore user directly loquitur after carrying out third operation, may
Appearance can not be indexed to user speak audio the case where.Therefore, language is included in order to allow users to perceive target electronic device
The time of sound, in the embodiment of the present application, S205 is specifically included:
S2051, target electronic device issue standby signal after receiving speech recognition request.
Specifically, standby signal at least may include vibration or playing alert tones.
S2052, the user speech after standby signal issues is included.
Specifically, generating the identification containing control instruction according to user speech after target electronic device includes user speech
As a result.
S206, target electronic device send recognition result and give AR glasses, and AR glasses receive recognition result and execute identification knot
Control instruction in fruit.
Wherein, the recognition result is for making the wearable device execute function corresponding with the recognition result.
Specifically, AR glasses receive the recognition result that target electronic device is sent, recognition result includes user speech input
Corresponding control instruction.AR glasses execute the control instruction in recognition result.AR glasses will be known after receiving recognition result
Other result is presented on prompt window, executes function corresponding with recognition result after user's confirmation.
Below in conjunction with example, audio recognition method provided by the embodiment of the present application is introduced: using mobile phone as AR
The target electronic device of glasses, AR glasses are matched with mobile phone by Bluetooth function.Mobile phone opens client application and passes through nothing
Line functionality scan peripheral equipment sends discovering device and requests to give AR glasses.The service of AR glasses running background, which receives discovery, asks
After asking, a facility information, the equipment that mobile phone is matched according to facility information selection are returned to cell phone client.AR glasses and mobile phone
After successful matching, when user needs to assign instruction to AR glasses, for example music player is opened.User is by specifically grasping
Make, for example nod, AR glasses can send speech recognition request to mobile phone, and mobile phone is in response to this speech recognition request and opens language
Sound includes function and starts typing voice.The speech recognition included is generated corresponding control by cell phone client after voice is included
System instructs and control instruction is sent to AR glasses.AR glasses execute control instruction after receiving control instruction.
It should be noted that in some implementations, other electronic equipment conducts can also be used in those skilled in the art
The language energy identification module of wearable device.For example, using tablet computer as VR glasses or the speech recognition mould of AR glasses
Block, can be with no restrictions to this application.
The embodiment of the present application also provides a kind of equipment of speech recognition, knows for realizing voice provided by above-described embodiment
The correlation function of wearable device in other method.Specifically, as shown in figure 4, wearable for one kind provided by the embodiments of the present application
The structural schematic diagram of equipment.Wherein, which includes the first transmission unit 401, the first receiving unit 402, first
Processing unit 403.Wherein:
First transmission unit 401, for sending speech recognition request to target electronic device, speech recognition request is for making
Target electronic device includes user speech and generates recognition result according to user speech.
First receiving unit 402, for receiving recognition result transmitted by target electronic device.
First processing units 403, for after the first receiving unit 402 receives recognition result, handling recognition result, and
Execute function corresponding with recognition result.
First receiving unit 402 is also used to receive the first operation, the second operation and third operation of user.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description
It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper
It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete
All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to
The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, as shown in figure 5, being the structure of another wearable device provided by the embodiments of the present application
Schematic diagram.Wherein, wearable device 50 includes: processor 501, memory 502, communication interface 503 and bus 504.Communication connects
Mouth 503 is for wearable device 50 and other equipment or network communication, and memory 502 is for storing one or more programs.It should
One or more programs include computer executed instructions.Processor 501 executes the computer execution that memory 502 stores and refers to
It enables, so that wearable device 50 executes the audio recognition method in above-described embodiment.
Wherein, above-mentioned processor 501 may be implemented or execute various exemplary in conjunction with described in present disclosure
Logic block, unit and circuit.The processor 501 can be central processing unit, general processor, digital signal processor,
Specific integrated circuit, field programmable gate array or other programmable logic device, transistor logic, hardware component or
Person's any combination thereof.It, which may be implemented or executes, combines various illustrative logic blocks described in present disclosure,
Unit and circuit.The processor 501 is also possible to realize the combination of computing function, such as includes one or more microprocessors
Combination, DSP and the combination of microprocessor etc..
Memory 502 may include volatile memory, such as random access memory.The memory 502 also can wrap
Include nonvolatile memory, such as read-only memory, flash memory, hard disk or solid state hard disk.Memory 502 includes above-mentioned kind
The combination of the memory of class.
Bus 504 can be expanding the industrial standard structure (Extended Industry Standard
Architecture, EISA) bus etc..Bus 504 can be divided into address bus, data/address bus, control bus etc..For convenient for table
Show, Fig. 5 is only indicated with a thick line, it is not intended that an only bus or a type of bus.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description
It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper
It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete
All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to
The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, the application also provides a kind of computer readable storage medium, computer-readable storage medium
Instruction is stored in matter, when computer executes the instruction, which executes method flow shown in above method embodiment
Each step performed by middle wearable device.
Wherein, computer readable storage medium, such as electricity, magnetic, optical, electromagnetic, infrared ray can be but not limited to or partly led
System, device or the device of body, or any above combination.The more specific example of computer readable storage medium is (non-poor
The list of act) it include: the electrical connection with one or more conducting wires, portable computer diskette, hard disk.Random access memory
(Random Access Memory, RAM), read-only memory (Read-Only Memory, ROM), erasable type may be programmed read-only
It is memory (Erasable Programmable Read Only Memory, EPROM), register, hard disk, optical fiber, portable
Compact disc read-only memory (Compact Disc Read-Only Memory, CD-ROM), light storage device, magnetic memory
Part or above-mentioned people are in appropriate combination or the computer readable storage medium of any other form of this field numerical value.
A kind of illustrative storage medium is coupled to processor, to enable a processor to from the read information, and can be to
Information is written in the storage medium.Certainly, storage medium is also possible to the component part of processor.Pocessor and storage media can be with
In application-specific IC (Application Specific Integrated Circuit, ASIC).In the application
In embodiment, computer readable storage medium can be any tangible medium for including or store program, which can be referred to
Enable execution system, device or device use or in connection.
In another embodiment, the application also provides a kind of computer program product, when instruction is on wearable device
When operation, so that wearable device executes step performed by wearable device in the method for speech recognition as shown in Figure 1.
The embodiment of the present application also provides a kind of equipment of speech recognition, knows for realizing voice provided by above-described embodiment
The correlation function of electronic equipment in other method.Specifically, as shown in fig. 6, being a kind of electronic equipment provided by the embodiments of the present application
Structural schematic diagram.Wherein, which includes the second transmission unit 601, the second receiving unit 602, second processing list
Member 603.
Wherein:
Second receiving unit 602, for receiving the speech recognition request of wearable device transmission, speech recognition request is used for
So that target electronic device is included user speech and recognition result is generated according to user speech.
Second transmission unit 601, for sending recognition result to user equipment.
The second processing unit 603, for including user's language after the second receiving unit 602 receives speech recognition request
Sound simultaneously handles user speech and is converted into corresponding text or control instruction.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description
It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper
It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete
All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to
The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, as shown in fig. 7, the structure for another electronic equipment provided by the embodiments of the present application is shown
It is intended to.Wherein, electronic equipment 70 includes: processor 701, memory 702 and communication interface 703.Wherein, communication interface 703 is used
In electronic equipment 70 and other equipment or network communication, memory is for storing one or more programs.The one or more journey
Sequence includes computer executed instructions, and when electronic equipment 70 is run, the computer execution that processor executes memory storage refers to
It enables, so that electronic equipment 70 executes the audio recognition method in above-described embodiment.
Wherein, above-mentioned processor 701 may be implemented or execute various exemplary in conjunction with described in present disclosure
Logic block, unit and circuit.The processor 701 can be central processing unit, general processor, digital signal processor,
Specific integrated circuit, field programmable gate array or other programmable logic device, transistor logic, hardware component or
Person's any combination thereof.It, which may be implemented or executes, combines various illustrative logic blocks described in present disclosure,
Unit and circuit.The processor 701 is also possible to realize the combination of computing function, such as includes one or more microprocessors
Combination, DSP and the combination of microprocessor etc..
Memory 702 may include volatile memory, such as random access memory.The memory 702 also can wrap
Include nonvolatile memory, such as read-only memory, flash memory, hard disk or solid state hard disk.Memory 702 includes above-mentioned kind
The combination of the memory of class.
Bus 704 can be expanding the industrial standard structure (Extended Industry Standard
Architecture, EISA) bus etc..Bus 704 can be divided into address bus, data/address bus, control bus etc..For convenient for table
Show, Fig. 7 is only indicated with a thick line, it is not intended that an only bus or a type of bus.
Through the above description of the embodiments, it is apparent to those skilled in the art that, for description
It is convenienct and succinct, only with the division progress of above-mentioned each functional unit for example, in practical application, can according to need and will be upper
It states function distribution to be completed by different functional units, i.e., the internal structure of device is divided into different functional units, to complete
All or part of function described above.The specific work process of the method for foregoing description, device and unit, before can referring to
The corresponding process in embodiment of the method is stated, details are not described herein.
In another embodiment, the application also provides a kind of computer readable storage medium, computer-readable storage medium
Instruction is stored in matter, when computer executes the instruction, which executes method flow shown in above method embodiment
Each step performed by middle electronic equipment.
Wherein, computer readable storage medium, such as electricity, magnetic, optical, electromagnetic, infrared ray can be but not limited to or partly led
System, device or the device of body, or any above combination.The more specific example of computer readable storage medium is (non-poor
The list of act) it include: the electrical connection with one or more conducting wires, portable computer diskette, hard disk.Random access memory
(Random Access Memory, RAM), read-only memory (Read-Only Memory, ROM), erasable type may be programmed read-only
It is memory (Erasable Programmable Read Only Memory, EPROM), register, hard disk, optical fiber, portable
Compact disc read-only memory (Compact Disc Read-Only Memory, CD-ROM), light storage device, magnetic memory
Part or above-mentioned people are in appropriate combination or the computer readable storage medium of any other form of this field numerical value.
A kind of illustrative storage medium is coupled to processor, to enable a processor to from the read information, and can be to
Information is written in the storage medium.Certainly, storage medium is also possible to the component part of processor.Pocessor and storage media can be with
In application-specific IC (Application Specific Integrated Circuit, ASIC).In the application
In embodiment, computer readable storage medium can be any tangible medium for including or store program, which can be referred to
Enable execution system, device or device use or in connection.
In another embodiment, the application also provides a kind of computer program product, when instruction is transported on an electronic device
When row, so that network side equipment executes step performed by electronic equipment in audio recognition method as shown in Figure 2.
By in an embodiment of the present invention audio recognition method and equipment, computer readable storage medium, computer journey
Sequence product can be applied to the above method, therefore, can be obtained technical effect see also above method embodiment, this hair
Details are not described herein for bright embodiment.
The above, the only specific embodiment of the application, but the protection scope of the application is not limited thereto, it is any
Change or replacement within the technical scope of the present application should all be covered within the scope of protection of this application.
In several embodiments provided herein, it should be understood that disclosed method and apparatus can pass through it
Its mode is realized.For example, apparatus embodiments described above are merely indicative, for example, the division of the unit, only
Only a kind of logical function partition, there may be another division manner in actual implementation, such as multiple units or components can be tied
Another system is closed or is desirably integrated into, or some features can be ignored or not executed.Another point, it is shown or discussed
Mutual coupling, direct-coupling or communication connection can be through some interfaces, the INDIRECT COUPLING or logical of equipment or unit
Letter connection can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit
The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple
In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme
's.
It, can also be in addition, each functional unit in each embodiment of the application can integrate in one processing unit
It is that each unit physically exists alone, can also be integrated in one unit with two or more units.
In the above-described embodiments, can come wholly or partly by software, hardware, firmware or any combination thereof real
It is existing.When being realized using software program, can entirely or partly realize in the form of a computer program product.The computer
Program product includes one or more computer instructions.On computers load and execute computer program instructions when, all or
It partly generates according to process or function described in the embodiment of the present application.The computer can be general purpose computer, dedicated meter
Calculation machine, computer network or other programmable devices.The computer instruction can store in computer readable storage medium
In, or from a computer readable storage medium to the transmission of another computer readable storage medium, for example, the computer
Instruction can pass through wired (such as coaxial cable, optical fiber, number from a web-site, computer, server or data center
Word user line (Digital Subscriber Line, DSL)) or wireless (such as infrared, wireless, microwave etc.) mode to another
A web-site, computer, server or data center are transmitted.The computer readable storage medium can be computer
Any usable medium that can be accessed either includes the numbers such as one or more server, data centers that medium can be used to integrate
According to storage equipment.The usable medium can be magnetic medium (for example, floppy disk, hard disk, tape), optical medium (for example, DVD),
Or semiconductor medium (such as solid state hard disk (Solid State Disk, SSD)) etc..
The above, the only specific embodiment of the application, but the protection scope of the application is not limited thereto, it is any
Those familiar with the art within the technical scope of the present application, can easily think of the change or the replacement, and should all contain
Lid is within the scope of protection of this application.Therefore, the protection scope of the application should be based on the protection scope of the described claims.
Claims (10)
1. a kind of audio recognition method is applied to wearable device, the wearable device includes Virtual Reality product or increasing
One of strong reality AR product, which is characterized in that the described method includes:
Speech recognition request is sent to target electronic device;The speech recognition request is for including the target electronic device
User speech simultaneously generates recognition result according to the user speech, and the recognition result is sent to the wearable device;
The recognition result that the target electronic device is sent is received, and executes function corresponding with the recognition result.
2. audio recognition method according to claim 1, which is characterized in that the method also includes:
Receive the first operation that user treats input area;The region to be entered includes: display circle of the wearable device
Need to input the region of text on face;
In response to first operation, shown on the display interface for prompting user to pass through the target electronic device
The prompt information of inputting word information;
The transmission speech recognition request is specifically included to target electronic device: in response to user to the of the prompt information
Two operations, send the speech recognition request;The speech recognition request, specifically for making the target electronic device include use
Family voice simultaneously generates corresponding text according to the user speech, and the text is sent to the wearable device;
The recognition result for receiving the target electronic device and sending, and execute function corresponding with the recognition result
Can, it specifically includes: receiving the text, and the text is inputted in the region to be entered.
3. audio recognition method according to claim 1, which is characterized in that the method also includes:
User is received to operate the third of the wearable device;
The transmission speech recognition request is specifically included to target electronic device: being operated in response to the third, is sent institute's predicate
Sound identification request;The speech recognition request, specifically for making the target electronic device include user speech and according to described
User speech generates corresponding control instruction, and the control instruction is sent to the wearable device;
The recognition result for receiving the target electronic device and sending, and execute function corresponding with the recognition result
Can, it specifically includes: receiving the control instruction, and execute the control instruction.
4. audio recognition method described in -3 any bars according to claim 1, which is characterized in that
The transmission speech recognition request is to target electronic device, comprising: the wearable device sends the speech recognition and asks
It asks to after the target electronic device, display reminding window in display interface;In the prompt window comprising cancel key, when with
Family instructs after clicking the cancellation key to target electronic device transmission cancellation, so that the target electronic device stops
Include user speech;
The recognition result for receiving the target electronic device and sending, and execute function corresponding with the recognition result
It can, comprising: recognition result is presented on the prompt window by the wearable device after receiving the recognition result,
Function corresponding with the recognition result is executed after user's confirmation.
5. a kind of audio recognition method is applied to target electronic device, which is characterized in that the described method includes:
Receive the speech recognition request that wearable device is sent;
In response to the speech recognition request, user speech is included;
Recognition result is generated according to the user speech;
The recognition result is sent to the wearable device, the recognition result is for making the wearable device execution and institute
State the corresponding function of recognition result.
6. audio recognition method according to claim 5, which is characterized in that
In response to the speech recognition request, user speech is included, is specifically included:
The target electronic device issues standby signal after receiving the speech recognition request;The standby signal is at least
Including vibration or playing alert tones;Include the user speech after the standby signal issues.
7. audio recognition method according to claim 5, which is characterized in that the method also includes:
When the wearable device and after the target electronic device successful connection, control the target electronic device enter it is specific
State;Wherein, under the particular state, the display interface of the target electronic device shows screen locking picture and the display
The brightness at interface is in minimum state;Also, under the particular state, the target electronic device is able to carry out the reception
Speech recognition request that the wearable device is sent, it is described include user speech, described generated according to the user speech is known
Other result, the transmission recognition result are to the wearable device.
8. according to audio recognition method described in claim 5-7 any bar, which is characterized in that receiving the wearable device
Before the speech recognition request of transmission, the method also includes:
User is received to the 4th of target electronic device operation, request voice recognition instruction is sent and wearable is set to described
It is standby;
The speech recognition request for receiving the wearable device and sending, comprising: request voice recognition instruction in the transmission
Afterwards, the speech recognition request that the wearable device is sent is received.
9. a kind of wearable device characterized by comprising processor, memory and communication interface;Wherein, communication interface is used
In the wearable device and other equipment or network communication;The memory is for storing one or more programs, and this
Or multiple programs include computer executed instructions, when wearable device operation, the processor executes the memory
Computer executed instructions of storage, so that the wearable device perform claim requires voice described in any one of 1-4 to know
Other method.
10. a kind of electronic equipment characterized by comprising processor, memory and communication interface;Wherein, communication interface is used for
The electronic equipment and other equipment or network communication;The memory is for storing one or more programs, and this or more
A program includes computer executed instructions, and when electronic equipment operation, the processor executes the memory storage
The computer executed instructions, so that the electronic equipment perform claim requires audio recognition method described in any one of 5-8.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910785213.2A CN110444211A (en) | 2019-08-23 | 2019-08-23 | A kind of audio recognition method and equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910785213.2A CN110444211A (en) | 2019-08-23 | 2019-08-23 | A kind of audio recognition method and equipment |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110444211A true CN110444211A (en) | 2019-11-12 |
Family
ID=68437396
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910785213.2A Pending CN110444211A (en) | 2019-08-23 | 2019-08-23 | A kind of audio recognition method and equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110444211A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111694605A (en) * | 2020-05-18 | 2020-09-22 | Oppo广东移动通信有限公司 | Voice information processing method and device, storage medium and electronic equipment |
CN111782053A (en) * | 2020-08-10 | 2020-10-16 | Oppo广东移动通信有限公司 | Model editing method, device, equipment and storage medium |
CN112885356A (en) * | 2021-01-29 | 2021-06-01 | 焦作大学 | Voice recognition method based on voiceprint |
CN113986011A (en) * | 2021-10-27 | 2022-01-28 | 北京乐驾科技有限公司 | Input method and device for AR glasses, storage medium and electronic device |
-
2019
- 2019-08-23 CN CN201910785213.2A patent/CN110444211A/en active Pending
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111694605A (en) * | 2020-05-18 | 2020-09-22 | Oppo广东移动通信有限公司 | Voice information processing method and device, storage medium and electronic equipment |
CN111782053A (en) * | 2020-08-10 | 2020-10-16 | Oppo广东移动通信有限公司 | Model editing method, device, equipment and storage medium |
CN112885356A (en) * | 2021-01-29 | 2021-06-01 | 焦作大学 | Voice recognition method based on voiceprint |
CN112885356B (en) * | 2021-01-29 | 2021-09-24 | 焦作大学 | Voice recognition method based on voiceprint |
CN113986011A (en) * | 2021-10-27 | 2022-01-28 | 北京乐驾科技有限公司 | Input method and device for AR glasses, storage medium and electronic device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110444211A (en) | A kind of audio recognition method and equipment | |
CN107851144B (en) | Request the user interface of the equipment of remote authorization | |
CN109656512A (en) | Exchange method, device, storage medium and terminal based on voice assistant | |
EP3584787A1 (en) | Headless task completion within digital personal assistants | |
CN106131057B (en) | Certification based on virtual reality scenario and device | |
CN107430858A (en) | The metadata of transmission mark current speaker | |
CN106575243A (en) | Hypervisor-hosted virtual machine forensics | |
CN109994108A (en) | Full-duplex communication technology for the session talk between chat robots and people | |
CN108133707A (en) | A kind of content share method and system | |
CN107967055A (en) | A kind of man-machine interaction method, terminal and computer-readable medium | |
CN106462713B (en) | The interface display method and terminal of terminal | |
CN107209905A (en) | For personalized and task completion service, correspondence spends theme and sorted out | |
CN107870711A (en) | Page navigation method, the method and client that user interface is provided | |
CN103902056B (en) | Dummy keyboard input method, equipment and system | |
CN108446320A (en) | A kind of data processing method, device and the device for data processing | |
CN108229535A (en) | Relate to yellow image audit method, apparatus, computer equipment and storage medium | |
CN109474658A (en) | Electronic equipment, server and the recording medium of task run are supported with external equipment | |
US11233897B1 (en) | Secure call center communications | |
CN108345667A (en) | A kind of searching method and relevant apparatus | |
CN110047484A (en) | A kind of speech recognition exchange method, system, equipment and storage medium | |
US20130120249A1 (en) | Electronic device | |
CN109144458A (en) | For executing the electronic equipment for inputting corresponding operation with voice | |
CN109448727A (en) | Voice interactive method and device | |
US20230131975A1 (en) | Music playing method and apparatus based on user interaction, and device and storage medium | |
CN109547836A (en) | Exchange method and device, electronic equipment, storage medium is broadcast live |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
CB02 | Change of applicant information | ||
CB02 | Change of applicant information |
Address after: 266555 Qingdao economic and Technological Development Zone, Shandong, Hong Kong Road, No. 218 Applicant after: Hisense Video Technology Co.,Ltd. Address before: 266555 Qingdao economic and Technological Development Zone, Shandong, Hong Kong Road, No. 218 Applicant before: QINGDAO HISENSE ELECTRONICS Co.,Ltd. |
|
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |