CN109120793A - Method of speech processing and device - Google Patents
Method of speech processing and device Download PDFInfo
- Publication number
- CN109120793A CN109120793A CN201811044257.1A CN201811044257A CN109120793A CN 109120793 A CN109120793 A CN 109120793A CN 201811044257 A CN201811044257 A CN 201811044257A CN 109120793 A CN109120793 A CN 109120793A
- Authority
- CN
- China
- Prior art keywords
- slide
- speech processes
- display interface
- current display
- voice
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
- H04M1/7243—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
- H04M1/72433—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2250/00—Details of telephonic subscriber devices
- H04M2250/22—Details of telephonic subscriber devices including a touch pad, a touch sensor or a touch detector
Landscapes
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- General Business, Economics & Management (AREA)
- Human Computer Interaction (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The disclosure is directed to method of speech processing and devices.This method comprises: display current display interface;The trigger action that monitoring touch object is inputted relative to current display interface;According to trigger action, corresponding speech processes operation is executed, wherein speech processes operation, which includes at least one of the following:, opens phonetic function, the voice that cancels this input, sends the voice of this input to Correspondent Node.Pass through the disclosure, make user can be by the directly i.e. executable speech processes operation of triggering current display interface, it just can be carried out speech processes operation for carrying out the designated button of speech processes operation without one is specially arranged in interface, so that user can quickly and easily carry out speech processes operation, improve the execution efficiency of speech processes operation, simultaneously, since the current display interface can be any interface of any application, thus, this further such that user can flexibly, quickly carry out speech processes operation.
Description
Technical field
This disclosure relates to voice technology field more particularly to method of speech processing and device.
Background technique
Currently, user can be realized at voice relevant with voice to other users voice communication etc. by instant messaging application
Reason operation.But in the related technology widespread practice be need user to click talk button, the designated buttons such as video button could be into
The operation of the speech processes such as row voice communication, and user can only be by clicking in instant messaging application (such as wechat, QQ) setting
Designated button just can be carried out speech processes operation, and can not be on the basis of not exiting other application directly on the boundary of other application
Speech processes operation is carried out in face.And this mode of operation is not available family not only and carries out at voice in non-instant communication applications
Reason operation and it is necessary to one specially be arranged in interface for carrying out the designated button of speech processes operation, in this way, not only not
Can flexibly, quickly carry out speech processes operation, and thus need additional to occupy interface due to needing to be arranged special button
In position and in operation not necessarily all user can successfully find very much the put position of the button so as to exist behaviour
Make inconvenient place.
Summary of the invention
The embodiment of the present disclosure provides method of speech processing and device.The technical solution is as follows:
According to the first aspect of the embodiments of the present disclosure, a kind of method of speech processing is provided, comprising:
Show current display interface;
The trigger action that monitoring touch object is inputted relative to the current display interface;
According to the trigger action, corresponding speech processes operation is executed, wherein the speech processes operation includes following
At least one of: it opens phonetic function, the voice for cancelling this input, send the voice of this input to Correspondent Node.
In one embodiment, described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action includes slide, the location dependent information of the slide is obtained, wherein described
Location dependent information includes at least one of the following: current location of the slide relative to screen, the slide phase
For the sliding distance of screen;
According to the location dependent information, speech processes operation corresponding with the location dependent information is executed.
In one embodiment, when the location dependent information includes sliding distance of the slide relative to screen
When, it is described according to the location dependent information, execute speech processes operation corresponding with the location dependent information, comprising:
When the slide is slide downward operation of the touch object relative to the current display interface, if
The distance of glide of the slide downward operation reaches the first pre-determined distance, then opens phonetic function, to allow user to input voice
Obtain this input voice;
When the slide is upward sliding operation of the touch object relative to the current display interface, if
The upslip of the upward sliding operation then cancels the voice of this input from the second pre-determined distance is reached.
In one embodiment, described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action is that the touch object leaves the operation of the current display interface, this input is sent
Voice to Correspondent Node.
In one embodiment, the method also includes:
User is prompted by default highlight, wherein the prompt includes: executable by the trigger action
Speech processes operation, speech processes operation will be executed, speech processes operation executes completion, described presets
Highlight will be hidden after preset duration.
In one embodiment, the default highlight is located at the top of the current display interface.
According to the second aspect of an embodiment of the present disclosure, a kind of voice processing apparatus is provided, comprising:
Display module, for showing current display interface;
Monitoring modular, the trigger action inputted for monitoring touch object relative to the current display interface;
Execution module, for executing corresponding speech processes operation, wherein at the voice according to the trigger action
Reason operation, which includes at least one of the following:, opens phonetic function, the voice that cancels this input, sends the voice of this input to logical
Believe opposite end.
In one embodiment, the execution module includes:
Acquisition submodule, for obtaining the position phase of the slide when the trigger action includes slide
Close information, wherein the location dependent information includes at least one of the following: present bit of the slide relative to screen
It sets, sliding distance of the slide relative to screen;
Implementation sub-module, for executing voice corresponding with the location dependent information according to the location dependent information
Processing operation.
In one embodiment, the implementation sub-module includes:
Opening unit, for when the location dependent information include the slide relative to the sliding distance of screen and
When the slide is slide downward operation of the touch object relative to the current display interface, if described to downslide
The distance of glide of dynamic operation reaches the first pre-determined distance, then opens phonetic function, defeated to allow user's input voice to obtain this
Enter voice;
Revocation unit, for being the touch object relative to the upward of the current display interface when the slide
When slide, if the upslip of upward sliding operation cancels the voice of this input from the second pre-determined distance is reached.
In one embodiment, the execution module includes:
Sending submodule, for being the operation that the touch object leaves the current display interface when the trigger action
When, the voice of this input is sent to Correspondent Node.
In one embodiment, described device further include:
Cue module, for being prompted by default highlight user, wherein the prompt includes: by described
The executable speech processes of trigger action operate, will execute the speech processes operation, speech processes operation is held
Row is completed, the default highlight will be hidden after preset duration.
The technical scheme provided by this disclosed embodiment can include the following benefits:
The trigger action inputted by monitoring touch object relative to current display interface, can be according to the trigger action, certainly
It is dynamic to execute speech processes operation corresponding with the trigger action, to keep user direct by triggering current display interface
Speech processes operation is executed, without one is specially arranged in interface for carrying out the designated button ability of speech processes operation
Speech processes operation is carried out, so that user can quickly and easily carry out speech processes operation, improves speech processes behaviour
The execution efficiency of work, simultaneously as the current display interface can be any interface of any application, thus, this further makes
User can flexibly, quickly carry out speech processes operation.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not
The disclosure can be limited.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows the implementation for meeting the disclosure
Example, and together with specification for explaining the principles of this disclosure.
Fig. 1 is a kind of flow chart of method of speech processing shown according to an exemplary embodiment.
Fig. 2 is the flow chart of another method of speech processing shown according to an exemplary embodiment.
Fig. 3 is the flow chart of another method of speech processing shown according to an exemplary embodiment.
Fig. 4 A is a kind of schematic diagram for carrying out touch control operation in the display interface shown according to an exemplary embodiment.
Fig. 4 B is a kind of progress touch-control behaviour on the display interface comprising highlight shown according to an exemplary embodiment
The schematic diagram of work.
Fig. 4 C is that another kind shown according to an exemplary embodiment carries out touch-control on the display interface comprising highlight
The schematic diagram of operation.
Fig. 4 D be it is shown according to an exemplary embodiment another touch-control is carried out on the display interface comprising highlight
The schematic diagram of operation.
Fig. 4 E be it is shown according to an exemplary embodiment another touch-control is carried out on the display interface comprising highlight
The schematic diagram of operation.
Fig. 5 is a kind of block diagram of voice processing apparatus shown according to an exemplary embodiment.
Fig. 6 is the block diagram of another voice processing apparatus shown according to an exemplary embodiment.
Fig. 7 is the block diagram of another voice processing apparatus shown according to an exemplary embodiment.
Fig. 8 is the block diagram of another voice processing apparatus shown according to an exemplary embodiment.
Specific embodiment
Example embodiments are described in detail here, and the example is illustrated in the accompanying drawings.Following description is related to
When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment
Described in embodiment do not represent all implementations consistent with this disclosure.On the contrary, they be only with it is such as appended
The example of the consistent device and method of some aspects be described in detail in claims, the disclosure.
In order to solve the above-mentioned technical problem, the embodiment of the present disclosure provides a kind of method of speech processing, and this method can be used for
In voice processing program, system or device, and the corresponding executing subject of this method can be the terminals such as mobile phone, computer, such as Fig. 1
It is shown, the method comprising the steps of S101 to step S103:
In step s101, current display interface is shown;
The current display interface can be any interface of any application, i.e. the current display interface can be instant messaging
Using in (such as wechat, QQ, microblogging) any display interface or non-instant communications applications (such as notepad appli-cation, video is answered
With, music application, shopping application etc.) in any display interface, it might even be possible to be desktop.
In step s 102, the trigger action that monitoring touch object is inputted relative to current display interface;
Trigger action can be the upward sliding operation relative to the current display interface, slide downward operation, slide to the left
Dynamic operation, slide, touch object leave the current display interface to the right (i.e. touch object unclamps the current display interface)
Operation, operation, the operation of the multi-hit current display interface of the long-pressing current display interface etc..
And touch object can be user's finger, intelligent touch pen etc..
The starting point of the trigger action can be any position in current display interface, as shown in Figure 4 A.
In step s 103, according to trigger action, corresponding speech processes operation is executed, wherein speech processes operation packet
It includes at least one of following: opening phonetic function, the voice for cancelling this input, sends the voice of this input to Correspondent Node.
The trigger action inputted by monitoring touch object relative to current display interface, can be according to the trigger action, certainly
It is dynamic to execute speech processes operation corresponding with the trigger action, to keep user direct by triggering current display interface
Speech processes operation is executed, without one is specially arranged in interface for carrying out the designated button ability of speech processes operation
Speech processes operation is carried out, so that user can quickly and easily carry out speech processes operation, improves speech processes behaviour
The execution efficiency of work, simultaneously as the current display interface can be any interface of any application, thus, this further makes
User can flexibly, quickly carry out speech processes operation.
As shown in Fig. 2, in one embodiment, above-mentioned step S103 shown in FIG. 1, i.e., according to trigger action, execution pair
The speech processes operation answered, may include step S201 and step S202, in which:
In step s 201, when trigger action includes slide, the location dependent information of slide is obtained,
In, location dependent information includes at least one of the following: slide, and relative to the current location of screen, (i.e. slide is absolute
Position, as touch point of the slide in current display interface relative to the top of screen, bottom, left or right side away from
From), slide relative to screen sliding distance (i.e. initial position of the slide in current display interface to terminate
The distance between position);
In step S202, according to location dependent information, speech processes operation corresponding with location dependent information is executed.
When touch control operation is slide, by obtaining the location dependent information, can execute automatically related to the position
The corresponding speech processes operation of information, thus it is not necessary that one is specially arranged in interface for carrying out the finger of speech processes operation
Determine button basis, without user trigger interface in designated button on the basis of, by relative to the current display interface into
Row slide can directly carry out speech processes operation, so that user can quickly and easily carry out speech processes behaviour
Make.
In one embodiment, above-mentioned when location dependent information includes sliding distance of the slide relative to screen
Step S202 in Fig. 2 executes speech processes operation corresponding with location dependent information that is, according to location dependent information, can quilt
It executes are as follows:
When slide is slide downward operation of the touch object relative to current display interface, if slide downward operates
Distance of glide reach the first pre-determined distance, then open phonetic function, with allow user input voice obtain this input voice;
When slide is upward sliding operation of the touch object relative to current display interface, if upward sliding operates
Upslip from the second pre-determined distance is reached, then cancel the voice of this input, wherein the first pre-determined distance and second it is default away from
From can be same or different.
When the slide is slide downward operation of the touch object relative to current display interface, if slide downward is grasped
The distance of glide of work reaches the first pre-determined distance, then can automatically opening voice function so that user can be by current relative to this
Display interface slide downward quickly and easily opens phonetic function, so that user inputs voice;If upward sliding operation is upper
Actual relative movement then cancels the voice of this input from the second pre-determined distance is reached automatically, so that user can be by current aobvious relative to this
Show that interface upward sliding can quickly and easily cancel the voice of this input, avoids accidentally sending voice.
Certainly, how those skilled in the art according to demand carries out speech processes operation if being also freely arranged, such as:
When slide is upward sliding operation of the touch object relative to current display interface, if upward sliding operates
Upslip from third pre-determined distance is reached, then open phonetic function, with allow user input voice obtain this input voice;
When slide is slide downward operation of the touch object relative to current display interface, if slide downward operates
Distance of glide reach the 4th pre-determined distance, then cancel this input voice.
Another example is: when slide is to the left slide of the touch object relative to current display interface, if to the left
The left actual relative movement of slide then opens phonetic function from the 5th pre-determined distance is reached, and obtains this to allow user to input voice
Input voice;
When slide is to the right slide of the touch object relative to current display interface, if slide to the right
Right actual relative movement from the 6th pre-determined distance is reached, then cancel this input voice.
Another example is: when slide is to the right slide of the touch object relative to current display interface, if to the right
The right actual relative movement of slide then opens phonetic function from the 7th pre-determined distance is reached, and obtains this to allow user to input voice
Input voice;
When slide is to the left slide of the touch object relative to current display interface, if slide to the left
Left actual relative movement from the 8th pre-determined distance is reached, then cancel this input voice.
As shown in figure 3, in one embodiment, the step S103 in above-mentioned Fig. 1 executes correspondence that is, according to trigger action
Speech processes operation, may include step S301:
In step S301, when trigger action is that touch object leaves the operation of current display interface, it is defeated to send this
The voice entered is to Correspondent Node.
When trigger action unclamps the current display interface for the operation such as finger that touch object leaves current display interface,
The voice of this input can be sent automatically to Correspondent Node, thus can be square when touch object leaves the current display interface
Just, the voice of this input is quickly sent to Correspondent Node.
The Correspondent Node can be the receiving end of default, or
The voice of this input can be subjected to automatic identification, phonetic incepting side be identified, by phonetic incepting side's conduct
The Correspondent Node.
In one embodiment, method further include:
User is prompted by default highlight, wherein prompt includes: at the voice that can be performed by trigger action
Reason operation, the speech processes operation that will be executed (that is: the next speech processes carried out in certain duration are operated, the duration
Can be 2 seconds, 3 seconds etc.), speech processes operation executes completion, preset highlight will be hidden after preset duration.
Automatic prompt is carried out to user by default highlight, family can be used to understand various states in time, with to user into
Row prompt in time, to promote user experience.
In one embodiment, the top that highlight is located at current display interface is preset.
Default highlight can be located at the top of current display interface, which can be suspended in current display
Region on interface, alternatively, the default highlight can be a part of region of current display interface.And in order to be carried out to user
Sufficiently prompt, touch object when carrying out slide relative to current display interface, the default highlight can it is static always not
It is dynamic.
Another example is: when the trigger action for opening phonetic function is to glide operation, and a distance that glided then can be
The top display reminding area at interface, and " gliding again, any will open phonetic function " is shown in the highlight at the top of interface, such as
Shown in Fig. 4 B.Or
Another example is: when the trigger action for closing phonetic function (being the voice for cancelling this input for accurate) is upper sliding behaviour
Work, the trigger action for opening phonetic function are when gliding operation, to close phonetic function if having gone up and having slided, can be at the top of interface
Highlight shows prompt information as shown in Figure 4 C.
Another example is: when sending the trigger action of voice to Correspondent Node of this input is to loose one's grip to operate, if user is
Screen (i.e. release current display interface) is unclamped, the voice of this input has been sent to Correspondent Node, then can push up at interface
The highlight in portion shows prompt information as shown in Figure 4 D, gradually to withdraw and (hide) highlight.
Another example is: (being to cancel for accurate when the trigger action for opening phonetic function operates for downslide, closes phonetic function
The voice of this input) trigger action be upper sliding operation, the voice to the Correspondent Node that send this input trigger action be
It looses one's grip when operating, prompt information as shown in Figure 4 E can be shown in the highlight at the top of interface.
Finally, it should be clear that: those skilled in the art can according to actual needs carry out above-mentioned multiple embodiments certainly
By combining.
The above-mentioned method of speech processing that the corresponding embodiment of the present disclosure provides, the embodiment of the present disclosure also provide a kind of speech processes
Device, as shown in figure 5, the device includes:
Display module 501 is configured as display current display interface;
Monitoring modular 502 is configured as the trigger action that monitoring touch object is inputted relative to current display interface;
Execution module 503 is configured as executing corresponding speech processes operation, wherein speech processes according to trigger action
Operation include at least one of the following: open phonetic function, cancel this input voice, send this input voice to communicate
Opposite end.
As shown in fig. 6, in one embodiment, execution module 503 may include:
Acquisition submodule 5031 is configured as obtaining the position phase of slide when trigger action includes slide
Close information, wherein location dependent information includes at least one of the following: that slide is grasped relative to the current location of screen, sliding
Make the sliding distance relative to screen;
Implementation sub-module 5032 is configured as executing voice corresponding with location dependent information according to location dependent information
Processing operation.
In one embodiment, implementation sub-module 5032 includes:
Opening unit is configured as when location dependent information including sliding distance and sliding of the slide relative to screen
When operation is slide downward operation of the touch object relative to current display interface, if the distance of glide of slide downward operation reaches
First pre-determined distance, then open phonetic function, obtains this input voice to allow user to input voice;
Revocation unit is configured as when slide being upward sliding operation of the touch object relative to current display interface
When, if the upslip of upward sliding operation cancels the voice of this input from the second pre-determined distance is reached.
As shown in fig. 7, in one embodiment, execution module 503 includes:
Sending submodule 5033 is configured as when trigger action is that touch object leaves the operation of current display interface,
The voice of this input is sent to Correspondent Node.
In one embodiment, device further include:
Cue module is configured as prompting user by default highlight, wherein prompt includes: to pass through triggering
The executable speech processes of operation operate, will execute speech processes operation, speech processes operation executes completion, default highlight
It will be hidden after preset duration.
Fig. 8 is a kind of block diagram for voice processing apparatus 800 shown according to an exemplary embodiment, which is applicable in
In terminal device.For example, device 800 can be mobile phone, computer, digital broadcasting terminal, messaging device, game control
Platform processed, tablet device, Medical Devices, body-building equipment, a number assistant etc..
Referring to Fig. 8, device 800 may include with next or at least two components: processing component 802, memory 804,
Power supply module 806, multimedia component 808, audio component 810, input/output (I/O) interface 812, sensor module 814, with
And communication component 816.
The integrated operation of the usual control device 800 of processing component 802, such as with display, telephone call, data communication, phase
Machine operation and record operate associated operation.Processing component 802 may include one or at least two processors 820 to execute
Instruction, to perform all or part of the steps of the methods described above.In addition, processing component 802 may include one or at least two
Module, convenient for the interaction between processing component 802 and other assemblies.For example, processing component 802 may include multi-media module,
To facilitate the interaction between multimedia component 808 and processing component 802.
Memory 804 is configured as storing various types of data to support the operation in device 800.These data are shown
Example includes instruction for any storage object or method that operate on device 800, contacts user data, telephone book data,
Message, picture, video etc..Memory 804 can by any kind of volatibility or non-volatile memory device or they
Combination is realized, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), it is erasable can
Program read-only memory (EPROM), programmable read only memory (PROM), read-only memory (ROM), magnetic memory, flash memory
Reservoir, disk or CD.
Power supply module 806 provides power supply for the various assemblies of device 800.Power supply module 806 may include power management system
System, one or at least two power supplys and other with for device 800 generate, manage, and distribute the associated component of power supply.
Multimedia component 808 includes the screen of one output interface of offer between described device 800 and user.One
In a little embodiments, screen may include liquid crystal display (LCD) and touch panel (TP).If screen includes touch panel, screen
Curtain may be implemented as touch screen, to receive input signal from the user.Touch panel includes that one or at least two touch
Sensor is to sense the gesture on touch, slide, and touch panel.The touch sensor can not only sense touch or sliding
The boundary of movement, but also detect duration and pressure associated with the touch or slide operation.In some embodiments,
Multimedia component 808 includes a front camera and/or rear camera.When device 800 is in operation mode, as shot mould
When formula or video mode, front camera and/or rear camera can receive external multi-medium data.Each preposition camera shooting
Head and rear camera can be a fixed optical lens system or have focusing and optical zoom capabilities.
Audio component 810 is configured as output and/or input audio signal.For example, audio component 810 includes a Mike
Wind (MIC), when device 800 is in operation mode, when such as call mode, recording mode, and voice recognition mode, microphone is matched
It is set to reception external audio signal.The received audio signal can be further stored in memory 804 or via communication set
Part 816 is sent.In some embodiments, audio component 810 further includes a loudspeaker, is used for output audio signal.
I/O interface 812 provides interface between processing component 802 and peripheral interface module, and above-mentioned peripheral interface module can
To be keyboard, click wheel, button etc..These buttons may include, but are not limited to: home button, volume button, start button and lock
Determine button.
Sensor module 814 includes one or at least two sensors, for providing the state of various aspects for device 800
Assessment.For example, sensor module 814 can detecte the state that opens/closes of device 800, the relative positioning of component, such as institute
The display and keypad that component is device 800 are stated, sensor module 814 can be with 800 1 groups of detection device 800 or device
The position change of part, the existence or non-existence that user contacts with device 800,800 orientation of device or acceleration/deceleration and device 800
Temperature change.Sensor module 814 may include proximity sensor, be configured to examine without any physical contact
Survey presence of nearby objects.Sensor module 814 can also include that optical sensor is used for such as CMOS or ccd image sensor
It is used in imaging applications.In some embodiments, which can also include acceleration transducer, and gyroscope passes
Sensor, Magnetic Sensor, pressure sensor or temperature sensor.
Communication component 816 is configured to facilitate the communication of wired or wireless way between device 800 and other equipment.Device
800 can access the wireless network based on communication standard, such as WiFi, 2G or 3G or their combination.In an exemplary implementation
In example, communication component 816 receives broadcast singal or broadcast related information from external broadcasting management system via broadcast channel.
In one exemplary embodiment, the communication component 816 further includes near-field communication (NFC) module, to promote short range communication.Example
Such as, NFC module can be based on radio frequency identification (RFID) technology, Infrared Data Association (IrDA) technology, ultra wide band (UWB) technology,
Bluetooth (BT) technology and other technologies are realized.
In the exemplary embodiment, device 800 can be by one or at least two application specific integrated circuits (ASIC), number
Word signal processor (DSP), digital signal processing appts (DSPD), programmable logic device (PLD), field programmable gate array
(FPGA), controller, microcontroller, microprocessor or other electronic building bricks are realized.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium including instruction, example are additionally provided
It such as include the memory 804 of instruction, above-metioned instruction can be executed by the processor 820 of device 800 to complete the above method.For example,
The non-transitorycomputer readable storage medium can be ROM, random access memory (RAM), CD-ROM, tape, floppy disk
With optical data storage devices etc..
Those skilled in the art will readily occur to its of the disclosure after considering specification and practicing disclosure disclosed herein
Its embodiment.This application is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or
Person's adaptive change follows the general principles of this disclosure and including the undocumented common knowledge in the art of the disclosure
Or conventional techniques.The description and examples are only to be considered as illustrative, and the true scope and spirit of the disclosure are by following
Claim is pointed out.
It should be understood that the present disclosure is not limited to the precise structures that have been described above and shown in the drawings, and
And various modifications and changes may be made without departing from the scope thereof.The scope of the present disclosure is only limited by the accompanying claims.
Claims (11)
1. a kind of method of speech processing characterized by comprising
Show current display interface;
The trigger action that monitoring touch object is inputted relative to the current display interface;
According to the trigger action, execute corresponding speech processes operation, wherein the speech processes operation include it is following at least
One: opening phonetic function, the voice for cancelling this input, send the voice of this input to Correspondent Node.
2. the method according to claim 1, wherein
It is described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action includes slide, the location dependent information of the slide is obtained, wherein the position
Relevant information include at least one of the following: the slide relative to the current location of screen, the slide relative to
The sliding distance of screen;
According to the location dependent information, speech processes operation corresponding with the location dependent information is executed.
3. according to the method described in claim 2, it is characterized in that,
It is described according to the position when the location dependent information includes sliding distance of the slide relative to screen
Relevant information executes speech processes operation corresponding with the location dependent information, comprising:
When the slide is slide downward operation of the touch object relative to the current display interface, if described
The distance of glide of slide downward operation reaches the first pre-determined distance, then opens phonetic function, is obtained with allowing user to input voice
This input voice;
When the slide is upward sliding operation of the touch object relative to the current display interface, if described
The upslip of upward sliding operation then cancels the voice of this input from the second pre-determined distance is reached.
4. the method according to claim 1, wherein
It is described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action is that the touch object leaves the operation of the current display interface, the language of this input is sent
Sound is to Correspondent Node.
5. method according to claim 1 to 4, which is characterized in that the method also includes:
User is prompted by default highlight, wherein the prompt includes: the institute that can be performed by the trigger action
Predicate sound processing operation, speech processes operation will be executed, speech processes operation executes completion, the default prompt
Area will be hidden after preset duration.
6. according to the method described in claim 5, it is characterized in that,
The default highlight is located at the top of the current display interface.
7. a kind of voice processing apparatus characterized by comprising
Display module, for showing current display interface;
Monitoring modular, the trigger action inputted for monitoring touch object relative to the current display interface;
Execution module, for executing corresponding speech processes operation according to the trigger action, wherein the speech processes behaviour
Include at least one of the following: open phonetic function, cancel this input voice, send this input voice to communication pair
End.
8. device according to claim 7, which is characterized in that
The execution module includes:
Acquisition submodule, the position correlation for when the trigger action includes slide, obtaining the slide are believed
Breath, wherein the location dependent information includes at least one of the following: current location of the slide relative to screen, institute
State sliding distance of the slide relative to screen;
Implementation sub-module, for executing speech processes corresponding with the location dependent information according to the location dependent information
Operation.
9. device according to claim 8, which is characterized in that
The implementation sub-module includes:
Opening unit, for including the slide relative to the sliding distance of screen and described when the location dependent information
When slide is slide downward operation of the touch object relative to the current display interface, if the slide downward is grasped
The distance of glide of work reaches the first pre-determined distance, then opens phonetic function, obtains this input language to allow user to input voice
Sound;
Revocation unit, for being upward sliding of the touch object relative to the current display interface when the slide
When operation, if the upslip of upward sliding operation cancels the voice of this input from the second pre-determined distance is reached.
10. device according to claim 7, which is characterized in that
The execution module includes:
Sending submodule, for when the trigger action be the touch object leave the operation of the current display interface when,
The voice of this input is sent to Correspondent Node.
11. device according to any one of claims 7 to 10, which is characterized in that described device further include:
Cue module, for being prompted by default highlight user, wherein the prompt includes: by the triggering
The executable speech processes of operation operate, will execute the speech processes operation, speech processes operation has executed
It will be hidden after preset duration at, the default highlight.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811044257.1A CN109120793A (en) | 2018-09-07 | 2018-09-07 | Method of speech processing and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811044257.1A CN109120793A (en) | 2018-09-07 | 2018-09-07 | Method of speech processing and device |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109120793A true CN109120793A (en) | 2019-01-01 |
Family
ID=64858922
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811044257.1A Pending CN109120793A (en) | 2018-09-07 | 2018-09-07 | Method of speech processing and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109120793A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110069200A (en) * | 2019-04-24 | 2019-07-30 | 努比亚技术有限公司 | Wearable device input control method, wearable device and storage medium |
CN110837334A (en) * | 2019-11-04 | 2020-02-25 | 北京字节跳动网络技术有限公司 | Method, device, terminal and storage medium for interactive control |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20140099123A (en) * | 2013-02-01 | 2014-08-11 | 엘지전자 주식회사 | Mobile terminal and control method thereof |
CN104144239A (en) * | 2013-09-25 | 2014-11-12 | 腾讯科技(深圳)有限公司 | Voice assist communication method and device |
CN104375702A (en) * | 2014-10-31 | 2015-02-25 | 北京搜狗科技发展有限公司 | Touch operation method and device |
CN105446489A (en) * | 2015-12-08 | 2016-03-30 | 广州神马移动信息科技有限公司 | Voice dual-mode control method and apparatus, and user terminal |
CN106959746A (en) * | 2016-01-12 | 2017-07-18 | 百度在线网络技术(北京)有限公司 | The processing method and processing device of speech data |
-
2018
- 2018-09-07 CN CN201811044257.1A patent/CN109120793A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20140099123A (en) * | 2013-02-01 | 2014-08-11 | 엘지전자 주식회사 | Mobile terminal and control method thereof |
CN104144239A (en) * | 2013-09-25 | 2014-11-12 | 腾讯科技(深圳)有限公司 | Voice assist communication method and device |
CN104375702A (en) * | 2014-10-31 | 2015-02-25 | 北京搜狗科技发展有限公司 | Touch operation method and device |
CN105446489A (en) * | 2015-12-08 | 2016-03-30 | 广州神马移动信息科技有限公司 | Voice dual-mode control method and apparatus, and user terminal |
CN106959746A (en) * | 2016-01-12 | 2017-07-18 | 百度在线网络技术(北京)有限公司 | The processing method and processing device of speech data |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110069200A (en) * | 2019-04-24 | 2019-07-30 | 努比亚技术有限公司 | Wearable device input control method, wearable device and storage medium |
CN110069200B (en) * | 2019-04-24 | 2024-02-09 | 努比亚技术有限公司 | Wearable device input control method, wearable device and storage medium |
CN110837334A (en) * | 2019-11-04 | 2020-02-25 | 北京字节跳动网络技术有限公司 | Method, device, terminal and storage medium for interactive control |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3163569B1 (en) | Method and device for controlling a smart device by voice, computer program and recording medium | |
CN105607805B (en) | The footmark processing method and processing device of application icon | |
CN106547663B (en) | Key management method and device | |
CN108509232A (en) | Screen recording method, device and computer readable storage medium | |
CN105975156B (en) | Application interface display methods and device | |
CN105704766B (en) | The control method and device of double-card mobile terminal | |
CN105160239A (en) | Application program access restriction method and apparatus | |
CN105224171B (en) | Icon display method, device and terminal | |
CN105335062B (en) | Information editing's operation processing method and device | |
CN104394137B (en) | A kind of method and device of prompting voice call | |
CN107608561A (en) | Touch-screen control method and device | |
CN106791921A (en) | The processing method and processing device of net cast | |
CN106791092A (en) | The searching method and device of contact person | |
CN107105517A (en) | Method for connecting network and device | |
CN109951379A (en) | Message treatment method and device | |
CN105468281B (en) | The method and apparatus for showing set interface | |
CN106990903A (en) | Display and the method and device of hide application program | |
CN106775377A (en) | The control method of gesture identifying device, equipment and gesture identifying device | |
CN105187671A (en) | Recording method and device | |
CN107566615B (en) | Message treatment method, device and computer readable storage medium | |
CN109120793A (en) | Method of speech processing and device | |
CN109325337A (en) | Unlocking method and device | |
CN106096442B (en) | Applications trigger method and device | |
CN106101372B (en) | Using prompting setting method and device | |
CN108766427A (en) | Sound control method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190101 |