CN109120793A - Method of speech processing and device - Google Patents

Method of speech processing and device Download PDF

Info

Publication number
CN109120793A
CN109120793A CN201811044257.1A CN201811044257A CN109120793A CN 109120793 A CN109120793 A CN 109120793A CN 201811044257 A CN201811044257 A CN 201811044257A CN 109120793 A CN109120793 A CN 109120793A
Authority
CN
China
Prior art keywords
slide
speech processes
display interface
current display
voice
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811044257.1A
Other languages
Chinese (zh)
Inventor
古栋辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wireless Living (hangzhou) Mdt Infotech Ltd
Original Assignee
Wireless Living (hangzhou) Mdt Infotech Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wireless Living (hangzhou) Mdt Infotech Ltd filed Critical Wireless Living (hangzhou) Mdt Infotech Ltd
Priority to CN201811044257.1A priority Critical patent/CN109120793A/en
Publication of CN109120793A publication Critical patent/CN109120793A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/22Details of telephonic subscriber devices including a touch pad, a touch sensor or a touch detector

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Human Computer Interaction (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The disclosure is directed to method of speech processing and devices.This method comprises: display current display interface;The trigger action that monitoring touch object is inputted relative to current display interface;According to trigger action, corresponding speech processes operation is executed, wherein speech processes operation, which includes at least one of the following:, opens phonetic function, the voice that cancels this input, sends the voice of this input to Correspondent Node.Pass through the disclosure, make user can be by the directly i.e. executable speech processes operation of triggering current display interface, it just can be carried out speech processes operation for carrying out the designated button of speech processes operation without one is specially arranged in interface, so that user can quickly and easily carry out speech processes operation, improve the execution efficiency of speech processes operation, simultaneously, since the current display interface can be any interface of any application, thus, this further such that user can flexibly, quickly carry out speech processes operation.

Description

Method of speech processing and device
Technical field
This disclosure relates to voice technology field more particularly to method of speech processing and device.
Background technique
Currently, user can be realized at voice relevant with voice to other users voice communication etc. by instant messaging application Reason operation.But in the related technology widespread practice be need user to click talk button, the designated buttons such as video button could be into The operation of the speech processes such as row voice communication, and user can only be by clicking in instant messaging application (such as wechat, QQ) setting Designated button just can be carried out speech processes operation, and can not be on the basis of not exiting other application directly on the boundary of other application Speech processes operation is carried out in face.And this mode of operation is not available family not only and carries out at voice in non-instant communication applications Reason operation and it is necessary to one specially be arranged in interface for carrying out the designated button of speech processes operation, in this way, not only not Can flexibly, quickly carry out speech processes operation, and thus need additional to occupy interface due to needing to be arranged special button In position and in operation not necessarily all user can successfully find very much the put position of the button so as to exist behaviour Make inconvenient place.
Summary of the invention
The embodiment of the present disclosure provides method of speech processing and device.The technical solution is as follows:
According to the first aspect of the embodiments of the present disclosure, a kind of method of speech processing is provided, comprising:
Show current display interface;
The trigger action that monitoring touch object is inputted relative to the current display interface;
According to the trigger action, corresponding speech processes operation is executed, wherein the speech processes operation includes following At least one of: it opens phonetic function, the voice for cancelling this input, send the voice of this input to Correspondent Node.
In one embodiment, described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action includes slide, the location dependent information of the slide is obtained, wherein described Location dependent information includes at least one of the following: current location of the slide relative to screen, the slide phase For the sliding distance of screen;
According to the location dependent information, speech processes operation corresponding with the location dependent information is executed.
In one embodiment, when the location dependent information includes sliding distance of the slide relative to screen When, it is described according to the location dependent information, execute speech processes operation corresponding with the location dependent information, comprising:
When the slide is slide downward operation of the touch object relative to the current display interface, if The distance of glide of the slide downward operation reaches the first pre-determined distance, then opens phonetic function, to allow user to input voice Obtain this input voice;
When the slide is upward sliding operation of the touch object relative to the current display interface, if The upslip of the upward sliding operation then cancels the voice of this input from the second pre-determined distance is reached.
In one embodiment, described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action is that the touch object leaves the operation of the current display interface, this input is sent Voice to Correspondent Node.
In one embodiment, the method also includes:
User is prompted by default highlight, wherein the prompt includes: executable by the trigger action Speech processes operation, speech processes operation will be executed, speech processes operation executes completion, described presets Highlight will be hidden after preset duration.
In one embodiment, the default highlight is located at the top of the current display interface.
According to the second aspect of an embodiment of the present disclosure, a kind of voice processing apparatus is provided, comprising:
Display module, for showing current display interface;
Monitoring modular, the trigger action inputted for monitoring touch object relative to the current display interface;
Execution module, for executing corresponding speech processes operation, wherein at the voice according to the trigger action Reason operation, which includes at least one of the following:, opens phonetic function, the voice that cancels this input, sends the voice of this input to logical Believe opposite end.
In one embodiment, the execution module includes:
Acquisition submodule, for obtaining the position phase of the slide when the trigger action includes slide Close information, wherein the location dependent information includes at least one of the following: present bit of the slide relative to screen It sets, sliding distance of the slide relative to screen;
Implementation sub-module, for executing voice corresponding with the location dependent information according to the location dependent information Processing operation.
In one embodiment, the implementation sub-module includes:
Opening unit, for when the location dependent information include the slide relative to the sliding distance of screen and When the slide is slide downward operation of the touch object relative to the current display interface, if described to downslide The distance of glide of dynamic operation reaches the first pre-determined distance, then opens phonetic function, defeated to allow user's input voice to obtain this Enter voice;
Revocation unit, for being the touch object relative to the upward of the current display interface when the slide When slide, if the upslip of upward sliding operation cancels the voice of this input from the second pre-determined distance is reached.
In one embodiment, the execution module includes:
Sending submodule, for being the operation that the touch object leaves the current display interface when the trigger action When, the voice of this input is sent to Correspondent Node.
In one embodiment, described device further include:
Cue module, for being prompted by default highlight user, wherein the prompt includes: by described The executable speech processes of trigger action operate, will execute the speech processes operation, speech processes operation is held Row is completed, the default highlight will be hidden after preset duration.
The technical scheme provided by this disclosed embodiment can include the following benefits:
The trigger action inputted by monitoring touch object relative to current display interface, can be according to the trigger action, certainly It is dynamic to execute speech processes operation corresponding with the trigger action, to keep user direct by triggering current display interface Speech processes operation is executed, without one is specially arranged in interface for carrying out the designated button ability of speech processes operation Speech processes operation is carried out, so that user can quickly and easily carry out speech processes operation, improves speech processes behaviour The execution efficiency of work, simultaneously as the current display interface can be any interface of any application, thus, this further makes User can flexibly, quickly carry out speech processes operation.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not The disclosure can be limited.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows the implementation for meeting the disclosure Example, and together with specification for explaining the principles of this disclosure.
Fig. 1 is a kind of flow chart of method of speech processing shown according to an exemplary embodiment.
Fig. 2 is the flow chart of another method of speech processing shown according to an exemplary embodiment.
Fig. 3 is the flow chart of another method of speech processing shown according to an exemplary embodiment.
Fig. 4 A is a kind of schematic diagram for carrying out touch control operation in the display interface shown according to an exemplary embodiment.
Fig. 4 B is a kind of progress touch-control behaviour on the display interface comprising highlight shown according to an exemplary embodiment The schematic diagram of work.
Fig. 4 C is that another kind shown according to an exemplary embodiment carries out touch-control on the display interface comprising highlight The schematic diagram of operation.
Fig. 4 D be it is shown according to an exemplary embodiment another touch-control is carried out on the display interface comprising highlight The schematic diagram of operation.
Fig. 4 E be it is shown according to an exemplary embodiment another touch-control is carried out on the display interface comprising highlight The schematic diagram of operation.
Fig. 5 is a kind of block diagram of voice processing apparatus shown according to an exemplary embodiment.
Fig. 6 is the block diagram of another voice processing apparatus shown according to an exemplary embodiment.
Fig. 7 is the block diagram of another voice processing apparatus shown according to an exemplary embodiment.
Fig. 8 is the block diagram of another voice processing apparatus shown according to an exemplary embodiment.
Specific embodiment
Example embodiments are described in detail here, and the example is illustrated in the accompanying drawings.Following description is related to When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment Described in embodiment do not represent all implementations consistent with this disclosure.On the contrary, they be only with it is such as appended The example of the consistent device and method of some aspects be described in detail in claims, the disclosure.
In order to solve the above-mentioned technical problem, the embodiment of the present disclosure provides a kind of method of speech processing, and this method can be used for In voice processing program, system or device, and the corresponding executing subject of this method can be the terminals such as mobile phone, computer, such as Fig. 1 It is shown, the method comprising the steps of S101 to step S103:
In step s101, current display interface is shown;
The current display interface can be any interface of any application, i.e. the current display interface can be instant messaging Using in (such as wechat, QQ, microblogging) any display interface or non-instant communications applications (such as notepad appli-cation, video is answered With, music application, shopping application etc.) in any display interface, it might even be possible to be desktop.
In step s 102, the trigger action that monitoring touch object is inputted relative to current display interface;
Trigger action can be the upward sliding operation relative to the current display interface, slide downward operation, slide to the left Dynamic operation, slide, touch object leave the current display interface to the right (i.e. touch object unclamps the current display interface) Operation, operation, the operation of the multi-hit current display interface of the long-pressing current display interface etc..
And touch object can be user's finger, intelligent touch pen etc..
The starting point of the trigger action can be any position in current display interface, as shown in Figure 4 A.
In step s 103, according to trigger action, corresponding speech processes operation is executed, wherein speech processes operation packet It includes at least one of following: opening phonetic function, the voice for cancelling this input, sends the voice of this input to Correspondent Node.
The trigger action inputted by monitoring touch object relative to current display interface, can be according to the trigger action, certainly It is dynamic to execute speech processes operation corresponding with the trigger action, to keep user direct by triggering current display interface Speech processes operation is executed, without one is specially arranged in interface for carrying out the designated button ability of speech processes operation Speech processes operation is carried out, so that user can quickly and easily carry out speech processes operation, improves speech processes behaviour The execution efficiency of work, simultaneously as the current display interface can be any interface of any application, thus, this further makes User can flexibly, quickly carry out speech processes operation.
As shown in Fig. 2, in one embodiment, above-mentioned step S103 shown in FIG. 1, i.e., according to trigger action, execution pair The speech processes operation answered, may include step S201 and step S202, in which:
In step s 201, when trigger action includes slide, the location dependent information of slide is obtained, In, location dependent information includes at least one of the following: slide, and relative to the current location of screen, (i.e. slide is absolute Position, as touch point of the slide in current display interface relative to the top of screen, bottom, left or right side away from From), slide relative to screen sliding distance (i.e. initial position of the slide in current display interface to terminate The distance between position);
In step S202, according to location dependent information, speech processes operation corresponding with location dependent information is executed.
When touch control operation is slide, by obtaining the location dependent information, can execute automatically related to the position The corresponding speech processes operation of information, thus it is not necessary that one is specially arranged in interface for carrying out the finger of speech processes operation Determine button basis, without user trigger interface in designated button on the basis of, by relative to the current display interface into Row slide can directly carry out speech processes operation, so that user can quickly and easily carry out speech processes behaviour Make.
In one embodiment, above-mentioned when location dependent information includes sliding distance of the slide relative to screen Step S202 in Fig. 2 executes speech processes operation corresponding with location dependent information that is, according to location dependent information, can quilt It executes are as follows:
When slide is slide downward operation of the touch object relative to current display interface, if slide downward operates Distance of glide reach the first pre-determined distance, then open phonetic function, with allow user input voice obtain this input voice;
When slide is upward sliding operation of the touch object relative to current display interface, if upward sliding operates Upslip from the second pre-determined distance is reached, then cancel the voice of this input, wherein the first pre-determined distance and second it is default away from From can be same or different.
When the slide is slide downward operation of the touch object relative to current display interface, if slide downward is grasped The distance of glide of work reaches the first pre-determined distance, then can automatically opening voice function so that user can be by current relative to this Display interface slide downward quickly and easily opens phonetic function, so that user inputs voice;If upward sliding operation is upper Actual relative movement then cancels the voice of this input from the second pre-determined distance is reached automatically, so that user can be by current aobvious relative to this Show that interface upward sliding can quickly and easily cancel the voice of this input, avoids accidentally sending voice.
Certainly, how those skilled in the art according to demand carries out speech processes operation if being also freely arranged, such as:
When slide is upward sliding operation of the touch object relative to current display interface, if upward sliding operates Upslip from third pre-determined distance is reached, then open phonetic function, with allow user input voice obtain this input voice;
When slide is slide downward operation of the touch object relative to current display interface, if slide downward operates Distance of glide reach the 4th pre-determined distance, then cancel this input voice.
Another example is: when slide is to the left slide of the touch object relative to current display interface, if to the left The left actual relative movement of slide then opens phonetic function from the 5th pre-determined distance is reached, and obtains this to allow user to input voice Input voice;
When slide is to the right slide of the touch object relative to current display interface, if slide to the right Right actual relative movement from the 6th pre-determined distance is reached, then cancel this input voice.
Another example is: when slide is to the right slide of the touch object relative to current display interface, if to the right The right actual relative movement of slide then opens phonetic function from the 7th pre-determined distance is reached, and obtains this to allow user to input voice Input voice;
When slide is to the left slide of the touch object relative to current display interface, if slide to the left Left actual relative movement from the 8th pre-determined distance is reached, then cancel this input voice.
As shown in figure 3, in one embodiment, the step S103 in above-mentioned Fig. 1 executes correspondence that is, according to trigger action Speech processes operation, may include step S301:
In step S301, when trigger action is that touch object leaves the operation of current display interface, it is defeated to send this The voice entered is to Correspondent Node.
When trigger action unclamps the current display interface for the operation such as finger that touch object leaves current display interface, The voice of this input can be sent automatically to Correspondent Node, thus can be square when touch object leaves the current display interface Just, the voice of this input is quickly sent to Correspondent Node.
The Correspondent Node can be the receiving end of default, or
The voice of this input can be subjected to automatic identification, phonetic incepting side be identified, by phonetic incepting side's conduct The Correspondent Node.
In one embodiment, method further include:
User is prompted by default highlight, wherein prompt includes: at the voice that can be performed by trigger action Reason operation, the speech processes operation that will be executed (that is: the next speech processes carried out in certain duration are operated, the duration Can be 2 seconds, 3 seconds etc.), speech processes operation executes completion, preset highlight will be hidden after preset duration.
Automatic prompt is carried out to user by default highlight, family can be used to understand various states in time, with to user into Row prompt in time, to promote user experience.
In one embodiment, the top that highlight is located at current display interface is preset.
Default highlight can be located at the top of current display interface, which can be suspended in current display Region on interface, alternatively, the default highlight can be a part of region of current display interface.And in order to be carried out to user Sufficiently prompt, touch object when carrying out slide relative to current display interface, the default highlight can it is static always not It is dynamic.
Another example is: when the trigger action for opening phonetic function is to glide operation, and a distance that glided then can be The top display reminding area at interface, and " gliding again, any will open phonetic function " is shown in the highlight at the top of interface, such as Shown in Fig. 4 B.Or
Another example is: when the trigger action for closing phonetic function (being the voice for cancelling this input for accurate) is upper sliding behaviour Work, the trigger action for opening phonetic function are when gliding operation, to close phonetic function if having gone up and having slided, can be at the top of interface Highlight shows prompt information as shown in Figure 4 C.
Another example is: when sending the trigger action of voice to Correspondent Node of this input is to loose one's grip to operate, if user is Screen (i.e. release current display interface) is unclamped, the voice of this input has been sent to Correspondent Node, then can push up at interface The highlight in portion shows prompt information as shown in Figure 4 D, gradually to withdraw and (hide) highlight.
Another example is: (being to cancel for accurate when the trigger action for opening phonetic function operates for downslide, closes phonetic function The voice of this input) trigger action be upper sliding operation, the voice to the Correspondent Node that send this input trigger action be It looses one's grip when operating, prompt information as shown in Figure 4 E can be shown in the highlight at the top of interface.
Finally, it should be clear that: those skilled in the art can according to actual needs carry out above-mentioned multiple embodiments certainly By combining.
The above-mentioned method of speech processing that the corresponding embodiment of the present disclosure provides, the embodiment of the present disclosure also provide a kind of speech processes Device, as shown in figure 5, the device includes:
Display module 501 is configured as display current display interface;
Monitoring modular 502 is configured as the trigger action that monitoring touch object is inputted relative to current display interface;
Execution module 503 is configured as executing corresponding speech processes operation, wherein speech processes according to trigger action Operation include at least one of the following: open phonetic function, cancel this input voice, send this input voice to communicate Opposite end.
As shown in fig. 6, in one embodiment, execution module 503 may include:
Acquisition submodule 5031 is configured as obtaining the position phase of slide when trigger action includes slide Close information, wherein location dependent information includes at least one of the following: that slide is grasped relative to the current location of screen, sliding Make the sliding distance relative to screen;
Implementation sub-module 5032 is configured as executing voice corresponding with location dependent information according to location dependent information Processing operation.
In one embodiment, implementation sub-module 5032 includes:
Opening unit is configured as when location dependent information including sliding distance and sliding of the slide relative to screen When operation is slide downward operation of the touch object relative to current display interface, if the distance of glide of slide downward operation reaches First pre-determined distance, then open phonetic function, obtains this input voice to allow user to input voice;
Revocation unit is configured as when slide being upward sliding operation of the touch object relative to current display interface When, if the upslip of upward sliding operation cancels the voice of this input from the second pre-determined distance is reached.
As shown in fig. 7, in one embodiment, execution module 503 includes:
Sending submodule 5033 is configured as when trigger action is that touch object leaves the operation of current display interface, The voice of this input is sent to Correspondent Node.
In one embodiment, device further include:
Cue module is configured as prompting user by default highlight, wherein prompt includes: to pass through triggering The executable speech processes of operation operate, will execute speech processes operation, speech processes operation executes completion, default highlight It will be hidden after preset duration.
Fig. 8 is a kind of block diagram for voice processing apparatus 800 shown according to an exemplary embodiment, which is applicable in In terminal device.For example, device 800 can be mobile phone, computer, digital broadcasting terminal, messaging device, game control Platform processed, tablet device, Medical Devices, body-building equipment, a number assistant etc..
Referring to Fig. 8, device 800 may include with next or at least two components: processing component 802, memory 804, Power supply module 806, multimedia component 808, audio component 810, input/output (I/O) interface 812, sensor module 814, with And communication component 816.
The integrated operation of the usual control device 800 of processing component 802, such as with display, telephone call, data communication, phase Machine operation and record operate associated operation.Processing component 802 may include one or at least two processors 820 to execute Instruction, to perform all or part of the steps of the methods described above.In addition, processing component 802 may include one or at least two Module, convenient for the interaction between processing component 802 and other assemblies.For example, processing component 802 may include multi-media module, To facilitate the interaction between multimedia component 808 and processing component 802.
Memory 804 is configured as storing various types of data to support the operation in device 800.These data are shown Example includes instruction for any storage object or method that operate on device 800, contacts user data, telephone book data, Message, picture, video etc..Memory 804 can by any kind of volatibility or non-volatile memory device or they Combination is realized, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), it is erasable can Program read-only memory (EPROM), programmable read only memory (PROM), read-only memory (ROM), magnetic memory, flash memory Reservoir, disk or CD.
Power supply module 806 provides power supply for the various assemblies of device 800.Power supply module 806 may include power management system System, one or at least two power supplys and other with for device 800 generate, manage, and distribute the associated component of power supply.
Multimedia component 808 includes the screen of one output interface of offer between described device 800 and user.One In a little embodiments, screen may include liquid crystal display (LCD) and touch panel (TP).If screen includes touch panel, screen Curtain may be implemented as touch screen, to receive input signal from the user.Touch panel includes that one or at least two touch Sensor is to sense the gesture on touch, slide, and touch panel.The touch sensor can not only sense touch or sliding The boundary of movement, but also detect duration and pressure associated with the touch or slide operation.In some embodiments, Multimedia component 808 includes a front camera and/or rear camera.When device 800 is in operation mode, as shot mould When formula or video mode, front camera and/or rear camera can receive external multi-medium data.Each preposition camera shooting Head and rear camera can be a fixed optical lens system or have focusing and optical zoom capabilities.
Audio component 810 is configured as output and/or input audio signal.For example, audio component 810 includes a Mike Wind (MIC), when device 800 is in operation mode, when such as call mode, recording mode, and voice recognition mode, microphone is matched It is set to reception external audio signal.The received audio signal can be further stored in memory 804 or via communication set Part 816 is sent.In some embodiments, audio component 810 further includes a loudspeaker, is used for output audio signal.
I/O interface 812 provides interface between processing component 802 and peripheral interface module, and above-mentioned peripheral interface module can To be keyboard, click wheel, button etc..These buttons may include, but are not limited to: home button, volume button, start button and lock Determine button.
Sensor module 814 includes one or at least two sensors, for providing the state of various aspects for device 800 Assessment.For example, sensor module 814 can detecte the state that opens/closes of device 800, the relative positioning of component, such as institute The display and keypad that component is device 800 are stated, sensor module 814 can be with 800 1 groups of detection device 800 or device The position change of part, the existence or non-existence that user contacts with device 800,800 orientation of device or acceleration/deceleration and device 800 Temperature change.Sensor module 814 may include proximity sensor, be configured to examine without any physical contact Survey presence of nearby objects.Sensor module 814 can also include that optical sensor is used for such as CMOS or ccd image sensor It is used in imaging applications.In some embodiments, which can also include acceleration transducer, and gyroscope passes Sensor, Magnetic Sensor, pressure sensor or temperature sensor.
Communication component 816 is configured to facilitate the communication of wired or wireless way between device 800 and other equipment.Device 800 can access the wireless network based on communication standard, such as WiFi, 2G or 3G or their combination.In an exemplary implementation In example, communication component 816 receives broadcast singal or broadcast related information from external broadcasting management system via broadcast channel. In one exemplary embodiment, the communication component 816 further includes near-field communication (NFC) module, to promote short range communication.Example Such as, NFC module can be based on radio frequency identification (RFID) technology, Infrared Data Association (IrDA) technology, ultra wide band (UWB) technology, Bluetooth (BT) technology and other technologies are realized.
In the exemplary embodiment, device 800 can be by one or at least two application specific integrated circuits (ASIC), number Word signal processor (DSP), digital signal processing appts (DSPD), programmable logic device (PLD), field programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic building bricks are realized.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium including instruction, example are additionally provided It such as include the memory 804 of instruction, above-metioned instruction can be executed by the processor 820 of device 800 to complete the above method.For example, The non-transitorycomputer readable storage medium can be ROM, random access memory (RAM), CD-ROM, tape, floppy disk With optical data storage devices etc..
Those skilled in the art will readily occur to its of the disclosure after considering specification and practicing disclosure disclosed herein Its embodiment.This application is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or Person's adaptive change follows the general principles of this disclosure and including the undocumented common knowledge in the art of the disclosure Or conventional techniques.The description and examples are only to be considered as illustrative, and the true scope and spirit of the disclosure are by following Claim is pointed out.
It should be understood that the present disclosure is not limited to the precise structures that have been described above and shown in the drawings, and And various modifications and changes may be made without departing from the scope thereof.The scope of the present disclosure is only limited by the accompanying claims.

Claims (11)

1. a kind of method of speech processing characterized by comprising
Show current display interface;
The trigger action that monitoring touch object is inputted relative to the current display interface;
According to the trigger action, execute corresponding speech processes operation, wherein the speech processes operation include it is following at least One: opening phonetic function, the voice for cancelling this input, send the voice of this input to Correspondent Node.
2. the method according to claim 1, wherein
It is described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action includes slide, the location dependent information of the slide is obtained, wherein the position Relevant information include at least one of the following: the slide relative to the current location of screen, the slide relative to The sliding distance of screen;
According to the location dependent information, speech processes operation corresponding with the location dependent information is executed.
3. according to the method described in claim 2, it is characterized in that,
It is described according to the position when the location dependent information includes sliding distance of the slide relative to screen Relevant information executes speech processes operation corresponding with the location dependent information, comprising:
When the slide is slide downward operation of the touch object relative to the current display interface, if described The distance of glide of slide downward operation reaches the first pre-determined distance, then opens phonetic function, is obtained with allowing user to input voice This input voice;
When the slide is upward sliding operation of the touch object relative to the current display interface, if described The upslip of upward sliding operation then cancels the voice of this input from the second pre-determined distance is reached.
4. the method according to claim 1, wherein
It is described according to the trigger action, execute corresponding speech processes operation, comprising:
When the trigger action is that the touch object leaves the operation of the current display interface, the language of this input is sent Sound is to Correspondent Node.
5. method according to claim 1 to 4, which is characterized in that the method also includes:
User is prompted by default highlight, wherein the prompt includes: the institute that can be performed by the trigger action Predicate sound processing operation, speech processes operation will be executed, speech processes operation executes completion, the default prompt Area will be hidden after preset duration.
6. according to the method described in claim 5, it is characterized in that,
The default highlight is located at the top of the current display interface.
7. a kind of voice processing apparatus characterized by comprising
Display module, for showing current display interface;
Monitoring modular, the trigger action inputted for monitoring touch object relative to the current display interface;
Execution module, for executing corresponding speech processes operation according to the trigger action, wherein the speech processes behaviour Include at least one of the following: open phonetic function, cancel this input voice, send this input voice to communication pair End.
8. device according to claim 7, which is characterized in that
The execution module includes:
Acquisition submodule, the position correlation for when the trigger action includes slide, obtaining the slide are believed Breath, wherein the location dependent information includes at least one of the following: current location of the slide relative to screen, institute State sliding distance of the slide relative to screen;
Implementation sub-module, for executing speech processes corresponding with the location dependent information according to the location dependent information Operation.
9. device according to claim 8, which is characterized in that
The implementation sub-module includes:
Opening unit, for including the slide relative to the sliding distance of screen and described when the location dependent information When slide is slide downward operation of the touch object relative to the current display interface, if the slide downward is grasped The distance of glide of work reaches the first pre-determined distance, then opens phonetic function, obtains this input language to allow user to input voice Sound;
Revocation unit, for being upward sliding of the touch object relative to the current display interface when the slide When operation, if the upslip of upward sliding operation cancels the voice of this input from the second pre-determined distance is reached.
10. device according to claim 7, which is characterized in that
The execution module includes:
Sending submodule, for when the trigger action be the touch object leave the operation of the current display interface when, The voice of this input is sent to Correspondent Node.
11. device according to any one of claims 7 to 10, which is characterized in that described device further include:
Cue module, for being prompted by default highlight user, wherein the prompt includes: by the triggering The executable speech processes of operation operate, will execute the speech processes operation, speech processes operation has executed It will be hidden after preset duration at, the default highlight.
CN201811044257.1A 2018-09-07 2018-09-07 Method of speech processing and device Pending CN109120793A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811044257.1A CN109120793A (en) 2018-09-07 2018-09-07 Method of speech processing and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811044257.1A CN109120793A (en) 2018-09-07 2018-09-07 Method of speech processing and device

Publications (1)

Publication Number Publication Date
CN109120793A true CN109120793A (en) 2019-01-01

Family

ID=64858922

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811044257.1A Pending CN109120793A (en) 2018-09-07 2018-09-07 Method of speech processing and device

Country Status (1)

Country Link
CN (1) CN109120793A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110069200A (en) * 2019-04-24 2019-07-30 努比亚技术有限公司 Wearable device input control method, wearable device and storage medium
CN110837334A (en) * 2019-11-04 2020-02-25 北京字节跳动网络技术有限公司 Method, device, terminal and storage medium for interactive control

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140099123A (en) * 2013-02-01 2014-08-11 엘지전자 주식회사 Mobile terminal and control method thereof
CN104144239A (en) * 2013-09-25 2014-11-12 腾讯科技(深圳)有限公司 Voice assist communication method and device
CN104375702A (en) * 2014-10-31 2015-02-25 北京搜狗科技发展有限公司 Touch operation method and device
CN105446489A (en) * 2015-12-08 2016-03-30 广州神马移动信息科技有限公司 Voice dual-mode control method and apparatus, and user terminal
CN106959746A (en) * 2016-01-12 2017-07-18 百度在线网络技术(北京)有限公司 The processing method and processing device of speech data

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140099123A (en) * 2013-02-01 2014-08-11 엘지전자 주식회사 Mobile terminal and control method thereof
CN104144239A (en) * 2013-09-25 2014-11-12 腾讯科技(深圳)有限公司 Voice assist communication method and device
CN104375702A (en) * 2014-10-31 2015-02-25 北京搜狗科技发展有限公司 Touch operation method and device
CN105446489A (en) * 2015-12-08 2016-03-30 广州神马移动信息科技有限公司 Voice dual-mode control method and apparatus, and user terminal
CN106959746A (en) * 2016-01-12 2017-07-18 百度在线网络技术(北京)有限公司 The processing method and processing device of speech data

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110069200A (en) * 2019-04-24 2019-07-30 努比亚技术有限公司 Wearable device input control method, wearable device and storage medium
CN110069200B (en) * 2019-04-24 2024-02-09 努比亚技术有限公司 Wearable device input control method, wearable device and storage medium
CN110837334A (en) * 2019-11-04 2020-02-25 北京字节跳动网络技术有限公司 Method, device, terminal and storage medium for interactive control

Similar Documents

Publication Publication Date Title
EP3163569B1 (en) Method and device for controlling a smart device by voice, computer program and recording medium
CN105607805B (en) The footmark processing method and processing device of application icon
CN106547663B (en) Key management method and device
CN108509232A (en) Screen recording method, device and computer readable storage medium
CN105975156B (en) Application interface display methods and device
CN105704766B (en) The control method and device of double-card mobile terminal
CN105160239A (en) Application program access restriction method and apparatus
CN105224171B (en) Icon display method, device and terminal
CN105335062B (en) Information editing's operation processing method and device
CN104394137B (en) A kind of method and device of prompting voice call
CN107608561A (en) Touch-screen control method and device
CN106791921A (en) The processing method and processing device of net cast
CN106791092A (en) The searching method and device of contact person
CN107105517A (en) Method for connecting network and device
CN109951379A (en) Message treatment method and device
CN105468281B (en) The method and apparatus for showing set interface
CN106990903A (en) Display and the method and device of hide application program
CN106775377A (en) The control method of gesture identifying device, equipment and gesture identifying device
CN105187671A (en) Recording method and device
CN107566615B (en) Message treatment method, device and computer readable storage medium
CN109120793A (en) Method of speech processing and device
CN109325337A (en) Unlocking method and device
CN106096442B (en) Applications trigger method and device
CN106101372B (en) Using prompting setting method and device
CN108766427A (en) Sound control method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190101