CN112230811A - Input method, device, equipment and storage medium - Google Patents
Input method, device, equipment and storage medium Download PDFInfo
- Publication number
- CN112230811A CN112230811A CN202011102752.0A CN202011102752A CN112230811A CN 112230811 A CN112230811 A CN 112230811A CN 202011102752 A CN202011102752 A CN 202011102752A CN 112230811 A CN112230811 A CN 112230811A
- Authority
- CN
- China
- Prior art keywords
- input
- text
- control icon
- voice
- target
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 73
- 230000001960 triggered effect Effects 0.000 claims abstract description 21
- 230000008569 process Effects 0.000 claims description 35
- 238000012549 training Methods 0.000 claims description 16
- 230000004044 response Effects 0.000 claims description 14
- 238000004590 computer program Methods 0.000 claims description 3
- 238000001514 detection method Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 16
- 238000004891 communication Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 4
- 238000013519 translation Methods 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000000903 blocking effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 241000209202 Bromus secalinus Species 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04817—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0486—Drag-and-drop
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/221—Announcement of recognition results
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The application discloses an input method, a device, equipment and a storage medium, the application displays a first control icon near the detected input focus position by detecting the position of the input focus in an information input interface, the first control icon is used for indicating voice input when being triggered, a user can perform voice input by triggering the first control icon, the application can respond to the triggering operation of the first control icon to acquire input voice and recognize the input voice as a text, and the recognized text is displayed at the position of the input focus in the information input interface The method is quick, and is particularly suitable for input scenes with small display screen size and inconvenient virtual keyboard setting.
Description
Technical Field
The present application relates to the field of information input technologies, and in particular, to an input method, an input apparatus, an input device, and a storage medium.
Background
The existing mobile equipment mostly adopts a virtual keyboard mode, and a user can perform interactive modes such as pinyin, handwriting, strokes, voice and the like through the virtual keyboard.
For some use scenarios, which are not suitable for inputting by means of a virtual keyboard, especially for some small-screen use scenarios, there is often no redundant space for providing a virtual keyboard display, such as wearing a device, and the existing solutions mainly have the following ways:
firstly, a free input mode is not provided, only built-in phrases are provided (a user can predefine in advance), and all reply scenes cannot be met when the user needs to input proper selection input;
and the second one is to force the use of virtual keyboard, which can be used for inputting pinyin, etc., but the keys are small and not easy to click.
Disclosure of Invention
In view of the foregoing problems, the present application provides an input method, apparatus, device and storage medium to achieve convenient and fast information input in a manner of separating from virtual keyboard input. The specific scheme is as follows:
an input method, comprising:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
Preferably, the method further comprises the following steps:
if the input text exists before the position of the input focus in the information input interface is determined, displaying a second control icon near the detected position of the input focus, wherein the second control icon is used for indicating that grid retreating operation is performed when the second control icon is triggered;
and responding to the triggering operation of the second control icon, and backing up the grid forwards by taking the input focus as a reference.
Preferably, the method further comprises the following steps:
hiding or destroying the first control icon in the process of responding to the triggering operation of the first control icon and acquiring input voice;
and after the end of the input voice is detected, redisplaying the first control icon.
Preferably, the method further comprises the following steps:
displaying a voice signal animation at the input focus position in the process of responding to the triggering operation of the first control icon and acquiring input voice, wherein the voice signal animation is used for indicating that the voice is currently input;
and after the end of the input voice is detected, hiding or destroying the voice signal animation.
Preferably, the method further comprises the following steps:
when the selected target text is detected to exist in the information input interface, displaying a third control icon near the target text, wherein the third control icon is used for indicating that replacement operation is performed when the third control icon is triggered;
responding to the triggering operation of the third control element icon, and displaying a candidate text popup, wherein the candidate text popup comprises at least one candidate text;
and responding to the selection operation of the target candidate text in the candidate text popup, and replacing the target text in the information input interface with the target candidate text.
Preferably, the process of determining candidate texts in the candidate text popup includes:
acquiring the target text and context information thereof;
determining candidate text for replacing the target text based on the context information.
Preferably, the determining the candidate text for replacing the target text based on the context information includes:
determining a target corpus matched with the context information based on the existing corpus knowledge;
determining a link corpus text linking the context information in the target corpus;
and determining candidate texts for replacing the target texts based on the connective corpus texts.
Preferably, the determining the candidate text for replacing the target text based on the context information includes:
predicting a candidate text used for connecting the context information based on the context information and a preset text prediction model, and determining the predicted candidate text as a candidate text used for replacing the target text;
the text prediction model is obtained by taking the context training data as a training sample and taking the candidate text connected with the context training data as a sample label for training.
Preferably, the method further comprises the following steps:
and hiding or destroying the candidate text popup in response to the selected operation of the target candidate text in the candidate text popup.
An input device, comprising:
an input focus detection unit for detecting a position of an input focus in the information input interface;
the first control icon display unit is used for displaying a first control icon near the detected position of the input focus, and the first control icon is used for indicating voice input when being triggered;
the first control icon triggering response unit is used for responding to triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and the recognized text display unit is used for displaying the recognized text at the position of the input focus in the information input interface.
An input device, comprising: a memory and a processor;
the memory is used for storing programs;
the processor is configured to execute the program to implement the steps of the input method.
A storage medium having stored thereon a computer program which, when executed by a processor, carries out the steps of the input method as described above.
By means of the technical scheme, the input method comprises the steps that the position of the input focus in the information input interface is detected, the first control icon is displayed near the detected position of the input focus and used for indicating voice input when the first control icon is triggered, a user can perform voice input by triggering the first control icon, then the input voice can be obtained and recognized as a text in response to the triggering operation of the first control icon, the recognized text is displayed at the position of the input focus in the information input interface, and based on the scheme provided by the application, the traditional virtual keyboard is removed, the input focus of the user is fixed, the input process of the whole information can be completed only by displaying the first control icon near the position of the input focus in cooperation with voice input and recognition, the input process is more convenient and rapid, and the method is particularly suitable for displaying the text with a small display screen size, It is inconvenient to set an input scene of the virtual keyboard.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the preferred embodiments. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the application. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
fig. 1 is a schematic flow chart of an input method according to an embodiment of the present disclosure;
FIG. 2 illustrates an information input interface diagram;
FIG. 3 illustrates a process diagram for a user triggering a first control icon;
FIG. 4 illustrates another information input interface diagram;
FIG. 5 illustrates a process diagram for a user triggering a second control icon;
FIG. 6 illustrates a process diagram for a user entering text at an information input interface;
FIG. 7 is a process diagram illustrating a user performing a replacement operation on an input text in an information input interface;
fig. 8 is a schematic structural diagram of an input device according to an embodiment of the present disclosure;
fig. 9 is a schematic structural diagram of an input device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The application provides an input scheme, a virtual keyboard is not required to be arranged in an information input interface, only the first control icon is displayed near the position of an input focus, and the whole information input process can be completed in a voice input mode.
This application scheme can be applied to various terminal equipment and various information input scenes, and especially be applicable to the terminal screen size less, be not convenient for set up virtual keyboard's input scene, like wearing equipment: and a scene needing information input under a small-size screen such as a bracelet.
The scheme can be realized based on a terminal with data processing capacity, and the terminal can be intelligent wearable equipment, a mobile phone, a computer and the like.
Next, as described in conjunction with fig. 1, the input method of the present application may include the steps of:
and S100, detecting the position of the focus in the information input interface.
Specifically, when the user inputs information at the terminal, the user may open an information input interface, which may be an interface that needs to input information, for example, a document interface may be opened when inputting information in a document, a browser search interface may be opened when inputting information in a browser search box, and the like.
When the information input interface inputs information, an input focus, such as an input cursor, is generally indicated. The input focus indicates a position at which information input is currently to be performed. The user can adjust the position of the input focus according to the input requirement. In this step, the terminal may detect an input focus in the input interface to obtain position information of the input focus.
Step S110 is to display a first control icon near the detected position of the input focus.
Specifically, the first control icon is used for indicating that voice input is performed when the first control icon is triggered.
In this step, in order to facilitate the user to perform voice input, a first control icon may be displayed near the position of the input focus, and a prompt such as "input" may be marked on the first control icon to prompt the user to implement voice input by triggering the first control icon.
The vicinity of the position of the input focus is understood to be each position region having a predetermined positional relationship with the reference position point, which is the input focus. The set position relationship may include a relative position direction and a relative position distance.
It should be noted that the specific meaning of "near the position of the input focus" mentioned below may refer to the explanation herein, and the explanation will not be repeated.
Based on the understanding of the proximity of the position of the input focus, the first control icon may be displayed according to a set positional relationship based on the position of the input focus, for example, may be displayed in the upper right corner, the lower right corner, or the same input line located at the right side of the input focus, and so on.
For example, when the first control icon is displayed at the upper right corner or the lower right corner of the input focus, to avoid the first control icon from blocking the input content, the size of the first control icon may be set based on the line spacing between the two text input lines, so that the first control icon is located directly between the two text input lines without blocking the input content.
In another optional case, in consideration of convenience of user triggering, the size of the first control icon may be determined based on a touch area when the user triggers, so as to ensure that the user can accurately trigger the first control icon.
Referring to FIG. 2, a schematic diagram of an information input interface is illustrated.
In the information input interface 10, an input focus 11 is detected. The first control icon 12 may then be displayed in the vicinity of the input focus 11. As shown in fig. 2, a first control icon 12 may be displayed at the upper right corner of the input focus 11.
And step S120, responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text.
Specifically, for a first control icon displayed in the information input interface, a user can trigger the first control icon, and the triggering mode can be clicking, double clicking, long pressing and other operations. By triggering the first control icon, the user can input voice, namely, text content to be expressed is input into the information input interface in a voice mode.
In the step, the terminal responds to the triggering operation of the user on the first control icon, so that input voice is obtained, and the input voice is identified and transcribed to obtain an identified text.
And step S130, displaying the recognized text at the position of the input focus in the information input interface.
Specifically, after the recognized text corresponding to the voice is obtained, the recognized text is displayed on the information input interface with the position of the input focus as a reference, so that the purpose of inputting text content into the information input interface is fulfilled.
Referring to FIG. 3, a process diagram of a user triggering a first control icon is illustrated.
The user can trigger the first control icon 12 and input text in the form of voice, and the terminal recognizes the input voice to obtain the input text "see he spit the last breath".
According to the input method, the position of the input focus in the information input interface is detected, the first control icon is displayed near the detected position of the input focus and used for indicating voice input when the first control icon is triggered, a user can perform voice input by triggering the first control icon, then the input voice can be obtained and recognized as a text in response to the triggering operation of the first control icon, the recognized text is displayed at the position of the input focus in the information input interface, based on the scheme provided by the application, the traditional virtual keyboard is removed, meanwhile, the input focus of the user is fixed, the input process of the whole information can be completed only by displaying the first control icon near the position of the input focus and matching with the voice input and recognition, the input process is more convenient and rapid, and the method is particularly suitable for displaying the input process with a small size display screen, It is inconvenient to set an input scene of the virtual keyboard.
In some embodiments of the present application, in order to facilitate the user to edit the input content, such as a delete operation, a solution is provided, as follows:
and if the input text exists before the position of the input focus in the information input interface is determined, displaying a second control icon near the detected position of the input focus.
And the second control icon is used for indicating that the backspace operation is performed when the second control icon is triggered.
It is to be appreciated that the second control icon may intersect the display location of the first control icon, e.g., the display locations may overlap. In addition, the display positions of the second control icon and the first control icon may also be separated, i.e., the display positions may not overlap. In the embodiment of the application, the second control icon and the first control icon can be set to support the movement of the display position under the operation of dragging by a user. That is, when it is detected that the user "drags" (e.g., selects and moves) the first or second control icon to the target position, the display position of the control icon to which the "dragging" operation is directed is moved to the target position.
By supporting the dragging of the first or second control icon, the display position of the control icon can be adjusted according to the operation requirement of the user, and the editing operation of the user is more convenient.
For example, the display positions of the second control icon and the first control icon are not overlapped, the second control icon and the first control icon may be arranged in sequence in the horizontal direction, or may be arranged in parallel in the longitudinal direction, or the first control icon and the second control icon are respectively located at different orientations of the input focus, for example, the first control icon and the second control icon are respectively located at the upper right corner and the lower right corner of the input focus.
Referring to FIG. 4, a schematic diagram of an information input interface is illustrated.
As can be seen from fig. 4, the input content exists before the input focus 11, and at this time, the user may have a need to delete the input content, and for this purpose, in this embodiment of the present application, the second control icon 13 may be further displayed while the first control icon 12 is displayed in the upper right corner of the input focus.
The second control icon 13 may be marked with a prompt such as "backspace" to prompt the user to implement backspace operation by triggering the second control icon 13.
On the basis, the terminal can respond to the triggering operation of the user on the second control icon, and retreat forward with the input focus as a reference.
It can be understood that if there is no selected text before the input focus, the text can be deleted by backing up in sequence according to the trigger operation of the user, for example, the text can be deleted by backing up one unit each time the second control icon is clicked. In addition, if the input focus is currently in a selection state, namely a plurality of text contents are currently selected, the second control icon is clicked, the selected text contents can be deleted firstly, and the selected text contents can be clicked again to retreat forward by one unit.
Referring to FIG. 5, a schematic diagram illustrating a process for a user to trigger the second control icon is shown.
The user clicks the second control icon 13 once, and the user will back over one unit, that is, delete the text content that originally existed.
In some embodiments of the present application, a voice input scheme is provided, and in a process of acquiring input voice in response to a trigger operation on a first control icon, the first control icon may be hidden or destroyed.
Further, the first control icon may be redisplayed after detecting that the input speech is over.
It can be understood that the form of triggering the first control icon by the user may be single-click, double-click, and the like, after the first control icon is triggered, the terminal performs the processes of acquiring and identifying the input voice, and at this time, the first control icon may not be displayed in the information input interface, so that the first control icon may be hidden or destroyed, and the visual influence of the first control icon on the input content may be avoided. After the input of the voice is finished, the first control icon may be redisplayed in order to facilitate the user to subsequently continue the voice input. Of course, the redisplayed first control icon is displayed based on the position of the latest input focus, and the display position may be different from the last display position.
Further optionally, in the process of acquiring the input voice in response to the triggering operation on the first control icon, on the basis of hiding or destroying the first control icon, in order to perform an interactive response on the input voice of the user, an operation of displaying a voice signal animation at the input focus position may be further added, where the voice signal animation is used to indicate that the voice is currently being input.
The displayed voice signal animation can be a fixed preset animation effect, and besides, the voice signal animation can also change along with the input voice, such as visual change which generates response along with the tone height, frequency size and the like of the input voice.
Further, after the end of the input voice is detected, the displayed voice signal animation can be hidden or destroyed, and the first control icon is further displayed.
By displaying the speech signal animation during the user's speech input, the user may be prompted that the speech input is currently in progress.
Referring to fig. 6, a schematic diagram of a process for a user to enter text in an information input interface is illustrated. The user can click the first control icon 12 once and the process of voice input can proceed. And the terminal responds to the triggered operation of the first control icon 12, hides the first control icon 12, and simultaneously displays the voice signal animation 14 at the input focus position. When the voice input is finished, the voice signal animation 14 is destroyed, and at the same time, the first control icon 12 is displayed near the position of the new input focus.
In some embodiments of the present application, in order to facilitate the user to edit the input content, such as modify, replace, etc., a solution is provided, as follows:
and when the selected target text is detected to exist in the information input interface, displaying a third control element icon near the target text.
Specifically, for the content that has been input in the information input interface, the user may select a text that needs to be edited, for example, select a text that needs to be edited by long-press dragging or the like. When detecting that the selected target text exists in the new input interface, the terminal can display a third control element icon near the target text. Wherein the third control icon is used for indicating that a replacement operation is performed when triggered.
The third control icon may be displayed at a set position, such as above or below the target text. It is understood that the display positions of the third control icon and the second control icon and the first control icon may or may not overlap. In addition, in the embodiment of the present application, a third control icon may be set to support movement of the display position under a "drag" operation of the user. For a specific implementation of the "drag" operation, reference may be made to the foregoing description, which is not described herein.
By supporting the dragging of the third control icon, the display position of the control icon can be adjusted according to the operation requirement of the user, and the editing operation of the user is more convenient.
The third control icon may be marked with a prompt such as "replace" to prompt the user that the replacement operation may be implemented by triggering the third control icon.
The user can perform a trigger operation on the displayed third control icon, such as clicking, double clicking, long pressing and the like. On the basis, the terminal can respond to the triggering operation of the third control element icon and display the candidate text popup. The candidate text popup comprises at least one candidate text.
And the candidate texts in the candidate text popup are obtained by the terminal and used for replacing the candidate texts of the target text.
On the basis, the user can select a target candidate text for replacing the target text from the candidate text popup, and then the terminal responds to the selection operation of the target candidate text in the candidate text popup and replaces the target text in the information input interface with the target candidate text.
According to the method and the device for replacing the target text, the third control icon and the candidate text popup window are arranged, a user can directly select the target candidate text from the candidate text popup window, and then the target text to be edited and operated is replaced, so that the process of text input is omitted, and the whole text replacement operation process is more convenient and faster.
Referring to fig. 7, a schematic diagram illustrating a process of performing a replacement operation on an input text by a user in an information input interface is shown.
As can be seen from fig. 7, when the user needs to modify and replace the target text "do not leave yourself" in the information input interface 10, he can select it. The terminal displays a third control icon 15 in the information input interface 10 upon detecting the presence of the selected target text. If the user can perform a trigger operation on the third control icon 15, the terminal responds to the trigger operation, and further displays a candidate text popup 16 in the information input interface 10, where multiple candidate texts for replacing the target text may be stored in the candidate text popup 16, and a process of determining the candidate text is described in detail later.
The user may select a desired target candidate text from the candidate text popup 16 and select the target candidate text, and the terminal may replace the target text in the information input interface with the selected target candidate text.
Alternatively, after replacing the target text with the target candidate text, the first control icon 12 and the second control icon 13 may be further displayed near the position of the input focus 11.
Further optionally, when the user selects the target candidate text in the candidate text popup, it indicates that the user has determined the content for replacing the target text, and at this time, the candidate text popup may not be displayed in the information input interface any more.
As shown in fig. 7, when the user selects a target candidate text in the candidate text popup 16, the terminal may hide or destroy the candidate text popup 16 in the information input interface 10 while replacing the target text with the target candidate text.
In some embodiments of the present application, a process for determining candidate texts in the candidate text popup is further described.
It is understood that the candidate text is text determined by the terminal to be likely to replace the selected target text, that is, the candidate text may be more suitable for display in the information input interface than the target text.
Based on this, the terminal may determine candidate texts in the candidate text popup based on the target text. Alternatively, the terminal may obtain the target text and its context information (if the target text only has the above or below information, the existing above or below information may be obtained, and if the target text does not have the context information, the context information may be null), and further determine a candidate text for replacing the target text based on the context information.
Specifically, the process of determining the candidate text may be processed by a local analysis engine of the terminal, or the terminal may send the target text, or the target text and the context information together to the server, and the server performs analysis processing and feeds back the result to the terminal.
Taking local analysis and processing of the terminal as an example:
first, the terminal may determine candidate texts based on the target text, such as:
1) the terminal can judge whether the target text contains wrongly written characters or not, if so, the wrongly written characters are changed, and the changed text is used as a candidate text.
2) The terminal can retrieve texts with the same or similar meanings as the target texts, and the retrieved texts are respectively used as candidate texts.
3) The terminal can perform multilingual translation on the target text and take each translation result as a candidate text respectively.
Secondly, the terminal can also determine the splicing text which is more matched with the context information by combining the context information, and further takes the determined splicing text as a candidate text.
Next, several alternative implementations are presented in which the terminal determines candidate texts for replacing the target text based on the context information.
1) An alternative implementation:
the method and the device can collect the corpus knowledge in advance, and then determine the target corpus matched with the context information based on the existing corpus knowledge.
For example, the corpus knowledge may be searched for matches with the context information as search criteria, or with keywords in the context information as search criteria, to find matching target corpora.
Further, determining a link corpus text linking the context information in the target corpus.
For example as follows:
the target text is: "do not leave by oneself", its upper message is "if life cheats you,". The above information can be used to search and match the existing corpus knowledge to find the matching target corpus as "if you were deceived in life, do not sadness, do not worry! … …'
And then, in the target corpus, determining that a link corpus text linked with the above information is: "Do not sadness, don't worry! "
It should be noted that the above-mentioned matched target corpus may be multiple pieces, and the obtained link corpus text may also be multiple pieces.
After determining the corpus text, the terminal may determine candidate texts for replacing the target text based on the corpus text.
Specifically, the concatenative corpus text itself may be used as a candidate text. In addition, expansion can be performed based on the linked corpus text, for example, an expanded text having the same or similar meaning as the linked corpus text is searched, or multilingual translation is performed on the linked corpus text, and then the expanded text and the translation result can be used as candidate texts respectively.
2) Another optional implementation manner:
the text prediction model can be trained in advance, namely the text prediction model is trained by taking the context training data as a training sample and taking a candidate text which is connected with the context training data as a sample label. Specifically, the method and the device can collect the corpus knowledge, further randomly mark partial text segments in the corpus knowledge as sample labels, mark the marked text segments with context information as training samples, and train the text prediction model based on the sample labels.
After the text prediction model is trained, the context information can be input into the model, and the candidate text predicted by the model and used for connecting the context information can be obtained.
Optionally, the first N results with higher confidence may be selected from the model prediction results, and each result is respectively used as a candidate text and displayed in a form of a candidate text popup.
The following describes an input device provided in an embodiment of the present application, and the input device described below and the input method described above may be referred to correspondingly.
Referring to fig. 8, fig. 8 is a schematic structural diagram of an input device disclosed in the embodiment of the present application.
As shown in fig. 8, the apparatus may include:
an input focus detection unit 11 for detecting a position of an input focus in the information input interface;
a first control icon display unit 12, configured to display a first control icon near the detected position of the input focus, where the first control icon is used to indicate that voice input is performed when triggered;
a first control icon trigger response unit 13, configured to respond to a trigger operation on the first control icon, acquire an input voice, and recognize the input voice as a text;
and the recognized text display unit 14 is used for displaying the recognized text at the position of the input focus in the information input interface.
Optionally, the input device of the present application may further include:
the second control icon display unit is used for displaying a second control icon near the detected position of the input focus if the input text exists before the position of the input focus in the information input interface is determined, wherein the second control icon is used for indicating that grid quitting operation is performed when the second control icon is triggered;
and the second control icon trigger response unit is used for responding to the trigger operation of the second control icon and retreating the grid forwards by taking the input focus as a reference.
Optionally, the input device of the present application may further include:
the first control icon processing unit is used for hiding or destroying the first control icon in the process of responding to the triggering operation of the first control icon and acquiring input voice;
the first control icon display unit is further used for displaying the first control icon again after the input voice is detected to be finished.
Optionally, the input device of the present application may further include:
the voice signal animation display unit is used for displaying voice signal animation at the position of the input focus in the process of responding to the triggering operation of the first control icon and acquiring input voice, and the voice signal animation is used for indicating that the voice is currently input;
and the voice signal animation processing unit is used for hiding or destroying the voice signal animation after the end of the input voice is detected.
Optionally, the input device of the present application may further include:
the third control icon display unit is used for displaying a third control icon near the target text when the selected target text is detected to exist in the information input interface, and the third control icon is used for indicating that the replacement operation is performed when the third control icon is triggered;
a third control icon trigger response unit, configured to respond to a trigger operation on the third control icon and display a candidate text popup, where the candidate text popup includes at least one candidate text;
and the candidate text popup trigger response unit is used for responding to the selection operation of the target candidate text in the candidate text popup and replacing the target text in the information input interface with the target candidate text.
Optionally, the input device of the present application may further include:
and the candidate text popup processing unit is used for responding to the selection operation of the target candidate text in the candidate text popup and hiding or destroying the candidate text popup.
Optionally, the input device of the present application may further include: the candidate text determining unit is used for acquiring the target text and the context information thereof; determining candidate text for replacing the target text based on the context information.
Optionally, the process of determining, by the candidate text determining unit, the candidate text for replacing the target text based on the context information may include:
determining a target corpus matched with the context information based on the existing corpus knowledge;
determining a link corpus text linking the context information in the target corpus;
and determining candidate texts for replacing the target texts based on the connective corpus texts.
Optionally, the process of determining, by the candidate text determining unit, the candidate text for replacing the target text based on the context information may include:
predicting a candidate text used for connecting the context information based on the context information and a preset text prediction model, and determining the predicted candidate text as a candidate text used for replacing the target text;
the text prediction model is obtained by taking the context training data as a training sample and taking the candidate text connected with the context training data as a sample label for training.
The input device provided by the embodiment of the application can be applied to input equipment such as intelligent wearable equipment, a mobile phone, a computer and the like. Alternatively, fig. 9 shows a block diagram of a hardware structure of the input device, and referring to fig. 9, the hardware structure of the input device may include: at least one processor 1, at least one communication interface 2, at least one memory 3 and at least one communication bus 4;
in the embodiment of the application, the number of the processor 1, the communication interface 2, the memory 3 and the communication bus 4 is at least one, and the processor 1, the communication interface 2 and the memory 3 complete mutual communication through the communication bus 4;
the processor 1 may be a central processing unit CPU, or an application Specific Integrated circuit asic, or one or more Integrated circuits configured to implement embodiments of the present invention, etc.;
the memory 3 may include a high-speed RAM memory, and may further include a non-volatile memory (non-volatile memory) or the like, such as at least one disk memory;
wherein the memory stores a program and the processor can call the program stored in the memory, the program for:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
Alternatively, the detailed function and the extended function of the program may be as described above.
Embodiments of the present application further provide a storage medium, where a program suitable for execution by a processor may be stored, where the program is configured to:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
Alternatively, the detailed function and the extended function of the program may be as described above.
Finally, it should also be noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The embodiments in the present description are described in a progressive manner, each embodiment focuses on differences from other embodiments, the embodiments may be combined as needed, and the same and similar parts may be referred to each other.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present application. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the application. Thus, the present application is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Claims (12)
1. An input method, comprising:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
2. The method of claim 1, further comprising:
if the input text exists before the position of the input focus in the information input interface is determined, displaying a second control icon near the detected position of the input focus, wherein the second control icon is used for indicating that grid retreating operation is performed when the second control icon is triggered;
and responding to the triggering operation of the second control icon, and backing up the grid forwards by taking the input focus as a reference.
3. The method of claim 1, further comprising:
hiding or destroying the first control icon in the process of responding to the triggering operation of the first control icon and acquiring input voice;
and after the end of the input voice is detected, redisplaying the first control icon.
4. The method of claim 3, further comprising:
displaying a voice signal animation at the input focus position in the process of responding to the triggering operation of the first control icon and acquiring input voice, wherein the voice signal animation is used for indicating that the voice is currently input;
and after the end of the input voice is detected, hiding or destroying the voice signal animation.
5. The method of claim 1, further comprising:
when the selected target text is detected to exist in the information input interface, displaying a third control icon near the target text, wherein the third control icon is used for indicating that replacement operation is performed when the third control icon is triggered;
responding to the triggering operation of the third control element icon, and displaying a candidate text popup, wherein the candidate text popup comprises at least one candidate text;
and responding to the selection operation of the target candidate text in the candidate text popup, and replacing the target text in the information input interface with the target candidate text.
6. The method of claim 5, wherein the determining of the candidate text in the candidate text popup comprises:
acquiring the target text and context information thereof;
determining candidate text for replacing the target text based on the context information.
7. The method of claim 6, wherein determining candidate text for replacing the target text based on the context information comprises:
determining a target corpus matched with the context information based on the existing corpus knowledge;
determining a link corpus text linking the context information in the target corpus;
and determining candidate texts for replacing the target texts based on the connective corpus texts.
8. The method of claim 6, wherein determining candidate text for replacing the target text based on the context information comprises:
predicting a candidate text used for connecting the context information based on the context information and a preset text prediction model, and determining the predicted candidate text as a candidate text used for replacing the target text;
the text prediction model is obtained by taking the context training data as a training sample and taking the candidate text connected with the context training data as a sample label for training.
9. The method of claim 5, further comprising:
and hiding or destroying the candidate text popup in response to the selected operation of the target candidate text in the candidate text popup.
10. An input device, comprising:
an input focus detection unit for detecting a position of an input focus in the information input interface;
the first control icon display unit is used for displaying a first control icon near the detected position of the input focus, and the first control icon is used for indicating voice input when being triggered;
the first control icon triggering response unit is used for responding to triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and the recognized text display unit is used for displaying the recognized text at the position of the input focus in the information input interface.
11. An input device, comprising: a memory and a processor;
the memory is used for storing programs;
the processor is used for executing the program and realizing the steps of the input method according to any one of claims 1-9.
12. A storage medium having a computer program stored thereon, wherein the computer program, when executed by a processor, performs the steps of the input method according to any one of claims 1 to 9.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011102752.0A CN112230811A (en) | 2020-10-15 | 2020-10-15 | Input method, device, equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011102752.0A CN112230811A (en) | 2020-10-15 | 2020-10-15 | Input method, device, equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112230811A true CN112230811A (en) | 2021-01-15 |
Family
ID=74113619
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011102752.0A Pending CN112230811A (en) | 2020-10-15 | 2020-10-15 | Input method, device, equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112230811A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112951232A (en) * | 2021-03-02 | 2021-06-11 | 深圳创维-Rgb电子有限公司 | Voice input method, device, equipment and computer readable storage medium |
CN113301416A (en) * | 2021-04-30 | 2021-08-24 | 当趣网络科技(杭州)有限公司 | Method for displaying voice frame |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102141889A (en) * | 2010-02-12 | 2011-08-03 | 微软公司 | Typing assistance for editing |
US20120004898A1 (en) * | 2007-02-12 | 2012-01-05 | Google Inc. | Contextual Input Method |
WO2014054922A1 (en) * | 2012-10-05 | 2014-04-10 | Samsung Electronics Co., Ltd. | User terminal apparatus, method for inputting text in user terminal apparauts, and computer-readable recording medium |
CN104508604A (en) * | 2012-07-30 | 2015-04-08 | 微软公司 | Generating string predictions using contexts |
US20150100537A1 (en) * | 2013-10-03 | 2015-04-09 | Microsoft Corporation | Emoji for Text Predictions |
CN107346229A (en) * | 2017-07-18 | 2017-11-14 | 珠海市魅族科技有限公司 | Pronunciation inputting method and device, computer installation and readable storage medium storing program for executing |
CN107679032A (en) * | 2017-09-04 | 2018-02-09 | 百度在线网络技术(北京)有限公司 | Voice changes error correction method and device |
CN109215660A (en) * | 2018-07-09 | 2019-01-15 | 维沃移动通信有限公司 | Text error correction method and mobile terminal after speech recognition |
CN109739425A (en) * | 2018-04-19 | 2019-05-10 | 北京字节跳动网络技术有限公司 | A kind of dummy keyboard, pronunciation inputting method, device and electronic equipment |
-
2020
- 2020-10-15 CN CN202011102752.0A patent/CN112230811A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120004898A1 (en) * | 2007-02-12 | 2012-01-05 | Google Inc. | Contextual Input Method |
CN102141889A (en) * | 2010-02-12 | 2011-08-03 | 微软公司 | Typing assistance for editing |
CN104508604A (en) * | 2012-07-30 | 2015-04-08 | 微软公司 | Generating string predictions using contexts |
WO2014054922A1 (en) * | 2012-10-05 | 2014-04-10 | Samsung Electronics Co., Ltd. | User terminal apparatus, method for inputting text in user terminal apparauts, and computer-readable recording medium |
US20150100537A1 (en) * | 2013-10-03 | 2015-04-09 | Microsoft Corporation | Emoji for Text Predictions |
CN107346229A (en) * | 2017-07-18 | 2017-11-14 | 珠海市魅族科技有限公司 | Pronunciation inputting method and device, computer installation and readable storage medium storing program for executing |
CN107679032A (en) * | 2017-09-04 | 2018-02-09 | 百度在线网络技术(北京)有限公司 | Voice changes error correction method and device |
CN109739425A (en) * | 2018-04-19 | 2019-05-10 | 北京字节跳动网络技术有限公司 | A kind of dummy keyboard, pronunciation inputting method, device and electronic equipment |
CN109215660A (en) * | 2018-07-09 | 2019-01-15 | 维沃移动通信有限公司 | Text error correction method and mobile terminal after speech recognition |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112951232A (en) * | 2021-03-02 | 2021-06-11 | 深圳创维-Rgb电子有限公司 | Voice input method, device, equipment and computer readable storage medium |
CN113301416A (en) * | 2021-04-30 | 2021-08-24 | 当趣网络科技(杭州)有限公司 | Method for displaying voice frame |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9703462B2 (en) | Display-independent recognition of graphical user interface control | |
US9098313B2 (en) | Recording display-independent computerized guidance | |
CN205427822U (en) | Electronic equipment and device that is used for editing text | |
KR102238809B1 (en) | Actionable content displayed on a touch screen | |
CN102999274B (en) | Semantic zoom animation | |
KR102005878B1 (en) | Managing real-time handwriting recognition | |
US8060841B2 (en) | Method and device for touchless media searching | |
US9405558B2 (en) | Display-independent computerized guidance | |
KR20180112031A (en) | Systems and methods for providing content selection | |
WO2016095689A1 (en) | Recognition and searching method and system based on repeated touch-control operations on terminal interface | |
US9996522B2 (en) | Dictionary device for determining a search method based on a type of a detected touch operation | |
JP2016524205A (en) | Permanent synchronization system for handwriting input | |
JP2011081778A (en) | Method and device for display-independent computerized guidance | |
CN104133815B (en) | The method and system of input and search | |
CN112230811A (en) | Input method, device, equipment and storage medium | |
US20110022956A1 (en) | Chinese Character Input Device and Method Thereof | |
CN113901186A (en) | Telephone recording marking method, device, equipment and storage medium | |
CN106970899B (en) | Text processing method and device | |
US9361859B2 (en) | Information processing device, method, and computer program product | |
KR20150097250A (en) | Sketch retrieval system using tag information, user equipment, service equipment, service method and computer readable medium having computer program recorded therefor | |
KR102159331B1 (en) | Sketch Retrieval system, user equipment, service equipment and service method based on meteorological phenomena information and computer readable medium having computer program recorded therefor | |
CN112764551A (en) | Vocabulary display method and device and electronic equipment | |
KR102138095B1 (en) | Voice command based virtual touch input apparatus | |
JP2002149679A (en) | System and method for browsing document and information recording medium | |
CN113297519A (en) | Comment data generation method and device, electronic equipment and computer storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210115 |