CN112230811A - Input method, device, equipment and storage medium - Google Patents

Input method, device, equipment and storage medium Download PDF

Info

Publication number
CN112230811A
CN112230811A CN202011102752.0A CN202011102752A CN112230811A CN 112230811 A CN112230811 A CN 112230811A CN 202011102752 A CN202011102752 A CN 202011102752A CN 112230811 A CN112230811 A CN 112230811A
Authority
CN
China
Prior art keywords
input
text
control icon
voice
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011102752.0A
Other languages
Chinese (zh)
Inventor
陶涛
余飞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
iFlytek Co Ltd
Original Assignee
iFlytek Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by iFlytek Co Ltd filed Critical iFlytek Co Ltd
Priority to CN202011102752.0A priority Critical patent/CN112230811A/en
Publication of CN112230811A publication Critical patent/CN112230811A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04817Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0486Drag-and-drop
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/221Announcement of recognition results

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses an input method, a device, equipment and a storage medium, the application displays a first control icon near the detected input focus position by detecting the position of the input focus in an information input interface, the first control icon is used for indicating voice input when being triggered, a user can perform voice input by triggering the first control icon, the application can respond to the triggering operation of the first control icon to acquire input voice and recognize the input voice as a text, and the recognized text is displayed at the position of the input focus in the information input interface The method is quick, and is particularly suitable for input scenes with small display screen size and inconvenient virtual keyboard setting.

Description

Input method, device, equipment and storage medium
Technical Field
The present application relates to the field of information input technologies, and in particular, to an input method, an input apparatus, an input device, and a storage medium.
Background
The existing mobile equipment mostly adopts a virtual keyboard mode, and a user can perform interactive modes such as pinyin, handwriting, strokes, voice and the like through the virtual keyboard.
For some use scenarios, which are not suitable for inputting by means of a virtual keyboard, especially for some small-screen use scenarios, there is often no redundant space for providing a virtual keyboard display, such as wearing a device, and the existing solutions mainly have the following ways:
firstly, a free input mode is not provided, only built-in phrases are provided (a user can predefine in advance), and all reply scenes cannot be met when the user needs to input proper selection input;
and the second one is to force the use of virtual keyboard, which can be used for inputting pinyin, etc., but the keys are small and not easy to click.
Disclosure of Invention
In view of the foregoing problems, the present application provides an input method, apparatus, device and storage medium to achieve convenient and fast information input in a manner of separating from virtual keyboard input. The specific scheme is as follows:
an input method, comprising:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
Preferably, the method further comprises the following steps:
if the input text exists before the position of the input focus in the information input interface is determined, displaying a second control icon near the detected position of the input focus, wherein the second control icon is used for indicating that grid retreating operation is performed when the second control icon is triggered;
and responding to the triggering operation of the second control icon, and backing up the grid forwards by taking the input focus as a reference.
Preferably, the method further comprises the following steps:
hiding or destroying the first control icon in the process of responding to the triggering operation of the first control icon and acquiring input voice;
and after the end of the input voice is detected, redisplaying the first control icon.
Preferably, the method further comprises the following steps:
displaying a voice signal animation at the input focus position in the process of responding to the triggering operation of the first control icon and acquiring input voice, wherein the voice signal animation is used for indicating that the voice is currently input;
and after the end of the input voice is detected, hiding or destroying the voice signal animation.
Preferably, the method further comprises the following steps:
when the selected target text is detected to exist in the information input interface, displaying a third control icon near the target text, wherein the third control icon is used for indicating that replacement operation is performed when the third control icon is triggered;
responding to the triggering operation of the third control element icon, and displaying a candidate text popup, wherein the candidate text popup comprises at least one candidate text;
and responding to the selection operation of the target candidate text in the candidate text popup, and replacing the target text in the information input interface with the target candidate text.
Preferably, the process of determining candidate texts in the candidate text popup includes:
acquiring the target text and context information thereof;
determining candidate text for replacing the target text based on the context information.
Preferably, the determining the candidate text for replacing the target text based on the context information includes:
determining a target corpus matched with the context information based on the existing corpus knowledge;
determining a link corpus text linking the context information in the target corpus;
and determining candidate texts for replacing the target texts based on the connective corpus texts.
Preferably, the determining the candidate text for replacing the target text based on the context information includes:
predicting a candidate text used for connecting the context information based on the context information and a preset text prediction model, and determining the predicted candidate text as a candidate text used for replacing the target text;
the text prediction model is obtained by taking the context training data as a training sample and taking the candidate text connected with the context training data as a sample label for training.
Preferably, the method further comprises the following steps:
and hiding or destroying the candidate text popup in response to the selected operation of the target candidate text in the candidate text popup.
An input device, comprising:
an input focus detection unit for detecting a position of an input focus in the information input interface;
the first control icon display unit is used for displaying a first control icon near the detected position of the input focus, and the first control icon is used for indicating voice input when being triggered;
the first control icon triggering response unit is used for responding to triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and the recognized text display unit is used for displaying the recognized text at the position of the input focus in the information input interface.
An input device, comprising: a memory and a processor;
the memory is used for storing programs;
the processor is configured to execute the program to implement the steps of the input method.
A storage medium having stored thereon a computer program which, when executed by a processor, carries out the steps of the input method as described above.
By means of the technical scheme, the input method comprises the steps that the position of the input focus in the information input interface is detected, the first control icon is displayed near the detected position of the input focus and used for indicating voice input when the first control icon is triggered, a user can perform voice input by triggering the first control icon, then the input voice can be obtained and recognized as a text in response to the triggering operation of the first control icon, the recognized text is displayed at the position of the input focus in the information input interface, and based on the scheme provided by the application, the traditional virtual keyboard is removed, the input focus of the user is fixed, the input process of the whole information can be completed only by displaying the first control icon near the position of the input focus in cooperation with voice input and recognition, the input process is more convenient and rapid, and the method is particularly suitable for displaying the text with a small display screen size, It is inconvenient to set an input scene of the virtual keyboard.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the preferred embodiments. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the application. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
fig. 1 is a schematic flow chart of an input method according to an embodiment of the present disclosure;
FIG. 2 illustrates an information input interface diagram;
FIG. 3 illustrates a process diagram for a user triggering a first control icon;
FIG. 4 illustrates another information input interface diagram;
FIG. 5 illustrates a process diagram for a user triggering a second control icon;
FIG. 6 illustrates a process diagram for a user entering text at an information input interface;
FIG. 7 is a process diagram illustrating a user performing a replacement operation on an input text in an information input interface;
fig. 8 is a schematic structural diagram of an input device according to an embodiment of the present disclosure;
fig. 9 is a schematic structural diagram of an input device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The application provides an input scheme, a virtual keyboard is not required to be arranged in an information input interface, only the first control icon is displayed near the position of an input focus, and the whole information input process can be completed in a voice input mode.
This application scheme can be applied to various terminal equipment and various information input scenes, and especially be applicable to the terminal screen size less, be not convenient for set up virtual keyboard's input scene, like wearing equipment: and a scene needing information input under a small-size screen such as a bracelet.
The scheme can be realized based on a terminal with data processing capacity, and the terminal can be intelligent wearable equipment, a mobile phone, a computer and the like.
Next, as described in conjunction with fig. 1, the input method of the present application may include the steps of:
and S100, detecting the position of the focus in the information input interface.
Specifically, when the user inputs information at the terminal, the user may open an information input interface, which may be an interface that needs to input information, for example, a document interface may be opened when inputting information in a document, a browser search interface may be opened when inputting information in a browser search box, and the like.
When the information input interface inputs information, an input focus, such as an input cursor, is generally indicated. The input focus indicates a position at which information input is currently to be performed. The user can adjust the position of the input focus according to the input requirement. In this step, the terminal may detect an input focus in the input interface to obtain position information of the input focus.
Step S110 is to display a first control icon near the detected position of the input focus.
Specifically, the first control icon is used for indicating that voice input is performed when the first control icon is triggered.
In this step, in order to facilitate the user to perform voice input, a first control icon may be displayed near the position of the input focus, and a prompt such as "input" may be marked on the first control icon to prompt the user to implement voice input by triggering the first control icon.
The vicinity of the position of the input focus is understood to be each position region having a predetermined positional relationship with the reference position point, which is the input focus. The set position relationship may include a relative position direction and a relative position distance.
It should be noted that the specific meaning of "near the position of the input focus" mentioned below may refer to the explanation herein, and the explanation will not be repeated.
Based on the understanding of the proximity of the position of the input focus, the first control icon may be displayed according to a set positional relationship based on the position of the input focus, for example, may be displayed in the upper right corner, the lower right corner, or the same input line located at the right side of the input focus, and so on.
For example, when the first control icon is displayed at the upper right corner or the lower right corner of the input focus, to avoid the first control icon from blocking the input content, the size of the first control icon may be set based on the line spacing between the two text input lines, so that the first control icon is located directly between the two text input lines without blocking the input content.
In another optional case, in consideration of convenience of user triggering, the size of the first control icon may be determined based on a touch area when the user triggers, so as to ensure that the user can accurately trigger the first control icon.
Referring to FIG. 2, a schematic diagram of an information input interface is illustrated.
In the information input interface 10, an input focus 11 is detected. The first control icon 12 may then be displayed in the vicinity of the input focus 11. As shown in fig. 2, a first control icon 12 may be displayed at the upper right corner of the input focus 11.
And step S120, responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text.
Specifically, for a first control icon displayed in the information input interface, a user can trigger the first control icon, and the triggering mode can be clicking, double clicking, long pressing and other operations. By triggering the first control icon, the user can input voice, namely, text content to be expressed is input into the information input interface in a voice mode.
In the step, the terminal responds to the triggering operation of the user on the first control icon, so that input voice is obtained, and the input voice is identified and transcribed to obtain an identified text.
And step S130, displaying the recognized text at the position of the input focus in the information input interface.
Specifically, after the recognized text corresponding to the voice is obtained, the recognized text is displayed on the information input interface with the position of the input focus as a reference, so that the purpose of inputting text content into the information input interface is fulfilled.
Referring to FIG. 3, a process diagram of a user triggering a first control icon is illustrated.
The user can trigger the first control icon 12 and input text in the form of voice, and the terminal recognizes the input voice to obtain the input text "see he spit the last breath".
According to the input method, the position of the input focus in the information input interface is detected, the first control icon is displayed near the detected position of the input focus and used for indicating voice input when the first control icon is triggered, a user can perform voice input by triggering the first control icon, then the input voice can be obtained and recognized as a text in response to the triggering operation of the first control icon, the recognized text is displayed at the position of the input focus in the information input interface, based on the scheme provided by the application, the traditional virtual keyboard is removed, meanwhile, the input focus of the user is fixed, the input process of the whole information can be completed only by displaying the first control icon near the position of the input focus and matching with the voice input and recognition, the input process is more convenient and rapid, and the method is particularly suitable for displaying the input process with a small size display screen, It is inconvenient to set an input scene of the virtual keyboard.
In some embodiments of the present application, in order to facilitate the user to edit the input content, such as a delete operation, a solution is provided, as follows:
and if the input text exists before the position of the input focus in the information input interface is determined, displaying a second control icon near the detected position of the input focus.
And the second control icon is used for indicating that the backspace operation is performed when the second control icon is triggered.
It is to be appreciated that the second control icon may intersect the display location of the first control icon, e.g., the display locations may overlap. In addition, the display positions of the second control icon and the first control icon may also be separated, i.e., the display positions may not overlap. In the embodiment of the application, the second control icon and the first control icon can be set to support the movement of the display position under the operation of dragging by a user. That is, when it is detected that the user "drags" (e.g., selects and moves) the first or second control icon to the target position, the display position of the control icon to which the "dragging" operation is directed is moved to the target position.
By supporting the dragging of the first or second control icon, the display position of the control icon can be adjusted according to the operation requirement of the user, and the editing operation of the user is more convenient.
For example, the display positions of the second control icon and the first control icon are not overlapped, the second control icon and the first control icon may be arranged in sequence in the horizontal direction, or may be arranged in parallel in the longitudinal direction, or the first control icon and the second control icon are respectively located at different orientations of the input focus, for example, the first control icon and the second control icon are respectively located at the upper right corner and the lower right corner of the input focus.
Referring to FIG. 4, a schematic diagram of an information input interface is illustrated.
As can be seen from fig. 4, the input content exists before the input focus 11, and at this time, the user may have a need to delete the input content, and for this purpose, in this embodiment of the present application, the second control icon 13 may be further displayed while the first control icon 12 is displayed in the upper right corner of the input focus.
The second control icon 13 may be marked with a prompt such as "backspace" to prompt the user to implement backspace operation by triggering the second control icon 13.
On the basis, the terminal can respond to the triggering operation of the user on the second control icon, and retreat forward with the input focus as a reference.
It can be understood that if there is no selected text before the input focus, the text can be deleted by backing up in sequence according to the trigger operation of the user, for example, the text can be deleted by backing up one unit each time the second control icon is clicked. In addition, if the input focus is currently in a selection state, namely a plurality of text contents are currently selected, the second control icon is clicked, the selected text contents can be deleted firstly, and the selected text contents can be clicked again to retreat forward by one unit.
Referring to FIG. 5, a schematic diagram illustrating a process for a user to trigger the second control icon is shown.
The user clicks the second control icon 13 once, and the user will back over one unit, that is, delete the text content that originally existed.
In some embodiments of the present application, a voice input scheme is provided, and in a process of acquiring input voice in response to a trigger operation on a first control icon, the first control icon may be hidden or destroyed.
Further, the first control icon may be redisplayed after detecting that the input speech is over.
It can be understood that the form of triggering the first control icon by the user may be single-click, double-click, and the like, after the first control icon is triggered, the terminal performs the processes of acquiring and identifying the input voice, and at this time, the first control icon may not be displayed in the information input interface, so that the first control icon may be hidden or destroyed, and the visual influence of the first control icon on the input content may be avoided. After the input of the voice is finished, the first control icon may be redisplayed in order to facilitate the user to subsequently continue the voice input. Of course, the redisplayed first control icon is displayed based on the position of the latest input focus, and the display position may be different from the last display position.
Further optionally, in the process of acquiring the input voice in response to the triggering operation on the first control icon, on the basis of hiding or destroying the first control icon, in order to perform an interactive response on the input voice of the user, an operation of displaying a voice signal animation at the input focus position may be further added, where the voice signal animation is used to indicate that the voice is currently being input.
The displayed voice signal animation can be a fixed preset animation effect, and besides, the voice signal animation can also change along with the input voice, such as visual change which generates response along with the tone height, frequency size and the like of the input voice.
Further, after the end of the input voice is detected, the displayed voice signal animation can be hidden or destroyed, and the first control icon is further displayed.
By displaying the speech signal animation during the user's speech input, the user may be prompted that the speech input is currently in progress.
Referring to fig. 6, a schematic diagram of a process for a user to enter text in an information input interface is illustrated. The user can click the first control icon 12 once and the process of voice input can proceed. And the terminal responds to the triggered operation of the first control icon 12, hides the first control icon 12, and simultaneously displays the voice signal animation 14 at the input focus position. When the voice input is finished, the voice signal animation 14 is destroyed, and at the same time, the first control icon 12 is displayed near the position of the new input focus.
In some embodiments of the present application, in order to facilitate the user to edit the input content, such as modify, replace, etc., a solution is provided, as follows:
and when the selected target text is detected to exist in the information input interface, displaying a third control element icon near the target text.
Specifically, for the content that has been input in the information input interface, the user may select a text that needs to be edited, for example, select a text that needs to be edited by long-press dragging or the like. When detecting that the selected target text exists in the new input interface, the terminal can display a third control element icon near the target text. Wherein the third control icon is used for indicating that a replacement operation is performed when triggered.
The third control icon may be displayed at a set position, such as above or below the target text. It is understood that the display positions of the third control icon and the second control icon and the first control icon may or may not overlap. In addition, in the embodiment of the present application, a third control icon may be set to support movement of the display position under a "drag" operation of the user. For a specific implementation of the "drag" operation, reference may be made to the foregoing description, which is not described herein.
By supporting the dragging of the third control icon, the display position of the control icon can be adjusted according to the operation requirement of the user, and the editing operation of the user is more convenient.
The third control icon may be marked with a prompt such as "replace" to prompt the user that the replacement operation may be implemented by triggering the third control icon.
The user can perform a trigger operation on the displayed third control icon, such as clicking, double clicking, long pressing and the like. On the basis, the terminal can respond to the triggering operation of the third control element icon and display the candidate text popup. The candidate text popup comprises at least one candidate text.
And the candidate texts in the candidate text popup are obtained by the terminal and used for replacing the candidate texts of the target text.
On the basis, the user can select a target candidate text for replacing the target text from the candidate text popup, and then the terminal responds to the selection operation of the target candidate text in the candidate text popup and replaces the target text in the information input interface with the target candidate text.
According to the method and the device for replacing the target text, the third control icon and the candidate text popup window are arranged, a user can directly select the target candidate text from the candidate text popup window, and then the target text to be edited and operated is replaced, so that the process of text input is omitted, and the whole text replacement operation process is more convenient and faster.
Referring to fig. 7, a schematic diagram illustrating a process of performing a replacement operation on an input text by a user in an information input interface is shown.
As can be seen from fig. 7, when the user needs to modify and replace the target text "do not leave yourself" in the information input interface 10, he can select it. The terminal displays a third control icon 15 in the information input interface 10 upon detecting the presence of the selected target text. If the user can perform a trigger operation on the third control icon 15, the terminal responds to the trigger operation, and further displays a candidate text popup 16 in the information input interface 10, where multiple candidate texts for replacing the target text may be stored in the candidate text popup 16, and a process of determining the candidate text is described in detail later.
The user may select a desired target candidate text from the candidate text popup 16 and select the target candidate text, and the terminal may replace the target text in the information input interface with the selected target candidate text.
Alternatively, after replacing the target text with the target candidate text, the first control icon 12 and the second control icon 13 may be further displayed near the position of the input focus 11.
Further optionally, when the user selects the target candidate text in the candidate text popup, it indicates that the user has determined the content for replacing the target text, and at this time, the candidate text popup may not be displayed in the information input interface any more.
As shown in fig. 7, when the user selects a target candidate text in the candidate text popup 16, the terminal may hide or destroy the candidate text popup 16 in the information input interface 10 while replacing the target text with the target candidate text.
In some embodiments of the present application, a process for determining candidate texts in the candidate text popup is further described.
It is understood that the candidate text is text determined by the terminal to be likely to replace the selected target text, that is, the candidate text may be more suitable for display in the information input interface than the target text.
Based on this, the terminal may determine candidate texts in the candidate text popup based on the target text. Alternatively, the terminal may obtain the target text and its context information (if the target text only has the above or below information, the existing above or below information may be obtained, and if the target text does not have the context information, the context information may be null), and further determine a candidate text for replacing the target text based on the context information.
Specifically, the process of determining the candidate text may be processed by a local analysis engine of the terminal, or the terminal may send the target text, or the target text and the context information together to the server, and the server performs analysis processing and feeds back the result to the terminal.
Taking local analysis and processing of the terminal as an example:
first, the terminal may determine candidate texts based on the target text, such as:
1) the terminal can judge whether the target text contains wrongly written characters or not, if so, the wrongly written characters are changed, and the changed text is used as a candidate text.
2) The terminal can retrieve texts with the same or similar meanings as the target texts, and the retrieved texts are respectively used as candidate texts.
3) The terminal can perform multilingual translation on the target text and take each translation result as a candidate text respectively.
Secondly, the terminal can also determine the splicing text which is more matched with the context information by combining the context information, and further takes the determined splicing text as a candidate text.
Next, several alternative implementations are presented in which the terminal determines candidate texts for replacing the target text based on the context information.
1) An alternative implementation:
the method and the device can collect the corpus knowledge in advance, and then determine the target corpus matched with the context information based on the existing corpus knowledge.
For example, the corpus knowledge may be searched for matches with the context information as search criteria, or with keywords in the context information as search criteria, to find matching target corpora.
Further, determining a link corpus text linking the context information in the target corpus.
For example as follows:
the target text is: "do not leave by oneself", its upper message is "if life cheats you,". The above information can be used to search and match the existing corpus knowledge to find the matching target corpus as "if you were deceived in life, do not sadness, do not worry! … …'
And then, in the target corpus, determining that a link corpus text linked with the above information is: "Do not sadness, don't worry! "
It should be noted that the above-mentioned matched target corpus may be multiple pieces, and the obtained link corpus text may also be multiple pieces.
After determining the corpus text, the terminal may determine candidate texts for replacing the target text based on the corpus text.
Specifically, the concatenative corpus text itself may be used as a candidate text. In addition, expansion can be performed based on the linked corpus text, for example, an expanded text having the same or similar meaning as the linked corpus text is searched, or multilingual translation is performed on the linked corpus text, and then the expanded text and the translation result can be used as candidate texts respectively.
2) Another optional implementation manner:
the text prediction model can be trained in advance, namely the text prediction model is trained by taking the context training data as a training sample and taking a candidate text which is connected with the context training data as a sample label. Specifically, the method and the device can collect the corpus knowledge, further randomly mark partial text segments in the corpus knowledge as sample labels, mark the marked text segments with context information as training samples, and train the text prediction model based on the sample labels.
After the text prediction model is trained, the context information can be input into the model, and the candidate text predicted by the model and used for connecting the context information can be obtained.
Optionally, the first N results with higher confidence may be selected from the model prediction results, and each result is respectively used as a candidate text and displayed in a form of a candidate text popup.
The following describes an input device provided in an embodiment of the present application, and the input device described below and the input method described above may be referred to correspondingly.
Referring to fig. 8, fig. 8 is a schematic structural diagram of an input device disclosed in the embodiment of the present application.
As shown in fig. 8, the apparatus may include:
an input focus detection unit 11 for detecting a position of an input focus in the information input interface;
a first control icon display unit 12, configured to display a first control icon near the detected position of the input focus, where the first control icon is used to indicate that voice input is performed when triggered;
a first control icon trigger response unit 13, configured to respond to a trigger operation on the first control icon, acquire an input voice, and recognize the input voice as a text;
and the recognized text display unit 14 is used for displaying the recognized text at the position of the input focus in the information input interface.
Optionally, the input device of the present application may further include:
the second control icon display unit is used for displaying a second control icon near the detected position of the input focus if the input text exists before the position of the input focus in the information input interface is determined, wherein the second control icon is used for indicating that grid quitting operation is performed when the second control icon is triggered;
and the second control icon trigger response unit is used for responding to the trigger operation of the second control icon and retreating the grid forwards by taking the input focus as a reference.
Optionally, the input device of the present application may further include:
the first control icon processing unit is used for hiding or destroying the first control icon in the process of responding to the triggering operation of the first control icon and acquiring input voice;
the first control icon display unit is further used for displaying the first control icon again after the input voice is detected to be finished.
Optionally, the input device of the present application may further include:
the voice signal animation display unit is used for displaying voice signal animation at the position of the input focus in the process of responding to the triggering operation of the first control icon and acquiring input voice, and the voice signal animation is used for indicating that the voice is currently input;
and the voice signal animation processing unit is used for hiding or destroying the voice signal animation after the end of the input voice is detected.
Optionally, the input device of the present application may further include:
the third control icon display unit is used for displaying a third control icon near the target text when the selected target text is detected to exist in the information input interface, and the third control icon is used for indicating that the replacement operation is performed when the third control icon is triggered;
a third control icon trigger response unit, configured to respond to a trigger operation on the third control icon and display a candidate text popup, where the candidate text popup includes at least one candidate text;
and the candidate text popup trigger response unit is used for responding to the selection operation of the target candidate text in the candidate text popup and replacing the target text in the information input interface with the target candidate text.
Optionally, the input device of the present application may further include:
and the candidate text popup processing unit is used for responding to the selection operation of the target candidate text in the candidate text popup and hiding or destroying the candidate text popup.
Optionally, the input device of the present application may further include: the candidate text determining unit is used for acquiring the target text and the context information thereof; determining candidate text for replacing the target text based on the context information.
Optionally, the process of determining, by the candidate text determining unit, the candidate text for replacing the target text based on the context information may include:
determining a target corpus matched with the context information based on the existing corpus knowledge;
determining a link corpus text linking the context information in the target corpus;
and determining candidate texts for replacing the target texts based on the connective corpus texts.
Optionally, the process of determining, by the candidate text determining unit, the candidate text for replacing the target text based on the context information may include:
predicting a candidate text used for connecting the context information based on the context information and a preset text prediction model, and determining the predicted candidate text as a candidate text used for replacing the target text;
the text prediction model is obtained by taking the context training data as a training sample and taking the candidate text connected with the context training data as a sample label for training.
The input device provided by the embodiment of the application can be applied to input equipment such as intelligent wearable equipment, a mobile phone, a computer and the like. Alternatively, fig. 9 shows a block diagram of a hardware structure of the input device, and referring to fig. 9, the hardware structure of the input device may include: at least one processor 1, at least one communication interface 2, at least one memory 3 and at least one communication bus 4;
in the embodiment of the application, the number of the processor 1, the communication interface 2, the memory 3 and the communication bus 4 is at least one, and the processor 1, the communication interface 2 and the memory 3 complete mutual communication through the communication bus 4;
the processor 1 may be a central processing unit CPU, or an application Specific Integrated circuit asic, or one or more Integrated circuits configured to implement embodiments of the present invention, etc.;
the memory 3 may include a high-speed RAM memory, and may further include a non-volatile memory (non-volatile memory) or the like, such as at least one disk memory;
wherein the memory stores a program and the processor can call the program stored in the memory, the program for:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
Alternatively, the detailed function and the extended function of the program may be as described above.
Embodiments of the present application further provide a storage medium, where a program suitable for execution by a processor may be stored, where the program is configured to:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
Alternatively, the detailed function and the extended function of the program may be as described above.
Finally, it should also be noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The embodiments in the present description are described in a progressive manner, each embodiment focuses on differences from other embodiments, the embodiments may be combined as needed, and the same and similar parts may be referred to each other.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present application. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the application. Thus, the present application is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (12)

1. An input method, comprising:
detecting the position of an input focus in an information input interface;
displaying a first control icon near the detected position of the input focus, wherein the first control icon is used for indicating that voice input is performed when being triggered;
responding to the triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and displaying the recognized text at the position of the input focus in the information input interface.
2. The method of claim 1, further comprising:
if the input text exists before the position of the input focus in the information input interface is determined, displaying a second control icon near the detected position of the input focus, wherein the second control icon is used for indicating that grid retreating operation is performed when the second control icon is triggered;
and responding to the triggering operation of the second control icon, and backing up the grid forwards by taking the input focus as a reference.
3. The method of claim 1, further comprising:
hiding or destroying the first control icon in the process of responding to the triggering operation of the first control icon and acquiring input voice;
and after the end of the input voice is detected, redisplaying the first control icon.
4. The method of claim 3, further comprising:
displaying a voice signal animation at the input focus position in the process of responding to the triggering operation of the first control icon and acquiring input voice, wherein the voice signal animation is used for indicating that the voice is currently input;
and after the end of the input voice is detected, hiding or destroying the voice signal animation.
5. The method of claim 1, further comprising:
when the selected target text is detected to exist in the information input interface, displaying a third control icon near the target text, wherein the third control icon is used for indicating that replacement operation is performed when the third control icon is triggered;
responding to the triggering operation of the third control element icon, and displaying a candidate text popup, wherein the candidate text popup comprises at least one candidate text;
and responding to the selection operation of the target candidate text in the candidate text popup, and replacing the target text in the information input interface with the target candidate text.
6. The method of claim 5, wherein the determining of the candidate text in the candidate text popup comprises:
acquiring the target text and context information thereof;
determining candidate text for replacing the target text based on the context information.
7. The method of claim 6, wherein determining candidate text for replacing the target text based on the context information comprises:
determining a target corpus matched with the context information based on the existing corpus knowledge;
determining a link corpus text linking the context information in the target corpus;
and determining candidate texts for replacing the target texts based on the connective corpus texts.
8. The method of claim 6, wherein determining candidate text for replacing the target text based on the context information comprises:
predicting a candidate text used for connecting the context information based on the context information and a preset text prediction model, and determining the predicted candidate text as a candidate text used for replacing the target text;
the text prediction model is obtained by taking the context training data as a training sample and taking the candidate text connected with the context training data as a sample label for training.
9. The method of claim 5, further comprising:
and hiding or destroying the candidate text popup in response to the selected operation of the target candidate text in the candidate text popup.
10. An input device, comprising:
an input focus detection unit for detecting a position of an input focus in the information input interface;
the first control icon display unit is used for displaying a first control icon near the detected position of the input focus, and the first control icon is used for indicating voice input when being triggered;
the first control icon triggering response unit is used for responding to triggering operation of the first control icon, acquiring input voice and recognizing the input voice as a text;
and the recognized text display unit is used for displaying the recognized text at the position of the input focus in the information input interface.
11. An input device, comprising: a memory and a processor;
the memory is used for storing programs;
the processor is used for executing the program and realizing the steps of the input method according to any one of claims 1-9.
12. A storage medium having a computer program stored thereon, wherein the computer program, when executed by a processor, performs the steps of the input method according to any one of claims 1 to 9.
CN202011102752.0A 2020-10-15 2020-10-15 Input method, device, equipment and storage medium Pending CN112230811A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011102752.0A CN112230811A (en) 2020-10-15 2020-10-15 Input method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011102752.0A CN112230811A (en) 2020-10-15 2020-10-15 Input method, device, equipment and storage medium

Publications (1)

Publication Number Publication Date
CN112230811A true CN112230811A (en) 2021-01-15

Family

ID=74113619

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011102752.0A Pending CN112230811A (en) 2020-10-15 2020-10-15 Input method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112230811A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112951232A (en) * 2021-03-02 2021-06-11 深圳创维-Rgb电子有限公司 Voice input method, device, equipment and computer readable storage medium
CN113301416A (en) * 2021-04-30 2021-08-24 当趣网络科技(杭州)有限公司 Method for displaying voice frame

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102141889A (en) * 2010-02-12 2011-08-03 微软公司 Typing assistance for editing
US20120004898A1 (en) * 2007-02-12 2012-01-05 Google Inc. Contextual Input Method
WO2014054922A1 (en) * 2012-10-05 2014-04-10 Samsung Electronics Co., Ltd. User terminal apparatus, method for inputting text in user terminal apparauts, and computer-readable recording medium
CN104508604A (en) * 2012-07-30 2015-04-08 微软公司 Generating string predictions using contexts
US20150100537A1 (en) * 2013-10-03 2015-04-09 Microsoft Corporation Emoji for Text Predictions
CN107346229A (en) * 2017-07-18 2017-11-14 珠海市魅族科技有限公司 Pronunciation inputting method and device, computer installation and readable storage medium storing program for executing
CN107679032A (en) * 2017-09-04 2018-02-09 百度在线网络技术(北京)有限公司 Voice changes error correction method and device
CN109215660A (en) * 2018-07-09 2019-01-15 维沃移动通信有限公司 Text error correction method and mobile terminal after speech recognition
CN109739425A (en) * 2018-04-19 2019-05-10 北京字节跳动网络技术有限公司 A kind of dummy keyboard, pronunciation inputting method, device and electronic equipment

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120004898A1 (en) * 2007-02-12 2012-01-05 Google Inc. Contextual Input Method
CN102141889A (en) * 2010-02-12 2011-08-03 微软公司 Typing assistance for editing
CN104508604A (en) * 2012-07-30 2015-04-08 微软公司 Generating string predictions using contexts
WO2014054922A1 (en) * 2012-10-05 2014-04-10 Samsung Electronics Co., Ltd. User terminal apparatus, method for inputting text in user terminal apparauts, and computer-readable recording medium
US20150100537A1 (en) * 2013-10-03 2015-04-09 Microsoft Corporation Emoji for Text Predictions
CN107346229A (en) * 2017-07-18 2017-11-14 珠海市魅族科技有限公司 Pronunciation inputting method and device, computer installation and readable storage medium storing program for executing
CN107679032A (en) * 2017-09-04 2018-02-09 百度在线网络技术(北京)有限公司 Voice changes error correction method and device
CN109739425A (en) * 2018-04-19 2019-05-10 北京字节跳动网络技术有限公司 A kind of dummy keyboard, pronunciation inputting method, device and electronic equipment
CN109215660A (en) * 2018-07-09 2019-01-15 维沃移动通信有限公司 Text error correction method and mobile terminal after speech recognition

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112951232A (en) * 2021-03-02 2021-06-11 深圳创维-Rgb电子有限公司 Voice input method, device, equipment and computer readable storage medium
CN113301416A (en) * 2021-04-30 2021-08-24 当趣网络科技(杭州)有限公司 Method for displaying voice frame

Similar Documents

Publication Publication Date Title
US9703462B2 (en) Display-independent recognition of graphical user interface control
US9098313B2 (en) Recording display-independent computerized guidance
CN205427822U (en) Electronic equipment and device that is used for editing text
KR102238809B1 (en) Actionable content displayed on a touch screen
CN102999274B (en) Semantic zoom animation
KR102005878B1 (en) Managing real-time handwriting recognition
US8060841B2 (en) Method and device for touchless media searching
US9405558B2 (en) Display-independent computerized guidance
KR20180112031A (en) Systems and methods for providing content selection
WO2016095689A1 (en) Recognition and searching method and system based on repeated touch-control operations on terminal interface
US9996522B2 (en) Dictionary device for determining a search method based on a type of a detected touch operation
JP2016524205A (en) Permanent synchronization system for handwriting input
JP2011081778A (en) Method and device for display-independent computerized guidance
CN104133815B (en) The method and system of input and search
CN112230811A (en) Input method, device, equipment and storage medium
US20110022956A1 (en) Chinese Character Input Device and Method Thereof
CN113901186A (en) Telephone recording marking method, device, equipment and storage medium
CN106970899B (en) Text processing method and device
US9361859B2 (en) Information processing device, method, and computer program product
KR20150097250A (en) Sketch retrieval system using tag information, user equipment, service equipment, service method and computer readable medium having computer program recorded therefor
KR102159331B1 (en) Sketch Retrieval system, user equipment, service equipment and service method based on meteorological phenomena information and computer readable medium having computer program recorded therefor
CN112764551A (en) Vocabulary display method and device and electronic equipment
KR102138095B1 (en) Voice command based virtual touch input apparatus
JP2002149679A (en) System and method for browsing document and information recording medium
CN113297519A (en) Comment data generation method and device, electronic equipment and computer storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210115