CN110502126B - Input method and electronic equipment - Google Patents

Input method and electronic equipment Download PDF

Info

Publication number
CN110502126B
CN110502126B CN201910661615.1A CN201910661615A CN110502126B CN 110502126 B CN110502126 B CN 110502126B CN 201910661615 A CN201910661615 A CN 201910661615A CN 110502126 B CN110502126 B CN 110502126B
Authority
CN
China
Prior art keywords
input
result
data
user
electronic device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910661615.1A
Other languages
Chinese (zh)
Other versions
CN110502126A (en
Inventor
张跃
吴海腾
李大伟
沈心忱
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Publication of CN110502126A publication Critical patent/CN110502126A/en
Application granted granted Critical
Publication of CN110502126B publication Critical patent/CN110502126B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • G06F3/0233Character input methods
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Abstract

The embodiment of the application discloses an input method and electronic equipment. Wherein the method comprises the following steps: receiving first data input by a user in a first input mode; determining a target result according to the first data, wherein the target result is a history selection result of a second data input by a user in a second input mode, syllable sequences corresponding to the first data and the second data are the same, and the target result is different from a universal result corresponding to the syllable sequences; the target result is displayed in a target input interface. The method can integrate the data input by the user in different input modes, so that the output result is more in line with the use habit of the user, and the input efficiency of the user is improved.

Description

Input method and electronic equipment
Technical Field
The present disclosure relates to the field of electronic devices, and in particular, to an input method and an electronic device.
Background
An input method application (input method application, IMA), which may also be referred to as an input method engine (input method engine, IME), or an input method editor (input method editor, IME), is software for implementing word input. At present, the input methods which are commonly used are applied as input methods, dog search input methods, hundred-degree input methods and the like. By installing the input method application in the terminal such as the mobile phone, the user can be assisted to realize the input of words in the terminal. The existing input method application summarization can generally provide multiple input modes for users, including a keyboard input mode, a voice input mode, an image scanning input mode and the like, and provides convenience for the users.
However, content input by a user in one input method may not be input accurately when the user inputs in another input method, resulting in low user input efficiency. For example, when the user inputs a personal name "chapter three" by using a keyboard input method, the user may recognize that the syllable sequence zhang 'san is "Zhang three" instead of "chapter three" desired by the user when the user inputs the syllable sequence zhang' san by using a voice input method. At this point the user needs to manually modify the "sheet" into "chapter" in the edit box. The process is complex in operation and poor in user experience.
Disclosure of Invention
The embodiment of the application provides an input method and electronic equipment, which can improve the input efficiency of a user.
In a first aspect, an embodiment of the present application provides an input method, including: receiving first data input by a user in a first input mode; determining a target result according to the first data, wherein the target result is a history selection result of the user inputting second data by adopting a second input mode, syllable sequences corresponding to the first data and the second data are the same, and the target result is different from a universal result corresponding to the syllable sequences; and displaying the target result in a target input interface.
According to the method and the device, the data input by the user in different input modes can be fused, so that the use habit of the user is not influenced by the different input modes, and under the condition that the use habit of the user is different from the use habit of the public, the output result can be more accordant with the use habit of the user no matter what input mode the user adopts, and the input efficiency of the user is improved.
In one possible implementation manner, the determining the target result according to the first data includes: searching a first language library according to the first data, and outputting a first result; the first language library is a language library corresponding to a plurality of input modes; searching a second language library according to the first data, and outputting a second result; wherein the second language library is a language library generated according to the second data; and determining a target result according to the first result and the second result.
According to the embodiment of the application, the data input by the users using the electronic equipment in different input modes can be fused, the data are shared in the different input modes, the processing capacity of the electronic equipment on the input data is enhanced, the output result is more in accordance with the use habit of the users compared with the prior art, and the input efficiency of the users is improved.
In one possible implementation manner, the first language library and the second language library each include a word, a syllable sequence corresponding to the word, and a word frequency corresponding to the word; the word with highest word frequency corresponding to the syllable sequence in the second language library is different from the word with highest word frequency corresponding to the syllable sequence in the first language library.
In the embodiment of the application, the second language library may record the use habit of the user, and the use habit of the user is different from the use habit of the public. Comprehensively considers the use habit of the user and the use habit of the public, and ensures that the output result is more accurate.
In one possible implementation, the syllable sequence corresponding to the word in the second language library is a part of the syllable sequence corresponding to the first data.
In one possible implementation, the syllable sequence corresponding to the first data is a part of the syllable sequence corresponding to the word in the second language library.
In one possible implementation, the first input mode is a voice input mode, and the second input mode is at least one of a keyboard input mode, an image input mode and a handwriting input mode.
According to the embodiment of the application, the data input by the user in the keyboard input mode, the image input mode and the handwriting input mode can be shared to the voice input mode, so that the accuracy of voice recognition can be improved, and the input efficiency of the user can be improved.
In another possible implementation, the first input mode is a keyboard input mode, and the second input mode is at least one of a voice input mode, an image input mode, and a handwriting input mode.
According to the embodiment of the application, the data input by the voice input mode, the image input mode and the handwriting input mode can be shared to the voice input mode, so that the accuracy of keyboard input can be improved, and the input efficiency of a user can be improved.
In another possible implementation, the second result is null and the target result is the first result.
Possibly, the second result being empty is that the second result is not present.
In the embodiment of the present application, when the second result does not exist, the output result is based on the first result. The first result is a result generated according to the use habits of a plurality of users, and accords with the use habits of most users.
In another possible implementation, the second result is not null and the target result is the second result.
In the embodiment of the application, in the case that the second result exists, the output result is based on the second result. The second result is a structure generated according to the use habit of the electronic equipment, the structure is more in line with the use habit of the user, the accuracy of the output result is ensured, the efficiency of user input is improved, and the user experience is improved.
In a second aspect, an embodiment of the present application provides an electronic device, including: one or more processors, memory, and wireless communication modules; the memory and the wireless communication module are coupled to one or more processors, the memory is configured to store computer program code, the computer program code includes computer instructions, and the electronic device performs, when the one or more processors execute the computer instructions: receiving data input by a user in a first input mode; determining a target result according to the first data, wherein the target result is a history selection result of the user inputting second data by adopting a second input mode, syllable sequences corresponding to the first data and the second data are the same, and the target result is different from a universal result corresponding to the syllable sequences; and displaying the target result in a target input interface.
In one possible implementation manner, the electronic device performs the specific performing when determining the target result according to the first data: searching a first language library according to the first data, and outputting a first result; the first language library is a language library corresponding to a plurality of input modes; searching a second language library according to the first data, and outputting a second result; the second language library is a language library generated according to the second data; and determining a target result according to the first result and the second result.
In one possible implementation manner, the first language library and the second language library each include a word, a syllable sequence corresponding to the word, and a word frequency corresponding to the word; the word with highest word frequency corresponding to the syllable sequence in the second language library is different from the word with highest word frequency corresponding to the syllable sequence in the first language library.
In one possible implementation, the syllable sequence corresponding to the word in the second language library is a part of the syllable sequence corresponding to the first data.
In one possible implementation, the syllable sequence corresponding to the first data is a part of the syllable sequence corresponding to the word in the second language library.
In one possible implementation, the first input mode is a voice input mode, and the second input mode is at least one of a keyboard input mode, an image input mode and a handwriting input mode.
In another possible implementation, the first input mode is a keyboard input mode, and the second input mode is at least one of a voice input mode, an image input mode, and a handwriting input mode.
In another possible implementation, the second result is null and the target result is the first result.
In another possible implementation, the second result is not null and the target result is the second result.
In a third aspect, an embodiment of the present application provides an electronic device, including: the device comprises an input module, a determining module and a display module. Wherein: the input module is used for receiving first data input by a user in a first input mode; the determining module is used for determining a target result according to the first data, wherein the target result is a history selection result of the user inputting second data by adopting a second input mode, syllable sequences corresponding to the first data and the second data are the same, and the target result is different from a universal result corresponding to the syllable sequences; the display module is used for displaying the target result in the target input interface.
In one possible implementation manner, the determining module includes a first searching unit, a second searching unit and a determining unit; the first searching unit is used for outputting a first result according to the first data received by the input module; the first language library is a language library corresponding to a plurality of input modes; the second searching unit is used for outputting a second result according to the first data received by the input module; the second language library is a language library generated according to the second data; the determining unit is used for determining a target result according to the first result and the second result.
In one possible implementation manner, the first language library and the second language library each include a word, a syllable sequence corresponding to the word, and a word frequency corresponding to the word; the word with highest word frequency corresponding to the syllable sequence in the second language library is different from the word with highest word frequency corresponding to the syllable sequence in the first language library.
In one possible implementation, the syllable sequence corresponding to the word in the second language library is a part of the syllable sequence corresponding to the first data.
In one possible implementation, the syllable sequence corresponding to the first data is a part of the syllable sequence corresponding to the word in the second language library.
In one possible implementation, the first input mode is a voice input mode, and the second input mode is at least one of a keyboard input mode, an image input mode and a handwriting input mode.
In another possible implementation, the first input mode is a keyboard input mode, and the second input mode is at least one of a voice input mode, an image input mode, and a handwriting input mode.
In another possible implementation, the second result is null and the target result is the first result.
In another possible implementation, the second result is not null and the target result is the second result.
In a fourth aspect, embodiments of the present application provide a computer storage medium comprising computer instructions which, when run on an electronic device, cause the electronic device to perform an input method as provided in the first aspect or any implementation of the first aspect.
In a fifth aspect, embodiments of the present application provide a computer program product for, when run on an electronic device, causing the electronic device to perform an input method as provided by the first aspect or any implementation of the first aspect.
It will be appreciated that the second aspect of the electronic device, the third aspect of the electronic device, the fourth aspect of the computer storage medium or the fifth aspect of the computer program product provided above are each adapted to perform the input method provided in the first aspect. Therefore, the advantages achieved by the method can be referred to as the advantages of the corresponding method, and will not be described herein.
Drawings
Fig. 1 is a schematic hardware structure of an electronic device according to an embodiment of the present application;
fig. 2A-2C are schematic diagrams of application scenarios according to embodiments of the present application;
fig. 3A to 3D are schematic views of a user interface of a keyboard input mode according to an embodiment of the present application;
FIG. 4 is a schematic diagram of voice input provided in an embodiment of the present application;
FIG. 5 is a schematic diagram of a voice input result provided in an embodiment of the present application;
FIGS. 6A-6G are schematic diagrams of a user interface of an image scan-in mode according to an embodiment of the present application;
fig. 7 is a schematic structural diagram of an electronic device according to another embodiment of the present application;
fig. 8 is a schematic flow chart of an input method according to an embodiment of the present application;
fig. 9 is a flowchart of a specific input method according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and thoroughly described below with reference to the accompanying drawings.
The embodiment of the application provides an input method, which can share data among all input modes of the input method and improve the recognition accuracy of voice input. That is, the text input by the keyboard or the text input by the image can be recognized when the text is input again by the voice, so that the recognition accuracy of the voice input is improved.
The electronic device involved in the embodiments of the present application may be a portable mobile terminal, such as a mobile phone, a tablet computer, a netbook, a personal digital assistant (Personal Digital Assistant, PDA), a wearable electronic device (such as a smart bracelet, a smart watch, etc.), a virtual reality device, etc.
The electronic device referred to in the embodiments of the present application may include an input method application. The input method application can be a system application or a third party application. One or more applications with input boxes may also be included in the electronic device. The application with input box may be a system application such as a short message, calendar, etc., or a third party application such as a Facebook (Facebook), amazon, map, music player, etc. The user may use the input method application to implement word input in an input box of the application having an input box.
Next, an exemplary electronic device provided in the following embodiments of the present application is described.
Fig. 1 shows a schematic configuration of an electronic device 100.
The electronic device 100 may include a processor 110, an internal memory 120, a usb interface 130, a charge management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a mobile communication module 150, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, keys 190, a camera 191, a display 192, and the like. The sensor module 180 may include a gyro sensor 180A, an acceleration sensor 180B, a distance sensor 180C, a fingerprint sensor 180D, a temperature sensor 180E, a touch sensor 180F, and the like.
It is to be understood that the structure illustrated in the embodiments of the present application does not constitute a specific limitation on the electronic device 100. In other embodiments of the present application, electronic device 100 may include more or fewer components than shown, or certain components may be combined, or certain components may be split, or different arrangements of components. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
The processor 110 may include one or more processing units, such as: the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a video codec, a digital signal processor (digital signal processor, DSP), a baseband processor, and/or a neural network processor (neural-network processing unit, NPU), etc. Wherein the different processing units may be separate devices or may be integrated in one or more processors.
The controller can generate operation control signals according to the instruction operation codes and the time sequence signals to finish the control of instruction fetching and instruction execution.
A memory may also be provided in the processor 110 for storing instructions and data. In some embodiments, the memory in the processor 110 is a cache memory. The memory may hold instructions or data that the processor 110 has just used or recycled. If the processor 110 needs to reuse the instruction or data, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 110 is reduced, thereby improving the efficiency of the system.
The charge management module 140 is configured to receive a charge input from a charger. The charging management module 140 may also supply power to the electronic device through the power management module 141 while charging the battery 142.
The power management module 141 is used for connecting the battery 142, and the charge management module 140 and the processor 110. The power management module 141 receives input from the battery 142 and/or the charge management module 140 and provides power to the processor 110, the internal memory 120, the display 192, the camera 191, the wireless communication module 160, and the like.
The wireless communication function of the electronic device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, a modem processor, a baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the electronic device 100 may be used to cover a single or multiple communication bands. Different antennas may also be multiplexed to improve the utilization of the antennas.
The mobile communication module 150 may provide a solution for wireless communication including 2G/3G/4G/5G, etc., applied to the electronic device 100. The mobile communication module 150 may include at least one filter, switch, power amplifier, low noise amplifier (low noise amplifier, LNA), etc. The mobile communication module 150 may receive electromagnetic waves from the antenna 1, perform processes such as filtering, amplifying, and the like on the received electromagnetic waves, and transmit the processed electromagnetic waves to the modem processor for demodulation. The mobile communication module 150 can amplify the signal modulated by the modem processor, and convert the signal into electromagnetic waves through the antenna 1 to radiate.
The wireless communication module 160 may provide solutions for wireless communication including wireless local area network (wireless local area networks, WLAN) (e.g., wireless fidelity (wireless fidelity, wi-Fi) network), bluetooth (BT), global navigation satellite system (global navigation satellite system, GNSS), frequency modulation (frequency modulation, FM), near field wireless communication technology (near field communication, NFC), infrared technology (IR), etc., as applied to the electronic device 100.
The electronic device 100 implements display functions through a GPU, a display screen 192, and an application processor, etc. The GPU is a microprocessor for image processing, and is connected to the display 192 and the application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. Processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
The display 192 is used to display images, videos, and the like. The display 192 includes a display panel. The display panel may employ a liquid crystal display (liquid crystal display, LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (AMOLED) or an active-matrix organic light-emitting diode (matrix organic light emitting diode), a flexible light-emitting diode (flex), a mini, a Micro led, a Micro-OLED, a quantum dot light-emitting diode (quantum dot light emitting diodes, QLED), or the like. In this embodiment of the present application, the display screen 192 may be used to display the result corresponding to the data input by the user using various input modes.
The camera 191 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image onto the photosensitive element. The photosensitive element may be a charge coupled device (charge coupled device, CCD) or a Complementary Metal Oxide Semiconductor (CMOS) phototransistor. The photosensitive element converts the optical signal into an electrical signal, which is then transferred to the ISP to be converted into a digital image signal. The ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into an image signal in a standard RGB, YUV, or the like format. The camera 191 in the embodiment of the present application may be used to capture a picture including text information input by a user in an image scanning input manner, so that the electronic device identifies the text information in the picture, and displays the text information on the display 192.
The NPU is a neural-network (NN) computing processor, and can rapidly process input information by referencing a biological neural network structure, for example, referencing a transmission mode between human brain neurons, and can also continuously perform self-learning. Applications such as intelligent awareness of the electronic device 100 may be implemented through the NPU, for example: image recognition, face recognition, speech recognition, text understanding, etc.
The internal memory 120 may be used to store computer executable program code including instructions. The internal memory 120 may include a storage program area and a storage data area. The storage program area may store an application program (such as a sound playing function, an image playing function, etc.) required for at least one function of the operating system, etc. The storage data area may store data created during use of the electronic device 100 (e.g., audio data, phonebook, etc.), and so on. In addition, the internal memory 120 may include a high-speed random access memory, and may also include a nonvolatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (universal flash storage, UFS), and the like. The processor 110 performs various functional applications and data processing of the electronic device 100 by executing instructions stored in the internal memory 120 and/or instructions stored in a memory provided in the processor. In this embodiment, the internal memory 120 may be configured to store a first language library and a second language library, so that the electronic device 100 searches for the first language library and the second language library according to data input by a user, to obtain a target result, and displays the target result through the display 192.
The electronic device 100 may implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, an application processor, and the like. Such as music playing, recording, etc.
The audio module 170 is used to convert digital audio information into an analog audio signal output and also to convert an analog audio input into a digital audio signal. The audio module 170 may also be used to encode and decode audio signals.
The speaker 170A, also referred to as a "horn," is used to convert audio electrical signals into sound signals. The electronic device 100 may listen to music, or to hands-free conversations, through the speaker 170A.
A receiver 170B, also referred to as a "earpiece", is used to convert the audio electrical signal into a sound signal. When electronic device 100 is answering a telephone call or voice message, voice may be received by placing receiver 170B in close proximity to the human ear.
Microphone 170C, also referred to as a "microphone" or "microphone", is used to convert sound signals into electrical signals. When making a call or transmitting voice information, the user can sound near the microphone 170C through the mouth, inputting a sound signal to the microphone 170C. In an embodiment of the present application, the microphone 170C may be used to collect audio signals input by a user, so that the electronic device inputs the audio signals into the acoustic model.
The earphone interface 170D is used to connect a wired earphone.
The gyro sensor 180A may be used to determine a motion gesture of the electronic device 100. In some embodiments, the angular velocity of electronic device 100 about three axes (i.e., x, y, and z axes) may be determined by gyro sensor 180A.
The acceleration sensor 180B may detect the magnitude of acceleration of the electronic device 100 in various directions (typically three axes).
A distance sensor 180C for measuring a distance. The electronic device 100 may measure the distance by infrared or laser.
The fingerprint sensor 180D is used to collect a fingerprint. The electronic device 100 may utilize the collected fingerprint feature to unlock the fingerprint, access the application lock, photograph the fingerprint, answer the incoming call, etc.
The temperature sensor 180E is used to detect temperature.
The touch sensor 180F, also referred to as a "touch device". The touch sensor 180F may be disposed on the display 192, and the touch sensor 180F and the display 192 form a touch screen, which is also referred to as a "touch screen". The touch sensor 180F is used to detect a touch operation acting thereon or thereabout. The touch sensor may communicate the detected touch operation to the application processor to determine the touch event type. Visual output related to touch operations may be provided through the display 192. In other embodiments, the touch sensor 180F may also be disposed on a surface of the electronic device 100 at a different location than the display 192. In this embodiment of the present application, the touch sensor 180F may be configured to detect data input by a user in a keyboard input manner or a handwriting input manner, so that the electronic device outputs a result corresponding to the data, and the result is displayed on the display screen 192.
Application scenarios related to embodiments of the present application are described next in connection with fig. 2A-2C.
Fig. 2A illustrates a keyboard input mode interface schematic diagram of an input method.
Fig. 2A illustrates a contact chat interface 30 for WeChat. Wherein, the WeChat is instant messaging software. As shown in fig. 2A, the user interface 30 may include: status bar 301, display area 302, input box 303, navigation bar 304. Wherein:
status bar 301 may include: carrier indicator (e.g., carrier name "china mobile"), one or more signal strength indicators of wireless high-fidelity (wireless fidelity, wi-Fi) signals, one or more signal strength indicators of mobile communication signals (also may be referred to as cellular signals), a time indicator, and a battery status indicator.
Display area 302 may be used to display historical chat content with the contact (Emmy).
The user may perform a click operation on the input box 303. In response to the click operation, the processor 110 of the electronic device 100 may invoke the input method application in the electronic device 110 and display the interface 305 of the input method application shown in FIG. 2B on the display 192. The interface is an input interface having an alphabetic combination key of 26 keys. The user may enter a string at the interface 305 of the input method application. The processor 110 of the electronic device 100 may search for one or more words corresponding to a character string according to the character string input by the user, and display the one or more words in the interface 305 of the input method application through the display 192. The user may select a word that he or she expects to input from one or more words corresponding to the string.
It will be appreciated that in fig. 2B, the interface 305 of the input method application is illustrated as an input interface with 26-key alphabetic combination keys, and the interface of the input method application may be an input interface with 9-key numeric combination keys, where a user may input a character string.
Navigation bar 304 may include: a back button 3041, a Home screen button 3042, an outgoing task history button 3043, and the like. The main interface is an interface displayed by the electronic device 100 after any one of the user interfaces detects a user operation acting on the main interface button 3042. When it is detected that the user clicks the back button 3041, the electronic device 100 may display a previous user interface of the current user interface. When it is detected that the user clicks the main interface button 3042, the electronic device 100 may display the main interface. When detecting that the user clicks the outgoing task history button 3043, the electronic device 100 may display a task that the user has recently opened. The names of the navigation keys may be other, for example, 3041 may be called Back Button,3042 may be called Home Button,3043 may be called Menu Button, which is not limited in this application. The navigation keys in the navigation bar 304 are not limited to virtual keys, but may be implemented as physical keys.
The interface 305 of the input method application may include an input mode switching control 3051.
An input mode switching control 3051 can be used to switch input modes. Specifically, the electronic device 100 may detect a touch operation (e.g., a click operation on the input mode switching control 3051) applied to the input mode switching control 3051, and in response to the touch operation, the electronic device 100 may display icons of other input modes in the interface 305 of the input method application, such as the icon 3052 of the text scan input mode, the icon 3053 of the speech input mode, and the icon 3054 of the handwriting input mode shown in fig. 2C. Wherein:
icon 3052 of text scan-in mode can be used to input words by scanning an image. Specifically, the electronic device 100 may detect a touch operation (such as a clicking operation on the icon 3052 in the text scan input mode) applied to the icon 3052 in the image scan input mode, and in response to the touch operation, the electronic device 100 starts the camera 191 to scan a picture, sends the scanned picture content to the processor 110, recognizes the text or the characters (letters, numbers, symbols, etc.) in the picture through the processor 110, and displays the recognized result in the interface 305 of the input method application through the display 192.
Icon 3053 of the speech input mode is used to input words by speech. Specifically, the electronic device 100 may detect a touch operation (such as a clicking operation on the icon 3053 in the voice input mode) applied to the icon 3053 in the voice input mode, and in response to the touch operation, the electronic device 100 may turn on the microphone 170C to collect a sound signal, send the collected sound signal to the processor 110, recognize a text or a character (letters, numbers, etc.) corresponding to the sound signal through the processor 110, and display the recognized result on the interface 305 of the input method application through the display 192.
Icon 3054 of handwriting input mode may be used to input words by handwriting of the user. Specifically, the electronic device 100 may detect a situation operation (such as a clicking operation on the icon 3054 in the handwriting input mode) acting on the icon 3054 in the handwriting input mode, and in response to the operation, the electronic device 100 may display a handwriting detection area in the interface 305 of the input method application, detect a touch track input by the user in the handwriting detection area, determine a word input by the user according to the touch track, and display the word in the interface 305 of the input method application through the display screen 192.
Next, the input method provided in several embodiments of the present application is described in conjunction with the foregoing application scenario.
Embodiment one: keyboard input-voice input, namely, data input by adopting a keyboard input mode is shared to a voice input mode.
Next, the input method provided by the first embodiment will be described in four sections.
A first part: and inputting data by adopting a keyboard input mode, and outputting a corresponding result.
As shown in fig. 3A, the electronic device 100 may detect a click operation of inputting a character string zhangsan in the interface 305 of the input method application, one or more letter keys (the one or more letter keys may be, for example, zhangsan in sequence), and in response to the click operation, the electronic device 100 may display an input result 401 (zhang' san) in the interface 305 of the input method application, and display a list 402 of to-be-selected items and a drop-down control 403 corresponding to the input result. The waiting option list 402 may include one or more waiting options 4021.
Wherein the drop down control 403 may be used to display more options to be selected. The electronic device 100 may detect a touch operation (e.g., a click operation on the drop-down control 403) on the drop-down control 403, and in response to the operation, the electronic device 100 may update the interface 305 displaying the input method application shown in fig. 3B.
As shown in fig. 3B, the interface 305 of the input method application may include a to-be-selected list 402, an up-turn control 404, a down-turn control 405, and a return control 406.
Both the up-scroll control 404 and the down-scroll control 405 may be used to update the to-be-selected items in the list 402. The electronic device 100 may detect a touch operation (e.g., a click operation on the scroll down control 405) on the scroll down control 405, and in response to the operation, the electronic device 100 may update the display of the to-be-selected items in the to-be-selected item list 402. The same applies to the flip-up control 404 and is not repeated here.
The return control 406 may be used to cause the electronic device 100 to display the user interface 30 shown in fig. 3A.
The to-be-selected item may be a to-be-selected item corresponding to the complete input result "zhang' san", or may be a to-be-selected item corresponding to the first half of the input result "zhang".
The user may click on the to-be-selected item in the to-be-selected item list 402 of fig. 3B to select the desired result. For example, the user may click on "chapter" in the to-be-selected list 402 to select the word corresponding to the input result "zhang". After selecting the word corresponding to the first half of the input result, the electronic device 100 may display the interface 305 of the input method application shown in fig. 3C.
The input method application interface 305 shown in fig. 3C is similar to the input method application interface 305 shown in fig. 3B. The difference is that the to-be-selected item corresponding to the complete input result "zhang' san" or the to-be-selected item corresponding to the first half "zhang" of the input result is displayed in the to-be-selected item list 402 in the interface 305 of the input method application in fig. 3B. And the to-be-selected item list 402 in the interface 305 of the input method application in fig. 3C shows to-be-selected items corresponding to the second half "san" of the input result.
The user may click on the to-be-selected item in the to-be-selected item list 402 of fig. 3C to select the desired result. For example, the user may click on "three" in the to-be-selected list 402 to select the word corresponding to the input result "san".
After the electronic device 100 determines the word corresponding to the complete input result "zhang' san", as shown in fig. 3D, the word "chapter three" corresponding to the word may be displayed in the input box 303.
The keyboard input method in the embodiment of the present application may be a pinyin input method, or may be an input method such as a five-stroke input method that is input through a keyboard. The interface 305 for the application of the input method may include an input interface of a 26-key letter combination key, an input interface of a 9-key number combination key, and the like.
Next, a procedure of determining a result corresponding to the character string by the electronic device 100 will be described.
After the electronic device 100 receives a character string input by a user, the character string may be combined into a plurality of syllables or syllable sequences, and find words corresponding to the plurality of syllables or syllable sequences.
First, the electronic device 100 may determine a plurality of syllables or syllable sequences consisting of the character string and determine the likelihood of each syllable or syllable sequence, respectively. The electronic device 100 may preferentially display the word corresponding to the syllable or syllable sequence with the highest likelihood in the list of alternatives 402.
The likelihood of each syllable or syllable sequence may be determined by collecting the usage habits of a large number of users, and may further be adjusted according to the usage habits of the users using the electronic device 100.
For example, the character string entered by the user may be a xin. The syllable or syllable sequence corresponding to the character string can be xi' an or xi. If the syllable or syllable sequence corresponding to the character string is selected to be xian for 600 times in 1000 times of the selections of the plurality of users, and the syllable sequence corresponding to the character string is selected to be xian for 400 times. The syllable or syllable sequence corresponding to the character string is 60% more likely than xian and the syllable sequence corresponding to the character string is 40% more likely than xian. The electronic device 100 may determine, according to the word finally selected by the user, the syllable or syllable sequence corresponding to the character string selected by the user. After determining the likelihood of various syllables or syllable sequences corresponding to the character string, the electronic device 100 may preferentially display the word corresponding to the syllable or syllable sequence with the highest likelihood in the to-be-selected list 402.
Further, if the syllable sequence corresponding to the character string is xi' an for 6 choices and the syllable sequence corresponding to the character string is xian for 4 choices among 10 choices of the user using the electronic device 100. The electronic device 100 may adjust according to the usage habit of the user using the electronic device 100 to determine that the probability that the syllable sequence corresponding to the character string is xi' an is 60% and the probability that the syllable corresponding to the character string is xian is 40%. The electronic device 100 can preferentially display the words corresponding to the syllable sequence xi' an in the list of to-be-selected items 402.
The word corresponding to the syllable or syllable sequence having the highest possibility is not limited to be preferentially displayed in the list of to-be-selected items 402, and the word corresponding to the syllable or syllable sequence most recently selected by the user may be preferentially displayed in the list of to-be-selected items 402. For example, if the syllable sequence corresponding to the character string xin selected by the user using the electronic device 100 is xi 'an, the electronic device 100 may preferentially display the word corresponding to the syllable sequence xi' an in the to-be-selected list 402 when the character string xin is input again by the user.
Second, the electronic device 100 can determine one or more words corresponding to syllables and determine the likelihood of each word separately. The electronic device 100 may preferentially display the most likely word in the list of alternatives 402.
It will be appreciated that there may be a variety of words to which syllables or syllable sequences correspond. For example, words corresponding to syllable sequence xi' an may be "xian", etc., and words corresponding to syllable xian may be "first", "now", "line", "fresh", etc.
The electronic device 100 may collect a plurality of user usage habits to obtain the possibility of various words corresponding to the syllable, and further may also adjust according to the usage habits of the user using the electronic device 100.
For example, for syllable xian, if the word corresponding to the syllable is selected "first" 400 times, the word corresponding to the syllable is selected "now" 300 times, the word corresponding to the syllable is selected "line" 200 times, and the word corresponding to the syllable is selected "fresh" 100 times among 1000 times of the multiple users. The likelihood that the word corresponding to the syllable is "first" is 40%, the likelihood that the word corresponding to the syllable is "present" is 30%, the likelihood that the word corresponding to the syllable is "line" is 20%, and the likelihood that the word corresponding to the syllable is "fresh" is 10%. After determining the likelihood of the various strings corresponding to the syllable, the electronic device 100 can preferentially display the most probable word in the list of to-be-selected items 402.
Further, if, among 10 selections by the user using the electronic device 100, 5 selections of the word corresponding to the syllable are "fresh", 3 selections of the word corresponding to the syllable are "line", one selection of the word corresponding to the syllable is "current", and one selection of the word corresponding to the syllable is "first". The electronic device 100 may adjust according to the usage habit of the user using the electronic device 100 to determine that the likelihood that the word corresponding to the syllable is "fresh" is 50%, the likelihood that the word corresponding to the syllable is "line" is 30%, the likelihood that the word corresponding to the syllable is "current" is 10%, and the likelihood that the word corresponding to the syllable is "first" is 10%. The electronic device 100 may preferentially display the word "fresh" in the list of to-be-selected 402.
The word with the highest possibility is not limited to be preferentially displayed in the list of to-be-selected items 402, and a word corresponding to the syllable recently selected by the user may be preferentially displayed in the list of to-be-selected items 402. For example, if the word corresponding to the syllable xian was selected as "Xian" last time by the user using the electronic device 100, the electronic device 100 may preferentially display the word "Xian" in the list of to-be-selected items 402 when the syllable xian is input again by the user.
The above-mentioned manner of preferentially displaying words in the list of to-be-selected items 402 may be to display the words in the leftmost column and the uppermost row of the list of to-be-selected items 402.
A second part: a personalized language library is generated.
Specifically, the electronic device 100 may store the character string input by the user using the keyboard and the word corresponding to the character string finally selected by the user in the personalized language library. Wherein the personalized language library may be a certain storage area in the internal memory 120. The electronic device 100 may determine the syllable or syllable sequence corresponding to the character string and the word corresponding to the syllable or syllable sequence according to the personalized language library and record the selected number of times (i.e., word frequency) of the word. Thus, the finally generated personalized language library may conform to the usage habits of the user using the electronic device 100.
For example, if the user inputs the character string zhangsan using the keyboard, and the user finally selects the word corresponding to the character string zhangsan to be "chapter three", the electronic device 100 may determine that the syllable sequence corresponding to the character string zhangsan is zhang 'san, and the word corresponding to the syllable sequence zhang' san is "chapter three", instead of "Zhang three".
Third section: and inputting data by adopting a voice input mode.
After the user clicks icon 3053 in fig. 2B, electronic device 100 may display application interface 305 of the input method shown in fig. 4, and turn on microphone 170C (the position of microphone 170C is exemplarily shown in fig. 4) to start collecting sound signals. At this time, the user may input an audio signal.
Specifically, the electronic device 100 may extract audio features of the audio signal input by the user, input the audio features into the acoustic model, and output syllables or syllable sequences. The fourth section below will be exemplarily described below using the output syllable sequence zhang' san as an example.
Fourth part: searching a universal language library and a personalized language library, and outputting a final result.
After determining that the syllable sequence output by the acoustic model is zhang' san, the electronic device 100 searches for a result corresponding to the syllable sequence in the universal language library and the personalized language library, determines a final result according to the result of searching for the universal language library and the result of searching for the personalized language library, and displays the final result in the input box 303 shown in fig. 5.
Specifically, the universal language library is a preset language library, and the language library can be obtained according to data input by a large number of users in various input modes. Thus, the universal language library may conform to the usage habits of most users. The general language library may be a language library of the input method when the user downloads the input method, or may be a language library of the electronic device 100.
Specifically, the universal language library may include, but is not limited to, a plurality of words, syllable sequences corresponding to each word, and word frequencies corresponding to the words. In addition, the generic language library may also be used to calculate the score of the candidate word in the sentence structure, such that the electronic device 100 preferentially displays the highest-scoring candidate word in the list of alternatives 402. Possibly, the same syllable sequence may correspond to a plurality of different words, and the word frequencies corresponding to the words may be different.
After determining the syllable or syllable sequence corresponding to the audio signal input by the user, the electronic device 100 may search the universal language library to determine the result with the highest likelihood of corresponding to the syllable or syllable sequence. The result with the highest possibility of searching the universal language library is called a first result.
The electronic device 100 may further search the personalized language library after determining the syllable or the syllable sequence corresponding to the audio signal input by the user, and determine the result with the highest likelihood corresponding to the syllable or the syllable sequence. The result with the highest possibility of searching the personalized language library is called a second result.
The most probable result in the personalized language library may be the most frequently selected result by the user, or may be the most recently selected result by the user, which is not limited in the embodiment of the present application.
The results of looking up the generic language library and the results of looking up the personalized language library may be in the following cases:
the first result and the second result are identical. The final output result of the voice input is the first result and the second result.
The second case, the first result and the second result are inconsistent. The final result of voice input is the second result. That is, in the case where the first result and the second result are inconsistent, the second result is prioritized.
In case three, the first result exists, the second result does not exist or the second result is empty. The final result of voice input is the first result.
In addition, the embodiment of the application does not limit the sequence of searching the universal language library and searching the personalized language library, and the time relationship between searching the universal language library and searching the personalized language library can be performed simultaneously.
For example, after the electronic device 100 determines that the syllable sequence corresponding to the audio signal input by the user is zhang' san, the result with the highest possibility of searching for the syllable in the universal language library is "Zhang three", and the result with the highest possibility of searching for the syllable in the personalized language library is "Zhang three". I.e., the first result is "Zhang San", the second result is "Zhang San", and the final result corresponding to the audio signal input by the user is "Zhang San", respectively.
By implementing the embodiment of the application, the data input by the user in the keyboard input mode can be shared to the voice input mode, and the input result can be determined according to the historical data input by the user in the keyboard input mode and the use habit of the public at the same time when the voice input mode is adopted. The processing capability of the electronic equipment on input data is enhanced, the accuracy of voice recognition is improved, and the input efficiency of a user is improved.
Embodiment two: image scan input-voice input, namely, data input by adopting an image scan input mode is shared to a voice input mode.
Next, the input method provided by the second embodiment will be described in four sections.
A first part: and inputting data by adopting an image scanning input mode, and outputting a corresponding result.
As shown in fig. 2C, the electronic device 100 may detect a touch operation (e.g., a click operation on the icon 3052 in the text scan input mode) applied to the icon 3052 in the image scan input mode, and in response to the touch operation, the electronic device 100 turns on the camera 191 to scan a picture and display a scan interface.
Illustratively, as shown in FIG. 6A, the scan interface 70 may include: a scan content preview area 701, a scan requirements prompt 702, a scan control 703, a return control 704, and a reference line 705. Wherein:
The scanned content preview area 701 may be used to display images captured by the camera 191.
The scan requirements prompt box 702 may be used to display scan requirements, prompting the user to scan as desired.
The scan control 703 can be used to determine the scan content. The electronic device 100 may detect a touch operation (e.g., a clicking operation on the scan control 703) applied to the scan control 703, and in response to the touch operation, the electronic device 100 may determine that an image captured by the camera 191 at the current moment is the scan content, and extract an area with text in the scan content. The area having text will be simply referred to as a text area in the following description.
The return control 704 may be used to cause the electronic device 100 to display the user interface 30 shown in fig. 2C.
The reference line 705 may be used to prompt a user to orient text in the scanned image parallel to the reference line 705 to facilitate recognition of text in the scanned image by the electronic device 100.
The electronic device 100 may also display the user interface 80 shown in fig. 6B in response to a touch operation on the scan control 703.
Illustratively, as shown in FIG. 6B, the user interface 80 may include: text field 801A, text field 801B, determination control 802, and return control 803.
The text area 801A and the text area 801B are areas with text extracted from the scanned content by the electronic device 100. Both text field 801A and text field 801B may be used to determine the text field to be selected. The electronic device 100 may detect a touch operation (e.g., a click operation on the text region 801A and/or the text region 801B) acting on the text region 801A and/or the text region 801B, and in response to the operation, the electronic device 100 may specifically display the text region 801A or the text region 801B to prompt the user that the text region 801A and/or the text region 801B is a text region to be selected. The special display mode may be, for example, but not limited to, thickening and displaying a border of the text area to be selected. Reference may be made specifically to text region 801B shown in fig. 6C.
The determination control 802 can be used to determine a target text region. The electronic device 100 may detect a touch operation (e.g., a click operation on the determination control 802) for the determination control 802, in response to which the electronic device 100 may determine the text region to be selected as the target text region and display the user interface 90 shown in fig. 6D.
The return control 803 may be used to cause the electronic device 100 to display the user interface 70 illustrated in fig. 6B.
As shown in fig. 6D, the user interface 90 may include: a target text area display box 901, a scan result display area 902, a send control 903, a copy control 904, and a return control 905. Wherein:
the target text area 901 may be used to display the target text area.
The scan result display area 902 may be used to display content extracted from the target text area by the electronic device 100, i.e., a scan result.
The send control 903 may be used to cause the scan result to be displayed in the input box 305. The electronic device 100 may detect a touch operation (e.g., a click operation on the send control 903) on the send control 903, and in response to this operation, the electronic device 100 may display the user interface 30 illustrated in fig. 6E, and the scan result may be displayed in the input box 303 of the user interface 30.
Copy control 904 can be used to copy scan results.
The return control 905 may be used to cause the electronic device 100 to display the user interface 80 shown in fig. 6C.
A second part: a personalized language library is generated.
Specifically, the electronic device 100 may store the text input by the user in the image scan input manner in the personalized language library. In addition, the electronic device 100 may store syllables or syllable sequences corresponding to characters input by the user through the image scan input method in the personalized language library. Thus, the finally generated personalized language library may conform to the usage habits of the user using the electronic device 100. Wherein the personalized language library may be a certain storage area in the internal memory 120. The results stored in the personalized language library may include a word, a syllable sequence corresponding to the word, a number of times the word is selected (i.e., word frequency), and so forth. Possibly, the same syllable sequence may correspond to a plurality of different words, and the word frequencies corresponding to the words may be different.
For example, the result input by the user in the image scanning manner is "the financial port B17 of the mountain and new area of the eastern lake of the martial arts, the hubei province, the result that the electronic device 100 may store in the personalized language library may be as follows:
table 1 results list in personalized language library
Word and word Syllable sequence Word frequency
Hubei province hu’bei’sheng 8
Wuhan city wu’han’shi 10
East lake Gaoxin district dong’hu’gao’xin’qu 13
Financial port jin’rong’gang 6
B17 span bi’shi’qi’dong 8
1006 room yi’ling’ling’liu’shi 9
Not limited to the words listed in table 1, the words stored in the personalized language library in the specific implementation may be other, and the embodiment of the present application is not limited thereto.
It is to be understood that the word frequencies corresponding to the words listed in table 1 are all exemplified, and are actually recorded according to the history data input by the user, and the above listed word frequencies do not limit the embodiments of the present application.
Third section: and inputting data by adopting a voice input mode.
Specifically, the third part in this embodiment is similar to the third part in the first embodiment.
Assume that the electronic device 100 extracts an audio feature of an audio signal input by a user, inputs the audio feature into an acoustic model, and outputs a syllable sequence of hu ' bei ' share ' wu ' han ' dong ' hu ' gao ' xin ' qu ' jin ' rong ' gang ' bi ' shi ' qi ' dong ' yi ' ling ' ling ' liu ' shi.
In particular implementations, the syllable sequence corresponding to the input audio signal may be a part of the syllable sequence corresponding to the result input by the image scanning method, for example, as shown in fig. 6F, the syllable sequence output according to the audio signal input by the user may be dong 'hu' gao 'xin' qu.
In contrast, the syllable sequence corresponding to the result input by the image scanning input method may be a part of the syllable sequence corresponding to the input audio signal, and for example, as shown in fig. 6G, the syllable sequence output according to the audio signal input by the user may be wo ' yao ' qu ' dong ' hu ' gao ' xin ' qu.
Fourth part: searching a universal language library and a personalized language library, and outputting a final result.
Specifically, the fourth part in this embodiment is similar to the fourth part in the first embodiment, and is not described here.
By implementing the embodiment of the application, the data input by the user in the image scanning mode can be shared to the voice input mode, and the input result can be determined according to the historical data input by the user in the image scanning mode and the use habit of the public at the same time when the voice input mode is adopted. The accuracy of voice recognition is improved, and the input efficiency of a user is improved.
Embodiment III: voice input-keyboard input, i.e. data input by adopting a voice input mode is shared to a keyboard input mode.
Next, the input method provided by the third embodiment will be described in four sections.
A first part: and inputting data by adopting a voice input mode, and outputting a corresponding result.
Specifically, the process of inputting data by the user in the voice input manner may refer to fig. 4, which is not described herein.
Assume that the electronic device 100 extracts an audio feature of an audio signal input by a user using a voice input method, inputs the audio feature into an acoustic model, and outputs a syllable sequence of wu ' he ' da ' dao. The electronic device 100 may display the highest likelihood that the syllable sequence corresponds as "wu-he-dao" in the input box 303. If the result does not meet the user's expectations, the user may edit the result again in the input box by using the keyboard input mode, for example, the "river" may be changed to "sum". The electronic device 100 may determine that the modified result "five and big meatus" is a word corresponding to the syllable sequence wu ' he ' da ' dao.
The method is not limited to changing partial results, for example, the "river" is changed into "sum", and in a specific implementation, the user can also edit all the results again by adopting a keyboard input mode in the input box, for example, edit the voice recognition result "five river channels" again by adopting a keyboard input mode after deleting all the results. The electronic device 100 can confirm that the result after the re-editing is a word corresponding to the syllable sequence wu ' he ' da ' dao. The embodiments of the present application are not limited in this regard.
A second part: a personalized language library is generated.
The electronic device 100 may store a syllable sequence (e.g., wu ' he ' da ' dao) corresponding to an audio signal input by a user using a voice input manner and a word (e.g., "five and big channels") corresponding to the syllable sequence that is finally determined in a personalized language library, and record the selected number of times (i.e., word frequency) of the word. Wherein the personalized language library may be a certain storage area in the internal memory 120. Thus, the finally generated personalized language library may conform to the usage habits of the user using the electronic device 100.
Third section: and inputting data by adopting a keyboard input mode.
Specifically, the process of inputting data by the user in the keyboard input manner may refer to fig. 3A, which is not described herein. The fourth section will be described by way of example using wuhendadao as the character string input by the user using the keyboard input method.
Fourth part: the general language library and the personalized language library are searched, the final result is determined, and the final result is preferentially displayed in the list to be selected 402.
Specifically, the process of determining the final result in this embodiment is similar to that in the fourth section of the first embodiment, and is not repeated here.
For example, the electronic device 100 may determine that the syllable sequence corresponding to the string wuhedadadao is wu ' he ' da ' dao. The electronic device 100 may look up the universal language library to determine that the syllable sequence corresponds to a "wu-he-dao", i.e. the first result is a "wu-he-dao". The electronic device 100 may look up the personalized language library and determine that the syllable sequence corresponds to a "five and big lane", i.e., the second result is a "five and big lane". The electronic device 100 may preferentially display the second result "five and major" in the to-be-selected list 402, and then display the first result "five major".
By implementing the embodiment of the application, the data input by the user in the voice input mode can be shared to the keyboard input mode, the options conforming to the voice input habit of the user are preferentially displayed in the to-be-selected item list in the keyboard input mode, the time for the user to search the target options in the to-be-selected item list is reduced, and the input efficiency of the user is improved.
The method is not limited to the above-listed embodiments, in a specific implementation, data input by an image scanning input mode may be shared to a keyboard input mode, data input by a handwriting input mode may be shared to a keyboard input mode or a voice input mode, and a specific implementation process may be similar to the above-mentioned three embodiments and will not be repeated here.
Not limited to the user interface shown in fig. 4, in a specific implementation, the user may input audio data through a voice assistant, and the user interface for inputting audio data is not limited in the embodiments of the present application. Specifically, the electronic device 100 may display any user interface when audio data is input through the voice assistant, or the electronic device 100 may be in a lock state or a quench state. The electronic device may display the results of the speech recognition in an interface of the speech assistant. Where the voice assistant may be a function that the electronic device 100 has, it is often necessary to wake up the function with a specific keyword, which may be, for example, but not limited to, "small art.
An electronic device provided in another embodiment of the present application is described next.
As shown in fig. 7, the electronic device 100 may include an input module 810, a universal language library 820, a personalized language library 830, a determination module 840, and a display module 850. Wherein:
the input module 810 may be used to receive data entered by a user.
Specifically, when the user inputs data using a voice input method, the input module 810 may be the microphone 170C. The input module 810 may be the touch sensor 180F when the user inputs data using a keyboard input method or a handwriting input method. When the user inputs data using an image scan input method, the input module 810 may be a camera 191.
The generic language library 820 may output a first result according to the data received by the input module 810.
Specifically, the universal language library 820 may be a preset language library, which may be obtained according to data input by a large number of users in various input manners. The generic language library may be referred to as a first language library in embodiments of the present application.
Specifically, the universal language library may include, but is not limited to, a plurality of words, syllable sequences corresponding to each word, and word frequencies corresponding to the words. In addition, the generic language library may also be used to calculate the score of the candidate word in the sentence structure, such that the electronic device 100 preferentially displays the highest-scoring candidate word in the list of alternatives 402. Possibly, the same syllable sequence may correspond to a plurality of different words, and the word frequencies corresponding to the words may be different.
The personalized language library 830 may output a second result based on the data received by the input module 810. The personalized language library may be referred to as a second language library in embodiments of the present application.
Specifically, when the user inputs data by using a keyboard input method, the personalized language library 830 may record data input by the user by other input methods (such as an image scan input method, a voice input method, and a handwriting input method) other than the keyboard input method.
When the user inputs data in a voice input manner, the personalized language library 830 may record data input by the user in other input manners (such as an image scan input manner, a keyboard input manner, and a handwriting input manner) other than the voice input manner.
The generic language library 820 and the personalized language library 830 may be a certain storage area in the internal memory 120.
The determining module 840 may be configured to receive a first result output by the universal language library 820 and a second result output by the personalized language library 830, and determine a final result according to the first result and the second result. The final result may be referred to as a target result in embodiments of the present application.
A display module 850 for displaying the final result in the target input interface. In particular, the display module 850 may be the display 192. Wherein the target input interface may be a user interface of an application having an input box. The applications with input boxes may be system applications such as short messages, calendars, etc., as well as third party applications such as facebooks (facebooks), amazon, maps, music players, etc. The target input interface may also be the interface of a voice assistant.
The electronic equipment provided by the embodiment of the application can fuse the user data of different input modes, share the data in the different input modes, and improve the input efficiency of the user.
Next, an input method provided in the embodiment of the present application is described.
As shown in fig. 8, the input method provided in the embodiment of the present application may at least include the following steps:
s901: and receiving first data input by a user in a first input mode.
Specifically, the first input mode may be a voice input mode or a keyboard input mode mentioned in the foregoing embodiments. The keyboard input mode can be a pinyin input mode or a five-stroke input mode.
The process of inputting data by the user in the voice input manner may refer to the description of the third part in the first embodiment or the second embodiment, which is not repeated here.
The process of inputting data by the user through the keyboard input mode may refer to the description of fig. 3A or the description of the third part in the third embodiment, which is not repeated here.
S902: and determining a target result according to the first data, wherein the target result is a historical selection result of the second data input by the user in a second input mode, the syllable sequence corresponding to the first data and the syllable sequence is the same, and the target result is different from the universal result corresponding to the syllable sequence.
The history selection result is a result selected by the user when the user inputs the second data by adopting other input modes before the history selection result. The general result is a result that conforms to most users' usage habits.
For example, if the first input mode is a voice input mode, the input first data is an audio signal, the syllable sequence output after the audio signal is input into the acoustic model according to the audio feature extracted from the audio signal may be zhang 'san, the second input mode is a keyboard input mode, the second data is a character string zhangsan, the history selection result of the user inputting the second data by using the second input mode is "zhang three", and the general result corresponding to the syllable sequence zhang' san is "zhang three". It is known that the target result determined from the first data is "chapter three", not "Zhang three".
That is, the target result does not conform to the usage habit of the public, but is associated with the data input by the user in other input modes, and conforms to the usage habit when the user inputs the data in other input modes.
In a specific implementation manner, the step S902 may specifically include the following steps:
s9021: and searching a first language library according to the first data input by the user, and outputting a first result.
Specifically, the first language library may be a general language library mentioned in the foregoing embodiment. The universal language library can be a preset language library, and the language library can be obtained according to data input by a large number of users in various input modes. The general language library may be a language library of the input method when the user downloads the input method, or may be a language library of the electronic device 100.
Specifically, the universal language library may include, but is not limited to, a plurality of words, syllable sequences corresponding to each word, and word frequencies corresponding to the words. In addition, the generic language library may also be used to calculate the score of the candidate word in the sentence structure, such that the electronic device 100 preferentially displays the highest-scoring candidate word in the list of alternatives 402. Possibly, the same syllable sequence may correspond to a plurality of different words, and the word frequencies corresponding to the words may be different.
S9022: and searching a second language library according to the first data input by the user, and outputting a second result.
In particular, the second language library may be a personalized language library as mentioned in the previous embodiments. The personalized language library may record data input by the user in other input modes than the first input mode.
Possibly, the word with the highest word frequency corresponding to the syllable sequence in the personalized language library is inconsistent with the word with the highest word frequency corresponding to the syllable sequence in the universal language library. Namely, the personalized language library records the own use habit of the user using the electronic equipment, and the general language library records the use habit of the public. The user's own usage habits are different from the usage habits of the public.
Possibly, the syllable sequence corresponding to the data input by the user in the first input mode may be a part of the syllable sequence corresponding to the data recorded in the personalized language library. For example, the syllable sequence corresponding to the data input by the user using the first input mode may be gao 'xin' qu, and the syllable sequence corresponding to the data recorded in the personalized language library may be dong 'hu' gao 'xin' qu.
Conversely, the syllable sequence corresponding to the data recorded in the personalized language library may be a part of the syllable sequence corresponding to the data input by the user in the first input mode. For example, the syllable sequence corresponding to the data input by the user using the first input method may be wo ' yao ' qu ' dong ' hu ' gao ' xin ' qu, and the syllable sequence corresponding to the data recorded in the personalized language library may be dong ' hu ' gao ' xin ' qu.
The input modes included in the electronic device 100 may include, but are not limited to: keyboard input mode, voice input mode, image scanning input mode and handwriting input mode.
If the first input mode is a voice input mode, the personalized language library can record the data input by the user in a keyboard input mode, an image scanning input mode and a handwriting input mode.
If the first input mode is a keyboard input mode, the personalized language library can record the data input by the user in the image scanning mode, the voice input mode and the handwriting input mode.
The second result may or may not be present. When the second result does not exist, the second result output by the second language library can be regarded as null.
The keyboard input mode in the embodiment of the present application may be a pinyin input mode or a wubi input mode. The interface 305 of the input method application may include an input interface of a 26-key letter combination key or an input interface of a 9-key number combination key.
It can be appreciated that the embodiment of the present application does not limit the implementation sequence of S9021 and S9022.
S9023: and determining a target result according to the first result and the second result.
Possibly, the first result and the second result are identical. The target result is either the first result or the second result.
Possibly, the first result and the second result are not identical and the second result is not empty. The target result is the second result.
Possibly, the first result and the second result are inconsistent and the second result is empty. The target result is the first result.
S903: and displaying the target result in the target input interface.
Specifically, after determining the target result, the target result may be displayed in the target input interface.
If the user inputs audio data through an application having an input box, such as a WeChat, the electronic device 100 may display the target result in the input box 303 in the WeChat interface. The target input interface is a user interface provided with the input box 303 and used for receiving voice data input by a user. Such as the user interface shown in fig. 5.
If the user, after waking up the voice assistant, inputs audio data through the voice assistant, the electronic device 100 may display the target result in the interface of the voice assistant. The interface of the voice assistant is the target input interface.
By implementing the embodiment of the application, the data input by users using different input modes can be fused, the data can be shared in the different input modes, the processing capacity of the electronic equipment on the input data is enhanced, the output result is more in line with the use habit of the users compared with the prior art, and the input efficiency of the users is improved.
In other embodiments, the data input by the user in the first input mode is searched in the first language library, and a first result corresponding to the complete input data is obtained. And searching in a second language library to obtain a second result of one part of the input data, and the second result of the other part of the input data is not present. The final result may be a concatenation of the result of another portion of the input data in the first language library (i.e., the result of the first result corresponding to the other portion) with the second result.
For example, the user inputs an audio signal by using a voice input method, and the corresponding syllable sequence obtained according to the audio signal is wo ' mini ' ian ' yao ' he ' xiao ' mini ' yi ' qi ' qu ' guang ' jie. The first result obtained by searching the syllable sequence in the first language library is' I am going to visit with a small street together. And in the second language library, obtaining that the second result of the syllable sequence xiao ' mini corresponding to a part of the input audio signal is ' small name ', and the second result of other syllable sequences does not exist. The final result may be that the result except syllable sequence xiao' mini in the first result is spliced with the second result, and the final result is "i am going to walk with the small name" is obtained.
The application program providing the first input method and the application program providing the second input method may be the same or different. For example, the application providing the keyboard input mode may be an input method, and the application providing the voice input mode may be a voice assistant or the like.
Next, a specific input method provided in the embodiments of the present application will be described.
As shown in fig. 9, the input method may include the following steps:
s1001: and receiving user operation of inputting the character string wuhedadao by adopting a keyboard input mode.
Specifically, the user may input the above character string in an application interface of the input method.
S1002: and receiving user operation for determining that the result corresponding to the character string is five and a large channel.
Specifically, the process of the user determining the result corresponding to the above character string may refer to the processes shown in fig. 3A to 3D.
In the embodiment of the application, the keyboard input mode is the second input mode, the data input by the user through the second input mode is the character string wuheidao, and the history selection result is five and a large channel.
S1003: and storing the character strings and the corresponding results of the character strings into a personalized language library.
S1004: an audio signal input by a voice input mode is received.
Specifically, the user may input the audio signal by using a voice assistant, the user may also input the audio signal by using a google map, and the user may also input the audio signal by using an input method. The embodiments of the present application are not limited in this regard.
In the embodiment of the application, the voice input mode is a first input mode, and the audio signal input by the user is first data.
S1005: and determining the syllable sequence corresponding to the audio signal as wu ' he ' da ' dao.
Specifically, the electronic device 100 may extract an audio feature of the audio signal, and obtain a syllable sequence corresponding to the audio signal after inputting the audio feature into the acoustic model.
S1006: searching a universal language library, and determining that the first result is five river channels.
In the embodiment of the application, the five river channel is a general result corresponding to the syllable sequence wu ' he ' da ' dao.
S1007: searching the personalized language library, and determining the second result as 'five and big ways'.
In the embodiment of the application, "five and big ways" are history selection results.
S1008: comparing the first result and the second result, and determining the target result as five and a large channel.
It can be seen that the target result is different from the general result and the same as the history selection result.
S1009: the target result "five and big lanes" is displayed in the target input interface.
Specifically, if the user inputs the audio signal using the voice assistant in S1004, the target input interface is the interface of the voice assistant; if the user inputs the audio signal by using the google map in S1004, the target input interface is the interface of the google map; if the user inputs the audio signal by using the input method in S1004, the target input interface may be a user interface for receiving the voice data input by the user, such as the user interface shown in fig. 5.
Specifically, in the case where the first result and the second result are inconsistent, the second result is prioritized, that is, the second result is prioritized as the target result. The second result is obtained by searching the personalized language library, and the personalized language library is obtained according to the data input by the user using the electronic equipment in other input modes. Therefore, the result output by the personalized language library accords with the use habit of the user, the accuracy is higher, and the input efficiency of the user is high.
Embodiments also provide a computer readable storage medium having instructions stored therein, which when run on a computer or processor, cause the computer or processor to perform one or more steps of any of the methods described above. The above-described constituent modules of the electronic apparatus may be stored in the computer-readable storage medium if implemented in the form of software functional units and sold or used as independent products.
In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, produces a flow or function in accordance with embodiments of the present application, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable apparatus. The computer instructions may be stored in or transmitted across a computer-readable storage medium. The computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by a wired (e.g., coaxial cable, fiber optic, digital Subscriber Line (DSL)), or wireless (e.g., infrared, wireless, microwave, etc.). The computer readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that contains an integration of one or more available media. The usable medium may be a magnetic medium (e.g., a floppy disk, a hard disk, a magnetic tape), an optical medium (e.g., a DVD), or a semiconductor medium (e.g., a Solid State Disk (SSD)), or the like.
Those of ordinary skill in the art will appreciate that implementing all or part of the above-described method embodiments may be accomplished by a computer program to instruct related hardware, the program may be stored in a computer readable storage medium, and the program may include the above-described method embodiments when executed. And the aforementioned storage medium includes: ROM or random access memory RAM, magnetic or optical disk, etc.

Claims (9)

1. An input method, comprising:
receiving first data input by a user in a first input mode;
determining a target result according to the first data, wherein the target result is a history selection result of the user inputting second data in a second input mode, the syllable sequence corresponding to the first data and the syllable sequence corresponding to the second data are the same, and the target result is different from a universal result corresponding to the syllable sequence;
displaying the target result in a target input interface;
the determining a target result according to the first data comprises:
searching a first language library according to the first data, and outputting a first result; the first language library is a universal language library;
Searching a second language library according to the first data, and outputting a second result; the second language library is a language library generated according to the second data;
determining a target result according to the first result and the second result; the second result is not null, and the target result is the second result;
the first language library and the second language library comprise words, syllable sequences corresponding to the words and word frequencies corresponding to the words;
and the word with highest word frequency corresponding to the syllable sequence in the second language library is different from the word with highest word frequency corresponding to the syllable sequence in the first language library.
2. The method of claim 1, wherein syllable sequences corresponding to words in the second language library are part of syllable sequences corresponding to the first data.
3. The method of claim 1, wherein the syllable sequence corresponding to the first data is part of a syllable sequence corresponding to a word in the second language library.
4. The method of any of claims 1-3, wherein the first input mode is a voice input mode and the second input mode is at least one of a keyboard input mode, an image scan input mode, and a handwriting input mode.
5. The method of any of claims 1-3, wherein the first input mode is a keyboard input mode and the second input mode is at least one of a voice input mode, an image input mode, and a handwriting input mode.
6. The method of any one of claims 1-5, wherein the second result is null and the target result is the first result.
7. An electronic device, comprising: one or more processors and memory;
the memory being coupled to the one or more processors, the memory being for storing computer program code comprising computer instructions which, when executed by the one or more processors, the electronic device performs the input method of any of claims 1-6.
8. A computer storage medium comprising computer instructions which, when run on an electronic device, cause the electronic device to perform the input method of any of claims 1-6.
9. A computer program product, characterized in that the computer program product, when run on an electronic device, causes the electronic device to perform the input method according to any of claims 1-6.
CN201910661615.1A 2019-05-28 2019-07-22 Input method and electronic equipment Active CN110502126B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2019104537333 2019-05-28
CN201910453733 2019-05-28

Publications (2)

Publication Number Publication Date
CN110502126A CN110502126A (en) 2019-11-26
CN110502126B true CN110502126B (en) 2023-12-29

Family

ID=68586664

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910661615.1A Active CN110502126B (en) 2019-05-28 2019-07-22 Input method and electronic equipment

Country Status (1)

Country Link
CN (1) CN110502126B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111627438A (en) * 2020-05-21 2020-09-04 四川虹美智能科技有限公司 Voice recognition method and device
CN112214115B (en) * 2020-09-25 2024-04-30 汉海信息技术(上海)有限公司 Input mode identification method and device, electronic equipment and storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01191199A (en) * 1988-01-27 1989-08-01 Toshiba Corp Voice input device
JPH0651939A (en) * 1992-07-29 1994-02-25 Nippon Telegr & Teleph Corp <Ntt> Voice input device
CN106774975A (en) * 2016-11-30 2017-05-31 百度在线网络技术(北京)有限公司 Input method and device
CN106873798A (en) * 2017-02-16 2017-06-20 北京百度网讯科技有限公司 For the method and apparatus of output information
CN106873799A (en) * 2017-02-16 2017-06-20 北京百度网讯科技有限公司 Input method and device
CN107731229A (en) * 2017-09-29 2018-02-23 百度在线网络技术(北京)有限公司 Method and apparatus for identifying voice
CN108428446A (en) * 2018-03-06 2018-08-21 北京百度网讯科技有限公司 Audio recognition method and device
CN109243430A (en) * 2017-07-04 2019-01-18 北京搜狗科技发展有限公司 A kind of audio recognition method and device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01191199A (en) * 1988-01-27 1989-08-01 Toshiba Corp Voice input device
JPH0651939A (en) * 1992-07-29 1994-02-25 Nippon Telegr & Teleph Corp <Ntt> Voice input device
CN106774975A (en) * 2016-11-30 2017-05-31 百度在线网络技术(北京)有限公司 Input method and device
CN106873798A (en) * 2017-02-16 2017-06-20 北京百度网讯科技有限公司 For the method and apparatus of output information
CN106873799A (en) * 2017-02-16 2017-06-20 北京百度网讯科技有限公司 Input method and device
CN109243430A (en) * 2017-07-04 2019-01-18 北京搜狗科技发展有限公司 A kind of audio recognition method and device
CN107731229A (en) * 2017-09-29 2018-02-23 百度在线网络技术(北京)有限公司 Method and apparatus for identifying voice
CN108428446A (en) * 2018-03-06 2018-08-21 北京百度网讯科技有限公司 Audio recognition method and device

Also Published As

Publication number Publication date
CN110502126A (en) 2019-11-26

Similar Documents

Publication Publication Date Title
JP5996783B2 (en) Method and terminal for updating voiceprint feature model
EP2440988B1 (en) Touch anywhere to speak
US9111538B2 (en) Genius button secondary commands
US20150213127A1 (en) Method for providing search result and electronic device using the same
CN103811005A (en) Mobile terminal and control method thereof
CN109599104B (en) Multi-beam selection method and device
CN111524501B (en) Voice playing method, device, computer equipment and computer readable storage medium
CN107133354B (en) Method and device for acquiring image description information
CN110659690B (en) Neural network construction method and device, electronic equipment and storage medium
CN110263131B (en) Reply information generation method, device and storage medium
CN111739517B (en) Speech recognition method, device, computer equipment and medium
CN111553464B (en) Image processing method and device based on super network and intelligent equipment
KR20110052898A (en) Method for setting background screen and mobile terminal using the same
CN110781813B (en) Image recognition method and device, electronic equipment and storage medium
CN112130714B (en) Keyword search method capable of learning and electronic equipment
US11335348B2 (en) Input method, device, apparatus, and storage medium
CN109685041B (en) Image analysis method and device, electronic equipment and storage medium
CN110502126B (en) Input method and electronic equipment
CN111339737A (en) Entity linking method, device, equipment and storage medium
CN114692639A (en) Text error correction method and electronic equipment
CN113298091A (en) Image processing method and device, electronic equipment and storage medium
CN111652107A (en) Object counting method and device, electronic equipment and storage medium
CN109447258B (en) Neural network model optimization method and device, electronic device and storage medium
CN112445906A (en) Method and device for generating reply message
CN112740148A (en) Method for inputting information into input box and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant