CN109407946B - Graphical interface target selection method based on voice recognition - Google Patents
Graphical interface target selection method based on voice recognition Download PDFInfo
- Publication number
- CN109407946B CN109407946B CN201811056705.XA CN201811056705A CN109407946B CN 109407946 B CN109407946 B CN 109407946B CN 201811056705 A CN201811056705 A CN 201811056705A CN 109407946 B CN109407946 B CN 109407946B
- Authority
- CN
- China
- Prior art keywords
- word
- marked
- user
- words
- tagged
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The invention relates to a graphical interface target selection method based on voice recognition, and belongs to the field of voice target selection. It comprises the following steps: creating a voice mark word stock and mark points; associating the tagged points with the tagged words; the user speaks the marked words and continuously sounds to generate a circle with the corresponding marked point as the circle center and the radius continuously increasing along with the continuous sound of the user; dividing the circle into a plurality of arc sections; associating the arc segment with the marker word; a user speaks a marker word, and a circle is generated by taking the center point of the corresponding arc segment as the center of a circle and the distance between the marker word and the adjacent arc segment as the radius; dividing the circle into a plurality of areas; associating the region with a tagged word; and (4) the user speaks the marked words and selects the target by taking the central point in the corresponding area as a selection point. The invention gives attention to visual feedback to the user, so that the user can clearly know when and what command to say, and does not need to learn the voice command for many times, thereby greatly facilitating the user to use the intelligent equipment.
Description
Technical Field
The invention relates to the field of voice target selection, in particular to a graphical interface target selection method based on voice recognition.
Background
Through the development of many years, the voice recognition technology gradually moves from a laboratory to practical application, starts to gradually become a landmark technology in the field of information industry, and gradually enters our daily life in man-machine interaction application, for example, the existing smart phone, tablet computer, smart television, vehicle-mounted tablet, smart bracelet, smart watch and the like are usually accompanied by a voice recognition function. The application program of the voice control intelligent device can be developed by utilizing the voice recognition technology, people do not need to carry out manual physical key operation on the intelligent device, and can realize the operation on the intelligent device only in a voice command mode, so that the method has important significance for a plurality of disabled people. However, the existing voice target selection method can only select a preset target, cannot select any target point appearing on the screen, and cannot give intuitive visual feedback to the user in the selection process, so that the user does not know when the user can speak the command, does not know whether the spoken command is valid, and does not know what the spoken command is, for example: when a certain target in a screen needs to be selected, if a plurality of targets are completely consistent, when a user says 'select a certain target', a voice selection system usually directly screens all matched targets, sets different target names for the consistent targets respectively, generates a new selectable command sentence, and waits for the user to say a new command, so that the user can not clearly know when the user can say the command, when the previous command is finished, and the user does not know what the new command is, thereby influencing the user experience.
Disclosure of Invention
The invention aims to solve the technical problem of providing a voice target selection method which is more intuitive, more efficient in selection and more convenient to use.
The technical scheme of the invention is as follows: a graphical interface target selection method based on voice recognition comprises the following steps:
step1, creating a voice marked word bank in the intelligent equipment, and setting at least one type of marked words in the voice marked word bank, wherein the type of marked words at least comprises one marked word;
step2, creating a plurality of mark points on the screen of the intelligent device and displaying the mark points on the screen of the intelligent device;
step3, respectively associating the mark points of the screen of the intelligent device with a certain type of mark words of the voice mark word library, and respectively displaying the mark words in the type around the corresponding mark points;
step4, judging whether the user speaks a tagged word displayed in a screen of the intelligent device in Step3, and judging whether the user speaks the tagged word continuously, if the user does not speak the tagged word, the user does not judge whether the user speaks the tagged word continuously, the system waits until the user speaks the tagged word, if the user speaks the tagged word and a target to be selected is just located at a tagged point corresponding to the tagged word, the user does not speak the tagged word continuously, the target is selected by taking the tagged point corresponding to the tagged word as a selection point, and if the user speaks the tagged word and the target to be selected is located outside the tagged point corresponding to the tagged word, the user speaks continuously, and a circle is generated, wherein the circle takes the tagged point corresponding to the tagged word as the circle center and the radius of the circle continuously increases along with the user's voice;
step5 cancels the association between the marked words and the marked points in Step3, clears the marked words displayed on the screen of the intelligent device in Step3, and clears the marked points created on the screen of the intelligent device in Step 2;
step6, dividing the circle generated on the screen of the intelligent device in Step4 into a plurality of arc segments, associating the arc segments with certain type of tagged words in the voice tagged word library respectively, and displaying the tagged words in the type of tagged words around the corresponding arc segments respectively;
step7, judging whether the user speaks a marker word displayed in the screen of the intelligent device in Step6, if the user does not speak the marker word, waiting until the user speaks the marker word, if the system waits for time out, returning to Step2, and if the user speaks the marker word, generating a circle by taking the center point of the arc segment corresponding to the marker word as the center of the circle and the distance between the center point and the intersection point of the adjacent arc segment of the arc segment as the radius;
step8 cancels the association of the marked words and the arc segments in Step6, clears the marked words displayed on the screen of the intelligent device in Step6, and clears the circle generated on the screen of the intelligent device in Step 4;
step9, dividing the circle generated on the screen of the intelligent device in Step7 into a plurality of areas, simultaneously associating each area with a certain type of tagged words in a voice tagged word library respectively, and displaying the tagged words in the type of tagged words in the corresponding area respectively;
step10 judges whether the user utters the tagged word displayed in the screen of the smart device in Step9, if the user has not uttered the tagged word, the system waits until the user utters the tagged word, if the system waits for time out, the system returns to Step2, and if the user utters the tagged word, the target selection is performed by taking the central point in the area corresponding to the tagged word as a selection point.
Specifically, the smart devices in the above method refer to computers and smart phones having a voice recognition function.
Specifically, the voice tagged word library in Step1 includes numeric tagged words, alphabetical tagged words, and text tagged words, or user-defined tagged words.
Specifically, the creation of the mark point in Step2 refers to: the screen of the intelligent device is divided into a plurality of blocks, and the central point of each block is taken as a mark point.
Specifically, the Step4 refers to repeating the sounding of the same tag word until the generated circle approaches the target to be selected, the user stops sounding, and the radius of the circle stops increasing.
Specifically, the association between Step3, Step6 and Step9 and a certain type of tagged word in the voice tagged word library is random, that is, the categories of the associated tagged words in the three steps can be the same or different, and the selection of the tagged words in the type of tagged words is random when the association is performed.
The invention has the beneficial effects that: the graphical interface target selection method based on the voice recognition emphasizes visual feedback to the user, generates a circle with a radius which is continuously enlarged through continuous vocalization of the user to acquire the position of the target point, and generates the circle based on the position to select the target, so that the visual feedback can be provided for the user, the user can clearly know what command is spoken when the user does not need to learn voice commands for many times, the user can conveniently use intelligent equipment, the time for selecting the target is shortened, and the precision for selecting the target is improved.
Detailed Description
The present invention will be described in further detail with reference to specific examples.
Example 1: a graphical interface target selection method based on voice recognition comprises the following steps:
step1, creating a voice mark word bank in the intelligent device, and setting at least one type of mark words in the voice mark word bank, wherein the type of mark words at least contains one mark word, and the voice mark word bank comprises digital mark words, letter mark words and character mark words, or user-defined mark words;
step2, creating a plurality of marking points on the screen of the intelligent equipment, dividing the screen of the intelligent equipment into a plurality of blocks, taking the central point of each block as a marking point, and displaying the marking point on the screen of the intelligent equipment;
step3, respectively associating the mark points of the screen of the intelligent device with a certain type of mark words of the voice mark word library, and respectively displaying the mark words in the type around the corresponding mark points;
step4, judging whether the user utters the marked word displayed in the screen of the intelligent device in Step3, and judging whether the user utters the marked word continuously, if the user does not utter the marked word, then not judging whether the user utters the marked word continuously, the system waits until the user utters the marked word, if the user utters the marked word and the target to be selected is just positioned at the marked point corresponding to the marked word, the user does not utter the marked word continuously, the target selection is carried out by taking the marked point corresponding to the marked word as the selection point, if the user utters the marked word and the target to be selected is positioned outside the marked point corresponding to the marked word, the user utters the marked word continuously, the continuous uttering means that the same marked word is uttered repeatedly, a circle with the marked point corresponding to the marked word as the center of the circle and the radius increasing continuously along with the utterance of the user is generated until the generated circle is close to the target to be selected, and the user stops uttering the voice, and the radius of the circle stops increasing.
Step5 cancels the association between the marked words and the marked points in Step3, clears the marked words displayed on the screen of the intelligent device in Step3, and clears the marked points created on the screen of the intelligent device in Step 2;
step6, dividing the circle generated on the screen of the intelligent device in Step4 into a plurality of arc segments, associating the arc segments with certain type of tagged words in the voice tagged word library respectively, and displaying the tagged words in the type of tagged words around the corresponding arc segments respectively;
step7, judging whether the user speaks a marker word displayed in the screen of the intelligent device in Step6, if the user does not speak the marker word, waiting until the user speaks the marker word, if the system waits for time out, returning to Step2, and if the user speaks the marker word, generating a circle by taking the center point of the arc segment corresponding to the marker word as the center of the circle and the distance between the center point and the intersection point of the adjacent arc segment of the arc segment as the radius;
step8 cancels the association of the marked words and the arc segments in Step6, clears the marked words displayed on the screen of the intelligent device in Step6, and clears the circle generated on the screen of the intelligent device in Step 4;
step9, dividing the circle generated on the screen of the intelligent device in Step7 into a plurality of areas, simultaneously associating each area with a certain type of tagged words in a voice tagged word library respectively, and displaying the tagged words in the type of tagged words in the corresponding area respectively;
step10 judges whether the user utters the tagged word displayed in the screen of the smart device in Step9, if the user has not uttered the tagged word, the system waits until the user utters the tagged word, if the system waits for time out, the system returns to Step2, and if the user utters the tagged word, the target selection is performed by taking the central point in the area corresponding to the tagged word as a selection point.
The intelligent equipment in the method refers to a computer and a smart phone with a voice recognition function. The Step3, Step6 and Step9 are all associated randomly with a certain type of tagged word in the voice tagged word library, that is, the categories of the associated tagged words in the three steps can be the same or different, and the selection of the tagged words in the type of tagged words is also random when the association is carried out.
Example 2: the following describes the method for selecting a target on a graphical interface based on voice recognition in further detail by taking the example that a user randomly appears a target folder in a voice selection screen when using a computer with a voice recognition function.
Firstly, a voice mark word bank is created in the intelligent equipment, and a digital mark word is set in the voice mark word bank: "1", "2", "3", "4", "5", "6", "7", "8", "9", "0";
secondly, equally dividing the length and the width of the screen of the intelligent equipment into three parts on the screen of the intelligent equipment, dividing the screen into nine rectangular areas with equal areas, taking the diagonal intersection point of the rectangular areas as a mark point, and displaying the 9 mark points on the screen of the intelligent equipment;
thirdly, associating 9 marking points of the screen of the intelligent device with the '1', '2', '3', '4', '5', '6', '7', '8' and '9' of the 'digital' type marking words of the voice marking word stock respectively, and setting the marking words as the background of a rectangular area for dividing the screen to be displayed;
fourthly, assuming that the folder randomly appearing in the screen is positioned at the upper right corner of the screen and just below the mark point corresponding to the mark word "3", the user speaks the mark word "3" in the screen of the intelligent device and pauses the sound production, the mark point corresponding to the mark word "3" spoken by the user is taken as the selection point for target selection, assuming that the folder randomly appearing in the screen is positioned at the upper right corner of the screen and is not positioned below the mark point corresponding to the mark word "3" but is positioned in the area where the mark point is positioned, the user speaks the mark word "3" in the screen of the intelligent device and continuously produces the sound, namely, "333333 …", the mark point corresponding to the mark word "3" spoken by the user generates a circle with the mark point corresponding to the mark word as the center of the circle and the radius continuously increasing with the sound production of the user, and when the arc of the circle is close to the target file, the user stops the sound production, and the radius of the circle stops increasing;
fifthly, canceling the association between the mark words and the mark points in the third step, clearing the mark words displayed on the screen of the intelligent equipment in the third step, and clearing the mark points created on the screen of the intelligent equipment in the second step;
sixthly, taking the point of the upper arc of the circle intersected by the straight line which is vertical and passes through the center of the circle as a starting point, averagely dividing the arc into nine sections, respectively associating the arc sections with the '1', '2', '3', '4', '5', '6', '7', '8' and '9' of the 'digital' type marking words of the voice marking thesaurus, and displaying the marking words above the corresponding arc sections;
seventhly, assuming that a file randomly appearing in the screen is clamped under the 2 nd arc segment of the circle in the sixth step, and the user speaks a marker word '2' in the screen of the intelligent device, generating a circle by taking the center point of the arc segment corresponding to the marker word as the center of a circle and the distance between the center point and the intersection point of the adjacent arc segment of the arc segment as a radius;
eighth step, canceling the association between the mark word and the arc segment in the sixth step, clearing the mark word displayed on the screen of the intelligent device in the sixth step, and clearing the circle generated on the screen of the intelligent device in the fourth step;
ninth, the left horizontal radius of the circle generated on the screen of the intelligent device in the seventh step is rotated clockwise by 40 degrees and rotated 9 times, the circle is evenly divided into nine sector areas, the sector areas and the concentric circle with the radius of 1/3 are intersected, a circular ring evenly divided into nine areas is obtained after the intersected parts are removed, the concentric circle is added, the circle is divided into ten areas, meanwhile, each area is respectively associated with the '1', '2', '3', '4', '5', '6', '7', '8', '9' and '0' of the 'digital' type mark words of the voice mark word library, and the mark words are set as the background of the area dividing the circle to be displayed;
and step ten, assuming that the file randomly appearing in the screen is clamped under the area of the circle corresponding to the marker word "1" in the step ninth, the user speaks the marker word "1" displayed in the screen of the intelligent device, and the center point in the area corresponding to the marker word is taken as a selection point to perform target selection.
When the target is selected, the corresponding mark point coordinates or the central point coordinates of the corresponding area are obtained after the mark words are spoken, and then the system receives the coordinate information and controls the cursor to move to the coordinate position, so that the cursor can be moved to a blank area to perform right key operation and the like according to the method when the target does not appear on the graphical interface, and the target can be selected more accurately after the area selection is gradually divided through the steps, namely the target selection accuracy is improved, and the selection time is also saved.
The graphical interface target selection method based on voice recognition in the embodiment is suitable for being used in the situation that a disabled user or a user with inconvenient hands selects a random target at any position in a screen.
The above examples are only for describing the preferred embodiments of the present invention, and are not intended to limit the scope of the present invention, and various modifications and improvements of the technical solution of the present invention by those skilled in the art should be made within the protection scope defined by the claims of the present invention without departing from the spirit of the present invention.
Claims (6)
1. A graphical interface target selection method based on voice recognition is characterized in that: the method comprises the following steps:
step1, creating a voice marked word bank in the intelligent equipment, and setting at least one type of marked words in the voice marked word bank, wherein the type of marked words at least comprises one marked word;
step2, creating a plurality of mark points on the screen of the intelligent device and displaying the mark points on the screen of the intelligent device;
step3, respectively associating the mark points of the screen of the intelligent device with a certain type of mark words of the voice mark word library, and respectively displaying the mark words in the type around the corresponding mark points;
step4, judging whether the user speaks the marked word displayed in the screen of the intelligent device in Step3, and judging whether the user speaks the marked word continuously, if the user does not speak the marked word, then not judging whether the user speaks the marked word continuously, waiting until the user speaks the marked word, if the user speaks the marked word and the target to be selected is just located at the marked point corresponding to the marked word, then the user does not speak continuously, selecting the target by taking the marked point corresponding to the marked word as a selection point, if the user speaks the marked word and the target to be selected is located outside the marked point corresponding to the marked word, then the user vocalizes continuously, generating a circle with the marked point corresponding to the marked word as the center of the circle and the radius increasing continuously along with the user vocalization, and until the generated circle is close to the target to be selected, stopping the user vocalizing and the radius of the circle increasing;
step5 cancels the association between the marked words and the marked points in Step3, clears the marked words displayed on the screen of the intelligent device in Step3, and clears the marked points created on the screen of the intelligent device in Step 2;
step6, dividing the circle generated on the screen of the intelligent device in Step4 into a plurality of arc segments, associating the arc segments with certain type of tagged words in the voice tagged word library respectively, and displaying the tagged words in the type of tagged words around the corresponding arc segments respectively;
step7, judging whether the user speaks a marker word displayed in the screen of the intelligent device in Step6, if the user does not speak the marker word, waiting until the user speaks the marker word, if the system waits for time out, returning to Step2, and if the user speaks the marker word, generating a circle by taking the center point of the arc segment corresponding to the marker word as the center of the circle and the distance between the center point and the intersection point of the adjacent arc segment of the arc segment as the radius;
step8 cancels the association of the marked words and the arc segments in Step6, clears the marked words displayed on the screen of the intelligent device in Step6, and clears the circle generated on the screen of the intelligent device in Step 4;
step9, dividing the circle generated on the screen of the intelligent device in Step7 into a plurality of areas, simultaneously associating each area with a certain type of tagged words in a voice tagged word library respectively, and displaying the tagged words in the type of tagged words in the corresponding area respectively;
step10 judges whether the user utters the tagged word displayed in the screen of the smart device in Step9, if the user has not uttered the tagged word, the system waits until the user utters the tagged word, if the system waits for time out, the system returns to Step2, and if the user utters the tagged word, the target selection is performed by taking the central point in the area corresponding to the tagged word as a selection point.
2. A method for voice recognition based graphical interface target selection as claimed in claim 1 wherein: the intelligent device in Step 1-Step 10 refers to a computer with a voice recognition function.
3. A method for selecting a graphical interface target based on speech recognition according to claim 1 or 2, characterized in that: the phonetic label word library in Step1 includes numeric label words, letter label words and text label words.
4. A method for selecting a graphical interface target based on speech recognition according to claim 1 or 2, characterized in that: the creation of the mark point in Step2 refers to: the screen of the intelligent device is divided into a plurality of blocks, and the central point of each block is taken as a mark point.
5. A method for selecting a graphical interface target based on speech recognition according to claim 1 or 2, characterized in that: the continuous sounding in Step4 refers to repeated sounding of the same tag word until the generated circle approaches the target to be selected, the user stops sounding, and the radius of the circle stops increasing.
6. A method for selecting a graphical interface target based on speech recognition according to claim 1 or 2, characterized in that: the Step3, Step6 and Step9 are all associated randomly with a certain type of tagged word in the voice tagged word library, that is, the categories of the associated tagged words in the three steps can be the same or different, and the selection of the tagged words in the type of tagged words is also random when the association is carried out.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811056705.XA CN109407946B (en) | 2018-09-11 | 2018-09-11 | Graphical interface target selection method based on voice recognition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811056705.XA CN109407946B (en) | 2018-09-11 | 2018-09-11 | Graphical interface target selection method based on voice recognition |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109407946A CN109407946A (en) | 2019-03-01 |
CN109407946B true CN109407946B (en) | 2021-05-14 |
Family
ID=65464748
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811056705.XA Active CN109407946B (en) | 2018-09-11 | 2018-09-11 | Graphical interface target selection method based on voice recognition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109407946B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113539253B (en) * | 2020-09-18 | 2024-05-14 | 厦门市和家健脑智能科技有限公司 | Audio data processing method and device based on cognitive assessment |
CN115248650B (en) * | 2022-06-24 | 2024-05-24 | 南京伟柏软件技术有限公司 | Screen reading method and device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1667700A (en) * | 2004-03-10 | 2005-09-14 | 微软公司 | New-word pronunciation learning using a pronunciation graph |
CN102547463A (en) * | 2011-12-15 | 2012-07-04 | Tcl集团股份有限公司 | Method and device for locating interface focus of TV set, and TV set |
CN103680498A (en) * | 2012-09-26 | 2014-03-26 | 华为技术有限公司 | Speech recognition method and speech recognition equipment |
CN103905636A (en) * | 2014-03-03 | 2014-07-02 | 联想(北京)有限公司 | Information processing method and electronic device |
CN105100460A (en) * | 2015-07-09 | 2015-11-25 | 上海斐讯数据通信技术有限公司 | Method and system for controlling intelligent terminal by use of sound |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8370746B2 (en) * | 1992-12-14 | 2013-02-05 | Monkeymedia, Inc. | Video player with seamless contraction |
-
2018
- 2018-09-11 CN CN201811056705.XA patent/CN109407946B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1667700A (en) * | 2004-03-10 | 2005-09-14 | 微软公司 | New-word pronunciation learning using a pronunciation graph |
CN102547463A (en) * | 2011-12-15 | 2012-07-04 | Tcl集团股份有限公司 | Method and device for locating interface focus of TV set, and TV set |
CN103680498A (en) * | 2012-09-26 | 2014-03-26 | 华为技术有限公司 | Speech recognition method and speech recognition equipment |
CN103905636A (en) * | 2014-03-03 | 2014-07-02 | 联想(北京)有限公司 | Information processing method and electronic device |
CN105100460A (en) * | 2015-07-09 | 2015-11-25 | 上海斐讯数据通信技术有限公司 | Method and system for controlling intelligent terminal by use of sound |
Non-Patent Citations (1)
Title |
---|
面向视频会议的多功能手写笔白板系统及协同工作的研究;丁怀东,殷继彬;《昆明理工大学学报( 理工版)》;20040831;正文第1-6页 * |
Also Published As
Publication number | Publication date |
---|---|
CN109407946A (en) | 2019-03-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9558737B2 (en) | System and method for audibly presenting selected text | |
CN110825340B (en) | Providing a pre-computed hotword model | |
US8239129B2 (en) | Method and system for improving speech recognition accuracy by use of geographic information | |
CN103000176B (en) | Speech recognition method and system | |
CN109407946B (en) | Graphical interface target selection method based on voice recognition | |
KR101819457B1 (en) | Voice recognition apparatus and system | |
CN103578467A (en) | Acoustic model building method, voice recognition method and electronic device | |
JP2016522903A (en) | Speech recognition system and method | |
KR20060037228A (en) | Methods, systems, and programming for performing speech recognition | |
US20080288260A1 (en) | Input/Output Apparatus Based on Voice Recognition, and Method Thereof | |
KR20200125735A (en) | Multi-party conversation recording/output method using speech recognition technology and device therefor | |
Liesenfeld et al. | Bottom-up discovery of structure and variation in response tokens (‘backchannels’) across diverse languages | |
JP6272496B2 (en) | Method and system for recognizing speech containing a sequence of words | |
JP2006031033A (en) | Information processor | |
KR102091684B1 (en) | Voice recognition text correction method and a device implementing the method | |
US20170301349A1 (en) | Speech recognition system | |
CN113327597A (en) | Speech recognition method, medium, device and computing equipment | |
JP2014235356A (en) | Candidate selection device and candidate selection method using voice recognition | |
KR102446676B1 (en) | Smart table for processing voice recognition and remote control based on ai and iot and operating method therefor | |
JP2006023572A (en) | Dialog system | |
CN107066080B (en) | Chinese character pronunciation, chinese character and symbol coding input method | |
KR101651909B1 (en) | Voice recognition text correction method and a device implementing the method | |
KR102605774B1 (en) | Smart Glass and Voice Recognition System having the same | |
JPH09288493A (en) | Voice recognizing method and voice recognition device and information processing method and information processor | |
JPH06110495A (en) | Speech recognition device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |