CN113539253B - Audio data processing method and device based on cognitive assessment - Google Patents
Audio data processing method and device based on cognitive assessment Download PDFInfo
- Publication number
- CN113539253B CN113539253B CN202010988651.1A CN202010988651A CN113539253B CN 113539253 B CN113539253 B CN 113539253B CN 202010988651 A CN202010988651 A CN 202010988651A CN 113539253 B CN113539253 B CN 113539253B
- Authority
- CN
- China
- Prior art keywords
- data
- text
- array
- audio data
- voice recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000001149 cognitive effect Effects 0.000 title claims abstract description 38
- 238000003672 processing method Methods 0.000 title claims abstract description 13
- 238000000034 method Methods 0.000 claims abstract description 40
- 238000012545 processing Methods 0.000 claims abstract description 24
- 238000011156 evaluation Methods 0.000 claims abstract description 23
- 230000008569 process Effects 0.000 claims abstract description 16
- 230000003930 cognitive ability Effects 0.000 claims abstract description 15
- 238000006243 chemical reaction Methods 0.000 claims abstract description 13
- 238000005516 engineering process Methods 0.000 claims abstract description 11
- 238000004590 computer program Methods 0.000 claims description 10
- 238000012854 evaluation process Methods 0.000 abstract description 8
- 208000010877 cognitive disease Diseases 0.000 abstract description 6
- 238000010586 diagram Methods 0.000 description 7
- 241000283973 Oryctolagus cuniculus Species 0.000 description 6
- 238000004891 communication Methods 0.000 description 6
- 230000006870 function Effects 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 4
- 230000019771 cognition Effects 0.000 description 3
- 244000099147 Ananas comosus Species 0.000 description 2
- 235000007119 Ananas comosus Nutrition 0.000 description 2
- 241000220324 Pyrus Species 0.000 description 2
- 235000014443 Pyrus communis Nutrition 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 230000003920 cognitive function Effects 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 208000020016 psychiatric disease Diseases 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 208000006888 Agnosia Diseases 0.000 description 1
- 241001047040 Agnosia Species 0.000 description 1
- 241000272525 Anas platyrhynchos Species 0.000 description 1
- 208000019901 Anxiety disease Diseases 0.000 description 1
- 241001248531 Euchloe <genus> Species 0.000 description 1
- 241001465754 Metazoa Species 0.000 description 1
- 235000010672 Monarda didyma Nutrition 0.000 description 1
- 244000179970 Monarda didyma Species 0.000 description 1
- 241000287420 Pyrus x nivalis Species 0.000 description 1
- 238000013019 agitation Methods 0.000 description 1
- 230000036506 anxiety Effects 0.000 description 1
- 201000007201 aphasia Diseases 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 238000013075 data extraction Methods 0.000 description 1
- 201000010099 disease Diseases 0.000 description 1
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 1
- 230000004064 dysfunction Effects 0.000 description 1
- 235000013399 edible fruits Nutrition 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007074 memory dysfunction Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/194—Calculation of difference between files
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The invention discloses an audio data processing method and device based on cognitive evaluation, which are characterized in that audio data input by a user according to preset voice recognition related content are collected, and the audio data are converted into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result. The difficulty in evaluation of cognitive dysfunction can be effectively reduced through processing the audio data, and more intelligent, efficient and quick experience is brought to the whole cognitive evaluation process. And the data acquired by the user in the cognitive evaluation process are more diversified and accurate, so that the data can be recorded and evaluated in real time, and the accuracy of the cognitive evaluation is effectively improved.
Description
Technical Field
The invention relates to the technical field of computers, in particular to an audio data processing method and device based on cognitive evaluation.
Background
At present, cognitive dysfunction is one of important diseases affecting the health and life quality of middle-aged and elderly people, and the cognitive dysfunction can be expressed not only by memory dysfunction, aphasia, agnosia, visual space dysfunction and the like, but also by affective behavior disorders such as anxiety, depression, agitation, impulse and the like, and the affective behavior disorders are also causes of disability of patients, so that a heavy burden is brought to society and families. Typically, doctors evaluate cognitive functions of patients during inquiry by conventional inquiry and paper scales, and doctors evaluate the cognitive functions mainly according to the performance of the patients and test results of the paper scales. The process has the advantages of huge workload, very much time and low efficiency, thus bringing great difficulty to the evaluation of the whole cognitive dysfunction, and needing to carry out more intelligent, efficient and rapid improvement on the whole cognitive evaluation process so as to facilitate accurate evaluation on patients.
In the prior art, doctors communicate with patients through a conventional inquiry means, judge according to the answers of the patients or answer according to questions in a paper scale, but voice answer information which can be received by the doctors is only received and judged instantaneously, so that the one-sided and subjective comparison is carried out, and more accurate recording and judging standards are lacked.
In view of this, it is important to propose a method and apparatus for audio data processing based on cognitive assessment.
Disclosure of Invention
The problems of one-sided and subjective comparison, lack of more accurate record and judgment standards and the like in the cognitive level evaluation process are solved. An object of an embodiment of the present application is to provide an audio data processing method and apparatus based on cognitive evaluation to solve the technical problems mentioned in the background section above.
In a first aspect, an embodiment of the present application provides an audio data processing method based on cognitive evaluation, including the steps of:
s1: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology;
s2: acquiring preset data generated by text conversion of voice recognition related content;
s3: comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
S4: and collecting time data of the user in the process of completing the voice recognition related content, and evaluating the cognitive ability of the user by combining the comparison result.
In some embodiments, the speech recognition related content comprises graphics or numbers, and the presentation mode of the speech recognition related content comprises displaying content or recording and playing content by a graphical interface. And displaying the content or recording and playing the content to show or guide the user to set forth the content of the preset part through the graphical interface.
In some embodiments, the preset data includes a first array, where the first array includes a one-dimensional array formed by text in the graphical interface display content, or a one-dimensional array obtained by performing a numerical operation on text in adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to graphics in the graphical interface display content and a classification thereof. The preset data is set according to the voice recognition related content in advance, and the preset data can be used for carrying out array matching on the audio data input by the user, so that the cognition level of the user is objectively reflected.
In some embodiments, step S3 specifically includes:
S31: matching one group of text information in the text data with the text in the first array by using a match method of the regular expression, if so, comparing the matched text information with the corresponding element in the first array, judging whether the comparison result is the same, if so, successfully matching, otherwise, not matching;
S32: and repeating the step S31 to match all the text information of the text data in turn, and obtaining the comparison result of each text information.
And obtaining the correct error condition in the comparison result by performing one comparison between one group of text information in the text data and the text in the first array.
In some embodiments, the positioning position of the cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching. The user can be guided to finish the input of the audio data required by the display content of the graphical interface by changing the positioning position of the cursor, so that the accuracy of array matching and the finishing efficiency are improved.
In some embodiments, the preset data includes a second array, where the second array includes a one-dimensional array formed by characters corresponding to the audio-recording playing content or a one-dimensional array generated by reversely valued characters corresponding to the audio-recording playing content. The user finishes inputting the audio data according to the characters or the requirements corresponding to the record playing content, and then performs comparison, so as to evaluate the cognitive ability of the user according to the comparison result.
In some embodiments, step S3 specifically includes:
S31': judging whether the text data and the elements in the second array belong to the same type or not through the regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
S32': converting the extracted text information into an array through a split algorithm, checking through an evaluation method, judging whether the extracted text information belongs to elements in a second array, if so, judging whether the extracted text information is consistent with the positions of the elements in the second array, and if so, successfully matching.
And judging whether the text data is matched with the corresponding elements in the second array or not through analysis and matching, thereby judging whether the result of the text data is correct or not.
In a second aspect, an embodiment of the present application further proposes an audio data processing device based on cognitive assessment, including:
the audio data acquisition module is configured to acquire audio data output by a user according to preset voice recognition related content, and convert the audio data into text data through a voice recognition technology;
The content data conversion module is configured to obtain preset data generated by text conversion of the voice recognition related content; and
The comparison module is configured to compare the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
And the time data acquisition module is configured to acquire time data of the user in the process of completing the voice recognition related content and evaluate the cognitive ability of the user by combining the comparison result.
In a third aspect, an embodiment of the present application provides an electronic device, including: one or more processors; and storage means for storing one or more programs which, when executed by the one or more processors, cause the one or more processors to implement the functions of the system as described in any implementation of the first aspect.
In a fourth aspect, embodiments of the present application provide a computer readable storage medium having stored thereon a computer program which, when executed by a processor, implements a system as described in any of the implementations of the first aspect.
The invention discloses an audio data processing method and device based on cognitive evaluation, which are characterized in that audio data input by a user according to preset voice recognition related content are collected, and the audio data are converted into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result. The difficulty in evaluation of cognitive dysfunction can be effectively reduced through processing the audio data, and more intelligent, efficient and quick experience is brought to the whole cognitive evaluation process. And the data acquired by the user in the cognitive evaluation process are more diversified and accurate, so that the data can be recorded and evaluated in real time, and the accuracy of the cognitive evaluation is effectively improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings that are needed in the description of the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is an exemplary device frame pattern to which an embodiment of the present application may be applied;
FIG. 2 is a flow chart of a method for processing audio data based on cognitive assessment according to an embodiment of the present invention;
FIG. 3 is a flow chart of step S3 in one embodiment of a cognitive assessment based audio data processing method according to an embodiment of the present invention;
Fig. 4 is a flowchart illustrating a step S3 of another embodiment of a cognitive evaluation-based audio data processing method according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of an audio data processing device based on cognitive assessment according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of a computer device suitable for use in implementing an embodiment of the application.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail below with reference to the accompanying drawings, and it is apparent that the described embodiments are only some embodiments of the present invention, not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Fig. 1 shows an exemplary device architecture 100 of an audio data processing method based on cognitive assessment or an audio data processing device based on cognitive assessment, to which embodiments of the present application may be applied.
As shown in fig. 1, the apparatus architecture 100 may include terminal devices 101, 102, 103, a network 104, and a server 105. The network 104 is used as a medium to provide communication links between the terminal devices 101, 102, 103 and the server 105. The network 104 may include various connection types, such as wired, wireless communication links, or fiber optic cables, among others.
The user may interact with the server 105 via the network 104 using the terminal devices 101, 102, 103 to receive or send messages or the like. Various applications, such as a data processing class application, a file processing class application, and the like, may be installed on the terminal devices 101, 102, 103.
The terminal devices 101, 102, 103 may be hardware or software. When the terminal devices 101, 102, 103 are hardware, they may be various electronic devices including, but not limited to, smartphones, tablets, laptop and desktop computers, and the like. When the terminal devices 101, 102, 103 are software, they can be installed in the above-listed electronic devices. Which may be implemented as multiple software or software modules (e.g., software or software modules for providing distributed services) or as a single software or software module. The present invention is not particularly limited herein.
The server 105 may be a server providing various services, such as a background data processing server processing files or data uploaded by the terminal devices 101, 102, 103. The background data processing server can process the acquired file or data to generate a processing result.
It should be noted that, the method for processing audio data based on cognitive evaluation provided in the embodiment of the present application may be executed by the server 105, or may be executed by the terminal devices 101, 102, 103, and accordingly, the audio data processing device based on cognitive evaluation may be set in the server 105, or may be set in the terminal devices 101, 102, 103.
It should be understood that the number of terminal devices, networks and servers in fig. 1 is merely illustrative. There may be any number of terminal devices, networks, and servers, as desired for implementation. In the case where the processed data does not need to be acquired from a remote location, the above-described apparatus architecture may not include a network, but only a server or terminal device.
With continued reference to fig. 2, there is shown a method of processing audio data based on cognitive assessment provided in an embodiment in accordance with the application, the method comprising the steps of:
s1: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology;
s2: acquiring preset data generated by text conversion of voice recognition related content;
s3: comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
S4: and collecting time data of the user in the process of completing the voice recognition related content, and evaluating the cognitive ability of the user by combining the comparison result.
In a specific embodiment, the voice recognition related content comprises graphics or numbers, and the display mode of the voice recognition related content comprises graphic interface display content or recording and playing content. And displaying the content or recording and playing the content to show or guide the user to set forth the content of the preset part through the graphical interface.
In a specific embodiment, the preset data includes a first array, where the first array includes a one-dimensional array formed by text in the graphical interface display content, or a one-dimensional array obtained by performing numerical operation on text in the adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to graphics in the graphical interface display content and classification thereof. The text or the image displayed by the display content of the graphical interface can be obtained, and the audio data input by the user is subjected to array matching, so that the cognition level of the user is objectively reflected.
In a specific embodiment, as shown in fig. 3, step S3 specifically includes:
S31: matching one group of text information in the text data with the text in the first array by using a match method of the regular expression, if so, comparing the matched text information with the corresponding element in the first array, judging whether the comparison result is the same, if so, successfully matching, otherwise, not matching;
S32: and repeating the step S31 to match all the text information of the text data in turn, and obtaining the comparison result of each text information.
In a preferred embodiment, the positioning position of the cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching. The user can be guided to finish the input of the audio data required by the display content of the graphical interface by changing the positioning position of the cursor, so that the accuracy of array matching and the finishing efficiency are improved.
When the image displayed by the display content of the graphical interface is an indication graph, for example, the preset data is an arrow graph, the direction of the corresponding arrow graph can be converted into a corresponding array, for example, the graph arrow ∈ is identified as a text array [ "upper face", "lower face", "upper face", "lower face" ], and therefore the text array becomes a first array. After the user sees the voice indicated by the arrow graph through the external voice dictation control, the voice is converted into audio data and then further converted into text data. Then matching the upper or lower part in the text data, if the array length returned by the match method of the regular expression is greater than 0, indicating that the matching is successful, otherwise, not finishing the matching. Data other than above or below in the text data can be filtered out. And finally, comparing the matched array with the corresponding elements in the first array, judging whether the matched array is the same element, if so, indicating that the matching is successful, otherwise, judging that the matching is not successful. At this time, the text data and the elements in the text array and the sequence thereof can be compared, and the correct quantity of the comparison result can be obtained according to the comparison result. In the process, the positioning position of a cursor in the display content of the graphical interface is determined according to the matching completion degree of each value in the text data. At the beginning, the cursor is positioned at the first position in the image displayed by the display content of the graphical interface, and when the cursor is matched to the upper surface or the lower surface for the first time, the cursor moves down by one position. And (3) turning the corresponding cursor to the next image after each value of the text data corresponding to each image displayed by the graphical interface display content is matched. In the process, time data of audio data input by a user according to the display content of the graphical interface and correct error conditions between text data corresponding to the display content of the graphical interface and preset data can be acquired. For example, the user can read 3 times according to the graph arrow, record the time of the voice required by reading the graph arrow in each time and the correct error condition of the result in each time, record the average time of reading the upper side and the average time of reading the lower side in each time, calculate the time of the attention control ability according to the time of the last time and the average time of the last three times, and finally comprehensively evaluate the cognitive ability of the user. The method can also be used for data acquisition when the images displayed by the graphical interface display content are other graphics. Compared with the traditional cognitive assessment mode, the method can acquire more dimensionality data, so that the cognitive ability of the user can be judged more accurately.
When the image displayed by the graphical interface display content is a picture with a plurality of specific nouns, the preset data can also be a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphical interface display content or a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphical interface display content and the classification thereof. And thus may also be performed in the manner described above to collect corresponding data. For example [ [ "bird" ], [ "ship", "boat" ], [ "pineapple", "pineapple" ], [ "bunny", "white rabbit", "white rabbit" ] ], the following are resolved: when a rabbit graph appears in the display content of the graphical interface, after the user sees the voice sent by the graph, the voice is converted into audio data and further converted into text data through an external voice dictation control, then the text data and each element in [ "little rabbit", "little white rabbit", "white rabbit" ] are circularly traversed, whether the text data is consistent with characters in preset data or not is determined, and finally whether the text data corresponding to each of a plurality of concrete noun pictures is consistent with the corresponding graph element in the two-bit array or not is determined.
When the preset data is a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphic interface display content and the classifications thereof, for example: daily necessities [ [ "writing brush", "paper", "chair" ] ], fruits [ [ "apple", "pear", "bergamot pear", "snow pear" ] ], animals [ [ "duck", "turkey", "garter" ] ]. The same can be performed in the manner described above to acquire corresponding data. When the image displayed by the display content of the graphical interface is a number, the preset data may be a one-dimensional array obtained by performing a number operation on the characters in the display content of the adjacent graphical interface, and in a preferred embodiment, the number operation is an addition. For example, if [ "14", "24" ] exists in the preset data, 5 is displayed in the first graphical interface display content, 9 is displayed in the second graphical interface display content, and the sum of the two is required to be calculated, the method is adopted to determine whether the value and the position of the corresponding array in the preset data are consistent with the value in the audio data input by the user.
In a specific embodiment, the preset data includes a second array, where the second array includes a one-dimensional array formed by characters corresponding to the audio recording and playing content or a one-dimensional array generated by reversely valued characters corresponding to the audio recording and playing content. The user finishes inputting the audio data according to the characters or the requirements corresponding to the record playing content, and then performs comparison, so as to evaluate the cognitive ability of the user according to the comparison result and the acquired time data.
In a specific embodiment, as shown in fig. 4, step S3 specifically includes:
S31': judging whether the text data and the elements in the second array belong to the same type or not through the regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
S32': converting the extracted text information into an array through a split algorithm, checking through an evaluation method, judging whether the extracted text information belongs to elements in a second array, if so, judging whether the extracted text information is consistent with the positions of the elements in the second array, and if so, successfully matching.
And judging whether the text data is matched with the corresponding elements in the second array or not through analysis and matching, thereby judging whether the result of the text data is correct or not. In particular embodiments, the speech recognition technique includes a random model approach or an artificial neural network approach. The voice recognition technology is mature, and the recognition efficiency is high.
When the second array in the preset data comprises a one-dimensional array formed by characters corresponding to the record playing content. The preset data can be digital, for example, set as [ "742", "285", "3419" ], after the record playing content is played, the playing times are recorded, and after the user sees the voice sent by the graph, the voice is converted into audio data and further converted into text data through an external voice dictation control. Firstly judging whether one of the numbers [0-9] exists in the text data through the regular expression, and if so, extracting the numbers in the text data through the match method of the regular expression. And then converting the extracted numbers into an array through a split algorithm, checking through an evaluation method, judging whether the extracted numbers belong to the preset numbers in the second array, if so, judging whether the extracted numbers are consistent with the positions of the numbers in the second array, and if so, successfully matching. Similarly, when the second array in the preset data includes a one-dimensional array generated by reversely valued the text corresponding to the record playing content. And generating a second array by reversely taking the value through the reserve based on the number, and then executing in the same way as the above to acquire data.
With further reference to fig. 5, as an implementation of the method shown in the foregoing figures, the present application provides an embodiment of a graphics rendering apparatus for cognitive assessment, where the apparatus embodiment corresponds to the method embodiment shown in fig. 2, and the apparatus may be specifically applied to various electronic devices.
As shown in fig. 5, the audio data processing apparatus based on cognitive evaluation of the present embodiment includes:
The audio data acquisition module 1 is configured to acquire audio data output by a user according to preset voice recognition related content, and convert the audio data into text data through a voice recognition technology;
A content data conversion module 2 configured to obtain preset data generated by text conversion of the voice recognition related content; and
The comparison module 3 is configured to compare the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
And the time data acquisition module 4 is configured to acquire time data of the user in the process of completing the voice recognition related content and evaluate the cognitive ability of the user by combining the comparison result.
In a specific embodiment, the voice recognition related content comprises graphics or numbers, and the display mode of the voice recognition related content comprises graphic interface display content or recording and playing content. And displaying the content or recording and playing the content to show or guide the user to set forth the content of the preset part through the graphical interface.
In a specific embodiment, the preset data includes a first array, where the first array includes a one-dimensional array formed by text in the graphical interface display content, or a one-dimensional array obtained by performing numerical operation on text in the adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to graphics in the graphical interface display content and classification thereof. The text or the image displayed by the display content of the graphical interface can be obtained, and the audio data input by the user is subjected to array matching, so that the cognition level of the user is objectively reflected.
In a specific embodiment, the comparison module 3 specifically includes:
The first matching module (not shown in the figure) is used for matching one group of text information in the text data with the text in the first array through the match method of the regular expression, if so, the matched text information is compared with the corresponding element in the first array, whether the comparison result is the same is judged, if so, the matching is successful, otherwise, the matching is not performed;
And the circular matching module (not shown in the figure) is used for repeatedly executing the first matching module (not shown in the figure) to sequentially match all the text information of the text data, and obtaining the comparison result of each text information.
In a preferred embodiment, the positioning position of the cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching. The user can be guided to finish the input of the audio data required by the display content of the graphical interface by changing the positioning position of the cursor, so that the accuracy of array matching and the finishing efficiency are improved.
In a specific embodiment, the preset data includes a second array, where the second array includes a one-dimensional array formed by characters corresponding to the audio recording and playing content or a one-dimensional array generated by reversely valued characters corresponding to the audio recording and playing content. The user finishes inputting the audio data according to the characters or the requirements corresponding to the record playing content, and then performs comparison, so as to evaluate the cognitive ability of the user according to the comparison result and the acquired time data.
In a specific embodiment, the comparison module 3 may specifically further include:
the data extraction module (not shown in the figure) is used for judging whether the text data and the elements in the second array belong to the same type through the regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
and the second matching module (not shown in the figure) is used for converting the extracted text information into an array through a split algorithm and checking through an evaluation method, judging whether the extracted text information belongs to the elements in the second array, if so, judging whether the extracted text information is consistent with the positions of the elements in the second array, and if so, successfully matching.
The invention discloses an audio data processing method and device based on cognitive evaluation, which are characterized in that audio data input by a user according to preset voice recognition related content are collected, and the audio data are converted into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result. The difficulty in evaluation of cognitive dysfunction can be effectively reduced through processing the audio data, and more intelligent, efficient and quick experience is brought to the whole cognitive evaluation process. And the data acquired by the user in the cognitive evaluation process are more diversified and accurate, so that the data can be recorded and evaluated in real time, and the accuracy of the cognitive evaluation is effectively improved.
Referring now to fig. 6, there is illustrated a schematic diagram of a computer apparatus 600 suitable for use in an electronic device (e.g., a server or terminal device as illustrated in fig. 1) for implementing an embodiment of the present application. The electronic device shown in fig. 6 is only an example and should not be construed as limiting the functionality and scope of use of the embodiments of the application.
As shown in fig. 6, the computer apparatus 600 includes a Central Processing Unit (CPU) 601 and a Graphics Processor (GPU) 602, which can perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 603 or a program loaded from a storage section 609 into a Random Access Memory (RAM) 606. In the RAM604, various programs and data required for the operation of the apparatus 600 are also stored. The CPU 601, GPU602, ROM 603, and RAM604 are connected to each other through a bus 605. An input/output (I/O) interface 606 is also connected to the bus 605.
The following components are connected to the I/O interface 606: an input portion 607 including a keyboard, a mouse, and the like; an output portion 608 including a speaker, such as a Liquid Crystal Display (LCD), etc.; a storage portion 609 including a hard disk and the like; and a communication section 610 including a network interface card such as a LAN card, a modem, or the like. The communication section 610 performs communication processing via a network such as the internet. The drive 611 may also be connected to the I/O interface 606 as needed. A removable medium 612 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is mounted on the drive 611 as necessary, so that a computer program read out therefrom is mounted into the storage section 609 as necessary.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method shown in the flowcharts. In such embodiments, the computer program may be downloaded and installed from a network via the communication portion 610, and/or installed from the removable medium 612. The above-described functions defined in the method of the present application are performed when the computer program is executed by a Central Processing Unit (CPU) 601 and a Graphics Processor (GPU) 602.
It should be noted that the computer readable medium according to the present application may be a computer readable signal medium or a computer readable medium, or any combination of the two. The computer readable medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor apparatus, device, or means, or a combination of any of the foregoing. More specific examples of the computer-readable medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution apparatus, device, or apparatus. In the present application, however, the computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may be any computer readable medium that is not a computer readable medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution apparatus, device, or apparatus. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wire, fiber optic cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations of the present application may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based devices which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The modules involved in the embodiments of the present application may be implemented in software or in hardware. The described modules may also be provided in a processor.
As another aspect, the present application also provides a computer-readable medium that may be contained in the electronic device described in the above embodiment; or may exist alone without being incorporated into the electronic device. The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result.
A description of the technical principles applied. It will be appreciated by persons skilled in the art that the scope of the application referred to in the present application is not limited to the specific combinations of the technical features described above, but also covers other technical features formed by any combination of the technical features described above or their equivalents without departing from the inventive concept described above. Such as the above-mentioned features and the technical features disclosed in the present application (but not limited to) having similar functions are replaced with each other.
Claims (7)
1. A method for processing audio data based on cognitive assessment, comprising the steps of:
s1: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology;
s2: acquiring preset data generated by text conversion of the voice recognition related content;
s3: comparing the text data with the preset data through a regular expression matching algorithm to obtain a comparison result; and
S4: collecting time data of a user in the process of completing the voice recognition related content, and evaluating the cognitive ability of the user by combining the comparison result;
The preset data comprises a first array, wherein the first array comprises a one-dimensional array formed by corresponding characters in the graphical interface display content, or a one-dimensional array obtained by carrying out numerical operation on the characters in the adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphical interface display content and classification thereof; the preset data comprises a second array, wherein the second array comprises a one-dimensional array formed by characters corresponding to the recording and playing contents or a one-dimensional array generated by reversely taking the characters corresponding to the recording and playing contents; the step S3 specifically includes:
S31: matching one group of text information in the text data with the text in the first array by using a match method of a regular expression, if so, comparing the matched text information with the corresponding element in the first array, judging whether the comparison result is the same, if so, successfully matching, otherwise, not matching;
S32: and repeating the step S31 to match all the text information of the text data in turn, and obtaining the comparison result of each text information.
2. The cognitive assessment-based audio data processing method according to claim 1, wherein the voice recognition related content comprises graphics or numbers, and the voice recognition related content is displayed in a manner comprising displaying content through a graphical interface or playing content through a sound recording.
3. The cognitive assessment-based audio data processing method according to claim 2, wherein the positioning position of a cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching.
4. The method for processing audio data based on cognitive assessment according to claim 3, wherein the step S3 specifically comprises:
s31': judging whether the text data and the elements in the second array belong to the same type or not through a regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
s32': converting the extracted text information into an array through a split algorithm, checking through an evaluation method, judging whether the extracted text information belongs to the element in the second array, if so, judging whether the extracted text information is consistent with the position of the element in the second array, and if so, successfully matching.
5. An audio data processing device based on cognitive assessment, comprising:
the audio data acquisition module is configured to acquire audio data output by a user according to preset voice recognition related content, and convert the audio data into text data through a voice recognition technology;
The content data conversion module is configured to obtain preset data generated by text conversion of the voice recognition related content; and
The comparison module is configured to compare the text data with the preset data through a regular expression matching algorithm to obtain a comparison result; and
And the time data acquisition module is configured to acquire time data of a user in the process of completing the voice recognition related content and is used for evaluating the cognitive ability of the user in combination with the comparison result.
6. An electronic device, comprising:
one or more processors;
Storage means for storing one or more programs,
When executed by the one or more processors, causes the one or more processors to implement the method of any of claims 1-4.
7. A computer readable storage medium, on which a computer program is stored, characterized in that the program, when being executed by a processor, implements the method according to any of claims 1-4.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010988651.1A CN113539253B (en) | 2020-09-18 | 2020-09-18 | Audio data processing method and device based on cognitive assessment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010988651.1A CN113539253B (en) | 2020-09-18 | 2020-09-18 | Audio data processing method and device based on cognitive assessment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113539253A CN113539253A (en) | 2021-10-22 |
CN113539253B true CN113539253B (en) | 2024-05-14 |
Family
ID=78094284
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010988651.1A Active CN113539253B (en) | 2020-09-18 | 2020-09-18 | Audio data processing method and device based on cognitive assessment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113539253B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115798718B (en) * | 2022-11-24 | 2024-03-29 | 广州市第一人民医院(广州消化疾病中心、广州医科大学附属市一人民医院、华南理工大学附属第二医院) | Cognitive test evaluation method and system |
CN116048282B (en) * | 2023-03-06 | 2023-08-04 | 中国医学科学院生物医学工程研究所 | Data processing method, system, device, equipment and storage medium |
Citations (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6278996B1 (en) * | 1997-03-31 | 2001-08-21 | Brightware, Inc. | System and method for message process and response |
WO2002037223A2 (en) * | 2000-11-06 | 2002-05-10 | Invention Machine Corporation | Computer based integrated text and graphic document analysis |
JP2004184535A (en) * | 2002-11-29 | 2004-07-02 | Fujitsu Ltd | Device and method for speech recognition |
RU2253365C1 (en) * | 2003-11-17 | 2005-06-10 | Государственное образовательное учреждение высшего профессионального образования Московская медицинская академия им. И.М. Сеченова МЗ РФ | Psycholinguistic method for diagnosing neurotic disorders |
KR20070019596A (en) * | 2005-08-12 | 2007-02-15 | 캐논 가부시끼가이샤 | Information processing method and information processing device |
KR20090000662A (en) * | 2007-03-16 | 2009-01-08 | 장성옥 | Speech studying game and system using the game |
CA2820599A1 (en) * | 2010-11-24 | 2012-05-31 | Digital Artefacts, Llc | Systems and methods to assess cognitive function |
CN103251418A (en) * | 2013-06-05 | 2013-08-21 | 清华大学 | Image cognition psychoanalysis system |
CN103400579A (en) * | 2013-08-04 | 2013-11-20 | 徐华 | Voice recognition system and construction method |
CN104021786A (en) * | 2014-05-15 | 2014-09-03 | 北京中科汇联信息技术有限公司 | Speech recognition method and speech recognition device |
KR101538317B1 (en) * | 2014-02-20 | 2015-07-29 | ㈜빅스바이트 | An automatic evaluation system for English literacy |
CN106446165A (en) * | 2016-09-26 | 2017-02-22 | 厦门吉信德宠物用品有限公司 | Big data processing based identification method |
CN108846119A (en) * | 2018-06-27 | 2018-11-20 | 清远墨墨教育科技有限公司 | A kind of arrangement method, storage device and the mobile terminal of word cognition degree |
CN109222882A (en) * | 2018-10-08 | 2019-01-18 | 浙江工业大学 | A kind of reading visual acuity test system and method |
CN109344231A (en) * | 2018-10-31 | 2019-02-15 | 广东小天才科技有限公司 | Method and system for completing corpus of semantic deformity |
CN109407946A (en) * | 2018-09-11 | 2019-03-01 | 昆明理工大学 | Graphical interfaces target selecting method based on speech recognition |
CN109933671A (en) * | 2019-01-31 | 2019-06-25 | 平安科技(深圳)有限公司 | Construct method, apparatus, computer equipment and the storage medium of personal knowledge map |
CN110473605A (en) * | 2018-05-09 | 2019-11-19 | 深圳市前海安测信息技术有限公司 | Alzheimer Disease patient figure cognitive ability assessment system and method |
CN111295141A (en) * | 2017-11-02 | 2020-06-16 | 松下知识产权经营株式会社 | Cognitive function evaluation device, cognitive function evaluation system, cognitive function evaluation method, and program |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004246184A (en) * | 2003-02-14 | 2004-09-02 | Eigyotatsu Kofun Yugenkoshi | Language learning system and method with visualized pronunciation suggestion |
US7429108B2 (en) * | 2005-11-05 | 2008-09-30 | Outland Research, Llc | Gaze-responsive interface to enhance on-screen user reading tasks |
JP4811507B2 (en) * | 2009-08-25 | 2011-11-09 | コニカミノルタビジネステクノロジーズ株式会社 | Image processing system, image processing apparatus, and information processing apparatus |
US9471715B2 (en) * | 2013-03-31 | 2016-10-18 | International Business Machines Corporation | Accelerated regular expression evaluation using positional information |
US11017774B2 (en) * | 2019-02-04 | 2021-05-25 | International Business Machines Corporation | Cognitive audio classifier |
-
2020
- 2020-09-18 CN CN202010988651.1A patent/CN113539253B/en active Active
Patent Citations (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6278996B1 (en) * | 1997-03-31 | 2001-08-21 | Brightware, Inc. | System and method for message process and response |
WO2002037223A2 (en) * | 2000-11-06 | 2002-05-10 | Invention Machine Corporation | Computer based integrated text and graphic document analysis |
JP2004184535A (en) * | 2002-11-29 | 2004-07-02 | Fujitsu Ltd | Device and method for speech recognition |
RU2253365C1 (en) * | 2003-11-17 | 2005-06-10 | Государственное образовательное учреждение высшего профессионального образования Московская медицинская академия им. И.М. Сеченова МЗ РФ | Psycholinguistic method for diagnosing neurotic disorders |
KR20070019596A (en) * | 2005-08-12 | 2007-02-15 | 캐논 가부시끼가이샤 | Information processing method and information processing device |
KR20090000662A (en) * | 2007-03-16 | 2009-01-08 | 장성옥 | Speech studying game and system using the game |
CA2820599A1 (en) * | 2010-11-24 | 2012-05-31 | Digital Artefacts, Llc | Systems and methods to assess cognitive function |
CN103251418A (en) * | 2013-06-05 | 2013-08-21 | 清华大学 | Image cognition psychoanalysis system |
CN103400579A (en) * | 2013-08-04 | 2013-11-20 | 徐华 | Voice recognition system and construction method |
KR101538317B1 (en) * | 2014-02-20 | 2015-07-29 | ㈜빅스바이트 | An automatic evaluation system for English literacy |
CN104021786A (en) * | 2014-05-15 | 2014-09-03 | 北京中科汇联信息技术有限公司 | Speech recognition method and speech recognition device |
CN106446165A (en) * | 2016-09-26 | 2017-02-22 | 厦门吉信德宠物用品有限公司 | Big data processing based identification method |
CN111295141A (en) * | 2017-11-02 | 2020-06-16 | 松下知识产权经营株式会社 | Cognitive function evaluation device, cognitive function evaluation system, cognitive function evaluation method, and program |
CN110473605A (en) * | 2018-05-09 | 2019-11-19 | 深圳市前海安测信息技术有限公司 | Alzheimer Disease patient figure cognitive ability assessment system and method |
CN108846119A (en) * | 2018-06-27 | 2018-11-20 | 清远墨墨教育科技有限公司 | A kind of arrangement method, storage device and the mobile terminal of word cognition degree |
CN109407946A (en) * | 2018-09-11 | 2019-03-01 | 昆明理工大学 | Graphical interfaces target selecting method based on speech recognition |
CN109222882A (en) * | 2018-10-08 | 2019-01-18 | 浙江工业大学 | A kind of reading visual acuity test system and method |
CN109344231A (en) * | 2018-10-31 | 2019-02-15 | 广东小天才科技有限公司 | Method and system for completing corpus of semantic deformity |
CN109933671A (en) * | 2019-01-31 | 2019-06-25 | 平安科技(深圳)有限公司 | Construct method, apparatus, computer equipment and the storage medium of personal knowledge map |
Non-Patent Citations (5)
Title |
---|
产品认知语境的类型及其应用;傅桂涛;潘荣;陈国东;陈思宇;;包装工程(第08期);全文 * |
元认知策略与大学英语阅读能力的提高;王莉;毕凤春;;高等农业教育;20060428(第04期);全文 * |
军人个体认知能力综合评估研究;安改红;王静;陈学伟;李超;陈佩延;安芳红;张文正;李正东;袭著革;马强;;人民军医(第01期);全文 * |
在校医学生对于老年健康的认知现状分析;高奎;张丽娜;涂虹;刘晓微;李婷;肖雄;;中国医药导报(第03期);全文 * |
基于PASS理论的五项认知能力与中文阅读理解能力的关系研究;张晴;刘巧云;杜晓新;黄昭鸣;祝亚平;;中国儿童保健杂志;20180105(第02期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN113539253A (en) | 2021-10-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110457432B (en) | Interview scoring method, interview scoring device, interview scoring equipment and interview scoring storage medium | |
CN110781413B (en) | Method and device for determining interest points, storage medium and electronic equipment | |
CN113539253B (en) | Audio data processing method and device based on cognitive assessment | |
CN111709630A (en) | Voice quality inspection method, device, equipment and storage medium | |
CN107481715B (en) | Method and apparatus for generating information | |
CN112116903B (en) | Speech synthesis model generation method and device, storage medium and electronic equipment | |
CN111651497A (en) | User label mining method and device, storage medium and electronic equipment | |
US11024329B2 (en) | Word repetition in separate conversations for detecting a sign of cognitive decline | |
CN114138960A (en) | User intention identification method, device, equipment and medium | |
CN114140814A (en) | Emotion recognition capability training method and device and electronic equipment | |
CN109101956B (en) | Method and apparatus for processing image | |
Shanthi et al. | An integrated approach for mental health assessment using emotion analysis and scales | |
CN117828355A (en) | Emotion quantitative model training method and emotion quantitative method based on multi-modal information | |
CN117786600A (en) | Cognitive evaluation method, device, electronic equipment and storage medium | |
KR20200084816A (en) | Method, apparatus and computer program for analyzing new contents for solving cold start | |
CN114240250A (en) | Intelligent management method and system for vocational evaluation | |
CN113361282B (en) | Information processing method and device | |
CN114613350A (en) | Test method, test device, electronic equipment and storage medium | |
CN111949860B (en) | Method and apparatus for generating a relevance determination model | |
CN109344289B (en) | Method and apparatus for generating information | |
CN112131378A (en) | Method and device for identifying categories of civil problems and electronic equipment | |
CN113220849B (en) | Customer service personnel emotion dispersion scheme searching method, electronic equipment and storage medium | |
CN115440198B (en) | Method, apparatus, computer device and storage medium for converting mixed audio signal | |
CN114219369B (en) | Prediction model training method and device, and user category prediction method and device | |
CN118467709B (en) | Evaluation method, device, medium and computer program product for visual question-answering task |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |