CN113539253B - Audio data processing method and device based on cognitive assessment - Google Patents

Audio data processing method and device based on cognitive assessment Download PDF

Info

Publication number
CN113539253B
CN113539253B CN202010988651.1A CN202010988651A CN113539253B CN 113539253 B CN113539253 B CN 113539253B CN 202010988651 A CN202010988651 A CN 202010988651A CN 113539253 B CN113539253 B CN 113539253B
Authority
CN
China
Prior art keywords
data
text
array
audio data
voice recognition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010988651.1A
Other languages
Chinese (zh)
Other versions
CN113539253A (en
Inventor
雷彪
姚凌翔
余滢
林腾达
肖衍
陈晓晓
章伏清
吴秀兰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Hejia Jiannao Intelligent Technology Co ltd
Original Assignee
Xiamen Hejia Jiannao Intelligent Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen Hejia Jiannao Intelligent Technology Co ltd filed Critical Xiamen Hejia Jiannao Intelligent Technology Co ltd
Priority to CN202010988651.1A priority Critical patent/CN113539253B/en
Publication of CN113539253A publication Critical patent/CN113539253A/en
Application granted granted Critical
Publication of CN113539253B publication Critical patent/CN113539253B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/194Calculation of difference between files
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention discloses an audio data processing method and device based on cognitive evaluation, which are characterized in that audio data input by a user according to preset voice recognition related content are collected, and the audio data are converted into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result. The difficulty in evaluation of cognitive dysfunction can be effectively reduced through processing the audio data, and more intelligent, efficient and quick experience is brought to the whole cognitive evaluation process. And the data acquired by the user in the cognitive evaluation process are more diversified and accurate, so that the data can be recorded and evaluated in real time, and the accuracy of the cognitive evaluation is effectively improved.

Description

Audio data processing method and device based on cognitive assessment
Technical Field
The invention relates to the technical field of computers, in particular to an audio data processing method and device based on cognitive evaluation.
Background
At present, cognitive dysfunction is one of important diseases affecting the health and life quality of middle-aged and elderly people, and the cognitive dysfunction can be expressed not only by memory dysfunction, aphasia, agnosia, visual space dysfunction and the like, but also by affective behavior disorders such as anxiety, depression, agitation, impulse and the like, and the affective behavior disorders are also causes of disability of patients, so that a heavy burden is brought to society and families. Typically, doctors evaluate cognitive functions of patients during inquiry by conventional inquiry and paper scales, and doctors evaluate the cognitive functions mainly according to the performance of the patients and test results of the paper scales. The process has the advantages of huge workload, very much time and low efficiency, thus bringing great difficulty to the evaluation of the whole cognitive dysfunction, and needing to carry out more intelligent, efficient and rapid improvement on the whole cognitive evaluation process so as to facilitate accurate evaluation on patients.
In the prior art, doctors communicate with patients through a conventional inquiry means, judge according to the answers of the patients or answer according to questions in a paper scale, but voice answer information which can be received by the doctors is only received and judged instantaneously, so that the one-sided and subjective comparison is carried out, and more accurate recording and judging standards are lacked.
In view of this, it is important to propose a method and apparatus for audio data processing based on cognitive assessment.
Disclosure of Invention
The problems of one-sided and subjective comparison, lack of more accurate record and judgment standards and the like in the cognitive level evaluation process are solved. An object of an embodiment of the present application is to provide an audio data processing method and apparatus based on cognitive evaluation to solve the technical problems mentioned in the background section above.
In a first aspect, an embodiment of the present application provides an audio data processing method based on cognitive evaluation, including the steps of:
s1: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology;
s2: acquiring preset data generated by text conversion of voice recognition related content;
s3: comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
S4: and collecting time data of the user in the process of completing the voice recognition related content, and evaluating the cognitive ability of the user by combining the comparison result.
In some embodiments, the speech recognition related content comprises graphics or numbers, and the presentation mode of the speech recognition related content comprises displaying content or recording and playing content by a graphical interface. And displaying the content or recording and playing the content to show or guide the user to set forth the content of the preset part through the graphical interface.
In some embodiments, the preset data includes a first array, where the first array includes a one-dimensional array formed by text in the graphical interface display content, or a one-dimensional array obtained by performing a numerical operation on text in adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to graphics in the graphical interface display content and a classification thereof. The preset data is set according to the voice recognition related content in advance, and the preset data can be used for carrying out array matching on the audio data input by the user, so that the cognition level of the user is objectively reflected.
In some embodiments, step S3 specifically includes:
S31: matching one group of text information in the text data with the text in the first array by using a match method of the regular expression, if so, comparing the matched text information with the corresponding element in the first array, judging whether the comparison result is the same, if so, successfully matching, otherwise, not matching;
S32: and repeating the step S31 to match all the text information of the text data in turn, and obtaining the comparison result of each text information.
And obtaining the correct error condition in the comparison result by performing one comparison between one group of text information in the text data and the text in the first array.
In some embodiments, the positioning position of the cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching. The user can be guided to finish the input of the audio data required by the display content of the graphical interface by changing the positioning position of the cursor, so that the accuracy of array matching and the finishing efficiency are improved.
In some embodiments, the preset data includes a second array, where the second array includes a one-dimensional array formed by characters corresponding to the audio-recording playing content or a one-dimensional array generated by reversely valued characters corresponding to the audio-recording playing content. The user finishes inputting the audio data according to the characters or the requirements corresponding to the record playing content, and then performs comparison, so as to evaluate the cognitive ability of the user according to the comparison result.
In some embodiments, step S3 specifically includes:
S31': judging whether the text data and the elements in the second array belong to the same type or not through the regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
S32': converting the extracted text information into an array through a split algorithm, checking through an evaluation method, judging whether the extracted text information belongs to elements in a second array, if so, judging whether the extracted text information is consistent with the positions of the elements in the second array, and if so, successfully matching.
And judging whether the text data is matched with the corresponding elements in the second array or not through analysis and matching, thereby judging whether the result of the text data is correct or not.
In a second aspect, an embodiment of the present application further proposes an audio data processing device based on cognitive assessment, including:
the audio data acquisition module is configured to acquire audio data output by a user according to preset voice recognition related content, and convert the audio data into text data through a voice recognition technology;
The content data conversion module is configured to obtain preset data generated by text conversion of the voice recognition related content; and
The comparison module is configured to compare the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
And the time data acquisition module is configured to acquire time data of the user in the process of completing the voice recognition related content and evaluate the cognitive ability of the user by combining the comparison result.
In a third aspect, an embodiment of the present application provides an electronic device, including: one or more processors; and storage means for storing one or more programs which, when executed by the one or more processors, cause the one or more processors to implement the functions of the system as described in any implementation of the first aspect.
In a fourth aspect, embodiments of the present application provide a computer readable storage medium having stored thereon a computer program which, when executed by a processor, implements a system as described in any of the implementations of the first aspect.
The invention discloses an audio data processing method and device based on cognitive evaluation, which are characterized in that audio data input by a user according to preset voice recognition related content are collected, and the audio data are converted into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result. The difficulty in evaluation of cognitive dysfunction can be effectively reduced through processing the audio data, and more intelligent, efficient and quick experience is brought to the whole cognitive evaluation process. And the data acquired by the user in the cognitive evaluation process are more diversified and accurate, so that the data can be recorded and evaluated in real time, and the accuracy of the cognitive evaluation is effectively improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings that are needed in the description of the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is an exemplary device frame pattern to which an embodiment of the present application may be applied;
FIG. 2 is a flow chart of a method for processing audio data based on cognitive assessment according to an embodiment of the present invention;
FIG. 3 is a flow chart of step S3 in one embodiment of a cognitive assessment based audio data processing method according to an embodiment of the present invention;
Fig. 4 is a flowchart illustrating a step S3 of another embodiment of a cognitive evaluation-based audio data processing method according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of an audio data processing device based on cognitive assessment according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of a computer device suitable for use in implementing an embodiment of the application.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail below with reference to the accompanying drawings, and it is apparent that the described embodiments are only some embodiments of the present invention, not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Fig. 1 shows an exemplary device architecture 100 of an audio data processing method based on cognitive assessment or an audio data processing device based on cognitive assessment, to which embodiments of the present application may be applied.
As shown in fig. 1, the apparatus architecture 100 may include terminal devices 101, 102, 103, a network 104, and a server 105. The network 104 is used as a medium to provide communication links between the terminal devices 101, 102, 103 and the server 105. The network 104 may include various connection types, such as wired, wireless communication links, or fiber optic cables, among others.
The user may interact with the server 105 via the network 104 using the terminal devices 101, 102, 103 to receive or send messages or the like. Various applications, such as a data processing class application, a file processing class application, and the like, may be installed on the terminal devices 101, 102, 103.
The terminal devices 101, 102, 103 may be hardware or software. When the terminal devices 101, 102, 103 are hardware, they may be various electronic devices including, but not limited to, smartphones, tablets, laptop and desktop computers, and the like. When the terminal devices 101, 102, 103 are software, they can be installed in the above-listed electronic devices. Which may be implemented as multiple software or software modules (e.g., software or software modules for providing distributed services) or as a single software or software module. The present invention is not particularly limited herein.
The server 105 may be a server providing various services, such as a background data processing server processing files or data uploaded by the terminal devices 101, 102, 103. The background data processing server can process the acquired file or data to generate a processing result.
It should be noted that, the method for processing audio data based on cognitive evaluation provided in the embodiment of the present application may be executed by the server 105, or may be executed by the terminal devices 101, 102, 103, and accordingly, the audio data processing device based on cognitive evaluation may be set in the server 105, or may be set in the terminal devices 101, 102, 103.
It should be understood that the number of terminal devices, networks and servers in fig. 1 is merely illustrative. There may be any number of terminal devices, networks, and servers, as desired for implementation. In the case where the processed data does not need to be acquired from a remote location, the above-described apparatus architecture may not include a network, but only a server or terminal device.
With continued reference to fig. 2, there is shown a method of processing audio data based on cognitive assessment provided in an embodiment in accordance with the application, the method comprising the steps of:
s1: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology;
s2: acquiring preset data generated by text conversion of voice recognition related content;
s3: comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
S4: and collecting time data of the user in the process of completing the voice recognition related content, and evaluating the cognitive ability of the user by combining the comparison result.
In a specific embodiment, the voice recognition related content comprises graphics or numbers, and the display mode of the voice recognition related content comprises graphic interface display content or recording and playing content. And displaying the content or recording and playing the content to show or guide the user to set forth the content of the preset part through the graphical interface.
In a specific embodiment, the preset data includes a first array, where the first array includes a one-dimensional array formed by text in the graphical interface display content, or a one-dimensional array obtained by performing numerical operation on text in the adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to graphics in the graphical interface display content and classification thereof. The text or the image displayed by the display content of the graphical interface can be obtained, and the audio data input by the user is subjected to array matching, so that the cognition level of the user is objectively reflected.
In a specific embodiment, as shown in fig. 3, step S3 specifically includes:
S31: matching one group of text information in the text data with the text in the first array by using a match method of the regular expression, if so, comparing the matched text information with the corresponding element in the first array, judging whether the comparison result is the same, if so, successfully matching, otherwise, not matching;
S32: and repeating the step S31 to match all the text information of the text data in turn, and obtaining the comparison result of each text information.
In a preferred embodiment, the positioning position of the cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching. The user can be guided to finish the input of the audio data required by the display content of the graphical interface by changing the positioning position of the cursor, so that the accuracy of array matching and the finishing efficiency are improved.
When the image displayed by the display content of the graphical interface is an indication graph, for example, the preset data is an arrow graph, the direction of the corresponding arrow graph can be converted into a corresponding array, for example, the graph arrow ∈ is identified as a text array [ "upper face", "lower face", "upper face", "lower face" ], and therefore the text array becomes a first array. After the user sees the voice indicated by the arrow graph through the external voice dictation control, the voice is converted into audio data and then further converted into text data. Then matching the upper or lower part in the text data, if the array length returned by the match method of the regular expression is greater than 0, indicating that the matching is successful, otherwise, not finishing the matching. Data other than above or below in the text data can be filtered out. And finally, comparing the matched array with the corresponding elements in the first array, judging whether the matched array is the same element, if so, indicating that the matching is successful, otherwise, judging that the matching is not successful. At this time, the text data and the elements in the text array and the sequence thereof can be compared, and the correct quantity of the comparison result can be obtained according to the comparison result. In the process, the positioning position of a cursor in the display content of the graphical interface is determined according to the matching completion degree of each value in the text data. At the beginning, the cursor is positioned at the first position in the image displayed by the display content of the graphical interface, and when the cursor is matched to the upper surface or the lower surface for the first time, the cursor moves down by one position. And (3) turning the corresponding cursor to the next image after each value of the text data corresponding to each image displayed by the graphical interface display content is matched. In the process, time data of audio data input by a user according to the display content of the graphical interface and correct error conditions between text data corresponding to the display content of the graphical interface and preset data can be acquired. For example, the user can read 3 times according to the graph arrow, record the time of the voice required by reading the graph arrow in each time and the correct error condition of the result in each time, record the average time of reading the upper side and the average time of reading the lower side in each time, calculate the time of the attention control ability according to the time of the last time and the average time of the last three times, and finally comprehensively evaluate the cognitive ability of the user. The method can also be used for data acquisition when the images displayed by the graphical interface display content are other graphics. Compared with the traditional cognitive assessment mode, the method can acquire more dimensionality data, so that the cognitive ability of the user can be judged more accurately.
When the image displayed by the graphical interface display content is a picture with a plurality of specific nouns, the preset data can also be a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphical interface display content or a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphical interface display content and the classification thereof. And thus may also be performed in the manner described above to collect corresponding data. For example [ [ "bird" ], [ "ship", "boat" ], [ "pineapple", "pineapple" ], [ "bunny", "white rabbit", "white rabbit" ] ], the following are resolved: when a rabbit graph appears in the display content of the graphical interface, after the user sees the voice sent by the graph, the voice is converted into audio data and further converted into text data through an external voice dictation control, then the text data and each element in [ "little rabbit", "little white rabbit", "white rabbit" ] are circularly traversed, whether the text data is consistent with characters in preset data or not is determined, and finally whether the text data corresponding to each of a plurality of concrete noun pictures is consistent with the corresponding graph element in the two-bit array or not is determined.
When the preset data is a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphic interface display content and the classifications thereof, for example: daily necessities [ [ "writing brush", "paper", "chair" ] ], fruits [ [ "apple", "pear", "bergamot pear", "snow pear" ] ], animals [ [ "duck", "turkey", "garter" ] ]. The same can be performed in the manner described above to acquire corresponding data. When the image displayed by the display content of the graphical interface is a number, the preset data may be a one-dimensional array obtained by performing a number operation on the characters in the display content of the adjacent graphical interface, and in a preferred embodiment, the number operation is an addition. For example, if [ "14", "24" ] exists in the preset data, 5 is displayed in the first graphical interface display content, 9 is displayed in the second graphical interface display content, and the sum of the two is required to be calculated, the method is adopted to determine whether the value and the position of the corresponding array in the preset data are consistent with the value in the audio data input by the user.
In a specific embodiment, the preset data includes a second array, where the second array includes a one-dimensional array formed by characters corresponding to the audio recording and playing content or a one-dimensional array generated by reversely valued characters corresponding to the audio recording and playing content. The user finishes inputting the audio data according to the characters or the requirements corresponding to the record playing content, and then performs comparison, so as to evaluate the cognitive ability of the user according to the comparison result and the acquired time data.
In a specific embodiment, as shown in fig. 4, step S3 specifically includes:
S31': judging whether the text data and the elements in the second array belong to the same type or not through the regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
S32': converting the extracted text information into an array through a split algorithm, checking through an evaluation method, judging whether the extracted text information belongs to elements in a second array, if so, judging whether the extracted text information is consistent with the positions of the elements in the second array, and if so, successfully matching.
And judging whether the text data is matched with the corresponding elements in the second array or not through analysis and matching, thereby judging whether the result of the text data is correct or not. In particular embodiments, the speech recognition technique includes a random model approach or an artificial neural network approach. The voice recognition technology is mature, and the recognition efficiency is high.
When the second array in the preset data comprises a one-dimensional array formed by characters corresponding to the record playing content. The preset data can be digital, for example, set as [ "742", "285", "3419" ], after the record playing content is played, the playing times are recorded, and after the user sees the voice sent by the graph, the voice is converted into audio data and further converted into text data through an external voice dictation control. Firstly judging whether one of the numbers [0-9] exists in the text data through the regular expression, and if so, extracting the numbers in the text data through the match method of the regular expression. And then converting the extracted numbers into an array through a split algorithm, checking through an evaluation method, judging whether the extracted numbers belong to the preset numbers in the second array, if so, judging whether the extracted numbers are consistent with the positions of the numbers in the second array, and if so, successfully matching. Similarly, when the second array in the preset data includes a one-dimensional array generated by reversely valued the text corresponding to the record playing content. And generating a second array by reversely taking the value through the reserve based on the number, and then executing in the same way as the above to acquire data.
With further reference to fig. 5, as an implementation of the method shown in the foregoing figures, the present application provides an embodiment of a graphics rendering apparatus for cognitive assessment, where the apparatus embodiment corresponds to the method embodiment shown in fig. 2, and the apparatus may be specifically applied to various electronic devices.
As shown in fig. 5, the audio data processing apparatus based on cognitive evaluation of the present embodiment includes:
The audio data acquisition module 1 is configured to acquire audio data output by a user according to preset voice recognition related content, and convert the audio data into text data through a voice recognition technology;
A content data conversion module 2 configured to obtain preset data generated by text conversion of the voice recognition related content; and
The comparison module 3 is configured to compare the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and
And the time data acquisition module 4 is configured to acquire time data of the user in the process of completing the voice recognition related content and evaluate the cognitive ability of the user by combining the comparison result.
In a specific embodiment, the voice recognition related content comprises graphics or numbers, and the display mode of the voice recognition related content comprises graphic interface display content or recording and playing content. And displaying the content or recording and playing the content to show or guide the user to set forth the content of the preset part through the graphical interface.
In a specific embodiment, the preset data includes a first array, where the first array includes a one-dimensional array formed by text in the graphical interface display content, or a one-dimensional array obtained by performing numerical operation on text in the adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to graphics in the graphical interface display content and classification thereof. The text or the image displayed by the display content of the graphical interface can be obtained, and the audio data input by the user is subjected to array matching, so that the cognition level of the user is objectively reflected.
In a specific embodiment, the comparison module 3 specifically includes:
The first matching module (not shown in the figure) is used for matching one group of text information in the text data with the text in the first array through the match method of the regular expression, if so, the matched text information is compared with the corresponding element in the first array, whether the comparison result is the same is judged, if so, the matching is successful, otherwise, the matching is not performed;
And the circular matching module (not shown in the figure) is used for repeatedly executing the first matching module (not shown in the figure) to sequentially match all the text information of the text data, and obtaining the comparison result of each text information.
In a preferred embodiment, the positioning position of the cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching. The user can be guided to finish the input of the audio data required by the display content of the graphical interface by changing the positioning position of the cursor, so that the accuracy of array matching and the finishing efficiency are improved.
In a specific embodiment, the preset data includes a second array, where the second array includes a one-dimensional array formed by characters corresponding to the audio recording and playing content or a one-dimensional array generated by reversely valued characters corresponding to the audio recording and playing content. The user finishes inputting the audio data according to the characters or the requirements corresponding to the record playing content, and then performs comparison, so as to evaluate the cognitive ability of the user according to the comparison result and the acquired time data.
In a specific embodiment, the comparison module 3 may specifically further include:
the data extraction module (not shown in the figure) is used for judging whether the text data and the elements in the second array belong to the same type through the regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
and the second matching module (not shown in the figure) is used for converting the extracted text information into an array through a split algorithm and checking through an evaluation method, judging whether the extracted text information belongs to the elements in the second array, if so, judging whether the extracted text information is consistent with the positions of the elements in the second array, and if so, successfully matching.
The invention discloses an audio data processing method and device based on cognitive evaluation, which are characterized in that audio data input by a user according to preset voice recognition related content are collected, and the audio data are converted into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result. The difficulty in evaluation of cognitive dysfunction can be effectively reduced through processing the audio data, and more intelligent, efficient and quick experience is brought to the whole cognitive evaluation process. And the data acquired by the user in the cognitive evaluation process are more diversified and accurate, so that the data can be recorded and evaluated in real time, and the accuracy of the cognitive evaluation is effectively improved.
Referring now to fig. 6, there is illustrated a schematic diagram of a computer apparatus 600 suitable for use in an electronic device (e.g., a server or terminal device as illustrated in fig. 1) for implementing an embodiment of the present application. The electronic device shown in fig. 6 is only an example and should not be construed as limiting the functionality and scope of use of the embodiments of the application.
As shown in fig. 6, the computer apparatus 600 includes a Central Processing Unit (CPU) 601 and a Graphics Processor (GPU) 602, which can perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 603 or a program loaded from a storage section 609 into a Random Access Memory (RAM) 606. In the RAM604, various programs and data required for the operation of the apparatus 600 are also stored. The CPU 601, GPU602, ROM 603, and RAM604 are connected to each other through a bus 605. An input/output (I/O) interface 606 is also connected to the bus 605.
The following components are connected to the I/O interface 606: an input portion 607 including a keyboard, a mouse, and the like; an output portion 608 including a speaker, such as a Liquid Crystal Display (LCD), etc.; a storage portion 609 including a hard disk and the like; and a communication section 610 including a network interface card such as a LAN card, a modem, or the like. The communication section 610 performs communication processing via a network such as the internet. The drive 611 may also be connected to the I/O interface 606 as needed. A removable medium 612 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is mounted on the drive 611 as necessary, so that a computer program read out therefrom is mounted into the storage section 609 as necessary.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method shown in the flowcharts. In such embodiments, the computer program may be downloaded and installed from a network via the communication portion 610, and/or installed from the removable medium 612. The above-described functions defined in the method of the present application are performed when the computer program is executed by a Central Processing Unit (CPU) 601 and a Graphics Processor (GPU) 602.
It should be noted that the computer readable medium according to the present application may be a computer readable signal medium or a computer readable medium, or any combination of the two. The computer readable medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor apparatus, device, or means, or a combination of any of the foregoing. More specific examples of the computer-readable medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution apparatus, device, or apparatus. In the present application, however, the computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may be any computer readable medium that is not a computer readable medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution apparatus, device, or apparatus. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wire, fiber optic cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations of the present application may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based devices which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The modules involved in the embodiments of the present application may be implemented in software or in hardware. The described modules may also be provided in a processor.
As another aspect, the present application also provides a computer-readable medium that may be contained in the electronic device described in the above embodiment; or may exist alone without being incorporated into the electronic device. The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology; acquiring preset data generated by text conversion of voice recognition related content; comparing the text data with preset data through a regular expression matching algorithm to obtain a comparison result; and collecting time data of the user in the process of completing the voice recognition related content for evaluating the cognitive ability of the user by combining the comparison result.
A description of the technical principles applied. It will be appreciated by persons skilled in the art that the scope of the application referred to in the present application is not limited to the specific combinations of the technical features described above, but also covers other technical features formed by any combination of the technical features described above or their equivalents without departing from the inventive concept described above. Such as the above-mentioned features and the technical features disclosed in the present application (but not limited to) having similar functions are replaced with each other.

Claims (7)

1. A method for processing audio data based on cognitive assessment, comprising the steps of:
s1: collecting audio data input by a user according to preset voice recognition related content, and converting the audio data into text data through a voice recognition technology;
s2: acquiring preset data generated by text conversion of the voice recognition related content;
s3: comparing the text data with the preset data through a regular expression matching algorithm to obtain a comparison result; and
S4: collecting time data of a user in the process of completing the voice recognition related content, and evaluating the cognitive ability of the user by combining the comparison result;
The preset data comprises a first array, wherein the first array comprises a one-dimensional array formed by corresponding characters in the graphical interface display content, or a one-dimensional array obtained by carrying out numerical operation on the characters in the adjacent graphical interface display content, or a two-dimensional array formed by a plurality of nouns corresponding to the graphics in the graphical interface display content and classification thereof; the preset data comprises a second array, wherein the second array comprises a one-dimensional array formed by characters corresponding to the recording and playing contents or a one-dimensional array generated by reversely taking the characters corresponding to the recording and playing contents; the step S3 specifically includes:
S31: matching one group of text information in the text data with the text in the first array by using a match method of a regular expression, if so, comparing the matched text information with the corresponding element in the first array, judging whether the comparison result is the same, if so, successfully matching, otherwise, not matching;
S32: and repeating the step S31 to match all the text information of the text data in turn, and obtaining the comparison result of each text information.
2. The cognitive assessment-based audio data processing method according to claim 1, wherein the voice recognition related content comprises graphics or numbers, and the voice recognition related content is displayed in a manner comprising displaying content through a graphical interface or playing content through a sound recording.
3. The cognitive assessment-based audio data processing method according to claim 2, wherein the positioning position of a cursor in the display content of the graphical interface is determined according to the matching completion degree of the text information participating in the matching.
4. The method for processing audio data based on cognitive assessment according to claim 3, wherein the step S3 specifically comprises:
s31': judging whether the text data and the elements in the second array belong to the same type or not through a regular expression, if so, extracting the corresponding text information in the text data, otherwise, not extracting;
s32': converting the extracted text information into an array through a split algorithm, checking through an evaluation method, judging whether the extracted text information belongs to the element in the second array, if so, judging whether the extracted text information is consistent with the position of the element in the second array, and if so, successfully matching.
5. An audio data processing device based on cognitive assessment, comprising:
the audio data acquisition module is configured to acquire audio data output by a user according to preset voice recognition related content, and convert the audio data into text data through a voice recognition technology;
The content data conversion module is configured to obtain preset data generated by text conversion of the voice recognition related content; and
The comparison module is configured to compare the text data with the preset data through a regular expression matching algorithm to obtain a comparison result; and
And the time data acquisition module is configured to acquire time data of a user in the process of completing the voice recognition related content and is used for evaluating the cognitive ability of the user in combination with the comparison result.
6. An electronic device, comprising:
one or more processors;
Storage means for storing one or more programs,
When executed by the one or more processors, causes the one or more processors to implement the method of any of claims 1-4.
7. A computer readable storage medium, on which a computer program is stored, characterized in that the program, when being executed by a processor, implements the method according to any of claims 1-4.
CN202010988651.1A 2020-09-18 2020-09-18 Audio data processing method and device based on cognitive assessment Active CN113539253B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010988651.1A CN113539253B (en) 2020-09-18 2020-09-18 Audio data processing method and device based on cognitive assessment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010988651.1A CN113539253B (en) 2020-09-18 2020-09-18 Audio data processing method and device based on cognitive assessment

Publications (2)

Publication Number Publication Date
CN113539253A CN113539253A (en) 2021-10-22
CN113539253B true CN113539253B (en) 2024-05-14

Family

ID=78094284

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010988651.1A Active CN113539253B (en) 2020-09-18 2020-09-18 Audio data processing method and device based on cognitive assessment

Country Status (1)

Country Link
CN (1) CN113539253B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115798718B (en) * 2022-11-24 2024-03-29 广州市第一人民医院(广州消化疾病中心、广州医科大学附属市一人民医院、华南理工大学附属第二医院) Cognitive test evaluation method and system
CN116048282B (en) * 2023-03-06 2023-08-04 中国医学科学院生物医学工程研究所 Data processing method, system, device, equipment and storage medium

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6278996B1 (en) * 1997-03-31 2001-08-21 Brightware, Inc. System and method for message process and response
WO2002037223A2 (en) * 2000-11-06 2002-05-10 Invention Machine Corporation Computer based integrated text and graphic document analysis
JP2004184535A (en) * 2002-11-29 2004-07-02 Fujitsu Ltd Device and method for speech recognition
RU2253365C1 (en) * 2003-11-17 2005-06-10 Государственное образовательное учреждение высшего профессионального образования Московская медицинская академия им. И.М. Сеченова МЗ РФ Psycholinguistic method for diagnosing neurotic disorders
KR20070019596A (en) * 2005-08-12 2007-02-15 캐논 가부시끼가이샤 Information processing method and information processing device
KR20090000662A (en) * 2007-03-16 2009-01-08 장성옥 Speech studying game and system using the game
CA2820599A1 (en) * 2010-11-24 2012-05-31 Digital Artefacts, Llc Systems and methods to assess cognitive function
CN103251418A (en) * 2013-06-05 2013-08-21 清华大学 Image cognition psychoanalysis system
CN103400579A (en) * 2013-08-04 2013-11-20 徐华 Voice recognition system and construction method
CN104021786A (en) * 2014-05-15 2014-09-03 北京中科汇联信息技术有限公司 Speech recognition method and speech recognition device
KR101538317B1 (en) * 2014-02-20 2015-07-29 ㈜빅스바이트 An automatic evaluation system for English literacy
CN106446165A (en) * 2016-09-26 2017-02-22 厦门吉信德宠物用品有限公司 Big data processing based identification method
CN108846119A (en) * 2018-06-27 2018-11-20 清远墨墨教育科技有限公司 A kind of arrangement method, storage device and the mobile terminal of word cognition degree
CN109222882A (en) * 2018-10-08 2019-01-18 浙江工业大学 A kind of reading visual acuity test system and method
CN109344231A (en) * 2018-10-31 2019-02-15 广东小天才科技有限公司 Method and system for completing corpus of semantic deformity
CN109407946A (en) * 2018-09-11 2019-03-01 昆明理工大学 Graphical interfaces target selecting method based on speech recognition
CN109933671A (en) * 2019-01-31 2019-06-25 平安科技(深圳)有限公司 Construct method, apparatus, computer equipment and the storage medium of personal knowledge map
CN110473605A (en) * 2018-05-09 2019-11-19 深圳市前海安测信息技术有限公司 Alzheimer Disease patient figure cognitive ability assessment system and method
CN111295141A (en) * 2017-11-02 2020-06-16 松下知识产权经营株式会社 Cognitive function evaluation device, cognitive function evaluation system, cognitive function evaluation method, and program

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004246184A (en) * 2003-02-14 2004-09-02 Eigyotatsu Kofun Yugenkoshi Language learning system and method with visualized pronunciation suggestion
US7429108B2 (en) * 2005-11-05 2008-09-30 Outland Research, Llc Gaze-responsive interface to enhance on-screen user reading tasks
JP4811507B2 (en) * 2009-08-25 2011-11-09 コニカミノルタビジネステクノロジーズ株式会社 Image processing system, image processing apparatus, and information processing apparatus
US9471715B2 (en) * 2013-03-31 2016-10-18 International Business Machines Corporation Accelerated regular expression evaluation using positional information
US11017774B2 (en) * 2019-02-04 2021-05-25 International Business Machines Corporation Cognitive audio classifier

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6278996B1 (en) * 1997-03-31 2001-08-21 Brightware, Inc. System and method for message process and response
WO2002037223A2 (en) * 2000-11-06 2002-05-10 Invention Machine Corporation Computer based integrated text and graphic document analysis
JP2004184535A (en) * 2002-11-29 2004-07-02 Fujitsu Ltd Device and method for speech recognition
RU2253365C1 (en) * 2003-11-17 2005-06-10 Государственное образовательное учреждение высшего профессионального образования Московская медицинская академия им. И.М. Сеченова МЗ РФ Psycholinguistic method for diagnosing neurotic disorders
KR20070019596A (en) * 2005-08-12 2007-02-15 캐논 가부시끼가이샤 Information processing method and information processing device
KR20090000662A (en) * 2007-03-16 2009-01-08 장성옥 Speech studying game and system using the game
CA2820599A1 (en) * 2010-11-24 2012-05-31 Digital Artefacts, Llc Systems and methods to assess cognitive function
CN103251418A (en) * 2013-06-05 2013-08-21 清华大学 Image cognition psychoanalysis system
CN103400579A (en) * 2013-08-04 2013-11-20 徐华 Voice recognition system and construction method
KR101538317B1 (en) * 2014-02-20 2015-07-29 ㈜빅스바이트 An automatic evaluation system for English literacy
CN104021786A (en) * 2014-05-15 2014-09-03 北京中科汇联信息技术有限公司 Speech recognition method and speech recognition device
CN106446165A (en) * 2016-09-26 2017-02-22 厦门吉信德宠物用品有限公司 Big data processing based identification method
CN111295141A (en) * 2017-11-02 2020-06-16 松下知识产权经营株式会社 Cognitive function evaluation device, cognitive function evaluation system, cognitive function evaluation method, and program
CN110473605A (en) * 2018-05-09 2019-11-19 深圳市前海安测信息技术有限公司 Alzheimer Disease patient figure cognitive ability assessment system and method
CN108846119A (en) * 2018-06-27 2018-11-20 清远墨墨教育科技有限公司 A kind of arrangement method, storage device and the mobile terminal of word cognition degree
CN109407946A (en) * 2018-09-11 2019-03-01 昆明理工大学 Graphical interfaces target selecting method based on speech recognition
CN109222882A (en) * 2018-10-08 2019-01-18 浙江工业大学 A kind of reading visual acuity test system and method
CN109344231A (en) * 2018-10-31 2019-02-15 广东小天才科技有限公司 Method and system for completing corpus of semantic deformity
CN109933671A (en) * 2019-01-31 2019-06-25 平安科技(深圳)有限公司 Construct method, apparatus, computer equipment and the storage medium of personal knowledge map

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
产品认知语境的类型及其应用;傅桂涛;潘荣;陈国东;陈思宇;;包装工程(第08期);全文 *
元认知策略与大学英语阅读能力的提高;王莉;毕凤春;;高等农业教育;20060428(第04期);全文 *
军人个体认知能力综合评估研究;安改红;王静;陈学伟;李超;陈佩延;安芳红;张文正;李正东;袭著革;马强;;人民军医(第01期);全文 *
在校医学生对于老年健康的认知现状分析;高奎;张丽娜;涂虹;刘晓微;李婷;肖雄;;中国医药导报(第03期);全文 *
基于PASS理论的五项认知能力与中文阅读理解能力的关系研究;张晴;刘巧云;杜晓新;黄昭鸣;祝亚平;;中国儿童保健杂志;20180105(第02期);全文 *

Also Published As

Publication number Publication date
CN113539253A (en) 2021-10-22

Similar Documents

Publication Publication Date Title
CN110457432B (en) Interview scoring method, interview scoring device, interview scoring equipment and interview scoring storage medium
CN110781413B (en) Method and device for determining interest points, storage medium and electronic equipment
CN113539253B (en) Audio data processing method and device based on cognitive assessment
CN111709630A (en) Voice quality inspection method, device, equipment and storage medium
CN107481715B (en) Method and apparatus for generating information
CN112116903B (en) Speech synthesis model generation method and device, storage medium and electronic equipment
CN111651497A (en) User label mining method and device, storage medium and electronic equipment
US11024329B2 (en) Word repetition in separate conversations for detecting a sign of cognitive decline
CN114138960A (en) User intention identification method, device, equipment and medium
CN114140814A (en) Emotion recognition capability training method and device and electronic equipment
CN109101956B (en) Method and apparatus for processing image
Shanthi et al. An integrated approach for mental health assessment using emotion analysis and scales
CN117828355A (en) Emotion quantitative model training method and emotion quantitative method based on multi-modal information
CN117786600A (en) Cognitive evaluation method, device, electronic equipment and storage medium
KR20200084816A (en) Method, apparatus and computer program for analyzing new contents for solving cold start
CN114240250A (en) Intelligent management method and system for vocational evaluation
CN113361282B (en) Information processing method and device
CN114613350A (en) Test method, test device, electronic equipment and storage medium
CN111949860B (en) Method and apparatus for generating a relevance determination model
CN109344289B (en) Method and apparatus for generating information
CN112131378A (en) Method and device for identifying categories of civil problems and electronic equipment
CN113220849B (en) Customer service personnel emotion dispersion scheme searching method, electronic equipment and storage medium
CN115440198B (en) Method, apparatus, computer device and storage medium for converting mixed audio signal
CN114219369B (en) Prediction model training method and device, and user category prediction method and device
CN118467709B (en) Evaluation method, device, medium and computer program product for visual question-answering task

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant