CN111353038A - Data display method and device, computer equipment and storage medium - Google Patents

Data display method and device, computer equipment and storage medium Download PDF

Info

Publication number
CN111353038A
CN111353038A CN202010451239.6A CN202010451239A CN111353038A CN 111353038 A CN111353038 A CN 111353038A CN 202010451239 A CN202010451239 A CN 202010451239A CN 111353038 A CN111353038 A CN 111353038A
Authority
CN
China
Prior art keywords
sentence
specified
information
data
recognition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010451239.6A
Other languages
Chinese (zh)
Inventor
陈杰
温平
杨汉丹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Youjie Zhixin Technology Co ltd
Original Assignee
Shenzhen Youjie Zhixin Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Youjie Zhixin Technology Co ltd filed Critical Shenzhen Youjie Zhixin Technology Co ltd
Priority to CN202010451239.6A priority Critical patent/CN111353038A/en
Publication of CN111353038A publication Critical patent/CN111353038A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/34Browsing; Visualisation therefor
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/221Announcement of recognition results

Abstract

The application discloses a data display method, a data display device, computer equipment and a storage medium, wherein the method comprises the following steps: receiving a viewing instruction for viewing a recording identification result corresponding to the recording file; acquiring all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file, and acquiring the recording start time of each piece of voice sentence data; the recording start time of the specified voice sentence data is presented, and the specified recognition sentence information corresponding to the specified voice sentence data is presented at a specified position corresponding to the recording start time of the specified voice sentence data. Through the method and the device, when the checking requirement for checking the recording identification result corresponding to the recording file is met, the user can not only look up the recording identification result of the recording file, but also look up more information related to the recording identification result of the recording file, the richness of data display of the recording identification result is effectively improved, and the user experience is improved.

Description

Data display method and device, computer equipment and storage medium
Technical Field
The present application relates to the field of information processing technologies, and in particular, to a data display method and apparatus, a computer device, and a storage medium.
Background
After the user has accomplished the input of recording data, can have the demand of looking over the recording recognition result that corresponds with the recording data usually, terminal equipment who has recording recognition function at present when showing the recording recognition result, generally can only show and carry out a holistic recording recognition result that obtains after speech recognition to the recording data for the user can only look up the recording recognition result of recording data, but can't learn other information relevant with the recording data except that the recording recognition result, user's use experience is not good. The existing data display of the recording identification result is single.
Disclosure of Invention
The application mainly aims to provide a data display method, a data display device, computer equipment and a storage medium, and aims to solve the technical problem that the existing data display of a recording identification result is single.
The application provides a data display method, which comprises the following steps:
receiving a viewing instruction for viewing a recording identification result corresponding to the recording file;
acquiring all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file, and acquiring the recording start time of each piece of voice sentence data, wherein one piece of voice sentence data corresponds to one piece of recognition sentence information;
displaying a recording start time of specified voice sentence data, and displaying specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, wherein the specified voice sentence data is any one piece of voice sentence data in all the voice sentence data.
Alternatively, the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data may be followed by:
obtaining sentence type information of the specified voice sentence data;
sentence type information of the specified speech sentence data is presented at a first position corresponding to the specified recognized sentence information.
Optionally, the step of presenting sentence type information of the specified speech sentence data at a first position corresponding to the specified recognized sentence information, includes:
judging whether the sentence type information of the specified voice sentence data is a silent sentence type;
if the sentence type information of the specified voice sentence data is judged to be the silent sentence type, generating a display switch button at a second position corresponding to the specified recognition sentence information;
judging whether a closing operation instruction for the display switch button is received or not;
and if the closing operation instruction of the display switch button is judged to be received, hiding all display information corresponding to the specified voice sentence data.
Alternatively, the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data may be followed by:
acquiring recording time length information of the specified voice sentence data; and the number of the first and second groups,
acquiring identification information of the appointed recognition sentence information;
displaying recording time length information of the specified voice sentence data at a third position corresponding to the specified voice sentence data; and the number of the first and second groups,
displaying identification information of the specified recognition sentence information at a fourth position corresponding to the specified recognition sentence information.
Alternatively, the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data may be followed by:
displaying an audio play button at a fifth position corresponding to the designated recognized sentence information;
judging whether a first operation instruction for the audio playing button is detected;
if the first operation instruction for the audio playing button is detected, responding to the first operation instruction, and extracting an audio file corresponding to the appointed recognition sentence information from the recording file;
and playing an audio file corresponding to the appointed recognition sentence information.
Alternatively, the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data may be followed by:
displaying an edit button on a sixth position corresponding to the specified recognition sentence information;
judging whether a second operation instruction for the editing button is detected;
if the second operation instruction for the editing button is detected, responding to the second operation instruction and receiving input editing information;
and according to the editing information, performing adjustment processing corresponding to the editing information on the appointed recognition sentence information.
Alternatively, the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data may be followed by:
calling a preset translation module to translate the appointed recognition sentence information to obtain translation information corresponding to the appointed recognition sentence information;
acquiring the translated text information output by the translation module;
displaying the translated version information of the specified recognition sentence information at a seventh position corresponding to the specified recognition sentence information.
The present application further provides a data display device, comprising:
the first receiving module is used for receiving a checking instruction for checking a recording identification result corresponding to the recording file;
a first obtaining module, configured to obtain all pieces of recognition sentence information obtained by performing speech recognition processing on each piece of speech sentence data in the recording file, and obtain a recording start time of each piece of speech sentence data, where one piece of speech sentence data corresponds to one piece of recognition sentence information;
the system comprises a first display module and a second display module, wherein the first display module is used for displaying the recording start time of specified voice sentence data and displaying the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, and the specified voice sentence data is any one section of voice sentence data in the voice sentence data.
The present application further provides a computer device, comprising a memory and a processor, wherein the memory stores a computer program, and the processor implements the steps of the above method when executing the computer program.
The present application also provides a computer-readable storage medium having stored thereon a computer program which, when being executed by a processor, carries out the steps of the above-mentioned method.
The data display method, the data display device, the computer equipment and the storage medium have the following beneficial effects:
according to the data display method, the data display device, the computer equipment and the storage medium, when a viewing instruction for viewing the recording recognition result corresponding to the recording file is received, all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file is obtained, and the recording start time of each piece of voice sentence data is obtained. Then, a recording start time of the specified voice sentence data is displayed, and specified recognition sentence information corresponding to the specified voice sentence data is displayed at a specified position corresponding to the recording start time of the specified voice sentence data, wherein the specified voice sentence data is any one piece of voice sentence data among all the voice sentence data. This application is when receiving the instruction of looking over the recording identification result that corresponds with the recording file, through the recording start time of every section of pronunciation sentence data in showing the recording file to the user to and the identification sentence information of every section of pronunciation sentence data one-to-one, make the user not only can look up the recording identification result of recording file, and can look up more information relevant with the recording identification result of recording file, the richness of the data display of recording identification result has been improved effectively, user's use experience has been improved.
Drawings
FIG. 1 is a schematic flow chart diagram illustrating a data presentation method according to an embodiment of the present application;
FIG. 2 is a schematic structural diagram of a data display device according to an embodiment of the present application;
fig. 3 is a schematic structural diagram of a computer device according to an embodiment of the present application.
The implementation, functional features and advantages of the objectives of the present application will be further explained with reference to the accompanying drawings.
Detailed Description
It should be understood that the specific embodiments described herein are merely illustrative of the present application and are not intended to limit the present application.
As used herein, the singular forms "a", "an", "the" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises" and/or "comprising," when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will be understood that when an element is referred to as being "connected" or "coupled" to another element, it can be directly connected or coupled to the other element or intervening elements may also be present. Further, "connected" or "coupled" as used herein may include wirelessly connected or wirelessly coupled. As used herein, the term "and/or" includes all or any element and all combinations of one or more of the associated listed items.
It will be understood by those skilled in the art that, unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the prior art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
Referring to fig. 1, a data display method according to an embodiment of the present application includes:
s1: receiving a viewing instruction for viewing a recording identification result corresponding to the recording file;
s2: acquiring all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file, and acquiring the recording start time of each piece of voice sentence data, wherein one piece of voice sentence data corresponds to one piece of recognition sentence information;
s3: displaying a recording start time of specified voice sentence data, and displaying specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, wherein the specified voice sentence data is any one piece of voice sentence data in all the voice sentence data.
As described in the above steps S1-S3, the main implementation of the present method is a data presentation apparatus. In practical applications, the data display device may be implemented by a virtual device, such as a software code, or an entity device in which a relevant execution code is written or integrated, and may perform human-computer interaction with a user through a keyboard, a mouse, a remote controller, a touch panel, or a voice control device, and the data display device may specifically include a terminal device having a recording recognition function. The data display device of this embodiment can effectively improve the richness of the data display of the recording identification result corresponding to the recording file, and improve the user's use experience. Specifically, a viewing instruction for viewing a recording identification result corresponding to a recording file is first received. The viewing instruction may include a single-click instruction or a double-click instruction input by a user for the recording identification result name associated with the recording identification result. The specific content included in the sound recording file is not limited, and the sound recording file may be sound recording file data generated by reading a specified text aloud by a user, and the sound recording recognition result may be character result information generated by performing a voice recognition process on the sound recording file. After receiving the viewing instruction, all recognition sentence information obtained by performing voice recognition processing on each piece of voice sentence data in the recording file is acquired, and the recording start time of each piece of voice sentence data is acquired. Wherein, a section of speech sentence data can correspond to a recognition sentence information, the process of carrying out recognition processing to the recording file comprises: the recording file is segmented into a plurality of sections of voice sentences. After obtaining the multiple sections of speech sentence data, performing speech recognition on each section of speech sentence data, and further obtaining recognition sentence information corresponding to each section of speech sentence data one by one. In addition, the recording file can be segmented based on a voice endpoint detection mode, a plurality of voice sentence data can be obtained through the segmentation processing, and simultaneously, the recording start time corresponding to each voice sentence data can be recorded. The recording start time may also be referred to as a recording start position, where the recording start time refers to start position information of voice sentence data in the entire recording file, the recording start times of different voice sentence data are different, and the precedence orders corresponding to different recording start times are also different, for example, the recording start time of a first section of voice sentence data, which is most front in the precedence order of the recording start times in the recording file, is 0 minutes 0 seconds, the recording start time of a second section of voice sentence data, which is second in the precedence order of the recording start times in the recording file, may be 0 minutes 10 seconds, and the precedence order of the recording start time of the first section of voice sentence data is ahead of the second section of voice sentence data. And finally, displaying the recording start time of the specified voice sentence data, and displaying the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data. The specified speech sentence data is any one of the speech sentence data, and the specified recognition sentence information is recognition sentence information obtained by performing speech recognition processing on the specified speech sentence data. Moreover, the information of the specified recognized sentence may have a case where there is no recognition data, that is, when the specified speech sentence data is a segment of silent sentence data, the recognition result corresponding to the silent sentence data is empty. The above-mentioned specified position corresponding to the recording start time of the specified voice sentence data is specifically a position in the same dimension (may be the same row or the same column) as the placement position of the recording start time of the specified voice sentence data, and is such that there is an association between the placement position of the recording start time of the specified voice sentence data and the placement position of the specified recognition sentence information (i.e. the above-mentioned specified position). In addition, usually, a recording file corresponds to a plurality of pieces of speech sentence data and a plurality of pieces of recognition sentence information, in order to improve the ordering of data display of the recording recognition result, the recording start time of each piece of speech sentence data in the recording file and the recognition sentence information corresponding to each piece of speech sentence data in the recording file one by one may be displayed according to a certain ordering rule, specifically, the recording start time of each piece of speech sentence data in the recording file may be sequentially displayed according to the sequence of the recording start time, and then the recognition sentence information corresponding to each piece of speech sentence data in the recording file one by one may be sequentially displayed at a position corresponding to the recording start time of each piece of speech sentence data in the recording file. The specified presentation format may be adopted to present the recording start time of each piece of speech sentence data in the recording file, and to present the recognition sentence information in one-to-one correspondence with each piece of speech sentence data in the recording file. Specifically, the specified presentation forms may include a list, an icon, a time axis, and the like. Taking a list form as an example for explanation, firstly, according to the sequence of the recording start time, placing the first recording start time of the first voice sentence data which is sequenced in the recording file in the first sequence (the sequence is the first), in the first cell of the first column corresponding to the preset recording start time label, and then placing the first recognition sentence information corresponding to the first voice sentence data, in the second cell of the first column corresponding to the recognition result label; placing a second recording start time of second voice sentence data which is sequenced in sequence in the recording file in a first cell of a second row corresponding to a preset recording start time label, and then placing second recognition sentence information corresponding to the second voice sentence data in a second cell of the second row corresponding to a recognition result label; and analogizing in turn until the recording start time of all the voice sentence data in the recording file and the identification sentence information corresponding to each voice sentence data in the recording file one by one are all placed in a list for displaying. Since the table form and the time axis form are similar to the list form, the table form and the time axis form can be referred to the list form, and the details are not repeated herein. When receiving a viewing instruction for viewing a recording identification result corresponding to a recording file, the embodiment enables a user to not only look up the recording identification result of the recording file, but also look up more information related to the recording identification result of the recording file by showing the recording start time of each piece of voice sentence data in the recording file and the identification sentence information corresponding to each piece of voice sentence data one to one, thereby effectively improving the richness of data display of the recording identification result and improving the use experience of the user.
Further, in an embodiment of the present application, after the step S3, the method includes:
s300: obtaining sentence type information of the specified voice sentence data;
s301: sentence type information of the specified speech sentence data is presented at a first position corresponding to the specified recognized sentence information.
As described in steps S300 to S301, when a viewing instruction for viewing the recording recognition result corresponding to the recording file is received, in addition to displaying the recording start time of each piece of speech sentence data and the recognition sentence information corresponding to each piece of speech sentence data one by one to the user, the sentence type information of each piece of speech sentence data may be further displayed, so that the user can clearly know the sentence type corresponding to each piece of speech sentence data. Specifically, after the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at the specified position corresponding to the recording start time of the specified voice sentence data, the method may further include: sentence type information of the above-mentioned specified speech sentence data is first acquired. When the recording file is segmented into a plurality of sections of voice sentence data, sentence type information of each section of voice sentence data can be recorded at the same time, and the sentence type information comprises a voice sentence type or a silent sentence type. And a data storage module can be further created, and the obtained sentence type information of each piece of recognition sentence information is stored in the data storage module. In addition, the speech sentence data corresponding to the voiced sentence type belongs to the speech signal, and the speech sentence data corresponding to the silent sentence type belongs to the non-speech signal. The voice endpoint detection can be called as voice activity detection, voice signals and non-voice signals can be separated from the recording files, the starting point and the ending point of the voice signals can be located, the starting point and the ending point can be called as endpoints, and therefore the endpoints are detected by adopting the voice endpoint detection mode and whether two sections of sounds belong to the same sentence or not is judged by combining with the endpoint interval, and therefore the sentence break segmentation of the recording files is achieved, and voice sentence data belonging to the type of the voice sentences and voice sentence data belonging to the type of the mute sentences can be obtained. In addition, the voice recognition sentence segmentation algorithm can be used for carrying out recognition judgment processing on the recording file so as to obtain sentence type information corresponding to each piece of voice sentence data. If it is recognized that a piece of sound data has been continuously muted, it can be determined that the piece of sound data belongs to a silent sentence. Alternatively, the piece of sound data may be recognized, and if the obtained recognition result is recognition failure, it may be determined that the piece of sound data belongs to a silent sentence, and if a valid recognition result can be recognized, it may be determined that the piece of sound data belongs to a voiced sentence. After the sentence type information of the specified speech sentence data is obtained, the sentence type information of the specified speech sentence data is displayed at a first position corresponding to the specified recognized sentence information. The first position corresponding to the designated recognized sentence information is a position in the same dimension (may be in the same row or the same column) as the position of the designated recognized sentence information, and the position of the sentence type information of the designated speech sentence data (i.e., the first position) is associated with the position of the designated recognized sentence information (i.e., the designated position). When a viewing instruction for viewing the recording identification result corresponding to the recording file is received, sentence type information of each piece of voice sentence data is further displayed on the basis of displaying the recording start time of each piece of voice sentence data in the recording file and the identification sentence information corresponding to each piece of voice sentence data one to one, so that a user can look up more information related to the recording identification result of the recording file, the richness of data display of the recording identification result is effectively improved, and the use experience of the user is improved.
Further, in an embodiment of the present application, after the step S301, the method includes:
s302: judging whether the sentence type information of the specified voice sentence data is a silent sentence type;
s303: if the sentence type information of the specified voice sentence data is judged to be the silent sentence type, generating a display switch button at a second position corresponding to the specified recognition sentence information;
s304: judging whether a closing operation instruction for the display switch button is received or not;
s305: and if the closing operation instruction of the display switch button is judged to be received, hiding all display information corresponding to the specified voice sentence data.
As described in the above steps S302 to S305, the sentence type information of the specified voice sentence data may include a voice sentence type or a silent sentence type, and when the sentence type information is the silent sentence type, a display switch button may be further provided for the specified voice sentence data belonging to the silent sentence type, so that the user can control whether to display the presentation information related to the specified voice sentence data by using the display switch button. Specifically, it is first determined whether the sentence type information of the above-mentioned specified voice sentence data is a silent sentence type. If the sentence type information of the specified speech sentence data is judged to be the silent sentence type, a display switch button is generated at a second position corresponding to the specified recognition sentence information. The second position corresponding to the specific recognized sentence information is not particularly limited, and may be set at any position in a cell containing the specific recognized sentence information. And then judging whether a closing operation instruction for the display switch button is received. The determination of whether the closing operation command is detected may be made by determining whether a first operation of closing the display switch button by a user is received. If the first operation action is received, the closing operation command is judged to be detected, and if the first operation action is not received, the closing operation command is judged not to be detected. And hiding all the display information corresponding to the specified voice sentence data if the closing operation instruction of the display switch button is received. The presentation information may include at least a recording start time of the specified speech sentence data, specified recognition sentence information, and sentence type information of the specified speech sentence data. According to the embodiment, the display switch button is arranged on the appointed recognition sentence information belonging to the silent sentence type, so that the user can control whether the display information related to the silent sentence needs to be displayed or not according to the actual requirement of the user, the user can be effectively helped to filter out some useless information, the user does not spend too much time to check the useless information, and the use experience of the user is improved. Further, the data presentation device may also automatically hide presentation information associated with the voice sentence data belonging to the silent sentence type. When the sentence type information of the specified voice sentence data is judged to be the silent sentence type, the data display device can default the information associated with the specified voice sentence data to be useless information and can directly hide all the display information corresponding to the specified voice sentence data, so that the input action of inputting the closing operation instruction of the display switch button is omitted for the user, and the use experience of the user is greatly improved.
Further, in an embodiment of the present application, after the step S3, the method includes:
s310: acquiring recording time length information of the specified voice sentence data; and the number of the first and second groups,
s311: acquiring identification information of the appointed recognition sentence information;
s312: displaying recording time length information of the specified voice sentence data at a third position corresponding to the specified voice sentence data; and the number of the first and second groups,
s313: displaying identification information of the specified recognition sentence information at a fourth position corresponding to the specified recognition sentence information.
As described in the above steps S310 to S313, when the viewing instruction for viewing the recording recognition result corresponding to the recording file is received, in addition to displaying the recording start time of each piece of speech sentence data and the recognition sentence information corresponding to each piece of speech sentence data one to the user, the recording time length information of each piece of speech sentence data and the identification information of each recognition sentence information may be further displayed so that the user can clearly know other information associated with the recording recognition result. Specifically, recording time length information of specified speech sentence data is first acquired, and identification information of the above-described specified recognition sentence information is acquired. When the recording file is segmented into a plurality of sections of voice sentence data, the recording time length information of each section of voice sentence data can be recorded at the same time. Also, the recording time length information of the specified voice sentence data may be obtained by acquiring a recording start time of the specified voice sentence data and a recording end time of the specified voice sentence data, and calculating a difference between the recording end time of the specified voice sentence data and the recording start time of the specified voice sentence data. Further, sentence length information specifying the recognized sentence information, which indicates the total number of words contained in one piece of speech sentence data, may be acquired instead of the recording time length information of the above-described specified speech sentence data. In addition, the identification information is unique identification information corresponding to the designated recognition sentence information, and the identification time may be specifically a recognition time. After the recording time length information of the specified speech sentence data is obtained, the recording time length information of the specified speech sentence data is presented at a third position corresponding to the specified speech sentence data. And displaying the identification information of the specified recognition sentence information at a fourth position corresponding to the specified recognition sentence information after the identification information of the specified recognition sentence information is obtained. The third position corresponding to the specified recognition sentence information is specifically a position in the same dimension (which may be the same row or the same column) as the placement position of the specified recognition sentence information, and a relationship exists between the placement position of the recording time length information of the specified speech sentence data (i.e., the first position) and the placement position of the specified recognition sentence information (i.e., the specified position); the setting manner of the fourth position corresponding to the information of the specified recognition sentence can refer to the third position, and is not described herein again. When a viewing instruction for viewing the recording recognition result corresponding to the recording file is received, the recording start time of each piece of voice sentence data in the viewing of the recording file is displayed, and the identification information of each piece of voice sentence data is further displayed on the basis of the identification sentence information corresponding to each piece of voice sentence data in a one-to-one manner, so that a user can look up more information related to the recording recognition result of the recording file, the richness of data display of the recording recognition result is effectively improved, and the use experience of the user is improved.
Further, in an embodiment of the present application, after the step S3, the method includes:
s320: displaying an audio play button at a fifth position corresponding to the designated recognized sentence information;
s321: judging whether a first operation instruction for the audio playing button is detected;
s322: if the first operation instruction for the audio playing button is detected, responding to the first operation instruction, and extracting an audio file corresponding to the appointed recognition sentence information from the recording file;
s323: and playing an audio file corresponding to the appointed recognition sentence information.
As described in the above steps S320 to S323, when the viewing instruction for viewing the recording recognition result corresponding to the recording file is received, in addition to displaying the recording start time of each piece of speech sentence data and the recognition sentence information corresponding to each piece of speech sentence data one to the user, the user may be further provided with a playing function for playing the audio file corresponding to the specified recognition sentence information. Specifically, first, an audio play button is displayed at a fifth position corresponding to the above-described specified recognition sentence information. The user can operate (for example, click or double click) the audio play button to input a first operation instruction for the audio play button to the data presentation device. The fifth position corresponding to the designated recognized sentence information is not particularly limited, and may be provided at any position in a cell containing the designated recognized sentence information, for example. And then judging whether a first operation instruction for the audio playing button is detected. The first operation command may be determined to be detected by determining whether or not a second operation in which the user operates the edit button is received. If the second operation action is received, it is determined that the first operation command is detected, and if the second operation action is not received, it is determined that the first operation command is not detected. And if the first operation instruction for the audio playing button is detected, responding to the first operation instruction, and acquiring the audio file corresponding to the appointed recognition sentence information from the recording file. The audio file corresponding to the specified recognized sentence information is the specified speech sentence data, and the audio file corresponding to the specified recognized sentence information is searched from the recording file according to the recording start time corresponding to the specified speech sentence data and the recording length information corresponding to the specified speech sentence data, which is the specified speech sentence data. In addition, after the recording file is segmented into a plurality of pieces of speech sentence data, each piece of speech sentence data can be stored in the pre-created data storage module, so that the designated speech sentence data corresponding to the designated recognition sentence information can be extracted from the data storage module. And after the audio file corresponding to the appointed identification sentence information is extracted, playing the audio file corresponding to the appointed identification sentence information. Wherein, the inquired audio file can be played through a loudspeaker. In the embodiment, the audio playing button is displayed at the fifth position corresponding to the appointed recognition sentence information, so that the user can use the audio playing button to screen out the interested audio file corresponding to the recognition sentence information from the recording file for playing, and the use experience of the user is effectively improved.
In an embodiment of the application, after the step S3, the method includes:
s330: displaying an edit button on a sixth position corresponding to the specified recognition sentence information;
s331: judging whether a second operation instruction for the editing button is detected;
s332: if the second operation instruction for the editing button is detected, responding to the second operation instruction and receiving input editing information;
s333: and according to the editing information, performing adjustment processing corresponding to the editing information on the appointed recognition sentence information.
As described in the above steps S330 to S333, upon receiving a viewing instruction for viewing the recording recognition result corresponding to the recording file, in addition to displaying the recording start time of each piece of speech sentence data and the recognition sentence information one-to-one corresponding to each piece of speech sentence data to the user, an editing function may be further configured for each recognition sentence information, so that the user can edit any one piece of recognition sentence information according to his own actual needs. Specifically, an edit button is first displayed at a sixth position corresponding to the specified recognition sentence information. Wherein, the user can operate (for example, click or double click) the editing button to input a second operation instruction for the editing button to the data presentation device. The sixth position corresponding to the designated recognized sentence information is not particularly limited, and may be provided at any position in a cell containing the designated recognized sentence information, for example. And then judging whether a second operation instruction for the editing button is detected. The determination of whether the second operation command is detected may be made by determining whether a third operation action of the user operating the edit button is received. And if the third operation action is received, determining that the second operation instruction is detected, and if the third operation action is not received, determining that the second operation instruction is not detected. And if the second operation instruction for the editing button is detected, receiving the editing information input by the user in response to the second operation instruction. The editing information may be modification information for modifying the specified recognition sentence information by the user, or may also be annotation information for adding annotation to the specified recognition sentence information by the user. And after receiving the editing information input by the user, correspondingly adjusting the information of the appointed recognition sentence according to the editing information. In the embodiment, the editing button is displayed at the sixth position corresponding to the appointed recognition sentence information, so that the user can correspondingly edit the appointed recognition sentence information according to the actual requirement of the user, and the use experience of the user is effectively improved.
Further, in an embodiment of the present application, after the step S3, the method includes:
s340: calling a preset translation module to translate the appointed recognition sentence information to obtain translation information corresponding to the appointed recognition sentence information;
s341: acquiring the translated text information output by the translation module;
s342: displaying the translated version information of the specified recognition sentence information at a seventh position corresponding to the specified recognition sentence information.
As described in the above steps S340 to S342, when the viewing instruction for viewing the recording recognition result corresponding to the recording file is received, in addition to the recording start time of each piece of speech sentence data and the recognition sentence information corresponding to each piece of speech sentence data, the translation information corresponding to each piece of recognition sentence information may be further displayed to the user, so that the user can clearly view the translation information related to any one piece of recognition sentence information. Specifically, a preset translation module is called to perform corresponding translation processing on the specified recognition sentence information to obtain translated text information corresponding to the specified recognition sentence information. The translation module is not particularly limited, and may be set according to actual requirements, for example, the translation module may be an english-to-english translation module, a middle-to-english translation module, or the like. And then acquiring the translated text information output by the translation module. And displaying the translated sentence information of the designated recognized sentence information at a seventh position corresponding to the designated recognized sentence information after obtaining the translated sentence information. The seventh position corresponding to the specified recognition sentence information is specifically a position in the same dimension (which may be the same row or the same column) as the placement position of the specified recognition sentence information, and a relationship exists between the placement position of the translation information of the specified speech sentence data (i.e., the seventh position) and the placement position of the specified recognition sentence information (i.e., the specified position). In addition, an edit button may be generated at the eighth position corresponding to the specified translated text information, and the user may perform relevant editing processing on the translated text information according to the actual needs of the user, for example, modify or add annotation information to the specified translated text information, and so on. The specified translated sentence information is translated sentence information corresponding to the specified recognized sentence information, and the eighth position corresponding to the specified translated sentence information is not particularly limited, and may be set at any position in a cell containing the specified translated sentence information. Furthermore, translation explanation information corresponding to the specified translation information can be acquired, and the translation explanation information is displayed at a ninth position corresponding to the specified translation information. In addition, the setting manner of the ninth position corresponding to the specified translation information may be according to the seventh position, and will not be described herein again. When a viewing instruction for viewing the recording identification result corresponding to the recording file is received, translation information corresponding to each piece of identification sentence information is further displayed on the basis of displaying the recording start time of each piece of voice sentence data in the recording file and displaying the identification sentence information corresponding to each piece of voice sentence data one to one, so that a user can look up more information related to the recording identification result of the recording file, the richness of data display of the recording identification result is effectively improved, and the use experience of the user is improved.
Further, there is an association relationship between the sound recording file and the specified text information, where the specified text information may include a reading text (or may also be referred to as a sound recording text) used by the user for recording voice, and the sound recording file is a file generated by the user by reading the reading text. When a viewing instruction for viewing the recording recognition result corresponding to the recording file is received, in addition to displaying the recording start time of each piece of speech sentence data and the recognition sentence information corresponding to each piece of speech sentence data one to the user, the usage learning times and the latest usage time corresponding to each piece of recognition sentence information may be further displayed so that the user can clearly know the learning information related to the recognition sentence information. Specifically, the number of usage learning times corresponding to the above-mentioned specified recognition sentence information is acquired first, and the most recent usage time corresponding to the above-mentioned specified recognition sentence information is acquired. And then displaying the use learning times and the latest use time at a ninth position corresponding to the appointed recognition sentence information so as to realize the function of providing learning reminding for the user. When a viewing instruction for viewing the recording identification result corresponding to the recording file is received, the recording start time of each piece of voice sentence data in the recording file is displayed, and the use learning times and the latest use time corresponding to each piece of voice sentence data are further displayed on the basis of the identification sentence information corresponding to each piece of voice sentence data one by one, so that a user can look up more information related to the recording identification result of the recording file, the richness of data display of the recording identification result is effectively improved, and the use experience of the user is improved.
Further, when the specified text information may include the reading text (or may also be referred to as recording text) used by the user for voice recording, the recording file is a file generated by the user by reading the reading text. When a viewing instruction for viewing the recording recognition result corresponding to the recording file is received, the recording start time of each piece of voice sentence data and the recognition sentence information respectively corresponding to each piece of voice sentence data can be displayed for the user, and voice evaluation and correction services can be further provided for the user. Specifically, the specified speech sentence data and specified recognition sentence information corresponding to the specified speech sentence data are acquired; inputting the specified voice sentence data and the specified recognition sentence information into a preset voice evaluating module, so as to evaluate and process the specified voice sentence data and the specified recognition sentence information through the voice evaluating module and output a corresponding evaluating result; and obtaining and displaying the evaluation result. The evaluation result generally comprises the following contents: the text of the request; sentence start time in audio, in seconds; sentence ending time in audio, the unit being seconds; scoring sentence integrity; sentence fluency score; a sentence accuracy score; speed of speech, word/minute; comprehensively scoring sentences; a word score array; a word; word start time in seconds; word end time in seconds; a word accuracy score; phonetic symbol array; phonetic symbols; phonetic symbol start time in seconds; phonetic symbol end time in seconds; judging whether the phoneme is wrong or not, wherein true is the pronunciation is correct, false is the pronunciation error, and simultaneously giving a prompt; if the pronunciation is wrong, prompting the user what the pronunciation looks like; the degree of accent, the higher the score, the more likely the current phonetic symbol is accent, the score is [0, 100 ]; if the vowel stress reference/standard answer is true, the reference answer is explained to be that the vowel is supposed to be stressed, and the consonant is meaningless; in a word, the user pronounces the phonetic symbol as an accent, and so on. In addition, the voice evaluating module is not particularly limited, for example, the voice evaluating module may use a smart voice evaluating module with channels. In the embodiment, when a viewing instruction for viewing the recording recognition result corresponding to the recording file is received, the recording start time of each piece of voice sentence data in the displayed recording file and the recognition sentence information corresponding to each piece of voice sentence data in a one-to-one manner are further evaluated and processed by the voice evaluation module and the corresponding evaluation result is displayed, so that a user can look up more information related to the recording recognition result of the recording file, the richness of data display of the recording recognition result is effectively improved, and the use experience of the user is improved.
Referring to fig. 2, an embodiment of the present application further provides a data display apparatus, including:
the first receiving module 1 is used for receiving a checking instruction for checking a recording identification result corresponding to the recording file;
a first obtaining module 2, configured to obtain all pieces of recognition sentence information obtained by performing speech recognition processing on each piece of speech sentence data in the recording file, and obtain a recording start time of each piece of speech sentence data, where one piece of speech sentence data corresponds to one piece of recognition sentence information;
the first presentation module 3 is configured to present a recording start time of specified speech sentence data, and present specified recognition sentence information corresponding to the specified speech sentence data at a specified position corresponding to the recording start time of the specified speech sentence data, where the specified speech sentence data is any one piece of speech sentence data in all the speech sentence data.
In this embodiment, the implementation processes of the functions and actions of the first receiving module, the first obtaining module and the first displaying module in the data displaying apparatus are specifically described in the implementation processes corresponding to steps S1 to S3 in the data displaying method, and are not described herein again.
Further, in an embodiment of the present application, the data display apparatus includes:
a second obtaining module, configured to obtain sentence type information of the specified speech sentence data;
a second presentation module for presenting sentence type information of the specified speech sentence data at a first position corresponding to the specified recognized sentence information.
In this embodiment, the implementation processes of the functions and actions of the second obtaining module and the second displaying module in the data displaying apparatus are specifically described in the implementation processes corresponding to steps S300 to S301 in the data displaying method, and are not described herein again.
Further, in an embodiment of the present application, the data display apparatus includes:
a first judging module, configured to judge whether sentence type information of the specified speech sentence data is a silent sentence type;
a generation module configured to generate a display switch button at a second position corresponding to the specified recognition sentence information if the sentence type information of the specified speech sentence data is judged to be a silent sentence type;
the second judgment module is used for judging whether a closing operation instruction for the display switch button is received or not;
and the hiding module is used for hiding all display information corresponding to the specified voice sentence data if the closing operation instruction of the display switch button is judged to be received.
In this embodiment, the implementation processes of the functions and functions of the first determining module, the generating module, the second determining module and the hiding module in the data displaying apparatus are specifically described in the implementation processes corresponding to steps S302 to S305 in the data displaying method, and are not described herein again.
Further, in an embodiment of the present application, the data display apparatus includes:
a third obtaining module, configured to obtain recording time length information of the specified speech sentence data; and the number of the first and second groups,
a fourth obtaining module, configured to obtain identification information of the specified recognition sentence information;
a third presentation module for presenting recording time length information of the specified voice sentence data at a third position corresponding to the specified voice sentence data; and the number of the first and second groups,
a fourth presentation module for presenting the identification information of the specified recognition sentence information at a fourth position corresponding to the specified recognition sentence information.
In this embodiment, the implementation processes of the functions and functions of the third obtaining module, the fourth obtaining module, the third displaying module and the fourth displaying module in the data displaying apparatus are specifically described in the implementation processes corresponding to steps S310 to S313 in the data displaying method, and are not described herein again.
Further, in an embodiment of the present application, the data display apparatus includes:
a first display module for displaying an audio play button at a fifth position corresponding to the designated recognized sentence information;
the third judging module is used for judging whether a first operation instruction for the audio playing button is detected or not;
the extraction module is used for responding to a first operation instruction if the first operation instruction for the audio playing button is detected, and extracting an audio file corresponding to the appointed recognition sentence information from the recording file;
and the playing module is used for playing the audio file corresponding to the appointed recognition sentence information.
In this embodiment, the implementation processes of the functions and functions of the first display module, the third determination module, the extraction module and the playing module in the data display device are specifically described in the implementation processes corresponding to steps S320 to S323 in the data display method, and are not described herein again.
Further, in an embodiment of the present application, the data display apparatus includes:
a second display module for displaying an edit button at a sixth position corresponding to the specified recognition sentence information;
the fourth judging module is used for judging whether a second operation instruction for the editing button is detected or not;
the second receiving module is used for responding to a second operation instruction and receiving input editing information if the second operation instruction for the editing button is detected;
and the editing module is used for adjusting the specified recognition sentence information corresponding to the editing information according to the editing information.
In this embodiment, the implementation processes of the functions and functions of the second display module, the fourth determination module, the second receiving module and the editing module in the data display device are specifically described in the implementation processes corresponding to steps S330 to S333 in the data display method, and are not described herein again.
Further, in an embodiment of the present application, the data display apparatus includes:
the calling module is used for calling a preset translation module to translate the appointed recognition sentence information to obtain translation information corresponding to the appointed recognition sentence information;
the fifth acquisition module is used for acquiring the translated text information output by the translation module;
and the fifth display module is used for displaying the translation information of the appointed recognition sentence information at a seventh position corresponding to the appointed recognition sentence information.
In this embodiment, the implementation processes of the functions and functions of the calling module, the fifth obtaining module and the fifth displaying module in the data displaying apparatus are specifically described in the implementation processes corresponding to steps S340 to S342 in the data displaying method, and are not described herein again.
Referring to fig. 3, a computer device, which may be a server and whose internal structure may be as shown in fig. 3, is also provided in the embodiment of the present application. The computer device includes a processor, a memory, a network interface, and a database connected by a system bus. Wherein the processor of the computer device is designed to provide computing and control capabilities. The memory of the computer device comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, a computer program, and a database. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The database of the computer device is used for storing data such as speech sentence data, recognition sentence information, and recording start time. The network interface of the computer device is used for communicating with an external terminal through a network connection. The computer program is executed by a processor to implement a data presentation method.
The processor executes the steps of the data display method:
receiving a viewing instruction for viewing a recording identification result corresponding to the recording file;
acquiring all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file, and acquiring the recording start time of each piece of voice sentence data, wherein one piece of voice sentence data corresponds to one piece of recognition sentence information;
displaying a recording start time of specified voice sentence data, and displaying specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, wherein the specified voice sentence data is any one piece of voice sentence data in all the voice sentence data.
Those skilled in the art will appreciate that the structure shown in fig. 3 is only a block diagram of a part of the structure related to the present application, and does not constitute a limitation to the apparatus and the computer device to which the present application is applied.
An embodiment of the present application further provides a computer-readable storage medium, on which a computer program is stored, where when the computer program is executed by a processor, the computer program implements a data display method, and specifically:
receiving a viewing instruction for viewing a recording identification result corresponding to the recording file;
acquiring all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file, and acquiring the recording start time of each piece of voice sentence data, wherein one piece of voice sentence data corresponds to one piece of recognition sentence information;
displaying a recording start time of specified voice sentence data, and displaying specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, wherein the specified voice sentence data is any one piece of voice sentence data in all the voice sentence data.
To sum up, the data presentation method, the data presentation device, the computer device, and the storage medium provided in the embodiments of the present application receive a viewing instruction for viewing a recording identification result corresponding to a recording file; acquiring all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file, and acquiring the recording start time of each piece of voice sentence data, wherein one piece of voice sentence data corresponds to one piece of recognition sentence information; displaying a recording start time of specified voice sentence data, and displaying specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, wherein the specified voice sentence data is any one piece of voice sentence data in all the voice sentence data. When receiving a viewing instruction for viewing a recording identification result corresponding to a recording file, the embodiment of the application enables a user to not only look up the recording identification result of the recording file, but also look up more information related to the recording identification result of the recording file by showing the recording start time of each piece of voice sentence data in the recording file and the identification sentence information corresponding to each piece of voice sentence data one to one, thereby effectively improving the richness of data display of the recording identification result and improving the use experience of the user.
It will be understood by those skilled in the art that all or part of the processes of the methods of the above embodiments may be implemented by hardware associated with instructions of a computer program, which may be stored on a non-volatile computer-readable storage medium, and when executed, may include processes of the above embodiments of the methods. Any reference to memory, storage, database, or other medium provided herein and used in the examples may include non-volatile and/or volatile memory. Non-volatile memory can include read-only memory (ROM), Programmable ROM (PROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), double-rate SDRAM (SSRSDRAM), Enhanced SDRAM (ESDRAM), synchronous link (Synchlink) DRAM (SLDRAM), Rambus Direct RAM (RDRAM), direct bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, apparatus, article, or method that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, apparatus, article, or method. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, apparatus, article, or method that includes the element.
The above description is only a preferred embodiment of the present application, and not intended to limit the scope of the present application, and all modifications of equivalent structures and equivalent processes, which are made by the contents of the specification and the drawings of the present application, or which are directly or indirectly applied to other related technical fields, are also included in the scope of the present application.

Claims (10)

1. A method for displaying data, comprising:
receiving a viewing instruction for viewing a recording identification result corresponding to the recording file;
acquiring all recognition sentence information obtained after voice recognition processing is carried out on each piece of voice sentence data in the recording file, and acquiring the recording start time of each piece of voice sentence data, wherein one piece of voice sentence data corresponds to one piece of recognition sentence information;
displaying a recording start time of specified voice sentence data, and displaying specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, wherein the specified voice sentence data is any one piece of voice sentence data in all the voice sentence data.
2. The data presentation method according to claim 1, wherein the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, is followed by:
obtaining sentence type information of the specified voice sentence data;
sentence type information of the specified speech sentence data is presented at a first position corresponding to the specified recognized sentence information.
3. The data presentation method according to claim 2, wherein said step of presenting sentence type information of the specified speech sentence data at a first position corresponding to the specified recognized sentence information is followed by:
judging whether the sentence type information of the specified voice sentence data is a silent sentence type;
if the sentence type information of the specified voice sentence data is judged to be the silent sentence type, generating a display switch button at a second position corresponding to the specified recognition sentence information;
judging whether a closing operation instruction for the display switch button is received or not;
and if the closing operation instruction of the display switch button is judged to be received, hiding all display information corresponding to the specified voice sentence data.
4. The data presentation method according to claim 1, wherein the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, is followed by:
acquiring recording time length information of the specified voice sentence data; and the number of the first and second groups,
acquiring identification information of the appointed recognition sentence information;
displaying recording time length information of the specified voice sentence data at a third position corresponding to the specified voice sentence data; and the number of the first and second groups,
displaying identification information of the specified recognition sentence information at a fourth position corresponding to the specified recognition sentence information.
5. The data presentation method according to claim 1, wherein the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, is followed by:
displaying an audio play button at a fifth position corresponding to the designated recognized sentence information;
judging whether a first operation instruction for the audio playing button is detected;
if the first operation instruction for the audio playing button is detected, responding to the first operation instruction, and extracting an audio file corresponding to the appointed recognition sentence information from the recording file;
and playing an audio file corresponding to the appointed recognition sentence information.
6. The data presentation method according to claim 1, wherein the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, is followed by:
displaying an edit button on a sixth position corresponding to the specified recognition sentence information;
judging whether a second operation instruction for the editing button is detected;
if the second operation instruction for the editing button is detected, responding to the second operation instruction and receiving input editing information;
and according to the editing information, performing adjustment processing corresponding to the editing information on the appointed recognition sentence information.
7. The data presentation method according to claim 1, wherein the step of presenting the recording start time of the specified voice sentence data and presenting the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, is followed by:
calling a preset translation module to translate the appointed recognition sentence information to obtain translation information corresponding to the appointed recognition sentence information;
acquiring the translated text information output by the translation module;
displaying the translated version information of the specified recognition sentence information at a seventh position corresponding to the specified recognition sentence information.
8. A data presentation device, comprising:
the first receiving module is used for receiving a checking instruction for checking a recording identification result corresponding to the recording file;
a first obtaining module, configured to obtain all pieces of recognition sentence information obtained by performing speech recognition processing on each piece of speech sentence data in the recording file, and obtain a recording start time of each piece of speech sentence data, where one piece of speech sentence data corresponds to one piece of recognition sentence information;
the system comprises a first display module and a second display module, wherein the first display module is used for displaying the recording start time of specified voice sentence data and displaying the specified recognition sentence information corresponding to the specified voice sentence data at a specified position corresponding to the recording start time of the specified voice sentence data, and the specified voice sentence data is any one section of voice sentence data in the voice sentence data.
9. A computer device comprising a memory and a processor, the memory having stored therein a computer program, characterized in that the processor, when executing the computer program, implements the steps of the method according to any one of claims 1 to 7.
10. A storage medium having a computer program stored thereon, wherein the computer program is
The program when executed by a processor implements the steps of the method of any of claims 1 to 7.
CN202010451239.6A 2020-05-25 2020-05-25 Data display method and device, computer equipment and storage medium Pending CN111353038A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010451239.6A CN111353038A (en) 2020-05-25 2020-05-25 Data display method and device, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010451239.6A CN111353038A (en) 2020-05-25 2020-05-25 Data display method and device, computer equipment and storage medium

Publications (1)

Publication Number Publication Date
CN111353038A true CN111353038A (en) 2020-06-30

Family

ID=71197771

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010451239.6A Pending CN111353038A (en) 2020-05-25 2020-05-25 Data display method and device, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN111353038A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112151080A (en) * 2020-10-28 2020-12-29 成都启英泰伦科技有限公司 Method for recording and processing training corpus

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101101590A (en) * 2006-07-04 2008-01-09 王建波 Sound and character correspondence relation table generation method and positioning method
CN109309751A (en) * 2017-07-28 2019-02-05 腾讯科技(深圳)有限公司 Voice recording method, electronic equipment and storage medium
CN109379641A (en) * 2018-11-14 2019-02-22 腾讯科技(深圳)有限公司 A kind of method for generating captions and device
CN110335612A (en) * 2019-07-11 2019-10-15 招商局金融科技有限公司 Minutes generation method, device and storage medium based on speech recognition
CN110534114A (en) * 2019-08-30 2019-12-03 上海互盾信息科技有限公司 A method of it first identifies when translating voice document on webpage and translates again

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101101590A (en) * 2006-07-04 2008-01-09 王建波 Sound and character correspondence relation table generation method and positioning method
CN109309751A (en) * 2017-07-28 2019-02-05 腾讯科技(深圳)有限公司 Voice recording method, electronic equipment and storage medium
CN109379641A (en) * 2018-11-14 2019-02-22 腾讯科技(深圳)有限公司 A kind of method for generating captions and device
CN110335612A (en) * 2019-07-11 2019-10-15 招商局金融科技有限公司 Minutes generation method, device and storage medium based on speech recognition
CN110534114A (en) * 2019-08-30 2019-12-03 上海互盾信息科技有限公司 A method of it first identifies when translating voice document on webpage and translates again

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112151080A (en) * 2020-10-28 2020-12-29 成都启英泰伦科技有限公司 Method for recording and processing training corpus
CN112151080B (en) * 2020-10-28 2021-08-03 成都启英泰伦科技有限公司 Method for recording and processing training corpus

Similar Documents

Publication Publication Date Title
US7054817B2 (en) User interface for speech model generation and testing
US20190196666A1 (en) Systems and Methods Document Narration
US8954328B2 (en) Systems and methods for document narration with multiple characters having multiple moods
US8793133B2 (en) Systems and methods document narration
US8155958B2 (en) Speech-to-text system, speech-to-text method, and speech-to-text program
JP5756555B1 (en) Utterance evaluation apparatus, utterance evaluation method, and program
US20020123894A1 (en) Processing speech recognition errors in an embedded speech recognition system
CN111541904B (en) Information prompting method, device, equipment and storage medium in live broadcast process
KR20150014236A (en) Apparatus and method for learning foreign language based on interactive character
Davel et al. Pronunciation dictionary development in resource-scarce environments
US20080177542A1 (en) Voice Recognition Program
WO2010133072A1 (en) Pronunciation evaluating device and method
CN112805734A (en) Speech assistance device for calling attention to speech inhibition words
KR102225435B1 (en) Language learning-training system based on speech to text technology
CN111353038A (en) Data display method and device, computer equipment and storage medium
US20230186895A1 (en) Method for performing synthetic speech generation operation on text
KR20140107067A (en) Apparatus and method for learning word by using native speakerpronunciation data and image data
JP2007071904A (en) Speaking learning support system by region
CN113393831B (en) Speech input operation method based on at least diphones and computer readable medium
JP2007122510A (en) Presenting device and presenting program
JPH0916597A (en) Sentence elaboration device and method therefor
JP2019211496A (en) Language learning support system
JP2020118872A (en) Information input system and method
CN117219062A (en) Training data generation method and device, electronic equipment and storage medium
CN107924641A (en) Utilize the interactive learning methods and its device of electronic pen

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200630

RJ01 Rejection of invention patent application after publication