WO2021153403A1 - Text information editing device and text information editing method - Google Patents

Text information editing device and text information editing method Download PDF

Info

Publication number
WO2021153403A1
WO2021153403A1 PCT/JP2021/001975 JP2021001975W WO2021153403A1 WO 2021153403 A1 WO2021153403 A1 WO 2021153403A1 JP 2021001975 W JP2021001975 W JP 2021001975W WO 2021153403 A1 WO2021153403 A1 WO 2021153403A1
Authority
WO
WIPO (PCT)
Prior art keywords
utterance
text
text information
editing
information editing
Prior art date
Application number
PCT/JP2021/001975
Other languages
French (fr)
Japanese (ja)
Inventor
太亮 尾崎
祐太 是枝
皓文 森下
Original Assignee
株式会社日立製作所
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社日立製作所 filed Critical 株式会社日立製作所
Publication of WO2021153403A1 publication Critical patent/WO2021153403A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • G06F40/35Discourse or dialogue representation

Definitions

  • the present invention relates to a text information editing device and a text information editing method.
  • Patent Document 1 As a background technology in this technical field, there is International Publication No. 2014/20298 (Patent Document 1).
  • Patent Document 1 “As a sentence classification device, a classification target section extraction unit that extracts and processes sections that contribute to sentence classification from input sentences that are textualized conversation contents to be classified based on clue sentences, and a classification target section.
  • a sentence classification unit for determining which classification the input sentence belongs to using the text of the section contributing to the sentence classification extracted by the extraction means is provided. ”(See summary).
  • Patent Document 1 it is necessary that the classification method and the viewpoint of classification for classifying sentences are defined in advance. Further, the invention described in Patent Document 1 is applicable only to a semi-standard dialogue text such as a response in a call center, and it is difficult to extract useful information from a general dialogue text.
  • one aspect of the present invention aims to easily obtain useful information based on an arbitrary viewpoint from the input dialogue text.
  • the text information editing device has a processor and a storage device, and is connected to a display device.
  • the storage device includes a dialogue text consisting of one or more spoken texts, and at least one of an important expression and an extraction algorithm of the important expression.
  • the processor retains the editing processing information indicating the above, extracts important expressions from the one or more spoken texts with reference to the editing processing information, and obtains the dialogue text and the extracted important expressions. It is displayed on the display device, accepts edit input for at least one of the utterance text included in the dialogue text and the comment to which the utterance text is attached, and executes the edit corresponding to the edit input.
  • useful information based on an arbitrary viewpoint can be easily obtained from the input dialogue text.
  • FIG. This is an example of dialogue text data output by the dialogue input reception unit in the first embodiment. This is an example of data output by the important expression detection unit in the first embodiment to the interactive editing unit. It is a flowchart which shows an example of the text information editing process in Example 1.
  • FIG. This is an example of a display screen for interactive editing in the first embodiment. This is an example of a display screen for interactive editing in the second embodiment. This is an example of the display screen for interactive editing in the third embodiment.
  • FIG. This is an example of dialogue text data output by the topic boundary identification unit in the fourth embodiment.
  • Example 4 It is a flowchart which shows an example of the text information editing process in Example 4. This is an example of a display screen for interactive editing in the fourth embodiment. It is an example of the display screen of the interactive editing in Example 5. This is an example of the display screen for interactive editing in the sixth embodiment. It is an example of the display screen of the interactive editing in Example 7.
  • This embodiment describes a text information editing device.
  • the text information editing device handles Japanese text
  • the language used for the text is not limited.
  • embodiments of the present invention will be described in detail with reference to the drawings.
  • the same components are designated by the same reference numerals, and the repeated description will be omitted.
  • the present embodiment is merely an example for realizing the present invention and does not limit the technical scope of the present invention.
  • FIG. 1 is a block diagram showing a configuration example of a text information editing device.
  • the text information editing device 100 is composed of, for example, a computer having a CPU (Central Processing Unit) 101, a memory 102, an auxiliary storage device 103, and a communication device 104.
  • CPU Central Processing Unit
  • the CPU 101 includes a processor and executes a program stored in the memory 102.
  • the memory 102 includes a ROM (Read Only Memory) which is a non-volatile storage element and a RAM (Random Access Memory) which is a volatile storage element.
  • the ROM stores an invariant program (for example, BIOS (Basic Input / Output System)) and the like.
  • BIOS Basic Input / Output System
  • the RAM is a high-speed and volatile storage element such as a DRAM (Dynamic Random Access Memory), and temporarily stores a program executed by the CPU 101 and data used when the program is executed.
  • the auxiliary storage device 103 is, for example, a large-capacity and non-volatile storage device such as a magnetic storage device (HDD (Hard Disk Drive)) or a flash memory (SSD (Solid State Drive)), and is a program and a program executed by the CPU 101. Stores the data used when executing. That is, the program is read from the auxiliary storage device 103, loaded into the memory 102, and executed by the CPU 101.
  • HDD Hard Disk Drive
  • SSD Solid State Drive
  • the text information editing device 100 may have an input interface 105 and an output interface 108.
  • the input interface 105 is an interface to which an input device such as a keyboard 106 or a mouse 107 is connected and receives input from an operator.
  • the output interface 108 is an interface to which a display device 109 such as a printer or a display is connected and outputs a program execution result in a format that can be visually recognized by an operator. Further, for example, a microphone for inputting a voice dialogue may be connected to the input interface 105 as an input device.
  • the communication device 104 is a network interface device that controls communication with other devices according to a predetermined protocol. Further, the communication device 104 may include a serial interface such as USB.
  • the program executed by the CPU 101 is provided to the text information editing device 100 via a removable medium (CD-ROM, flash memory, etc.) or a network, and is stored in the non-volatile auxiliary storage device 103, which is a non-temporary storage medium. .. Therefore, the text information editing device 100 may have an interface for reading data from removable media.
  • the text information editing device 100 is a computer system composed of physically one computer or a plurality of computers logically or physically configured, and operates in separate threads on the same computer. It may operate on a virtual computer built on a plurality of physical computer resources. The same applies to the terminal 200.
  • the input device connected to the text information editing device 100 inputs to the text information editing device 100, but another computer such as a tablet terminal owned by each participant in the dialogue, for example. Input may be made from.
  • the calculation result by the text information editing device 100 is output to the display device 109 connected to the text information editing device 100, but for example, a tablet terminal owned by each participant in the dialogue. The calculation result may be output to another computer.
  • the CPU 101 includes, for example, an interactive input receiving unit 111, an important expression detecting unit 112, an editing input receiving unit 113, an interactive editing unit 114, an output information generating unit 115, and an output unit 116.
  • the dialogue input reception unit 111 receives the input of the dialogue text via the input device connected to the input interface 105, and divides the dialogue text into utterance units.
  • Texts transcribed from dialogues and conferences, text chats, texts in web bulletin boards and forums, and texts in e-mail and SMS are all examples of dialogue texts.
  • the dialogue text in the present embodiment is composed of utterance texts by a plurality of humans, but a text composed of one or more utterance texts by one or more humans may be used as the dialogue text.
  • the important expression detection unit 112 detects important expressions such as important utterances, phrases, and a plurality of utterances from the utterance text.
  • the edit input reception unit 113 accepts the input of the annotation attached in the utterance text.
  • the interactive editing unit 114 adds the annotation input to the editing input receiving unit 113 to the utterance text.
  • the output information generation unit 115 generates information for outputting the annotated utterance text to the display device 109.
  • the output unit 116 outputs the utterance received by the dialogue input reception unit 111, the important expression extracted by the important expression detection unit 112, the editing result report by the interactive editing unit 114, and the like to the display device 109.
  • the output unit 116 provides the user with a user output result in a form that can be used by the user, such as a GUI (Graphical User Interface), a text file, an image file, or an audio file.
  • GUI Graphic User Interface
  • the CPU 101 functions as the dialogue input reception unit 111 by operating according to the dialogue input program loaded in the memory 102, and operates according to the important expression detection program loaded in the memory 102, so that the important expression detection unit 112 Functions as.
  • the relationship between the program and the functional unit is the same for the other functional units included in the CPU 101.
  • CPU 101 may be realized by hardware such as ASIC (Application Specific Integrated Circuit) or FPGA (Field-Programmable Gate Array).
  • ASIC Application Specific Integrated Circuit
  • FPGA Field-Programmable Gate Array
  • the auxiliary storage device 103 holds, for example, the utterance annotation DB (DataBase) 121 and the editing processing information DB 122. Note that some or all of the information stored in the auxiliary storage device 103 may be described in the program stored in the memory 102.
  • the utterance annotation DB 121 holds the dialogue text input to the dialogue input receiving unit 111 and the annotation given by the interactive editing unit 114.
  • the editing processing information DB 122 holds information such as an algorithm for detecting an important expression from the uttered text and a correspondence between a handy word that can be an important expression and the content of an annotation to be given.
  • FIG. 2 is an example of dialogue text data output by the dialogue input reception unit 111.
  • the dialogue input reception unit 111 receives the input of the dialogue text via the input interface 105, divides the input dialogue text into utterance texts, assigns an utterance ID that can uniquely identify the utterance, and sets the utterance text and the utterance ID. Correspondence is transmitted to the output unit 116.
  • the output unit 116 displays the dialogue text data in which the utterance ID and the utterance text are associated with each other on the display device 109.
  • FIG. 3 is an example of data output by the important expression detection unit 112 to the interactive editing unit 114.
  • the important expression detection unit 112 adds an important expression detected from each utterance text as an annotation.
  • the number of important expressions included in one utterance does not have to be one or less, and may be plural, or important expressions may be extracted across a plurality of utterances.
  • the utterance annotation DB 121 retains the correspondence between the utterance ID, the utterance text, and the annotation shown in FIG. Further, the utterance annotation DB 121 may include a dialogue ID indicating which dialogue text includes each utterance text.
  • FIG. 4 is a flowchart showing an example of the text information editing process.
  • the text information editing process is executed, for example, every time the dialogue input receiving unit 111 accepts the input of the dialogue text.
  • the dialogue input reception unit 111 may input, for example, the start timing and the end timing of the dialogue, or the dialogue is performed after a predetermined time has elapsed from the previous utterance. It is determined that it is the end timing of the dialogue, and when the first utterance after the end timing of the dialogue starts, it is determined to be the start timing of the dialogue.
  • the dialogue input receiving unit 111 determines that the text in the text file is one dialogue, for example.
  • the dialogue input reception unit 111 generates an utterance text by dividing the input dialogue text into utterance units, assigns a utterance ID to each of the generated utterance texts, and utters the utterance text and the utterance ID. It is stored in the DB 121 (S401).
  • the dialogue input reception unit 111 divides the dialogue text at a point where the utterance time interval is more than a predetermined time when the dialogue text by voice is input. Further, for example, the dialogue input reception unit 111 may receive a predetermined character code (for example, a punctuation mark, an exclamation mark, a space, or a space) in the text when the dialogue text is input by the text file. Divide the dialogue text before and after (enter, etc.). In addition, the dialogue text divided for each utterance may be input to the dialogue input reception unit 111 in advance, and in this case, the process of step S401 may not be performed.
  • a predetermined character code for example, a punctuation mark, an exclamation mark, a space, or a space
  • the dialogue input reception unit 111 may transmit the utterance ID, the utterance text, and the output unit 116, and the output unit 116 may output the received utterance ID and the utterance text, for example, as shown in FIG.
  • the important expression detection unit 112 detects the important expression by referring to each utterance text (S402). Specifically, for example, the important expression detection unit 112 extracts the clue word stored in the editing processing information DB 122 from the utterance text as an important expression. For example, in the utterance text in which the utterance ID of FIG. 3 is "1", the expressions "example” and “for example” are important expressions defined in advance in the editing processing information DB 122 (or described in the program). When a representation is given as an annotation, the annotation is an "example”. Further, for example, the important expression detection unit 112 may detect the important expression by applying a machine learning discriminator, TF-IDF, a statistical model, or the like to the dialogue text and / or the utterance text.
  • TF-IDF machine learning discriminator
  • the important expression detection unit 112 may add the detected important expression to the annotation of the utterance text and store it in the utterance annotation DB 121. In this case, the user deletes the important expression from the annotation. Remove the important expression from the annotation when instructed to do so.
  • the important expression detection unit 112 transmits the utterance ID, the utterance text, and the detected important expression to the output unit 116, and the output unit 116 sends the received utterance ID, the utterance text, and the important expression to the user. It is displayed on the display device 109 in an editable manner (S403). When the process of step S403 is completed, the user can edit the annotation, sort the utterances, search the utterances, and so on. On the display screen where editing input is possible, for example, an utterance in which an utterance ID is assigned to a text box on the GUI is displayed so that the user can easily recognize the text to be edited.
  • the interactive editing unit 114 determines whether or not the editing input receiving unit 113 has accepted the editing input from the user (S404).
  • the edit input receiving unit 113 may accept the user's input via the image and video, the sensor data, and the GUI such as click and tap in the dedicated application, in addition to the text and voice. The specific editing contents by the user will be described later.
  • the interactive editing unit 114 determines that the editing input receiving unit 113 has received the editing input from the user (S404: YES)
  • the interactive editing unit 114 performs editing according to the user's editing input, and stores the edited utterance and annotation in the utterance annotation DB 121. (S405).
  • the interactive editorial unit 114 can add, delete, and replace lines, and add and delete annotations to the data shown in FIG.
  • the information output in step S403 may continue to be output as it is.
  • the output information generation unit 115 generates information for displaying the edited utterance and annotation, and the output unit 116 displays the information on the display device 109 (S406).
  • the output information generating unit 115 provides information for displaying unedited utterances and comments.
  • the output unit 116 generates and displays the information on the display device 109 (S406).
  • the edit input receiving unit 113 may receive a plurality of times and a plurality of types of edit inputs in one text editing process. In that case, the interactive editing unit 114 performs sequential editing, and the output unit 116 thereof. The editing result may be output sequentially. As a result, the user can continue editing after confirming the editing results one by one.
  • FIG. 5 is an example of a display screen for interactive editing.
  • the display screen 500 shows a text box 501 and a text box 502 for displaying the spoken text, a text box 503 for showing the important expression detected by the important expression detection unit 112, and a text box 504 for showing the result of editing input by the user.
  • step S403 A text box 503 indicates whether the text should be annotated. Then, in step S404, the user makes an edit input indicating that the expression is an important expression in the text box 504.
  • the user can also input an annotation as an important expression for the utterance text in which the important expression is not detected by the important expression detection unit 112 in the edit input.
  • the annotation as an important expression input by the user is displayed in the text box 504.
  • the display screen may include the emotion icon 505.
  • the important expression detection unit 112 may perform the sentiment analysis of the dialogue from the text of the dialogue text according to a predetermined algorithm (and when the dialogue text is given by voice, from the frequency of the speaker's voice or the like. (Sentiment analysis may be performed), in this case, the result of the sentiment analysis is presented to the user by the sentiment icon 505.
  • the emotion icon 505 makes it easier for the user to determine whether or not the expression contained in the presented dialogue text is an important expression.
  • the emotion indicated by the emotion icon 505 may be determined for the entire dialogue text, or may be determined for each one or more utterances.
  • the important expression detection unit 112 when the result of the sentiment analysis for a certain utterance is a predetermined emotion (for example, "joy" that promotes discussion), the important expression detection unit 112 includes important information in the immediately preceding utterance. It can be regarded as an important expression and the utterance immediately before the said can be annotated.
  • the user may specify an emotion in the edit input and input the emotion.
  • the emotion icon 505 indicates the input emotion
  • the important expression detection unit 112 inputs the important expression based on the input emotion. It may be detected.
  • the user can make an utterance (edit input) as displayed in the text box 504 to express the important expression.
  • the user can input that there is an unnecessary utterance in the edit input (for example, the unnecessary utterance can be deleted from the utterance commentary DB 121 by performing an edit input for inputting the utterance ID of the unnecessary utterance).
  • the additional explanation is recorded in the utterance corresponding to the unclear utterance in the utterance comment DB. ) Can also be done.
  • the text information editing device 100 can proofread, summarize, and extract information from the dialogue text from the user's point of view.
  • FIG. 6 is an example of the interactive editing display screen output in step S406.
  • the display screen 500 includes a text box 501 and a text box 502 for displaying the spoken text, and a button 601 for rearranging the order of the spoken text.
  • the user can change the order of utterances by selecting button 601 in the edit input.
  • the output unit 116 has a display screen for inputting the order of each utterance text, or a position (for example,) in which the utterance text selected by the user is specified by the user. , Top or bottom) to display the display screen.
  • important utterances can be moved higher and less important utterances can be moved lower, and the dialogue text can be proofread, summarized, and information from the user's point of view. Extraction can be performed.
  • FIG. 7 is an example of the interactive editing display screen output in step S406.
  • the display screen 500 includes a text box 501 and a text box 502 for displaying the spoken text, and a card 701 for displaying important expressions.
  • the important expression detected by the important expression detection unit 112 is displayed on the card 701.
  • the user can save important utterances as annotations from any viewpoint such as homework, and proofread, summarize, and extract information from the dialogue text according to the user's viewpoint. It can be performed.
  • the card 701 may display an important expression specified by the user in the edit input.
  • FIG. 8 is a block diagram showing a configuration example of the text information editing device 100 of this embodiment.
  • the text information editing device 100 of the present embodiment is different from the text information editing device 100 of the first embodiment in that the CPU 101 includes the topic boundary identification unit 117.
  • the topic boundary identification unit 117 can identify one or more utterance texts (hereinafter, also referred to as utterance groups) belonging to the same topic by dividing the utterances included in the dialogue text by the topic boundary. For example, in a dialogue between a plurality of people, an utterance containing a clue word indicating an aizuchi or consent (for example, a word such as "yes” or "yes”) is made on the same topic as the immediately preceding utterance. Therefore, it is considered that the utterance showing the aizuchi or consent and the utterance immediately before it belong to the same group of topics.
  • utterance groups for example, a clue word indicating an aizuchi or consent
  • utterances that include clues and conjunctions that indicate a causal relationship are also conducted on the same topics as the immediately preceding utterance. Therefore, it is considered that the utterance showing the aizuchi or consent and the utterance immediately before it belong to the same group of topics.
  • the topic boundary identification unit 117 uses morphological analysis / syntactic analysis to identify an utterance that states a question and an utterance that states an answer corresponding to the question, and these utterances are the same. Determined to belong to the utterance group. Further, for example, the topic boundary identification unit 117 calculates the similarity between utterances by using parsing, and the utterances whose similarity is equal to or higher than a predetermined value have overlapping contents, so that the same utterance group is used. It may be determined that the utterances that have a similarity less than the predetermined value belong to a different utterance group (that is, a topic boundary is generated).
  • the topic boundary identification unit 117 may calculate the amount of information of the utterance and determine that the utterance with a small amount of information belongs to the same utterance group as the previous or subsequent utterance. For the amount of utterance information, for example, a predetermined weight may be applied to each speaker. Further, for example, the topic boundary identification unit 117 may specify the utterance group (topic boundary) by applying a classifier by machine learning or the like to the dialogue text and / or the utterance text.
  • the topic boundary identification unit 117 may specify the utterance group (topic boundary) by applying a predetermined weight to a part or all of the above-mentioned utterance group (topic boundary).
  • the topic boundary identification unit 117 can identify the topic of each utterance group by extracting the words and phrases that characterize the utterance group by using, for example, morphological analysis and syntactic analysis.
  • Information necessary to generate topic boundaries such as clues and algorithms that identify the boundaries of utterances as described above, clues and algorithms that identify that multiple utterances belong to the same utterance group, and topics.
  • Information necessary for specifying a topic such as an algorithm for specifying the above may be stored in advance in the editing process information DB 122, or may be described as a program.
  • the boundary between utterance groups is the topic boundary, but the utterance group indicated by this topic boundary is exclusive, that is, the utterances do not have to overlap between the utterance groups, and one utterance becomes a plurality of utterance groups. May belong.
  • FIG. 9 is an example of dialogue text data output by the topic boundary identification unit 117.
  • the dialogue input reception unit 111 receives the utterance ID and the utterance text, identifies the topic boundary from the utterance text, and transmits the topic boundary to the output unit 116 via the output information generation unit 115.
  • the dialogue text data of FIG. 9 includes a topic boundary that can identify the utterance at which the topic starts and the utterance at which the topic ends.
  • FIG. 10 is a flowchart showing an example of the text information editing process.
  • the topic boundary identification unit 117 identifies the topic boundary in the dialogue text by the method described above (S1001).
  • the topic boundary identification unit 117 transmits the topic boundary to the output information generation unit 115, the output information generation unit 115 generates output information indicating the topic boundary, and the output unit 116 displays the output information 109. It may be displayed in.
  • the process of step S1001 is completed, the process proceeds to step S402.
  • FIG. 11 is an example of the display screen output in step S406.
  • the display screen 500 includes the utterance group box 1101 and the utterance group box 1104, and the dialogue text is divided and displayed for each utterance group.
  • the utterance group box 1101 includes a text box 501 and a text box 502 that display the text of the utterance included in the utterance group, a topic box 1102 that indicates the topic of the utterance group, and a tag that characterizes the topic with respect to the topic. Includes button 1103 for presentation.
  • the topic box 1102 not the topic but the first utterance of the utterance group, the important expression in the utterance group detected by the important expression detection unit 112, and the like may be displayed.
  • a tag that characterizes the topic specified by the topic boundary identification unit 117 is presented, and the user can input information indicating whether or not the tag is appropriate in the edit input. It should be noted that only one tag may be attached to one utterance group, a plurality of tags may be attached, or no tag may be attached.
  • a pop-up display or the like may be performed using the button 1103 so that the user can freely input a tag indicating the topic according to the pop-up display in the edit input, for example, and the user can input a tag that characterizes the topic. You can freely save it as an annotation.
  • Tags can describe, for example, topical topics and special notes such as homework in free words and sentences.
  • the text information editing device 100 can investigate the utterance group in which a similar topic is spoken, for example, by using the feature amount of the utterance in the utterance group.
  • the feature amount the frequency of words appearing in the utterance group and the feature amount applicable to any sentence such as TF-IDF can be used.
  • the user can efficiently edit the dialogue text for similar utterance groups, and provide necessary information based on the topics and homework annotated as tags. It can be easily extracted and summarized. Further, the user can delete the tag attached to the similar utterance group by editing input.
  • the text information editing device 100 configured in this way, it is possible to proofread, summarize, and extract information from the dialogue text from the user's point of view.
  • FIG. 12 is an example of the interactive editing display screen output in step S406.
  • the display screen 500 includes the utterance group box 1101 and displays the dialogue text divided for each utterance group. Further, the display screen 500 includes a search result display box 1204.
  • the utterance group box 1101 includes a text box 501 and a text box 502 that display the text of the utterance included in the topic, a topic box 1102 that indicates the topic of the utterance group, a search box 1202, and a search button 1203.
  • the user can extract and rank utterance groups and utterances related to the utterance group by inputting a query such as a keyword in the search box 1202 and selecting the search button 1203 in the edit input.
  • a query such as a keyword in the search box 1202 and selecting the search button 1203 in the edit input.
  • the search result display box 1204 includes snippet 1205 and snippet 1206. Snippets 1205 and snippets 1206 can display important expressions extracted as search results as headings. As a result, the user can efficiently search for utterances and topic boundary ranges related to the utterance group.
  • the utterance or topic boundary range related to the utterance group can be linked to the utterance group, and the interactive editorial unit 114 can connect the result to the utterance group, for example. It can be stored in the utterance annotation DB 121 in the form of an annotation or the like.
  • FIG. 13 is an example of the interactive editing display screen output in step S406.
  • the display screen 500 includes the utterance group box 1101 and the utterance group box 1104, and the dialogue text is divided and displayed for each utterance group.
  • the utterance group box 1101 includes a text box 501 and a text box 502 for displaying the utterance text included in the topic, and a similar utterance display box 1301.
  • the topic boundary identification unit 117 extracts, for example, utterances that are desirable to belong to the utterance group from the utterances other than the series of utterances included in the utterance group box 1101 as similar utterances, and displays them in the similar utterance display box 1301. Specifically, for example, the topic boundary identification unit 117 extracts utterances other than the series of utterances as similar utterances if the degree of similarity with the series of utterances included in the utterance group box 1101 is equal to or higher than a predetermined value.
  • the interactive editorial unit 114 adds the similar utterance to the utterance group and changes the topic of the similar utterance to the topic corresponding to the utterance group. Add and save in the utterance annotation DB 121.
  • the text information editing device 100 configured in this way, it is possible to proofread, summarize, and extract information from the dialogue text from the user's point of view.
  • FIG. 14 is an example of the interactive editing display screen output in step S406.
  • the display screen 500 includes a topic box 1102 and a topic box 1105, and a text box 501 and a text box 502.
  • the utterance of the text box 501 displayed directly under the topic box 1102 is an utterance belonging to the utterance group indicating the topic of the topic box 1102, and the utterance of the text box 502 displayed directly under the topic box 1105 is a topic.
  • Box 1105 An utterance belonging to an utterance group indicating a topic.
  • the topic boundary identification unit 117 determines that the topic in the topic box 1102 and the utterance in the text box 501 are related, they are connected by a solid line edge. Similarly, since the topic boundary identification unit 117 determines that the topic of the topic box 1107 and the utterance of the text box 502 are related, they are connected by a solid line edge.
  • the display screen 500 includes an edge addition button 1401 and an edge deletion button 1402.
  • the user can add a dotted edge by selecting two topic boxes or text boxes and selecting the edge addition button 1401 in the edit input, and the interactive editorial unit 114 can add the two topics or utterances.
  • the utterances connected by the edge may be registered in the utterance commentary DB 121, or the utterances connected by the edge may be included in the same utterance group.
  • the user can delete the edge by selecting the edge and selecting the edge delete button 1402 in the edit input, and the interactive editorial unit 114 may delete the information about the edge from the utterance annotation DB 121. However, the utterance with the edge removed may be excluded from the utterance group.
  • the text information editing device 100 can connect utterances and utterance groups that are considered to be related from the user's point of view at an edge, and can accumulate this as a new annotation or change the utterance group.
  • the interactive editorial unit 114 edits so as to continuously display, for example, related utterances according to the user's designation.
  • Related utterances and utterance groups can be extracted as shown in Example 4 and Example 5.
  • the present invention is not limited to the above-described embodiment, and includes various modifications.
  • the above-described embodiment has been described in detail in order to explain the present invention in an easy-to-understand manner, and is not necessarily limited to the one including all the described configurations. It is also possible to replace a part of the configuration of one embodiment with the configuration of another embodiment, and it is also possible to add the configuration of another embodiment to the configuration of one embodiment. Further, it is possible to add / delete / replace a part of the configuration of each embodiment with another configuration.
  • each of the above configurations, functions, processing units, processing means, etc. may be realized by hardware by designing a part or all of them by, for example, an integrated circuit. Further, each of the above configurations, functions, and the like may be realized by software by the processor interpreting and executing a program that realizes each function. Information such as programs, tables, and files that realize each function can be placed in a memory, a hard disk, a recording device such as an SSD (Solid State Drive), or a recording medium such as an IC card, an SD card, or a DVD.
  • SSD Solid State Drive
  • control lines and information lines indicate those that are considered necessary for explanation, and not all control lines and information lines are necessarily indicated on the product. In practice, it can be considered that almost all configurations are interconnected.

Abstract

A text information editing device is connected to a display device, holds dialog text comprising one or more utterance text and editing processing information that indicates at least one of an important expression and an important expression extraction algorithm, refers to the editing processing information and extracts an important expression from the one or more utterance text pieces, displays the dialog text and the extracted important expression on the display device, accepts editing input to at least one of the utterance text included in the dialog text and an annotation to which the utterance text is added, and executes editing that corresponds to the editing input.

Description

テキスト情報編集装置及びテキスト情報編集方法Text information editing device and text information editing method
 本出願は、2020年1月27日に出願された日本特許出願第2020-010977号の優先権を主張し、その内容を参照することにより、本出願に取り込む。 This application claims the priority of Japanese Patent Application No. 2020-010977 filed on January 27, 2020, and incorporates it into this application by referring to its contents.
 本発明は、テキスト情報編集装置及びテキスト情報編集方法に関する。 The present invention relates to a text information editing device and a text information editing method.
 会議や議論など音声を用いてなされる対話の書き起こしやテキストチャットなどの対話テキストに存在する有用な情報を、後に参照し活用することは種々の業務において重要である。しかし、一般的な文章および書類などが予め読者を想定して十分な推敲の上で記されることが多いのに対して、それらの対話テキストは話者間で対話が成立することが重要視される。この結果、対話テキストにおいては、有用な情報が明瞭な文体及び表現で記述されているとは限らず、複数人の対話や過去の発話等を加味して一つの有用な情報が構成される場合が多い。 It is important in various tasks to refer to and utilize useful information that exists in dialogue texts such as transcriptions of dialogues made by voice such as meetings and discussions and text chats. However, while general sentences and documents are often written with sufficient elaboration assuming the reader in advance, it is important that dialogue is established between the speakers in those dialogue texts. Will be done. As a result, in the dialogue text, useful information is not always described in a clear style and expression, and one useful information is constructed in consideration of dialogues of a plurality of people and past utterances. There are many.
 本技術分野の背景技術として、国際公開第2014/208298号(特許文献1)がある。この公報には「文章分類装置として、分類対象である会話内容をテキスト化した入力文章から、文章分類に寄与する区間を、手がかり文を基礎に抽出処理する分類対象区間抽出部と、分類対象区間抽出手段で抽出された前記文章分類に寄与する区間のテキストを用いて該入力文章が何れの分類に属するかを判定処理する文章分類部とを設ける。」と記載されている(要約参照)。 As a background technology in this technical field, there is International Publication No. 2014/20298 (Patent Document 1). In this publication, "As a sentence classification device, a classification target section extraction unit that extracts and processes sections that contribute to sentence classification from input sentences that are textualized conversation contents to be classified based on clue sentences, and a classification target section. A sentence classification unit for determining which classification the input sentence belongs to using the text of the section contributing to the sentence classification extracted by the extraction means is provided. ”(See summary).
国際公開第2014/208298号International Publication No. 2014/20298
 しかしながら、特許文献1に記載の発明では、文章を分類するための予め分類方法や分類の観点が定義されている必要がある。また、特許文献1に記載の発明は、コールセンタにおける応答などの半定型的な対話テキストのみに対して適用可能であり、一般的な対話テキストから有用な情報を抽出することは困難である。 However, in the invention described in Patent Document 1, it is necessary that the classification method and the viewpoint of classification for classifying sentences are defined in advance. Further, the invention described in Patent Document 1 is applicable only to a semi-standard dialogue text such as a response in a call center, and it is difficult to extract useful information from a general dialogue text.
 例えば、「今日は天気が良かったので、歩いてこちらまで来ました」という発話は、「話者の交通手段」についての観点と、「発話がなされた場所の天気」についての観点と、からなる2つの観点に関する情報を含む。しかしながら、交通手段と天気の一方又は双方が有用か否かは、この発話テキストを含む対話テキストを利用する人物の観点によって決定される。従って、一般的な文脈において分類による有用情報抽出は、すべての情報要求に対応できるわけではない。 For example, the utterance "I came here on foot because the weather was fine today" comes from the perspective of "the means of transportation of the speaker" and the perspective of "the weather at the place where the utterance was made". Contains information on two perspectives. However, whether one or both of the means of transportation and the weather are useful depends on the perspective of the person using the dialogue text, including this utterance text. Therefore, in the general context, useful information extraction by classification cannot meet all information requests.
 そこで、本発明の一態様は、入力された対話テキストから、任意の観点に立脚した有用な情報を容易に取得することを目的とする。 Therefore, one aspect of the present invention aims to easily obtain useful information based on an arbitrary viewpoint from the input dialogue text.
 上記課題を解決するために、本発明の一態様は以下の構成を採用する。テキスト情報編集装置は、プロセッサと記憶装置とを有し、表示装置に接続され、前記記憶装置は、1以上の発話テキストからなる対話テキストと、重要表現と前記重要表現の抽出アルゴリズムとの少なくとも一方を示す編集処理情報と、を保持し、前記プロセッサは、前記編集処理情報を参照して前記1以上の発話テキストから重要表現を抽出し、前記対話テキストと、前記抽出した重要表現と、を前記表示装置に表示し、前記対話テキストに含まれる発話テキスト及び当該発話テキストが付与される注釈の少なくとも一方に対する編集入力を受け付け、前記編集入力に対応する編集を実行する。 In order to solve the above problems, one aspect of the present invention adopts the following configuration. The text information editing device has a processor and a storage device, and is connected to a display device. The storage device includes a dialogue text consisting of one or more spoken texts, and at least one of an important expression and an extraction algorithm of the important expression. The processor retains the editing processing information indicating the above, extracts important expressions from the one or more spoken texts with reference to the editing processing information, and obtains the dialogue text and the extracted important expressions. It is displayed on the display device, accepts edit input for at least one of the utterance text included in the dialogue text and the comment to which the utterance text is attached, and executes the edit corresponding to the edit input.
 本発明の一態様によれば、入力された対話テキストから、任意の観点に立脚した有用な情報を容易に取得することができる。 According to one aspect of the present invention, useful information based on an arbitrary viewpoint can be easily obtained from the input dialogue text.
 上記した以外の課題、構成及び効果は、以下の実施形態の説明により明らかにされる。 Issues, configurations and effects other than those described above will be clarified by the explanation of the following embodiments.
実施例1におけるテキスト情報編集装置の構成例を示すブロック図である。It is a block diagram which shows the structural example of the text information editing apparatus in Example 1. FIG. 実施例1における対話入力受付部が出力する対話テキストデータの一例である。This is an example of dialogue text data output by the dialogue input reception unit in the first embodiment. 実施例1における重要表現検出部がインタラクティブ編集部へ出力するデータの一例である。This is an example of data output by the important expression detection unit in the first embodiment to the interactive editing unit. 実施例1におけるテキスト情報編集処理の一例を示すフローチャートである。It is a flowchart which shows an example of the text information editing process in Example 1. FIG. 実施例1におけるインタラクティブ編集の表示画面の一例である。This is an example of a display screen for interactive editing in the first embodiment. 実施例2におけるインタラクティブ編集の表示画面の一例である。This is an example of a display screen for interactive editing in the second embodiment. 実施例3におけるインタラクティブ編集の表示画面の一例である。This is an example of the display screen for interactive editing in the third embodiment. 実施例4におけるテキスト情報編集装置の構成例を示すブロック図である。It is a block diagram which shows the structural example of the text information editing apparatus in Example 4. FIG. 実施例4における話題境界識別部が出力する対話テキストデータの一例である。This is an example of dialogue text data output by the topic boundary identification unit in the fourth embodiment. 実施例4におけるテキスト情報編集処理の一例を示すフローチャートである。It is a flowchart which shows an example of the text information editing process in Example 4. 実施例4におけるインタラクティブ編集の表示画面の一例である。This is an example of a display screen for interactive editing in the fourth embodiment. 実施例5におけるインタラクティブ編集の表示画面の一例である。It is an example of the display screen of the interactive editing in Example 5. 実施例6におけるインタラクティブ編集の表示画面の一例である。This is an example of the display screen for interactive editing in the sixth embodiment. 実施例7におけるインタラクティブ編集の表示画面の一例である。It is an example of the display screen of the interactive editing in Example 7.
 本実施形態は、テキスト情報編集装置を説明する。本実施形態では、テキスト情報編集装置が日本語のテキストを扱う例を説明するが、テキストに用いられる言語に制限はない。以下、本発明の実施形態を図面に基づいて詳細に説明する。本実施形態において、同一の構成には原則として同一の符号を付け、繰り返しの説明は省略する。なお、本実施形態は本発明を実現するための一例に過ぎず、本発明の技術的範囲を限定するものではないことに注意すべきである。 This embodiment describes a text information editing device. In this embodiment, an example in which the text information editing device handles Japanese text will be described, but the language used for the text is not limited. Hereinafter, embodiments of the present invention will be described in detail with reference to the drawings. In the present embodiment, in principle, the same components are designated by the same reference numerals, and the repeated description will be omitted. It should be noted that the present embodiment is merely an example for realizing the present invention and does not limit the technical scope of the present invention.
 図1は、テキスト情報編集装置の構成例を示すブロック図である。テキスト情報編集装置100は、例えば、例えば、CPU(Central Processing Unit)101、メモリ102、補助記憶装置103、及び通信装置104を有する計算機によって構成される。 FIG. 1 is a block diagram showing a configuration example of a text information editing device. The text information editing device 100 is composed of, for example, a computer having a CPU (Central Processing Unit) 101, a memory 102, an auxiliary storage device 103, and a communication device 104.
 CPU101は、プロセッサを含み、メモリ102に格納されたプログラムを実行する。メモリ102は、不揮発性の記憶素子であるROM(Read Only Memory)及び揮発性の記憶素子であるRAM(Random Access Memory)を含む。ROMは、不変のプログラム(例えば、BIOS(Basic Input/Output System))などを格納する。RAMは、DRAM(Dynamic Random Access Memory)のような高速かつ揮発性の記憶素子であり、CPU101が実行するプログラム及びプログラムの実行時に使用されるデータを一時的に格納する。 The CPU 101 includes a processor and executes a program stored in the memory 102. The memory 102 includes a ROM (Read Only Memory) which is a non-volatile storage element and a RAM (Random Access Memory) which is a volatile storage element. The ROM stores an invariant program (for example, BIOS (Basic Input / Output System)) and the like. The RAM is a high-speed and volatile storage element such as a DRAM (Dynamic Random Access Memory), and temporarily stores a program executed by the CPU 101 and data used when the program is executed.
 補助記憶装置103は、例えば、磁気記憶装置(HDD(Hard Disk Drive))、フラッシュメモリ(SSD(Solid State Drive))等の大容量かつ不揮発性の記憶装置であり、CPU101が実行するプログラム及びプログラムの実行時に使用されるデータを格納する。すなわち、プログラムは、補助記憶装置103から読み出されて、メモリ102にロードされて、CPU101によって実行される。 The auxiliary storage device 103 is, for example, a large-capacity and non-volatile storage device such as a magnetic storage device (HDD (Hard Disk Drive)) or a flash memory (SSD (Solid State Drive)), and is a program and a program executed by the CPU 101. Stores the data used when executing. That is, the program is read from the auxiliary storage device 103, loaded into the memory 102, and executed by the CPU 101.
 テキスト情報編集装置100は、入力インターフェース105及び出力インターフェース108を有してもよい。入力インターフェース105は、キーボード106やマウス107などの入力装置が接続され、オペレータからの入力を受けるインターフェースである。出力インターフェース108は、プリンタやディスプレイなどの表示装置109が接続され、プログラムの実行結果をオペレータが視認可能な形式で出力するインターフェースである。また、入力インターフェース105には、例えば、音声による対話を入力するためのマイクが、入力装置として接続されていてもよい。 The text information editing device 100 may have an input interface 105 and an output interface 108. The input interface 105 is an interface to which an input device such as a keyboard 106 or a mouse 107 is connected and receives input from an operator. The output interface 108 is an interface to which a display device 109 such as a printer or a display is connected and outputs a program execution result in a format that can be visually recognized by an operator. Further, for example, a microphone for inputting a voice dialogue may be connected to the input interface 105 as an input device.
 通信装置104は、所定のプロトコルに従って、他の装置との通信を制御するネットワークインターフェース装置である。また、通信装置104は、例えば、USB等のシリアルインターフェースを含んでもよい。 The communication device 104 is a network interface device that controls communication with other devices according to a predetermined protocol. Further, the communication device 104 may include a serial interface such as USB.
 CPU101が実行するプログラムは、リムーバブルメディア(CD-ROM、フラッシュメモリなど)又はネットワークを介してテキスト情報編集装置100に提供され、非一時的記憶媒体である不揮発性の補助記憶装置103に格納される。このため、テキスト情報編集装置100は、リムーバブルメディアからデータを読み込むインターフェースを有するとよい。 The program executed by the CPU 101 is provided to the text information editing device 100 via a removable medium (CD-ROM, flash memory, etc.) or a network, and is stored in the non-volatile auxiliary storage device 103, which is a non-temporary storage medium. .. Therefore, the text information editing device 100 may have an interface for reading data from removable media.
 テキスト情報編集装置100は、物理的に一つの計算機上で、又は、論理的又は物理的に構成された複数の計算機上で構成される計算機システムであり、同一の計算機上で別個のスレッドで動作してもよく、複数の物理的計算機資源上に構築された仮想計算機上で動作してもよい。端末200についても同様である。 The text information editing device 100 is a computer system composed of physically one computer or a plurality of computers logically or physically configured, and operates in separate threads on the same computer. It may operate on a virtual computer built on a plurality of physical computer resources. The same applies to the terminal 200.
 なお、本実施形態ではテキスト情報編集装置100に接続された入力装置から、テキスト情報編集装置100に対する入力が行われるが、例えば、対話の各参加者が保有するタブレット端末等のような別の計算機から入力が行われてもよい。同様に、本実施形態ではテキスト情報編集装置100に接続された表示装置109に、テキスト情報編集装置100による演算結果が出力されるが、例えば、対話の各参加者が保有するタブレット端末等のような別の計算機に当該演算結果が出力されてもよい。 In the present embodiment, the input device connected to the text information editing device 100 inputs to the text information editing device 100, but another computer such as a tablet terminal owned by each participant in the dialogue, for example. Input may be made from. Similarly, in the present embodiment, the calculation result by the text information editing device 100 is output to the display device 109 connected to the text information editing device 100, but for example, a tablet terminal owned by each participant in the dialogue. The calculation result may be output to another computer.
 CPU101は、例えば、対話入力受付部111、重要表現検出部112、編集入力受付部113、インタラクティブ編集部114、出力情報生成部115、及び出力部116を含む。 The CPU 101 includes, for example, an interactive input receiving unit 111, an important expression detecting unit 112, an editing input receiving unit 113, an interactive editing unit 114, an output information generating unit 115, and an output unit 116.
 対話入力受付部111は、入力インターフェース105に接続された入力装置を介して対話テキストの入力を受け付け、対話テキストを発話単位に分割する。対話及び会議における発話が書き起こしされたテキスト、テキストチャット、Web掲示板及及びWebフォーラム内のテキスト、並びにe-mail及びSMS内のテキストは、いずれも対話テキストの一例である。なお、本実施形態における対話テキストは、複数の人間による発話テキストからなるが、1以上の人間による1以上の発話テキストからなるテキストを対話テキストとしてもよい。 The dialogue input reception unit 111 receives the input of the dialogue text via the input device connected to the input interface 105, and divides the dialogue text into utterance units. Texts transcribed from dialogues and conferences, text chats, texts in web bulletin boards and forums, and texts in e-mail and SMS are all examples of dialogue texts. The dialogue text in the present embodiment is composed of utterance texts by a plurality of humans, but a text composed of one or more utterance texts by one or more humans may be used as the dialogue text.
 重要表現検出部112は、発話テキストから重要な発話、語句、及び複数の発話などの表現である重要表現を検出する。編集入力受付部113は、発話テキスト内に付される注釈の入力を受け付ける。インタラクティブ編集部114は、編集入力受付部113に入力された注釈を発話テキストに付与する。出力情報生成部115は、注釈を付与された発話テキストを表示装置109に出力するための情報を生成する。 The important expression detection unit 112 detects important expressions such as important utterances, phrases, and a plurality of utterances from the utterance text. The edit input reception unit 113 accepts the input of the annotation attached in the utterance text. The interactive editing unit 114 adds the annotation input to the editing input receiving unit 113 to the utterance text. The output information generation unit 115 generates information for outputting the annotated utterance text to the display device 109.
 出力部116は、対話入力受付部111が入力を受け付けた発話、重要表現検出部112が抽出した重要表現、及びインタラクティブ編集部114による編集結果報等を、表示装置109に出力する。出力部116は、例えばGUI(Grafical User Interface)や、テキストファイル、画像ファイル、音声ファイル等のユーザが利用可能な形態でユーザ出力結果をユーザに提供する。 The output unit 116 outputs the utterance received by the dialogue input reception unit 111, the important expression extracted by the important expression detection unit 112, the editing result report by the interactive editing unit 114, and the like to the display device 109. The output unit 116 provides the user with a user output result in a form that can be used by the user, such as a GUI (Graphical User Interface), a text file, an image file, or an audio file.
 例えば、CPU101は、メモリ102にロードされた対話入力プログラムに従って動作することで、対話入力受付部111として機能し、メモリ102にロードされた重要表現検出プログラムに従って動作することで、重要表現検出部112として機能する。CPU101に含まれる他の機能部についても、プログラムと機能部の関係は同様である。 For example, the CPU 101 functions as the dialogue input reception unit 111 by operating according to the dialogue input program loaded in the memory 102, and operates according to the important expression detection program loaded in the memory 102, so that the important expression detection unit 112 Functions as. The relationship between the program and the functional unit is the same for the other functional units included in the CPU 101.
 なお、CPU101に含まれる機能部による機能の一部又は全部が、例えば、ASIC(Application Specific Integrated Circuit)やFPGA(Field-Programmable Gate Array)等のハードウェアによって実現されてもよい。 Note that some or all of the functions of the functional unit included in the CPU 101 may be realized by hardware such as ASIC (Application Specific Integrated Circuit) or FPGA (Field-Programmable Gate Array).
 補助記憶装置103は、例えば、発話注釈DB(DataBase)121及び編集処理情報DB122を保持する。なお、補助記憶装置103に格納されている一部又は全部の情報は、メモリ102に格納されているプログラム内に記述されていてもよい。 The auxiliary storage device 103 holds, for example, the utterance annotation DB (DataBase) 121 and the editing processing information DB 122. Note that some or all of the information stored in the auxiliary storage device 103 may be described in the program stored in the memory 102.
 発話注釈DB121は、対話入力受付部111に入力された対話テキスト、及びインタラクティブ編集部114によって付与された注釈を保持する。編集処理情報DB122は、発話テキストから重要表現を検出するためのアルゴリズムや、重要表現となり得る手がかかり語と付与される注釈の内容との対応等の情報を保持する。 The utterance annotation DB 121 holds the dialogue text input to the dialogue input receiving unit 111 and the annotation given by the interactive editing unit 114. The editing processing information DB 122 holds information such as an algorithm for detecting an important expression from the uttered text and a correspondence between a handy word that can be an important expression and the content of an annotation to be given.
 図2は、対話入力受付部111が出力する対話テキストデータの一例である。対話入力受付部111は、入力インターフェース105を介して対話テキストの入力を受け付け入力された対話テキストを発話テキストに分割し、発話を一意に識別可能な発話IDを付与し、発話テキストと発話IDとの対応を出力部116へ送信する。出力部116は、発話IDと、発話テキストと、が対応付けられた対話テキストデータを表示装置109に表示する。 FIG. 2 is an example of dialogue text data output by the dialogue input reception unit 111. The dialogue input reception unit 111 receives the input of the dialogue text via the input interface 105, divides the input dialogue text into utterance texts, assigns an utterance ID that can uniquely identify the utterance, and sets the utterance text and the utterance ID. Correspondence is transmitted to the output unit 116. The output unit 116 displays the dialogue text data in which the utterance ID and the utterance text are associated with each other on the display device 109.
 図3は、重要表現検出部112がインタラクティブ編集部114へ出力するデータの一例である。重要表現検出部112は、各発話テキストから検出した重要表現を注釈として付与する。なお、1つの発話に含まれる重要表現は1つ以下である必要はなく複数であってもよいし、複数の発話に跨って重要表現が抽出されてもよい。なお、発話注釈DB121は、図3に示す、発話IDと発話テキストと注釈との対応を保持する。さらに、発話注釈DB121には、各発話テキストがどの対話テキストに含まれるかを示す対話IDを含んでもよい。 FIG. 3 is an example of data output by the important expression detection unit 112 to the interactive editing unit 114. The important expression detection unit 112 adds an important expression detected from each utterance text as an annotation. The number of important expressions included in one utterance does not have to be one or less, and may be plural, or important expressions may be extracted across a plurality of utterances. The utterance annotation DB 121 retains the correspondence between the utterance ID, the utterance text, and the annotation shown in FIG. Further, the utterance annotation DB 121 may include a dialogue ID indicating which dialogue text includes each utterance text.
 図4は、テキスト情報編集処理の一例を示すフローチャートである。テキスト情報編集処理は、例えば、対話入力受付部111が対話テキストの入力を受け付ける度に実行される。例えば、対話入力受付部111は、音声による対話テキストが入力された場合には、例えば、ユーザが対話の開始タイミング及び終了タイミングを入力してもよいし、前回の発話から所定時間が経過したら対話の終了タイミングと判定し、対話の終了タイミング以降の最初の発話が開始したら対話の開始タイミングと判定する。また、例えば、対話入力受付部111は、テキストファイルによる対話テキストが入力された場合には、例えば、当該テキストファイル内のテキストが1つの対話であると判定する。 FIG. 4 is a flowchart showing an example of the text information editing process. The text information editing process is executed, for example, every time the dialogue input receiving unit 111 accepts the input of the dialogue text. For example, when the dialogue text is input by voice, the dialogue input reception unit 111 may input, for example, the start timing and the end timing of the dialogue, or the dialogue is performed after a predetermined time has elapsed from the previous utterance. It is determined that it is the end timing of the dialogue, and when the first utterance after the end timing of the dialogue starts, it is determined to be the start timing of the dialogue. Further, for example, when the dialogue input by the text file is input, the dialogue input receiving unit 111 determines that the text in the text file is one dialogue, for example.
 まず、対話入力受付部111は入力された対話テキストを発話単位に分割することにより発話テキストを生成し、生成した発話テキストそれぞれに対して発話IDを付与し、発話テキストと発話IDとを発話注釈DB121に格納する(S401)。 First, the dialogue input reception unit 111 generates an utterance text by dividing the input dialogue text into utterance units, assigns a utterance ID to each of the generated utterance texts, and utters the utterance text and the utterance ID. It is stored in the DB 121 (S401).
 具体的には、例えば、対話入力受付部111は、音声による対話テキストが入力された場合であれば、発話の時間間隔が所定時間以上空いた点において、対話テキストを分割する。また、例えば、対話入力受付部111は、対話入力受付部111は、テキストファイルよる対話テキストが入力された場合であれば、テキスト内における所定の文字コード(例えば、句読点、感嘆符、スペース、又はエンター等)の前後において対話テキストを分割する。なお、対話入力受付部111には、予め発話ごとに分割された対話テキストが入力されてもよく、この場合、ステップS401の処理が行われなくてもよい。 Specifically, for example, the dialogue input reception unit 111 divides the dialogue text at a point where the utterance time interval is more than a predetermined time when the dialogue text by voice is input. Further, for example, the dialogue input reception unit 111 may receive a predetermined character code (for example, a punctuation mark, an exclamation mark, a space, or a space) in the text when the dialogue text is input by the text file. Divide the dialogue text before and after (enter, etc.). In addition, the dialogue text divided for each utterance may be input to the dialogue input reception unit 111 in advance, and in this case, the process of step S401 may not be performed.
 なお、対話入力受付部111は、発話IDと発話テキストと出力部116に送信し、出力部116は受信した発話IDと発話テキストと、例えば図2のように出力してもよい。 The dialogue input reception unit 111 may transmit the utterance ID, the utterance text, and the output unit 116, and the output unit 116 may output the received utterance ID and the utterance text, for example, as shown in FIG.
 重要表現検出部112は、各発話テキストを参照して重要表現を検出する(S402)。具体的には、例えば、重要表現検出部112は、編集処理情報DB122に格納されている手掛かり語を、重要表現として発話テキストから抽出する。例えば、図3の発話IDが「1」である発話テキストにおいて「例」や「例えば」という表現が予め編集処理情報DB122において定義された(又はプログラム内に記述された)重要表現であり、当該表現が注釈として付与される場合、その注釈は「例」である。また、例えば、重要表現検出部112は、対話テキスト及び/又は発話テキストに対して、機械学習による識別器、TF-IDF、統計モデル等を適用して、重要表現を検出してもよい。 The important expression detection unit 112 detects the important expression by referring to each utterance text (S402). Specifically, for example, the important expression detection unit 112 extracts the clue word stored in the editing processing information DB 122 from the utterance text as an important expression. For example, in the utterance text in which the utterance ID of FIG. 3 is "1", the expressions "example" and "for example" are important expressions defined in advance in the editing processing information DB 122 (or described in the program). When a representation is given as an annotation, the annotation is an "example". Further, for example, the important expression detection unit 112 may detect the important expression by applying a machine learning discriminator, TF-IDF, a statistical model, or the like to the dialogue text and / or the utterance text.
 なお、ステップS402において、重要表現検出部112は、検出した重要表現を、発話テキストの注釈に付与して発話注釈DB121に格納してもよく、この場合、ユーザによって当該重要表現を注釈から削除する指示を受けた場合に、当該重要表現を注釈から削除する。 In step S402, the important expression detection unit 112 may add the detected important expression to the annotation of the utterance text and store it in the utterance annotation DB 121. In this case, the user deletes the important expression from the annotation. Remove the important expression from the annotation when instructed to do so.
 重要表現検出部112は、発話IDと、発話テキストと、検出した重要表現と、を出力部116に送信し、出力部116は受信した発話IDと、発話テキストと、重要表現と、をユーザが編集可能な態様で表示装置109に表示する(S403)。ステップS403の処理が終了すると、ユーザが注釈を編集したり、発話をソートしたり、発話を検索したりする等の編集入力をすることができるようになる。編集入力が可能な表示画面には、例えば、ユーザが編集対象のテキストを容易に認識できるように、GUI上のテキストボックスに発話IDを付与した発話等が表示される。 The important expression detection unit 112 transmits the utterance ID, the utterance text, and the detected important expression to the output unit 116, and the output unit 116 sends the received utterance ID, the utterance text, and the important expression to the user. It is displayed on the display device 109 in an editable manner (S403). When the process of step S403 is completed, the user can edit the annotation, sort the utterances, search the utterances, and so on. On the display screen where editing input is possible, for example, an utterance in which an utterance ID is assigned to a text box on the GUI is displayed so that the user can easily recognize the text to be edited.
 インタラクティブ編集部114は、編集入力受付部113がユーザからの編集入力を受け付けたか否かを判定する(S404)。なお、ステップS404において、編集入力受付部113はテキスト及び音声に加えて、画像及び映像、センサデータ、並びに専用アプリケーションにおけるクリック及びタップなどのGUIを介して、ユーザの入力を受け付けてもよい。ユーザによる具体的な編集内容については後述する。 The interactive editing unit 114 determines whether or not the editing input receiving unit 113 has accepted the editing input from the user (S404). In step S404, the edit input receiving unit 113 may accept the user's input via the image and video, the sensor data, and the GUI such as click and tap in the dedicated application, in addition to the text and voice. The specific editing contents by the user will be described later.
 インタラクティブ編集部114は、編集入力受付部113がユーザからの編集入力を受け付けたと判定した場合(S404:YES)、ユーザの編集入力に従う編集を行い、編集後の発話及び注釈を発話注釈DB121に格納する(S405)。具体的には、例えば、インタラクティブ編集部114は、図3のデータに対して、行の追加、削除、及び入れ替え、並びに注釈の付与及び削除などを行うことができる。なお、例えば、編集入力と関係ない発話テキスト等については、ステップS403で出力された情報がそのまま出力され続けていてもよい。 When the interactive editing unit 114 determines that the editing input receiving unit 113 has received the editing input from the user (S404: YES), the interactive editing unit 114 performs editing according to the user's editing input, and stores the edited utterance and annotation in the utterance annotation DB 121. (S405). Specifically, for example, the interactive editorial unit 114 can add, delete, and replace lines, and add and delete annotations to the data shown in FIG. For example, for the utterance text or the like that is not related to the edit input, the information output in step S403 may continue to be output as it is.
 続いて、出力情報生成部115は編集後の発話及び注釈を表示するための情報を生成し、出力部116は表示装置109に当該情報を表示する(S406)。 Subsequently, the output information generation unit 115 generates information for displaying the edited utterance and annotation, and the output unit 116 displays the information on the display device 109 (S406).
 インタラクティブ編集部114は、編集入力受付部113がユーザからの編集入力を受け付けていないと判定した場合(S404:YES)、出力情報生成部115は未編集の発話及び注釈を表示するための情報を生成し、出力部116は表示装置109に当該情報を表示する(S406)。 When the interactive editing unit 114 determines that the editing input receiving unit 113 does not accept the editing input from the user (S404: YES), the output information generating unit 115 provides information for displaying unedited utterances and comments. The output unit 116 generates and displays the information on the display device 109 (S406).
 また、編集入力受付部113は、1回のテキスト編集処理において、複数回及び複数種類の編集入力を受け付けてもよく、その場合、インタラクティブ編集部114は、逐次編集を行い、出力部116はその編集結果を逐次出力してもよい。これにより、ユーザは編集結果を逐次確認した上で編集を続けることができる。 Further, the edit input receiving unit 113 may receive a plurality of times and a plurality of types of edit inputs in one text editing process. In that case, the interactive editing unit 114 performs sequential editing, and the output unit 116 thereof. The editing result may be output sequentially. As a result, the user can continue editing after confirming the editing results one by one.
 図5は、インタラクティブ編集の表示画面の一例である。表示画面500は、発話のテキストを表示するテキストボックス501及びテキストボックス502、重要表現検出部112が検出した重要表現を示すテキストボックス503、並びにユーザによる編集入力の結果を示すテキストボックス504を示す。 FIG. 5 is an example of a display screen for interactive editing. The display screen 500 shows a text box 501 and a text box 502 for displaying the spoken text, a text box 503 for showing the important expression detected by the important expression detection unit 112, and a text box 504 for showing the result of editing input by the user.
 図5の例では、重要表現検出部112がテキストボックス502内のテキストに「入力を見て対話的に処理する」という表現が重要表現であると判定したため、ステップS403において当該表現を重要表現の注釈として当該テキストに付与すべきかをテキストボックス503によって提示した。そして、ステップS404において、ユーザはテキストボックス504において当該表現が重要表現であることを示す編集入力をした。 In the example of FIG. 5, since the important expression detection unit 112 determines that the expression "seeing the input and processing it interactively" in the text in the text box 502 is an important expression, the expression is referred to as an important expression in step S403. A text box 503 indicates whether the text should be annotated. Then, in step S404, the user makes an edit input indicating that the expression is an important expression in the text box 504.
 なお、ユーザは、編集入力において、重要表現検出部112によって重要表現が検出されなかった発話テキストに対して重要表現としての注釈を入力することもできる。この場合、ユーザが入力した重要表現としての注釈がテキストボックス504に表示される。 Note that the user can also input an annotation as an important expression for the utterance text in which the important expression is not detected by the important expression detection unit 112 in the edit input. In this case, the annotation as an important expression input by the user is displayed in the text box 504.
 また、表示画面は感情アイコン505を含んでもよい。例えば、重要表現検出部112が、所定のアルゴリズムに従って対話テキストの文章から対話の感情分析を行ってもよく(また、対話テキストが音声で与えられている場合には話者の音声の周波数等から感情分析をしてもよい)、この場合、感情分析の結果を感情アイコン505によってユーザに提示する。 Further, the display screen may include the emotion icon 505. For example, the important expression detection unit 112 may perform the sentiment analysis of the dialogue from the text of the dialogue text according to a predetermined algorithm (and when the dialogue text is given by voice, from the frequency of the speaker's voice or the like. (Sentiment analysis may be performed), in this case, the result of the sentiment analysis is presented to the user by the sentiment icon 505.
 感情アイコン505によって、ユーザは提示されている対話テキストに含まれる表現が重要表現であるか否かを判断しやすくなる。なお、感情アイコン505が示す感情は、対話テキスト全体に対して決定されてもよいし、1以上の発話ごとに対して決定されてもよい。 The emotion icon 505 makes it easier for the user to determine whether or not the expression contained in the presented dialogue text is an important expression. The emotion indicated by the emotion icon 505 may be determined for the entire dialogue text, or may be determined for each one or more utterances.
 また、例えば、重要表現検出部112は、ある発話に対する感情分析の結果が所定の感情(例えば議論を促進する「歓喜」)であった場合は、直前の発話に重要な情報が含まれているとみなし、当該直前の発話を重要表現として注釈することができる。 Further, for example, when the result of the sentiment analysis for a certain utterance is a predetermined emotion (for example, "joy" that promotes discussion), the important expression detection unit 112 includes important information in the immediately preceding utterance. It can be regarded as an important expression and the utterance immediately before the said can be annotated.
 また、編集入力においてユーザが感情を指定して入力してもよく、この場合、感情アイコン505は入力された感情を示し、重要表現検出部112は、入力された感情に基づいて、重要表現を検出してもよい。 Further, the user may specify an emotion in the edit input and input the emotion. In this case, the emotion icon 505 indicates the input emotion, and the important expression detection unit 112 inputs the important expression based on the input emotion. It may be detected.
 なお、例えば、テキストボックス503に表示された重要表現が会議における宿題事項などを表している場合は、ユーザはテキストボックス504に表示されているような発話(編集入力)を行うことで、重要表現を宿題事項として注釈することができるし、さらに会議が終了した後に宿題事項を出力情報生成部115及び出力部116を介して抽出して参照することができる。 For example, when the important expression displayed in the text box 503 represents homework matters in the meeting, the user can make an utterance (edit input) as displayed in the text box 504 to express the important expression. Can be annotated as homework items, and homework items can be extracted and referred to via the output information generation unit 115 and the output unit 116 after the meeting is completed.
 また、例えば、ユーザは編集入力において、不要な発話があることを入力したり(例えば不要な発話の発話IDを入力する編集入力を行うことで当該不要な発話を発話注釈DB121から削除できる)、不明瞭な発話に説明を追加したり(例えば不明瞭な発話の発話IDと説明を入力する編集入力を行うことで、発話注釈DBにおいて当該不明瞭な発話に対応する注釈に追加説明を記録する)することなどもできる。 Further, for example, the user can input that there is an unnecessary utterance in the edit input (for example, the unnecessary utterance can be deleted from the utterance commentary DB 121 by performing an edit input for inputting the utterance ID of the unnecessary utterance). By adding an explanation to the unclear utterance (for example, by performing an edit input to input the utterance ID and the explanation of the unclear utterance, the additional explanation is recorded in the utterance corresponding to the unclear utterance in the utterance comment DB. ) Can also be done.
 重要表現検出部112が検出した重要表現が宿題事項に該当するか否か、及び対話に不要な発話が含まれているか否か、はユーザの観点に依拠している場合があるが、以上のようにテキスト情報編集装置100が構成されていることにより、テキスト情報編集装置100はユーザの観点に即して対話テキストを校正、要約、及び情報抽出を行うことができる。 Whether or not the important expression detected by the important expression detection unit 112 corresponds to the homework matter and whether or not the dialogue contains unnecessary utterances may depend on the user's point of view. By configuring the text information editing device 100 as described above, the text information editing device 100 can proofread, summarize, and extract information from the dialogue text from the user's point of view.
 実施例2以降の実施例については、当該実施例よりも前の実施例との相違点を説明し、当該前の実施例と同様の構成については原則的に説明を省略する。図6は、ステップS406で出力されるインタラクティブ編集の表示画面の一例である。表示画面500は、発話のテキストを表示するテキストボックス501及びテキストボックス502、並びに発話テキストの順序を並び替えるためにボタン601を含む。 For the examples after the second embodiment, the differences from the examples prior to the embodiment will be explained, and the description of the same configuration as that of the previous embodiment will be omitted in principle. FIG. 6 is an example of the interactive editing display screen output in step S406. The display screen 500 includes a text box 501 and a text box 502 for displaying the spoken text, and a button 601 for rearranging the order of the spoken text.
 ユーザは、編集入力においてボタン601を選択することで発話の順序を変更することができる。具体的には、例えば、ボタン601が選択されると、出力部116は、各発話テキストの順番を入力するための表示画面や、ユーザによって選択された発話テキストをユーザによって指定される位置(例えば、最上位又は最下位)へ移動させるための表示画面を表示する。本実施例によれば、例えば、重要な発話を上位に移動させたり、重要ではない発話を下位に移動させたりすることができ、ユーザの観点に即して対話テキストを校正、要約、及び情報抽出を行うことができる。 The user can change the order of utterances by selecting button 601 in the edit input. Specifically, for example, when the button 601 is selected, the output unit 116 has a display screen for inputting the order of each utterance text, or a position (for example,) in which the utterance text selected by the user is specified by the user. , Top or bottom) to display the display screen. According to this embodiment, for example, important utterances can be moved higher and less important utterances can be moved lower, and the dialogue text can be proofread, summarized, and information from the user's point of view. Extraction can be performed.
 図7は、ステップS406で出力されるインタラクティブ編集の表示画面の一例である。表示画面500は、発話のテキストを表示するテキストボックス501及びテキストボックス502、並びに重要表現を表示するためのカード701を含む。 FIG. 7 is an example of the interactive editing display screen output in step S406. The display screen 500 includes a text box 501 and a text box 502 for displaying the spoken text, and a card 701 for displaying important expressions.
 カード701には、重要表現検出部112が検出した重要表現が表示される。ユーザは、編集入力においてカード701を選択することで、重要な発話を宿題事項など任意の観点で注釈として保存することができ、ユーザの観点に即して対話テキストを校正、要約、及び情報抽出を行うことができる。なお、カード701には、ユーザが編集入力において指定した重要表現が表示されてもよい。 The important expression detected by the important expression detection unit 112 is displayed on the card 701. By selecting card 701 in the edit input, the user can save important utterances as annotations from any viewpoint such as homework, and proofread, summarize, and extract information from the dialogue text according to the user's viewpoint. It can be performed. The card 701 may display an important expression specified by the user in the edit input.
 図8は、本実施例のテキスト情報編集装置100の構成例を示すブロック図である。本実施例のテキスト情報編集装置100は、CPU101が話題境界識別部117を含む点において、実施例1のテキスト情報編集装置100と異なる。 FIG. 8 is a block diagram showing a configuration example of the text information editing device 100 of this embodiment. The text information editing device 100 of the present embodiment is different from the text information editing device 100 of the first embodiment in that the CPU 101 includes the topic boundary identification unit 117.
 話題境界識別部117は、対話テキストに含まれる発話の間を話題境界で分割することにより、同じ話題に属する1以上の発話テキスト(以下、発話グループとも呼ぶ)を特定することができる。例えば、複数人の対話において、相槌や同意を示す手掛かり語(例えば、「ええ」や「そうですね」等の語)を含む発話は、直前の発話と同様な話題について行われるものである。従って、相槌や同意を示す発話とその直前の発話は同じ話題のまとまりに属していると考えられる。 The topic boundary identification unit 117 can identify one or more utterance texts (hereinafter, also referred to as utterance groups) belonging to the same topic by dividing the utterances included in the dialogue text by the topic boundary. For example, in a dialogue between a plurality of people, an utterance containing a clue word indicating an aizuchi or consent (for example, a word such as "yes" or "yes") is made on the same topic as the immediately preceding utterance. Therefore, it is considered that the utterance showing the aizuchi or consent and the utterance immediately before it belong to the same group of topics.
 また、因果関係を示す手掛かり語や接続詞を含む発話についても、同様に、直前の発話と同様な話題について行われるものである。従って、相槌や同意を示す発話とその直前の発話は同じ話題のまとまりに属していると考えられる。 In addition, utterances that include clues and conjunctions that indicate a causal relationship are also conducted on the same topics as the immediately preceding utterance. Therefore, it is considered that the utterance showing the aizuchi or consent and the utterance immediately before it belong to the same group of topics.
 また、例えば、話題境界識別部117は、形態素解析・構文解析を用いて、質問を述べている発話と、当該質問に対応する回答を述べている発話と、を特定し、これらの発話が同じ発話グループに属していると判定する。また、例えば、話題境界識別部117は、構文解析を用いて、発話同士の類似度を算出し、当該類似度が所定値以上である発話は内容がオーバーラップしているために、同じ発話グループに属すると判定し、類似度が当該所定値未満である発話は異なる発話グループに属すると判定してもよい(即ち話題境界が生成される)。 Further, for example, the topic boundary identification unit 117 uses morphological analysis / syntactic analysis to identify an utterance that states a question and an utterance that states an answer corresponding to the question, and these utterances are the same. Determined to belong to the utterance group. Further, for example, the topic boundary identification unit 117 calculates the similarity between utterances by using parsing, and the utterances whose similarity is equal to or higher than a predetermined value have overlapping contents, so that the same utterance group is used. It may be determined that the utterances that have a similarity less than the predetermined value belong to a different utterance group (that is, a topic boundary is generated).
 また、例えば、話題境界識別部117は、発話の情報量を算出し、情報量の少ない発話が前又は後の発話と同じ発話グループに属すると判定してもよい。発話の情報量については、例えば、発話者ごとに予め定められた重みを適用してもよい。また、例えば、話題境界識別部117は、対話テキスト及び/又は発話テキストに対して、機械学習などによる識別器を適用して、発話グループ(話題境界)を特定してもよい。 Further, for example, the topic boundary identification unit 117 may calculate the amount of information of the utterance and determine that the utterance with a small amount of information belongs to the same utterance group as the previous or subsequent utterance. For the amount of utterance information, for example, a predetermined weight may be applied to each speaker. Further, for example, the topic boundary identification unit 117 may specify the utterance group (topic boundary) by applying a classifier by machine learning or the like to the dialogue text and / or the utterance text.
 また、例えば、話題境界識別部117は、上記した発話グループ(話題境界)の一部または全部それぞれに所定の重みを適用して、発話グループ(話題境界)を特定してもよい。 Further, for example, the topic boundary identification unit 117 may specify the utterance group (topic boundary) by applying a predetermined weight to a part or all of the above-mentioned utterance group (topic boundary).
 また、話題境界識別部117は、例えば、形態素解析・構文解析を用いて、発話グループを特徴づける語句を抽出することによって、各発話グループの話題を特定することができる。 Further, the topic boundary identification unit 117 can identify the topic of each utterance group by extracting the words and phrases that characterize the utterance group by using, for example, morphological analysis and syntactic analysis.
 上記したような発話の境界を特定する手掛かり語やアルゴリズム、複数の発話が同じ発話グループに属していることを特定する手掛かり語やアルゴリズム等の、話題境界を生成するために必要な情報、及び話題を特定するためのアルゴリズム等の話題を特定するために必要な情報は編集処理情報DB122に予め格納されていてもよいし、プログラムとして記述されていてもよい。 Information necessary to generate topic boundaries, such as clues and algorithms that identify the boundaries of utterances as described above, clues and algorithms that identify that multiple utterances belong to the same utterance group, and topics. Information necessary for specifying a topic such as an algorithm for specifying the above may be stored in advance in the editing process information DB 122, or may be described as a program.
 なお、上記では発話グループ間の境界が話題境界としたが、この話題境界が示す発話グループは排他的、即ち発話グループ間で発話の重なりなくてもよいし、1つの発話が複数の発話グループに属してもよい。 In the above, the boundary between utterance groups is the topic boundary, but the utterance group indicated by this topic boundary is exclusive, that is, the utterances do not have to overlap between the utterance groups, and one utterance becomes a plurality of utterance groups. May belong.
 図9は、話題境界識別部117が出力する対話テキストデータの一例である。話題境界識別部117は対話入力受付部111が発話IDと発話テキストとを受信し、発話テキストから話題境界を特定し、出力情報生成部115を介して出力部116に送信する。図9の対話テキストデータは、図2のデータに加えて、話題が開始した発話と話題が終了した発話とを特定可能な話題境界を含む。 FIG. 9 is an example of dialogue text data output by the topic boundary identification unit 117. In the topic boundary identification unit 117, the dialogue input reception unit 111 receives the utterance ID and the utterance text, identifies the topic boundary from the utterance text, and transmits the topic boundary to the output unit 116 via the output information generation unit 115. In addition to the data of FIG. 2, the dialogue text data of FIG. 9 includes a topic boundary that can identify the utterance at which the topic starts and the utterance at which the topic ends.
 図10は、テキスト情報編集処理の一例を示すフローチャートである。ステップS401に続いて、話題境界識別部117は、上記した方法で、対話テキスト内の話題境界を識別する(S1001)。このとき、話題境界識別部117は、出力情報生成部115に話題境界を送信し、出力情報生成部115は話題境界を示す出力情報を生成して、出力部116が当該出力情報を表示装置109に表示してもよい。ステップS1001の処理が終了するとステップS402に遷移する。 FIG. 10 is a flowchart showing an example of the text information editing process. Following step S401, the topic boundary identification unit 117 identifies the topic boundary in the dialogue text by the method described above (S1001). At this time, the topic boundary identification unit 117 transmits the topic boundary to the output information generation unit 115, the output information generation unit 115 generates output information indicating the topic boundary, and the output unit 116 displays the output information 109. It may be displayed in. When the process of step S1001 is completed, the process proceeds to step S402.
 図11は、ステップS406で出力される表示画面の一例である。表示画面500は、発話グループボックス1101及び発話グループボックス1104を含み、発話グループごとに対話テキストが分割されて表示されている。発話グループボックス1101は、当該発話グループに含まれる発話のテキストを表示するテキストボックス501及びテキストボックス502、当該発話グループの話題を示す話題ボックス1102、並びに当該話題に対して当該話題を特徴づけるタグを提示するためのボタン1103を含む。なお、話題ボックス1102には、話題ではなく当該発話グループの最初の発話や、重要表現検出部112が検出した当該発話グループにおける重要表現等が表示されてもよい。 FIG. 11 is an example of the display screen output in step S406. The display screen 500 includes the utterance group box 1101 and the utterance group box 1104, and the dialogue text is divided and displayed for each utterance group. The utterance group box 1101 includes a text box 501 and a text box 502 that display the text of the utterance included in the utterance group, a topic box 1102 that indicates the topic of the utterance group, and a tag that characterizes the topic with respect to the topic. Includes button 1103 for presentation. In the topic box 1102, not the topic but the first utterance of the utterance group, the important expression in the utterance group detected by the important expression detection unit 112, and the like may be displayed.
 ボタン1103の横の領域には、話題境界識別部117が特定した話題を特徴づけるタグが提示され、ユーザは編集入力において当該タグが適切であるか否かを示す情報を入力することができる。なお、タグは1つ発話グループに対して、1つだけ付与されてもよいし、複数付与されてもよいし、1つも付与されなくてもよい。 In the area next to the button 1103, a tag that characterizes the topic specified by the topic boundary identification unit 117 is presented, and the user can input information indicating whether or not the tag is appropriate in the edit input. It should be noted that only one tag may be attached to one utterance group, a plurality of tags may be attached, or no tag may be attached.
 また、ボタン1103を用いてポップアップ表示などを行い、ユーザが、例えば、編集入力において、ポップアップ表示従って話題を示すタグを自由に入力できるようにしてもよく、これにより話題を特徴づけるタグをユーザが自由に注釈として保存することができる。タグは、例えば話題のトピックや宿題事項などの特記事項などを自由な語や文章で記述することができる。 Further, a pop-up display or the like may be performed using the button 1103 so that the user can freely input a tag indicating the topic according to the pop-up display in the edit input, for example, and the user can input a tag that characterizes the topic. You can freely save it as an annotation. Tags can describe, for example, topical topics and special notes such as homework in free words and sentences.
 このように、タグが付与されると、テキスト情報編集装置100は、例えば、発話グループ内の発話の特徴量を用いて、類似の話題が発言されている発話グループを調査することができる。この特徴量として、発話グループ内の出現語の頻度や、TF-IDF等の任意の文章に対して適用できる特徴量を用いることができる。 In this way, when the tag is attached, the text information editing device 100 can investigate the utterance group in which a similar topic is spoken, for example, by using the feature amount of the utterance in the utterance group. As this feature amount, the frequency of words appearing in the utterance group and the feature amount applicable to any sentence such as TF-IDF can be used.
 類似する発話グループが抽出された場合、ユーザは類似する発話グループに対して、効率的に対話テキストを編集することもできるし、タグとして注釈されたトピックや宿題事項に基づいて、必要な情報を容易に抽出及び要約することができる。さらに、ユーザは、類似する発話グループに付与されたタグに対して、編集入力によってタグを削除することもできる。 When similar utterance groups are extracted, the user can efficiently edit the dialogue text for similar utterance groups, and provide necessary information based on the topics and homework annotated as tags. It can be easily extracted and summarized. Further, the user can delete the tag attached to the similar utterance group by editing input.
 このようにテキスト情報編集装置100が構成されていることにより、ユーザの観点に即して対話テキストを校正、要約、及び情報抽出を行うことができる。 With the text information editing device 100 configured in this way, it is possible to proofread, summarize, and extract information from the dialogue text from the user's point of view.
 図12は、ステップS406で出力されるインタラクティブ編集の表示画面の一例である。表示画面500は、発話グループボックス1101を含み、発話グループごとに対話テキストが分割されて表示されている。また、表示画面500は、検索結果表示ボックス1204を含む。発話グループボックス1101は、当該話題に含まれる発話のテキストを表示するテキストボックス501及びテキストボックス502、当該発話グループの話題を示す話題ボックス1102、検索ボックス1202、並びに検索ボタン1203を含む。 FIG. 12 is an example of the interactive editing display screen output in step S406. The display screen 500 includes the utterance group box 1101 and displays the dialogue text divided for each utterance group. Further, the display screen 500 includes a search result display box 1204. The utterance group box 1101 includes a text box 501 and a text box 502 that display the text of the utterance included in the topic, a topic box 1102 that indicates the topic of the utterance group, a search box 1202, and a search button 1203.
 ユーザは、編集入力おいて、検索ボックス1202にキーワード等のクエリを入力して検索ボタン1203を選択することにより、当該発話グループに関連する発話グループや発話を抽出及びランキング化することができる。 The user can extract and rank utterance groups and utterances related to the utterance group by inputting a query such as a keyword in the search box 1202 and selecting the search button 1203 in the edit input.
 検索結果表示ボックス1204は、スニペット1205及びスニペット1206を含む。スニペット1205及ぶスニペット1206は、検索結果として抽出された重要表現を見出しとして表示することができる。これにより、ユーザは、当該発話グループに関連する発話、及び話題境界範囲を効率的に探すことができる。 The search result display box 1204 includes snippet 1205 and snippet 1206. Snippets 1205 and snippets 1206 can display important expressions extracted as search results as headings. As a result, the user can efficiently search for utterances and topic boundary ranges related to the utterance group.
 また、ユーザが編集入力において、スニペット1205及びスニペット1206を選択することにより、当該発話グループに関連する発話や話題境界範囲を、当該発話グループに結び付けることができ、インタラクティブ編集部114はその結果を例えば注釈などの形式で発話注釈DB121に蓄積することができる。
 このようにテキスト情報編集装置100が構成されていることにより、ユーザの観点に即して対話テキストを校正、要約、及び情報抽出を行うことができる。
Further, when the user selects snippet 1205 and snippet 1206 in the edit input, the utterance or topic boundary range related to the utterance group can be linked to the utterance group, and the interactive editorial unit 114 can connect the result to the utterance group, for example. It can be stored in the utterance annotation DB 121 in the form of an annotation or the like.
With the text information editing device 100 configured in this way, it is possible to proofread, summarize, and extract information from the dialogue text from the user's point of view.
 図13は、ステップS406で出力されるインタラクティブ編集の表示画面の一例である。表示画面500は、発話グループボックス1101及び発話グループボックス1104を含み、発話グループごとに対話テキストが分割されて表示されている。発話グループボックス1101は、当該話題に含まれる発話のテキストを表示するテキストボックス501及びテキストボックス502、類似発話表示ボックス1301を含む。 FIG. 13 is an example of the interactive editing display screen output in step S406. The display screen 500 includes the utterance group box 1101 and the utterance group box 1104, and the dialogue text is divided and displayed for each utterance group. The utterance group box 1101 includes a text box 501 and a text box 502 for displaying the utterance text included in the topic, and a similar utterance display box 1301.
 話題境界識別部117は、発話グループボックス1101に含まれる一連の発話以外の発話から、例えば、当該発話グループに属することが望ましい発話を類似発話として抽出し、類似発話表示ボックス1301に表示する。具体的には、例えば、話題境界識別部117は、発話グループボックス1101に含まれる一連の発話との類似度が所定値以上であれば当該一連の発話以外の発話を類似発話として抽出する。 The topic boundary identification unit 117 extracts, for example, utterances that are desirable to belong to the utterance group from the utterances other than the series of utterances included in the utterance group box 1101 as similar utterances, and displays them in the similar utterance display box 1301. Specifically, for example, the topic boundary identification unit 117 extracts utterances other than the series of utterances as similar utterances if the degree of similarity with the series of utterances included in the utterance group box 1101 is equal to or higher than a predetermined value.
 ユーザは、編集入力において、類似発話表示ボックスを選択することにより、インタラクティブ編集部114は、類似発話を、当該発話グループに追加し、当該類似発話の話題を当該発話グループに対応する話題に変更又は追加して、発話注釈DB121に保存する。 When the user selects the similar utterance display box in the edit input, the interactive editorial unit 114 adds the similar utterance to the utterance group and changes the topic of the similar utterance to the topic corresponding to the utterance group. Add and save in the utterance annotation DB 121.
 このようにテキスト情報編集装置100が構成されていることにより、ユーザの観点に即して対話テキストを校正、要約、及び情報抽出を行うことができる。 With the text information editing device 100 configured in this way, it is possible to proofread, summarize, and extract information from the dialogue text from the user's point of view.
 図14は、ステップS406で出力されるインタラクティブ編集の表示画面の一例である。表示画面500は、話題ボックス1102及び話題ボックス1105、並びにテキストボックス501及びテキストボックス502を含む。 FIG. 14 is an example of the interactive editing display screen output in step S406. The display screen 500 includes a topic box 1102 and a topic box 1105, and a text box 501 and a text box 502.
 話題ボックス1102の直下に表示されているテキストボックス501の発話は、話題ボックス1102の話題を示す発話グループに属する発話であり、話題ボックス1105の直下に表示されているテキストボックス502の発話は、話題ボックス1105話題を示す発話グループに属する発話である。 The utterance of the text box 501 displayed directly under the topic box 1102 is an utterance belonging to the utterance group indicating the topic of the topic box 1102, and the utterance of the text box 502 displayed directly under the topic box 1105 is a topic. Box 1105 An utterance belonging to an utterance group indicating a topic.
 話題ボックス1102の話題とテキストボックス501の発話は関連するものであると話題境界識別部117によって判定されるため、実線のエッジで結ばれている。同様に話題ボックス1107の話題とテキストボックス502の発話は関連するものであると話題境界識別部117によって判定されるため、実線のエッジで結ばれている。 Since the topic boundary identification unit 117 determines that the topic in the topic box 1102 and the utterance in the text box 501 are related, they are connected by a solid line edge. Similarly, since the topic boundary identification unit 117 determines that the topic of the topic box 1107 and the utterance of the text box 502 are related, they are connected by a solid line edge.
 また、表示画面500は、エッジ追加ボタン1401とエッジ削除ボタン1402を含む。ユーザは、編集入力において、話題ボックス又はテキストボックスを2つ選択して、エッジ追加ボタン1401を選択することにより、点線のエッジを追加することができ、インタラクティブ編集部114は当該2つの話題又は発話がエッジで結ばれたことを発話注釈DB121に登録してもよいし、エッジで結ばれた発話を同じ発話グループに含めてもよい。 Further, the display screen 500 includes an edge addition button 1401 and an edge deletion button 1402. The user can add a dotted edge by selecting two topic boxes or text boxes and selecting the edge addition button 1401 in the edit input, and the interactive editorial unit 114 can add the two topics or utterances. The utterances connected by the edge may be registered in the utterance commentary DB 121, or the utterances connected by the edge may be included in the same utterance group.
 ユーザは、編集入力において、エッジを選択して、エッジ削除ボタン1402を選択することにより、エッジを削除することができ、インタラクティブ編集部114は当該エッジに関する情報を発話注釈DB121から削除してもよいし、エッジが削除された発話を当該発話グループから除外してもよい。 The user can delete the edge by selecting the edge and selecting the edge delete button 1402 in the edit input, and the interactive editorial unit 114 may delete the information about the edge from the utterance annotation DB 121. However, the utterance with the edge removed may be excluded from the utterance group.
 これにより、テキスト情報編集装置100は、ユーザの観点で関連すると考えられる発話や発話グループをエッジで結びつけることができ、これを新たな注釈として蓄積したり、発話グループを変更したりできる。 As a result, the text information editing device 100 can connect utterances and utterance groups that are considered to be related from the user's point of view at an edge, and can accumulate this as a new annotation or change the utterance group.
 また、ユーザが、編集入力において、関連する発話及び発話グループを指定した場合に、インタラクティブ編集部114は、ユーザの指定に合わせて、例えば、関連する発話を連続して表示するように編集したり、実施例4や実施例5で示したように関連する発話や発話グループを抽出したりできる。
 このようにテキスト情報編集装置100が構成されていることにより、ユーザの観点に即して対話テキストを校正、要約、及び情報抽出を行うことができる。
Further, when the user specifies related utterances and utterance groups in the edit input, the interactive editorial unit 114 edits so as to continuously display, for example, related utterances according to the user's designation. , Related utterances and utterance groups can be extracted as shown in Example 4 and Example 5.
With the text information editing device 100 configured in this way, it is possible to proofread, summarize, and extract information from the dialogue text from the user's point of view.
 なお、本発明は上記した実施例に限定されるものではなく、様々な変形例が含まれる。例えば、上記した実施例は本発明を分かりやすく説明するために詳細に説明したものであり、必ずしも説明した全ての構成を備えるものに限定されるものではない。また、ある実施例の構成の一部を他の実施例の構成に置き換えることも可能であり、また、ある実施例の構成に他の実施例の構成を加えることも可能である。また、各実施例の構成の一部について、他の構成の追加・削除・置換をすることが可能である。 The present invention is not limited to the above-described embodiment, and includes various modifications. For example, the above-described embodiment has been described in detail in order to explain the present invention in an easy-to-understand manner, and is not necessarily limited to the one including all the described configurations. It is also possible to replace a part of the configuration of one embodiment with the configuration of another embodiment, and it is also possible to add the configuration of another embodiment to the configuration of one embodiment. Further, it is possible to add / delete / replace a part of the configuration of each embodiment with another configuration.
 また、上記の各構成、機能、処理部、処理手段等は、それらの一部又は全部を、例えば集積回路で設計する等によりハードウェアで実現してもよい。また、上記の各構成、機能等は、プロセッサがそれぞれの機能を実現するプログラムを解釈し、実行することによりソフトウェアで実現してもよい。各機能を実現するプログラム、テーブル、ファイル等の情報は、メモリや、ハードディスク、SSD(Solid State Drive)等の記録装置、または、ICカード、SDカード、DVD等の記録媒体に置くことができる。 Further, each of the above configurations, functions, processing units, processing means, etc. may be realized by hardware by designing a part or all of them by, for example, an integrated circuit. Further, each of the above configurations, functions, and the like may be realized by software by the processor interpreting and executing a program that realizes each function. Information such as programs, tables, and files that realize each function can be placed in a memory, a hard disk, a recording device such as an SSD (Solid State Drive), or a recording medium such as an IC card, an SD card, or a DVD.
 また、制御線や情報線は説明上必要と考えられるものを示しており、製品上必ずしも全ての制御線や情報線を示しているとは限らない。実際には殆ど全ての構成が相互に接続されていると考えてもよい。 In addition, the control lines and information lines indicate those that are considered necessary for explanation, and not all control lines and information lines are necessarily indicated on the product. In practice, it can be considered that almost all configurations are interconnected.

Claims (13)

  1.  テキスト情報編集装置であって、
     プロセッサと記憶装置とを有し、
     表示装置に接続され、
     前記記憶装置は、
     1以上の発話テキストからなる対話テキストと、
     重要表現と前記重要表現の抽出アルゴリズムとの少なくとも一方を示す編集処理情報と、を保持し、
     前記プロセッサは、
     前記編集処理情報を参照して前記1以上の発話テキストから重要表現を抽出し、
     前記対話テキストと、前記抽出した重要表現と、を前記表示装置に表示し、
     前記対話テキストに含まれる発話テキスト及び当該発話テキストが付与される注釈の少なくとも一方に対する編集入力を受け付け、
     前記編集入力に対応する編集を実行する、テキスト情報編集装置。
    It is a text information editing device
    It has a processor and a storage device,
    Connected to the display device,
    The storage device is
    Dialogue text consisting of one or more utterance texts and
    The editing processing information indicating at least one of the important expression and the extraction algorithm of the important expression is retained.
    The processor
    An important expression is extracted from the one or more utterance texts with reference to the editing processing information.
    The dialogue text and the extracted important expressions are displayed on the display device.
    Accepts edit input for at least one of the utterance text included in the dialogue text and the annotation to which the utterance text is attached.
    A text information editing device that executes editing corresponding to the editing input.
  2.  請求項1に記載のテキスト情報編集装置であって、
     前記プロセッサは、前記編集入力において、前記表示装置に表示した重要表現を、当該重要表現を含む発話テキストに前記注釈として対応づける指示を受け付けた場合、当該発話テキストと、当該重要表現と、を前記注釈として対応付けて、前記記憶装置に格納する、テキスト情報編集装置。
    The text information editing device according to claim 1.
    When the processor receives an instruction to associate the important expression displayed on the display device with the utterance text including the important expression as the annotation in the editing input, the utterance text and the important expression are described. A text information editing device that is associated as an annotation and stored in the storage device.
  3.  請求項1に記載のテキスト情報編集装置であって、
     前記プロセッサは、
     前記編集入力として、前記対話テキストに含まれるテキストの指定を受け付け、
     前記指定を受け付けたテキスト、当該テキストが含まれる発話テキストと、を前記注釈として対応付けて、前記記憶装置に格納する、テキスト情報編集装置。
    The text information editing device according to claim 1.
    The processor
    As the edit input, the specification of the text included in the dialogue text is accepted.
    A text information editing device that associates a text that has received the designation and an utterance text that includes the text as the annotation and stores the text in the storage device.
  4.  請求項1に記載のテキスト情報編集装置であって、
     前記プロセッサは、前記編集入力として、前記対話テキストに含まれる前記1以上の発話テキストの並び替え指示を受け付ける、テキスト情報編集装置。
    The text information editing device according to claim 1.
    The processor is a text information editing device that receives a rearrangement instruction of one or more utterance texts included in the dialogue text as the editing input.
  5.  請求項1に記載のテキスト情報編集装置であって、
     前記編集処理情報は、前記対話テキストを、同一の話題に関する発話テキストである発話グループに分類するアルゴリズムを示し、
     前記プロセッサは、
     前記編集処理情報を参照して、前記対話テキストを1以上の発話グループに分類し、
     前記重要表現を、前記重要表現を含む発話グループに対応づけて、前記記憶装置に格納し、前記表示装置に表示し、
     前記1以上の発話グループの分類結果を示す情報を前記注釈として、テキスト情報編集装置。
    The text information editing device according to claim 1.
    The editing processing information indicates an algorithm for classifying the dialogue text into utterance groups which are utterance texts related to the same topic.
    The processor
    With reference to the editing process information, the dialogue text is classified into one or more utterance groups.
    The important expression is stored in the storage device in association with the utterance group including the important expression, and displayed on the display device.
    A text information editing device using the information indicating the classification result of the one or more utterance groups as the annotation.
  6.  請求項5に記載のテキスト情報編集装置であって、
     前記編集処理情報は、発話グループの話題を特定するアルゴリズムを示し、
     前記プロセッサは、
     前記編集処理情報を参照して、前記1以上の発話グループの話題を特定し、
     前記1以上の発話グループを、前記特定した話題と、対応付けて、前記表示装置に表示し、
     前記特定した話題を前記注釈として前記記憶装置に格納する、テキスト情報編集装置。
    The text information editing device according to claim 5.
    The editing processing information indicates an algorithm for identifying the topic of the utterance group.
    The processor
    With reference to the editing process information, the topic of the one or more utterance groups is identified, and the topic is identified.
    The one or more utterance groups are displayed on the display device in association with the specified topic.
    A text information editing device that stores the specified topic as the annotation in the storage device.
  7.  請求項5に記載のテキスト情報編集装置であって、
     前記編集処理情報は、前記対話テキストを前記1以上の発話グループに分類するための手掛かり語を保持し、
     前記プロセッサは、前記手掛かり語を含む発話テキストと、当該発話テキストの前又は後ろの発話テキストは別の発話グループに属すると判定する、テキスト情報編集装置。
    The text information editing device according to claim 5.
    The editing process information holds a clue word for classifying the dialogue text into the one or more utterance groups.
    The processor is a text information editing device that determines that the utterance text including the clue word and the utterance text before or after the utterance text belong to another utterance group.
  8.  請求項7に記載のテキスト情報編集装置であって、
     前記手掛かり語は、相槌を示す表現、接続詞、及び因果関係を示す表現の少なくとも1つを含む、テキスト情報編集装置。
    The text information editing device according to claim 7.
    The clue word is a text information editing device including at least one of an expression indicating an aizuchi, a conjunction, and an expression indicating a causal relationship.
  9.  請求項5に記載のテキスト情報編集装置であって、
     前記プロセッサは、
     前記編集入力として、第1発話グループに対するタグの入力を受け付け、
     前記タグを、前記第1発話グループと、対応付けて、前記記憶装置に格納する、テキスト情報編集装置。
    The text information editing device according to claim 5.
    The processor
    As the edit input, the input of the tag for the first utterance group is accepted.
    A text information editing device that stores the tag in the storage device in association with the first utterance group.
  10.  請求項5に記載のテキスト情報編集装置であって、
     前記プロセッサは、
     前記編集入力として、第2発話グループに対する検索クエリの入力を受け付け、
     前記検索クエリによる前記第2発話グループに対する検索結果を、前記表示装置に表示する、テキスト情報編集装置。
    The text information editing device according to claim 5.
    The processor
    As the edit input, the input of the search query for the second utterance group is accepted.
    A text information editing device that displays a search result for the second utterance group by the search query on the display device.
  11.  請求項5に記載のテキスト情報編集装置であって、
     前記プロセッサは、
     第3発話グループに含まれる発話テキストとの類似度が所定値以上である前記第3発話グループに含まれない発話テキストを特定し、
     前記特定した発話テキストを、前記第3発話グループと対応付けて前記表示装置に表示し、
     前記編集入力として、前記特定した発話テキストを、前記第3発話グループに含める指示を受け付けた場合、前記特定した発話テキストが前記第3発話グループに含まれることを示す前記注釈を、前記記憶装置に格納する、テキスト情報編集装置。
    The text information editing device according to claim 5.
    The processor
    Identify the utterance texts that are not included in the third utterance group and whose similarity with the utterance texts included in the third utterance group is equal to or greater than a predetermined value.
    The identified utterance text is displayed on the display device in association with the third utterance group.
    When an instruction to include the specified utterance text in the third utterance group is received as the edit input, the annotation indicating that the specified utterance text is included in the third utterance group is sent to the storage device. A text information editing device that stores.
  12.  請求項5に記載のテキスト情報編集装置であって、
     前記プロセッサは、前記編集入力として、異なる発話グループに含まれる発話テキストをエッジで結ぶ指示を受け付けた場合、前記エッジで結ばれた発話テキストを関連付けて、前記記憶装置に格納し、前記表示装置に表示する、テキスト情報編集装置。
    The text information editing device according to claim 5.
    When the processor receives an instruction to connect utterance texts included in different utterance groups at an edge as the edit input, the processor associates the utterance texts connected at the edge, stores the utterance text in the storage device, and stores the utterance text in the display device. A text information editing device to display.
  13.  テキスト情報編集装置がテキスト情報を編集する方法であって、
     前記テキスト情報編集装置は、表示装置に接続され、
     前記テキスト情報編集装置は、
     1以上の発話テキストからなる対話テキストと、
     重要表現と前記重要表現の抽出アルゴリズムとの少なくとも一方を示す編集処理情報と、を保持し、
     前記方法は、
     前記テキスト情報編集装置が、前記編集処理情報を参照して前記1以上の発話テキストから重要表現を抽出し、
     前記テキスト情報編集装置が、前記対話テキストと、前記抽出した重要表現と、を前記表示装置に表示し、
     前記テキスト情報編集装置が、前記対話テキストに含まれる発話テキスト及び当該発話テキストが付与される注釈の少なくとも一方に対する編集入力を受け付け、
     前記テキスト情報編集装置が、前記編集入力に対応する編集を実行する、テキスト情報編集方法。
    The text information editing device is a method of editing text information.
    The text information editing device is connected to the display device and is connected to the display device.
    The text information editing device is
    Dialogue text consisting of one or more utterance texts and
    The editing processing information indicating at least one of the important expression and the extraction algorithm of the important expression is retained.
    The method is
    The text information editing device extracts important expressions from the one or more utterance texts with reference to the editing processing information.
    The text information editing device displays the dialogue text and the extracted important expression on the display device.
    The text information editing device accepts edit input for at least one of the utterance text included in the dialogue text and the annotation to which the utterance text is attached.
    A text information editing method in which the text information editing device executes editing corresponding to the editing input.
PCT/JP2021/001975 2020-01-27 2021-01-21 Text information editing device and text information editing method WO2021153403A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2020-010977 2020-01-27
JP2020010977A JP2021117759A (en) 2020-01-27 2020-01-27 Text information editing device and text information editing method

Publications (1)

Publication Number Publication Date
WO2021153403A1 true WO2021153403A1 (en) 2021-08-05

Family

ID=77079872

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/001975 WO2021153403A1 (en) 2020-01-27 2021-01-21 Text information editing device and text information editing method

Country Status (2)

Country Link
JP (1) JP2021117759A (en)
WO (1) WO2021153403A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2023079353A (en) * 2021-11-29 2023-06-08 株式会社日立ソリューションズ Dialogue management device, dialogue management system, and dialogue management method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004213260A (en) * 2002-12-27 2004-07-29 Csk Corp Speech sentence acquiring system and program, attribute giving device and program, document generating device and program, and information processing method
JP2014527651A (en) * 2011-07-07 2014-10-16 インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation System and method for determining interpersonal influence information using text content from interpersonal dialogue
JP2018185431A (en) * 2017-04-26 2018-11-22 シャープ株式会社 Dialog system, dialog device, response controller, control method of dialog device, control method of response controller, and control program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004213260A (en) * 2002-12-27 2004-07-29 Csk Corp Speech sentence acquiring system and program, attribute giving device and program, document generating device and program, and information processing method
JP2014527651A (en) * 2011-07-07 2014-10-16 インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation System and method for determining interpersonal influence information using text content from interpersonal dialogue
JP2018185431A (en) * 2017-04-26 2018-11-22 シャープ株式会社 Dialog system, dialog device, response controller, control method of dialog device, control method of response controller, and control program

Also Published As

Publication number Publication date
JP2021117759A (en) 2021-08-10

Similar Documents

Publication Publication Date Title
JP4218758B2 (en) Subtitle generating apparatus, subtitle generating method, and program
JP5257330B2 (en) Statement recording device, statement recording method, program, and recording medium
US10783314B2 (en) Emphasizing key points in a speech file and structuring an associated transcription
US7996227B2 (en) System and method for inserting a description of images into audio recordings
US20080077869A1 (en) Conference supporting apparatus, method, and computer program product
EP3115907A1 (en) Common data repository for improving transactional efficiencies of user interactions with a computing device
JP2009140466A (en) Method and system for providing conversation dictionary services based on user created dialog data
WO2021153403A1 (en) Text information editing device and text information editing method
JP2010198247A (en) Support device, support program and support method
JP3444831B2 (en) Editing processing device and storage medium storing editing processing program
JP5423993B2 (en) Text processing apparatus, text processing method, and program
US20230069113A1 (en) Text Summarization Method and Text Summarization System
KR102390009B1 (en) Ai-based syntax analysis research note system
JP6903365B1 (en) Server and data allocation method
JP7416665B2 (en) Dialogue system and control method for dialogue system
JP3471253B2 (en) Document classification method, document classification device, and recording medium recording document classification program
JP5916666B2 (en) Apparatus, method, and program for analyzing document including visual expression by text
JP3721397B2 (en) Device for converting spoken language into written language
WO2022215433A1 (en) Information representation structure analysis device, and information representation structure analysis method
US20230359837A1 (en) Multilingual summarization of episodes using longformers
US20220114202A1 (en) Summary generation apparatus, control method, and system
US20240111941A1 (en) Framework agnostic summarization of multi-channel communication
CN110969026A (en) Translation output method and device, electronic equipment and storage medium
WO2010106660A1 (en) Keyword presentation device and keyword presentation program
Ribeiro et al. Improving Speech-to-Text Summarization by Using Additional Information Sources

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21747476

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21747476

Country of ref document: EP

Kind code of ref document: A1