WO2016067348A1 - Presentation support method, presentation support program, and presentation support device - Google Patents

Presentation support method, presentation support program, and presentation support device Download PDF

Info

Publication number
WO2016067348A1
WO2016067348A1 PCT/JP2014/078533 JP2014078533W WO2016067348A1 WO 2016067348 A1 WO2016067348 A1 WO 2016067348A1 JP 2014078533 W JP2014078533 W JP 2014078533W WO 2016067348 A1 WO2016067348 A1 WO 2016067348A1
Authority
WO
WIPO (PCT)
Prior art keywords
word
area
display
presentation support
region
Prior art date
Application number
PCT/JP2014/078533
Other languages
French (fr)
Japanese (ja)
Inventor
田中 正清
村瀬 健太郎
Original Assignee
富士通株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 富士通株式会社 filed Critical 富士通株式会社
Priority to PCT/JP2014/078533 priority Critical patent/WO2016067348A1/en
Priority to JP2016556070A priority patent/JP6304396B2/en
Publication of WO2016067348A1 publication Critical patent/WO2016067348A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/10Speech classification or search using distance or distortion measures between unknown speech and reference templates
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Definitions

  • the present invention relates to a presentation support method, a presentation support program, and a presentation support apparatus.
  • a display device As an example of a technology that supports presentation, there is a technique that presents a presenter or a listener with a part that the presenter is explaining. For example, a display device has been proposed that aims to suppress skipping of a document. This display device recognizes a phrase uttered by a speaker, identifies a read-out portion of a document being displayed on a display panel based on the recognized phrase, and displays a display state of the identified portion as a first state. The display state is changed to a second display state different from the first display state, for example, highlight display such as blinking.
  • speech recognition is used to obtain a phrase uttered by a speaker.
  • a portion where the speaker is not explaining due to the misrecognition is highlighted, and as a result, the speaker's explanation portion may not be highlighted.
  • the display device cannot present the explanation part to the speaker or the listener, and may disturb the presentation.
  • an object of the present invention is to provide a presentation support method, a presentation support program, and a presentation support apparatus that can suppress a situation where an explanation part of a presenter is not highlighted.
  • a computer executes a process of extracting a first word from a character string included in each area of a document file including a page displayed on a screen basis when the page is divided. To do. Further, the computer executes speech recognition, and for each region in the page being displayed on the predetermined display unit, the first word extracted from the region and the second word obtained as a result of the speech recognition The process of calculating the degree of association from the above is executed. Further, the computer sets a higher speed for the highlight display of the region as the region having a higher degree of relevance calculated for each region, or advances the highlight display of the region as the region having a lower relevance level. Execute processing to set the speed lower. Further, the computer executes a process of controlling the highlight display in the page according to the speed set for each area.
  • FIG. 1 is a diagram illustrating the configuration of the presentation support system according to the first embodiment.
  • FIG. 2 is a block diagram illustrating a functional configuration of the presentation support apparatus according to the first embodiment.
  • FIG. 3 is a diagram illustrating an example of extracted word data.
  • FIG. 4 is a diagram illustrating an example of a temporal change related to the progress of highlight display.
  • FIG. 5 is a diagram illustrating a transition example of the slide screen.
  • FIG. 6 is a diagram illustrating a transition example of the slide screen.
  • FIG. 7 is a flowchart illustrating the procedure of the weighting process according to the first embodiment.
  • FIG. 8 is a flowchart illustrating the procedure of the speech recognition process according to the first embodiment.
  • FIG. 9 is a flowchart illustrating the procedure of the display control process according to the first embodiment.
  • FIG. 10 is a diagram illustrating a hardware configuration example of a computer that executes the presentation support program according to the first embodiment and the second embodiment.
  • FIG. 1 is a diagram illustrating the configuration of the presentation support system according to the first embodiment.
  • the presentation support system 1 shown in FIG. 1 highlights an area including a word obtained as a result of recognition of speech input from the microphone 3 in a presentation screen on which a document file is displayed on the display device 5. I will provide a.
  • the presentation support system 1 realizes display control that increases the speed of highlighting as the area having a higher degree of relevance to a word and lowers the speed of highlighting as the area having a lower degree of relevance. To do. This suppresses the situation where the presenter's explanation part is not highlighted.
  • the presentation support system 1 accommodates a microphone 3, a display device 5, an input device 7, and a presentation support device 10.
  • the peripheral devices such as the microphone 3, the display device 5 and the input device 7 and the presentation support device 10 are connected by wire or wirelessly.
  • the microphone 3 is a device that converts sound into an electrical signal, and is sometimes called a microphone.
  • the microphone 3 can be attached to a presenter who performs a presentation.
  • a headset-type or tie-pin type microphone can be attached to a predetermined position of the presenter's body or clothes, or a hand-type microphone can be carried by the presenter.
  • the microphone 3 can also be installed at a predetermined position in a range where the utterance of the presenter can be collected.
  • the microphone 3 may be an attachment type or a stationary type microphone.
  • a microphone having any type of directivity can be adopted as the microphone 3, but sounds other than the presenter's utterance, for example, the utterance of the listener and the noise such as noise are collected.
  • the sensitivity of the microphone can be limited to the speaking direction of the presenter.
  • the microphone 3 can employ any conversion method such as a dynamic type, an electret capacitor type, or a capacitor type.
  • the analog signal obtained by collecting sound in the microphone 3 is converted into a digital signal and then input to the presentation support apparatus 10.
  • the display device 5 is a device that displays various types of information.
  • the display device 5 may be a liquid crystal display or an organic EL (electroluminescence) display that realizes display by light emission, or a projector that realizes display by projection.
  • the number of installed display devices 5 is not necessarily limited to one, and a plurality of display devices 5 may be provided.
  • a liquid crystal display can be mounted as a display device for a presenter or a person concerned, and a projector and a screen for projecting an image projected by the projector can be mounted as a display device shared by the presenter and the audience.
  • a dedicated liquid crystal display may be mounted on each listener.
  • the display device 5 displays a presentation screen according to an instruction from the presentation support device 10 as an example.
  • the display device 5 displays a slide of a document file opened by presentation software that operates on the presentation support device 10.
  • the display device 5 can display any slide specified by the presenter via the input device 7 among the slides included in the document file, and the slide show function of the presentation software is set to the ON state.
  • the slides included in the document file can be switched and displayed in the order in which the slides are created.
  • the input device 7 is a device that receives instruction inputs for various types of information.
  • a mouse or a keyboard or a touch sensor bonded on the liquid crystal display can be adopted as the input device 7.
  • a laser pointer indicating the position on the screen projected on the screen can be mounted as the input device 7. That is, among laser pointers, there is also a laser pointer with a remote control function provided with an operation unit such as various buttons for advancing and returning a slide page.
  • the operation unit of the laser pointer with a remote control function can be used as the input device 7.
  • an image sensor that senses the position of the light spot pointed by the laser pointer can be mounted as the input device 7.
  • the input device 7 accepts a specification of a document file to be executed by the presentation software on the presentation support device 10, an operation of advancing a slide page, an operation of returning a slide page, and the like.
  • the operation accepted through the input device 7 in this way is output to the presentation support device 10.
  • the presentation support apparatus 10 is a computer on which presentation software is executed.
  • the presentation support apparatus 10 may employ an information processing apparatus such as a desktop or notebook personal computer.
  • the presentation support apparatus 10 can employ not only a stationary terminal such as the personal computer but also various portable terminal apparatuses.
  • a stationary terminal such as the personal computer
  • portable terminal apparatuses for example, as an example of the mobile terminal device, mobile communication terminals such as smartphones, mobile phones and PHS (Personal Handyphone System), and slate terminals such as PDA (Personal Digital Assistants) are included in the category.
  • PHS Personal Handyphone System
  • slate terminals such as PDA (Personal Digital Assistants) are included in the category.
  • the presentation support apparatus 10 provides the above-described presentation support service in a stand-alone manner that independently executes the above-described presentation software without depending on external resources.
  • the presentation support service is not limited to the implementation provided in a stand-alone manner.
  • a client server system can be constructed by providing a server that provides the presentation support service to a client that executes presentation software.
  • FIG. 2 is a block diagram illustrating a functional configuration of the presentation support apparatus 10 according to the first embodiment.
  • the presentation support apparatus 10 includes an input / output I / F (InterFace) unit 11, a storage unit 13, and a control unit 15.
  • I / F InterFace
  • the input / output I / F unit 11 is an interface for performing input / output with peripheral devices such as the microphone 3, the display device 5, and the input device 7.
  • the input / output I / F unit 11 outputs the audio data input from the microphone 3 to the control unit 15. Further, the input / output I / F unit 11 outputs the slide image data output from the control unit 15 to the display device 5, highlights an area included in the slide output from the control unit 15, or cancels the instruction. An instruction is output to the display device 5. The input / output I / F unit 11 outputs various operations input from the input device 7 to the control unit 15.
  • the storage unit 13 is a device that stores data used for various programs such as an OS (Operating System) and presentation software executed by the control unit 15 and application programs.
  • OS Operating System
  • presentation software executed by the control unit 15 and application programs.
  • the storage unit 13 is implemented as a main storage device in the presentation support apparatus 10.
  • various semiconductor memory elements such as RAM (Random Access Memory) and flash memory can be employed for the storage unit 13.
  • the storage unit 13 can also be implemented as an auxiliary storage device. In this case, HDD (Hard Disk Drive), optical disk, SSD (Solid State Drive), etc. can be adopted.
  • the storage unit 13 stores document data 13a, extracted word data 13b, and recognized word data 13c as an example of data used in a program executed by the control unit 15. Note that the extracted word data 13b and the recognized word data 13c other than the document data 13a are intermediate data generated through processing by the control unit 15, and will be described together with the description of the control unit 15. In addition to the above data, the storage unit 13 can also store other electronic data such as a presentation timetable.
  • Document data 13a is data relating to a document.
  • a document file in which one or a plurality of slides are created using presentation software can be adopted as the document data 13a.
  • Such slides can be imported with text and graphics as well as content created by other application programs. For example, you can import documents created with word processing software, tables and graphs created with spreadsheet software, images and movies taken with an imaging device, and images and movies edited with image editing software. And can be imported.
  • meta information including a character string such as an explanatory word or a description of the content is added to the content other than the text before the presentation is started. Can do.
  • the control unit 15 has an internal memory for storing various programs and control data, and executes various processes using these.
  • control unit 15 is implemented as a central processing unit, a so-called CPU (Central Processing Unit).
  • CPU Central Processing Unit
  • the control unit 15 does not necessarily have to be implemented as a central processing unit, and may be implemented as an MPU (Micro Processing Unit).
  • MPU Micro Processing Unit
  • the control unit 15 can be realized by hard wired logic such as ASIC (Application Specific Integrated Circuit) or FPGA (Field Programmable Gate Array).
  • the control unit 15 virtually implements the following processing unit by executing various programs.
  • the control unit 15 includes a dividing unit 15a, an extracting unit 15b, an adding unit 15c, a recognizing unit 15d, a calculating unit 15e, a setting unit 15f, and a display control unit 15g. Have.
  • the dividing unit 15a is a processing unit that divides a slide into a plurality of regions.
  • the dividing unit 15a reads a document file that has received a designation first among document files included in the document data 13a stored in the storage unit 13.
  • the document file acquisition path is not limited thereto.
  • the dividing unit 15a can also acquire an image from an auxiliary storage device such as a hard disk or an optical disk, or a removable medium such as a memory card or a USB (Universal Serial Bus) memory.
  • the acquisition unit 15a can also acquire an image by receiving it from an external device via a network.
  • the dividing unit 15a divides the slide included in the previously read document file into a plurality of areas. For example, the dividing unit 15a divides the slide in units of one sentence, line, paragraph, and the like. In this case, the dividing unit 15a scans a character string included in the slide, detects a delimiter character corresponding to a space, a punctuation point, or a line feed, and sets the delimiter character as a boundary of the region. The dividing unit 15a divides a character string included in the slide before and after the boundary. Thus, the slide is divided into a plurality of areas for each delimiter character. In addition, the dividing unit 15a assigns an index for identifying the area to the area obtained by dividing the slide.
  • the slide may be manually divided by designating the boundary of the area via the input device 7 or the like.
  • the extraction unit 15b is a processing unit that extracts words from the character string included in the region.
  • the extraction unit 15b selects one of a plurality of areas after dividing the slide. Subsequently, the extraction unit 15b extracts a word by executing natural language processing on the character string included in the previously selected region. For example, the extraction unit 15b extracts a word whose part of speech is a noun from morphemes obtained by executing morphological analysis or the like on a character string in the region. And the extraction part 15b provides the index allocated to the area
  • the assigning unit 15c is a processing unit that assigns a weight to each word.
  • the assigning unit 15 c calculates the appearance frequency f k of the word k for each word included in the slide after the extraction unit 15 b extracts the words from all the regions. As an example of such appearance frequency, the assigning unit 15c calculates the total number of appearances for each word by counting the number of times the word k appears on the same slide. Then, applying section 15c imparts the weights w k of the word corresponding to the frequency f k, which is calculated for each word before. In this case, the assigning unit 15 c uses a weight calculation formula in which the weight w k becomes smaller as the appearance frequency f k becomes higher.
  • FIG. 3 is a diagram illustrating an example of the extracted word data 13b.
  • FIG. 3 shows extracted word data relating to one slide out of a plurality of slides.
  • the word “a”, since frequency is "2”, 0.25 is assigned as a weight to compute the 1/2 2.
  • 0.25 is given as a weight.
  • FIG. 3 illustrates the extracted word data related to one slide. However, in the other slides, the value of each item is different, but the computer can identify the word region and weight as in the example of FIG. Extracted word data is stored.
  • the assigning unit 15c can calculate the word weight fk using a factor other than the above total number of appearances, or add another factor to the above total number of appearances and add the word weight fk. k can be calculated.
  • the recognition unit 15d is a processing unit that performs voice recognition.
  • the recognition unit 15d is activated when the presentation software receives a presentation start instruction with the document file opened, and waits until an audio signal having a predetermined time length is input from the microphone 3. For example, it waits for an audio signal having a time length of at least one frame, for example, 10 msec.
  • the recognizing unit 15d performs voice recognition such as word spotting on the voice signal every time a voice signal having a predetermined time length is input from the microphone 3. At this time, the recognizing unit 15d extracts the extracted word data related to the slide that is included in the document file that is being executed by the presentation software among the extracted word data 13b stored in the storage unit 13 and that is being displayed on the display device 5. Apply to word spotting.
  • the recognition unit 15d recognizes whether or not a word extracted from each region included in the slide being displayed exists in the utterance of the presenter. Then, when a word is recognized from the audio signal, the recognition unit 15 d registers the recognition word data 13 c in which the word and the time when the word is recognized are associated with each other in the storage unit 13. When the same word is recognized a plurality of times as time passes, the last, that is, the latest recognized time is registered in the storage unit 13.
  • the recognizing unit 15d determines whether or not there is a word for which a predetermined period has elapsed since it was registered in the storage unit 13 in the recognized word data 13c stored in the storage unit 13. For example, for each word included in the recognized word data 13c, the recognizing unit 15d has a difference between the time registered in association with the word and the time when the recognizing unit 15d refers to the recognized word data 13c, that is, the current time. It is determined whether or not a predetermined threshold is exceeded. At this time, the recognizing unit 15d can change the threshold value used for the above determination according to the unit in which the slide is divided by the dividing unit 15a, for example, one sentence, line, paragraph, or the like.
  • the slide when the slide is divided in units of lines, it can be assumed that the number of characters read out in one area is approximately 20 to 30 characters. In this case, 5 to 10 seconds can be used as an example of the threshold value. Further, when the slide is divided in units of paragraphs, it can be assumed that a time longer than the line unit is devoted to reading. In this case, 20 to 30 seconds can be used as an example of the above threshold.
  • the recognition unit 15d deletes the record related to the word from the recognized word data 13c stored in the storage unit 13.
  • the recognition unit 15d leaves the word included in the recognition word data 13c stored in the storage unit 13 without deleting it.
  • the recognition unit 15d determines whether or not the slide page displayed on the display device 5 has been changed. For example, the recognizing unit 15d determines whether a slide is switched by a slide show or an operation for advancing a slide page or an operation for returning a slide page is received via the input device 7. At this time, when the slide page displayed on the display device 5 is changed, it is highly possible that the description of the presenter is switched from the slide of the page before the change to the slide of the page after the change. In this case, the recognition unit 15d deletes the recognized word data 13c stored in the storage unit 13. On the other hand, when the slide page displayed on the display device 5 is not changed, there is a high possibility that the page explained by the presenter will not change. In this case, the recognition unit 15d leaves the word included in the recognized word data 13c stored in the storage unit 13 without deleting it.
  • the recognizing unit 15d recognizes a word that is highly likely to be explained by the presenter in the displayed slide.
  • a word included in the extracted word data 13b is described as an “extracted word”
  • a word included in the recognized word data 13c is described as a “recognized word” to distinguish the labels from each other. is there.
  • the calculation unit 15e is a processing unit that calculates the degree of association between the region in the slide being displayed and the word obtained as a speech recognition result.
  • the calculation unit 15e selects one index among the indexes of the area included in the slide that is being displayed on the display device 5. Subsequently, the calculation unit 15e calculates the area from the weight assigned to the extracted word that matches the recognized word of the recognized word data 13c among the extracted words of the extracted word data 13b associated with the previously selected index area. The relevance of is calculated. For example, when calculating the degree of association r x of the region x using the above-described word weight w k , the calculation unit 15e adds the weights w k given to the extracted words that match the recognized word, thereby calculating the degree of association. r x can be calculated.
  • the relevance of the area is calculated as zero.
  • the degree of relevance between the description contents of each area in the slide and the utterance contents of the presenter is obtained as the “relevance degree”.
  • the setting unit 15f is a processing unit that sets the speed at which highlight display of the area in the slide is advanced.
  • the speed at which highlight display proceeds is sometimes referred to as “highlight speed”.
  • the setting unit 15f sets a higher highlight speed for a region with a higher relevance or a lower highlight speed for a region with a lower relevance each time the relevance is calculated by the calculator 15e.
  • the above related to the calculation formula of the highlight velocity v x 'v x V ⁇ r x " It can be calculated by substituting the degree r x .
  • “V” included in the calculation formula is a predetermined fixed value. That is, by using the calculation formula of the highlights velocity v x, it is possible to calculate a highlight velocity v x proportional to the value of relevance r x.
  • the display control unit 15g is a processing unit that executes display control for the display device 5.
  • the display control unit 15g when a document file is opened by presentation software, the display control unit 15g causes the display device 5 to display a slide included in the document file. At this time, the display control unit 15g may display the slide of the first page among the slides included in the document file, or may display the slide of the page edited last.
  • the display control unit 15g executes the following process every time the setting unit 15f sets the highlight speed of each area. That is, the display control unit 15g advances the highlight display according to the highlight speed set for each area included in the slide being displayed. That is, the display control unit 15g does not always complete the highlight display immediately because a value greater than zero is set as the highlight speed of the region. That is, the display control unit 15g advances the highlight display toward the completion at the highlight speed set by the setting unit 15f. As a result, the highlight display is advanced toward a display form different from the display form set at the time of creating the slide in the region where the highlight speed greater than zero is set.
  • the degree of progress of highlight display of an area toward completion may be referred to as “progress”.
  • the display control unit 15g can execute arbitrary highlight display.
  • the display control unit 15g can realize the highlight display by increasing the luminance of the character string included in the region or the background of the character string, rather than the luminance set to the region at the time of creating the slide.
  • the display control unit 15g may change the font of the character string, or change the background display color or fill.
  • the display control unit 15g can also realize highlighting by highlighting the area.
  • the display control unit 15g monitors the presence / absence of a region where the progress of highlight display is equal to or greater than a predetermined threshold.
  • a predetermined threshold When there is an area where the progress of highlight display is equal to or greater than a predetermined threshold, the area is higher than the area where the progress of highlight display is less than the threshold, in other words, the sum of relevance. It can be judged that the average was maintained at a high level.
  • the display control unit 15g maintains the setting of the highlight speed of the area where the progress of highlight display is equal to or higher than the threshold, and the highlight display of the area where the progress of highlight display is less than the threshold is changed to the original.
  • the highlight speed of the area where the progress of highlight display is less than the threshold is reset to zero.
  • highlight display is executed by narrowing down to a region where it can be determined that the presenter is an explanation location based on the passage of time or the accumulation of extracted words that match the recognized word over time.
  • the display control unit 15g determines whether or not the degree of association calculated by the calculation unit 15e decreases in an area where the progress degree of highlight display is equal to or greater than the threshold value. For example, if the current relevance level calculated this time is lower than the past relevance level calculated last time, the number of extracted words that match the recognized word decreases over time, or only extracted words with low weight are recognized. It can be determined that there is a change in situation such as not matching the word. In this case, the display control unit 15g returns the highlight display of the area where the degree of relevance is reduced to the original state, and resets the highlight speed to zero. In this example, the highlight display is canceled when the current relevance level is lower than the past relevance level. However, the highlight display is performed when the current relevance level is lower than the past relevance level by a certain value. You can also cancel.
  • the display control unit 15g when the display control unit 15g receives a page switching instruction via the input device 7, the display control unit 15g changes the slide to be displayed on the display device 5. For example, when an operation for advancing a page is received, the display control unit 15g causes the display device 5 to display a slide on the next page of the slide being displayed. Further, when the operation of returning the page is received, the display control unit 15g causes the display device 5 to display the slide of the previous page of the slide being displayed.
  • FIG. 4 is a diagram illustrating an example of a temporal change related to the progress of highlight display.
  • 5 and 6 are diagrams showing an example of transition of the slide screen.
  • region which the slide currently displayed on the display apparatus 5 includes using the extracted word data 13b shown in FIG. 3 is illustrated.
  • 5 and 6 exemplify a case in which the slide being displayed on the display device 5 includes three areas of the indices idx1 to idx3 according to the example of the extracted word data 13b illustrated in FIG.
  • the highlight display a case where reverse display of each area is executed according to the highlight speed set for the area is illustrated.
  • the relevance of each region is calculated as follows. That is, as shown in FIG. 3, since the extracted word includes the recognized word “a” in which the weight “0.25” is set in the two areas of the index idx1 and the index idx2, the degree of association is “0”. .25 ". On the other hand, since no recognized word is included in the area of index idx3, the degree of association is calculated as “0”. As a result, a highlight speed proportional to the relevance “0.25” is set in the two areas of index idx1 and index idx2, and the highlight speed is set to zero in the area of index idx3. .
  • the slide screen shown second from the top in FIG. 5 is displayed. That is, in the two areas of the index idx1 and the index idx2, the highlight display progresses to the same extent, while the area of the index idx3 transitions to a state where the highlight display does not progress.
  • the word “b” is recognized at the time t2.
  • the area of the index idx1 two recognition words of a recognition word “a” with a weight “0.25” set as an extracted word and a recognition word “b” with a weight “0.25” set Therefore, the relevance is calculated as “0.5”.
  • the extracted word includes the recognized word “a” with the weight “0.25” set in the area of the index idx2, the degree of association is calculated as “0.25”. Is done.
  • the recognized word “b” with the weight “0.25” set in the extracted word is included in the area of the index idx3, the degree of association is calculated as “0.25”.
  • a highlight speed proportional to the degree of association “0.5” is set in the index idx1
  • a highlight speed proportional to the degree of association “0.25” is set in the area of the index idx2.
  • the screen transitions to the third slide screen from the top in FIG. That is, a difference starts in the progress of highlight display between the two areas of index idx1 and index idx2.
  • the area with the index idx1 is highlighted more greatly than the area with the index idx2.
  • the area of the index idx3 is delayed from the progress of the highlight display of the index idx2 by the amount that the highlight speed is set later than the index idx2.
  • the word “c” is recognized at the time t3.
  • the recognized word “a” with the weight “0.25”, the recognized word “b” with the weight “0.25”, and the weight “0.25” are set.
  • the extracted word includes three recognition words “c” for which “is set”. For this reason, the relevance is calculated as “0.75” by the calculation of “0.25 + 0.25 + 0.25”.
  • the area of index idx2 includes a recognized word “a” in which a weight “0.25” is set for the extracted word and a recognized word “c” in which the weight “0.25” is set for the extracted word. Therefore, the relevance is calculated as “0.5”.
  • the extracted word includes the recognized word “b” having the weight “0.25”, so the degree of association is calculated as “0.25”. Is done.
  • a highlight speed proportional to the relevance degree “0.75” is set in the index idx1
  • a highlight speed proportional to the relevance degree “0.5” is set in the area of the index idx2.
  • a highlight speed proportional to the relevance “0.25” is set. That is, at time t3, the highlight speed of each region is “idx1> idx2> idx3”.
  • the screen changes to the fourth slide screen from the top in FIG. That is, the process proceeds until the highlight display of the area of index idx1 reaches the threshold value.
  • the two areas of the index idx2 and the index idx3 have a difference in their progress, the progress does not reach the threshold.
  • the screen transitions to the slide screen shown first from the top in FIG. That is, the area of the index idx1 in which the highlight display has progressed to the threshold is maintained as it is, while the area of the index idx2 and the index idx3 in which the highlight display has not progressed to the threshold is highlighted. Canceled.
  • the transition is made with the first slide screen from the top in FIG. 6 until time t5, and the recognized words “a”, “b”, and “c” are deleted from the recognized word data 13c at time t5. If it does, it will change to the slide screen shown to the 2nd from the top of FIG. That is, since the three regions from index idx1 to index idx3 do not include the recognized word, the relevance is calculated as “0”. As a result, the degree of relevance of the area with the index idx1 in which the highlight display has progressed to the threshold value is reduced, so that the highlight display of the area with the index idx1 is canceled. For this reason, all three areas of the index idx1 to the index idx3 return to the default display form set when the slide is created. The second slide screen from the top in FIG. 6 is maintained until the word “e” is recognized at time t6.
  • the relevance of each region is calculated as follows.
  • the index idx3 is the only area in which the recognized word “e” is included in the extracted word among the three areas.
  • the weight “1” set for the recognized word “e” is calculated as the degree of association in the area of the index idx3.
  • the degree of association is calculated as “0”.
  • the highlight speed is set to zero in the two areas of index idx1 and index idx2, and the highlight speed proportional to the degree of association “1” is set in the area of index idx3.
  • the highlight display of the area of index idx3 proceeds to the threshold value.
  • a transition is made to the third slide screen from the top in FIG.
  • the area of index idx3 where the highlight display has progressed to the threshold is maintained as it is, while the areas of index idx2 and index idx3 both have the default display form set when the slide is created. Maintained.
  • the period until the progress of highlight display in the area of index idx3 reaches the threshold is shorter than the period until the progress of highlight display of the area of index idx1 reaches the threshold.
  • a highlight speed proportional to the relevance degree “1” is set, and therefore, the relevance degree “0.75” is the highest among the time points of time t1, time t2, and time t3.
  • the period until the degree of progress reaches the threshold is shorter than the highlight display in the area of the index idx1 in which the highlight speed proportional to is set.
  • the presenter when the presenter explains the area of index idx1 from time t1 to time t5, the area including index idx1 can be highlighted. That is, at the stage of time t1 and time t2, just because the recognized word is included in the extracted word in the area of index idx2 or index idx3, only the highlight display of the area of index idx2 or index idx3 does not proceed. The highlight display also proceeds in the area of index idx1. Therefore, it is possible to suppress the situation where the presenter's explanation part is not highlighted.
  • FIG. 7 is a flowchart illustrating the procedure of the weighting process according to the first embodiment.
  • This process can be started automatically or manually. For example, when starting automatically, when the presentation software saves the document file in the storage unit 13 and then closes, or when the document file is saved in the storage unit 13 while editing the document file via the presentation, Can be activated.
  • the processing can be activated when an instruction to execute presentation pre-processing is received via the input device 7. In any case, the processing is started by reading out the document file corresponding to the save or execution instruction from the document files included in the document data 13a stored in the storage unit 13.
  • the dividing unit 15a divides the slide included in the document file into a plurality of areas in units of one sentence, line, paragraph, or the like (step S101). Subsequently, the dividing unit 15a assigns an index for identifying each area to the area obtained in step S101 (step S102).
  • the extraction part 15b selects one index among the indexes allocated by step S102 (step S103). Subsequently, the extraction unit 15b extracts a word whose part of speech is a noun from morphemes obtained by executing morphological analysis or the like on the character string in the index area selected in step S103 (step S104). Thereafter, the extraction unit 15b gives each word extracted in step S104 an index assigned to the area including the word (step S105).
  • the extraction unit 15b repeatedly executes the processes from step S103 to step S105 until all the indexes assigned in step S102 are selected (No in step S106).
  • step S106 Yes when all the indexes assigned in step S101 are selected (step S106 Yes), the assigning unit 15c calculates the appearance frequency f k of the word k for each word included in the slide (step S107). Then, the assigning unit 15c assigns a word weight w k corresponding to the appearance frequency f k calculated for each word in step S107 (step S108). On top of that, imparting unit 15c, the word k, and registers the extracted word data 13b index idx and weights w k is associated to the storage unit 13 (step S109), and ends the process.
  • FIG. 8 is a flowchart illustrating the procedure of voice recognition processing according to the first embodiment. This process is started when the presentation software receives a presentation start instruction with the document file opened, and is repeatedly executed until a presentation end instruction is received.
  • the recognition unit 15d waits until an audio signal having a predetermined time length is input from the microphone 3 until an audio signal having a time length of, for example, at least one frame, for example, 10 msec is input (Step S15). S301).
  • Step S301 when an audio signal having a predetermined time length is input from the microphone 3 (Yes at Step S301), the recognition unit 15d performs voice recognition such as word spotting on the audio signal (Step S302).
  • word spotting is executed in step S302
  • the slide is included in the document file being executed by the presentation software among the extracted word data 13b stored in the storage unit 13 and is being displayed on the display device 5.
  • Extracted word data relating to a certain slide is applied as dictionary data.
  • the recognition unit 15d stores the recognized word data 13c in which the word recognized in step S302 and the time when the word is recognized are associated with each other. (Step S304), and the process proceeds to step S305.
  • Step S301 or Step S303 when a voice signal having a predetermined time length is not input from the microphone 3 or when a word is not recognized from the voice signal (No in Step S301 or Step S303), the subsequent processing is skipped and the processing proceeds to Step S305. .
  • the recognizing unit 15d determines whether or not there is a word for which a predetermined period has elapsed since registration in the storage unit 13 among the recognized word data 13c stored in the storage unit 13 (step S305). If there is a word for which a predetermined period has elapsed since registration in the storage unit 13 (Yes in step S305), the recognition unit 15d deletes the record related to the word from the recognized word data 13c stored in the storage unit 13. (Step S306). If there is no word for which a predetermined period has elapsed since registration in the storage unit 13 (No in step S305), the process of step S306 is skipped and the process proceeds to step S307.
  • the recognition unit 15d determines whether or not the slide page displayed on the display device 5 has been changed (step S307). At this time, when the slide page displayed on the display device 5 is changed (Yes in step S307), the recognition unit 15d deletes the recognized word data 13c stored in the storage unit 13 (step S308), and in step S301. Return to processing. If the slide page displayed on the display device 5 has not been changed (No at Step S307), the process returns to Step S301 without executing Step S308.
  • FIG. 9 is a flowchart illustrating the procedure of the display control process according to the first embodiment. This process is executed in parallel with the voice recognition process shown in FIG. 8, and is started when the presentation software receives a presentation start instruction with the document file opened, and gives a presentation end instruction. It is repeatedly executed until it is accepted.
  • the cycle in which the execution of the process is repeated may be the same as or different from the voice recognition process shown in FIG. 8, and is executed in synchronization with the voice recognition process shown in FIG. It can also be executed asynchronously.
  • the calculation unit 15e selects one index from among the indexes of the area included in the slide being displayed on the display device 5 (step S501). Subsequently, the calculation unit 15e calculates the degree of association of the area from the weights assigned to the extracted words that match the recognized word among the extracted words of the extracted word data 13b associated with the index area selected in step S501. Calculate (step S502).
  • the setting unit 15f sets the highlight speed higher as the relevance calculated in step S502 is higher for the index area selected in step S501, or sets the highlight speed lower as the relevance is lower. (Step S503).
  • step S501 to step S503 are repeatedly executed until all indexes are selected (No in step S504).
  • the processing from step S501 to step S503 is repeatedly executed until the highlight speed is set for all regions.
  • step S504 Yes the presence / absence of an area where the progress of highlight display is equal to or greater than a predetermined threshold is monitored (step S505).
  • the display control unit 15g executes the following process. That is, the display control unit 15g advances the highlight display of each area according to the highlight speed set for each area in step S503 (step S506), and ends the process.
  • the display control unit 15g executes the following process. That is, the display control unit 15g maintains the setting of the highlight speed in the area where the progress of highlight display is equal to or higher than the threshold, and the highlight display of the area where the progress of highlight display is less than the threshold is in the original state. The highlight display is canceled by returning to (step S507), and the highlight speed of the area where the progress of highlight display is less than the threshold is reset to zero (step S508).
  • the display control unit 15g determines whether or not the current relevance level is less than the past relevance level in an area where the progress degree of highlight display is equal to or greater than the threshold (step S509). At this time, if the current relevance level is equal to or higher than the past relevance level (No in step S509), the display control unit 15g advances the highlight display of each area according to the highlight speed set in each area in step S508. (Step S506), the process ends.
  • step S509 the display control unit 15g cancels the highlight display by returning the highlight display of the area with the low relevance level to the original state.
  • Step S510 the highlight speed is reset to zero (Step S511), and the process ends.
  • the presentation support apparatus 10 when the presentation support apparatus 10 according to the present embodiment highlights an area including a word obtained as a speech recognition result on the presentation screen, the presentation support apparatus 10 The speed of highlighting is increased as the display is made, and the speed of highlighting is lowered as the display is in a region having a lower degree of association.
  • the area where the highlight display is executed is not necessarily limited to the alternative. For this reason, when a word is detected across a plurality of areas by voice recognition, highlight display is executed in each area. Therefore, it is possible to increase the possibility that the presenting part of the presenter is included in the area where the highlight display is executed. Therefore, according to the presentation support apparatus 10 according to the present embodiment, it is possible to suppress the situation where the presenting part of the presenter is not highlighted.
  • the speed of highlighting a region varies depending on the degree of association between words and regions obtained as a result of speech recognition. For example, even when a plurality of areas are highlighted, the response of highlight display regarding an area that is highly likely to be an explanation part of the presenter is increased. As a result, it is possible to easily draw attention to a region that is highly likely to be an explanation location of the presenter. At the same time, the response of highlight display relating to the region that is unlikely to be an explanation location of the presenter is lowered. As a result, it is possible to delay the perception of an area where the possibility of not being the presenting part of the presenter is higher than other areas. As described above, it is possible to suppress a situation in which attention is focused on a highlight display that may cause an error while suppressing a decrease in response from the utterance of the presenter to the highlight display.
  • the presentation support device 10 generates extracted word data in the same manner as in the first embodiment by extracting words from character strings included in meta information set in graphs, tables, images, and moving images. be able to.
  • the display control unit 15g can switch between display and non-display of the highlight depending on whether or not the relevance of the region or the highlight speed of the region is equal to or higher than a predetermined threshold. In this case, the display and non-display of the highlight can be controlled before the highlight display progresses with time.
  • the appearance frequency is not necessarily limited to the total number of appearances.
  • the number of appearances between regions in which the number of times the word k appears between the regions can be counted can be used as the appearance frequency.
  • the frequency of appearance between the areas is 1/3, so the weight of the word k is 1 / ( 1/3) 2 is given. At this time, the number of times that the word k appears in one region is not counted as the total number of times, and the same weight is given even if it appears multiple times.
  • the assigning unit 15c can assign a weight to the word k according to the number of mora of the word k. Specifically, the assigning unit 15c can assign a greater weight as the number of mora of the word k increases. For example, the following two formulas can be used as an example of a weight calculation formula.
  • the accuracy of speech recognition tends to decrease as the number of mora phonemes decreases. Therefore, according to the above weighting method, a larger weight is given when the number of mora phonemes of the word k is larger than when the number of mora phonemes of the word k is small. As a result that can be given, the calculation accuracy of the relevance can be improved.
  • the weighting method described in this section can be implemented by using it alone instead of the weighting method described in the first embodiment, or the weighting method described in the first embodiment. It can also be implemented in combination with other weighting methods described in the second embodiment.
  • the presentation support apparatus 10 can give a weight to a recognized word after voice recognition is performed. That is, when speech recognition is performed, the likelihood that the recognized word is a correct answer such as learning data, that is, a so-called score is often calculated together with the recognized word. For this reason, the presentation support apparatus 10 can also give a weight to the recognized word according to the score. It should be noted that the weighting method described in this section can be implemented by using it alone instead of the weighting method described in the first embodiment, or the weighting method described in the first embodiment. It can also be implemented in combination with other weighting methods described in the second embodiment.
  • the above-mentioned ratio is used as it is as the degree of relevance, or the total value obtained by adding the weights of the extracted words that match the recognized word is normalized using the above-mentioned ratio, thereby increasing the calculation accuracy of the degree of relevance.
  • the weighting method described in this section can be implemented by using it alone instead of the weighting method described in the first embodiment, or the weighting method described in the first embodiment. It can also be implemented in combination with other weighting methods described in the second embodiment.
  • the presentation support apparatus 10 can accept an instruction to accelerate or cancel highlight display via the input device 7 or the like.
  • an instruction to accelerate highlight display or an instruction to cancel highlight display is given to a predetermined key included in a keyboard, a predetermined button of a mouse, or a predetermined button of a laser pointer with a remote control function. Assign keys and buttons to accept.
  • the presentation support apparatus 10 receives an instruction for accelerating highlight display, the presentation support apparatus 10 accelerates the highlight speed in a region where the degree of progress of highlight display is the highest when the instruction is received.
  • the presentation support apparatus 10 also includes raising the progress of highlight display in a region where the progress of highlight display is the highest to a threshold value at a time.
  • the presentation support apparatus 10 cancels highlight display of the area where the progress of highlight display is the highest when the instruction is received and highlights the area. You can also reset the speed.
  • the presenter creates slides that the presenter himself uses for the presentation, and the presenter himself assembles the slide description order and logical configuration as preparations for the presentation. It is likely that you will notice the highlighted area on the slide. This increases the possibility that the listener can accept an instruction to accelerate the highlight display or notice an instruction to cancel the highlight display before noticing the highlight display. Therefore, it is possible to increase the response of highlight display to the listener, or to suppress a situation where the listener notices an erroneous highlight display.
  • the presentation support apparatus 10 provides the presentation support service in a stand-alone manner in which the presentation software 10 is independently executed without depending on an external resource.
  • a client server system can be constructed by providing a server that provides the presentation support service to a client that executes presentation software.
  • the server device can be implemented by installing a presentation support program for realizing the above presentation support service as package software or online software.
  • the server device 10 may be implemented as a Web server that provides the presentation support service, or may be implemented as a cloud that provides the presentation support service by outsourcing.
  • the presentation support program is added to the presentation software. However, when a request for referring to the presentation support program as a library is received from a client having a license authority, the presentation support program is It can also be plugged in.
  • FIG. 10 is a diagram illustrating a hardware configuration example of a computer that executes the presentation support program according to the first and second embodiments.
  • the computer 100 includes an operation unit 110a, a speaker 110b, a camera 110c, a display 120, and a communication unit 130. Further, the computer 100 includes a CPU 150, a ROM 160, an HDD 170, and a RAM 180. These units 110 to 180 are connected via a bus 140.
  • the HDD 170 is similar to the dividing unit 15a, the extracting unit 15b, the adding unit 15c, the recognizing unit 15d, the calculating unit 15e, the setting unit 15f, and the display control unit 15g described in the first embodiment.
  • a presentation support program 170a that performs the function is stored.
  • This presentation support program 170a is integrated or separated in the same manner as each component of the dividing unit 15a, extracting unit 15b, adding unit 15c, recognizing unit 15d, calculating unit 15e, setting unit 15f and display control unit 15g shown in FIG. It doesn't matter. That is, the HDD 170 does not necessarily have to store all the data shown in the first embodiment, and data used for processing may be stored in the HDD 170.
  • the CPU 150 reads the presentation support program 170a from the HDD 170 and develops it on the RAM 180.
  • the presentation support program 170a functions as a presentation support process 180a as shown in FIG.
  • the presentation support process 180a expands various data read from the HDD 170 in an area allocated to the presentation support process 180a in the storage area of the RAM 180, and executes various processes using the expanded data.
  • examples of processing executed by the presentation support process 180a include the processing shown in FIGS.
  • the CPU 150 does not necessarily operate all the processing units described in the first embodiment, and the processing unit corresponding to the process to be executed may be virtually realized.
  • the presentation support program 170a may not necessarily be stored in the HDD 170 or the ROM 160 from the beginning.
  • each program is stored in a “portable physical medium” such as a flexible disk inserted into the computer 100, so-called FD, CD-ROM, DVD disk, magneto-optical disk, or IC card. Then, the computer 100 may acquire and execute each program from these portable physical media.
  • each program is stored in another computer or server device connected to the computer 100 via a public line, the Internet, a LAN, a WAN, etc., and the computer 100 acquires and executes each program from these. It may be.

Abstract

A presentation support device (10) extracts a first word from a character string in each of the regions into which a page of a document file is divided, each of the pages of the document file being displayed on a per screen basis during display. The presentation support device (10) carries out speech recognition, and calculates a degree of relevance from the first word extracted from each region in the page being displayed on a display device 5 and a second word acquired as a result of the speech recognition. The presentation support device (10) sets a faster speed for progressing the highlighting of a region the higher the degree of relevance calculated for that region is or sets a slower speed for progressing the highlighting of a region the lower the degree of relevance for that region is. The presentation support device (10) controls the highlight display on a page in accordance with a speed set for each region.

Description

プレゼンテーション支援方法、プレゼンテーション支援プログラム及びプレゼンテーション支援装置Presentation support method, presentation support program, and presentation support apparatus
 本発明は、プレゼンテーション支援方法、プレゼンテーション支援プログラム及びプレゼンテーション支援装置に関する。 The present invention relates to a presentation support method, a presentation support program, and a presentation support apparatus.
 プレゼンテーションを支援する技術の一例として、プレゼンタが説明中である箇所をプレゼンタや聴講者に提示するものがある。例えば、原稿の読み飛ばしを抑制することを目的とする表示装置が提案されている。この表示装置では、話者が発声した語句を認識し、認識した語句をもとに、表示パネルに表示中の原稿のうち読み上げられた部分を特定し、この特定した部分の表示状態を、第1の表示状態とは異なる第2の表示状態、例えば点滅等のハイライト表示に変化させる。 As an example of a technology that supports presentation, there is a technique that presents a presenter or a listener with a part that the presenter is explaining. For example, a display device has been proposed that aims to suppress skipping of a document. This display device recognizes a phrase uttered by a speaker, identifies a read-out portion of a document being displayed on a display panel based on the recognized phrase, and displays a display state of the identified portion as a first state. The display state is changed to a second display state different from the first display state, for example, highlight display such as blinking.
特開2009-271814号公報JP 2009-271814 A 特開2005-208292号公報JP 2005-208292 A 特開2002-268667号公報JP 2002-268667 A 特開昭61-036853号公報Japanese Patent Laid-Open No. 61-036853
 しかしながら、上記の技術では、次に説明するように、プレゼンタの説明箇所がハイライト表示されない場合がある。 However, in the above technique, there are cases where the explanation part of the presenter is not highlighted as described below.
 すなわち、上記の表示装置では、話者が発声する語句を得るために音声認識が用いられる。ところが、音声認識で誤認識が発生する場合、誤認識に伴って話者が説明中でない箇所がハイライト表示される結果、話者の説明箇所がハイライト表示されない場合がある。この場合、表示装置は、話者や聴講者に説明箇所を提示できず、プレゼンテーションを妨げてしまうことがある。 That is, in the above display device, speech recognition is used to obtain a phrase uttered by a speaker. However, when misrecognition occurs in voice recognition, a portion where the speaker is not explaining due to the misrecognition is highlighted, and as a result, the speaker's explanation portion may not be highlighted. In this case, the display device cannot present the explanation part to the speaker or the listener, and may disturb the presentation.
 1つの側面では、プレゼンタの説明箇所がハイライト表示されない事態を抑制できるプレゼンテーション支援方法、プレゼンテーション支援プログラム及びプレゼンテーション支援装置を提供することを目的とする。 In one aspect, an object of the present invention is to provide a presentation support method, a presentation support program, and a presentation support apparatus that can suppress a situation where an explanation part of a presenter is not highlighted.
 一態様のプレゼンテーション支援方法は、コンピュータが、表示時に画面単位で表示されるページを含む文書ファイルのページが分割された領域ごとに当該領域が含む文字列から第1の単語を抽出する処理を実行する。さらに、前記コンピュータが、音声認識を実行し、所定の表示部に表示中であるページ内の領域ごとに当該領域から抽出された第1の単語と前記音声認識の結果として得られる第2の単語とから関連度を算出する処理を実行する。さらに、前記コンピュータが、前記領域ごとに算出された関連度が高い領域ほど前記領域のハイライト表示を進行させる速度を高く設定し、あるいは関連度が低い領域ほど前記領域のハイライト表示を進行させる速度を低く設定する処理を実行する。さらに、前記コンピュータが、前記領域ごとに設定された速度にしたがって前記ページ内のハイライト表示を制御する処理を実行する。 According to one aspect of the presentation support method, a computer executes a process of extracting a first word from a character string included in each area of a document file including a page displayed on a screen basis when the page is divided. To do. Further, the computer executes speech recognition, and for each region in the page being displayed on the predetermined display unit, the first word extracted from the region and the second word obtained as a result of the speech recognition The process of calculating the degree of association from the above is executed. Further, the computer sets a higher speed for the highlight display of the region as the region having a higher degree of relevance calculated for each region, or advances the highlight display of the region as the region having a lower relevance level. Execute processing to set the speed lower. Further, the computer executes a process of controlling the highlight display in the page according to the speed set for each area.
 プレゼンタの説明箇所がハイライト表示されない事態を抑制できる。 ∙ The situation where the explanation part of the presenter is not highlighted can be suppressed.
図1は、実施例1に係るプレゼンテーション支援システムの構成を示す図である。FIG. 1 is a diagram illustrating the configuration of the presentation support system according to the first embodiment. 図2は、実施例1に係るプレゼンテーション支援装置の機能的構成を示すブロック図である。FIG. 2 is a block diagram illustrating a functional configuration of the presentation support apparatus according to the first embodiment. 図3は、抽出単語データの一例を示す図である。FIG. 3 is a diagram illustrating an example of extracted word data. 図4は、ハイライト表示の進行度に関する時間変化の一例を示す図である。FIG. 4 is a diagram illustrating an example of a temporal change related to the progress of highlight display. 図5は、スライド画面の遷移例を示す図である。FIG. 5 is a diagram illustrating a transition example of the slide screen. 図6は、スライド画面の遷移例を示す図である。FIG. 6 is a diagram illustrating a transition example of the slide screen. 図7は、実施例1に係る重み付与処理の手順を示すフローチャートである。FIG. 7 is a flowchart illustrating the procedure of the weighting process according to the first embodiment. 図8は、実施例1に係る音声認識処理の手順を示すフローチャートである。FIG. 8 is a flowchart illustrating the procedure of the speech recognition process according to the first embodiment. 図9は、実施例1に係る表示制御処理の手順を示すフローチャートである。FIG. 9 is a flowchart illustrating the procedure of the display control process according to the first embodiment. 図10は、実施例1及び実施例2に係るプレゼンテーション支援プログラムを実行するコンピュータのハードウェア構成例を示す図である。FIG. 10 is a diagram illustrating a hardware configuration example of a computer that executes the presentation support program according to the first embodiment and the second embodiment.
 以下に添付図面を参照して本願に係るプレゼンテーション支援方法、プレゼンテーション支援プログラム及びプレゼンテーション支援装置について説明する。なお、この実施例は開示の技術を限定するものではない。そして、各実施例は、処理内容を矛盾させない範囲で適宜組み合わせることが可能である。 Hereinafter, the presentation support method, presentation support program, and presentation support apparatus according to the present application will be described with reference to the accompanying drawings. Note that this embodiment does not limit the disclosed technology. Each embodiment can be appropriately combined within a range in which processing contents are not contradictory.
[システム構成]
 図1は、実施例1に係るプレゼンテーション支援システムの構成を示す図である。図1に示すプレゼンテーション支援システム1は、文書ファイルが表示装置5に表示されたプレゼンテーション画面のうちマイク3から入力される音声の認識結果として得られた単語を含む領域をハイライト表示させるプレゼンテーション支援サービスを提供する。
[System configuration]
FIG. 1 is a diagram illustrating the configuration of the presentation support system according to the first embodiment. The presentation support system 1 shown in FIG. 1 highlights an area including a word obtained as a result of recognition of speech input from the microphone 3 in a presentation screen on which a document file is displayed on the display device 5. I will provide a.
 かかるプレゼンテーション支援サービスの一環として、プレゼンテーション支援システム1は、単語との関連度が高い領域の表示ほどハイライトの速度を上げ、関連度が低い領域の表示ほどハイライトの速度を下げる表示制御を実現する。これをもって、プレゼンタの説明箇所がハイライト表示されない事態を抑制する。 As part of this presentation support service, the presentation support system 1 realizes display control that increases the speed of highlighting as the area having a higher degree of relevance to a word and lowers the speed of highlighting as the area having a lower degree of relevance. To do. This suppresses the situation where the presenter's explanation part is not highlighted.
 ここで、以下では、一例として、上記の表示制御に関する機能がプレゼンテーションソフトにアドオンされる場合を想定し、当該プレゼンテーションソフトを用いて作成された文書ファイルが含む1または複数のスライドを表示装置5に表示させることによってプレゼンテーションが進行される場合を想定する。かかるスライドには、テキストや図形を始め、他のアプリケーションプログラムによって作成されたコンテンツをインポートすることができる。例えば、ワープロソフトで作成された文書、表計算ソフトで作成された表やグラフをインポートしたり、撮像デバイスで撮像された画像や動画、さらには、画像編集ソフトで編集された画像や動画などをインポートしたりすることができる。 Here, in the following, as an example, it is assumed that the above display control function is added to presentation software, and one or more slides included in a document file created using the presentation software are displayed on the display device 5. Assume that the presentation is progressed by displaying. Such slides can be imported with text and graphics as well as content created by other application programs. For example, you can import documents created with word processing software, tables and graphs created with spreadsheet software, images and movies taken with an imaging device, and images and movies edited with image editing software. And can be imported.
 図1に示すように、プレゼンテーション支援システム1には、マイク3と、表示装置5と、入力装置7と、プレゼンテーション支援装置10とが収容される。これらマイク3、表示装置5及び入力装置7などの周辺機器と、プレゼンテーション支援装置10との間は、有線または無線により接続される。 As shown in FIG. 1, the presentation support system 1 accommodates a microphone 3, a display device 5, an input device 7, and a presentation support device 10. The peripheral devices such as the microphone 3, the display device 5 and the input device 7 and the presentation support device 10 are connected by wire or wirelessly.
 マイク3は、音声を電気信号に変換する装置であり、マイクロフォンと呼ばれることもある。例えば、マイク3は、プレゼンテーションを実施するプレゼンタに装着させることができる。この場合、ヘッドセット型やタイピン型のマイクをプレゼンタの身体や衣服の所定位置に装着させたり、ハンド型のマイクをプレゼンタに携帯させたりすることができる。また、マイク3は、プレゼンタの発話が集音できる範囲の所定位置に設置することもできる。この場合、マイク3には、取付け型や据置き型のマイクを採用することもできる。これらいずれの場合においても、マイク3には、任意のタイプの指向性を持つマイクを採用できるが、プレゼンタの発話以外の音声、例えば聴講者等の発話や騒音などの雑音が集音されるのを抑制するために、マイクの感度をプレゼンタの発声方向に限定することもできる。なお、マイク3には、ダイナミック型、エレクトレットコンデンサ型、コンデンサ型などの任意の変換方式を採用することができる。 The microphone 3 is a device that converts sound into an electrical signal, and is sometimes called a microphone. For example, the microphone 3 can be attached to a presenter who performs a presentation. In this case, a headset-type or tie-pin type microphone can be attached to a predetermined position of the presenter's body or clothes, or a hand-type microphone can be carried by the presenter. The microphone 3 can also be installed at a predetermined position in a range where the utterance of the presenter can be collected. In this case, the microphone 3 may be an attachment type or a stationary type microphone. In any of these cases, a microphone having any type of directivity can be adopted as the microphone 3, but sounds other than the presenter's utterance, for example, the utterance of the listener and the noise such as noise are collected. In order to suppress this, the sensitivity of the microphone can be limited to the speaking direction of the presenter. The microphone 3 can employ any conversion method such as a dynamic type, an electret capacitor type, or a capacitor type.
 このマイク3に音声を採取することにより得られたアナログ信号は、デジタル信号へ変換された上でプレゼンテーション支援装置10へ入力される。 The analog signal obtained by collecting sound in the microphone 3 is converted into a digital signal and then input to the presentation support apparatus 10.
 表示装置5は、各種の情報を表示する装置である。例えば、表示装置5には、発光により表示を実現する液晶ディスプレイや有機EL(electroluminescence)ディスプレイなどを採用することもできるし、投影により表示を実現するプロジェクタを採用することもできる。また、表示装置5の設置台数は、必ずしも1台に限定されずともよく、複数の台数であってかまわない。例えば、プレゼンタもしくはその関係者用の表示装置として液晶ディスプレイを実装すると共に、プレゼンタ及び聴講者の共用の表示装置としてプロジェクタ及びプロジェクタが投影する画像を映すスクリーンを実装することもできる。また、各聴講者に専用の液晶ディスプレイを実装することとしてもかまわない。 The display device 5 is a device that displays various types of information. For example, the display device 5 may be a liquid crystal display or an organic EL (electroluminescence) display that realizes display by light emission, or a projector that realizes display by projection. Further, the number of installed display devices 5 is not necessarily limited to one, and a plurality of display devices 5 may be provided. For example, a liquid crystal display can be mounted as a display device for a presenter or a person concerned, and a projector and a screen for projecting an image projected by the projector can be mounted as a display device shared by the presenter and the audience. In addition, a dedicated liquid crystal display may be mounted on each listener.
 この表示装置5は、一例として、プレゼンテーション支援装置10からの指示にしたがってプレゼンテーション画面を表示する。例えば、表示装置5は、プレゼンテーション支援装置10上で動作するプレゼンテーションソフトが開く文書ファイルのスライドを表示する。この場合、表示装置5は、文書ファイルが含むスライドのうちプレゼンタが入力装置7を介して指定する任意のスライドを表示させることもできるし、プレゼンテーションソフトが有するスライドショーの機能がON状態に設定された場合、各スライドが作成されたページ順に文書ファイルが含むスライドを切り替えて表示させることもできる。 The display device 5 displays a presentation screen according to an instruction from the presentation support device 10 as an example. For example, the display device 5 displays a slide of a document file opened by presentation software that operates on the presentation support device 10. In this case, the display device 5 can display any slide specified by the presenter via the input device 7 among the slides included in the document file, and the slide show function of the presentation software is set to the ON state. In this case, the slides included in the document file can be switched and displayed in the order in which the slides are created.
 入力装置7は、各種の情報に対する指示入力を受け付ける装置である。例えば、表示装置5が液晶ディスプレイとして実装される場合、入力装置7には、マウスやキーボードを採用したり、液晶ディスプレイ上に貼り合わせられたタッチセンサを採用したりすることもできる。また、表示装置5がプロジェクタとして実装される場合、スクリーンに映し出された画面上の位置を指し示すレーザポインタを入力装置7として実装することもできる。すなわち、レーザポインタの中には、スライドのページを進めたり、戻したりする各種のボタンなどの操作部が設けられたリモコン機能付きのレーザポインタも存在する。このリモコン機能付きのレーザポインタが有する操作部を入力装置7として援用することもできる。さらには、レーザポインタによって指し示された光点の位置をセンシングする画像センサを入力装置7として実装することもできる。 The input device 7 is a device that receives instruction inputs for various types of information. For example, when the display device 5 is mounted as a liquid crystal display, a mouse or a keyboard or a touch sensor bonded on the liquid crystal display can be adopted as the input device 7. Further, when the display device 5 is mounted as a projector, a laser pointer indicating the position on the screen projected on the screen can be mounted as the input device 7. That is, among laser pointers, there is also a laser pointer with a remote control function provided with an operation unit such as various buttons for advancing and returning a slide page. The operation unit of the laser pointer with a remote control function can be used as the input device 7. Furthermore, an image sensor that senses the position of the light spot pointed by the laser pointer can be mounted as the input device 7.
 この入力装置7は、一例として、プレゼンテーション支援装置10上でプレゼンテーションソフトに実行させる文書ファイルの指定、スライドのページを進める操作やスライドのページを戻す操作などを受け付ける。このように入力装置7を介して受け付けられた操作は、プレゼンテーション支援装置10へ出力されることになる。 As an example, the input device 7 accepts a specification of a document file to be executed by the presentation software on the presentation support device 10, an operation of advancing a slide page, an operation of returning a slide page, and the like. The operation accepted through the input device 7 in this way is output to the presentation support device 10.
 プレゼンテーション支援装置10は、プレゼンテーションソフトが実行されるコンピュータである。 The presentation support apparatus 10 is a computer on which presentation software is executed.
 一実施形態として、プレゼンテーション支援装置10には、デスクトップ型またはノート型のパーソナルコンピュータなどの情報処理装置を採用することができる。この他、プレゼンテーション支援装置10には、上記のパーソナルコンピュータなどの据置き型の端末のみならず、各種の携帯端末装置を採用することもできる。例えば、携帯端末装置の一例として、スマートフォン、携帯電話機やPHS(Personal Handyphone System)などの移動体通信端末、さらには、PDA(Personal Digital Assistants)などのスレート端末などがその範疇に含まれる。 As an embodiment, the presentation support apparatus 10 may employ an information processing apparatus such as a desktop or notebook personal computer. In addition, the presentation support apparatus 10 can employ not only a stationary terminal such as the personal computer but also various portable terminal apparatuses. For example, as an example of the mobile terminal device, mobile communication terminals such as smartphones, mobile phones and PHS (Personal Handyphone System), and slate terminals such as PDA (Personal Digital Assistants) are included in the category.
 なお、本実施例では、あくまで一例として、プレゼンテーション支援装置10が上記のプレゼンテーションソフトを外部のリソースに依存せずに単独で実行するスタンドアローンで上記のプレゼンテーション支援サービスを提供する場合を想定する。詳細は後述するが、上記のプレゼンテーション支援サービスは、スタンドアローンで提供される実装に限定されない。例えば、プレゼンテーションソフトを実行するクライアントに対し、上記のプレゼンテーション支援サービスを提供するサーバを設けることによってクライアントサーバシステムとして構築することもできる。 In the present embodiment, as an example, it is assumed that the presentation support apparatus 10 provides the above-described presentation support service in a stand-alone manner that independently executes the above-described presentation software without depending on external resources. Although the details will be described later, the presentation support service is not limited to the implementation provided in a stand-alone manner. For example, a client server system can be constructed by providing a server that provides the presentation support service to a client that executes presentation software.
[プレゼンテーション支援装置10の構成]
 続いて、本実施例に係るプレゼンテーション支援装置10の機能的構成について説明する。図2は、実施例1に係るプレゼンテーション支援装置10の機能的構成を示すブロック図である。図2に示すように、プレゼンテーション支援装置10は、入出力I/F(InterFace)部11と、記憶部13と、制御部15とを有する。
[Configuration of Presentation Support Device 10]
Next, the functional configuration of the presentation support apparatus 10 according to the present embodiment will be described. FIG. 2 is a block diagram illustrating a functional configuration of the presentation support apparatus 10 according to the first embodiment. As shown in FIG. 2, the presentation support apparatus 10 includes an input / output I / F (InterFace) unit 11, a storage unit 13, and a control unit 15.
 入出力I/F部11は、マイク3、表示装置5及び入力装置7などの周辺機器との間で入出力を行うインタフェースである。 The input / output I / F unit 11 is an interface for performing input / output with peripheral devices such as the microphone 3, the display device 5, and the input device 7.
 一実施形態として、入出力I/F部11は、マイク3から入力された音声データを制御部15へ出力する。また、入出力I/F部11は、制御部15から出力されたスライドの画像データを表示装置5へ出力したり、制御部15から出力されたスライドに含まれる領域に対するハイライト指示またはそのキャンセル指示を表示装置5へ出力したりする。また、入出力I/F部11は、入力装置7から入力された各種の操作を制御部15へ出力する。 As one embodiment, the input / output I / F unit 11 outputs the audio data input from the microphone 3 to the control unit 15. Further, the input / output I / F unit 11 outputs the slide image data output from the control unit 15 to the display device 5, highlights an area included in the slide output from the control unit 15, or cancels the instruction. An instruction is output to the display device 5. The input / output I / F unit 11 outputs various operations input from the input device 7 to the control unit 15.
 記憶部13は、制御部15で実行されるOS(Operating System)やプレゼンテーションソフトを始め、アプリケーションプログラムなどの各種プログラムに用いられるデータを記憶するデバイスである。 The storage unit 13 is a device that stores data used for various programs such as an OS (Operating System) and presentation software executed by the control unit 15 and application programs.
 一実施形態として、記憶部13は、プレゼンテーション支援装置10における主記憶装置として実装される。例えば、記憶部13には、各種の半導体メモリ素子、例えばRAM(Random Access Memory)やフラッシュメモリを採用できる。また、記憶部13は、補助記憶装置として実装することもできる。この場合、HDD(Hard Disk Drive)、光ディスクやSSD(Solid State Drive)などを採用できる。 As an embodiment, the storage unit 13 is implemented as a main storage device in the presentation support apparatus 10. For example, various semiconductor memory elements such as RAM (Random Access Memory) and flash memory can be employed for the storage unit 13. The storage unit 13 can also be implemented as an auxiliary storage device. In this case, HDD (Hard Disk Drive), optical disk, SSD (Solid State Drive), etc. can be adopted.
 記憶部13は、制御部15で実行されるプログラムに用いられるデータの一例として、文書データ13a、抽出単語データ13b及び認識単語データ13cを記憶する。なお、上記の文書データ13a以外の抽出単語データ13b及び認識単語データ13cは、制御部15による処理を経て生成される中間データであるので、制御部15の説明で併せて説明することとする。また、記憶部13には、上記のデータ以外にも、他の電子データ、例えばプレゼンテーションの時間割なども併せて記憶することもできるのは言うまでもない。 The storage unit 13 stores document data 13a, extracted word data 13b, and recognized word data 13c as an example of data used in a program executed by the control unit 15. Note that the extracted word data 13b and the recognized word data 13c other than the document data 13a are intermediate data generated through processing by the control unit 15, and will be described together with the description of the control unit 15. In addition to the above data, the storage unit 13 can also store other electronic data such as a presentation timetable.
 文書データ13aは、文書に関するデータである。 Document data 13a is data relating to a document.
 一実施形態として、文書データ13aには、プレゼンテーションソフトを用いて1または複数のスライドが作成された文書ファイルを採用できる。かかるスライドには、テキストや図形を始め、他のアプリケーションプログラムによって作成されたコンテンツをインポートすることができる。例えば、ワープロソフトで作成された文書、表計算ソフトで作成された表やグラフをインポートしたり、撮像デバイスで撮像された画像や動画、さらには、画像編集ソフトで編集された画像や動画などをインポートしたりすることができる。このように、テキスト以外のコンテンツには、音声認識によるキーワード検索を実現するために、プレゼンテーションの開始前までに当該コンテンツの説明語句や説明文などの文字列を含むメタ情報を付与しておくことができる。 As an embodiment, a document file in which one or a plurality of slides are created using presentation software can be adopted as the document data 13a. Such slides can be imported with text and graphics as well as content created by other application programs. For example, you can import documents created with word processing software, tables and graphs created with spreadsheet software, images and movies taken with an imaging device, and images and movies edited with image editing software. And can be imported. In this way, in order to realize a keyword search by voice recognition, meta information including a character string such as an explanatory word or a description of the content is added to the content other than the text before the presentation is started. Can do.
 制御部15は、各種のプログラムや制御データを格納する内部メモリを有し、これらによって種々の処理を実行するものである。 The control unit 15 has an internal memory for storing various programs and control data, and executes various processes using these.
 一実施形態として、制御部15は、中央処理装置、いわゆるCPU(Central Processing Unit)として実装される。なお、制御部15は、必ずしも中央処理装置として実装されずともよく、MPU(Micro Processing Unit)として実装されることとしてもよい。また、制御部15は、ASIC(Application Specific Integrated Circuit)やFPGA(Field Programmable Gate Array)などのハードワイヤードロジックによっても実現できる。 As an embodiment, the control unit 15 is implemented as a central processing unit, a so-called CPU (Central Processing Unit). Note that the control unit 15 does not necessarily have to be implemented as a central processing unit, and may be implemented as an MPU (Micro Processing Unit). Further, the control unit 15 can be realized by hard wired logic such as ASIC (Application Specific Integrated Circuit) or FPGA (Field Programmable Gate Array).
 制御部15は、各種のプログラムを実行することによって下記の処理部を仮想的に実現する。例えば、制御部15は、図2に示すように、分割部15aと、抽出部15bと、付与部15cと、認識部15dと、算出部15eと、設定部15fと、表示制御部15gとを有する。 The control unit 15 virtually implements the following processing unit by executing various programs. For example, as shown in FIG. 2, the control unit 15 includes a dividing unit 15a, an extracting unit 15b, an adding unit 15c, a recognizing unit 15d, a calculating unit 15e, a setting unit 15f, and a display control unit 15g. Have.
 分割部15aは、スライドを複数の領域に分割する処理部である。 The dividing unit 15a is a processing unit that divides a slide into a plurality of regions.
 一実施形態として、分割部15aは、記憶部13に記憶された文書データ13aが含む文書ファイルのうち先に指定を受け付けた文書ファイルを読み出す。ここでは、一例として、分割部15aが記憶部13から文書ファイルを読み出す場合を例示したが、文書ファイルの入手経路はこれに限定されない。例えば、分割部15aは、ハードディスクや光ディスクなどの補助記憶装置またはメモリカードやUSB(Universal Serial Bus)メモリなどのリムーバブルメディアから画像を取得することもできる。また、取得部15aは、外部装置からネットワークを介して受信することによって画像を取得することもできる。 As an embodiment, the dividing unit 15a reads a document file that has received a designation first among document files included in the document data 13a stored in the storage unit 13. Here, as an example, the case where the dividing unit 15a reads the document file from the storage unit 13 is illustrated, but the document file acquisition path is not limited thereto. For example, the dividing unit 15a can also acquire an image from an auxiliary storage device such as a hard disk or an optical disk, or a removable medium such as a memory card or a USB (Universal Serial Bus) memory. The acquisition unit 15a can also acquire an image by receiving it from an external device via a network.
 続いて、分割部15aは、先に読み出した文書ファイルに含まれるスライドを複数の領域へ分割する。例えば、分割部15aは、一文、行、段落などの単位でスライドを分割する。この場合、分割部15aは、スライドが含む文字列を走査して、スペース、句点または改行に対応する区切り文字を検出し、当該区切り文字を領域の境界に設定する。かかる境界を前後に、分割部15aは、スライドが含む文字列を区切る。これによって、スライドが複数の領域へ区切り文字ごとに分割される。その上で、分割部15aは、スライドの分割によって得られた領域に当該領域を識別するインデックスを割り当てる。なお、ここでは、スライドを自動的に分割する場合を例示したが、入力装置7等を介して領域の境界を指定させることによってスライドを手動設定で分割することとしてもよい。 Subsequently, the dividing unit 15a divides the slide included in the previously read document file into a plurality of areas. For example, the dividing unit 15a divides the slide in units of one sentence, line, paragraph, and the like. In this case, the dividing unit 15a scans a character string included in the slide, detects a delimiter character corresponding to a space, a punctuation point, or a line feed, and sets the delimiter character as a boundary of the region. The dividing unit 15a divides a character string included in the slide before and after the boundary. Thus, the slide is divided into a plurality of areas for each delimiter character. In addition, the dividing unit 15a assigns an index for identifying the area to the area obtained by dividing the slide. Here, the case where the slide is automatically divided is illustrated, but the slide may be manually divided by designating the boundary of the area via the input device 7 or the like.
 抽出部15bは、領域が含む文字列から単語を抽出する処理部である。 The extraction unit 15b is a processing unit that extracts words from the character string included in the region.
 一実施形態として、抽出部15bは、スライドの分割後に、複数の領域のうち領域を1つ選択する。続いて、抽出部15bは、先に選択された領域が含む文字列に対し、自然言語処理を実行することによって単語を抽出する。例えば、抽出部15bは、領域内の文字列に形態素解析等を実行することにより得られた形態素のうち品詞が名詞である単語を抽出する。そして、抽出部15bは、先に抽出された各単語に当該単語が含まれる領域に割り当てられたインデックスを付与する。その後、抽出部15bは、スライドが含む領域が全て選択されるまで上記の単語の抽出及び上記のインデックスの付与を繰返し実行する。なお、ここでは、スライドが含む領域を1つずつ順番に処理する場合を例示したが、各領域を並列して処理することができるのは言うまでもない。 As one embodiment, the extraction unit 15b selects one of a plurality of areas after dividing the slide. Subsequently, the extraction unit 15b extracts a word by executing natural language processing on the character string included in the previously selected region. For example, the extraction unit 15b extracts a word whose part of speech is a noun from morphemes obtained by executing morphological analysis or the like on a character string in the region. And the extraction part 15b provides the index allocated to the area | region where the said word is included to each word extracted previously. Thereafter, the extraction unit 15b repeatedly executes the extraction of the word and the assignment of the index until all the regions included in the slide are selected. In addition, although the case where the area | region which a slide contains was processed one by one was illustrated here, it cannot be overemphasized that each area | region can be processed in parallel.
 付与部15cは、各単語に重みを付与する処理部である。 The assigning unit 15c is a processing unit that assigns a weight to each word.
 一実施形態として、付与部15cは、抽出部15bにより全ての領域から単語が抽出された後に、スライドに含まれる単語ごとに当該単語kの出現頻度fを算出する。かかる出現頻度の一例として、付与部15cは、単語kが同一のスライドに出現する回数を集計することによって単語別の総出現回数を算出する。そして、付与部15cは、先に単語別に算出された出現頻度fに対応する単語の重みwを付与する。この場合、付与部15cは、出現頻度fが高くなるにしたがって重みwが小さい値となる重みの算出式を用いる。例えば、付与部15cは、重みの算出式「w=1/f 」に出現頻度fを代入することによって計算された重みwを単語kに付与する。その上で、付与部15cは、単語k、インデックスidx及び重みwが対応付けられた抽出単語データ13bを記憶部13へ登録する。 As an embodiment, the assigning unit 15 c calculates the appearance frequency f k of the word k for each word included in the slide after the extraction unit 15 b extracts the words from all the regions. As an example of such appearance frequency, the assigning unit 15c calculates the total number of appearances for each word by counting the number of times the word k appears on the same slide. Then, applying section 15c imparts the weights w k of the word corresponding to the frequency f k, which is calculated for each word before. In this case, the assigning unit 15 c uses a weight calculation formula in which the weight w k becomes smaller as the appearance frequency f k becomes higher. For example, the assigning unit 15 c assigns the weight w k calculated by substituting the appearance frequency f k to the weight calculation formula “w k = 1 / f k 2 ” to the word k. Then, the assigning unit 15 c registers the extracted word data 13 b in which the word k, the index idx, and the weight w k are associated with the storage unit 13.
 図3は、抽出単語データ13bの一例を示す図である。図3には、複数のスライドのうち1つのスライドに関する抽出単語データが抜粋して示されている。図3に示す抽出単語データ13bの例で言えば、単語“a”が領域「idx1」及び領域「idx2」の2つの領域に出現することを意味する。さらに、単語“a”は、出現頻度が「2」であるので、1/2の計算により0.25が重みとして付与されている。また、単語“b”が領域「idx1」及び領域「idx3」の2つの領域に出現することを意味する。さらに、単語“b”も出現頻度が「2」であるので、0.25が重みとして付与されている。また、単語“c”が領域「idx1」及び領域「idx2」の2つの領域に出現することを意味する。さらに、単語“c”も出現頻度が「2」であるので、0.25が重みとして付与されている。また、単語“d”が領域「idx2」及び領域「idx3」の2つの領域に出現することを意味する。さらに、単語“d”も出現頻度が「2」であるので、0.25が重みとして付与されている。最後に、単語“e”が領域「idx3」の1つの領域に出現することを意味する。そして、単語“e”は、出現頻度が「1」であるので、1/1の計算により1が重みとして付与されている。なお、図3には、1つのスライドに関する抽出単語データを例示したが、他のスライドについても各項目の値は異なれども図3の例と同様にコンピュータが単語の領域や重みを識別できる状態で抽出単語データが記憶される。 FIG. 3 is a diagram illustrating an example of the extracted word data 13b. FIG. 3 shows extracted word data relating to one slide out of a plurality of slides. In the example of the extracted word data 13b shown in FIG. 3, this means that the word “a” appears in two areas, the area “idx1” and the area “idx2”. Further, the word "a", since frequency is "2", 0.25 is assigned as a weight to compute the 1/2 2. Further, it means that the word “b” appears in two areas of the area “idx1” and the area “idx3”. Furthermore, since the appearance frequency of the word “b” is “2”, 0.25 is given as a weight. Further, it means that the word “c” appears in two areas of the area “idx1” and the area “idx2”. Furthermore, since the appearance frequency of the word “c” is “2”, 0.25 is given as a weight. Further, it means that the word “d” appears in two areas of the area “idx2” and the area “idx3”. Furthermore, since the appearance frequency of the word “d” is “2”, 0.25 is given as a weight. Finally, it means that the word “e” appears in one area of the area “idx3”. Then, the word "e", since the frequency of occurrence is "1", 1 is given as a weight to the calculation of the 1/1 2. FIG. 3 illustrates the extracted word data related to one slide. However, in the other slides, the value of each item is different, but the computer can identify the word region and weight as in the example of FIG. Extracted word data is stored.
 なお、単語の重みfの算出方法に関する応用例の詳細については後述するが、単語の重みfは上記の例に限定されない。すなわち、付与部15cは、上記の総出現回数以外の他の因子を用いて単語の重みfを算出することもできるし、あるいは上記の総出現回数に他の因子を加えて単語の重みfを算出することができる。 Note that details of an application example regarding the method of calculating the word weight fk will be described later, but the word weight fk is not limited to the above example. That is, the assigning unit 15c can calculate the word weight fk using a factor other than the above total number of appearances, or add another factor to the above total number of appearances and add the word weight fk. k can be calculated.
 認識部15dは、音声認識を実行する処理部である。 The recognition unit 15d is a processing unit that performs voice recognition.
 一実施形態として、認識部15dは、プレゼンテーションソフトが文書ファイルを開いた状態でプレゼンテーションの開始指示を受け付けた場合に起動し、マイク3から所定時間長の音声信号が入力されるまで待機する。例えば、少なくとも1フレーム分の時間長、例えば10msecの音声信号が入力されるのを待機する。そして、認識部15dは、マイク3から所定時間長の音声信号が入力される度に、当該音声信号にワードスポッティングなどの音声認識を実行する。このとき、認識部15dは、記憶部13に記憶された抽出単語データ13bのうちプレゼンテーションソフトが実行中である文書ファイルが含むスライドであり、かつ表示装置5に表示中であるスライドに関する抽出単語データをワードスポッティングに適用する。これによって、認識部15dは、プレゼンタの発話の中に表示中のスライドに含まれる各領域から抽出された単語が存在するか否かを認識する。そして、認識部15dは、音声信号から単語が認識された場合、当該単語及びその単語が認識された時間が対応付けられた認識単語データ13cを記憶部13へ登録する。なお、同一の単語が時間経過に伴って複数回にわたって認識される場合には、最後、すなわち最新に認識された時刻が記憶部13へ登録される。 As an embodiment, the recognition unit 15d is activated when the presentation software receives a presentation start instruction with the document file opened, and waits until an audio signal having a predetermined time length is input from the microphone 3. For example, it waits for an audio signal having a time length of at least one frame, for example, 10 msec. The recognizing unit 15d performs voice recognition such as word spotting on the voice signal every time a voice signal having a predetermined time length is input from the microphone 3. At this time, the recognizing unit 15d extracts the extracted word data related to the slide that is included in the document file that is being executed by the presentation software among the extracted word data 13b stored in the storage unit 13 and that is being displayed on the display device 5. Apply to word spotting. Thereby, the recognition unit 15d recognizes whether or not a word extracted from each region included in the slide being displayed exists in the utterance of the presenter. Then, when a word is recognized from the audio signal, the recognition unit 15 d registers the recognition word data 13 c in which the word and the time when the word is recognized are associated with each other in the storage unit 13. When the same word is recognized a plurality of times as time passes, the last, that is, the latest recognized time is registered in the storage unit 13.
 その後、認識部15dは、記憶部13に記憶された認識単語データ13cのうち記憶部13へ登録されてから所定の期間が経過した単語が存在するか否かを判定する。例えば、認識部15dは、認識単語データ13cに含まれる単語ごとに、当該単語に対応付けて登録された時間と、認識部15dが認識単語データ13cを参照する時間、すなわち現時間との差が所定の閾値を超過するか否かを判定する。このとき、認識部15dは、分割部15aによってスライドが分割された単位、例えば一文、行や段落などによって上記の判定に用いる閾値を変えることができる。例えば、スライドが行単位で分割される場合、1つの領域で読み上げられる文字数はおよそ20~30文字であると想定できる。この場合、上記の閾値の一例として、5~10秒を用いることができる。また、スライドが段落単位で分割される場合、行単位よりも長い時間が読み上げに割かれると想定できる。この場合、上記の閾値の一例として、20~30秒を用いることができる。 Thereafter, the recognizing unit 15d determines whether or not there is a word for which a predetermined period has elapsed since it was registered in the storage unit 13 in the recognized word data 13c stored in the storage unit 13. For example, for each word included in the recognized word data 13c, the recognizing unit 15d has a difference between the time registered in association with the word and the time when the recognizing unit 15d refers to the recognized word data 13c, that is, the current time. It is determined whether or not a predetermined threshold is exceeded. At this time, the recognizing unit 15d can change the threshold value used for the above determination according to the unit in which the slide is divided by the dividing unit 15a, for example, one sentence, line, paragraph, or the like. For example, when the slide is divided in units of lines, it can be assumed that the number of characters read out in one area is approximately 20 to 30 characters. In this case, 5 to 10 seconds can be used as an example of the threshold value. Further, when the slide is divided in units of paragraphs, it can be assumed that a time longer than the line unit is devoted to reading. In this case, 20 to 30 seconds can be used as an example of the above threshold.
 ここで、記憶部13へ登録されてから所定の期間が経過した単語が存在する場合、当該単語を含むスライドの領域に関する説明が終了している可能性が高まる。このような単語を残しておくと、説明が終了している領域がハイライトで表示される可能性も高まる。よって、認識部15dは、記憶部13に記憶された認識単語データ13cから当該単語に関するレコードを削除する。一方、記憶部13へ登録されてから所定の期間が経過した単語が存在しない場合、認識単語データ13cに含まれる単語が出現するスライドの領域に関する説明が終了していない可能性が高まる。この場合、説明が終了している領域がハイライトで表示される可能性は低い。よって、認識部15dは、記憶部13に記憶された認識単語データ13cに含まれる単語を削除せずにそのまま残す。 Here, when there is a word for which a predetermined period has elapsed since it was registered in the storage unit 13, there is a high possibility that the explanation regarding the slide area including the word has ended. If such a word is left, the possibility that the area where the explanation has ended will be highlighted. Therefore, the recognition unit 15d deletes the record related to the word from the recognized word data 13c stored in the storage unit 13. On the other hand, when there is no word for which a predetermined period has elapsed since registration in the storage unit 13, there is a high possibility that the description regarding the slide area where the word included in the recognized word data 13c appears has not ended. In this case, there is a low possibility that the area for which the explanation has been completed is displayed with highlight. Therefore, the recognition unit 15d leaves the word included in the recognition word data 13c stored in the storage unit 13 without deleting it.
 また、認識部15dは、表示装置5に表示されるスライドのページが変更されたか否かを判定する。例えば、認識部15dは、スライドショーによりスライドが切り替えられたり、入力装置7を介してスライドのページを進める操作またはスライドのページを戻す操作を受け付けたりしたかを判定する。このとき、表示装置5に表示されるスライドのページが変更された場合、プレゼンタの説明も変更前のページのスライドから変更後のページのスライドへ切り替わった可能性が高い。この場合、認識部15dは、記憶部13に記憶された認識単語データ13cを削除する。一方、表示装置5に表示されるスライドのページが変更されていない場合、プレゼンタが説明するページにも変りがない可能性が高い。この場合、認識部15dは、記憶部13に記憶された認識単語データ13cに含まれる単語を削除せずにそのまま残す。 Further, the recognition unit 15d determines whether or not the slide page displayed on the display device 5 has been changed. For example, the recognizing unit 15d determines whether a slide is switched by a slide show or an operation for advancing a slide page or an operation for returning a slide page is received via the input device 7. At this time, when the slide page displayed on the display device 5 is changed, it is highly possible that the description of the presenter is switched from the slide of the page before the change to the slide of the page after the change. In this case, the recognition unit 15d deletes the recognized word data 13c stored in the storage unit 13. On the other hand, when the slide page displayed on the display device 5 is not changed, there is a high possibility that the page explained by the presenter will not change. In this case, the recognition unit 15d leaves the word included in the recognized word data 13c stored in the storage unit 13 without deleting it.
 これら一連の動作により、認識部15dは、表示中であるスライドの中でプレゼンタが説明中である可能性が高い単語を認識する。以下では、抽出単語データ13bに含まれる単語のことを「抽出単語」と記載すると共に、認識単語データ13cに含まれる単語のことを「認識単語」と記載し、互いのラベルを区別する場合がある。 Through these series of operations, the recognizing unit 15d recognizes a word that is highly likely to be explained by the presenter in the displayed slide. In the following, a word included in the extracted word data 13b is described as an “extracted word”, and a word included in the recognized word data 13c is described as a “recognized word” to distinguish the labels from each other. is there.
 算出部15eは、表示中であるスライド内の領域と、音声認識結果として得られた単語との関連度を算出する処理部である。 The calculation unit 15e is a processing unit that calculates the degree of association between the region in the slide being displayed and the word obtained as a speech recognition result.
 一実施形態として、算出部15eは、表示装置5に表示中であるスライドが含む領域のインデックスのうちインデックスを1つ選択する。続いて、算出部15eは、先に選択されたインデックスの領域に対応付けられた抽出単語データ13bの抽出単語のうち認識単語データ13cの認識単語と一致する抽出単語に付与された重みから当該領域の関連度を算出する。例えば、上記の単語の重みwを用いて領域xの関連度rを算出する場合、算出部15eは、認識単語と一致する抽出単語に付与された重みwを合計することによって関連度rを算出できる。このとき、インデックスの領域に対応付けられた抽出単語の中に認識単語と一致する単語が存在しない場合、当該領域の関連度はゼロと算出されることになる。このような算出ロジックによって、スライド内の各領域の記述内容がプレゼンタの発話内容との間で関連している度合いを上記の「関連度」として求める。 As one embodiment, the calculation unit 15e selects one index among the indexes of the area included in the slide that is being displayed on the display device 5. Subsequently, the calculation unit 15e calculates the area from the weight assigned to the extracted word that matches the recognized word of the recognized word data 13c among the extracted words of the extracted word data 13b associated with the previously selected index area. The relevance of is calculated. For example, when calculating the degree of association r x of the region x using the above-described word weight w k , the calculation unit 15e adds the weights w k given to the extracted words that match the recognized word, thereby calculating the degree of association. r x can be calculated. At this time, if there is no word that matches the recognized word among the extracted words associated with the index area, the relevance of the area is calculated as zero. By such calculation logic, the degree of relevance between the description contents of each area in the slide and the utterance contents of the presenter is obtained as the “relevance degree”.
 設定部15fは、スライド内の領域のハイライト表示を進行させる速度を設定する処理部である。以下では、ハイライト表示を進行させる速度のことを「ハイライト速度」と記載する場合がある。 The setting unit 15f is a processing unit that sets the speed at which highlight display of the area in the slide is advanced. Hereinafter, the speed at which highlight display proceeds is sometimes referred to as “highlight speed”.
 一実施形態として、設定部15fは、算出部15eにより関連度が算出される度に、関連度が高い領域ほどハイライト速度を高く設定し、あるいは関連度が低い領域ほどハイライト速度を低く設定する。例えば、上記の関連度rを用いて領域xのハイライト速度vを設定する場合、設定部15fは、ハイライト速度vの算出式「v=V×r」に上記の関連度rを代入することによって算出することができる。この算出式に含まれる「V」は、予め定められた固定値である。つまり、上記のハイライト速度vの算出式を用いれば、関連度rの値と比例するハイライト速度vを算出することが可能となる。 As an embodiment, the setting unit 15f sets a higher highlight speed for a region with a higher relevance or a lower highlight speed for a region with a lower relevance each time the relevance is calculated by the calculator 15e. To do. For example, to set the highlight velocity v x of the region x with the degree of association r x, setting unit 15f, the above related to the calculation formula of the highlight velocity v x 'v x = V × r x " It can be calculated by substituting the degree r x . “V” included in the calculation formula is a predetermined fixed value. That is, by using the calculation formula of the highlights velocity v x, it is possible to calculate a highlight velocity v x proportional to the value of relevance r x.
 表示制御部15gは、表示装置5に対する表示制御を実行する処理部である。 The display control unit 15g is a processing unit that executes display control for the display device 5.
 一実施形態として、表示制御部15gは、プレゼンテーションソフトにより文書ファイルが開かれた場合、当該文書ファイルが含むスライドを表示装置5に表示させる。このとき、表示制御部15gは、文書ファイルが含むスライドのうち最初のページのスライドを表示させることとしてもよいし、最後に編集が行われたページのスライドを表示させることとしてもよい。 As one embodiment, when a document file is opened by presentation software, the display control unit 15g causes the display device 5 to display a slide included in the document file. At this time, the display control unit 15g may display the slide of the first page among the slides included in the document file, or may display the slide of the page edited last.
 その後、表示制御部15gは、プレゼンテーションの開始指示を受け付けた場合、設定部15fにより各領域のハイライト速度が設定される度に、次のような処理を実行する。すなわち、表示制御部15gは、表示中のスライドが含む領域ごとに当該領域に設定されたハイライト速度にしたがってハイライト表示を進行させる。すなわち、表示制御部15gは、領域のハイライト速度にゼロよりも大きな値が設定されたからといってハイライト表示を直ちに完了させるとは限らない。つまり、表示制御部15gは、設定部15fにより設定されたハイライト速度でハイライト表示を完了へ向けて進行させる。これによって、ゼロよりも大きいハイライト速度が設定された領域に対し、スライドの作成時に設定された表示形態とは異なる表示形態へ向けてハイライト表示が進行される。以下では、領域のハイライト表示が完了へ向けて進行している度合いのことを「進行度」と記載する場合がある。 After that, when receiving a presentation start instruction, the display control unit 15g executes the following process every time the setting unit 15f sets the highlight speed of each area. That is, the display control unit 15g advances the highlight display according to the highlight speed set for each area included in the slide being displayed. That is, the display control unit 15g does not always complete the highlight display immediately because a value greater than zero is set as the highlight speed of the region. That is, the display control unit 15g advances the highlight display toward the completion at the highlight speed set by the setting unit 15f. As a result, the highlight display is advanced toward a display form different from the display form set at the time of creating the slide in the region where the highlight speed greater than zero is set. Hereinafter, the degree of progress of highlight display of an area toward completion may be referred to as “progress”.
 ここで、表示制御部15gは、任意のハイライト表示を実行することができる。例えば、表示制御部15gは、スライドの作成時に領域へ設定された輝度よりも領域が含む文字列または文字列の背景の輝度を上昇させることによって強調表示を実現することができる。また、表示制御部15gは、文字列のフォントを変えたり、背景の表示色や塗りつぶしを変えることとしてもかまわない。この他、表示制御部15gは、領域を反転表示させることによって強調表示を実現することもできる。 Here, the display control unit 15g can execute arbitrary highlight display. For example, the display control unit 15g can realize the highlight display by increasing the luminance of the character string included in the region or the background of the character string, rather than the luminance set to the region at the time of creating the slide. Further, the display control unit 15g may change the font of the character string, or change the background display color or fill. In addition, the display control unit 15g can also realize highlighting by highlighting the area.
 また、表示制御部15gは、ハイライト表示の進行度が所定の閾値以上である領域の有無を監視する。そして、ハイライト表示の進行度が所定の閾値以上である領域が存在する場合、当該領域は、ハイライト表示の進行度が閾値未満である領域よりもハイライト速度、言い換えれば関連度の総和の平均が高い状態で維持されたと判断できる。この場合、表示制御部15gは、ハイライト表示の進行度が閾値以上である領域のハイライト速度の設定は維持し、ハイライト表示の進行度が閾値未満である領域のハイライト表示を元の状態に戻すと共にハイライト表示の進行度が閾値未満である領域のハイライト速度をゼロにリセットする。これによって、時間経過もしくは認識単語と一致する抽出単語の時間経過による積み重なりから、プレゼンタの説明箇所であると判断できる領域に絞ってハイライト表示を実行させる。 Further, the display control unit 15g monitors the presence / absence of a region where the progress of highlight display is equal to or greater than a predetermined threshold. When there is an area where the progress of highlight display is equal to or greater than a predetermined threshold, the area is higher than the area where the progress of highlight display is less than the threshold, in other words, the sum of relevance. It can be judged that the average was maintained at a high level. In this case, the display control unit 15g maintains the setting of the highlight speed of the area where the progress of highlight display is equal to or higher than the threshold, and the highlight display of the area where the progress of highlight display is less than the threshold is changed to the original. Returning to the state, the highlight speed of the area where the progress of highlight display is less than the threshold is reset to zero. As a result, highlight display is executed by narrowing down to a region where it can be determined that the presenter is an explanation location based on the passage of time or the accumulation of extracted words that match the recognized word over time.
 その後、表示制御部15gは、ハイライト表示の進行度が閾値以上である領域で算出部15eにより算出される関連度が低下するか否かを判定する。例えば、今回に算出されるカレントの関連度が前回に算出されたパストの関連度よりも低い場合、時間経過によって認識単語と一致する抽出単語の数が減ったり、あるいは重みの低い抽出単語しか認識単語と一致しなかったりという状況変化があると判断できる。この場合、表示制御部15gは、関連度が低下した領域のハイライト表示を元の状態に戻すと共にハイライト速度をゼロにリセットする。なお、ここでは、カレントの関連度がパストの関連度よりも低い場合にハイライト表示を取り消す場合を例示したが、カレントの関連度がパストの関連度よりも一定値にわたって低い場合にハイライト表示を取り消すこととしてもかまわない。 After that, the display control unit 15g determines whether or not the degree of association calculated by the calculation unit 15e decreases in an area where the progress degree of highlight display is equal to or greater than the threshold value. For example, if the current relevance level calculated this time is lower than the past relevance level calculated last time, the number of extracted words that match the recognized word decreases over time, or only extracted words with low weight are recognized. It can be determined that there is a change in situation such as not matching the word. In this case, the display control unit 15g returns the highlight display of the area where the degree of relevance is reduced to the original state, and resets the highlight speed to zero. In this example, the highlight display is canceled when the current relevance level is lower than the past relevance level. However, the highlight display is performed when the current relevance level is lower than the past relevance level by a certain value. You can also cancel.
 また、表示制御部15gは、入力装置7を介してページの切替え指示を受け付けた場合、表示装置5に表示させるスライドを変更する。例えば、ページを進める操作を受け付けた場合、表示制御部15gは、表示中のスライドの次ページのスライドを表示装置5に表示させる。また、ページを戻る操作を受け付けた場合、表示制御部15gは、表示中のスライドの前ページのスライドを表示装置5に表示させる。 Further, when the display control unit 15g receives a page switching instruction via the input device 7, the display control unit 15g changes the slide to be displayed on the display device 5. For example, when an operation for advancing a page is received, the display control unit 15g causes the display device 5 to display a slide on the next page of the slide being displayed. Further, when the operation of returning the page is received, the display control unit 15g causes the display device 5 to display the slide of the previous page of the slide being displayed.
[具体例]
 次に、図4~図6を用いて、プレゼンテーション支援方法の具体例について説明する。図4は、ハイライト表示の進行度に関する時間変化の一例を示す図である。図5及び図6は、スライド画面の遷移例を示す図である。これら図4には、図3に示した抽出単語データ13bを用いて、表示装置5に表示中であるスライドが含む各領域の関連度が算出される場合が例示されている。また、図5及び図6には、図3に示した抽出単語データ13bの例にしたがって表示装置5に表示中であるスライドがインデックスidx1~idx3の3つの領域を含む場合が例示されている。なお、ここでは、ハイライト表示の一例として、領域に設定されたハイライト速度にしたがって各領域の反転表示が実行される場合を例示する。
[Concrete example]
Next, a specific example of the presentation support method will be described with reference to FIGS. FIG. 4 is a diagram illustrating an example of a temporal change related to the progress of highlight display. 5 and 6 are diagrams showing an example of transition of the slide screen. In these FIG. 4, the case where the relevance degree of each area | region which the slide currently displayed on the display apparatus 5 includes using the extracted word data 13b shown in FIG. 3 is illustrated. 5 and 6 exemplify a case in which the slide being displayed on the display device 5 includes three areas of the indices idx1 to idx3 according to the example of the extracted word data 13b illustrated in FIG. Here, as an example of the highlight display, a case where reverse display of each area is executed according to the highlight speed set for the area is illustrated.
 図4に示すように、時刻t1の時点までは、いずれの単語も認識されていないので、インデックスidx1~idx3のいずれの領域でもハイライト表示が実行されていない。すなわち、図5の最上段に示すように、スライドの作成時に設定された表示形態のままでインデックスidx1~idx3の領域が表示される。 As shown in FIG. 4, since no word has been recognized until time t1, highlight display has not been executed in any of the areas of the index idx1 to idx3. That is, as shown in the uppermost part of FIG. 5, the areas of the indices idx1 to idx3 are displayed with the display form set when the slide is created.
 ここで、時刻t1の時点で単語“a”が認識されたとしたとき、各領域の関連度は、次のように算出される。すなわち、図3に示したように、インデックスidx1及びインデックスidx2の2つの領域には、抽出単語に重み「0.25」が設定された認識単語“a”が含まれるので、関連度が「0.25」と算出される。一方、インデックスidx3の領域には、いずれの認識単語も含まれないので、関連度が「0」と算出される。この結果、インデックスidx1及びインデックスidx2の2つの領域には、関連度「0.25」に比例するハイライト速度が設定されると共に、インデックスidx3の領域には、ハイライト速度がゼロに設定される。 Here, assuming that the word “a” is recognized at time t1, the relevance of each region is calculated as follows. That is, as shown in FIG. 3, since the extracted word includes the recognized word “a” in which the weight “0.25” is set in the two areas of the index idx1 and the index idx2, the degree of association is “0”. .25 ". On the other hand, since no recognized word is included in the area of index idx3, the degree of association is calculated as “0”. As a result, a highlight speed proportional to the relevance “0.25” is set in the two areas of index idx1 and index idx2, and the highlight speed is set to zero in the area of index idx3. .
 そして、時刻t2の時点では、図5の上から2番目に示すスライド画面となる。すなわち、インデックスidx1及びインデックスidx2の2つの領域は、ハイライト表示が同程度に進行する一方で、インデックスidx3の領域は、ハイライト表示が進行していない状態に遷移する。 At the time t2, the slide screen shown second from the top in FIG. 5 is displayed. That is, in the two areas of the index idx1 and the index idx2, the highlight display progresses to the same extent, while the area of the index idx3 transitions to a state where the highlight display does not progress.
 この時刻t2の時点で単語“b”が認識されたとする。この場合、インデックスidx1の領域には、抽出単語に重み「0.25」が設定された認識単語“a”と重み「0.25」が設定された認識単語“b”との2つの認識単語が含まれるので、関連度が「0.5」と算出される。また、インデックスidx2の領域には、時刻t1の時点と同様に、抽出単語に重み「0.25」が設定された認識単語“a”が含まれるので、関連度が「0.25」と算出される。一方、インデックスidx3の領域には、抽出単語に重み「0.25」が設定された認識単語“b”が含まれるので、関連度が「0.25」と算出される。この結果、インデックスidx1には、関連度「0.5」に比例するハイライト速度が設定され、インデックスidx2の領域には、関連度「0.25」に比例するハイライト速度が設定され、インデックスidx3の領域には、関連度「0.25」に比例するハイライト速度が設定される。つまり、時刻t2の時点で、各領域のハイライト速度は、「idx1>idx2=idx3」となる。 Suppose that the word “b” is recognized at the time t2. In this case, in the area of the index idx1, two recognition words of a recognition word “a” with a weight “0.25” set as an extracted word and a recognition word “b” with a weight “0.25” set Therefore, the relevance is calculated as “0.5”. In addition, since the extracted word includes the recognized word “a” with the weight “0.25” set in the area of the index idx2, the degree of association is calculated as “0.25”. Is done. On the other hand, since the recognized word “b” with the weight “0.25” set in the extracted word is included in the area of the index idx3, the degree of association is calculated as “0.25”. As a result, a highlight speed proportional to the degree of association “0.5” is set in the index idx1, and a highlight speed proportional to the degree of association “0.25” is set in the area of the index idx2. In the area of idx3, a highlight speed proportional to the relevance “0.25” is set. That is, at time t2, the highlight speed of each region is “idx1> idx2 = idx3”.
 その後、時刻t3の時点では、図5の上から3番目に示すスライド画面へ遷移する。すなわち、インデックスidx1及びインデックスidx2の2つの領域の間で、ハイライト表示の進行度に差が生じ始める。つまり、インデックスidx1の領域は、インデックスidx2の領域に比べて大きくハイライト表示が進行する。さらに、インデックスidx3の領域は、インデックスidx2よりも遅れてハイライト速度が設定された分、インデックスidx2のハイライト表示の進行度との間に遅れがある。 Then, at time t3, the screen transitions to the third slide screen from the top in FIG. That is, a difference starts in the progress of highlight display between the two areas of index idx1 and index idx2. In other words, the area with the index idx1 is highlighted more greatly than the area with the index idx2. Further, the area of the index idx3 is delayed from the progress of the highlight display of the index idx2 by the amount that the highlight speed is set later than the index idx2.
 この時刻t3の時点で単語“c”が認識されたとする。この場合、インデックスidx1の領域には、重み「0.25」が設定された認識単語“a”、重み「0.25」が設定された認識単語“b”、及び、重み「0.25」が設定された認識単語“c”の3つが抽出単語に含まれる。このため、関連度は、「0.25+0.25+0.25」の計算により、「0.75」と算出される。また、インデックスidx2の領域には、抽出単語に重み「0.25」が設定された認識単語“a”と抽出単語に重み「0.25」が設定された認識単語“c”とが含まれるので、関連度が「0.5」と算出される。一方、インデックスidx3の領域には、時刻t2の場合と同様に、抽出単語に重み「0.25」が設定された認識単語“b”が含まれるので、関連度が「0.25」と算出される。この結果、インデックスidx1には、関連度「0.75」に比例するハイライト速度が設定され、インデックスidx2の領域には、関連度「0.5」に比例するハイライト速度が設定され、インデックスidx3の領域には、関連度「0.25」に比例するハイライト速度が設定される。つまり、時刻t3の時点で、各領域のハイライト速度は、「idx1>idx2>idx3」となる。 Suppose that the word “c” is recognized at the time t3. In this case, in the area of the index idx1, the recognized word “a” with the weight “0.25”, the recognized word “b” with the weight “0.25”, and the weight “0.25” are set. The extracted word includes three recognition words “c” for which “is set”. For this reason, the relevance is calculated as “0.75” by the calculation of “0.25 + 0.25 + 0.25”. The area of index idx2 includes a recognized word “a” in which a weight “0.25” is set for the extracted word and a recognized word “c” in which the weight “0.25” is set for the extracted word. Therefore, the relevance is calculated as “0.5”. On the other hand, in the area of index idx3, as in the case of time t2, the extracted word includes the recognized word “b” having the weight “0.25”, so the degree of association is calculated as “0.25”. Is done. As a result, a highlight speed proportional to the relevance degree “0.75” is set in the index idx1, and a highlight speed proportional to the relevance degree “0.5” is set in the area of the index idx2. In the area of idx3, a highlight speed proportional to the relevance “0.25” is set. That is, at time t3, the highlight speed of each region is “idx1> idx2> idx3”.
 その後、時刻t4の時点では、図5の上から4番目に示すスライド画面へ遷移する。すなわち、インデックスidx1の領域のハイライト表示が閾値に達する段階まで進行する。一方で、インデックスidx2及びインデックスidx3の2つの領域は、互いの進行度に差はあるものの、進行度は閾値までは達していない。 After that, at time t4, the screen changes to the fourth slide screen from the top in FIG. That is, the process proceeds until the highlight display of the area of index idx1 reaches the threshold value. On the other hand, although the two areas of the index idx2 and the index idx3 have a difference in their progress, the progress does not reach the threshold.
 このようにインデックスidx1の領域のハイライト表示が閾値まで進行した場合、図6の上から1番目に示すスライド画面へ遷移する。つまり、ハイライト表示が閾値まで進行したインデックスidx1の領域は、そのままハイライト表示が維持される一方で、ハイライト表示が閾値まで進行していないインデックスidx2及びインデックスidx3の領域は、ハイライト表示がキャンセルされる。 In this way, when the highlight display of the area of the index idx1 proceeds to the threshold value, the screen transitions to the slide screen shown first from the top in FIG. That is, the area of the index idx1 in which the highlight display has progressed to the threshold is maintained as it is, while the area of the index idx2 and the index idx3 in which the highlight display has not progressed to the threshold is highlighted. Canceled.
 その後、時刻t5の時点になるまで図6の上から1番目に示すスライド画面のまま遷移し、時刻t5になって認識単語データ13cから認識単語“a”、“b”及び“c”が削除されると、図6の上から2番目に示すスライド画面へ遷移する。すなわち、インデックスidx1~インデックスidx3の3つの領域には、いずれも認識単語が含まれなくなるので、関連度が「0」と算出される。この結果、ハイライト表示が閾値まで進行したインデックスidx1の領域の関連度が低下するので、インデックスidx1の領域のハイライト表示がキャンセルされる。このため、インデックスidx1~インデックスidx3の3つの領域は、いずれもスライドの作成時に設定されたデフォルトの表示形態に戻る。そして、図6の上から2番目に示すスライド画面は、時刻t6の時点になって単語“e”が認識されるまで維持される。 Thereafter, the transition is made with the first slide screen from the top in FIG. 6 until time t5, and the recognized words “a”, “b”, and “c” are deleted from the recognized word data 13c at time t5. If it does, it will change to the slide screen shown to the 2nd from the top of FIG. That is, since the three regions from index idx1 to index idx3 do not include the recognized word, the relevance is calculated as “0”. As a result, the degree of relevance of the area with the index idx1 in which the highlight display has progressed to the threshold value is reduced, so that the highlight display of the area with the index idx1 is canceled. For this reason, all three areas of the index idx1 to the index idx3 return to the default display form set when the slide is created. The second slide screen from the top in FIG. 6 is maintained until the word “e” is recognized at time t6.
 その後、時刻t6の時点になって単語“e”が認識されると、各領域の関連度は、次のように算出される。すなわち、3つの領域のうち抽出単語に認識単語“e”が含まれる領域は、インデックスidx3だけである。このため、インデックスidx3の領域には、認識単語“e”に設定された重み「1」が関連度として算出される。一方、インデックスidx1及びインデックスidx2の領域は、いずれの認識単語も含まれないので、関連度が「0」と算出される。この結果、インデックスidx1及びインデックスidx2の2つの領域には、ハイライト速度がゼロに設定されると共に、インデックスidx3の領域には、関連度「1」に比例するハイライト速度が設定される。 After that, when the word “e” is recognized at time t6, the relevance of each region is calculated as follows. In other words, the index idx3 is the only area in which the recognized word “e” is included in the extracted word among the three areas. For this reason, the weight “1” set for the recognized word “e” is calculated as the degree of association in the area of the index idx3. On the other hand, since the areas of index idx1 and index idx2 do not include any recognized word, the degree of association is calculated as “0”. As a result, the highlight speed is set to zero in the two areas of index idx1 and index idx2, and the highlight speed proportional to the degree of association “1” is set in the area of index idx3.
 この結果、時刻t7の時点では、インデックスidx3の領域のハイライト表示が閾値まで進行する。この場合、図6の上から3番目に示すスライド画面へ遷移する。つまり、ハイライト表示が閾値まで進行したインデックスidx3の領域は、そのままハイライト表示が維持される一方で、インデックスidx2及びインデックスidx3の領域は、いずれもスライドの作成時に設定されたデフォルトの表示形態が維持される。 As a result, at time t7, the highlight display of the area of index idx3 proceeds to the threshold value. In this case, a transition is made to the third slide screen from the top in FIG. In other words, the area of index idx3 where the highlight display has progressed to the threshold is maintained as it is, while the areas of index idx2 and index idx3 both have the default display form set when the slide is created. Maintained.
 このとき、インデックスidx3の領域のハイライト表示の進行度が閾値に達するまでの期間は、インデックスidx1の領域のハイライト表示の進行度が閾値に達するまでの期間よりも短い。これは、インデックスidx3の領域には、関連度「1」に比例するハイライト速度が設定されるので、時刻t1、時刻t2及び時刻t3の各時点のうち最高でも、関連度「0.75」に比例するハイライト速度が設定されたインデックスidx1の領域のハイライト表示よりも、進行度が閾値に達するまでの期間は短くなるからである。 At this time, the period until the progress of highlight display in the area of index idx3 reaches the threshold is shorter than the period until the progress of highlight display of the area of index idx1 reaches the threshold. This is because, in the area of index idx3, a highlight speed proportional to the relevance degree “1” is set, and therefore, the relevance degree “0.75” is the highest among the time points of time t1, time t2, and time t3. This is because the period until the degree of progress reaches the threshold is shorter than the highlight display in the area of the index idx1 in which the highlight speed proportional to is set.
 これら図5及び図6に示すハイライト表示によって、時刻t1~時刻t5にかけてプレゼンタがインデックスidx1の領域に関する説明を行う場合に、インデックスidx1の領域を含めてハイライト表示できる。すなわち、時刻t1及び時刻t2の段階で、インデックスidx2やインデックスidx3の領域の抽出単語に認識単語が含まれるからといってインデックスidx2やインデックスidx3の領域のハイライト表示だけが進行される訳ではなく、インデックスidx1の領域もハイライト表示が進行される。したがって、プレゼンタの説明箇所がハイライト表示されない事態を抑制できる。 5 and FIG. 6, when the presenter explains the area of index idx1 from time t1 to time t5, the area including index idx1 can be highlighted. That is, at the stage of time t1 and time t2, just because the recognized word is included in the extracted word in the area of index idx2 or index idx3, only the highlight display of the area of index idx2 or index idx3 does not proceed. The highlight display also proceeds in the area of index idx1. Therefore, it is possible to suppress the situation where the presenter's explanation part is not highlighted.
 さらに、プレゼンタが時刻t6からインデックスidx3の領域に関する説明を行う場合には、インデックスidx3の領域の抽出単語にしか含まれない単語が認識される。このようにプレゼンタがインデックスidx3の領域に関する説明を確度が高い場合、インデックスidx3の領域のハイライト表示の進行度を高めることもできる。したがって、プレゼンタの発話からハイライト表示までのレスポンスの低下を抑制しつつ、誤りのおそれがあるハイライト表示に注意が集まる事態を抑制できる。 Furthermore, when the presenter explains the area of index idx3 from time t6, words that are included only in the extracted word of the area of index idx3 are recognized. As described above, when the presenter has a high degree of accuracy in explaining the area with the index idx3, the progress of highlight display in the area with the index idx3 can be increased. Accordingly, it is possible to suppress a situation where attention is focused on a highlight display that may cause an error while suppressing a decrease in response from the utterance of the presenter to the highlight display.
[処理の流れ]
 次に、本実施例に係るプレゼンテーション支援装置10の処理の流れについて説明する。なお、ここでは、プレゼンテーション支援装置10によって実行される(1)重み付与処理、(2)音声認識処理、(3)表示制御処理の順に説明することとする。
[Process flow]
Next, a processing flow of the presentation support apparatus 10 according to the present embodiment will be described. Here, description will be made in the order of (1) weighting processing, (2) speech recognition processing, and (3) display control processing executed by the presentation support apparatus 10.
(1)重み付与処理
 図7は、実施例1に係る重み付与処理の手順を示すフローチャートである。この処理は、自動的に開始することもできるし、手動設定で開始することもできる。例えば、自動的に開始する場合、プレゼンテーションソフトが文書ファイルを記憶部13に保存した上で閉じる場合、あるいはプレゼンテーションを介する文書ファイルの編集中に文書ファイルが記憶部13に保存された場合に、処理を起動させることができる。また、手動設定で開始する場合、入力装置7を介してプレゼンテーションの前処理の実行指示を受け付けた場合に、処理を起動させることができる。いずれの場合においても、記憶部13に記憶された文書データ13aが含む文書ファイルのうち、保存または実行指示に対応する文書ファイルを読み出すことによって処理が開始される。
(1) Weighting Process FIG. 7 is a flowchart illustrating the procedure of the weighting process according to the first embodiment. This process can be started automatically or manually. For example, when starting automatically, when the presentation software saves the document file in the storage unit 13 and then closes, or when the document file is saved in the storage unit 13 while editing the document file via the presentation, Can be activated. In addition, when starting with manual setting, the processing can be activated when an instruction to execute presentation pre-processing is received via the input device 7. In any case, the processing is started by reading out the document file corresponding to the save or execution instruction from the document files included in the document data 13a stored in the storage unit 13.
 図7に示すように、分割部15aは、文書ファイルに含まれるスライドを一文、行または段落などの単位で複数の領域へ分割する(ステップS101)。続いて、分割部15aは、ステップS101で得られた領域に各領域を識別するインデックスを割り当てる(ステップS102)。 As shown in FIG. 7, the dividing unit 15a divides the slide included in the document file into a plurality of areas in units of one sentence, line, paragraph, or the like (step S101). Subsequently, the dividing unit 15a assigns an index for identifying each area to the area obtained in step S101 (step S102).
 そして、抽出部15bは、ステップS102で割り当てられたインデックスのうちインデックスを1つ選択する(ステップS103)。続いて、抽出部15bは、ステップS103で選択されたインデックスの領域内の文字列に形態素解析等を実行することにより得られた形態素のうち品詞が名詞である単語を抽出する(ステップS104)。その後、抽出部15bは、ステップS104で抽出された各単語に当該単語が含まれる領域に割り当てられたインデックスを付与する(ステップS105)。 And the extraction part 15b selects one index among the indexes allocated by step S102 (step S103). Subsequently, the extraction unit 15b extracts a word whose part of speech is a noun from morphemes obtained by executing morphological analysis or the like on the character string in the index area selected in step S103 (step S104). Thereafter, the extraction unit 15b gives each word extracted in step S104 an index assigned to the area including the word (step S105).
 そして、抽出部15bは、ステップS102で割り当てられたインデックスが全て選択されるまで(ステップS106No)、上記のステップS103~ステップS105までの処理を繰返し実行する。 Then, the extraction unit 15b repeatedly executes the processes from step S103 to step S105 until all the indexes assigned in step S102 are selected (No in step S106).
 その後、ステップS101で割り当てられたインデックスが全て選択された場合(ステップS106Yes)、付与部15cは、スライドに含まれる単語ごとに当該単語kの出現頻度fを算出する(ステップS107)。そして、付与部15cは、ステップS107で単語別に算出された出現頻度fに対応する単語の重みwを付与する(ステップS108)。その上で、付与部15cは、単語k、インデックスidx及び重みwが対応付けられた抽出単語データ13bを記憶部13へ登録し(ステップS109)、処理を終了する。 After that, when all the indexes assigned in step S101 are selected (step S106 Yes), the assigning unit 15c calculates the appearance frequency f k of the word k for each word included in the slide (step S107). Then, the assigning unit 15c assigns a word weight w k corresponding to the appearance frequency f k calculated for each word in step S107 (step S108). On top of that, imparting unit 15c, the word k, and registers the extracted word data 13b index idx and weights w k is associated to the storage unit 13 (step S109), and ends the process.
(2)音声認識処理
 図8は、実施例1に係る音声認識処理の手順を示すフローチャートである。この処理は、プレゼンテーションソフトが文書ファイルを開いた状態でプレゼンテーションの開始指示を受け付けた場合に起動し、プレゼンテーションの終了指示を受け付けるまで繰返し実行される。
(2) Voice Recognition Processing FIG. 8 is a flowchart illustrating the procedure of voice recognition processing according to the first embodiment. This process is started when the presentation software receives a presentation start instruction with the document file opened, and is repeatedly executed until a presentation end instruction is received.
 図8に示すように、認識部15dは、マイク3から所定時間長の音声信号が入力されるまで、例えば少なくとも1フレーム分の時間長、例えば10msecの音声信号が入力されるまで待機する(ステップS301)。 As shown in FIG. 8, the recognition unit 15d waits until an audio signal having a predetermined time length is input from the microphone 3 until an audio signal having a time length of, for example, at least one frame, for example, 10 msec is input (Step S15). S301).
 そして、マイク3から所定時間長の音声信号が入力されると(ステップS301Yes)、認識部15dは、当該音声信号にワードスポッティングなどの音声認識を実行する(ステップS302)。かかるステップS302でワードスポッティングが実行される場合には、記憶部13に記憶された抽出単語データ13bのうちプレゼンテーションソフトが実行中である文書ファイルが含むスライドであり、かつ表示装置5に表示中であるスライドに関する抽出単語データが辞書データとして適用される。 Then, when an audio signal having a predetermined time length is input from the microphone 3 (Yes at Step S301), the recognition unit 15d performs voice recognition such as word spotting on the audio signal (Step S302). When word spotting is executed in step S302, the slide is included in the document file being executed by the presentation software among the extracted word data 13b stored in the storage unit 13 and is being displayed on the display device 5. Extracted word data relating to a certain slide is applied as dictionary data.
 このとき、音声信号から単語が認識された場合(ステップS303Yes)、認識部15dは、ステップS302で認識された単語及びその単語が認識された時間が対応付けられた認識単語データ13cを記憶部13へ登録し(ステップS304)、ステップS305の処理へ移行する。 At this time, when a word is recognized from the audio signal (Yes in step S303), the recognition unit 15d stores the recognized word data 13c in which the word recognized in step S302 and the time when the word is recognized are associated with each other. (Step S304), and the process proceeds to step S305.
 一方、マイク3から所定時間長の音声信号が入力されていない場合、あるいは音声信号から単語が認識されなかった場合(ステップS301NoまたはステップS303No)、以降の処理を飛ばしてステップS305の処理へ移行する。 On the other hand, when a voice signal having a predetermined time length is not input from the microphone 3 or when a word is not recognized from the voice signal (No in Step S301 or Step S303), the subsequent processing is skipped and the processing proceeds to Step S305. .
 ここで、認識部15dは、記憶部13に記憶された認識単語データ13cのうち記憶部13へ登録されてから所定の期間が経過した単語が存在するか否かを判定する(ステップS305)。そして、記憶部13へ登録されてから所定の期間が経過した単語が存在する場合(ステップS305Yes)、認識部15dは、記憶部13に記憶された認識単語データ13cから当該単語に関するレコードを削除する(ステップS306)。なお、記憶部13へ登録されてから所定の期間が経過した単語が存在しない場合(ステップS305No)には、ステップS306の処理を飛ばしてステップS307の処理へ移行する。 Here, the recognizing unit 15d determines whether or not there is a word for which a predetermined period has elapsed since registration in the storage unit 13 among the recognized word data 13c stored in the storage unit 13 (step S305). If there is a word for which a predetermined period has elapsed since registration in the storage unit 13 (Yes in step S305), the recognition unit 15d deletes the record related to the word from the recognized word data 13c stored in the storage unit 13. (Step S306). If there is no word for which a predetermined period has elapsed since registration in the storage unit 13 (No in step S305), the process of step S306 is skipped and the process proceeds to step S307.
 その後、認識部15dは、表示装置5に表示されるスライドのページが変更されたか否かを判定する(ステップS307)。このとき、表示装置5に表示されるスライドのページが変更された場合(ステップS307Yes)、認識部15dは、記憶部13に記憶された認識単語データ13cを削除し(ステップS308)、ステップS301の処理へ戻る。なお、表示装置5に表示されるスライドのページが変更されていない場合(ステップS307No)、ステップS308の処理を実行せずにステップS301の処理へ戻る。 Thereafter, the recognition unit 15d determines whether or not the slide page displayed on the display device 5 has been changed (step S307). At this time, when the slide page displayed on the display device 5 is changed (Yes in step S307), the recognition unit 15d deletes the recognized word data 13c stored in the storage unit 13 (step S308), and in step S301. Return to processing. If the slide page displayed on the display device 5 has not been changed (No at Step S307), the process returns to Step S301 without executing Step S308.
(3)表示制御処理
 図9は、実施例1に係る表示制御処理の手順を示すフローチャートである。この処理は、図8に示した音声認識処理と並行して実行される処理であり、プレゼンテーションソフトが文書ファイルを開いた状態でプレゼンテーションの開始指示を受け付けた場合に起動し、プレゼンテーションの終了指示を受け付けるまで繰返し実行される。なお、処理の実行が繰り返される周期は、図8に示した音声認識処理と同様であってもよいし、異なってもよく、図8に示した音声認識処理と同期して実行されることとしてもよいし、非同期で実行されることとしてもかまわない。
(3) Display Control Process FIG. 9 is a flowchart illustrating the procedure of the display control process according to the first embodiment. This process is executed in parallel with the voice recognition process shown in FIG. 8, and is started when the presentation software receives a presentation start instruction with the document file opened, and gives a presentation end instruction. It is repeatedly executed until it is accepted. The cycle in which the execution of the process is repeated may be the same as or different from the voice recognition process shown in FIG. 8, and is executed in synchronization with the voice recognition process shown in FIG. It can also be executed asynchronously.
 図9に示すように、算出部15eは、表示装置5に表示中であるスライドが含む領域のインデックスのうちインデックスを1つ選択する(ステップS501)。続いて、算出部15eは、ステップS501で選択されたインデックスの領域に対応付けられた抽出単語データ13bの抽出単語のうち認識単語と一致する抽出単語に付与された重みから当該領域の関連度を算出する(ステップS502)。 As shown in FIG. 9, the calculation unit 15e selects one index from among the indexes of the area included in the slide being displayed on the display device 5 (step S501). Subsequently, the calculation unit 15e calculates the degree of association of the area from the weights assigned to the extracted words that match the recognized word among the extracted words of the extracted word data 13b associated with the index area selected in step S501. Calculate (step S502).
 そして、設定部15fは、ステップS501で選択されたインデックスの領域に対し、ステップS502で算出された関連度が高いほどハイライト速度を高く設定し、あるいは関連度が低いほどハイライト速度を低く設定する(ステップS503)。 Then, the setting unit 15f sets the highlight speed higher as the relevance calculated in step S502 is higher for the index area selected in step S501, or sets the highlight speed lower as the relevance is lower. (Step S503).
 その後、全てのインデックスが選択されるまで(ステップS504No)、上記のステップS501~ステップS503までの処理を繰返し実行する。これによって、全ての領域にハイライト速度が設定されるまで上記のステップS501~ステップS503までの処理が繰返し実行されることになる。 Thereafter, the processes from step S501 to step S503 are repeatedly executed until all indexes are selected (No in step S504). As a result, the processing from step S501 to step S503 is repeatedly executed until the highlight speed is set for all regions.
 その後、全てのインデックスが選択された場合(ステップS504Yes)、ハイライト表示の進行度が所定の閾値以上である領域の有無を監視する(ステップS505)。このとき、ハイライト表示の進行度が所定の閾値以上である領域が存在しない場合(ステップS505No)、表示制御部15gは、次のような処理を実行する。すなわち、表示制御部15gは、ステップS503で各領域に設定されたハイライト速度にしたがって各領域のハイライト表示を進行させ(ステップS506)、処理を終了する。 Thereafter, when all the indexes are selected (step S504 Yes), the presence / absence of an area where the progress of highlight display is equal to or greater than a predetermined threshold is monitored (step S505). At this time, when there is no region where the progress degree of highlight display is equal to or greater than the predetermined threshold (No in step S505), the display control unit 15g executes the following process. That is, the display control unit 15g advances the highlight display of each area according to the highlight speed set for each area in step S503 (step S506), and ends the process.
 一方、ハイライト表示の進行度が所定の閾値以上である領域が存在する場合(ステップS505Yes)、表示制御部15gは、次のような処理を実行する。すなわち、表示制御部15gは、ハイライト表示の進行度が閾値以上である領域のハイライト速度の設定は維持し、ハイライト表示の進行度が閾値未満である領域のハイライト表示を元の状態に戻すことによりハイライト表示をキャンセルし(ステップS507)、ハイライト表示の進行度が閾値未満である領域のハイライト速度をゼロにリセットする(ステップS508)。 On the other hand, when there is an area where the progress degree of highlight display is equal to or greater than the predetermined threshold (Yes in step S505), the display control unit 15g executes the following process. That is, the display control unit 15g maintains the setting of the highlight speed in the area where the progress of highlight display is equal to or higher than the threshold, and the highlight display of the area where the progress of highlight display is less than the threshold is in the original state. The highlight display is canceled by returning to (step S507), and the highlight speed of the area where the progress of highlight display is less than the threshold is reset to zero (step S508).
 続いて、表示制御部15gは、ハイライト表示の進行度が閾値以上である領域でカレントの関連度がパストの関連度未満であるか否かを判定する(ステップS509)。このとき、カレントの関連度がパストの関連度以上である場合(ステップS509No)、表示制御部15gは、ステップS508で各領域に設定されたハイライト速度にしたがって各領域のハイライト表示を進行させ(ステップS506)、処理を終了する。 Subsequently, the display control unit 15g determines whether or not the current relevance level is less than the past relevance level in an area where the progress degree of highlight display is equal to or greater than the threshold (step S509). At this time, if the current relevance level is equal to or higher than the past relevance level (No in step S509), the display control unit 15g advances the highlight display of each area according to the highlight speed set in each area in step S508. (Step S506), the process ends.
 一方、カレントの関連度がパストの関連度未満である場合(ステップS509Yes)、表示制御部15gは、関連度が低下した領域のハイライト表示を元の状態に戻すことによりハイライト表示をキャンセルし(ステップS510)、ハイライト速度をゼロにリセットし(ステップS511)、処理を終了する。 On the other hand, if the current relevance level is less than the past relevance level (Yes in step S509), the display control unit 15g cancels the highlight display by returning the highlight display of the area with the low relevance level to the original state. (Step S510), the highlight speed is reset to zero (Step S511), and the process ends.
[効果の一側面]
 上述してきたように、本実施例に係るプレゼンテーション支援装置10は、プレゼンテーション画面のうち音声の認識結果として得られた単語を含む領域をハイライト表示させる場合に、単語との関連度が高い領域の表示ほどハイライトの速度を上げ、関連度が低い領域の表示ほどハイライトの速度を下げる。
[One aspect of effect]
As described above, when the presentation support apparatus 10 according to the present embodiment highlights an area including a word obtained as a speech recognition result on the presentation screen, the presentation support apparatus 10 The speed of highlighting is increased as the display is made, and the speed of highlighting is lowered as the display is in a region having a lower degree of association.
 このように、本実施例に係るプレゼンテーション支援装置10では、ハイライト表示が実行される領域が必ずしも択一に限定されない。このため、音声認識で単語が複数の領域にまたがって検出された場合、各領域でハイライト表示が実行される。それ故、ハイライト表示が実行される領域の中にプレゼンタの説明箇所が含まれる可能性を高めることができる。したがって、本実施例に係るプレゼンテーション支援装置10によれば、プレゼンタの説明箇所がハイライト表示されない事態を抑制できる。 Thus, in the presentation support apparatus 10 according to the present embodiment, the area where the highlight display is executed is not necessarily limited to the alternative. For this reason, when a word is detected across a plurality of areas by voice recognition, highlight display is executed in each area. Therefore, it is possible to increase the possibility that the presenting part of the presenter is included in the area where the highlight display is executed. Therefore, according to the presentation support apparatus 10 according to the present embodiment, it is possible to suppress the situation where the presenting part of the presenter is not highlighted.
 さらに、本実施例に係るプレゼンテーション支援装置10では、音声認識の結果として得た単語及び領域の関連度の高低によって領域をハイライトさせる速度が変化する。例えば、複数の領域がハイライト表示される場合でも、プレゼンタの説明箇所である可能性が高い領域に関するハイライト表示のレスポンスを上げる。これによって、プレゼンタの説明箇所である可能性が高い領域を注目させやすくできる。これと共に、プレゼンタの説明箇所である可能性が低い領域に関するハイライト表示のレスポンスを下げる。これによって、プレゼンタの説明箇所でない可能性が他の領域よりも高い領域が知覚されるのを遅らせることができる。このように、プレゼンタの発話からハイライト表示までのレスポンスの低下を抑制しつつ、誤りのおそれがあるハイライト表示に注意が集まる事態を抑制できる。 Furthermore, in the presentation support apparatus 10 according to the present embodiment, the speed of highlighting a region varies depending on the degree of association between words and regions obtained as a result of speech recognition. For example, even when a plurality of areas are highlighted, the response of highlight display regarding an area that is highly likely to be an explanation part of the presenter is increased. As a result, it is possible to easily draw attention to a region that is highly likely to be an explanation location of the presenter. At the same time, the response of highlight display relating to the region that is unlikely to be an explanation location of the presenter is lowered. As a result, it is possible to delay the perception of an area where the possibility of not being the presenting part of the presenter is higher than other areas. As described above, it is possible to suppress a situation in which attention is focused on a highlight display that may cause an error while suppressing a decrease in response from the utterance of the presenter to the highlight display.
 さて、これまで開示の装置に関する実施例について説明したが、本発明は上述した実施例以外にも、種々の異なる形態にて実施されてよいものである。そこで、以下では、本発明に含まれる他の実施例を説明する。 Now, although the embodiments related to the disclosed device have been described so far, the present invention may be implemented in various different forms other than the above-described embodiments. Therefore, another embodiment included in the present invention will be described below.
[テキスト以外のハイライト表示]
 上記の実施例1では、スライドのうち文字列を含む領域をハイライト表示させる場合を例示したが、スライドには、文字列以外にも、グラフ、表、画像や動画などが含まれていてもかまわない。この場合、例えば、プレゼンテーション支援装置10は、グラフ、表、画像や動画に設定されたメタ情報に含まれる文字列から単語を抽出することによって上記の実施例1と同様に抽出単語データを生成することができる。
[Highlights other than text]
In the first embodiment, the case where the region including the character string in the slide is highlighted is illustrated. However, the slide may include a graph, a table, an image, a movie, or the like in addition to the character string. It doesn't matter. In this case, for example, the presentation support device 10 generates extracted word data in the same manner as in the first embodiment by extracting words from character strings included in meta information set in graphs, tables, images, and moving images. be able to.
[ハイライト表示の対象]
 上記の実施例1では、ハイライト表示の進行度が閾値以上であるか否かによってハイライトの表示及び非表示を切り替える場合を例示したが、他の因子によってハイライトの表示及び非表示を切り替えることもできる。例えば、表示制御部15gは、領域の関連度または領域のハイライト速度が所定の閾値以上であるか否かによってハイライトの表示及び非表示を切り替えることもできる。この場合、時間経過によってハイライト表示が進行するよりも前にハイライトの表示及び非表示を制御できる。
[Target of highlight display]
In the first embodiment, the case where highlight display and non-display are switched depending on whether or not the progress of highlight display is equal to or higher than the threshold is illustrated. However, the display and non-display of highlight are switched depending on other factors. You can also. For example, the display control unit 15g can switch between display and non-display of the highlight depending on whether or not the relevance of the region or the highlight speed of the region is equal to or higher than a predetermined threshold. In this case, the display and non-display of the highlight can be controlled before the highlight display progresses with time.
[出現頻度の応用例]
 上記の実施例1では、単語kが同一のスライドに出現する回数が集計された総出現回数を出現頻度として用いる場合を例示したが、必ずしも出現頻度は総出現回数に限定されない。例えば、単語kが各領域間で出現する回数が集計された領域間の出現回数を出現頻度として用いることもできる。一例を挙げれば、インデックスidx1~idx3の3つの領域のうち単語kが1つの領域に出現される場合には、領域間の出現頻度が1/3であるので、単語kの重みとして1/(1/3)を付与する。このとき、単語kが1つの領域に出現する回数はのべ回数では集計されず、複数回にわたって出現しても同様の重みが付与される。
[Application example of appearance frequency]
In the first embodiment, the case where the total number of appearances in which the number of times the word k appears on the same slide is counted is used as the appearance frequency. However, the appearance frequency is not necessarily limited to the total number of appearances. For example, the number of appearances between regions in which the number of times the word k appears between the regions can be counted can be used as the appearance frequency. As an example, when the word k appears in one area among the three areas of the indices idx1 to idx3, the frequency of appearance between the areas is 1/3, so the weight of the word k is 1 / ( 1/3) 2 is given. At this time, the number of times that the word k appears in one region is not counted as the total number of times, and the same weight is given even if it appears multiple times.
[重み付与方法の応用例1]
 また、上記の実施例1では、単語kの出現頻度に応じて単語kに重みを付与する場合を例示したが、出現頻度以外の因子によって単語kに重みを付与することもできる。例えば、付与部15cは、単語kのモーラ数によって単語kに重みを付与することもできる。具体的には、付与部15cは、単語kのモーラ数が多いほど大きい重みを付与することができる。例えば、重みの算出式の一例として、次の2式を用いることができる。すなわち、単語kのモーラ数mの重みをwとしたとき、モーラ数mが固定値M、例えば「6」よりも大きい場合には、「w=1(m>M)」を用いる一方で、モーラ数mが固定値M以下である場合には、「w=m/M(m≦M)」を用いる。なお、ここでは、モーラ音素を用いる場合を例示したが、他の音素を用いることができるのは言うまでもない。
[Application example 1 of weighting method]
Further, in the first embodiment, the case where the weight is given to the word k according to the appearance frequency of the word k is exemplified, but the weight can be given to the word k by a factor other than the appearance frequency. For example, the assigning unit 15c can assign a weight to the word k according to the number of mora of the word k. Specifically, the assigning unit 15c can assign a greater weight as the number of mora of the word k increases. For example, the following two formulas can be used as an example of a weight calculation formula. That is, when the weight of the mora number m of the word k is w m and the mora number m is larger than a fixed value M, for example, “6”, “w m = 1 (m> M)” is used. When the mora number m is a fixed value M or less, “w m = m / M (m ≦ M)” is used. In addition, although the case where a mora phoneme was used was illustrated here, it cannot be overemphasized that another phoneme can be used.
 一般に、モーラ音素が少ないほど音声認識の精度は低下する傾向にある。それ故、上記の重み付与方法により、単語kのモーラ音素が少ない場合よりも単語kのモーラ音素が多い場合の方が大きい重みを付与することで、音声認識の精度が高くなるにつれて大きな重みを付与できる結果、関連度の算出精度も高めることができる。なお、本項で説明した重み付与方法は、上記の実施例1で説明した重み付与方法の代わりに単独で用いて実施することもできるし、また、上記の実施例1で説明した重み付与方法や実施例2で説明する他の重み付与方法と組み合わせて実施することもできる。 In general, the accuracy of speech recognition tends to decrease as the number of mora phonemes decreases. Therefore, according to the above weighting method, a larger weight is given when the number of mora phonemes of the word k is larger than when the number of mora phonemes of the word k is small. As a result that can be given, the calculation accuracy of the relevance can be improved. It should be noted that the weighting method described in this section can be implemented by using it alone instead of the weighting method described in the first embodiment, or the weighting method described in the first embodiment. It can also be implemented in combination with other weighting methods described in the second embodiment.
[重み付与方法の応用例2]
 また、上記の実施例1では、音声認識が実行される前に抽出単語に重みを付与する場合を例示したが、重み付与方法はこれに限定されない。例えば、プレゼンテーション支援装置10は、音声認識が実行された後に認識単語に重みを付与することもできる。すなわち、音声認識が実行される場合には認識単語と共に認識単語が学習データ等の正解である尤もらしさ、いわゆるスコアが算出されることが多い。このため、プレゼンテーション支援装置10は、認識単語にスコアの多寡に応じて重みを付与することもできる。なお、本項で説明した重み付与方法は、上記の実施例1で説明した重み付与方法の代わりに単独で用いて実施することもできるし、また、上記の実施例1で説明した重み付与方法や実施例2で説明する他の重み付与方法と組み合わせて実施することもできる。
[Application example 2 of weighting method]
In the first embodiment, the case where the extracted word is given a weight before the speech recognition is executed is illustrated, but the weighting method is not limited to this. For example, the presentation support apparatus 10 can give a weight to a recognized word after voice recognition is performed. That is, when speech recognition is performed, the likelihood that the recognized word is a correct answer such as learning data, that is, a so-called score is often calculated together with the recognized word. For this reason, the presentation support apparatus 10 can also give a weight to the recognized word according to the score. It should be noted that the weighting method described in this section can be implemented by using it alone instead of the weighting method described in the first embodiment, or the weighting method described in the first embodiment. It can also be implemented in combination with other weighting methods described in the second embodiment.
[関連度の算出方法の応用例1]
 上記の実施例1では、認識単語と一致する抽出単語の数により関連度を領域ごとに算出する場合を例示したが、関連度の算出方法はこれに限定されない。すなわち、上記の実施例1では、認識単語と一致する抽出単語の重みを足し合わせる場合を例示したが、抽出単語の総数に対する、認識単語と一致する抽出単語の数の割合により、関連度を領域ごとに算出することもできる。このような関連度の算出方法を用いる理由は、各領域から抽出される抽出単語の総数が必ずしも同一または略同一であるとは限らないからである。このため、認識単語と一致する抽出単語の重みを足し合わせる場合、抽出単語の総数の少ない領域の方が抽出単語の総数の多い領域よりも関連度が不当に低く算出されることもある。このことから、上記の割合をそのまま関連度として用いたり、上記の割合を用いて認識単語と一致する抽出単語の重みが足し合わされた合計値を正規化することにより、関連度の算出精度を高めることができる。なお、本項で説明した重み付与方法は、上記の実施例1で説明した重み付与方法の代わりに単独で用いて実施することもできるし、また、上記の実施例1で説明した重み付与方法や実施例2で説明する他の重み付与方法と組み合わせて実施することもできる。
[Application Example 1 of Relevance Calculation Method]
In the first embodiment, the case in which the degree of association is calculated for each region based on the number of extracted words that match the recognized word is exemplified, but the method for calculating the degree of association is not limited to this. That is, in the first embodiment, the case where the weights of the extracted words that match the recognized words are added is illustrated. It can also be calculated for each. The reason for using such a relevance calculation method is that the total number of extracted words extracted from each region is not necessarily the same or substantially the same. For this reason, when adding the weights of the extracted words that match the recognized words, the region having a smaller total number of extracted words may be calculated with an unreasonably lower relevance than the region having a large total number of extracted words. For this reason, the above-mentioned ratio is used as it is as the degree of relevance, or the total value obtained by adding the weights of the extracted words that match the recognized word is normalized using the above-mentioned ratio, thereby increasing the calculation accuracy of the degree of relevance. be able to. It should be noted that the weighting method described in this section can be implemented by using it alone instead of the weighting method described in the first embodiment, or the weighting method described in the first embodiment. It can also be implemented in combination with other weighting methods described in the second embodiment.
[プレゼンタの指示操作]
 例えば、プレゼンテーション支援装置10は、入力装置7等を介してハイライト表示の加速指示またはキャンセル指示を受け付けることもできる。例えば、キーボードに含まれる所定のキー、マウスが有する所定のボタン、あるいはリモコン機能付きのレーザポインタが有する所定のボタンに、ハイライト表示を加速する指示を受け付けたり、ハイライト表示をキャンセルする指示を受け付けたりするキーやボタンを割り当てる。そして、プレゼンテーション支援装置10は、ハイライト表示を加速する指示を受け付けた場合に、当該指示を受け付けた段階でハイライト表示の進行度が最も大きい領域のハイライト速度を加速させる。このとき、プレゼンテーション支援装置10は、ハイライト表示の進行度が最も大きい領域のハイライト表示の進行度を一度に閾値まで引き上げることも上記の加速の範疇に含まれる。一方、プレゼンテーション支援装置10は、ハイライト表示を取り消す指示を受け付けた場合に、当該指示を受け付けた段階でハイライト表示の進行度が最も大きい領域のハイライト表示をキャンセルすると共に当該領域のハイライト速度をリセットすることもできる。
[Presenter instructions]
For example, the presentation support apparatus 10 can accept an instruction to accelerate or cancel highlight display via the input device 7 or the like. For example, an instruction to accelerate highlight display or an instruction to cancel highlight display is given to a predetermined key included in a keyboard, a predetermined button of a mouse, or a predetermined button of a laser pointer with a remote control function. Assign keys and buttons to accept. When the presentation support apparatus 10 receives an instruction for accelerating highlight display, the presentation support apparatus 10 accelerates the highlight speed in a region where the degree of progress of highlight display is the highest when the instruction is received. At this time, the presentation support apparatus 10 also includes raising the progress of highlight display in a region where the progress of highlight display is the highest to a threshold value at a time. On the other hand, when the presentation support apparatus 10 receives an instruction to cancel highlight display, the presentation support apparatus 10 cancels highlight display of the area where the progress of highlight display is the highest when the instruction is received and highlights the area. You can also reset the speed.
 ここで、プレゼンタは、プレゼンタ自身がプレゼンテーションに用いるスライドを作成している点、さらには、プレゼンタ自身がプレゼンテーションの前準備としてスライドの説明順序や論理構成を組み立てている点などから、聴講者よりもスライドでハイライト表示された領域を気付くことができる可能性が高い。このため、聴講者がハイライト表示に気付く前にハイライト表示を加速する指示を受け付けたり、ハイライト表示をキャンセルする指示を受け付けたりできる可能性が高まる。したがって、聴講者に対するハイライト表示のレスポンスを高めたり、聴講者が誤ったハイライト表示に気付く事態を抑制したりすることができる。 Here, the presenter creates slides that the presenter himself uses for the presentation, and the presenter himself assembles the slide description order and logical configuration as preparations for the presentation. It is likely that you will notice the highlighted area on the slide. This increases the possibility that the listener can accept an instruction to accelerate the highlight display or notice an instruction to cancel the highlight display before noticing the highlight display. Therefore, it is possible to increase the response of highlight display to the listener, or to suppress a situation where the listener notices an erroneous highlight display.
[関連度の算出範囲]
 上記の実施例1では、表示中のスライド内の領域に絞って各領域の関連度を算出する場合を例示したが、必ずしも関連度の算出範囲はこれに限定されない。例えば、表示中のスライド以外の領域についても関連度の算出範囲に含めることもできる。このとき、表示中のスライド以外の領域の関連度が表示中のスライド内の領域の関連度よりも高い場合、プレゼンテーション支援装置10は、表示装置5に表示されるスライドを関連度が最高である領域を持つスライドに切り替えた上で当該領域に関するハイライト表示を進行させることもできる。
[Relation degree calculation range]
In the above-described first embodiment, the case where the degree of association of each area is calculated by focusing on the area in the slide being displayed is illustrated, but the calculation range of the degree of association is not necessarily limited thereto. For example, an area other than the slide being displayed can also be included in the relevance calculation range. At this time, when the relevance level of the region other than the slide being displayed is higher than the relevance level of the region within the slide being displayed, the presentation support apparatus 10 has the highest relevance level for the slide displayed on the display device 5. After switching to a slide having an area, highlight display regarding the area can be advanced.
[文書ファイルの応用例]
 上記の実施例1では、プレゼンテーションソフトによって作成された文書ファイルを用いる場合を例示したが、他のアプリケーションプログラムによって作成された文書ファイルを用いることができる。すなわち、表示時に画面単位で表示されるページを含む文書ファイルであれば、ワープロソフトの文書ファイルが有するページをスライドに読み替えたり、表計算ソフトの文書ファイルが有するシートをスライドに読み替えることによって図7~図9に示した処理を同様に適用できる。
[Application examples of document files]
In the first embodiment, the case where the document file created by the presentation software is used has been exemplified. However, the document file created by another application program can be used. That is, if the document file includes a page that is displayed in units of screens at the time of display, the page of the word processing software document file is replaced with a slide, or the sheet of the spreadsheet software document file is replaced with a slide. The processing shown in FIG. 9 can be similarly applied.
[他の実装例]
 上記の実施例1では、プレゼンテーション支援装置10が上記のプレゼンテーションソフトを外部のリソースに依存せずに単独で実行するスタンドアローンで上記のプレゼンテーション支援サービスを提供する場合を例示したが、他の実装形態を採用することもできる。例えば、プレゼンテーションソフトを実行するクライアントに対し、上記のプレゼンテーション支援サービスを提供するサーバを設けることによってクライアントサーバシステムとして構築することもできる。この場合、パッケージソフトウェアやオンラインソフトウェアとして上記のプレゼンテーション支援サービスを実現するプレゼンテーション支援プログラムをインストールさせることによってサーバ装置を実装できる。例えば、サーバ装置10は、上記のプレゼンテーション支援サービスを提供するWebサーバとして実装することとしてもよいし、アウトソーシングによって上記のプレゼンテーション支援サービスを提供するクラウドとして実装することとしてもかまわない。なお、上記の実施例1では、プレゼンテーション支援プログラムがプレゼンテーションソフトにアドオンされる場合を想定したが、ライセンス権限を有するクライアントからプレゼンテーション支援プログラムをライブラリとして参照する要求を受け付けた場合に、プレゼンテーション支援プログラムをプラグインさせることもできる。
[Other implementation examples]
In the first embodiment, the case where the presentation support apparatus 10 provides the presentation support service in a stand-alone manner in which the presentation software 10 is independently executed without depending on an external resource is exemplified. Can also be adopted. For example, a client server system can be constructed by providing a server that provides the presentation support service to a client that executes presentation software. In this case, the server device can be implemented by installing a presentation support program for realizing the above presentation support service as package software or online software. For example, the server device 10 may be implemented as a Web server that provides the presentation support service, or may be implemented as a cloud that provides the presentation support service by outsourcing. In the first embodiment, it is assumed that the presentation support program is added to the presentation software. However, when a request for referring to the presentation support program as a library is received from a client having a license authority, the presentation support program is It can also be plugged in.
[プレゼンテーション支援プログラム]
 また、上記の実施例で説明した各種の処理は、予め用意されたプログラムをパーソナルコンピュータやワークステーションなどのコンピュータで実行することによって実現することができる。そこで、以下では、図10を用いて、上記の実施例と同様の機能を有するプレゼンテーション支援プログラムを実行するコンピュータの一例について説明する。
[Presentation support program]
The various processes described in the above embodiments can be realized by executing a prepared program on a computer such as a personal computer or a workstation. In the following, an example of a computer that executes a presentation support program having the same function as that of the above embodiment will be described with reference to FIG.
 図10は、実施例1及び実施例2に係るプレゼンテーション支援プログラムを実行するコンピュータのハードウェア構成例を示す図である。図10に示すように、コンピュータ100は、操作部110aと、スピーカ110bと、カメラ110cと、ディスプレイ120と、通信部130とを有する。さらに、このコンピュータ100は、CPU150と、ROM160と、HDD170と、RAM180とを有する。これら110~180の各部はバス140を介して接続される。 FIG. 10 is a diagram illustrating a hardware configuration example of a computer that executes the presentation support program according to the first and second embodiments. As illustrated in FIG. 10, the computer 100 includes an operation unit 110a, a speaker 110b, a camera 110c, a display 120, and a communication unit 130. Further, the computer 100 includes a CPU 150, a ROM 160, an HDD 170, and a RAM 180. These units 110 to 180 are connected via a bus 140.
 HDD170には、図10に示すように、上記の実施例1で示した分割部15a、抽出部15b、付与部15c、認識部15d、算出部15e、設定部15f及び表示制御部15gと同様の機能を発揮するプレゼンテーション支援プログラム170aが記憶される。このプレゼンテーション支援プログラム170aは、図2に示した分割部15a、抽出部15b、付与部15c、認識部15d、算出部15e、設定部15f及び表示制御部15gの各構成要素と同様、統合又は分離してもかまわない。すなわち、HDD170には、必ずしも上記の実施例1で示した全てのデータが格納されずともよく、処理に用いるデータがHDD170に格納されればよい。 As shown in FIG. 10, the HDD 170 is similar to the dividing unit 15a, the extracting unit 15b, the adding unit 15c, the recognizing unit 15d, the calculating unit 15e, the setting unit 15f, and the display control unit 15g described in the first embodiment. A presentation support program 170a that performs the function is stored. This presentation support program 170a is integrated or separated in the same manner as each component of the dividing unit 15a, extracting unit 15b, adding unit 15c, recognizing unit 15d, calculating unit 15e, setting unit 15f and display control unit 15g shown in FIG. It doesn't matter. That is, the HDD 170 does not necessarily have to store all the data shown in the first embodiment, and data used for processing may be stored in the HDD 170.
 このような環境の下、CPU150は、HDD170からプレゼンテーション支援プログラム170aを読み出した上でRAM180へ展開する。この結果、プレゼンテーション支援プログラム170aは、図10に示すように、プレゼンテーション支援プロセス180aとして機能する。このプレゼンテーション支援プロセス180aは、RAM180が有する記憶領域のうちプレゼンテーション支援プロセス180aに割り当てられた領域にHDD170から読み出した各種データを展開し、この展開した各種データを用いて各種の処理を実行する。例えば、プレゼンテーション支援プロセス180aが実行する処理の一例として、図7~図9に示す処理などが含まれる。なお、CPU150では、必ずしも上記の実施例1で示した全ての処理部が動作せずともよく、実行対象とする処理に対応する処理部が仮想的に実現されればよい。 Under such an environment, the CPU 150 reads the presentation support program 170a from the HDD 170 and develops it on the RAM 180. As a result, the presentation support program 170a functions as a presentation support process 180a as shown in FIG. The presentation support process 180a expands various data read from the HDD 170 in an area allocated to the presentation support process 180a in the storage area of the RAM 180, and executes various processes using the expanded data. For example, examples of processing executed by the presentation support process 180a include the processing shown in FIGS. Note that the CPU 150 does not necessarily operate all the processing units described in the first embodiment, and the processing unit corresponding to the process to be executed may be virtually realized.
 なお、上記のプレゼンテーション支援プログラム170aは、必ずしも最初からHDD170やROM160に記憶されておらずともかまわない。例えば、コンピュータ100に挿入されるフレキシブルディスク、いわゆるFD、CD-ROM、DVDディスク、光磁気ディスク、ICカードなどの「可搬用の物理媒体」に各プログラムを記憶させる。そして、コンピュータ100がこれらの可搬用の物理媒体から各プログラムを取得して実行するようにしてもよい。また、公衆回線、インターネット、LAN、WANなどを介してコンピュータ100に接続される他のコンピュータまたはサーバ装置などに各プログラムを記憶させておき、コンピュータ100がこれらから各プログラムを取得して実行するようにしてもよい。 Note that the presentation support program 170a may not necessarily be stored in the HDD 170 or the ROM 160 from the beginning. For example, each program is stored in a “portable physical medium” such as a flexible disk inserted into the computer 100, so-called FD, CD-ROM, DVD disk, magneto-optical disk, or IC card. Then, the computer 100 may acquire and execute each program from these portable physical media. In addition, each program is stored in another computer or server device connected to the computer 100 via a public line, the Internet, a LAN, a WAN, etc., and the computer 100 acquires and executes each program from these. It may be.
   1  プレゼンテーション支援システム
   3  マイク
   5  表示装置
   7  入力装置
  10  プレゼンテーション支援装置
  11  入出力I/F部
  13  記憶部
  15  制御部
  15a 分割部
  15b 抽出部
  15c 付与部
  15d 認識部
  15e 算出部
  15f 設定部
  15g 表示制御部
DESCRIPTION OF SYMBOLS 1 Presentation support system 3 Microphone 5 Display apparatus 7 Input device 10 Presentation support apparatus 11 Input / output I / F part 13 Storage part 15 Control part 15a Division part 15b Extraction part 15c Assignment part 15d Recognition part 15e Calculation part 15f Setting part 15g Display control Part

Claims (13)

  1.  コンピュータが、
     表示時に画面単位で表示されるページを含む文書ファイルのページが分割された領域ごとに当該領域が含む文字列から第1の単語を抽出し、
     音声認識を実行し、
     所定の表示部に表示中であるページ内の領域ごとに当該領域から抽出された第1の単語と前記音声認識の結果として得られる第2の単語とから関連度を算出し、
     前記領域ごとに算出された関連度が高い領域ほど前記領域のハイライト表示を進行させる速度を高く設定し、あるいは関連度が低い領域ほど前記領域のハイライト表示を進行させる速度を低く設定し、
     前記領域ごとに設定された速度にしたがって前記ページ内のハイライト表示を制御する
     処理を実行することを特徴とするプレゼンテーション支援方法。
    Computer
    A first word is extracted from a character string included in the area for each area into which a page of a document file including a page displayed in units of screen is divided;
    Perform speech recognition,
    Calculating the degree of association from the first word extracted from the region and the second word obtained as a result of the speech recognition for each region in the page being displayed on the predetermined display unit;
    The higher the degree of relevance calculated for each area, the higher the speed at which highlight display of the area proceeds, or the lower the degree of relevance, the lower the speed at which highlight display of the area proceeds,
    A presentation support method, comprising: executing a process of controlling highlight display in the page according to a speed set for each area.
  2.  前記コンピュータが、
     前記音声認識の結果として得られる第2の単語を記憶部に登録してから所定の期間にわたって保存する処理をさらに実行し、
     前記算出する処理は、前記記憶部に記憶された第2の単語を用いて前記関連度を前記領域ごとに算出し、
     前記設定する処理は、前記領域ごとに前記関連度が算出される度に、各領域のハイライト表示の速度を設定することを特徴とする請求項1に記載のプレゼンテーション支援方法。
    The computer is
    Further executing a process of storing the second word obtained as a result of the speech recognition for a predetermined period after being registered in the storage unit;
    The calculating process calculates the relevance for each region using the second word stored in the storage unit,
    The presentation support method according to claim 1, wherein the setting processing sets a highlight display speed of each area each time the relevance is calculated for each area.
  3.  前記制御する処理は、前記関連度または前記ハイライト表示の進行度が所定の閾値以上である領域に関するハイライト表示を実行することを特徴とする請求項1または2に記載のプレゼンテーション支援方法。 3. The presentation support method according to claim 1, wherein the controlling process executes highlight display regarding an area where the relevance level or the progress level of the highlight display is a predetermined threshold value or more.
  4.  前記制御する処理は、前記関連度または前記ハイライト表示の進行度が所定の閾値未満である領域に関するハイライト表示をキャンセルすることを特徴とする請求項3に記載のプレゼンテーション支援方法。 4. The presentation support method according to claim 3, wherein the control processing cancels highlight display related to an area where the degree of association or the progress of highlight display is less than a predetermined threshold.
  5.  前記コンピュータが
     前記領域ごとに抽出される第1の単語に重みを付与する処理をさらに実行することを特徴とする請求項1に記載のプレゼンテーション支援方法。
    The presentation support method according to claim 1, wherein the computer further executes a process of assigning a weight to the first word extracted for each region.
  6.  前記付与する処理は、前記ページ内における第1の単語の出現頻度を用いて、前記第1の単語に重みを付与することを特徴とする請求項5に記載のプレゼンテーション支援方法。 The presentation support method according to claim 5, wherein the assigning process assigns a weight to the first word using an appearance frequency of the first word in the page.
  7.  前記付与する処理は、前記第1の単語のモーラ数を用いて、前記第1の単語に重みを付与することを特徴とする請求項5に記載のプレゼンテーション支援方法。 6. The presentation support method according to claim 5, wherein the assigning process assigns a weight to the first word using the number of mora of the first word.
  8.  前記算出する処理は、前記第2の単語と一致する第1の単語の数により、前記関連度を前記領域ごとに算出することを特徴とする請求項1に記載のプレゼンテーション支援方法。 The presentation support method according to claim 1, wherein the calculating process calculates the relevance for each region based on the number of first words that match the second word.
  9.  前記算出する処理は、前記領域から抽出された第1の単語の数に対する、前記第2の単語と一致する第1の単語の数の割合により、前記関連度を前記領域ごとに算出することを特徴とする請求項1に記載のプレゼンテーション支援方法。 The calculation processing includes calculating the degree of association for each region based on a ratio of the number of first words matching the second word to the number of first words extracted from the region. The presentation support method according to claim 1, wherein:
  10.  前記コンピュータが、
     前記ハイライト表示を加速する指示を受け付け、
     前記制御する処理は、前記指示を受け付けた場合に、前記ハイライト表示の進行度が最も高い領域のハイライト表示を加速させることを特徴とする請求項1に記載のプレゼンテーション支援方法。
    The computer is
    Accepts instructions to accelerate the highlighting;
    2. The presentation support method according to claim 1, wherein when the instruction is received, the controlling process accelerates highlight display in an area where the progress of highlight display is the highest. 3.
  11.  前記コンピュータが、
     前記ハイライト表示をキャンセルする指示を受け付け、
     前記制御する処理は、前記指示を受け付けた場合に、前記ハイライト表示の進行度が最も高い領域以外の領域のハイライト表示をキャンセルすることを特徴とする請求項1に記載のプレゼンテーション支援方法。
    The computer is
    Accepting an instruction to cancel the highlighting;
    2. The presentation support method according to claim 1, wherein when the instruction is received, the control processing cancels highlight display in an area other than the area where the highlight display progress is highest.
  12.  コンピュータに、
     表示時に画面単位で表示されるページを含む文書ファイルのページが分割された領域ごとに当該領域が含む文字列から第1の単語を抽出し、
     音声認識を実行し、
     所定の表示部に表示中であるページ内の領域ごとに当該領域から抽出された第1の単語と前記音声認識の結果として得られる第2の単語とから関連度を算出し、
     前記領域ごとに算出された関連度が高い領域ほど前記領域のハイライト表示を進行させる速度を高く設定し、あるいは関連度が低い領域ほど前記領域のハイライト表示を進行させる速度を低く設定し、
     前記領域ごとに設定された速度にしたがって前記ページ内のハイライト表示を制御する
     処理を実行させることを特徴とするプレゼンテーション支援プログラム。
    On the computer,
    A first word is extracted from a character string included in the area for each area into which a page of a document file including a page displayed in units of screen is divided;
    Perform speech recognition,
    Calculating the degree of association from the first word extracted from the region and the second word obtained as a result of the speech recognition for each region in the page being displayed on the predetermined display unit;
    The higher the degree of relevance calculated for each area, the higher the speed at which highlight display of the area proceeds, or the lower the degree of relevance, the lower the speed at which highlight display of the area proceeds,
    A presentation support program for executing a process of controlling highlight display in the page according to a speed set for each area.
  13.  表示時に画面単位で表示されるページを含む文書ファイルのページが分割された領域ごとに当該領域が含む文字列から第1の単語を抽出する抽出部と、
     音声認識を実行する認識部と、
     所定の表示部に表示中であるページ内の領域ごとに当該領域から抽出された第1の単語と前記音声認識の結果として得られる第2の単語とから関連度を算出する算出部と、
     前記領域ごとに算出された関連度が高い領域ほど前記領域のハイライト表示を進行させる速度を高く設定し、あるいは関連度が低い領域ほど前記領域のハイライト表示を進行させる速度を低く設定する設定部と、
     前記領域ごとに設定された速度にしたがって前記ページ内のハイライト表示を制御する表示制御部と
     を有することを特徴とするプレゼンテーション支援装置。
    An extraction unit that extracts a first word from a character string included in each area of the divided page of a document file including a page displayed in screen units when displayed;
    A recognition unit that performs speech recognition;
    A calculation unit that calculates the degree of association from the first word extracted from the region and the second word obtained as a result of the speech recognition for each region in the page being displayed on the predetermined display unit;
    A setting for setting a higher speed for the highlight display of the area to be set higher for an area having a higher relevance calculated for each area, or a setting for setting a lower speed for the highlight display of the area for a lower relevance level. And
    A display control unit that controls highlight display in the page according to a speed set for each region.
PCT/JP2014/078533 2014-10-27 2014-10-27 Presentation support method, presentation support program, and presentation support device WO2016067348A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/JP2014/078533 WO2016067348A1 (en) 2014-10-27 2014-10-27 Presentation support method, presentation support program, and presentation support device
JP2016556070A JP6304396B2 (en) 2014-10-27 2014-10-27 Presentation support method, presentation support program, and presentation support apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2014/078533 WO2016067348A1 (en) 2014-10-27 2014-10-27 Presentation support method, presentation support program, and presentation support device

Publications (1)

Publication Number Publication Date
WO2016067348A1 true WO2016067348A1 (en) 2016-05-06

Family

ID=55856743

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/078533 WO2016067348A1 (en) 2014-10-27 2014-10-27 Presentation support method, presentation support program, and presentation support device

Country Status (2)

Country Link
JP (1) JP6304396B2 (en)
WO (1) WO2016067348A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017211546A (en) * 2016-05-26 2017-11-30 富士通株式会社 Idle talk detection device, image display system, idle talk detection method, and idle talk detection program
JP2019124750A (en) * 2018-01-12 2019-07-25 株式会社日立ソリューションズ Method of displaying presentation material

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002023716A (en) * 2000-07-05 2002-01-25 Pfu Ltd Presentation system and recording medium
JP2011065467A (en) * 2009-09-17 2011-03-31 Sharp Corp Conference relay device and computer program
JP2012185567A (en) * 2011-03-03 2012-09-27 Fujitsu Ltd Display control device, display control method and display control program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002023716A (en) * 2000-07-05 2002-01-25 Pfu Ltd Presentation system and recording medium
JP2011065467A (en) * 2009-09-17 2011-03-31 Sharp Corp Conference relay device and computer program
JP2012185567A (en) * 2011-03-03 2012-09-27 Fujitsu Ltd Display control device, display control method and display control program

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017211546A (en) * 2016-05-26 2017-11-30 富士通株式会社 Idle talk detection device, image display system, idle talk detection method, and idle talk detection program
JP2019124750A (en) * 2018-01-12 2019-07-25 株式会社日立ソリューションズ Method of displaying presentation material

Also Published As

Publication number Publication date
JPWO2016067348A1 (en) 2017-05-25
JP6304396B2 (en) 2018-04-04

Similar Documents

Publication Publication Date Title
USRE49762E1 (en) Method and device for performing voice recognition using grammar model
US10114809B2 (en) Method and apparatus for phonetically annotating text
JP6432405B2 (en) Presentation support device, presentation support method, and presentation support program
JP7111682B2 (en) Speech command matching during testing of a speech-assisted application prototype for languages using non-phonetic writing systems
US9548052B2 (en) Ebook interaction using speech recognition
US9196253B2 (en) Information processing apparatus for associating speaker identification information to speech data
EP2849054A1 (en) Apparatus and method for selecting a control object by voice recognition
JP2019528470A (en) Acoustic model training using corrected terms
JP2020003926A (en) Interaction system control method, interaction system and program
JP2009042968A (en) Information selection system, information selection method, and program for information selection
US20220121712A1 (en) Interactive representation of content for relevance detection and review
JP6304396B2 (en) Presentation support method, presentation support program, and presentation support apparatus
JP2018005011A (en) Presentation support device, presentation support system, presentation support method and presentation support program
US20130179165A1 (en) Dynamic presentation aid
US20210165540A1 (en) Information processing device, information processing method, and program
CN110890095A (en) Voice detection method, recommendation method, device, storage medium and electronic equipment
JP6372577B2 (en) Presentation support method, presentation support program, and presentation support apparatus
TW201506685A (en) Apparatus and method for selecting a control object by voice recognition
JP2018045193A (en) Communication terminal, voice conversion method, and program
JP6399221B2 (en) Presentation support device, presentation support method, and presentation support program
JPWO2020116001A1 (en) Information processing device and information processing method
JP6350682B2 (en) Presentation support device, presentation support method, and presentation support program
JP6651985B2 (en) Chat detection apparatus, image display system, chat detection method, and chat detection program
JP6567372B2 (en) Editing support apparatus, editing support method, and program
JP2022163217A (en) Content editing support method and system based on real time generation of synthetic sound for video content

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14904876

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2016556070

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14904876

Country of ref document: EP

Kind code of ref document: A1