WO2006049249A1 - ディジタル映像再生装置 - Google Patents
ディジタル映像再生装置 Download PDFInfo
- Publication number
- WO2006049249A1 WO2006049249A1 PCT/JP2005/020300 JP2005020300W WO2006049249A1 WO 2006049249 A1 WO2006049249 A1 WO 2006049249A1 JP 2005020300 W JP2005020300 W JP 2005020300W WO 2006049249 A1 WO2006049249 A1 WO 2006049249A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- character information
- digital video
- dictionary
- index
- code
- Prior art date
Links
- 239000000284 extract Substances 0.000 abstract 1
- 230000006870 function Effects 0.000 description 12
- 238000000034 method Methods 0.000 description 9
- 239000000872 buffer Substances 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7844—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using original textual content or text extracted from visual content or transcript of audio data
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42203—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/426—Internal components of the client ; Characteristics thereof
- H04N21/42646—Internal components of the client ; Characteristics thereof for reading from or writing on a non-volatile solid state storage medium, e.g. DVD, CD-ROM
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/432—Content retrieval operation from a local storage medium, e.g. hard-disk
- H04N21/4325—Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
- H04N21/4394—Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/84—Television signal recording using optical recording
- H04N5/85—Television signal recording using optical recording on discs or drums
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/87—Regeneration of colour television signals
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2562—DVDs [digital versatile discs]; Digital video discs; MMCDs; HDCDs
Definitions
- the present invention relates to a digital video reproduction apparatus for reproducing digital video content.
- a list of chapters is displayed as an index image at the start of content playback, and a key provided on the device main body or remote control is operated. To select chapters. As a result, the user can reproduce the content with the desired chapter power.
- some recent BS digital no-vision televisions have a voice recognition function, and the channel can be switched by inputting a broadcast station name, a channel name, or the like by voice.
- Patent Document 1 JP 11 41538 A
- an object of the present invention is to provide a digital video reproduction apparatus capable of designating a reproduction position of digital video content by voice.
- a first aspect of the present invention is a digital video reproduction device (1) for reproducing digital video content, which can be used for voice recognition processing based on a signal of the digital video content.
- a dictionary creation means (102) for creating a dictionary, a speech recognition means (104) for performing speech recognition processing using the dictionary created by the dictionary creation means, and a speech recognition result of the speech recognition means Playback control means (105) for controlling playback of digital video content based on the playback control.
- a second aspect of the present invention is the playback position selection for causing the user to select the playback position of the digital video content included in the video signal of the digital video content in the first aspect.
- Character information recognizing means (101) for recognizing character information from an image (FIGS. 2, 4, and 5) is further provided, and the dictionary creating means uses the character information recognized by the character information recognizing means.
- an index dictionary is created for allowing the user to select a playback position by voice.
- the reproduction position selection image when the reproduction position selection image is displayed on a display device, the reproduction position selection image is registered in the index dictionary based on the reproduction position selection image.
- the image display device further comprises index display means (106) for highlighting the character string portion corresponding to the written character information based on the recognition result of the character information recognition means.
- the character information recognized by the character information recognition means is stored in a nonvolatile manner as chapter information of digital video content used in the past.
- Non-volatile storage means (107) is further provided.
- a fifth aspect of the present invention is the digital video content according to the fourth aspect, wherein the non-volatile storage means converts the character information recognized by the character information recognition means to the character information corresponding to the character information. Is stored together with history information indicating whether or not this part has been reproduced in the past.
- a sixth aspect of the present invention is that, in the fifth aspect, the nonvolatile storage means corresponds to the character information, together with the history information, the character information recognized by the character information recognition means.
- the digital video content part is recorded together with the date and time when it was reproduced.
- a seventh aspect of the present invention is that, in the second aspect, the dictionary creation unit is configured such that the number of characters of the character information recognized by the character information recognition unit exceeds a predetermined threshold.
- the index dictionary is created by using a reduced number of characters of the character information.
- a language determining means (101) for determining whether or not the character information recognized by the character information recognizing means is Japanese, and the character information If the character information is not recognized by the recognition means or if the character information recognized by the character information recognition means is in English, the reproduction position selection image is determined according to a predetermined rule.
- Code addition means (103) for adding a code
- index display means (106) for causing the display device (5) to display the reproduction position selection image added with the code by the code addition means;
- the creation means uses the English character information recognized by the character information recognition means and the code added to the reproduction position selection image by the code addition means to select the reproduction position by voice to the user.
- An index dictionary is created for the purpose.
- a ninth aspect of the present invention is the playback position selection for causing the user to select the playback position of the digital video content included in the video signal of the digital video content in the first aspect.
- a code adding means (103) for adding a code to an image (FIGS. 2, 4, and 5) according to a predetermined rule and a reproduction position selection image to which a code is added by the code adding means are displayed.
- the dictionary creating means is for causing the user to select a playback position of the digital video content included in the signal of the digital video content.
- an index dictionary is created to allow the user to select the playback position by voice.
- the user can view chapter information of a medium such as a DVD that has been used in the past without setting the medium.
- the user can check which chapter has been played back in the past and which chapter has not been played back.
- a chapter viewed in the past can be selected based on a past reproduction date and time.
- FIG. 1 is a diagram showing a configuration of a digital video playback apparatus according to an embodiment of the present invention.
- FIG. 2 is a first example of an index image.
- FIG. 3 is a diagram showing functions of the system controller 4.
- FIG. 4 is a second example of the index image.
- FIG. 5 is a third example of an index image.
- FIG. 6 is a first example of an index dictionary.
- FIG. 7 shows a second example of the index dictionary.
- FIG. 8 shows a first example of an index image covered by the index display means 106.
- FIG. 9 is a second example of an index image covered by the index display means 106.
- FIG. 10 is a third example of the index image cached by the index display means 106.
- FIG. 11 shows a third example of the index dictionary.
- FIG. 12 is an example of information stored in the storage unit 107 together with the index dictionary.
- the digital video playback device 1 includes a DVD drive 2 that can play back DVD video, a buffer 3 that temporarily stores video played back by the DVD drive 2, and digital video playback.
- a system controller 4 that controls the entire device 1, a display device 5 that displays video, a button 6 that the user presses to start voice input, a microphone 8 that inputs voice, and a microphone ADC (Analog to Digital Converter) 7 that converts analog audio signals from 8 into digital audio signals.
- the DVD drive 2 and microphone 8 are not necessarily built in the digital video playback device 1.
- DVD video When a user records a DVD on which digital video content (here, referred to as a DVD video) is inserted into the DVD drive 2, an indented image as shown in FIG. Displayed in 5.
- DVD video has multiple Selection buttons 21 to 25 for selecting each chapter are displayed on the index image.
- the selection button is not necessarily limited to the button shown in FIG. 2, and when the outline of the selection button is not displayed (that is, the same color as the background), or a reduced image of the first video of the chapter is used as the selection button. There is also a case.
- 20 in FIG. 2 is not a selection button but a part of the background display.
- Figure 2 shows an example where a DVD video consists of five chapters. The user can play a DVD video from the chapter by selecting a desired chapter listing power using an input device (not shown).
- Selection names 21 to 25 The chapter names (such as “an entrance ceremony” and “an excursion”) are not included in the DVD video as text information (character code).
- the index image including the buttons 21 to 25 is merely stored in the DVD video as image information. Information for distinguishing which area of the index image corresponds to which selection button is stored in the DVD video.
- System controller 4 is realized by a combination of hardware such as a microcomputer, DSP, and memory, and software stored in ROM, etc., and performs character recognition processing in image information, voice recognition processing, and chapter playback processing.
- the system controller 4 functions as character information recognition means 101, index dictionary creation means 102, code addition means 103, voice recognition means 104, reproduction control means 105, index display means 106, and storage means 107.
- each means will be described in detail.
- the character information recognition means 101 recognizes characters from the selection buttons 21 to 25 (that is, the image area for selecting each chapter) in FIG. 2 included in the index image read from the DVD drive 2. Character information is extracted by processing.
- the character information recognition unit 101 has a function of determining whether or not the character information extracted by the character recognition process is English. Here, depending on whether or not the English character information is extracted, The processing procedure after this changes.
- the explanation is based on the assumption that the digital video playback device 1 is used in the English-speaking area. However, for example, the digital video playback device 1 is used in the Japanese-speaking area. If this is the case, the subsequent processing procedure will change depending on whether or not Japanese character information is extracted. That is, more generally, the character information recognition unit 101 only needs to have a function of determining whether or not the character information extracted by the character recognition process is a specific language.
- the character information recognizing means 101 When English character information is extracted by the character information recognizing means 101, the character information recognizing means 101 outputs the extracted character information to the index dictionary creating means 102 as text data. On the other hand, when the English character information is not extracted (that is, when no character information is extracted or when non-English character information is extracted), the character information recognition means 101 reads the code information. The sign adding process is requested to the adding means 103. For example, when the chapter name is displayed in Japanese on the selection button as shown in FIG. 4 or when the selection button is a reduced image (summail) of the first video of the chapter as shown in FIG. The character information recognizing unit 101 requests the code adding unit 103 for the code adding process.
- the sign adding means 103 assigns a code such as a number to a selection button that does not include English character information in accordance with a predetermined rule, and a text corresponding to the code assigned to each selection button.
- the data is output to the index dictionary creation means 102.
- the index dictionary creation means 102 is text data ("entrance ceremony”, “excursion”, etc.) output from the character information recognition means 101 or text data ("1", “excursion”, etc.) output from the sign addition means 103. Based on “2” and the like, an index dictionary that can be used for speech recognition processing by the speech recognition means 104 is created. Hereinafter, the processing of the index dictionary creating means 102 will be described more specifically.
- the index dictionary creation means 102 first checks whether there is any identical text data in the text data output from the character information recognition means 101 or the signed calorie means 103, and there is identical text data. In such a case, numbers (more precisely, text data corresponding to the numbers) are added after the text data so that the text data can be distinguished. For example, the text data input to the index dictionary creation means 102 If there are two “an excursion”, convert one to "excursion 'one" and the other to "excursion'two".
- the index dictionary creation means 102 uses these text data to create an index dictionary for speech recognition processing. Created and stored in the storage means 107.
- the storage unit 107 may be a non-volatile storage unit that retains the stored contents even when the power is turned off, or may be a volatile storage unit that loses the stored contents when the power is turned off.
- the display position in the index image of the corresponding selection button is also recorded in association with the text data.
- FIG. 6 shows the contents of the index dictionary created based on the index image of FIG.
- FIG. 7 shows the contents of the index dictionary created based on the index image of FIG. 4 or FIG.
- the index dictionary creating means 102 has a plurality of the same text data as described above, and therefore, for those in which numbers etc. are added to the text data, those after addition of the numbers etc. are converted into text. At the same time as recording as data, information (flag) indicating that the recorded text data is obtained by adding numbers or the like to the original text data is recorded in the index dictionary in association with the text data.
- the index display means 106 performs display control for presenting the user with words to be uttered in order to select each selection button included in the index image. More specifically, for the character string recognized as English by the character information recognition means 101, the display position of the character string in the index image is acquired from the character information recognition means 101 as necessary, and the character string is recognized as shown in FIG. Underline 51-55 to the character string as shown in the figure, surround the character string with a frame, and utterance marks 61-65 indicating that the character string is to be uttered near the character string The character string is emphasized by adding it.
- the index dictionary creating means 102 for a character string in which a number or the like is added to the text data, a number (71, 72) is additionally displayed after the character string as shown in FIG.
- the codes (81 to 85) assigned to the character strings by the code addition means 103 are additionally displayed as shown in FIG. To do.
- voice recognition start signal S 1 is output from button 6.
- the voice recognition means 104 receives the voice recognition start signal S1
- the voice recognition means 104 performs voice recognition processing using the voice data D1 fetched through the ADC 7 and the index dictionary stored in the storage means 107.
- the speech recognition means 104 displays a selection button corresponding to the text data. The position information is acquired and output to the reproduction control means 105.
- the reproduction control means 105 specifies a chapter from the display position information received from the voice recognition means 104, and starts reproduction of the chapter.
- the character information recognition means 101 has a function of determining whether or not the character information extracted by the character recognition process is Japanese, and is included in the index image in Japanese.
- the index dictionary creating means 102 first starts from the character information recognizing means 101 or the sign adding means 103.
- the output text data is converted into katakana data (such as “Two Yugakushiki”, “Yensoku”, “Ichi”, “Two”).
- katakana data such as “Two Yugakushiki”, “Yensoku”, “Ichi”, “Two”.
- the text data output from the character information recognition unit 101 and the code addition unit 103 and the display position of the corresponding selection button in the index image are also recorded in association with the katakana data.
- Fig. 11 shows the contents of the index dictionary created based on the index image shown in Fig. 4.
- the user selects an arbitrary selection button by voice from the selection buttons included in the index image, and views a chapter corresponding to the selection button. Can start.
- chapters can be selected by voice, which is particularly useful as an in-vehicle DVD playback device.
- a navigation apparatus having a digital video content playback function may be used.
- non-volatile storage means is used as the storage means 107, and the index dictionary created by the index dictionary creation means 102 is retained even after the digital video reproduction apparatus 1 is turned off.
- the data stored in the storage means 107 can be used when the digital video playback apparatus 1 is turned off and then turned on again, the usability of the digital video playback apparatus 1 can be improved. it can.
- an application example using a non-volatile storage unit as the storage unit 107 will be described.
- the index dictionary created by the index dictionary creation means 102 is stored in association with DVD video titles, DVD-specific information, and other additional information. Recorded in 107.
- Information specific to a DVD includes, for example, the size of data recorded on the DVD, the video time, and the number of chapters.
- the playback control means 105 stores the playback date and time in the storage means 107 as playback history information in association with the display position information of the selection button corresponding to the chapter.
- the reproduction control unit 105 stores the time counter value of the video at the time of the stop in the storage unit 107 as the reproduction history information in association with the reproduction date and time.
- the system controller 4 first checks whether a DVD is inserted in the DV drive 2! If a DVD has been inserted into the DVD drive 2, the system controller 4 refers to the information stored in the storage means 107 as shown in FIG. Check if the DVD has been inserted in the past. If the DVD currently inserted in DVD drive 2 is a DVD that has been inserted in the past, voice recognition processing is performed using the index dictionary created in the past, and it is inserted for the first time. In the case of a DVD, the index image power is also created in an index dictionary and stored in the storage means 107, and voice recognition processing is executed using this index dictionary. This saves you the trouble of creating a new index dictionary for DVDs that have been used in the past, so it takes less time to insert a DVD into the DV D drive 2 and select a chapter by force. be able to.
- each chapter Information indicating whether or not the force has been reproduced in the past can be added to the index image by the index display means 106 and the force can be displayed on the display device 5.
- the case of playing back digital video content recorded on a DVD has been described as an example.
- the present invention is not limited to this, and the digital video recorded on another recording medium is used.
- the present invention can be similarly applied to the case of playing back content, the case of playing back digital video content supplied through a communication line, and the case of playing back digital video content supplied by broadcast waves. .
- the character information recognition means 101 does not perform character recognition processing.
- the index dictionary creating means 102 should create an index dictionary using the text information.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Television Signal Processing For Recording (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
- Indexing, Searching, Synchronizing, And The Amount Of Synchronization Travel Of Record Carriers (AREA)
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/666,678 US7953602B2 (en) | 2004-11-08 | 2005-11-04 | Digital video reproducing apparatus for recognizing and reproducing a digital video content |
EP05805411A EP1811776B1 (en) | 2004-11-08 | 2005-11-04 | Digital video reproduction apparatus |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004-323970 | 2004-11-08 | ||
JP2004323970A JP3824168B2 (ja) | 2004-11-08 | 2004-11-08 | ディジタル映像再生装置 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006049249A1 true WO2006049249A1 (ja) | 2006-05-11 |
Family
ID=36319249
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/020300 WO2006049249A1 (ja) | 2004-11-08 | 2005-11-04 | ディジタル映像再生装置 |
Country Status (5)
Country | Link |
---|---|
US (1) | US7953602B2 (ja) |
EP (1) | EP1811776B1 (ja) |
JP (1) | JP3824168B2 (ja) |
CN (1) | CN100536552C (ja) |
WO (1) | WO2006049249A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101606384B (zh) * | 2007-12-07 | 2011-01-12 | 索尼株式会社 | 图像处理装置、运动图像再现装置及其处理方法 |
US7929764B2 (en) | 2007-06-15 | 2011-04-19 | Microsoft Corporation | Identifying character information in media content |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101558640A (zh) * | 2006-12-14 | 2009-10-14 | 皇家飞利浦电子股份有限公司 | 再现和显示信息的系统和方法 |
CN101472082B (zh) * | 2007-12-25 | 2012-07-25 | 新奥特(北京)视频技术有限公司 | 一种场记系统和方法 |
KR20120080069A (ko) * | 2011-01-06 | 2012-07-16 | 삼성전자주식회사 | 디스플레이 장치 및 그 음성 제어 방법 |
KR101295711B1 (ko) * | 2011-02-15 | 2013-08-16 | 주식회사 팬택 | 음성 인식으로 애플리케이션의 실행 가능한 이동통신 단말 장치 및 방법 |
CN105340003B (zh) * | 2013-06-20 | 2019-04-05 | 株式会社东芝 | 语音合成字典创建装置以及语音合成字典创建方法 |
JP6739907B2 (ja) * | 2015-06-18 | 2020-08-12 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 機器特定方法、機器特定装置及びプログラム |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001101805A (ja) * | 1999-10-04 | 2001-04-13 | L & G Kikaku:Kk | 音声認識自在な記録再生装置 |
JP2002041081A (ja) * | 2000-07-28 | 2002-02-08 | Sharp Corp | 音声認識用辞書作成装置および音声認識用辞書作成方法、音声認識装置、携帯端末器、並びに、プログラム記録媒体 |
JP2003230094A (ja) * | 2002-02-06 | 2003-08-15 | Nec Corp | チャプター作成装置及びデータ再生装置及びその方法並びにプログラム |
JP2006018336A (ja) * | 2004-06-30 | 2006-01-19 | Toshiba Corp | メタデータ生成装置および方法、メタデータ生成プログラム |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5031206A (en) * | 1987-11-30 | 1991-07-09 | Fon-Ex, Inc. | Method and apparatus for identifying words entered on DTMF pushbuttons |
JPH02253369A (ja) * | 1989-03-28 | 1990-10-12 | Canon Inc | 電子辞書 |
US5890123A (en) * | 1995-06-05 | 1999-03-30 | Lucent Technologies, Inc. | System and method for voice controlled video screen display |
US5809471A (en) | 1996-03-07 | 1998-09-15 | Ibm Corporation | Retrieval of additional information not found in interactive TV or telephony signal by application using dynamically extracted vocabulary |
KR100217179B1 (ko) * | 1997-03-25 | 1999-09-01 | 윤종용 | 디지털비디오디스크 재생장치에서의 메뉴재생방법 |
JPH1141538A (ja) | 1997-07-17 | 1999-02-12 | Nec Home Electron Ltd | 音声認識文字表示装置 |
DE69712485T2 (de) * | 1997-10-23 | 2002-12-12 | Sony Int Europe Gmbh | Sprachschnittstelle für ein Hausnetzwerk |
US6408128B1 (en) * | 1998-11-12 | 2002-06-18 | Max Abecassis | Replaying with supplementary information a segment of a video |
US6314398B1 (en) * | 1999-03-01 | 2001-11-06 | Matsushita Electric Industrial Co., Ltd. | Apparatus and method using speech understanding for automatic channel selection in interactive television |
US6643620B1 (en) * | 1999-03-15 | 2003-11-04 | Matsushita Electric Industrial Co., Ltd. | Voice activated controller for recording and retrieving audio/video programs |
JP2001028722A (ja) * | 1999-07-13 | 2001-01-30 | Matsushita Electric Ind Co Ltd | 動画像管理装置及び動画像管理システム |
EP1189206B1 (en) * | 2000-09-19 | 2006-05-31 | Thomson Licensing | Voice control of electronic devices |
JP2004521392A (ja) * | 2001-06-22 | 2004-07-15 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 発話制御手段と、発話制御手段の機能をテストするためのテスト手段とを有する装置 |
US20030069734A1 (en) * | 2001-10-05 | 2003-04-10 | Everhart Charles Allen | Technique for active voice recognition grammar adaptation for dynamic multimedia application |
US20040177317A1 (en) * | 2003-03-07 | 2004-09-09 | John Bradstreet | Closed caption navigation |
US7437296B2 (en) * | 2003-03-13 | 2008-10-14 | Matsushita Electric Industrial Co., Ltd. | Speech recognition dictionary creation apparatus and information search apparatus |
US7508984B2 (en) * | 2003-07-31 | 2009-03-24 | Ricoh Company, Ltd. | Language recognition method, system and software |
US20060075429A1 (en) * | 2004-04-30 | 2006-04-06 | Vulcan Inc. | Voice control of television-related information |
US20070143117A1 (en) * | 2005-12-21 | 2007-06-21 | Conley Kevin M | Voice controlled portable memory storage device |
-
2004
- 2004-11-08 JP JP2004323970A patent/JP3824168B2/ja active Active
-
2005
- 2005-11-04 WO PCT/JP2005/020300 patent/WO2006049249A1/ja active Application Filing
- 2005-11-04 US US11/666,678 patent/US7953602B2/en active Active
- 2005-11-04 CN CNB2005800381479A patent/CN100536552C/zh active Active
- 2005-11-04 EP EP05805411A patent/EP1811776B1/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001101805A (ja) * | 1999-10-04 | 2001-04-13 | L & G Kikaku:Kk | 音声認識自在な記録再生装置 |
JP2002041081A (ja) * | 2000-07-28 | 2002-02-08 | Sharp Corp | 音声認識用辞書作成装置および音声認識用辞書作成方法、音声認識装置、携帯端末器、並びに、プログラム記録媒体 |
JP2003230094A (ja) * | 2002-02-06 | 2003-08-15 | Nec Corp | チャプター作成装置及びデータ再生装置及びその方法並びにプログラム |
JP2006018336A (ja) * | 2004-06-30 | 2006-01-19 | Toshiba Corp | メタデータ生成装置および方法、メタデータ生成プログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP1811776A4 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7929764B2 (en) | 2007-06-15 | 2011-04-19 | Microsoft Corporation | Identifying character information in media content |
CN101606384B (zh) * | 2007-12-07 | 2011-01-12 | 索尼株式会社 | 图像处理装置、运动图像再现装置及其处理方法 |
Also Published As
Publication number | Publication date |
---|---|
JP3824168B2 (ja) | 2006-09-20 |
US7953602B2 (en) | 2011-05-31 |
EP1811776A1 (en) | 2007-07-25 |
EP1811776B1 (en) | 2011-06-01 |
CN100536552C (zh) | 2009-09-02 |
CN101057497A (zh) | 2007-10-17 |
JP2006134506A (ja) | 2006-05-25 |
US20080208576A1 (en) | 2008-08-28 |
EP1811776A4 (en) | 2009-10-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2006049249A1 (ja) | ディジタル映像再生装置 | |
JP3248981B2 (ja) | 計算機 | |
US5999695A (en) | Multilingual recording medium and reproduction apparatus | |
US20090119108A1 (en) | Audio-book playback method and apparatus | |
US20040102955A1 (en) | DVD driver for language study and method of processing audio streams thereof | |
JP2007503747A (ja) | リアルタイムのメディア辞書 | |
US20070016846A1 (en) | Apparatus and method for reproducing text file in digital video device | |
US20070087312A1 (en) | Method for separating sentences in audio-video display system | |
JP3036430B2 (ja) | 文章読み上げ装置 | |
JP2004325905A (ja) | 外国語学習装置および外国語学習プログラム | |
JP4970392B2 (ja) | 音声の録音と再生方法及び該方法を用いる電子辞書 | |
JP4994182B2 (ja) | オーディオ装置 | |
JP2007149163A (ja) | コンテンツ再生装置 | |
JP7009338B2 (ja) | 情報処理装置、情報処理システム、および映像装置 | |
KR101074018B1 (ko) | 학습 보조 장치 | |
JP5295699B2 (ja) | 車載用オーディオ装置 | |
KR970071761A (ko) | 비디오 콤팩트 디스크 플레이어에서의 어학 학습을 위한 재생 방법 | |
KR20010054297A (ko) | 디지탈동영상 맵핑방법 및 이것을 이용한 반복재생방법 | |
KR20050106246A (ko) | 엠펙 플레이어에 있어서 데이터 검색 방법 | |
JP2001312288A (ja) | 音楽データ処理装置 | |
KR20040062317A (ko) | 오디오 데이터 파일의 부가 정보 음성 안내방법 | |
KR970071662A (ko) | 비디오 콤팩트 디스크 플레이어에서의 어학 학습을 위한 재생방법 | |
JP2006208514A (ja) | 二ヶ国語で入力し得るカラオケ選曲用キーボードを備えたカラオケ装置及びその装置における選曲方法 | |
US20080126419A1 (en) | Method for providing file information according to selection of language and file reproducing apparatus using the same | |
KR970071661A (ko) | 비디오 콤팩트 디스크 플레이어에서의 어학 학습을 위한 재생 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS KE KG KM KN KP KR KZ LC LK LR LS LT LU LV LY MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU LV MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 11666678 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005805411 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200580038147.9 Country of ref document: CN |
|
WWP | Wipo information: published in national office |
Ref document number: 2005805411 Country of ref document: EP |