EP2945052B1 - Voice recognition device, voice recognition program, and voice recognition method - Google Patents
Voice recognition device, voice recognition program, and voice recognition method Download PDFInfo
- Publication number
- EP2945052B1 EP2945052B1 EP13870530.6A EP13870530A EP2945052B1 EP 2945052 B1 EP2945052 B1 EP 2945052B1 EP 13870530 A EP13870530 A EP 13870530A EP 2945052 B1 EP2945052 B1 EP 2945052B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- screen
- option
- voice
- options
- speech recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims description 16
- 238000012545 processing Methods 0.000 claims description 61
- 230000007704 transition Effects 0.000 claims description 28
- 238000010586 diagram Methods 0.000 description 36
- 238000004891 communication Methods 0.000 description 8
- 230000008859 change Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 5
- 238000013459 approach Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3605—Destination input or retrieval
- G01C21/3608—Destination input or retrieval using speech input, e.g. using speech recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0412—Digitisers structurally integrated in a display
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3626—Details of the output of route guidance instructions
- G01C21/3629—Guidance using speech or audio output, e.g. text-to-speech
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Definitions
- the present invention relates to a technology for a speech recognition device.
- the present invention claims priority to Japanese Patent Application No. 2013-1373 filed on January 8, 2013
- an electronic device including: detection means for detecting a state relating to the electronic device; and determination means for determining based on at least a part of the detected state whether or not to start speech recognition or whether or not to end the speech recognition, in which it is determined based on a determination result thereof whether to start or end the speech recognition, the speech recognition is conducted, and the electronic device is caused to conduct a predetermined operation based on a recognition result thereof.
- detection means for detecting a state relating to the electronic device
- determination means for determining based on at least a part of the detected state whether or not to start speech recognition or whether or not to end the speech recognition, in which it is determined based on a determination result thereof whether to start or end the speech recognition, the speech recognition is conducted, and the electronic device is caused to conduct a predetermined operation based on a recognition result thereof.
- JP 2005-053331 A there is furthermore shown an approach for selection of options from a screen in a vehicle.
- US 2006/0224260 A1 discloses an approach where small segments of songs from a music collection are played to a user, one by one.
- the option reading unit may further conduct, when the option received by the voice instruction reception unit designates a narrowing-down condition for narrowing down the options on a transition destination screen to which a transition is made from the predetermined screen, the voice outputs of the options narrowed down by the narrowing-down condition on the transition destination screen.
- the option reading unit may conduct, when the option received by the voice instruction reception unit designates a determination condition for determining a processing target for predetermined processing, the predetermined processing for the processing target identified by the determination condition.
- the option reading unit may conduct the voice output by excluding the option that has been displayed among the options on the predetermined screen .
- the speech recognition device may further include a history creation unit for updating the number of selected times within the selection history information for the option for which the instruction has been received by the touch instruction reception unit and the voice instruction reception unit.
- the speech recognition device may be mounted to a moving object, and the speech recognition device may further include an input reception switching unit for restricting, when the moving object starts moving at a predetermined speed or faster, the reception of the instruction conducted by the touch instruction reception unit.
- a speech recognition program for causing a computer to execute a speech recognition procedure according to claim 7.
- a speech recognition method to be performed by a speech recognition device according to claim 8.
- FIG. 1 is an overall configuration diagram of the navigation device 100.
- the navigation device 100 is a so-called navigation device capable of displaying map information and presenting a spot indicating a present location of the navigation device 100 and information that guides a user along a route to a set destination.
- the navigation device 100 includes an arithmetic processing unit 1, a display 2, a storage device 3, a voice input/output device 4 (including a microphone 41 as a voice input device and a speaker 42 as a voice output device), an input device 5, a read only memory (ROM) device 6, a vehicle speed sensor 7, a gyro sensor 8, a global positioning system (GPS) receiver 9, an FM multiplex broadcast receiver 10, a beacon receiver 11, and an in-vehicle network communication device 12.
- ROM read only memory
- GPS global positioning system
- the arithmetic processing unit 1 is a main unit for conducting various kinds of processing. For example, the arithmetic processing unit 1 calculates the present location based on information output from the respective sensors 7 and 8, the GPS receiver 9, the FM multiplex broadcast receiver 10, and the like. Further, based on information on the obtained present location, map data necessary for display is read from the storage device 3 or the ROM device 6.
- the arithmetic processing unit 1 transforms the read map data into graphics, and displays the graphics on the display 2 with the graphics overlaid with a mark indicating the present location.
- the map data or the like stored in the storage device 3 or the ROM device 6 is used to search for a recommended route that is an optimal route, which connects the present location or a point of departure specified by a user to the destination (or transit point or drop-by point).
- the speaker 42 or the display 2 is used to guide the user.
- the arithmetic processing unit 1 of the navigation device 100 the respective devices are connected to one another through a bus 25.
- the arithmetic processing unit 1 includes: a central processing unit (CPU) 21 for executing various kinds of processing such as an numerical value arithmetic operation and control of each device; a random access memory (RAM) 22 for storing the map data read from the storage device 3, arithmetic operation data, and the like; a ROM 23 for storing a program and data; and an interface (I/F) 24 for connection between various kinds of hardware and the arithmetic processing unit 1.
- CPU central processing unit
- RAM random access memory
- ROM 23 for storing a program and data
- I/F interface
- the display 2 is a unit for displaying graphics information generated by the arithmetic processing unit 1 or the like.
- the display2 is formed of a liquid crystal display, an organic EL display, or the like.
- the storage device 3 is formed of a storage medium, which is at least readable and writable, such as a hard disk drive (HDD) or a nonvolatile memory card.
- a storage medium which is at least readable and writable, such as a hard disk drive (HDD) or a nonvolatile memory card.
- This storage medium stores: a link table 200, which is the map data (including link data on a link forming a road on a map) necessary for a general route search device; a screen definition table 300, which is definition information on a screen displayed on the navigation device 100; and a selection history table 400, which associates the number of times that an option serving as a candidate to be selected on each screen has been actually selected with each option in units of screens.
- the storage medium of the storage device 3 stores: one, two, or more song files; and information relating to a playlist, which defines identification information identifying a plurality of song files to be played back and a playback order of the song files.
- each song file includes, as meta information, attribute information such as information identifying an artist of a song, a composer thereof, a genre thereof, and an album name containing the song.
- FIG. 2 is a diagram for showing a configuration of the link table 200.
- the link table 200 For each identification code (mesh ID) 201 of a mesh that is an area segmented on the map, the link table 200 includes link data 202 on each link forming a road included in a mesh area thereof.
- the link data 202 For each link ID 211 serving as the identifier of the link, the link data 202 includes coordinate information 222 on two nodes (start node and end node) forming the link, a road type 223 indicating a type of road including the link, a link length 224 indicating a length of the link, a link travel time 225 stored in advance, a start connection link and an end connection link 226, and a speed limit 227 indicating a speed limit of the road including the link.
- the start connection link and the end connection link 226 are information identifying a start connection link serving as a link connecting to the start node of the link and an end connection link serving as a link connecting to the end node of the link.
- the two nodes forming the link an upward direction and a downward direction of the same road are managed as mutually different links by distinguishing between the start node and the end node, but the present invention is not limited thereto.
- the two nodes forming the link may have no distinction between the start node and the end node.
- FIG. 3 is a diagram for showing a configuration of the screen definition table 300.
- the screen definition table 300 includes information in which a screen ID 301, a screen tier 302, an upper-tier screen 303, an in-screen page ID 304, a lower-tier screen 305, and a voice operation handover allowability 306 are associated with one another.
- the screen ID 301 is information identifying the screen.
- the screen tier 302 is information identifying a tier in which the screen identified by the screen ID 301 is positioned within a screen transition system.
- the upper-tier screen 303 is information identifying a screen in the immediately upper tier with respect to the screen identified by the screen ID 301.
- the in-screen page ID 304 is information identifying a split page in a case where the screen identified by the screen ID 301 is configured to be displayed by being split into a plurality of pages when the number of options increases.
- the lower-tier screen 305 is information identifying a screen in the immediately lower tier with respect to the screen identified by the screen ID 301.
- the voice operation handover allowability 306 is information identifying whether or not the current page is a page for which an input method is handed over to voice operation when a manual operation is no longer received while the screen identified by the screen ID 301 is being displayed.
- FIG. 4 is a diagram for showing a configuration of the selection history table 400.
- the selection history table 400 includes information in which a screen ID 401, an option 402, and a selection count 403 are associated with one another.
- the screen ID 401 is information identifying the screen.
- the option 402 is information identifying the option displayed on the screen identified by the screen ID 401.
- the option 402 includes a determination condition for finally identifying a target to be operated, for example, information identifying a file name of the song file to be played back or a facility name of a facility to be set as the destination.
- the option 402 also includes, instead of the determination condition itself, a narrowing-down condition for narrowing down the determination conditions, for example, information identifying the artist of the song file to be played back or a category of the facility to be set as the destination.
- the option 402 also includes information for receiving the manual operations such as "Back", "OK", and "cancel" buttons.
- the selection count 403 is information identifying the number of times that the option 402 has been actually selected. For example, assuming that one of the options has been selected on a given screen five times, information identifying that the number of selected times is "5" is stored in the selection count 403 corresponding to the option.
- the voice input/output device 4 includes the microphone 41 as the voice input device and the speaker 42 as the voice output device.
- the microphone 41 acquires a voice outside the navigation device 100 such as a voice uttered by the user or another vehicle occupant, and receives the voice operation.
- the speaker 42 vocally outputs a message for the user generated by the arithmetic processing unit 1.
- the microphone 41 and the speaker 42 are separately arranged in predetermined sites of a vehicle, but may be received in a single housing.
- the navigation device 100 can include a plurality of microphones 41 and a plurality of speakers 42.
- the input device 5 is a device for receiving an instruction from the user through the manual operation conducted by the user.
- the input device 5 is formed of a touch panel 51, a dial switch 52, and other hardware switches (not shown) such as a scroll key and a scale change key.
- the input device 5 includes a remote control capable of remotely instructing the navigation device 100 to conduct an operation.
- the remote control includes a dial switch, a scroll key, and a scale change key, and can send information indicating that each key or switch is operated to the navigation device 100.
- the touch panel 51 is mounted on a display surface side of the display 2, and allows the display screen to be seen therethrough.
- the touch panel 51 identifies a touched position in which the manual operation is performed, which corresponds to XY coordinates of an image displayed on the display 2, and converts the touched position into coordinates, to output the coordinates.
- the touch panel 51 is formed of a pressure-sensitive or electrostatic input detection element or the like. Note that, the touch panel 51 may be one that realizes multitouch capable of simultaneously detecting a plurality of touched positions.
- the dial switch 52 is configured so as to be able to rotate clockwise and counterclockwise, and generates a pulse signal for each rotation by a predetermined angle, to output the pulse signal to the arithmetic processing unit 1.
- the arithmetic processing unit 1 obtains a rotation angle from the number of pulse signals.
- the ROM device 6 is formed of at least a readable storage medium, for example, a ROM such as a CD-ROM or a DVD-ROM, or an integrated circuit (IC) card.
- This storage medium stores, for example, video data and audio data.
- the vehicle speed sensor 7, the gyro sensor 8, and the GPS receiver 9 are used by the navigation device 100 to detect the present location (for example, location of own vehicle) .
- the vehicle speed sensor 7 is a sensor for outputting a value used to calculate a vehicle speed.
- the gyro sensor 8 is formed of, for example, a fibre optic gyroscope or a vibrating structure gyroscope, and detects an angular velocity of a moving object produced by rotation thereof.
- the GPS receiver 9 receives a signal from a GPS satellite, and measures a distance between the moving object and the GPS satellite and a rate of change in the distance for three or more satellites, to thereby measure the present location, a traveling speed, and a traveling azimuth of the moving object.
- the FM multiplex broadcast receiver 10 receives an FM multiplex broadcast signal transmitted from an FM broadcast station.
- An FM multiplex broadcast includes: vehicle information communication system (VICS: trademark) information including overall current traffic information, regulation information, service area/parking area (SA/PA) information, parking lot information, and weather information; and text information provided by a radio station as FM multiplex general information.
- VICS vehicle information communication system
- SA/PA service area/parking area
- the beacon receiver 11 receives, for example, the VICS information including the overall current traffic information, the regulation information, the service area/parking area (SA/PA) information, the parking lot information, the weather information, and an emergency alarm.
- the beacon receiver 11 is a receiver such as an optical beacon for communications using light, a radio wave beacon for communications using a radio wave, or the like.
- the in-vehicle network communication device 12 is a device for connecting the navigation device 100 to a network compatible with a controller area network (CAN) or other such control network standards for a vehicle (not shown) and conducting communications by exchanging a CAN message with an electronic control unit (ECU) that is another vehicle control device connected to the network.
- CAN controller area network
- ECU electronice control unit
- FIG. 5 is a diagram for illustrating a configuration example of screen transitions relating to an operation screen according to this embodiment.
- the screen transitions are expressed by a hierarchical structure, and the screen in a deeper tier is designed as a screen serving to input/output more concrete information than the screen in a shallower tier, that is, the upper tier, or as a screen presenting a processing result.
- the screens having no direct transition relationship are different in degree of concreteness.
- a song selection screen subjected to narrowing down through the screen for selecting the artist and a song selection screen that is not subjected to narrowing down, which are both screens for selecting a song may be different in tier for the screen transition.
- each screen can receive an operation of both the manual operation and the voice operation in a state in which the manual operation is not restricted by an input restriction unit 105, and can receive the voice operation in a state in which the manual operation is restricted by the input restriction unit 105.
- a menu screen 511 exists in a zeroth tier 501, which is the uppermost tier, and includes, as options, buttons or the like for each receiving an instruction to conduct a transition to any one of an artist selection screen 521, a playlist selection screen 522, and an album selection screen 523 in a first tier 502, which is the lower tier with respect to the menu screen 511.
- the artist selection screen 521 is a screen for receiving an input of the narrowing-down condition for, when the meta information included in a song file stored in the storage device 3 or the ROM device 6 includes information identifying an artist regarding the song, narrowing down songs to songs of the artist in distinction from songs of another artist. Further, the artist selection screen 521 displays an option for identifying the artist involved in performance or the like of the song. Whichever option for the artist is selected, a transition is made to an artist/song selection screen 531 in a second tier 503, which is the lower tier.
- the playlist selection screen 522 is a screen for receiving, when the storage device 3 or the ROM device 6 includes playlist information identifying the playback order of the song files stored in the storage device 3 or the like, an input of an instruction to play back songs within the playlist, that is, an input of the determination condition.
- the album selection screen 523 is a screen for receiving an input of the narrowing-down condition for, when the meta information included in the song file stored in the storage device 3 or the ROM device 6 includes information identifying an album, narrowing down the songs to songs within the album in distinction from songs within another album. Further, the album selection screen 523 displays an option for specifying an album serving as a unit in which one or a plurality of songs are managed by being grouped in a predetermined order . Whichever option for the album is selected, a transition is made to an album/song selection screen 533 in the second tier 503, which is the lower tier.
- the artist/song selection screen 531 which has transitioned from the artist selection screen 521, is a screen for presenting the songs obtained by being narrowed down to the songs of the selected artist in such a manner that allows selection thereof and for receiving an input of the determination condition for specifying the song file. Further, the artist/song selection screen 531 displays an option for specifying the song. Whichever option for the song is selected, a transition is made to a song playback screen 541 in a third tier 504, which is the lower tier.
- an artist/song selection (page 2) 532 is added as a screen for splitting the artist/song selection screen 531 into a plurality of pages to be displayed, and the artist/song selection screen (page 1) 531 and the artist/song selection screen (page 2) are alternately displayed so as to be movable backward and forward.
- an operation for changing a display range between the pages may be configured to switch between the pages before and after the change, or the change in the display range may be enabled by continuously changing the options included in the respective pages by an operation such as scrolling.
- the album/song selection screen 533 which has transitioned from the album selection screen 523, is a screen for presenting the songs obtained by being narrowed down to the songs of the selected album in such a manner that allows selection thereof and for receiving an input of the determination condition for specifying the song file. Further, the album/song selection screen 533 displays an option for specifying the song. Whichever option for the song is selected, a transition is made to a song playback screen 542 in the third tier 504, which is the lower tier. Note that, in the same manner as the addition to the above-mentioned artist/song selection screens 531 and 532, a page is added to the album/song selection screen 533 when there are too many options for the songs to be displayed in one screen.
- the song playback screen 541 which has transitioned from the artist/song selection screen (page 1) 531 or the artist/song selection screen (page 2) 532, is a screen for presenting information relating to the sound file for which the determination condition has been input.
- the song playback screen 541 displays a moving image or a still image relating to the playback of the song file, displays a length of a played-back part relative to a length of the song by using an indicator, displays an operation panel or the like including as options playback, stop, pause, fast forward, rewind, and output volume adjustment for the song, and conducts other such display.
- the song playback screen 542 which has transitioned from the album/song selection screen 533, is a screen for presenting information relating to the sound file for which the determination condition has been input.
- the song playback screen 542 displays a moving image or a still image relating to the song file, displays a length of a played-back part relative to a length of the song by using an indicator, displays an operation panel or the like including as options playback, stop, pause, fast forward, rewind, and output volume adjustment for the song, and conducts other such display.
- FIG. 6 is a functional diagram of the arithmetic processing unit 1. As illustrated in FIG. 6 , the arithmetic processing unit 1 includes a basic control unit 101, an input reception unit 102, an output processing unit 103, an operation history creation unit 104, an input restriction unit 105, an input reception switching unit 106, and an option reading unit 107.
- the arithmetic processing unit 1 includes a basic control unit 101, an input reception unit 102, an output processing unit 103, an operation history creation unit 104, an input restriction unit 105, an input reception switching unit 106, and an option reading unit 107.
- the basic control unit 101 is a main functional unit for conducting various kinds of processing, and controls an operation of another functional unit based on processing contents. Further, information is acquired from the respective sensors, the GPS receiver 9, and the like, and the present location is identified by conducting map matching processing or the like. Further, as the need arises, a traveling history is stored in the storage device 3 for each link by associating a date, time, and location at which traveling has taken place with one another. In addition, a present time is output in response to a request from each processing unit.
- the basic control unit 101 searches for the recommended route that is an optimal route, which connects the present location or the point of departure specified by the user to the destination (or transit point or drop-by point) .
- a route search logic such as the Dijkstra's algorithm is used to search for a route that minimizes a link cost based on the link cost set in advance for a predetermined segment (link) of the road.
- the basic control unit 101 uses the speaker 42 or the display 2 to guide the user while displaying the recommended route so as to prevent the present location from departing from the recommended route.
- the input reception unit 102 receives the manual operation or the voice operation input by the user through the input device 5 or the microphone 41, and transmits, to the basic control unit 101, an instruction to execute processing corresponding to a request content together with sound information and a coordinate position of a touch that is information relating to the voice operation. For example, when the user requests to search for the recommended route, a request instruction thereof is requested from the basic control unit 101. That is, the input reception unit 102 can be regarded as a touch instruction reception unit for receiving the instruction through a manual operation accompanied by touching. Further, the input reception unit 102 can also be regarded as a voice instruction reception unit for receiving the instruction through an operation using a voice (voice operation).
- the output processing unit 103 receives information used to form the screen to be displayed such as polygon information, and converts the information into a signal for conducting drawing on the display 2, to instruct the display 2 to conduct the drawing.
- the operation history creation unit 104 creates a history of an input of the received narrowing-down condition or determination condition for predetermined processing of the navigation device 100 such as execution of the song file or setting of the destination. Specifically, the operation history creation unit 104 counts the number of times that the execution is carried out (input of selection is instructed) for each option that is the narrowing-down condition or the determination condition the input of which is received at a time of execution (playback) of the song file or at a time of destination setting for the route search, and stored in the storage device 3 as the selection count 403 of the selection history table 400.
- the input restriction unit 105 determines that the input is to be restricted in accordance with the state of the vehicle or the like on which the navigation device 100 is mounted. Specifically, the input restriction unit 105 receives an operation with respect to the input reception unit 102 based on both the manual operation through the touch panel 51 or the dial switch 52 and the voice operation through the microphone 41 while the vehicle is stopped, but while the vehicle is traveling at a fixed speed or faster, the input restriction unit 105 determines that the manual operation through the touch panel 51 or the dial switch 52 with respect to the input reception unit 102 is restricted. Further, when a gear for moving the vehicle is selected, that is, for example, when a parking gear is not selected, the input restriction unit 105 determines that the manual operation through the touch panel 51 or the dial switch 52 with respect to the input reception unit 102 is restricted.
- the input reception switching unit 106 switches the input method by instructing the output processing unit 103 to display a predetermined screen operation disabling message such as "traveling" and instructing the input reception unit 102 to restrict the manual operation through the touch panel 51 or the dial switch 52 and to receive the voice operation through the voice input/output device 4.
- the option reading unit 107 vocally outputs the options on the screen that was displayed at a time point of the switching and the options on the subsequent transition screens through the speaker 42 or the like in an order corresponding to the selected count.
- the option reading unit 107 can be regarded as vocally outputting the options on a predetermined screen in the order corresponding to the selected count when the reception of the manual operation is restricted by the input restriction unit 105 on the predetermined screen.
- the option reading unit 107 sets a voice operation reception period that is a predetermined period for receiving the voice operation for each option, and receives the voice operation through the input reception unit 102 during the period.
- a predetermined voice operation for example, voice operation with a positive meaning such as "hai”, "OK”, or "yes
- the option reading unit 107 assumes that the option corresponding to the voice operation reception period has been selected and input, and identifies the options on a transition destination screen (lower-tier screen or the like), to start reading the identified options and receiving a selection input.
- the option reading unit 107 vocally outputs the subsequent options through the speaker 42 or the like, and sets a predetermined voice operation reception period, to receive the voice operation through the input reception unit 102 during the period.
- the option reading unit 107 further vocally outputs the options narrowed down by the narrowing-down condition on the transition destination screen.
- the option reading unit 107 conducts predetermined processing for the processing target specified by the determination condition.
- the option reading unit 107 conducts a voice output by excluding the option that has been displayed among the options on the predetermined screen.
- the respective functional units of the arithmetic processing unit 1 described above that is, the basic control unit 101, the input reception unit 102, the output processing unit 103, the operation history creation unit 104, the input restriction unit 105, the input reception switching unit 106, and the option reading unit 107 are constructed by the CPU 21 reading and executing a predetermined program. Therefore, the RAM 22 stores the program for implementing the processing of the respective functional units.
- the above-mentioned respective components are obtained by classifying the configuration of the navigation device 100 based on main processing contents in order to facilitate an understanding thereof. Therefore, the present invention is not limited by the classification method of the components and the names thereof.
- the configuration of the navigation device 100 can be classified into more components based on the processing contents. Alternatively, the configuration can be classified so that one component executes more pieces of processing.
- the respective functional units may be constructed by hardware (such as ASIC or GPU). Further, the processing of the respective functional units may be executed by one piece of hardware, or may be executed by a plurality of pieces of hardware.
- FIG. 7 is a flowchart for illustrating the voice operation handover processing carried out by the navigation device 100. This flow is carried out when the restriction of the manual operation is determined by the input restriction unit 105 in a case where, for example, the vehicle on which the navigation device 100 is mounted starts traveling after the navigation device 100 is started up, and when the input reception switching unit 106 switches the input method from the input method for receiving both the manual operation and the voice operation to the input method for receiving the voice operation with the reception of the manual operation being restricted.
- the option reading unit 107 identifies the screen ID at a time of operation restriction (Step S001). Specifically, when the screen that was displayed in the state in which the manual operation was restricted by the input restriction unit 105 is the screen display for a predetermined function activated from a menu screen, the option reading unit 107 identifies the screen ID that was displayed for the predetermined function.
- the option reading unit 107 identifies selection candidates on the screen (Step S002). Specifically, the option reading unit 107 identifies, as the selection candidates, the options that were displayed in a selectable manner on the screen identified by the screen ID identified in Step S001. Note that, the option reading unit 107 may refer to the voice operation handover allowability 306 regarding the screen, and may finish the operation for the voice operation handover processing when handover is not allowed.
- the option reading unit 107 identifies the past selection count for each selection candidate (Step S003). Specifically, the option reading unit 107 reads the selection count 403 associated in the selection history table 400 with each of the options that are the selection candidates identified in Step S002 to identify the selection count.
- the option reading unit 107 identifies the in-screen page ID being displayed at the time of operation restriction (Step S004). Specifically, when the operation for changing the display range between the pages was carried out on the screen that was displayed in a situation in which the manual operation was restricted by the input restriction unit 105, the option reading unit 107 identifies the page that has finished being referred to, that is, the page that has been excluded from the display range after being displayed. Note that, the option reading unit 107 identifies the page that has finished being referred to, that is, the options that have been excluded from the display range after being displayed when the operation for changing the display range between the pages was carried out by scrolling of the like on the screen that was displayed in the state in which the input was restricted by the input restriction unit 105.
- the option reading unit 107 extracts the candidates included in the pages subsequent to the page within the screen being displayed from among the selection candidates (Step S005). Specifically, the option reading unit 107 extracts the selection candidates by excluding the selection candidates included in the page that has finished being referred to (or the selection candidate excluded from the display range in the case of scrolling), which is identified in Step S004, from among the selection candidates identified in Step S002.
- the option reading unit 107 conducts intro sound playing or reading of the candidates for the extracted selection candidates in descending order of the past selection count (Step S006) . Specifically, the option reading unit 107 sorts the selection candidates extracted in Step S005 in descending order of the past selection count identified in Step S003, and conducts the reading of the selection candidates having a large selection count. In the processing for the reading, when the selection candidate is the determination condition, the option reading unit 107 starts a part of the processing executed for the selection candidate when the determination condition is received, and vocally outputs a name or the like of the option when the selection candidate is the narrowing-down condition.
- the option reading unit 107 outputs a sound by playing back the song for a predetermined time period (for example, 3 seconds) from a beginning thereof. Further, for example, in a case where the selection candidate is an artist, which corresponds to the narrowing-down condition, the option reading unit 107 vocally outputs a name of the artist by text-to-speech (TS) or the like.
- TS text-to-speech
- the option reading unit 107 determines whether or not a voice operation for instructing the navigation device 100 to make a selection has been received (Step S007). Specifically, the option reading unit 107 determines whether or not the voice operation for instructing the navigation device 100 to make a selection with a positive or negative meaning has been received in regard to candidates read in Step S006 through the input reception unit 102. When the voice operation for instructing the navigation device 100 to make a selection is not received, the option reading unit 107 determines repeatedly whether or not the voice operation for instructing the navigation device 100 to make a selection has been received during the predetermined voice operation reception period (for example, after the reading of the option is started and within 2 seconds after the reading of the option is finished) .
- the predetermined voice operation reception period for example, after the reading of the option is started and within 2 seconds after the reading of the option is finished
- the option reading unit 107 receives the selection of a candidate that was output at a time point at which a voice for instructing the navigation device 100 to make a selection was recognized (Step S008) . Specifically, when the voice for instructing the navigation device 100 to make a selection has a positive meaning, the option reading unit 107 identifies the option that was read in Step S006, and receives the option as one that has been selected and input. When the voice for instructing the navigation device 100 to make a selection does not have a positive meaning, the option reading unit 107 ignores the voice, and executes processing of Step S006 for the option having the next largest selection count among the options that have not been read yet.
- the option reading unit 107 causes the display to transition to the transition destination screen, and executes the file the selection of which has been received (Step S009). Specifically, the option reading unit 107 identifies the lower-tier screen 305 regarding the option that has been selected and input, and executes the file of the option when the option is the determination condition. In other words, when the song is received as the one that has been selected and input, the option reading unit 107 starts the playback of the song. When the option is the narrowing-down condition, the option reading unit 107 identifies the lower-tier screen 305 regarding the option that has been selected and input, and carries out the voice operation handover processing on the assumption that the operation is restricted when the lower-tier screen is displayed.
- the input through the voice operation can be continued when the restriction of the manual operation is carried out during the manual operation or during the voice operation.
- FIG. 8 is a diagram for illustrating an output screen example of a touch operation screen displayed when a selection target is the narrowing-down condition. Specifically, FIG. 8 is a diagram for illustrating an exemplary screen 600 of the artist selection screen 521 that is a screen for receiving the input of artist selection, which is displayed on the navigation device 100.
- the exemplary screen 600 includes a back button area 600A for receiving an instruction to return to the upper tier and an artist selection button area 600B for receiving the selection input of the artist, and each of artist names displayed in the artist selection button area 600B corresponds to the option for uniquely receiving the selection input of the artist name.
- FIG. 9 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the narrowing-down condition. Specifically, FIG. 9 is a diagram for illustrating the exemplary screen 600 displayed when the restriction of the manual operation is carried out for the artist selection screen 521 that is the screen for receiving the input of the artist selection, which is displayed on the navigation device 100.
- the back button area 600A in which the options are displayed under the state of the manual operation being disabled
- the artist selection button area 600B in which the options are displayed under the state of the manual operation being disabled
- the exemplary screen 600 displays a message area 610 indicating that the manual operation is restricted due to the traveling, in which a message of "traveling" is being displayed.
- the navigation device 100 is in a state in which the manual operation is not received through the input device 5.
- a voice guidance 620 is vocally output simultaneously with the display of the screen.
- FIG. 10 is a diagram for illustrating an output screen example of a touch operation screen displayed when the selection target is the determination condition. Specifically, FIG. 10 is a diagram for illustrating an exemplary screen 700 of the artist/song selection screen 531 that is a screen for receiving the input of song selection, which is displayed on the navigation device 100.
- the exemplary screen 700 includes a back button area 700A for receiving an instruction to return to the upper tier and an artist/song selection button area 700B for receiving the selection input of the song, and each of song names displayed in the artist/song selection button area 700B corresponds to the option for uniquely receiving the selection input of the song.
- FIG. 11 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the narrowing-down condition. Specifically, FIG. 11 is a diagram for illustrating the exemplary screen 700 displayed when the restriction of the manual operation is carried out for the artist/song selection screen 531 that is the screen for receiving the input of the artist/song selection, which is displayed on the navigation device 100.
- the back button area 700A in which the options are displayed under the state of the manual operation being disabled
- the artist/song selection button area 700B in which the options are displayed under the state of the manual operation being disabled
- the exemplary screen 700 displays a message area 710 indicating that the manual operation is restricted due to the traveling, in which the message of "traveling" is being displayed.
- the navigation device 100 is in a state in which the manual operation is not received through the input device 5.
- a voice guidance 720 is vocally output simultaneously with the display of the screen.
- a song name that is the option is vocally output, and then the message of "Do you want to play back from it?" for prompting the user to issue the instruction is read by voice.
- the determination condition relating to "Song-0005" has been specified, and the song playback screen 541 indicating detailed information at the time of the playback of the song is displayed while the song is played back to output a sound.
- FIG. 12 is a diagram for illustrating another output screen example of the touch operation screen displayed when the selection target is the narrowing-down condition. Specifically, FIG. 12 is a diagram for illustrating an exemplary screen 800 for receiving the input of destination selection, which is displayed on the navigation device 100.
- the exemplary screen 800 includes a back button area 800A for receiving an instruction to return to the upper tier and a genre selection button area 800B for receiving the selection input of the genre, and each of genre names displayed in the genre selection button area 800B corresponds to the option for uniquely receiving the selection input of the genre.
- FIG. 13 is a diagram for illustrating another output screen example of the touch operation disabled screen displayed when the selection target is the narrowing-down condition. Specifically, FIG. 13 is a diagram for illustrating the exemplary screen 800 displayed when the restriction of the manual operation is carried out for the genre selection screen that is the screen for receiving the input of the genre selection, which is displayed on the navigation device 100.
- the back button area 800A in which the options are displayed under the state of the manual operation being disabled
- the genre selection button area 800B in which the options are displayed under the state of the manual operation being disabled
- the exemplary screen 800 displays a message area 810 indicating that the manual operation is restricted due to the traveling, in which the message of "traveling" is being displayed.
- the navigation device 100 is in a state in which the manual operation is not received through the input device 5.
- a voice guidance 820 is vocally output simultaneously with the display of the screen.
- FIG. 14 is a diagram for illustrating an output screen example of the touch operation screen displayed when the selection target is the determination condition. Specifically, FIG. 14 is a diagram for illustrating an exemplary screen 900 for receiving the input of facility selection, which is displayed on the navigation device 100.
- the exemplary screen 900 includes a back button area 900A for receiving an instruction to return to the upper tier and a facility selection button area 900B for receiving the selection input of the facility, and each of facility names displayed in the facility selection button area 900B corresponds to the option for uniquely receiving the selection input of the facility.
- FIG. 15 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the determination condition. Specifically, FIG. 15 is a diagram for illustrating the exemplary screen 900 displayed when the restriction of the manual operation is carried out for the facility selection screen that is the screen for receiving the input of the facility selection, which is displayed on the navigation device 100.
- the back button area 900A in which the options are displayed under the state of the manual operation being disabled
- the facility selection button area 900B in which the options are displayed under the state of the manual operation being disabled
- the exemplary screen 900 displays a message area 910 indicating that the manual operation is restricted due to the traveling, in which the message of "traveling" is being displayed.
- the navigation device 100 is in a state in which the manual operation is not received through the input device 5.
- a voice guidance 920 is vocally output simultaneously with the display of the screen.
- the present invention is not limited to the above-mentioned embodiment.
- Various modifications can be made to the above-mentioned embodiment within the scope defined by the appended claims.
- the screen transition is expressed by the hierarchical structure, the screen in the deeper tier is designed as a screen serving to input/output more concrete information than the screen in the shallower tier, that is, the upper tier, or as the screen presenting the processing result, but the present invention is not limited thereto.
- the input screen may have a structure involving transitions among a plurality of screens.
- the input screen may have a structure involving transitions among a plurality of screens.
- the voice operation is used to receive the input of the option of the narrowing-down condition, but the present invention is not limited thereto.
- the song may be played back when the input of the voice for identifying the song that is the determination condition is received.
- the voice operation of a predetermined reserved word such as "usual" is received
- the songs may be narrowed down by the narrowing-down condition that has already been received on the screen before the transition, and the intro playback may be started in descending order of the playback count. With such a modification, it is possible to further increase the convenience.
- the selection history table 400 may be provided in a storage area accessible through the network depending on the user, and the selection count may be acquired from the navigation device 100 through communications.
- a plurality of navigation devices 100 can share a selection history.
- the present invention has been described above mainly with reference to the embodiment.
- the above-mentioned embodiment assumes the navigation device 100 that can be mounted to an automobile, but the present invention is not limited thereto, and can be applied to the navigation device for a general moving object or a device for the general moving object.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Automation & Control Theory (AREA)
- Computational Linguistics (AREA)
- Navigation (AREA)
- User Interface Of Digital Computer (AREA)
Description
- The present invention relates to a technology for a speech recognition device. The present invention claims priority to
Japanese Patent Application No. 2013-1373 filed on January 8, 2013 - Hitherto, there has been a technology for an electronic device including: detection means for detecting a state relating to the electronic device; and determination means for determining based on at least a part of the detected state whether or not to start speech recognition or whether or not to end the speech recognition, in which it is determined based on a determination result thereof whether to start or end the speech recognition, the speech recognition is conducted, and the electronic device is caused to conduct a predetermined operation based on a recognition result thereof. In
Patent Literature 1, there is disclosed a technology regarding such a device. - In
JP 2005-053331 A -
US 2006/0224260 A1 discloses an approach where small segments of songs from a music collection are played to a user, one by one. - [PTL 1]
JP 2003-195891 A - With a device as described with
PTL 1, even after speech recognition is started, in a case where, for example, a user forgets a name or the like of an instruction target or only remembers the instruction target incorrectly, a voice instruction through utterance may not be appropriate, which may inhibit an intended operation. - . It is an object of the present invention to provide a technology for a speech recognition device having higher convenience.
- In order to solve the above-mentioned problems, according to one embodiment of the present invention, there is provided a speech recognition device according to
claim 1. - Further, in the speech recognition device, the option reading unit may further conduct, when the option received by the voice instruction reception unit designates a narrowing-down condition for narrowing down the options on a transition destination screen to which a transition is made from the predetermined screen, the voice outputs of the options narrowed down by the narrowing-down condition on the transition destination screen.
- Further, in the speech recognition device, the option reading unit may conduct, when the option received by the voice instruction reception unit designates a determination condition for determining a processing target for predetermined processing, the predetermined processing for the processing target identified by the determination condition.
- Further, in the speech recognition device, the option reading unit may conduct the voice output by excluding the option that has been displayed among the options on the predetermined screen .
- Further, the speech recognition device may further include a history creation unit for updating the number of selected times within the selection history information for the option for which the instruction has been received by the touch instruction reception unit and the voice instruction reception unit.
- Further, in the speech recognition device, the speech recognition device may be mounted to a moving object, and the speech recognition device may further include an input reception switching unit for restricting, when the moving object starts moving at a predetermined speed or faster, the reception of the instruction conducted by the touch instruction reception unit.
- Further, according to one embodiment of the present invention, there is provided a speech recognition program for causing a computer to execute a speech recognition procedure according to
claim 7. - Further, according to one embodiment of the present invention, there is provided a speech recognition method to be performed by a speech recognition device according to
claim 8. - According to the one embodiment of the present invention, it is possible to provide the technology for the speech recognition device having higher convenience.
-
-
FIG. 1 is a schematic configuration diagram of a navigation device. -
FIG. 2 is a diagram for showing a configuration of a link table. -
FIG. 3 is a diagram for showing a configuration of a screen definition table. -
FIG. 4 is a diagram for showing a configuration example of a selection history table. -
FIG. 5 is a diagram for illustrating a configuration example of screen transitions. -
FIG. 6 is a functional diagram of an arithmetic processing unit of the navigation device. -
FIG. 7 is a flowchart for illustrating voice operation handover processing. -
FIG. 8 is a diagram for illustrating an output screen example of a touch operation screen displayed when a selection target is a narrowing-down condition. -
FIG. 9 is a diagram for illustrating an output screen example of a touch operation disabled screen displayed when the selection target is the narrowing-down condition. -
FIG. 10 is a diagram for illustrating an output screen example of the touch operation screen displayed when the selection target is a determination condition. -
FIG. 11 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the determination condition. -
FIG. 12 is a diagram for illustrating an output screen example of the touch operation screen displayed when the selection target is the narrowing-down condition. -
FIG. 13 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the narrowing-down condition. -
FIG. 14 is a diagram for illustrating an output screen example of the touch operation screen displayed when the selection target is the determination condition. -
FIG. 15 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the determination condition. - Now, a
navigation device 100 according to the present invention is described with reference to the accompanying drawings. -
FIG. 1 is an overall configuration diagram of thenavigation device 100. Thenavigation device 100 is a so-called navigation device capable of displaying map information and presenting a spot indicating a present location of thenavigation device 100 and information that guides a user along a route to a set destination. - The
navigation device 100 includes anarithmetic processing unit 1, adisplay 2, astorage device 3, a voice input/output device 4 (including amicrophone 41 as a voice input device and aspeaker 42 as a voice output device), aninput device 5, a read only memory (ROM) device 6, avehicle speed sensor 7, agyro sensor 8, a global positioning system (GPS)receiver 9, an FMmultiplex broadcast receiver 10, abeacon receiver 11, and an in-vehiclenetwork communication device 12. - The
arithmetic processing unit 1 is a main unit for conducting various kinds of processing. For example, thearithmetic processing unit 1 calculates the present location based on information output from therespective sensors GPS receiver 9, the FMmultiplex broadcast receiver 10, and the like. Further, based on information on the obtained present location, map data necessary for display is read from thestorage device 3 or the ROM device 6. - Further, the
arithmetic processing unit 1 transforms the read map data into graphics, and displays the graphics on thedisplay 2 with the graphics overlaid with a mark indicating the present location. Further, the map data or the like stored in thestorage device 3 or the ROM device 6 is used to search for a recommended route that is an optimal route, which connects the present location or a point of departure specified by a user to the destination (or transit point or drop-by point). Further, thespeaker 42 or thedisplay 2 is used to guide the user. - In the
arithmetic processing unit 1 of thenavigation device 100, the respective devices are connected to one another through abus 25. Thearithmetic processing unit 1 includes: a central processing unit (CPU) 21 for executing various kinds of processing such as an numerical value arithmetic operation and control of each device; a random access memory (RAM) 22 for storing the map data read from thestorage device 3, arithmetic operation data, and the like; aROM 23 for storing a program and data; and an interface (I/F) 24 for connection between various kinds of hardware and thearithmetic processing unit 1. - The
display 2 is a unit for displaying graphics information generated by thearithmetic processing unit 1 or the like. The display2 is formed of a liquid crystal display, an organic EL display, or the like. - The
storage device 3 is formed of a storage medium, which is at least readable and writable, such as a hard disk drive (HDD) or a nonvolatile memory card. - This storage medium stores: a link table 200, which is the map data (including link data on a link forming a road on a map) necessary for a general route search device; a screen definition table 300, which is definition information on a screen displayed on the
navigation device 100; and a selection history table 400, which associates the number of times that an option serving as a candidate to be selected on each screen has been actually selected with each option in units of screens. Further, for example, the storage medium of thestorage device 3 stores: one, two, or more song files; and information relating to a playlist, which defines identification information identifying a plurality of song files to be played back and a playback order of the song files. Note that, each song file includes, as meta information, attribute information such as information identifying an artist of a song, a composer thereof, a genre thereof, and an album name containing the song. -
FIG. 2 is a diagram for showing a configuration of the link table 200. For each identification code (mesh ID) 201 of a mesh that is an area segmented on the map, the link table 200 includeslink data 202 on each link forming a road included in a mesh area thereof. - For each
link ID 211 serving as the identifier of the link, thelink data 202 includescoordinate information 222 on two nodes (start node and end node) forming the link, aroad type 223 indicating a type of road including the link, alink length 224 indicating a length of the link, alink travel time 225 stored in advance, a start connection link and anend connection link 226, and aspeed limit 227 indicating a speed limit of the road including the link. Note that, the start connection link and theend connection link 226 are information identifying a start connection link serving as a link connecting to the start node of the link and an end connection link serving as a link connecting to the end node of the link. - Note that, in this case, in regard to the two nodes forming the link, an upward direction and a downward direction of the same road are managed as mutually different links by distinguishing between the start node and the end node, but the present invention is not limited thereto. For example, the two nodes forming the link may have no distinction between the start node and the end node.
-
FIG. 3 is a diagram for showing a configuration of the screen definition table 300. The screen definition table 300 includes information in which ascreen ID 301, ascreen tier 302, an upper-tier screen 303, an in-screen page ID 304, a lower-tier screen 305, and a voiceoperation handover allowability 306 are associated with one another. - The
screen ID 301 is information identifying the screen. Thescreen tier 302 is information identifying a tier in which the screen identified by thescreen ID 301 is positioned within a screen transition system. The upper-tier screen 303 is information identifying a screen in the immediately upper tier with respect to the screen identified by thescreen ID 301. The in-screen page ID 304 is information identifying a split page in a case where the screen identified by thescreen ID 301 is configured to be displayed by being split into a plurality of pages when the number of options increases. The lower-tier screen 305 is information identifying a screen in the immediately lower tier with respect to the screen identified by thescreen ID 301. The voiceoperation handover allowability 306 is information identifying whether or not the current page is a page for which an input method is handed over to voice operation when a manual operation is no longer received while the screen identified by thescreen ID 301 is being displayed. -
FIG. 4 is a diagram for showing a configuration of the selection history table 400. The selection history table 400 includes information in which ascreen ID 401, anoption 402, and aselection count 403 are associated with one another. - The
screen ID 401 is information identifying the screen. Theoption 402 is information identifying the option displayed on the screen identified by thescreen ID 401. Note that, theoption 402 includes a determination condition for finally identifying a target to be operated, for example, information identifying a file name of the song file to be played back or a facility name of a facility to be set as the destination. Further, theoption 402 also includes, instead of the determination condition itself, a narrowing-down condition for narrowing down the determination conditions, for example, information identifying the artist of the song file to be played back or a category of the facility to be set as the destination. Further, theoption 402 also includes information for receiving the manual operations such as "Back", "OK", and "cancel" buttons. - The
selection count 403 is information identifying the number of times that theoption 402 has been actually selected. For example, assuming that one of the options has been selected on a given screen five times, information identifying that the number of selected times is "5" is stored in theselection count 403 corresponding to the option. - The description is made referring back to
FIG. 1 . The voice input/output device 4 includes themicrophone 41 as the voice input device and thespeaker 42 as the voice output device. Themicrophone 41 acquires a voice outside thenavigation device 100 such as a voice uttered by the user or another vehicle occupant, and receives the voice operation. - The
speaker 42 vocally outputs a message for the user generated by thearithmetic processing unit 1. Themicrophone 41 and thespeaker 42 are separately arranged in predetermined sites of a vehicle, but may be received in a single housing. Thenavigation device 100 can include a plurality ofmicrophones 41 and a plurality ofspeakers 42. - The
input device 5 is a device for receiving an instruction from the user through the manual operation conducted by the user. Theinput device 5 is formed of atouch panel 51, adial switch 52, and other hardware switches (not shown) such as a scroll key and a scale change key. Further, theinput device 5 includes a remote control capable of remotely instructing thenavigation device 100 to conduct an operation. The remote control includes a dial switch, a scroll key, and a scale change key, and can send information indicating that each key or switch is operated to thenavigation device 100. - The
touch panel 51 is mounted on a display surface side of thedisplay 2, and allows the display screen to be seen therethrough. Thetouch panel 51 identifies a touched position in which the manual operation is performed, which corresponds to XY coordinates of an image displayed on thedisplay 2, and converts the touched position into coordinates, to output the coordinates. Thetouch panel 51 is formed of a pressure-sensitive or electrostatic input detection element or the like. Note that, thetouch panel 51 may be one that realizes multitouch capable of simultaneously detecting a plurality of touched positions. - The
dial switch 52 is configured so as to be able to rotate clockwise and counterclockwise, and generates a pulse signal for each rotation by a predetermined angle, to output the pulse signal to thearithmetic processing unit 1. Thearithmetic processing unit 1 obtains a rotation angle from the number of pulse signals. - The ROM device 6 is formed of at least a readable storage medium, for example, a ROM such as a CD-ROM or a DVD-ROM, or an integrated circuit (IC) card. This storage medium stores, for example, video data and audio data.
- The
vehicle speed sensor 7, thegyro sensor 8, and theGPS receiver 9 are used by thenavigation device 100 to detect the present location (for example, location of own vehicle) . Thevehicle speed sensor 7 is a sensor for outputting a value used to calculate a vehicle speed. Thegyro sensor 8 is formed of, for example, a fibre optic gyroscope or a vibrating structure gyroscope, and detects an angular velocity of a moving object produced by rotation thereof. TheGPS receiver 9 receives a signal from a GPS satellite, and measures a distance between the moving object and the GPS satellite and a rate of change in the distance for three or more satellites, to thereby measure the present location, a traveling speed, and a traveling azimuth of the moving object. - The FM
multiplex broadcast receiver 10 receives an FM multiplex broadcast signal transmitted from an FM broadcast station. An FM multiplex broadcast includes: vehicle information communication system (VICS: trademark) information including overall current traffic information, regulation information, service area/parking area (SA/PA) information, parking lot information, and weather information; and text information provided by a radio station as FM multiplex general information. - The
beacon receiver 11 receives, for example, the VICS information including the overall current traffic information, the regulation information, the service area/parking area (SA/PA) information, the parking lot information, the weather information, and an emergency alarm. For example, thebeacon receiver 11 is a receiver such as an optical beacon for communications using light, a radio wave beacon for communications using a radio wave, or the like. - The in-vehicle
network communication device 12 is a device for connecting thenavigation device 100 to a network compatible with a controller area network (CAN) or other such control network standards for a vehicle (not shown) and conducting communications by exchanging a CAN message with an electronic control unit (ECU) that is another vehicle control device connected to the network. -
FIG. 5 is a diagram for illustrating a configuration example of screen transitions relating to an operation screen according to this embodiment. In this embodiment, the screen transitions are expressed by a hierarchical structure, and the screen in a deeper tier is designed as a screen serving to input/output more concrete information than the screen in a shallower tier, that is, the upper tier, or as a screen presenting a processing result. However, there is no problem even if the screens having no direct transition relationship are different in degree of concreteness. For example, a song selection screen subjected to narrowing down through the screen for selecting the artist and a song selection screen that is not subjected to narrowing down, which are both screens for selecting a song, may be different in tier for the screen transition. Further, each screen can receive an operation of both the manual operation and the voice operation in a state in which the manual operation is not restricted by aninput restriction unit 105, and can receive the voice operation in a state in which the manual operation is restricted by theinput restriction unit 105. - As exemplified in
FIG. 5 , in this embodiment, amenu screen 511 exists in azeroth tier 501, which is the uppermost tier, and includes, as options, buttons or the like for each receiving an instruction to conduct a transition to any one of anartist selection screen 521, aplaylist selection screen 522, and analbum selection screen 523 in afirst tier 502, which is the lower tier with respect to themenu screen 511. - In this case, the
artist selection screen 521 is a screen for receiving an input of the narrowing-down condition for, when the meta information included in a song file stored in thestorage device 3 or the ROM device 6 includes information identifying an artist regarding the song, narrowing down songs to songs of the artist in distinction from songs of another artist. Further, theartist selection screen 521 displays an option for identifying the artist involved in performance or the like of the song. Whichever option for the artist is selected, a transition is made to an artist/song selection screen 531 in asecond tier 503, which is the lower tier. - Further, the
playlist selection screen 522 is a screen for receiving, when thestorage device 3 or the ROM device 6 includes playlist information identifying the playback order of the song files stored in thestorage device 3 or the like, an input of an instruction to play back songs within the playlist, that is, an input of the determination condition. - The
album selection screen 523 is a screen for receiving an input of the narrowing-down condition for, when the meta information included in the song file stored in thestorage device 3 or the ROM device 6 includes information identifying an album, narrowing down the songs to songs within the album in distinction from songs within another album. Further, thealbum selection screen 523 displays an option for specifying an album serving as a unit in which one or a plurality of songs are managed by being grouped in a predetermined order . Whichever option for the album is selected, a transition is made to an album/song selection screen 533 in thesecond tier 503, which is the lower tier. - The artist/
song selection screen 531, which has transitioned from theartist selection screen 521, is a screen for presenting the songs obtained by being narrowed down to the songs of the selected artist in such a manner that allows selection thereof and for receiving an input of the determination condition for specifying the song file. Further, the artist/song selection screen 531 displays an option for specifying the song. Whichever option for the song is selected, a transition is made to asong playback screen 541 in athird tier 504, which is the lower tier. Further, when there are too many options for the songs to be displayed in one screen in the artist/song selection screen 531, an artist/song selection (page 2) 532 is added as a screen for splitting the artist/song selection screen 531 into a plurality of pages to be displayed, and the artist/song selection screen (page 1) 531 and the artist/song selection screen (page 2) are alternately displayed so as to be movable backward and forward. Note that, an operation for changing a display range between the pages may be configured to switch between the pages before and after the change, or the change in the display range may be enabled by continuously changing the options included in the respective pages by an operation such as scrolling. - The album/
song selection screen 533, which has transitioned from thealbum selection screen 523, is a screen for presenting the songs obtained by being narrowed down to the songs of the selected album in such a manner that allows selection thereof and for receiving an input of the determination condition for specifying the song file. Further, the album/song selection screen 533 displays an option for specifying the song. Whichever option for the song is selected, a transition is made to asong playback screen 542 in thethird tier 504, which is the lower tier. Note that, in the same manner as the addition to the above-mentioned artist/song selection screens 531 and 532, a page is added to the album/song selection screen 533 when there are too many options for the songs to be displayed in one screen. - The
song playback screen 541, which has transitioned from the artist/song selection screen (page 1) 531 or the artist/song selection screen (page 2) 532, is a screen for presenting information relating to the sound file for which the determination condition has been input. For example, thesong playback screen 541 displays a moving image or a still image relating to the playback of the song file, displays a length of a played-back part relative to a length of the song by using an indicator, displays an operation panel or the like including as options playback, stop, pause, fast forward, rewind, and output volume adjustment for the song, and conducts other such display. - The
song playback screen 542, which has transitioned from the album/song selection screen 533, is a screen for presenting information relating to the sound file for which the determination condition has been input. For example, thesong playback screen 542 displays a moving image or a still image relating to the song file, displays a length of a played-back part relative to a length of the song by using an indicator, displays an operation panel or the like including as options playback, stop, pause, fast forward, rewind, and output volume adjustment for the song, and conducts other such display. -
FIG. 6 is a functional diagram of thearithmetic processing unit 1. As illustrated inFIG. 6 , thearithmetic processing unit 1 includes abasic control unit 101, aninput reception unit 102, anoutput processing unit 103, an operationhistory creation unit 104, aninput restriction unit 105, an inputreception switching unit 106, and anoption reading unit 107. - The
basic control unit 101 is a main functional unit for conducting various kinds of processing, and controls an operation of another functional unit based on processing contents. Further, information is acquired from the respective sensors, theGPS receiver 9, and the like, and the present location is identified by conducting map matching processing or the like. Further, as the need arises, a traveling history is stored in thestorage device 3 for each link by associating a date, time, and location at which traveling has taken place with one another. In addition, a present time is output in response to a request from each processing unit. - Further, the
basic control unit 101 searches for the recommended route that is an optimal route, which connects the present location or the point of departure specified by the user to the destination (or transit point or drop-by point) . In the route search, a route search logic such as the Dijkstra's algorithm is used to search for a route that minimizes a link cost based on the link cost set in advance for a predetermined segment (link) of the road. - Further, the
basic control unit 101 uses thespeaker 42 or thedisplay 2 to guide the user while displaying the recommended route so as to prevent the present location from departing from the recommended route. - The
input reception unit 102 receives the manual operation or the voice operation input by the user through theinput device 5 or themicrophone 41, and transmits, to thebasic control unit 101, an instruction to execute processing corresponding to a request content together with sound information and a coordinate position of a touch that is information relating to the voice operation. For example, when the user requests to search for the recommended route, a request instruction thereof is requested from thebasic control unit 101. That is, theinput reception unit 102 can be regarded as a touch instruction reception unit for receiving the instruction through a manual operation accompanied by touching. Further, theinput reception unit 102 can also be regarded as a voice instruction reception unit for receiving the instruction through an operation using a voice (voice operation). - The
output processing unit 103 receives information used to form the screen to be displayed such as polygon information, and converts the information into a signal for conducting drawing on thedisplay 2, to instruct thedisplay 2 to conduct the drawing. - The operation
history creation unit 104 creates a history of an input of the received narrowing-down condition or determination condition for predetermined processing of thenavigation device 100 such as execution of the song file or setting of the destination. Specifically, the operationhistory creation unit 104 counts the number of times that the execution is carried out (input of selection is instructed) for each option that is the narrowing-down condition or the determination condition the input of which is received at a time of execution (playback) of the song file or at a time of destination setting for the route search, and stored in thestorage device 3 as theselection count 403 of the selection history table 400. - The
input restriction unit 105 determines that the input is to be restricted in accordance with the state of the vehicle or the like on which thenavigation device 100 is mounted. Specifically, theinput restriction unit 105 receives an operation with respect to theinput reception unit 102 based on both the manual operation through thetouch panel 51 or thedial switch 52 and the voice operation through themicrophone 41 while the vehicle is stopped, but while the vehicle is traveling at a fixed speed or faster, theinput restriction unit 105 determines that the manual operation through thetouch panel 51 or thedial switch 52 with respect to theinput reception unit 102 is restricted. Further, when a gear for moving the vehicle is selected, that is, for example, when a parking gear is not selected, theinput restriction unit 105 determines that the manual operation through thetouch panel 51 or thedial switch 52 with respect to theinput reception unit 102 is restricted. - In response to the determination of the
input restriction unit 105, the inputreception switching unit 106 switches the input method by instructing theoutput processing unit 103 to display a predetermined screen operation disabling message such as "traveling" and instructing theinput reception unit 102 to restrict the manual operation through thetouch panel 51 or thedial switch 52 and to receive the voice operation through the voice input/output device 4. - When the input method is switched by the input
reception switching unit 106, theoption reading unit 107 vocally outputs the options on the screen that was displayed at a time point of the switching and the options on the subsequent transition screens through thespeaker 42 or the like in an order corresponding to the selected count. In other words, theoption reading unit 107 can be regarded as vocally outputting the options on a predetermined screen in the order corresponding to the selected count when the reception of the manual operation is restricted by theinput restriction unit 105 on the predetermined screen. - Further, in the processing for vocally outputting the options, the
option reading unit 107 sets a voice operation reception period that is a predetermined period for receiving the voice operation for each option, and receives the voice operation through theinput reception unit 102 during the period. When a predetermined voice operation (for example, voice operation with a positive meaning such as "hai", "OK", or "yes") is received, theoption reading unit 107 assumes that the option corresponding to the voice operation reception period has been selected and input, and identifies the options on a transition destination screen (lower-tier screen or the like), to start reading the identified options and receiving a selection input. - When the predetermined voice operation is not received (for example, when there is no reaction, there is no sound, or a voice operation with a negative meaning such as "iie", "tsugi", "next", or "no" is received), the
option reading unit 107 vocally outputs the subsequent options through thespeaker 42 or the like, and sets a predetermined voice operation reception period, to receive the voice operation through theinput reception unit 102 during the period. - Further, when the option received through the voice operation designates the narrowing-down condition for narrowing down the options on the transition destination screen to which a transition is made from a predetermined screen, the
option reading unit 107 further vocally outputs the options narrowed down by the narrowing-down condition on the transition destination screen. - Further, when the option received through the voice operation designates the determination condition for determining a processing target for predetermined processing, the
option reading unit 107 conducts predetermined processing for the processing target specified by the determination condition. - Further, the
option reading unit 107 conducts a voice output by excluding the option that has been displayed among the options on the predetermined screen. - The respective functional units of the
arithmetic processing unit 1 described above, that is, thebasic control unit 101, theinput reception unit 102, theoutput processing unit 103, the operationhistory creation unit 104, theinput restriction unit 105, the inputreception switching unit 106, and theoption reading unit 107 are constructed by theCPU 21 reading and executing a predetermined program. Therefore, theRAM 22 stores the program for implementing the processing of the respective functional units. - Note that, the above-mentioned respective components are obtained by classifying the configuration of the
navigation device 100 based on main processing contents in order to facilitate an understanding thereof. Therefore, the present invention is not limited by the classification method of the components and the names thereof. The configuration of thenavigation device 100 can be classified into more components based on the processing contents. Alternatively, the configuration can be classified so that one component executes more pieces of processing. - Further, the respective functional units may be constructed by hardware (such as ASIC or GPU). Further, the processing of the respective functional units may be executed by one piece of hardware, or may be executed by a plurality of pieces of hardware.
- [Description of operation] Now, a description is made of an operation for voice operation handover processing carried out by the
navigation device 100.FIG. 7 is a flowchart for illustrating the voice operation handover processing carried out by thenavigation device 100. This flow is carried out when the restriction of the manual operation is determined by theinput restriction unit 105 in a case where, for example, the vehicle on which thenavigation device 100 is mounted starts traveling after thenavigation device 100 is started up, and when the inputreception switching unit 106 switches the input method from the input method for receiving both the manual operation and the voice operation to the input method for receiving the voice operation with the reception of the manual operation being restricted. - First, the
option reading unit 107 identifies the screen ID at a time of operation restriction (Step S001). Specifically, when the screen that was displayed in the state in which the manual operation was restricted by theinput restriction unit 105 is the screen display for a predetermined function activated from a menu screen, theoption reading unit 107 identifies the screen ID that was displayed for the predetermined function. - Then, the
option reading unit 107 identifies selection candidates on the screen (Step S002). Specifically, theoption reading unit 107 identifies, as the selection candidates, the options that were displayed in a selectable manner on the screen identified by the screen ID identified in Step S001. Note that, theoption reading unit 107 may refer to the voiceoperation handover allowability 306 regarding the screen, and may finish the operation for the voice operation handover processing when handover is not allowed. - Then, the
option reading unit 107 identifies the past selection count for each selection candidate (Step S003). Specifically, theoption reading unit 107 reads theselection count 403 associated in the selection history table 400 with each of the options that are the selection candidates identified in Step S002 to identify the selection count. - Then, the
option reading unit 107 identifies the in-screen page ID being displayed at the time of operation restriction (Step S004). Specifically, when the operation for changing the display range between the pages was carried out on the screen that was displayed in a situation in which the manual operation was restricted by theinput restriction unit 105, theoption reading unit 107 identifies the page that has finished being referred to, that is, the page that has been excluded from the display range after being displayed. Note that, theoption reading unit 107 identifies the page that has finished being referred to, that is, the options that have been excluded from the display range after being displayed when the operation for changing the display range between the pages was carried out by scrolling of the like on the screen that was displayed in the state in which the input was restricted by theinput restriction unit 105. - Then, the
option reading unit 107 extracts the candidates included in the pages subsequent to the page within the screen being displayed from among the selection candidates (Step S005). Specifically, theoption reading unit 107 extracts the selection candidates by excluding the selection candidates included in the page that has finished being referred to (or the selection candidate excluded from the display range in the case of scrolling), which is identified in Step S004, from among the selection candidates identified in Step S002. - Then, the
option reading unit 107 conducts intro sound playing or reading of the candidates for the extracted selection candidates in descending order of the past selection count (Step S006) . Specifically, theoption reading unit 107 sorts the selection candidates extracted in Step S005 in descending order of the past selection count identified in Step S003, and conducts the reading of the selection candidates having a large selection count. In the processing for the reading, when the selection candidate is the determination condition, theoption reading unit 107 starts a part of the processing executed for the selection candidate when the determination condition is received, and vocally outputs a name or the like of the option when the selection candidate is the narrowing-down condition. For example, in a case where the selection candidate is a song, which corresponds to the determination condition, theoption reading unit 107 outputs a sound by playing back the song for a predetermined time period (for example, 3 seconds) from a beginning thereof. Further, for example, in a case where the selection candidate is an artist, which corresponds to the narrowing-down condition, theoption reading unit 107 vocally outputs a name of the artist by text-to-speech (TS) or the like. - Then, the
option reading unit 107 determines whether or not a voice operation for instructing thenavigation device 100 to make a selection has been received (Step S007). Specifically, theoption reading unit 107 determines whether or not the voice operation for instructing thenavigation device 100 to make a selection with a positive or negative meaning has been received in regard to candidates read in Step S006 through theinput reception unit 102. When the voice operation for instructing thenavigation device 100 to make a selection is not received, theoption reading unit 107 determines repeatedly whether or not the voice operation for instructing thenavigation device 100 to make a selection has been received during the predetermined voice operation reception period (for example, after the reading of the option is started and within 2 seconds after the reading of the option is finished) . - When the voice operation for instructing the
navigation device 100 to make a selection is received (when "Yes" in Step S007), theoption reading unit 107 receives the selection of a candidate that was output at a time point at which a voice for instructing thenavigation device 100 to make a selection was recognized (Step S008) . Specifically, when the voice for instructing thenavigation device 100 to make a selection has a positive meaning, theoption reading unit 107 identifies the option that was read in Step S006, and receives the option as one that has been selected and input. When the voice for instructing thenavigation device 100 to make a selection does not have a positive meaning, theoption reading unit 107 ignores the voice, and executes processing of Step S006 for the option having the next largest selection count among the options that have not been read yet. - Then, the
option reading unit 107 causes the display to transition to the transition destination screen, and executes the file the selection of which has been received (Step S009). Specifically, theoption reading unit 107 identifies the lower-tier screen 305 regarding the option that has been selected and input, and executes the file of the option when the option is the determination condition. In other words, when the song is received as the one that has been selected and input, theoption reading unit 107 starts the playback of the song. When the option is the narrowing-down condition, theoption reading unit 107 identifies the lower-tier screen 305 regarding the option that has been selected and input, and carries out the voice operation handover processing on the assumption that the operation is restricted when the lower-tier screen is displayed. - The processing flow of the voice operation handover processing has been described above. According to the voice operation handover processing, the input through the voice operation can be continued when the restriction of the manual operation is carried out during the manual operation or during the voice operation.
-
FIG. 8 is a diagram for illustrating an output screen example of a touch operation screen displayed when a selection target is the narrowing-down condition. Specifically,FIG. 8 is a diagram for illustrating anexemplary screen 600 of theartist selection screen 521 that is a screen for receiving the input of artist selection, which is displayed on thenavigation device 100. - The
exemplary screen 600 includes aback button area 600A for receiving an instruction to return to the upper tier and an artist selection button area 600B for receiving the selection input of the artist, and each of artist names displayed in the artist selection button area 600B corresponds to the option for uniquely receiving the selection input of the artist name. -
FIG. 9 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the narrowing-down condition. Specifically,FIG. 9 is a diagram for illustrating theexemplary screen 600 displayed when the restriction of the manual operation is carried out for theartist selection screen 521 that is the screen for receiving the input of the artist selection, which is displayed on thenavigation device 100. - On the
exemplary screen 600, theback button area 600A, in which the options are displayed under the state of the manual operation being disabled, and the artist selection button area 600B, in which the options are displayed under the state of the manual operation being disabled, are displayed by being grayed out. In addition, theexemplary screen 600 displays amessage area 610 indicating that the manual operation is restricted due to the traveling, in which a message of "traveling" is being displayed. When the screen is being displayed, thenavigation device 100 is in a state in which the manual operation is not received through theinput device 5. Further, avoice guidance 620 is vocally output simultaneously with the display of the screen. - In the
voice guidance 620, "Artist-0005", which is the option having the largest selection count, is first read by voice, and then a message of "Do you want to play back from it?" for prompting the user to issue the instruction is read by voice. In this case, when the positive voice operation is conducted, it is assumed that the narrowing-down condition relating to "Artist-0005" has been specified, and the options on the artist/song selection screen 531 that is the next screen for selecting the song relating to the artist is read by voice in the same manner (seeFIG. 11 ). When the positive voice operation is not conducted, "Artist-0033" having the next largest playback count is further read by voice. When the positive voice operation is not conducted, "Artist-0084" having the third largest playback count is read by voice. -
FIG. 10 is a diagram for illustrating an output screen example of a touch operation screen displayed when the selection target is the determination condition. Specifically,FIG. 10 is a diagram for illustrating anexemplary screen 700 of the artist/song selection screen 531 that is a screen for receiving the input of song selection, which is displayed on thenavigation device 100. - The
exemplary screen 700 includes aback button area 700A for receiving an instruction to return to the upper tier and an artist/song selection button area 700B for receiving the selection input of the song, and each of song names displayed in the artist/song selection button area 700B corresponds to the option for uniquely receiving the selection input of the song. -
FIG. 11 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the narrowing-down condition. Specifically,FIG. 11 is a diagram for illustrating theexemplary screen 700 displayed when the restriction of the manual operation is carried out for the artist/song selection screen 531 that is the screen for receiving the input of the artist/song selection, which is displayed on thenavigation device 100. - On the
exemplary screen 700, theback button area 700A, in which the options are displayed under the state of the manual operation being disabled, and the artist/song selection button area 700B, in which the options are displayed under the state of the manual operation being disabled, are displayed by being grayed out. In addition, theexemplary screen 700 displays amessage area 710 indicating that the manual operation is restricted due to the traveling, in which the message of "traveling" is being displayed. When the screen is being displayed, thenavigation device 100 is in a state in which the manual operation is not received through theinput device 5. Further, avoice guidance 720 is vocally output simultaneously with the display of the screen. - In the
voice guidance 720, the sound in an opening part (for example, 3 seconds of opening or introduction part) of "Song-0005", which is the option having the largest playback count, is first played back (intro playback). At the same time, a song name that is the option is vocally output, and then the message of "Do you want to play back from it?" for prompting the user to issue the instruction is read by voice. In this case, when the positive voice operation is conducted, it is assumed that the determination condition relating to "Song-0005" has been specified, and thesong playback screen 541 indicating detailed information at the time of the playback of the song is displayed while the song is played back to output a sound. When the positive voice operation is not conducted, the sound in the opening part of "Song-0001" having the next largest playback count is further played back. When the positive voice operation is not conducted, the sound in the opening part of "Song-0012" having the third largest playback count is played back. -
FIG. 12 is a diagram for illustrating another output screen example of the touch operation screen displayed when the selection target is the narrowing-down condition. Specifically,FIG. 12 is a diagram for illustrating anexemplary screen 800 for receiving the input of destination selection, which is displayed on thenavigation device 100. - The
exemplary screen 800 includes aback button area 800A for receiving an instruction to return to the upper tier and a genreselection button area 800B for receiving the selection input of the genre, and each of genre names displayed in the genreselection button area 800B corresponds to the option for uniquely receiving the selection input of the genre. -
FIG. 13 is a diagram for illustrating another output screen example of the touch operation disabled screen displayed when the selection target is the narrowing-down condition. Specifically,FIG. 13 is a diagram for illustrating theexemplary screen 800 displayed when the restriction of the manual operation is carried out for the genre selection screen that is the screen for receiving the input of the genre selection, which is displayed on thenavigation device 100. - On the
exemplary screen 800, theback button area 800A, in which the options are displayed under the state of the manual operation being disabled, and the genreselection button area 800B, in which the options are displayed under the state of the manual operation being disabled, are displayed by being grayed out. In addition, theexemplary screen 800 displays amessage area 810 indicating that the manual operation is restricted due to the traveling, in which the message of "traveling" is being displayed. When the screen is being displayed, thenavigation device 100 is in a state in which the manual operation is not received through theinput device 5. Further, avoice guidance 820 is vocally output simultaneously with the display of the screen. - In the
voice guidance 820, "Genre-0007", which is the option having the largest selection count is first read by voice, and then the message of "Do you want to select from it?" for prompting the user to issue the instruction is read by voice. In this case, when the positive voice operation is conducted, it is assumed that the narrowing-down condition relating to "Genre-0007" has been specified, and the options on the next screen for selecting the facility relating to the genre is read by voice in the same manner (seeFIG. 15 ) . When the positive voice operation is not conducted, "Genre-0021" having the next largest selection count is further read by voice. When the positive voice operation is not conducted, "Genre-0077" having the third largest selection count is read by voice. -
FIG. 14 is a diagram for illustrating an output screen example of the touch operation screen displayed when the selection target is the determination condition. Specifically,FIG. 14 is a diagram for illustrating anexemplary screen 900 for receiving the input of facility selection, which is displayed on thenavigation device 100. - The
exemplary screen 900 includes aback button area 900A for receiving an instruction to return to the upper tier and a facilityselection button area 900B for receiving the selection input of the facility, and each of facility names displayed in the facilityselection button area 900B corresponds to the option for uniquely receiving the selection input of the facility. -
FIG. 15 is a diagram for illustrating an output screen example of the touch operation disabled screen displayed when the selection target is the determination condition. Specifically,FIG. 15 is a diagram for illustrating theexemplary screen 900 displayed when the restriction of the manual operation is carried out for the facility selection screen that is the screen for receiving the input of the facility selection, which is displayed on thenavigation device 100. - On the
exemplary screen 900, theback button area 900A, in which the options are displayed under the state of the manual operation being disabled, and the facilityselection button area 900B, in which the options are displayed under the state of the manual operation being disabled, are displayed by being grayed out. In addition, theexemplary screen 900 displays amessage area 910 indicating that the manual operation is restricted due to the traveling, in which the message of "traveling" is being displayed. When the screen is being displayed, thenavigation device 100 is in a state in which the manual operation is not received through theinput device 5. Further, avoice guidance 920 is vocally output simultaneously with the display of the screen. - In the
voice guidance 920, "Facility-0090", which is the option having the largest selection count is first read by voice, and then the message of "Do you want to select from it?" for prompting the user to issue the instruction is read by voice. In this case, when the positive voice operation is conducted, it is assumed that the determination condition relating to "Facility-0090" has been specified, and a route display screen including the facility as the destination is displayed, to set the route as the recommended route. When the positive voice operation is not conducted, "Facility-0038" having the next largest selection count is further read by voice. When the positive voice operation is not conducted, "Facility-0002" having the third largest selection count is read by voice. - The embodiment of the present invention has been described above. According to the above-mentioned embodiment of the present invention, it is possible to provide the speech recognition device having higher convenience.
- The present invention is not limited to the above-mentioned embodiment. Various modifications can be made to the above-mentioned embodiment within the scope defined by the appended claims. For example, in the above-mentioned embodiment, it is assumed that the screen transition is expressed by the hierarchical structure, the screen in the deeper tier is designed as a screen serving to input/output more concrete information than the screen in the shallower tier, that is, the upper tier, or as the screen presenting the processing result, but the present invention is not limited thereto.
- For example, when a screen or the like having a large number of input items is included, the input screen may have a structure involving transitions among a plurality of screens. In other words, according to the above-mentioned embodiment, it is conceivable that an appropriate input using a voice is possible even when the screen that has already been subjected to the input operation exists within the transitions.
- Further, for example, in the above-mentioned embodiment, when the manual operation is restricted in the selection of the option of the narrowing-down condition, the voice operation is used to receive the input of the option of the narrowing-down condition, but the present invention is not limited thereto. For example, the song may be played back when the input of the voice for identifying the song that is the determination condition is received. Further, when the voice operation of a predetermined reserved word such as "usual" is received, the songs may be narrowed down by the narrowing-down condition that has already been received on the screen before the transition, and the intro playback may be started in descending order of the playback count. With such a modification, it is possible to further increase the convenience.
- Further, for example, the selection history table 400 according to the above-mentioned embodiment may be provided in a storage area accessible through the network depending on the user, and the selection count may be acquired from the
navigation device 100 through communications. With this configuration, a plurality ofnavigation devices 100 can share a selection history. - The present invention has been described above mainly with reference to the embodiment. Note that, the above-mentioned embodiment assumes the
navigation device 100 that can be mounted to an automobile, but the present invention is not limited thereto, and can be applied to the navigation device for a general moving object or a device for the general moving object. - 1 ··· arithmetic processing unit, 2 ··· display, 3 ··· storage device, 4 ··· voice input/output device, 5 ··· input device, 6··· ROM device, 7 ··· vehicle speed sensor, 8 ··· gyro sensor, 9 ··· GPS receiver, 10 ··· FM multiplex broadcast receiver, 11 ··· beacon receiver, 12 ··· in-vehicle network communication device, 21 ··· CPU, 22 ··· RAM, 23 ··· ROM, 24 ··· I/F, 25 ··· bus, 41 ··· microphone, 42 ··· speaker, 51 ··· touch panel, 52 ··· dial switch, 100 ··· navigation device, 101 ··· basic control unit, 102 ··· input reception unit, 103 ··· output processing unit, 104 ··· operation history creation unit, 105 ··· input restriction unit, 106 ··· input reception switching unit, 107 ··· option reading unit, 200 ··· link table, 300 ··· screen definition table, 400 ··· selection history table
Claims (8)
- A speech recognition device, comprising:a storage unit (3) for storing screen definition information (300), in which a screen is associated with an option on the screen, and selection history information (400) identifying a number of selection times for each of the options;a touch instruction reception unit (51, 102) for receiving an instruction through touching operation;a voice instruction reception unit (41, 102) for receiving an instruction through an operation using voice; andan option reading unit (42, 107) for conducting, when reception of the instruction conducted by the touch instruction reception unit is restricted on a predetermined screen, voice output of the options on the predetermined screen in an order corresponding to the number of selection times,wherein the voice instruction reception unit (41, 102) is adapted to receive an instruction regarding any one of the options output by the option reading unit;each of the options on the predetermined screen identifies a predetermined song file; andthe option reading unit (42, 107) is adapted to conduct the voice output of the options by playing back, for each song file, at least a part of a song regarding the song file.
- A speech recognition device according to claim 1, wherein the option reading unit (42, 107) is further adapted to conduct, when the option received by the voice instruction reception unit (41, 102) designates a narrowing-down condition for narrowing down the options on a transition destination screen to which a transition is made from the predetermined screen, the voice output of the options narrowed down by the narrowing-down condition on the transition destination screen.
- A speech recognition device according to claim 1 or 2, wherein the option reading unit (42, 107) is adapted to conduct, when the option received by the voice instruction reception unit (41, 102) designates a determination condition for determining a processing target for predetermined processing, the predetermined processing for the processing target identified by the determination condition.
- A speech recognition device according to any one of claims 1 to 3, wherein the option reading unit (42, 107) is adapted to conduct the voice output by excluding an option that has been displayed among the options on the predetermined screen.
- A speech recognition device according to any one of claims 1 to 4, further comprising a history creation unit (104) for updating the number of selection times within the selection history information (400) for the option for which the instruction has been received by the touch instruction reception unit and the voice instruction reception unit.
- A speech recognition device according to any one of claims 1 to 5, wherein:the speech recognition device is mounted to a moving object; andthe speech recognition device further comprises an input reception switching unit (106) for restricting, when the moving object starts moving at a predetermined speed or faster, the reception of the instruction conducted by the touch instruction reception unit (51, 102).
- A speech recognition program for causing a computer to execute a speech recognition procedure, the speech recognition program further causing the computer to function as:control means (101);touch instruction reception means (51, 102) for receiving an instruction through touching operation;voice instruction reception means (41, 102) for receiving an instruction through an operation using voice; andstorage means (3) for storing screen definition information (300), in which a screen is associated with an option on the screen, and selection history information (400) identifying a number of selection times for each of the options, wherein:the speech recognition program is further adapted to cause the control means (101) to execute an option reading procedure (S006, 42, 107) of conducting, when reception of the instruction conducted by the touch instruction reception means is restricted on a predetermined screen, voice output of the options on the predetermined screen in an order corresponding to the number of selection times;the speech recognition program is further adapted to cause the voice instruction reception means (41, 102) to receive an instruction regarding any one of the options output in the option reading procedure;each of the options on the predetermined screen identifies a predetermined song file; andthe option reading procedure (S006, 42, 107) conducts the voice output of the options by playing back, for each song file, at least a part of a song regarding the song file.
- A speech recognition method performed by a speech recognition device, comprising:storing screen definition information (300), in which a screen is associated with an option on the screen, and selection history information (400) identifying a number of selection times for each of the options by a storage unit (3) of the speech recognition device; receiving an instruction through touching operation by a touch instruction reception unit (51, 102) of the speech recognition device; andreceiving an instruction through an operation using voice by a voice instruction reception unit (41, 102) of the speech recognition device, the speech recognition method comprising:an option reading step (S006) of conducting, by the speech recognition device, when reception of the instruction conducted by the touch instruction reception unit is restricted on a predetermined screen, voice output of the options on the predetermined screen in an order corresponding to the number of selection times; anda step of receiving (S008), by the voice instruction reception unit (41, 102) of the speech recognition device, an instruction regarding any one of the options output in the option reading step,wherein each of the options on the predetermined screen identifies a predetermined song file; andthe option reading step (S006) conducts the voice output of the options by playing back, for each song file, at least a part of a song regarding the song file.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013001373 | 2013-01-08 | ||
PCT/JP2013/078498 WO2014109104A1 (en) | 2013-01-08 | 2013-10-21 | Voice recognition device, voice recognition program, and voice recognition method |
Publications (3)
Publication Number | Publication Date |
---|---|
EP2945052A1 EP2945052A1 (en) | 2015-11-18 |
EP2945052A4 EP2945052A4 (en) | 2016-08-10 |
EP2945052B1 true EP2945052B1 (en) | 2017-12-20 |
Family
ID=51166769
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP13870530.6A Active EP2945052B1 (en) | 2013-01-08 | 2013-10-21 | Voice recognition device, voice recognition program, and voice recognition method |
Country Status (5)
Country | Link |
---|---|
US (1) | US20150348555A1 (en) |
EP (1) | EP2945052B1 (en) |
JP (1) | JPWO2014109104A1 (en) |
CN (1) | CN104903846B (en) |
WO (1) | WO2014109104A1 (en) |
Families Citing this family (82)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10417037B2 (en) | 2012-05-15 | 2019-09-17 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
EP4138075A1 (en) | 2013-02-07 | 2023-02-22 | Apple Inc. | Voice trigger for a digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
EP3008641A1 (en) | 2013-06-09 | 2016-04-20 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
CN105453026A (en) | 2013-08-06 | 2016-03-30 | 苹果公司 | Auto-activating smart responses based on activities from remote devices |
CN106471570B (en) | 2014-05-30 | 2019-10-01 | 苹果公司 | Order single language input method more |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10460227B2 (en) | 2015-05-15 | 2019-10-29 | Apple Inc. | Virtual assistant in a communication session |
US10200824B2 (en) | 2015-05-27 | 2019-02-05 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device |
US20160378747A1 (en) | 2015-06-29 | 2016-12-29 | Apple Inc. | Virtual assistant for media playback |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10740384B2 (en) | 2015-09-08 | 2020-08-11 | Apple Inc. | Intelligent automated assistant for media search and playback |
US10331312B2 (en) | 2015-09-08 | 2019-06-25 | Apple Inc. | Intelligent automated assistant in a media environment |
JP2017062300A (en) * | 2015-09-24 | 2017-03-30 | セイコーエプソン株式会社 | Semiconductor device, system, electronic equipment, and voice recognizing method |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10956666B2 (en) | 2015-11-09 | 2021-03-23 | Apple Inc. | Unconventional virtual assistant interactions |
FR3044436B1 (en) * | 2015-11-27 | 2017-12-01 | Thales Sa | METHOD FOR USING A MAN-MACHINE INTERFACE DEVICE FOR AN AIRCRAFT HAVING A SPEECH RECOGNITION UNIT |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | User interface for correcting recognition errors |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK180048B1 (en) | 2017-05-11 | 2020-02-04 | Apple Inc. | MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | Low-latency intelligent automated assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK201770411A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Multi-modal interfaces |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
US20180336892A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Detecting a trigger of a digital assistant |
CN107342082A (en) * | 2017-06-29 | 2017-11-10 | 北京小米移动软件有限公司 | Audio-frequency processing method, device and audio-frequence player device based on interactive voice |
US10448762B2 (en) | 2017-09-15 | 2019-10-22 | Kohler Co. | Mirror |
US10887125B2 (en) | 2017-09-15 | 2021-01-05 | Kohler Co. | Bathroom speaker |
US11314214B2 (en) | 2017-09-15 | 2022-04-26 | Kohler Co. | Geographic analysis of water conditions |
US11093554B2 (en) | 2017-09-15 | 2021-08-17 | Kohler Co. | Feedback for water consuming appliance |
US11099540B2 (en) | 2017-09-15 | 2021-08-24 | Kohler Co. | User identity in household appliances |
JP6911730B2 (en) * | 2017-11-29 | 2021-07-28 | 京セラドキュメントソリューションズ株式会社 | Display device, image processing device, processing execution method, processing execution program |
CN111801249A (en) * | 2018-02-22 | 2020-10-20 | 三菱电机株式会社 | Operation restriction control device and operation restriction control method |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US11231848B2 (en) * | 2018-06-28 | 2022-01-25 | Hewlett-Packard Development Company, L.P. | Non-positive index values of panel input sources |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
DK201970509A1 (en) | 2019-05-06 | 2021-01-15 | Apple Inc | Spoken notifications |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
DK201970510A1 (en) | 2019-05-31 | 2021-02-11 | Apple Inc | Voice identification in digital assistant systems |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
DK180129B1 (en) | 2019-05-31 | 2020-06-02 | Apple Inc. | User activity shortcut suggestions |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11227599B2 (en) | 2019-06-01 | 2022-01-18 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
CN112802474A (en) * | 2019-10-28 | 2021-05-14 | 中国移动通信有限公司研究院 | Voice recognition method, device, equipment and storage medium |
US11038934B1 (en) | 2020-05-11 | 2021-06-15 | Apple Inc. | Digital assistant hardware abstraction |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
US11756544B2 (en) * | 2020-12-15 | 2023-09-12 | Google Llc | Selectively providing enhanced clarification prompts in automated assistant interactions |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060224260A1 (en) * | 2005-03-04 | 2006-10-05 | Hicken Wendell T | Scan shuffle for building playlists |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001125766A (en) * | 1999-10-28 | 2001-05-11 | Sumitomo Electric Ind Ltd | Device and method for controlling apparatus loaded on vehicle |
JP2002311986A (en) * | 2001-04-17 | 2002-10-25 | Alpine Electronics Inc | Navigator |
JP3951705B2 (en) | 2001-12-27 | 2007-08-01 | 株式会社デンソー | Electronics |
KR100996212B1 (en) * | 2002-09-06 | 2010-11-24 | 보이스 시그널 테크놀로지스, 인코포레이티드. | Methods, systems, and programming for performing speech recognition |
JP2005053331A (en) * | 2003-08-04 | 2005-03-03 | Nissan Motor Co Ltd | Information presenting device for vehicular instrument |
US7870142B2 (en) * | 2006-04-04 | 2011-01-11 | Johnson Controls Technology Company | Text to grammar enhancements for media files |
JP2012234341A (en) * | 2011-04-28 | 2012-11-29 | Toshiba Corp | Video display device and menu screen display method |
US8811938B2 (en) * | 2011-12-16 | 2014-08-19 | Microsoft Corporation | Providing a user interface experience based on inferred vehicle state |
-
2013
- 2013-10-21 CN CN201380069575.2A patent/CN104903846B/en active Active
- 2013-10-21 WO PCT/JP2013/078498 patent/WO2014109104A1/en active Application Filing
- 2013-10-21 JP JP2014556325A patent/JPWO2014109104A1/en active Pending
- 2013-10-21 EP EP13870530.6A patent/EP2945052B1/en active Active
- 2013-10-21 US US14/759,537 patent/US20150348555A1/en not_active Abandoned
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060224260A1 (en) * | 2005-03-04 | 2006-10-05 | Hicken Wendell T | Scan shuffle for building playlists |
Also Published As
Publication number | Publication date |
---|---|
JPWO2014109104A1 (en) | 2017-01-19 |
EP2945052A4 (en) | 2016-08-10 |
CN104903846A (en) | 2015-09-09 |
WO2014109104A1 (en) | 2014-07-17 |
CN104903846B (en) | 2017-07-28 |
US20150348555A1 (en) | 2015-12-03 |
EP2945052A1 (en) | 2015-11-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2945052B1 (en) | Voice recognition device, voice recognition program, and voice recognition method | |
JP4551961B2 (en) | VOICE INPUT SUPPORT DEVICE, ITS METHOD, ITS PROGRAM, RECORDING MEDIUM RECORDING THE PROGRAM, AND NAVIGATION DEVICE | |
JP5522273B2 (en) | In-vehicle information equipment operating device | |
JP6098419B2 (en) | Traffic information guidance system, traffic information guidance device, traffic information guidance method, and computer program | |
JP6595293B2 (en) | In-vehicle device, in-vehicle system, and notification information output method | |
WO2017111075A1 (en) | On-board device, display area dividing method, program, and information control device | |
JP2013101535A (en) | Information retrieval device and information retrieval method | |
JP5506252B2 (en) | Route search apparatus and route search method | |
JP2011232270A (en) | Navigation device and help presentation method thereof | |
JP5018671B2 (en) | Vehicle navigation device | |
JP2013171502A (en) | Warning device and warning method | |
JP2019138669A (en) | Information processing device and information processing method | |
JP2019020682A (en) | Information control device and information control method | |
JP2011086210A (en) | In-vehicle device | |
JP2011185667A (en) | Navigation system, program, and display method | |
JP2013199241A (en) | Device and method of vehicle control and computer program | |
JP2011080851A (en) | Navigation system and map image display method | |
JP2010249642A (en) | Navigation device for vehicle | |
US9459109B2 (en) | Navigation device | |
JP2012042481A (en) | Navigation device | |
JP2010127770A (en) | Navigation device and navigation method | |
JP2008298522A (en) | Navigation apparatus, and method and program for retrieval of navigation apparatus | |
JP7252045B2 (en) | Information control device, information control method and program | |
JP5028045B2 (en) | In-vehicle display device | |
JP2011227002A (en) | On-vehicle navigation device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20150721 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAX | Request for extension of the european patent (deleted) | ||
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G01C 21/26 20060101ALI20160630BHEP Ipc: G06F 3/16 20060101AFI20160630BHEP Ipc: G10L 15/00 20130101ALI20160630BHEP Ipc: G10L 13/00 20060101ALI20160630BHEP Ipc: G01C 21/36 20060101ALI20160630BHEP Ipc: G10L 13/02 20130101ALI20160630BHEP Ipc: G10L 15/22 20060101ALI20160630BHEP |
|
A4 | Supplementary search report drawn up and despatched |
Effective date: 20160707 |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G01C 21/36 20060101ALI20170602BHEP Ipc: G06F 3/16 20060101AFI20170602BHEP Ipc: G10L 13/02 20130101ALI20170602BHEP Ipc: G01C 21/26 20060101ALI20170602BHEP Ipc: G10L 13/00 20060101ALI20170602BHEP Ipc: G10L 15/00 20130101ALI20170602BHEP Ipc: G10L 15/22 20060101ALI20170602BHEP |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
INTG | Intention to grant announced |
Effective date: 20170719 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: REF Ref document number: 956940 Country of ref document: AT Kind code of ref document: T Effective date: 20180115 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 602013031230 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: MP Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: NO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180320 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG4D |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: MK05 Ref document number: 956940 Country of ref document: AT Kind code of ref document: T Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180321 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180320 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: RS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20180420 Ref country code: SM Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 6 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 602013031230 Country of ref document: DE |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20180921 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
REG | Reference to a national code |
Ref country code: BE Ref legal event code: MM Effective date: 20181031 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20181021 Ref country code: MC Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: MM4A |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20181031 Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20181031 Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20181031 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20181021 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20181021 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MK Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20171220 Ref country code: HU Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO Effective date: 20131021 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20171220 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20230920 Year of fee payment: 11 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20230920 Year of fee payment: 11 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20230920 Year of fee payment: 11 |