WO2007145331A1 - Camera control apparatus, camera control method, camera control program, and recording medium - Google Patents

Camera control apparatus, camera control method, camera control program, and recording medium Download PDF

Info

Publication number
WO2007145331A1
WO2007145331A1 PCT/JP2007/062151 JP2007062151W WO2007145331A1 WO 2007145331 A1 WO2007145331 A1 WO 2007145331A1 JP 2007062151 W JP2007062151 W JP 2007062151W WO 2007145331 A1 WO2007145331 A1 WO 2007145331A1
Authority
WO
WIPO (PCT)
Prior art keywords
camera
candidate
passenger
information
input
Prior art date
Application number
PCT/JP2007/062151
Other languages
French (fr)
Japanese (ja)
Inventor
Hiroaki Shibasaki
Original Assignee
Pioneer Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pioneer Corporation filed Critical Pioneer Corporation
Priority to JP2008521277A priority Critical patent/JP4660592B2/en
Publication of WO2007145331A1 publication Critical patent/WO2007145331A1/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • CAMERA CONTROL DEVICE CAMERA CONTROL METHOD
  • CAMERA CONTROL PROGRAM AND RECORDING MEDIUM
  • the present invention relates to a camera control device, a camera control method, a camera control program, and a recording medium that control a shooting direction of a camera.
  • the use of the present invention is not limited to the above-described camera control device, camera control method, camera control program, and recording medium. Background art
  • Patent Document 1 Japanese Patent Application Laid-Open No. 2000-244885
  • a camera control device includes: an acquisition unit that acquires audio around a camera; and the audio acquired by the acquisition unit, Discriminating means for discriminating a word (hereinafter, referred to as “specific word”) that identifies an object to be photographed by the camera, and control means for controlling the shooting direction of the power camera based on the specific word or phrase discriminated by the discriminating means It is characterized by providing.
  • an acquisition step of acquiring audio around the camera, and a phrase for specifying a subject to be imaged by the camera from the audio acquired by the acquisition step And a control step for controlling the shooting direction of the camera based on the specific phrase determined by the determination step.
  • a camera control program according to the invention of claim 8 causes a computer to execute the camera control method according to claim 7.
  • a recording medium according to the invention of claim 9 is readable by a computer having the camera control program of claim 8 recorded thereon.
  • FIG. 1 is a block diagram showing a functional configuration of a camera control device.
  • FIG. 2 is a flowchart showing a procedure of camera control processing by the camera control device.
  • FIG. 3 is an explanatory view showing the vicinity of a dashboard of a vehicle in which a navigation device is installed.
  • FIG. 4 is a block diagram showing a hardware configuration of the navigation device.
  • FIG. 5 is an explanatory diagram schematically showing the contents of a database created by the navigation device.
  • FIG. 6 is a flowchart showing a procedure of in-vehicle shooting processing of the navigation device.
  • Fig. 7 is a flowchart showing a procedure of in-vehicle shooting processing of the navigation device.
  • FIG. 8 is an explanatory diagram showing an example of an input screen for prospective passenger information.
  • FIG. 1 is a block diagram showing a functional configuration of the camera control apparatus.
  • Camera control device 10
  • 0 is configured by an acquisition unit 101, a determination unit 102, a control unit 103, an input unit 104, and a detection unit 105.
  • the acquisition unit 101 acquires audio around the camera 110.
  • the sound around the camera 110 is, for example, an utterance made by a person located around the camera 110.
  • the acquisition unit 101 acquires the sound around the camera 110 using, for example, a microphone.
  • the discriminating unit 102 discriminates a phrase (hereinafter referred to as a specific phrase) that identifies a subject to be photographed by the camera 110 from the sound acquired by the acquiring unit 101.
  • the discriminating unit 102 discriminates, for example, information relating to a candidate for photographing input to the input unit 104 described later as a specific phrase.
  • the control unit 103 controls the shooting direction of the camera 110 based on the specific phrase determined by the determination unit 102. For example, when the determination unit 102 determines a word or phrase that substantially matches information related to the candidate for the shooting target, the control unit 103 determines the shooting direction of the camera 110 at the position of the candidate for the shooting target detected by the detection unit 105 described later. Turn. Words that match A phrase is a phrase that is the same as or similar to a phrase that is input as information about a candidate to be photographed.
  • the input unit 104 receives input of information related to a candidate for photographing.
  • the information regarding the candidates for the imaging target includes, for example, name information (name, nickname, etc.) of the candidate for the imaging target and attribute information.
  • the input unit 104 may receive a candidate image or sound to be captured.
  • the detection unit 105 detects the position of a candidate for imaging.
  • the detection unit 105 collates the image of the photographing target candidate input to the input unit 104 with the image photographed by the camera 110, and detects the position of the photographing target candidate.
  • the shooting target position is the position of the seat where the shooting target is seated, and the detection unit 105 determines which seat the passenger is seated on. Is detected.
  • the position to be imaged may be a relative direction from the camera 110, a relative direction, or the like. Specifically, for example, when an object whose similarity with a candidate image to be captured is reflected in a video captured by the camera 110, the detection unit 105 detects the position of the object. It is set as the position of a candidate for photographing.
  • the detection unit 105 detects the position of the candidate for the photographing target by collating the voice of the candidate for the photographing target input to the input unit 104 with the voice acquired by the acquiring unit 101, for example. Specifically, for example, if the voice acquired by the acquisition unit 101 includes a voice whose similarity with the candidate voice to be shot is a predetermined value or more! /, The voice is pronounced. Therefore, this position is set as the position of the candidate for the photographing target.
  • the camera 110 may be installed inside the vehicle.
  • the acquisition unit 101 acquires the utterance of the passenger of the vehicle
  • the determination unit 102 determines information about the passenger to be photographed as a specific phrase
  • the control unit 103 specifies the specific phrase by the specific phrase. Aim the shooting direction of the camera 110 at the boarding position of the passenger. Further, for example, information related to the passenger is input to the input unit 104, and the detection unit 105 detects which seat each passenger is seated in.
  • FIG. 2 is a flowchart showing a procedure of camera control processing by the camera control apparatus.
  • Figure 2 In the flowchart, first, information relating to a candidate for an imaging target is input to the input unit 104 (step S201). Further, the position of the candidate for photographing is detected by the detection unit 105 (step S 202).
  • step S 203 the sound around the camera 110 is acquired by the acquisition unit 101 (step S 203).
  • the determination unit 102 determines a specific phrase from the voice acquired in step S203 (step S204).
  • the specific word / phrase at this time is information on the candidate to be photographed.
  • a photographing target to be photographed is specified from among photographing target candidates.
  • the control unit 103 controls the shooting direction of the camera 110 to the position of the shooting target detected in step S202 (step S205), and the process according to this flowchart ends.
  • the sound force around the camera also specifies the shooting target and controls the shooting direction of the camera toward the shooting target. Thereby, it is possible to change the shooting direction of the camera 110 in consideration of the audio content around the camera.
  • the camera control device 100 specifies the shooting target using information relating to the shooting target candidates input in advance as specific words, the shooting target can be specified with higher accuracy. Furthermore, since the position of the shooting target is detected from the image and sound of the shooting target candidate, the shooting direction of the camera 110 can be controlled with higher accuracy.
  • FIG. 3 is an explanatory diagram showing the vicinity of the dashboard of the vehicle in which the navigation device is installed.
  • the navigation device 300 is installed on the dashboard of the vehicle.
  • the navigation device 300 includes a main body M and a display unit (display) D.
  • the display unit D displays the current location of the vehicle, map information, current time, and the like.
  • the navigation apparatus 300 is connected to an in-vehicle camera 311 installed in the vicinity of the knock mirror and an in-vehicle microphone 312 installed in the sun-noiser.
  • Car camera 311 The shooting direction can be changed, and images are taken in front of the vehicle and in the vehicle (passengers, etc.).
  • the in-vehicle camera 311 may be a camera capable of capturing only still images, assuming that the camera 311 is capable of capturing moving images and still images.
  • the in-vehicle microphone 312 is used when an in-vehicle sound is input and an operation by the sound input of the navigation device 300 or a state in the vehicle is recorded. It should be noted that the position of the in-vehicle microphone 312 is not limited to the sun visor as long as it can efficiently input voice in the vehicle. Further, a plurality of in-vehicle cameras 311 and in-vehicle microphones 312 may be installed in the vehicle, and may be movable rather than fixed. In the present embodiment, it is assumed that the in-vehicle microphone 312 is provided for each passenger's seat.
  • the navigation device 300 has an in-vehicle shooting function for recording a state in the vehicle being driven, in addition to searching for a route to the destination and recording information.
  • the in-vehicle shooting function records in-vehicle video and audio with the in-vehicle camera 311 and the in-vehicle microphone 312.
  • the video and audio recorded by the in-vehicle shooting function are recorded on a recording medium (magnetic disk 405 and optical disk 407 described later) of the navigation device 300. Further, the recorded video and sound may be recorded on an external recording medium so that they can be enjoyed on a home TV or the like.
  • FIG. 4 is a block diagram showing the hardware configuration of the navigation device.
  • the navigation device 300 includes a CPU 401, a ROM 402, a RAM (memory) 403, a magnetic disk drive 404, a magnetic disk 405, an optical disk drive 406, an optical disk 407, and an audio IZF (interface) 408.
  • Each component 40 1-418 is connected by a bus 420.
  • the CPU 401 governs overall control of the navigation device 300.
  • the ROM 402 records programs such as a boot program, a communication program, a database creation program, and a data analysis program.
  • RAM403 is the work area of CPU401 used.
  • the magnetic disk drive 404 controls reading and writing of data to the magnetic disk 405 according to the control of the CPU 401.
  • the magnetic disk 405 records data written under the control of the magnetic disk drive 404.
  • the magnetic disk 405 for example, HD (node disk) or FD (flexible disk) can be used.
  • the optical disk drive 406 controls data reading / writing to the optical disk 407 according to the control of the CPU 401.
  • the optical disk 407 is a detachable recording medium from which data is read according to the control of the optical disk drive 406.
  • the optical disc 407 can also use a writable recording medium.
  • the removable recording medium may be a power MO of the optical disc 407, a memory card, or the like.
  • map data used for route search and route guidance.
  • the map data includes background data representing features (features) such as buildings, rivers, and the ground surface, and road shape data representing the shape of the road, and is displayed in two or three dimensions on the display screen of the display 414. It is drawn.
  • the navigation device 300 is guiding a route, the map data and the current location of the host vehicle acquired by the GPS unit 416 described later are displayed in an overlapping manner.
  • Audio IZF 408 is connected to audio input microphone 409 (for example, in-vehicle microphone 312 in FIG. 3) and audio output speaker 410.
  • the sound received by the microphone 409 is AZD converted in the sound IZF 408.
  • sound is output from the speaker 410. Note that the sound input from the microphone 409 can be recorded on the magnetic disk 405 or the optical disk 407 as sound data.
  • Examples of the input device 411 include a remote controller, a keyboard, a mouse, and a touch panel that are provided with a plurality of keys for inputting characters, numerical values, various instructions, and the like. Further, the input device 411 can be connected to another information processing terminal such as a digital camera or a mobile phone terminal to input / output data.
  • the video IZF 412 is connected to a video input camera 413 (for example, the in-vehicle camera 311 in FIG. 3) and a video output display 414.
  • Video IZF412 is specifically a graphic controller that controls the entire display 414 and an immediate display, for example. It consists of a buffer memory such as VRAM (Video RAM) that temporarily records possible image information and a control IC that controls display 414 based on image data output from the graphic controller.
  • VRAM Video RAM
  • the camera 413 captures images (including moving images) inside and outside the vehicle and outputs them as image data.
  • An image taken by the camera 413 can be recorded on the magnetic disk 405 or the optical disk 407 as image data.
  • This image data can be used by other information processing terminals by outputting it on the display 414, recording it on a recording medium, or transmitting it via a network.
  • the display 414 displays icons, cursors, menus, windows, or various data such as characters and images.
  • this display 414 for example, a CRT, a TFT liquid crystal display, a plasma display, or the like can be adopted.
  • Communication IZF 415 is connected to a communication network such as the Internet via radio, and functions as an interface between this communication network and CPU 401.
  • Communication networks include LAN, WAN, public line network and mobile phone network.
  • 0-3 416 receives a radio wave from a GPS satellite and outputs information indicating the current position of the vehicle (the current position of the navigation device 300).
  • the output information of the GPS unit 416 is used when the CPU 401 calculates the current position of the vehicle together with output values of various sensors 417 described later.
  • the information indicating the current location is information specifying one point on the map data, for example, latitude'longitude and altitude.
  • Various sensors 417 output information that can determine the position and behavior of the vehicle, such as a vehicle speed sensor, an acceleration sensor, and an angular velocity sensor.
  • the output values of the various sensors 417 are used for calculation of the current position by the CCU 401 and measurement of speed and direction change.
  • the IZF 418 for external connection is an interface for connecting to an external device such as an audio device or a vehicle air conditioner.
  • the external connection IZF 418 is configured by, for example, a dedicated connection cable port and an infrared communication port.
  • the acquisition unit 101 is a voice I / F 408, a microphone 409, a half IJ additional U ⁇ 102, a detection ⁇ is a CPU 401, and an ffilj control unit 103.
  • the input unit 104 is connected to the input device 411 by the CPU 401 and the video IZF 412. To realize each function.
  • the in-vehicle camera 311 and the in-vehicle microphone 312 are provided in the vehicle in which the navigation device 300 is installed, and the state in the vehicle during driving can be recorded.
  • simply shooting the inside of the vehicle simply reduces the opportunity to view the recorded video that is not attractive as content.
  • the navigation apparatus 300 identifies the shooting target based on the content of the utterance by the passenger, and changes the shooting direction of the in-vehicle camera 311. Specifically, when a speech corresponding to a specific passenger is included in the speaker's utterance, the vehicle-mounted camera 311 is pointed at the passenger and photographed. This makes it possible to focus on the passenger at the center of the topic and take pictures of the situation inside the vehicle, improving the attractiveness of the recorded video content.
  • the navigation device 300 in order to identify the passenger to be imaged, information on the passenger is registered in advance, and the vehicle-mounted camera 311 is controlled based on the registered information. ing.
  • the navigation device 300 is a storage-type database (planned passenger database) in which information on the possibility of boarding a vehicle or a person who has a plan (hereinafter referred to as “planned passenger”) is accumulated.
  • An updated database current passenger database that is updated each time a passenger gets on or off the vehicle is created.
  • FIG. 5 is an explanatory diagram schematically showing the contents of a database created by the navigation device.
  • text information 511 such as name information 522, nickname information 522, attribute information 523, keyword 524 corresponding to the prospective passenger, and the face of the passenger Image data 512 and voiceprint data 513 are registered.
  • the navigation device 300 identifies the passenger who has boarded this time from the image or voiceprint of the passenger. And scheduled to board The information (text information 541) corresponding to the passenger stored in the passenger information database 510 is stored in the passenger database 530 this time.
  • the text information 541 includes name information 551, nickname information 552, attribute information 553, a keyword 554 corresponding to the boarding person, and the like.
  • the boarding position information 542 is information on the boarding positions of the passengers who have boarded this time, such as “passenger seat” and “right rear seat”.
  • the control table 560 of the in-vehicle camera 311 corresponding to the position of each seat is recorded in the ROM 402 of the navigation device 300.
  • the control table 500 defines a change amount for changing the shooting direction of the in-vehicle camera 311 when shooting each seat. Specifically, for example, if the reference shooting direction (0 ° direction) is perpendicular to the rear window, the driver's seat is 45 ° left from the reference shooting direction, and the passenger seat is the reference shooting direction. 45 ° to the right, right rear seat 30 ° to the left from the standard shooting direction, and left rear seat 30 ° to the right from the standard shooting direction.
  • the navigation apparatus 300 controls the shooting direction of the in-vehicle camera 311.
  • FIG. 6 and FIG. 7 are flowcharts showing the procedure of the in-vehicle shooting process of the navigation device.
  • the navigation apparatus 300 first causes the person who is scheduled to board the vehicle on which the apparatus is installed to input boarding person information (step S601).
  • the boarding person information corresponds to the text information 511 in FIG. 5 and is information such as the name, nickname, attribute, and keyword of the passenger.
  • FIG. 8 is an explanatory diagram showing an example of an input screen for prospective passenger information.
  • FIG. 8 illustrates an example of registering a person who is scheduled to board using the input device 411 of the navigation apparatus 300.
  • a new registration screen 800 of the person who is scheduled to board is displayed on the new registration screen 800.
  • a name input part 811 for inputting the name of the prospective boarder a nickname input part 812 for entering the nickname of the prospective boarder
  • an attribute input part 813 for entering the attributes of the prospective boarder
  • a keyword input unit 814 for inputting one word is displayed.
  • the newly scheduled passenger who enters the board inputs his / her information in these input sections. Then, the user presses the shooting button 821 and shoots his / her face image using the in-vehicle camera 311. In addition, the sound collection button 822 is pressed and the in-vehicle microphone 312 is used to pick up its own voice. Note that only one of the face image shooting and the sound collection may be used.
  • the person who is registered as a boarding person is not limited to a human but may be an animal such as a dog or a cat. In this case, a person registers in the boarding person information database 510 instead.
  • the new registration screen is displayed on the input screen and the prospective boarder information is input in characters.
  • the prospective boarder information is spoken by voice conversation, and the voice is spoken. It is good also as inputting.
  • audio data can be acquired simultaneously.
  • the face image of the person who will be boarding may be taken automatically by triggering the opening / closing of the door of the vehicle.
  • the navigation apparatus 300 stores the occupant information input in step S601 in the planned passenger database 510 (step S602). Specifically, text information 511 such as the name, nickname, and attribute of the prospective boarder is associated with the face image data 512 of the prospective boarder and voiceprint data 513 (which may be either one) that also extracts voice power. And store it in the planned passenger database.
  • navigation device 300 waits until the passenger gets in the vehicle (step S603: No loop). Whether or not the passenger has boarded the vehicle is determined based on, for example, whether or not the vehicle engine has been started and whether or not the vehicle door has been opened or closed.
  • the in-vehicle camera 311 captures an image of the entire interior of the vehicle, and the in-vehicle microphone 312 collects the sound of the passenger who has boarded each seat (step S604).
  • the shooting direction of the in-vehicle camera 311 is controlled to shoot a range where the faces of the passengers in all seats are located.
  • the passenger is instructed to speak to the in-vehicle microphone 312 provided in each seat. Note that only one of image shooting and sound pickup may be performed.
  • the navigation apparatus 300 extracts the face image and voiceprint of the passenger who has boarded this time from the image and sound captured and collected in step S604 (step S605). And Then, any one of the face images and voiceprints extracted in step S605 is collated with the planned passenger database 510 (step S606), and the face image data 512 and voiceprint data 513 similar to the face images and voiceprints are checked.
  • Step S607 the facial image and voiceprint feature points of the passenger who boarded this time, and the facial image data 5 stored in the planned passenger database 5 ⁇ ⁇ 2 and the feature points of voiceprint data 513 are compared, and it is determined whether or not the face image data 512 or voiceprint data 513 having a similarity higher than a predetermined value is stored.
  • step S607 If similar face image data 512 and voiceprint data 513 are accumulated (step S607: Yes), the text information 511 of the planned passenger corresponding to the face image data 512 and voiceprint data 513 is stored.
  • the text information 541 is stored in the current passenger database 530, and the boarding position information 542 of the passenger is stored in the current passenger database (step S608).
  • the text information 511 stored in the planned passenger database 510 is also retained as it is.
  • step S607 if similar face image data 512 and voiceprint data 513 are not accumulated (step S607: No), the passenger information is input to the passenger (step S609). Then, the input information is stored in the boarding person database 510 as boarding person information (step S610). The passenger information input in step S609 is stored as text information 541 in the current passenger database 530, and the boarding position information 542 of the passenger is stored in the current passenger database (step S611).
  • step S612 Until all face images and voiceprints are collated (step S612: No), the process returns to step S606 and the subsequent processing is repeated. When all face images and voiceprints are collated (step S612: Yes), the process proceeds to step S613 in FIG. Through the processing up to this point, the scheduled passenger database 510 and the current passenger database 530 are created.
  • the arrangement of the passenger may change during traveling, a predetermined timing during traveling, for example, when the door is opened or closed after the vehicle stops is detected
  • the passenger database 530 may be updated this time.
  • the navigation device 300 performs the process from step S604 onward, so that the current passenger database 53 Update 0.
  • the navigation device 300 monitors the voice in the vehicle (step S613), and the words / phrases included in the text information 541 stored in the passenger database 530 this time in the passenger's utterance It is determined whether or not (specific phrase) is included (step S614).
  • Voice monitoring is performed using commonly used voice recognition technology. Specifically, the voice picked up by the in-vehicle microphone 312 is converted into text, and it is determined whether or not the phrase included in the text information 541 is included in the utterance of the passenger.
  • the navigation device 300 determines whether or not there is a priority camera operation (step S616).
  • the priority camera operation is a process that should be performed with priority over the passenger photographing process among processes using the in-vehicle camera 311 such as a drive recorder function and a passenger recognition process.
  • the drive recorder function is prioritized because the possibility of an accident is higher than normal.
  • the priority camera operation condition setting may be determined in advance or may be set by the user. Until this time, the in-vehicle camera 311 may be in a stopped (power off) state.
  • step S616: Yes If there is a priority camera operation (step S616: Yes), the process waits until the camera operation ends. If there is no priority camera operation (step S616: No), refer to the control table 560 recorded in ROM 402 and the shooting direction of the in-vehicle camera 311 at the boarding position of the passenger corresponding to the spoken phrase (Step S617) and photograph the passenger (step S618). On the other hand, when the phrase included in the text information 541 is not uttered (step S614: No), the process returns to step S613, and the voice monitoring is continued.
  • a specific keyword (shooting instruction keyword) is added in addition to the phrase included in the text information 541 that does not necessarily photograph the passenger. You may shoot only when) is spoken.
  • Shooting instruction keywords are, for example, “shooting” and “take a picture”. In addition to the words shown, it is a word that indirectly instructs the photographer, such as “Look at Mr. OO” and “Towards Mr. OO”.
  • Shooting instruction keywords may be determined in advance or registered by the user!
  • shooting may be started when an utterance including a shooting instruction key word is made.
  • the shooting direction of the in-vehicle camera 311 may be returned to the initial state as a timeout.
  • shooting starts when an utterance that includes a shooting instruction keyword is made, and if a phrase included in text information 541 is uttered during shooting, the phrase that corresponds to the spoken phrase is handled. It may be directed toward the passenger. Furthermore, when the utterance containing the shooting instruction keyword is made, the shooting direction is directed toward the speaker, and then, when the phrase included in the text information 541 is spoken, the phrase corresponding to the spoken phrase is dealt with. The shooting direction may be directed toward the passenger to start shooting.
  • the number of passengers to be photographed is not limited to one.
  • the corresponding passenger Shoot sequentially.
  • the passengers corresponding to each word are photographed sequentially.
  • the shooting priority order of each passenger may be determined.
  • the navigation device 300 when instructed to end the shooting (step S619: Yes) or when the shooting has started and the force has elapsed for a predetermined time (step S620: Yes), End (step S621).
  • the shooting end instruction includes, for example, words such as “stop” and “end” that directly indicate the end of shooting, as well as utterances of specific keywords and button operations. If there is no instruction to end the shooting (Step S619: No) and the predetermined time has not elapsed since the start of shooting (Step S620: No), the process returns to Step S618 and the passenger's shooting is continued.
  • step S621 After shooting is completed in step S621, the shooting direction of the in-vehicle camera 311 is returned to the original position (step S622). Until the vehicle finishes running (Step S623: No), Return to step S613 to continue the following processing. When the vehicle finishes traveling (step S623: Yes), the current passenger database is deleted (step S624), and the process of this flow chart is terminated.
  • the planned passenger information is stored in the planned passenger information database 510.
  • the planned passenger information database 510 is not created, and the It is also possible to simply generate the passenger database 530 this time by inputting information.
  • the registration screen as shown in FIG. 8 is displayed for each run, and all the passengers on board at that time enter the passenger information and store it in the current passenger database 530 together with the boarding position information. . Since the passenger database 530 is deleted every time traveling is completed, the passenger needs to input information every time he / she gets on the board, but the navigation device 300 does not need to hold the passenger information.
  • the power of photographing the state in the vehicle can be similarly applied to, for example, a video conference system.
  • the boarding person in the present embodiment is a meeting attendee and the boarding position is the seating position in the meeting room.
  • the shooting target is specified from the sound in the vehicle, and the shooting direction of the in-vehicle camera 311 is controlled toward the shooting target. Accordingly, the shooting direction of the in-vehicle camera 311 can be changed in consideration of the content of the passenger's utterance. For example, a child sitting in the back seat can be called out, the shooting direction of the in-vehicle camera 311 can be directed to the child, and the video can be output to the display 414. This makes it possible to confirm the child in the rear seat without looking back while driving.
  • the navigation device 300 extracts the position of the occupant from the video captured by the in-vehicle camera 311 and the sound collected from the in-vehicle microphone 312.
  • the boarding position of the passenger can be specified without having the passenger input the boarding position.
  • the boarding person information is stored in the boarding person database, it is possible to take a picture without performing complicated processing for the passenger without having to input information about the boarder each time.
  • the camera control method described in the present embodiment is a program prepared in advance. Can be realized by executing it on a computer such as a personal computer or a workstation.
  • This program is recorded on a computer-readable recording medium such as a hard disk, a flexible disk, a CD-ROM, an MO, and a DVD, and is executed by reading the recording medium force by the computer.
  • the program may be a transmission medium that can be distributed through a network such as the Internet.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Studio Devices (AREA)
  • Navigation (AREA)
  • Accessories Of Cameras (AREA)

Abstract

A camera control apparatus (100) comprises an acquiring section (101), a judging section (102), a control section (103), an input section (104), and a detecting section (105). The acquiring section (101) acquires sound around a camera (110). The judging section (102) judges a word or phrase (hereinafter referred to as a specific word or phrase) for specifying a subject to be captured by the camera (110). The control section (103) controls the direction in which the camera (110) captures the subject according to the specific word or and phrase judged by the judging section (102). The input section (104) receives information on candidates for the subject. The detecting section (105) detects the positions of the candidates for the subject.

Description

明 細 書  Specification
カメラ制御装置、カメラ制御方法、カメラ制御プログラムおよび記録媒体 技術分野  CAMERA CONTROL DEVICE, CAMERA CONTROL METHOD, CAMERA CONTROL PROGRAM, AND RECORDING MEDIUM
[0001] この発明は、カメラの撮影方向を制御するカメラ制御装置、カメラ制御方法、カメラ 制御プログラムおよび記録媒体に関する。ただし、この発明の利用は、上述したカメ ラ制御装置、カメラ制御方法、カメラ制御プログラムおよび記録媒体に限られない。 背景技術  The present invention relates to a camera control device, a camera control method, a camera control program, and a recording medium that control a shooting direction of a camera. However, the use of the present invention is not limited to the above-described camera control device, camera control method, camera control program, and recording medium. Background art
[0002] 従来、インターネットや電話回線を用いたテレビ会議システムなどでは、音声が発 せられた方向にカメラを向けて発言者を撮影し、違和感なく会議を進行できるようにし たものがある。このようなテレビ会議システムでは、たとえば、音声発生方向がカメラの 現在の画角内にな 、場合は、雲台による方向変更により画角内に入る力否か判定し 、入ると判定したときは、画角内に入るように雲台を駆動し、画像を表示し、入らない と判定したときは、画角内に入るように、画角を広げ、かつ雲台を駆動し、画像を表示 する(たとえば、下記特許文献 1参照。 ) o  Conventionally, there are video conference systems using the Internet and telephone lines that allow a camera to be pointed at a camera in the direction in which the voice is emitted so that the conference can proceed without a sense of incongruity. In such a video conference system, for example, if the direction of sound generation is within the current angle of view of the camera, it is determined whether or not the force is within the angle of view by changing the direction due to the pan head, Drive the pan head to enter the angle of view, display an image, and if it is determined that it does not enter, widen the angle of view to enter the angle of view and drive the pan head to display the image. (For example, see Patent Document 1 below.) O
[0003] 特許文献 1:特開 2000— 244885号公報  [0003] Patent Document 1: Japanese Patent Application Laid-Open No. 2000-244885
発明の開示  Disclosure of the invention
発明が解決しょうとする課題  Problems to be solved by the invention
[0004] し力しながら、上述した従来技術によれば、発言者による発言の内容を考慮した撮 影をおこなうことができな 、と 、う問題点が一例として挙げられる。従来技術にお!、て 、カメラによって撮影されるのは音声を発している発言者である力 発話の内容によ つては、発言者を撮影対象とするのが必ずしも適切ではない場合がある。たとえば、 カメラの撮影可能範囲内に、発言内容に関連する人物力 ^、る場合には、その人物を 撮影する方が好ま ヽ場合がある。  [0004] However, according to the above-described conventional technology, it is not possible to take a picture in consideration of the content of the speech made by the speaker. According to the prior art, it is not always appropriate to photograph a speaker depending on the content of the power utterance, which is a speaker who is uttering a voice. For example, if the person's power related to the content of the remark is within the camera's shooting range, it may be preferable to shoot that person.
[0005] また、上述した従来技術によれば、発言者以外を撮影対象とした!/ヽ場合は、操作者 によるマニュアル操作によってカメラの撮影方向を変更する必要があるという問題点 がー例として挙げられる。この場合、カメラを操作する操作者が必要となり、撮影が煩 雑になってしまうという問題点が一例として挙げられる。 課題を解決するための手段 [0005] In addition, according to the above-described prior art, in the case of shooting targets other than the speaker! / ヽ, there is a problem that it is necessary to change the shooting direction of the camera by manual operation by the operator. Can be mentioned. In this case, an operator who operates the camera is required, and the problem that photographing becomes complicated is an example. Means for solving the problem
[0006] 上述した課題を解決し、目的を達成するため、請求項 1の発明にかかるカメラ制御 装置は、カメラ周辺の音声を取得する取得手段と、前記取得手段によって取得され た音声から、前記カメラによる撮影対象を特定する語句 (以下、特定語句という)を判 別する判別手段と、前記判別手段によって判別された特定語句に基づいて、前記力 メラの撮影方向を制御する制御手段と、を備えることを特徴とする。  [0006] In order to solve the above-described problems and achieve the object, a camera control device according to the invention of claim 1 includes: an acquisition unit that acquires audio around a camera; and the audio acquired by the acquisition unit, Discriminating means for discriminating a word (hereinafter, referred to as “specific word”) that identifies an object to be photographed by the camera, and control means for controlling the shooting direction of the power camera based on the specific word or phrase discriminated by the discriminating means It is characterized by providing.
[0007] また、請求項 7の発明に力かるカメラ制御方法は、カメラ周辺の音声を取得する取 得工程と、前記取得工程によって取得された音声から、前記カメラによる撮影対象を 特定する語句 (以下、特定語句という)を判別する判別工程と、前記判別工程によつ て判別された特定語句に基づいて、前記カメラの撮影方向を制御する制御工程と、 を含んだことを特徴とする。  [0007] Further, in the camera control method according to the invention of claim 7, an acquisition step of acquiring audio around the camera, and a phrase for specifying a subject to be imaged by the camera from the audio acquired by the acquisition step ( And a control step for controlling the shooting direction of the camera based on the specific phrase determined by the determination step.
[0008] また、請求項 8の発明に力かるカメラ制御プログラムは、請求項 7に記載のカメラ制 御方法をコンピュータに実行させることを特徴とする。  [0008] Further, a camera control program according to the invention of claim 8 causes a computer to execute the camera control method according to claim 7.
[0009] また、請求項 9の発明に力かる記録媒体は、請求項 8に記載のカメラ制御プロダラ ムを記録したコンピュータに読み取り可能なことを特徴とする。  [0009] Further, a recording medium according to the invention of claim 9 is readable by a computer having the camera control program of claim 8 recorded thereon.
図面の簡単な説明  Brief Description of Drawings
[0010] [図 1]図 1は、カメラ制御装置の機能的構成を示すブロック図である。 FIG. 1 is a block diagram showing a functional configuration of a camera control device.
[図 2]図 2は、カメラ制御装置によるカメラ制御処理の手順を示すフローチャートであ る。  FIG. 2 is a flowchart showing a procedure of camera control processing by the camera control device.
[図 3]図 3は、ナビゲーシヨン装置が設置された車両のダッシュボード付近を示す説明 図である。  FIG. 3 is an explanatory view showing the vicinity of a dashboard of a vehicle in which a navigation device is installed.
[図 4]図 4は、ナビゲーシヨン装置のハードウェア構成を示すブロック図である。  FIG. 4 is a block diagram showing a hardware configuration of the navigation device.
[図 5]図 5は、ナビゲーシヨン装置が作成するデータベースの内容を模式的に示す説 明図である。  FIG. 5 is an explanatory diagram schematically showing the contents of a database created by the navigation device.
[図 6]図 6は、ナビゲーシヨン装置の車内撮影処理の手順を示すフローチャートである [図 7]図 7は、ナビゲーシヨン装置の車内撮影処理の手順を示すフローチャートである [図 8]図 8は、搭乗者予定者情報の入力画面の一例を示す説明図である。 [Fig. 6] Fig. 6 is a flowchart showing a procedure of in-vehicle shooting processing of the navigation device. [Fig. 7] Fig. 7 is a flowchart showing a procedure of in-vehicle shooting processing of the navigation device. FIG. 8 is an explanatory diagram showing an example of an input screen for prospective passenger information.
符号の説明  Explanation of symbols
[0011] 100 カメラ制御装置 [0011] 100 camera control device
101 取得部  101 Acquisition Department
102 判別部  102 Discriminator
103 制御部  103 Control unit
104 入力部  104 Input section
105 検出部  105 Detector
110 カメラ  110 Camera
発明を実施するための最良の形態  BEST MODE FOR CARRYING OUT THE INVENTION
[0012] 以下に添付図面を参照して、この発明にかかるカメラ制御装置、カメラ制御方法、 カメラ制御プログラムおよび記録媒体の好適な実施の形態を詳細に説明する。 Exemplary embodiments of a camera control device, a camera control method, a camera control program, and a recording medium according to the present invention will be described below in detail with reference to the accompanying drawings.
[0013] (実施の形態) [0013] (Embodiment)
はじめに、実施の形態に力かるカメラ制御装置 100の機能的構成について説明す る。図 1は、カメラ制御装置の機能的構成を示すブロック図である。カメラ制御装置 10 First, a functional configuration of the camera control device 100 that is useful for the embodiment will be described. FIG. 1 is a block diagram showing a functional configuration of the camera control apparatus. Camera control device 10
0は、取得部 101、判別部 102、制御部 103、入力部 104、検出部 105によって構成 される。 0 is configured by an acquisition unit 101, a determination unit 102, a control unit 103, an input unit 104, and a detection unit 105.
[0014] 取得部 101は、カメラ 110周辺の音声を取得する。カメラ 110周辺の音声とは、たと えば、カメラ 110周辺に位置する人物による発話である。取得部 101は、たとえば、マ イクなどによってカメラ 110周辺の音声を取得する。  The acquisition unit 101 acquires audio around the camera 110. The sound around the camera 110 is, for example, an utterance made by a person located around the camera 110. The acquisition unit 101 acquires the sound around the camera 110 using, for example, a microphone.
[0015] 判別部 102は、取得部 101によって取得された音声から、カメラ 110による撮影対 象を特定する語句(以下、特定語句という)を判別する。判別部 102は、たとえば、後 述する入力部 104に入力される撮影対象の候補に関する情報を、特定語句として判 別する。  The discriminating unit 102 discriminates a phrase (hereinafter referred to as a specific phrase) that identifies a subject to be photographed by the camera 110 from the sound acquired by the acquiring unit 101. The discriminating unit 102 discriminates, for example, information relating to a candidate for photographing input to the input unit 104 described later as a specific phrase.
[0016] 制御部 103は、判別部 102によって判別された特定語句に基づいて、カメラ 110の 撮影方向を制御する。制御部 103は、たとえば、判別部 102によって撮影対象の候 補に関する情報と略一致する語句が判別された場合、後述する検出部 105によって 検出された撮影対象の候補の位置にカメラ 110の撮影方向を向ける。略一致する語 句とは、撮影対象の候補に関する情報として入力された語句と同一または類似する 語句である。 The control unit 103 controls the shooting direction of the camera 110 based on the specific phrase determined by the determination unit 102. For example, when the determination unit 102 determines a word or phrase that substantially matches information related to the candidate for the shooting target, the control unit 103 determines the shooting direction of the camera 110 at the position of the candidate for the shooting target detected by the detection unit 105 described later. Turn. Words that match A phrase is a phrase that is the same as or similar to a phrase that is input as information about a candidate to be photographed.
[0017] 入力部 104は、撮影対象の候補に関する情報の入力を受け付ける。撮影対象の候 補に関する情報とは、たとえば、撮影対象の候補の名称情報 (氏名や愛称など)ゃ属 性情報などである。また、入力部 104には、撮影対象の候補の画像や音声が入力さ れてもよい。  The input unit 104 receives input of information related to a candidate for photographing. The information regarding the candidates for the imaging target includes, for example, name information (name, nickname, etc.) of the candidate for the imaging target and attribute information. In addition, the input unit 104 may receive a candidate image or sound to be captured.
[0018] 検出部 105は、撮影対象の候補の位置を検出する。検出部 105は、たとえば、入 力部 104に入力された撮影対象の候補の画像と、カメラ 110によって撮影された映 像とを照合して、撮影対象の候補の位置を検出する。ここで、撮影対象の位置とは、 たとえば、カメラ 110が車両に設置されている場合には、撮影対象の着席した座席の 位置であり、検出部 105は、搭乗者がどの座席に着席したかを検出する。また、撮影 対象の位置とは、カメラ 110からの相対方向、相対方位などであってもよい。検出部 1 05は、具体的には、たとえば、カメラ 110によって撮影された映像中に、撮影対象の 候補の画像との類似度が所定値以上の物体が映っていた場合、その物体の位置を 撮影対象の候補の位置とする。  [0018] The detection unit 105 detects the position of a candidate for imaging. The detection unit 105, for example, collates the image of the photographing target candidate input to the input unit 104 with the image photographed by the camera 110, and detects the position of the photographing target candidate. Here, for example, when the camera 110 is installed in a vehicle, the shooting target position is the position of the seat where the shooting target is seated, and the detection unit 105 determines which seat the passenger is seated on. Is detected. Further, the position to be imaged may be a relative direction from the camera 110, a relative direction, or the like. Specifically, for example, when an object whose similarity with a candidate image to be captured is reflected in a video captured by the camera 110, the detection unit 105 detects the position of the object. It is set as the position of a candidate for photographing.
[0019] また、検出部 105は、たとえば、入力部 104に入力された撮影対象の候補の音声と 、取得部 101によって取得された音声とを照合して撮影対象の候補の位置を検出す る。具体的には、たとえば、取得部 101によって取得された音声中に、撮影対象の候 補の音声との類似度が所定値以上の音声が含まれて!/、た場合、その音声が発音さ れて 、る位置を撮影対象の候補の位置とする。  In addition, the detection unit 105 detects the position of the candidate for the photographing target by collating the voice of the candidate for the photographing target input to the input unit 104 with the voice acquired by the acquiring unit 101, for example. . Specifically, for example, if the voice acquired by the acquisition unit 101 includes a voice whose similarity with the candidate voice to be shot is a predetermined value or more! /, The voice is pronounced. Therefore, this position is set as the position of the candidate for the photographing target.
[0020] また、カメラ 110は、車両内部に設置されていてもよい。この場合、たとえば、取得 部 101は、車両の搭乗者の発話を取得し、判別部 102は、撮影対象となる搭乗者に 関する情報を特定語句として判別し、制御部 103は、特定語句によって特定される 搭乗者の乗車位置にカメラ 110の撮影方向を向ける。また、入力部 104には、たとえ ば、搭乗者に関する情報が入力され、検出部 105は、各搭乗者がどの座席に着席し たかを検出する。  [0020] The camera 110 may be installed inside the vehicle. In this case, for example, the acquisition unit 101 acquires the utterance of the passenger of the vehicle, the determination unit 102 determines information about the passenger to be photographed as a specific phrase, and the control unit 103 specifies the specific phrase by the specific phrase. Aim the shooting direction of the camera 110 at the boarding position of the passenger. Further, for example, information related to the passenger is input to the input unit 104, and the detection unit 105 detects which seat each passenger is seated in.
[0021] つぎに、カメラ制御装置 100によるカメラ 110の制御処理について説明する。図 2は 、カメラ制御装置によるカメラ制御処理の手順を示すフローチャートである。図 2のフ ローチャートにおいて、まず、入力部 104に対して、撮影対象の候補に関する情報の 入力がおこなわれる (ステップ S201)。また、検出部 105によって、撮影対象の候補 の位置を検出する (ステップ S 202)。 Next, control processing of the camera 110 by the camera control device 100 will be described. FIG. 2 is a flowchart showing a procedure of camera control processing by the camera control apparatus. Figure 2 In the flowchart, first, information relating to a candidate for an imaging target is input to the input unit 104 (step S201). Further, the position of the candidate for photographing is detected by the detection unit 105 (step S 202).
[0022] つぎに、取得部 101によって、カメラ 110周辺の音声を取得する(ステップ S 203)。  Next, the sound around the camera 110 is acquired by the acquisition unit 101 (step S 203).
つづいて、判別部 102によって、ステップ S203で取得された音声から、特定語句を 判別する (ステップ S 204)。このときの特定語句とは、撮影対象の候補に関する情報 である。これにより、撮影対象の候補の中から撮影すべき撮影対象が特定される。そ して、制御部 103によって、ステップ S202で検出した撮影対象の位置にカメラ 110 の撮影方向を制御して (ステップ S205)、本フローチャートによる処理を終了する。  Subsequently, the determination unit 102 determines a specific phrase from the voice acquired in step S203 (step S204). The specific word / phrase at this time is information on the candidate to be photographed. Thereby, a photographing target to be photographed is specified from among photographing target candidates. Then, the control unit 103 controls the shooting direction of the camera 110 to the position of the shooting target detected in step S202 (step S205), and the process according to this flowchart ends.
[0023] 以上説明したように、カメラ制御装置 100によれば、カメラ周辺の音声力も撮影対象 を特定し、撮影対象に向けてカメラの撮影方向を制御する。これにより、カメラ周辺の 音声の内容を考慮して、カメラ 110の撮影方向を変更することができる。  [0023] As described above, according to the camera control device 100, the sound force around the camera also specifies the shooting target and controls the shooting direction of the camera toward the shooting target. Thereby, it is possible to change the shooting direction of the camera 110 in consideration of the audio content around the camera.
[0024] また、カメラ制御装置 100は、あらかじめ入力された撮影対象の候補に関する情報 を特定語句として撮影対象を特定するので、より精度良く撮影対象を特定することが できる。さらに、撮影対象の候補の画像や音声から撮影対象の位置を検出するので 、より精度良くカメラ 110の撮影方向を制御することができる。  [0024] In addition, since the camera control device 100 specifies the shooting target using information relating to the shooting target candidates input in advance as specific words, the shooting target can be specified with higher accuracy. Furthermore, since the position of the shooting target is detected from the image and sound of the shooting target candidate, the shooting direction of the camera 110 can be controlled with higher accuracy.
実施例  Example
[0025] つぎに、上述した実施の形態に力かるカメラ制御装置 100の実施例について説明 する。以下の実施例においては、カメラ制御装置 100を、車両に搭載されたナビゲー シヨン装置 300に適用した場合について説明する。  Next, an example of the camera control device 100 that works on the above-described embodiment will be described. In the following embodiment, a case where the camera control device 100 is applied to a navigation device 300 mounted on a vehicle will be described.
[0026] (ナビゲーシヨン装置 300の周辺機器構成)  [0026] (Peripheral device configuration of the navigation device 300)
はじめに、ナビゲーシヨン装置 300の周辺機器構成について説明する。図 3は、ナ ピゲーシヨン装置が設置された車両のダッシュボード付近を示す説明図である。ナビ ゲーシヨン装置 300は、車両のダッシュボードに設置されている。ナビゲーシヨン装置 300は、本体部 Mおよび表示部(ディスプレイ) Dによって構成され、表示部 Dには車 両の現在地点や地図情報、現在時刻などが表示される。  First, the peripheral device configuration of the navigation device 300 will be described. FIG. 3 is an explanatory diagram showing the vicinity of the dashboard of the vehicle in which the navigation device is installed. The navigation device 300 is installed on the dashboard of the vehicle. The navigation device 300 includes a main body M and a display unit (display) D. The display unit D displays the current location of the vehicle, map information, current time, and the like.
[0027] また、ナビゲーシヨン装置 300には、ノ ックミラー周辺に設置された車載用カメラ 31 1、サンノ ィザ一に設置された車載用マイク 312が接続されている。車載用カメラ 311 は、撮影方向を変更可能であり、車外前方および車両内の各部 (搭乗者など)を撮影 する。以下、車載用カメラ 311は、動画および静止画を撮影可能なカメラであるものと する力 静止画のみを撮影可能なカメラであってもよい。 In addition, the navigation apparatus 300 is connected to an in-vehicle camera 311 installed in the vicinity of the knock mirror and an in-vehicle microphone 312 installed in the sun-noiser. Car camera 311 The shooting direction can be changed, and images are taken in front of the vehicle and in the vehicle (passengers, etc.). Hereinafter, the in-vehicle camera 311 may be a camera capable of capturing only still images, assuming that the camera 311 is capable of capturing moving images and still images.
[0028] 車載用マイク 312は、車両内の音声が入力され、ナビゲーシヨン装置 300の音声入 力による操作や車両内の様子を記録する際などに用いられる。なお、車載用マイク 3 12の位置は、サンバイザーに限ることなぐ車両内の音声を効率的に入力できる位 置にあればよい。また、車載用カメラ 311および車載用マイク 312は、車両に複数設 置されていてもよいし、さらに、固定式ではなく可動式であってもよい。本実施例では 、車載用マイク 312は、各搭乗者の座席ごとに設けられているものとする。  [0028] The in-vehicle microphone 312 is used when an in-vehicle sound is input and an operation by the sound input of the navigation device 300 or a state in the vehicle is recorded. It should be noted that the position of the in-vehicle microphone 312 is not limited to the sun visor as long as it can efficiently input voice in the vehicle. Further, a plurality of in-vehicle cameras 311 and in-vehicle microphones 312 may be installed in the vehicle, and may be movable rather than fixed. In the present embodiment, it is assumed that the in-vehicle microphone 312 is provided for each passenger's seat.
[0029] ナビゲーシヨン装置 300は、 目的地点までの経路探索および情報記録をおこなう他 、ドライブ中の車両内の様子を記録する車内撮影機能を有している。車内撮影機能 は、車載用カメラ 311や車載用マイク 312で車両内の映像および音声を記録する。 車内撮影機能によって記録された映像および音声はナビゲーシヨン装置 300の記録 媒体 (後述する磁気ディスク 405、光ディスク 407)に記録される。また、記録された映 像および音声を外部記録媒体に記録して、自宅のテレビなどで楽しめるようにしても よい。  [0029] The navigation device 300 has an in-vehicle shooting function for recording a state in the vehicle being driven, in addition to searching for a route to the destination and recording information. The in-vehicle shooting function records in-vehicle video and audio with the in-vehicle camera 311 and the in-vehicle microphone 312. The video and audio recorded by the in-vehicle shooting function are recorded on a recording medium (magnetic disk 405 and optical disk 407 described later) of the navigation device 300. Further, the recorded video and sound may be recorded on an external recording medium so that they can be enjoyed on a home TV or the like.
[0030] (ナビゲーシヨン装置 300のハードウェア構成)  [0030] (Hardware configuration of navigation device 300)
つぎに、ナビゲーシヨン装置 300のハードウェア構成について説明する。図 4は、ナ ピゲーシヨン装置のハードウェア構成を示すブロック図である。図 4において、ナビゲ ーシヨン装置 300は、 CPU401と、 ROM402と、 RAM (メモリ) 403と、磁気ディスク ドライブ 404と、磁気ディスク 405と、光ディスクドライブ 406と、光ディスク 407と、音 声 IZF (インターフェース) 408と、マイク 409と、スピーカ 410と、入力デバイス 411と 、映像 IZF412と、カメラ 413と、ディスプレイ 414と、通信 IZF415と、 GPSユニット 416と、各種センサ 417と、外部接続用 IZF418とを備えている。また、各構成部 40 1-418はバス 420によってそれぞれ接続されて 、る。  Next, the hardware configuration of the navigation device 300 will be described. FIG. 4 is a block diagram showing the hardware configuration of the navigation device. In FIG. 4, the navigation device 300 includes a CPU 401, a ROM 402, a RAM (memory) 403, a magnetic disk drive 404, a magnetic disk 405, an optical disk drive 406, an optical disk 407, and an audio IZF (interface) 408. A microphone 409, a speaker 410, an input device 411, a video IZF 412, a camera 413, a display 414, a communication IZF 415, a GPS unit 416, various sensors 417, and an external connection IZF 418. . Each component 40 1-418 is connected by a bus 420.
[0031] まず、 CPU401は、ナビゲーシヨン装置 300の全体の制御を司る。 ROM402は、 ブートプログラム、通信プログラム、データベース作成プログラム、データ解析プログ ラムなどのプログラムを記録している。 RAM403は、 CPU401のワークエリアとして 使用される。 First, the CPU 401 governs overall control of the navigation device 300. The ROM 402 records programs such as a boot program, a communication program, a database creation program, and a data analysis program. RAM403 is the work area of CPU401 used.
[0032] 磁気ディスクドライブ 404は、 CPU401の制御に従って磁気ディスク 405に対する データの読み取り Z書き込みを制御する。磁気ディスク 405は、磁気ディスクドライブ 404の制御で書き込まれたデータを記録する。磁気ディスク 405としては、たとえば、 HD (ノヽードディスク)や FD (フレキシブルディスク)を用いることができる。  The magnetic disk drive 404 controls reading and writing of data to the magnetic disk 405 according to the control of the CPU 401. The magnetic disk 405 records data written under the control of the magnetic disk drive 404. As the magnetic disk 405, for example, HD (node disk) or FD (flexible disk) can be used.
[0033] 光ディスクドライブ 406は、 CPU401の制御に従って光ディスク 407に対するデー タの読み取り Z書き込みを制御する。光ディスク 407は、光ディスクドライブ 406の制 御に従ってデータが読み出される着脱自在な記録媒体である。光ディスク 407は、書 き込み可能な記録媒体を利用することもできる。また、この着脱可能な記録媒体とし て、光ディスク 407のほ力 MO、メモリカードなどであってもよい。  The optical disk drive 406 controls data reading / writing to the optical disk 407 according to the control of the CPU 401. The optical disk 407 is a detachable recording medium from which data is read according to the control of the optical disk drive 406. The optical disc 407 can also use a writable recording medium. Further, the removable recording medium may be a power MO of the optical disc 407, a memory card, or the like.
[0034] 磁気ディスク 405または光ディスク 407に記録される情報の一例として、経路探索' 経路誘導などに用いる地図データが挙げられる。地図データは、建物、河川、地表 面などの地物 (フィーチャ)を表す背景データと、道路の形状を表す道路形状データ とを有しており、ディスプレイ 414の表示画面において 2次元または 3次元に描画され る。ナビゲーシヨン装置 300が経路誘導中の場合は、地図データと後述する GPSュ ニット 416によって取得された自車の現在地点とが重ねて表示されることとなる。  [0034] As an example of information recorded on the magnetic disk 405 or the optical disk 407, there is map data used for route search and route guidance. The map data includes background data representing features (features) such as buildings, rivers, and the ground surface, and road shape data representing the shape of the road, and is displayed in two or three dimensions on the display screen of the display 414. It is drawn. When the navigation device 300 is guiding a route, the map data and the current location of the host vehicle acquired by the GPS unit 416 described later are displayed in an overlapping manner.
[0035] 音声 IZF408は、音声入力用のマイク 409 (たとえば、図 3の車載用マイク 312)お よび音声出力用のスピーカ 410に接続される。マイク 409に受音された音声は、音声 IZF408内で AZD変換される。また、スピーカ 410からは音声が出力される。なお、 マイク 409から入力された音声は、音声データとして磁気ディスク 405ある 、は光ディ スク 407に記録可能である。  Audio IZF 408 is connected to audio input microphone 409 (for example, in-vehicle microphone 312 in FIG. 3) and audio output speaker 410. The sound received by the microphone 409 is AZD converted in the sound IZF 408. In addition, sound is output from the speaker 410. Note that the sound input from the microphone 409 can be recorded on the magnetic disk 405 or the optical disk 407 as sound data.
[0036] 入力デバイス 411は、文字、数値、各種指示などの入力のための複数のキーを備 えたリモコン、キーボード、マウス、タツチパネルなどが挙げられる。さらに、入力デバ イス 411は、デジタルカメラや携帯電話端末などの他の情報処理端末を接続し、デ ータの入出力をおこなうことができる。  [0036] Examples of the input device 411 include a remote controller, a keyboard, a mouse, and a touch panel that are provided with a plurality of keys for inputting characters, numerical values, various instructions, and the like. Further, the input device 411 can be connected to another information processing terminal such as a digital camera or a mobile phone terminal to input / output data.
[0037] 映像 IZF412は、映像入力用のカメラ 413 (たとえば、図 3の車載用カメラ 311)お よび映像出力用のディスプレイ 414と接続される。映像 IZF412は、具体的には、た とえば、ディスプレイ 414全体の制御をおこなうグラフィックコントローラと、即時表示 可能な画像情報を一時的に記録する VRAM (Video RAM)などのバッファメモリと 、グラフィックコントローラから出力される画像データに基づいて、ディスプレイ 414を 表示制御する制御 ICなどによって構成される。 The video IZF 412 is connected to a video input camera 413 (for example, the in-vehicle camera 311 in FIG. 3) and a video output display 414. Video IZF412 is specifically a graphic controller that controls the entire display 414 and an immediate display, for example. It consists of a buffer memory such as VRAM (Video RAM) that temporarily records possible image information and a control IC that controls display 414 based on image data output from the graphic controller.
[0038] カメラ 413は、車両内外の画像 (動画を含む)を撮影し、画像データとして出力する 。カメラ 413で撮影された画像は、画像データとして磁気ディスク 405あるいは光ディ スク 407に記録することができる。この画像データは、ディスプレイ 414で出力する他 、記録媒体に記録したり、ネットワークを介して送信するなどして、他の情報処理端末 で利用することができる。  [0038] The camera 413 captures images (including moving images) inside and outside the vehicle and outputs them as image data. An image taken by the camera 413 can be recorded on the magnetic disk 405 or the optical disk 407 as image data. This image data can be used by other information processing terminals by outputting it on the display 414, recording it on a recording medium, or transmitting it via a network.
[0039] ディスプレイ 414には、アイコン、カーソル、メニュー、ウィンドウ、あるいは文字や画 像などの各種データが表示される。このディスプレイ 414は、たとえば、 CRT、 TFT 液晶ディスプレイ、プラズマディスプレイなどを採用することができる。  The display 414 displays icons, cursors, menus, windows, or various data such as characters and images. As this display 414, for example, a CRT, a TFT liquid crystal display, a plasma display, or the like can be adopted.
[0040] 通信 IZF415は、無線を介してインターネットなどの通信網に接続され、この通信 網と CPU401とのインターフェースとして機能する。通信網には、 LAN, WAN,公 衆回線網や携帯電話網などがある。  Communication IZF 415 is connected to a communication network such as the Internet via radio, and functions as an interface between this communication network and CPU 401. Communication networks include LAN, WAN, public line network and mobile phone network.
[0041] 0?3ュ-ット416は、 GPS衛星からの電波を受信し、車両の現在地点(ナビゲーシ ヨン装置 300の現在地点)を示す情報を出力する。 GPSユニット 416の出力情報は、 後述する各種センサ 417の出力値とともに、 CPU401による車両の現在地点の算出 に際して利用される。現在地点を示す情報は、たとえば緯度'経度、高度などの、地 図データ上の 1点を特定する情報である。  [0041] 0-3 416 receives a radio wave from a GPS satellite and outputs information indicating the current position of the vehicle (the current position of the navigation device 300). The output information of the GPS unit 416 is used when the CPU 401 calculates the current position of the vehicle together with output values of various sensors 417 described later. The information indicating the current location is information specifying one point on the map data, for example, latitude'longitude and altitude.
[0042] 各種センサ 417は、車速センサや加速度センサ、角速度センサなどの、車両の位 置や挙動を判断することが可能な情報を出力する。各種センサ 417の出力値は、 CP U401による現在地点の算出や、速度や方位の変化量の測定に用いられる。  [0042] Various sensors 417 output information that can determine the position and behavior of the vehicle, such as a vehicle speed sensor, an acceleration sensor, and an angular velocity sensor. The output values of the various sensors 417 are used for calculation of the current position by the CCU 401 and measurement of speed and direction change.
[0043] 外部接続用 IZF418は、オーディオ装置や車内空調装置など、外部の機器と接続 するためのインターフェース類である。外部接続用 IZF418は、たとえば、専用の接 続ケーブルのポート、赤外線通信用ポートなどによって構成される。  [0043] The IZF 418 for external connection is an interface for connecting to an external device such as an audio device or a vehicle air conditioner. The external connection IZF 418 is configured by, for example, a dedicated connection cable port and an infrared communication port.
[0044] また、実施の形態に力かるカメラ制御装置 100の構成のうち、取得部 101は音声 I /F408、マイク 409によって、半 IJ另 U咅 102、検出咅 は CPU401によって、 ffilj御 部 103は CPU401、映像 IZF412によって、入力部 104は入力デバイス 411によつ て、それぞれの機能を実現する。 [0044] In addition, in the configuration of the camera control apparatus 100 that is relevant to the embodiment, the acquisition unit 101 is a voice I / F 408, a microphone 409, a half IJ additional U 咅 102, a detection 咅 is a CPU 401, and an ffilj control unit 103. The input unit 104 is connected to the input device 411 by the CPU 401 and the video IZF 412. To realize each function.
[0045] (ナビゲーシヨン装置 300による車内撮影処理)  [0045] (In-car shooting process using navigation device 300)
つづいて、ナビゲーシヨン装置 300による車内撮影処理について説明する。前述の ように、ナビゲーシヨン装置 300が設置された車両内には、車載用カメラ 311および 車載用マイク 312が設けられており、ドライブ中の車両内の様子を記録することがで きる。ここで、車両内の様子を撮影する際に、ただ漫然と車両内を撮影するのみでは 、コンテンツとしての魅力に乏しぐ記録された映像を視聴する機会は低下してしまう  Next, in-vehicle shooting processing by the navigation device 300 will be described. As described above, the in-vehicle camera 311 and the in-vehicle microphone 312 are provided in the vehicle in which the navigation device 300 is installed, and the state in the vehicle during driving can be recorded. Here, when shooting inside the vehicle, simply shooting the inside of the vehicle simply reduces the opportunity to view the recorded video that is not attractive as content.
[0046] このため、ナビゲーシヨン装置 300は、搭乗者による発話の内容に基づいて撮影対 象を特定し、車載用カメラ 311の撮影方向を変更する。具体的には、発言者の発話 に特定の搭乗者に対応する言葉が含まれる場合に、その搭乗者に車載用カメラ 311 を向けて撮影する。これにより、話題の中心にいる搭乗者に焦点をあてて車両内の 様子を撮影することができ、記録された映像のコンテンツとしての魅力を向上させるこ とがでさる。 For this reason, the navigation apparatus 300 identifies the shooting target based on the content of the utterance by the passenger, and changes the shooting direction of the in-vehicle camera 311. Specifically, when a speech corresponding to a specific passenger is included in the speaker's utterance, the vehicle-mounted camera 311 is pointed at the passenger and photographed. This makes it possible to focus on the passenger at the center of the topic and take pictures of the situation inside the vehicle, improving the attractiveness of the recorded video content.
[0047] ここで、ナビゲーシヨン装置 300では、撮影対象となる搭乗者を特定するため、あら 力じめ搭乗者に関する情報を登録しておき、その登録情報に基づいて車載用カメラ 311を制御している。具体的には、ナビゲーシヨン装置 300は、車両に搭乗する可能 性や予定がある人物 (以下、「搭乗予定者」という)に関する情報が累積的に蓄積され る蓄積型データベース (搭乗予定者データベース)と、車両に搭乗者が乗降するごと に更新される更新型データベース (今回搭乗者データベース)を作成する。  Here, in the navigation device 300, in order to identify the passenger to be imaged, information on the passenger is registered in advance, and the vehicle-mounted camera 311 is controlled based on the registered information. ing. Specifically, the navigation device 300 is a storage-type database (planned passenger database) in which information on the possibility of boarding a vehicle or a person who has a plan (hereinafter referred to as “planned passenger”) is accumulated. An updated database (current passenger database) that is updated each time a passenger gets on or off the vehicle is created.
[0048] 図 5は、ナビゲーシヨン装置が作成するデータベースの内容を模式的に示す説明 図である。搭乗予定者データベース 510には、車両への搭乗者予定者の氏名情報 5 21や愛称情報 522、属性情報 523、その搭乗予定者に対応するキーワード 524など のテキスト情報 511と、その搭乗者の顔画像データ 512、声紋データ 513が登録され る。  FIG. 5 is an explanatory diagram schematically showing the contents of a database created by the navigation device. In the prospective passenger database 510, text information 511 such as name information 522, nickname information 522, attribute information 523, keyword 524 corresponding to the prospective passenger, and the face of the passenger Image data 512 and voiceprint data 513 are registered.
[0049] 今回搭乗者データベース 530には、今回搭乗した搭乗者に対応したテキスト情報 5 41と、その乗車位置情報 542とが登録される。後述するように、ナビゲーシヨン装置 3 00は、搭乗者の画像や声紋から今回搭乗した搭乗者を特定する。そして、搭乗予定 者情報データベース 510に格納されている、その搭乗者に対応する情報 (テキスト情 報 541)を今回搭乗者データベース 530に格納する。 [0049] In the current passenger database 530, text information 541 corresponding to the passenger who has boarded this time and the boarding position information 542 are registered. As will be described later, the navigation device 300 identifies the passenger who has boarded this time from the image or voiceprint of the passenger. And scheduled to board The information (text information 541) corresponding to the passenger stored in the passenger information database 510 is stored in the passenger database 530 this time.
[0050] テキスト情報 541は、搭乗予定者データベース 510と同様に、氏名情報 551や愛 称情報 552、属性情報 553、その搭乗予定者に対応するキーワード 554などによつ て構成される。また、乗車位置情報 542は、今回搭乗した搭乗者の乗車位置、たとえ ば、「助手席」や「右後部座席」などの情報である。  [0050] Like the boarding person database 510, the text information 541 includes name information 551, nickname information 552, attribute information 553, a keyword 554 corresponding to the boarding person, and the like. The boarding position information 542 is information on the boarding positions of the passengers who have boarded this time, such as “passenger seat” and “right rear seat”.
[0051] 乗車位置情報 542に関連して、ナビゲーシヨン装置 300の ROM402には、各座席 の位置に対応する車載用カメラ 311の制御テーブル 560が記録されている。制御テ 一ブル 500には、各座席を撮影する際に車載用カメラ 311の撮影方向を変更する変 更量が定義されている。具体的には、たとえば、基準となる撮影方向(0° 方向)をリ ァウィンド一に対して垂直とすると、運転席は基準となる撮影方向から左 45° 方向、 助手席は基準となる撮影方向から右 45° 方向、右後部座席は基準となる撮影方向 から左 30° 方向、左後部座席は基準となる撮影方向から右 30° 方向となる。  [0051] In relation to the boarding position information 542, the control table 560 of the in-vehicle camera 311 corresponding to the position of each seat is recorded in the ROM 402 of the navigation device 300. The control table 500 defines a change amount for changing the shooting direction of the in-vehicle camera 311 when shooting each seat. Specifically, for example, if the reference shooting direction (0 ° direction) is perpendicular to the rear window, the driver's seat is 45 ° left from the reference shooting direction, and the passenger seat is the reference shooting direction. 45 ° to the right, right rear seat 30 ° to the left from the standard shooting direction, and left rear seat 30 ° to the right from the standard shooting direction.
[0052] 以上のようなデータベースを用いて、ナビゲーシヨン装置 300は、車載用カメラ 311 の撮影方向を制御する。  Using the database as described above, the navigation apparatus 300 controls the shooting direction of the in-vehicle camera 311.
[0053] 図 6および図 7は、ナビゲーシヨン装置の車内撮影処理の手順を示すフローチヤ一 トである。図 6のフローチャートにおいて、ナビゲーシヨン装置 300は、まず、自装置が 搭載された車両の搭乗予定者に、搭乗予定者情報を入力させる (ステップ S601)。 搭乗予定者情報は、図 5のテキスト情報 511に対応し、搭乗者の氏名や愛称、属性、 キーワードなどの情報である。搭乗者予定者情報の入力は、ナビゲーシヨン装置 300 の入力デバイス 411でおこなう他、各搭乗予定者の自宅の PC (パーソナルコンビュ →)や携帯電話端末など力もネットワークを介しておこなうこととしてもょ 、。  FIG. 6 and FIG. 7 are flowcharts showing the procedure of the in-vehicle shooting process of the navigation device. In the flowchart of FIG. 6, the navigation apparatus 300 first causes the person who is scheduled to board the vehicle on which the apparatus is installed to input boarding person information (step S601). The boarding person information corresponds to the text information 511 in FIG. 5 and is information such as the name, nickname, attribute, and keyword of the passenger. In addition to the input of prospective passenger information using the input device 411 of the navigation device 300, it is also possible to use the power of each prospective passenger's home PC (Personal Review →) or mobile phone terminal via the network. .
[0054] 図 8は、搭乗者予定者情報の入力画面の一例を示す説明図である。図 8は、ナビゲ ーシヨン装置 300の入力デバイス 411を用 、て搭乗予定者の登録をおこなう場合を 例として説明する。図 8において、ナビゲーシヨン装置 300のディスプレイ 414には、 搭乗予定者の新規登録画面 800が表示されている。新規登録画面 800には、搭乗 予定者の氏名を入力する氏名入力部 811、搭乗予定者の愛称を入力する愛称入力 部 812、搭乗予定者の属性を入力する属性入力部 813、搭乗予定者と関連するキ 一ワードを入力するキーワード入力部 814が表示されている。 FIG. 8 is an explanatory diagram showing an example of an input screen for prospective passenger information. FIG. 8 illustrates an example of registering a person who is scheduled to board using the input device 411 of the navigation apparatus 300. In FIG. 8, on the display 414 of the navigation device 300, a new registration screen 800 of the person who is scheduled to board is displayed. On the new registration screen 800, a name input part 811 for inputting the name of the prospective boarder, a nickname input part 812 for entering the nickname of the prospective boarder, an attribute input part 813 for entering the attributes of the prospective boarder, Related key A keyword input unit 814 for inputting one word is displayed.
[0055] 新規に登録する搭乗予定者は、これらの入力部に自己の情報を入力する。そして 、撮影ボタン 821を押下して、車載用カメラ 311を用いて自己の顔画像を撮影する。 また、収音ボタン 822押下して、車載用マイク 312を用いて自己の音声を収音する。 なお、顔画像の撮影と音声の収音は、どちらか一方のみとしてもよい。また、搭乗予 定者として登録するのは、人間には限らず、犬や猫などの動物であってもよい。この 場合、搭乗予定者情報データベース 510への登録は、人間が代わりにおこなう。  [0055] The newly scheduled passenger who enters the board inputs his / her information in these input sections. Then, the user presses the shooting button 821 and shoots his / her face image using the in-vehicle camera 311. In addition, the sound collection button 822 is pressed and the in-vehicle microphone 312 is used to pick up its own voice. Note that only one of the face image shooting and the sound collection may be used. The person who is registered as a boarding person is not limited to a human but may be an animal such as a dog or a cat. In this case, a person registers in the boarding person information database 510 instead.
[0056] なお、上述したように、入力画面に新規登録画面を表示して搭乗予定者情報を文 字で入力させる他、たとえば、音声対話によって、搭乗予定者情報を発話させて、音 声で入力することとしてもよい。この方法によれば、音声データも同時に取得すること ができる。また、搭乗予定者の顔画像は、車両のドアの開閉などをトリガーとして、自 動的に撮影することとしてちよい。  [0056] As described above, the new registration screen is displayed on the input screen and the prospective boarder information is input in characters. For example, the prospective boarder information is spoken by voice conversation, and the voice is spoken. It is good also as inputting. According to this method, audio data can be acquired simultaneously. In addition, the face image of the person who will be boarding may be taken automatically by triggering the opening / closing of the door of the vehicle.
[0057] 図 6説明に戻り、ナビゲーシヨン装置 300は、ステップ S601で入力された搭乗者情 報を搭乗予定者データベース 510に格納する (ステップ S602)。具体的には、搭乗 予定者の氏名や愛称、属性などのテキスト情報 511と、搭乗予定者の顔画像データ 512および音声力も抽出した声紋データ 513 (どちらか一方のみでもよい)とを関連 付けて、搭乗予定者データベースに格納する。  Returning to the description of FIG. 6, the navigation apparatus 300 stores the occupant information input in step S601 in the planned passenger database 510 (step S602). Specifically, text information 511 such as the name, nickname, and attribute of the prospective boarder is associated with the face image data 512 of the prospective boarder and voiceprint data 513 (which may be either one) that also extracts voice power. And store it in the planned passenger database.
[0058] つぎに、ナビゲーシヨン装置 300は、車両に搭乗者が乗車するまで待機する (ステ ップ S603 :Noのループ)。搭乗者が乗車したか否かは、たとえば、車両のエンジン が始動したか否かや、車両のドアが開閉した力否力などによって判断する。搭乗者が 乗車すると (ステップ S603 : Yes)、車載用カメラ 311で車両内全体の画像を撮影し、 車載用マイク 312で各座席に搭乗した搭乗者の音声を収音する (ステップ S604)。 具体的には、車載用カメラ 311の撮影方向を制御して、全座席の搭乗者の顔が位置 する範囲を撮影する。また、各座席に設けられている車載用マイク 312に向カゝつて発 話するよう、搭乗者に指示する。なお、画像の撮影と音声の収音は、一方のみをおこ なってもよい。  Next, navigation device 300 waits until the passenger gets in the vehicle (step S603: No loop). Whether or not the passenger has boarded the vehicle is determined based on, for example, whether or not the vehicle engine has been started and whether or not the vehicle door has been opened or closed. When the passenger gets on (step S603: Yes), the in-vehicle camera 311 captures an image of the entire interior of the vehicle, and the in-vehicle microphone 312 collects the sound of the passenger who has boarded each seat (step S604). Specifically, the shooting direction of the in-vehicle camera 311 is controlled to shoot a range where the faces of the passengers in all seats are located. In addition, the passenger is instructed to speak to the in-vehicle microphone 312 provided in each seat. Note that only one of image shooting and sound pickup may be performed.
[0059] つぎに、ナビゲーシヨン装置 300は、ステップ S604で撮影 ·収音した画像および音 声から、今回搭乗した搭乗者の顔画像および声紋を抽出する (ステップ S605)。そし て、ステップ S605で抽出された顔画像および声紋のうち、任意の 1つを搭乗予定者 データベース 510と照合し (ステップ S606)、その顔画像および声紋と類似する顔画 像データ 512および声紋データ 513が蓄積されているか否かを判断する (ステップ S 607) o具体的には、今回搭乗した搭乗者の顔画像や声紋の特徴点と、搭乗予定者 データベースに格納されて 、る顔画像データ 5丄 2や声紋データ 513の特徴点とを比 較して、類似度が所定値以上の顔画像データ 512や声紋データ 513が蓄積されて いる力否かを判断する。 Next, the navigation apparatus 300 extracts the face image and voiceprint of the passenger who has boarded this time from the image and sound captured and collected in step S604 (step S605). And Then, any one of the face images and voiceprints extracted in step S605 is collated with the planned passenger database 510 (step S606), and the face image data 512 and voiceprint data 513 similar to the face images and voiceprints are checked. (Step S607) o Specifically, the facial image and voiceprint feature points of the passenger who boarded this time, and the facial image data 5 stored in the planned passenger database 5特 徴 2 and the feature points of voiceprint data 513 are compared, and it is determined whether or not the face image data 512 or voiceprint data 513 having a similarity higher than a predetermined value is stored.
[0060] 類似した顔画像データ 512や声紋データ 513が蓄積されている場合は (ステップ S 607 : Yes)、その顔画像データ 512や声紋データ 513に対応する搭乗予定者のテキ スト情報 511を、テキスト情報 541として今回搭乗者データベース 530に格納するとと もに、その搭乗者の搭乗位置情報 542を今回搭乗者データベースに格納する (ステ ップ S608)。なお、搭乗予定者データベース 510に格納されているテキスト情報 511 も、そのまま保持される。  [0060] If similar face image data 512 and voiceprint data 513 are accumulated (step S607: Yes), the text information 511 of the planned passenger corresponding to the face image data 512 and voiceprint data 513 is stored. The text information 541 is stored in the current passenger database 530, and the boarding position information 542 of the passenger is stored in the current passenger database (step S608). The text information 511 stored in the planned passenger database 510 is also retained as it is.
[0061] 一方、類似した顔画像データ 512や声紋データ 513が蓄積されていない場合は( ステップ S607 :No)、その搭乗者に対して搭乗者予定者情報を入力させる (ステップ S609)。そして、入力された情報を搭乗予定者情報として搭乗予定者データベース 510に格納する(ステップ S610)。そして、ステップ S609で入力された搭乗者情報を 、テキスト情報 541として今回搭乗者データベース 530に格納するとともに、その搭 乗者の搭乗位置情報 542を今回搭乗者データベースに格納する (ステップ S611)。  On the other hand, if similar face image data 512 and voiceprint data 513 are not accumulated (step S607: No), the passenger information is input to the passenger (step S609). Then, the input information is stored in the boarding person database 510 as boarding person information (step S610). The passenger information input in step S609 is stored as text information 541 in the current passenger database 530, and the boarding position information 542 of the passenger is stored in the current passenger database (step S611).
[0062] すべての顔画像および声紋について照合をおこなうまでは(ステップ S612 : No)、 ステップ S606〖こ戻り、以降の処理を繰り返す。そして、すべての顔画像および声紋 につ ヽて照合をおこなうと(ステップ S612: Yes)、図 7のステップ S613に移行する。 ここまでの処理によって、搭乗予定者データベース 510および今回搭乗者データべ ース 530が作成される。  [0062] Until all face images and voiceprints are collated (step S612: No), the process returns to step S606 and the subsequent processing is repeated. When all face images and voiceprints are collated (step S612: Yes), the process proceeds to step S613 in FIG. Through the processing up to this point, the scheduled passenger database 510 and the current passenger database 530 are created.
[0063] なお、走行中に搭乗者の配置が変更する場合もあるため、走行中の所定のタイミン グ、たとえば、停車後にドアの開閉があった場合など、搭乗者の乗降を検知した際に 、今回搭乗者データベース 530を更新してもよい。この場合、ナビゲーシヨン装置 30 0は、ステップ S604以降の処理をおこなうことによって、今回搭乗者データベース 53 0を更新する。 [0063] Since the arrangement of the passenger may change during traveling, a predetermined timing during traveling, for example, when the door is opened or closed after the vehicle stops is detected The passenger database 530 may be updated this time. In this case, the navigation device 300 performs the process from step S604 onward, so that the current passenger database 53 Update 0.
[0064] 図 7の説明に移り、ナビゲーシヨン装置 300は、車両内の音声を監視し (ステップ S6 13)、搭乗者の発話に今回搭乗者データベース 530に格納されたテキスト情報 541 に含まれる語句(特定語句)が含まれて 、るか否かを判断する (ステップ S614)。音 声の監視は、一般に用いられている音声認識技術を用いておこなう。具体的には、 車載用マイク 312で収音された音声をテキスト変換し、テキスト情報 541に含まれる 語句が搭乗者の発話に含まれているか否かを判断する。  [0064] Turning to the description of FIG. 7, the navigation device 300 monitors the voice in the vehicle (step S613), and the words / phrases included in the text information 541 stored in the passenger database 530 this time in the passenger's utterance It is determined whether or not (specific phrase) is included (step S614). Voice monitoring is performed using commonly used voice recognition technology. Specifically, the voice picked up by the in-vehicle microphone 312 is converted into text, and it is determined whether or not the phrase included in the text information 541 is included in the utterance of the passenger.
[0065] テキスト情報 541に含まれる語句が発話された場合は (ステップ S614 : Yes)、今回 搭乗者データベース 530を照合し、発話された語句に対応する搭乗者の搭乗位置 情報 542を取得する(ステップ S615)。つづいて、ナビゲーシヨン装置 300は、優先 するカメラ動作があるか否かを判断する (ステップ S616)。優先するカメラ動作とは、 たとえば、ドライブレコーダ機能や搭乗者認識処理など、車載用カメラ 311を利用す る処理のうち、搭乗者撮影処理より優先しておこなうべき処理である。たとえば、車両 の走行速度が所定速度以上の場合や、車両の現在位置が交差点から所定距離以 内の場合は、事故が発生する可能性が通常より高いとして、ドライブレコーダ機能を 優先して動作させる。優先するカメラ動作の条件設定は、あらかじめ定められていて もよいし、ユーザによって設定できることとしてもよい。また、このときまで、車載用カメ ラ 311は、停止 (電源オフ)状態であってもよい。  [0065] When the phrase included in the text information 541 is uttered (step S614: Yes), the passenger database 530 is checked this time, and the boarding position information 542 of the passenger corresponding to the spoken phrase is acquired ( Step S615). Subsequently, the navigation device 300 determines whether or not there is a priority camera operation (step S616). The priority camera operation is a process that should be performed with priority over the passenger photographing process among processes using the in-vehicle camera 311 such as a drive recorder function and a passenger recognition process. For example, if the vehicle's running speed is greater than or equal to a predetermined speed, or if the current position of the vehicle is within a predetermined distance from the intersection, the drive recorder function is prioritized because the possibility of an accident is higher than normal. . The priority camera operation condition setting may be determined in advance or may be set by the user. Until this time, the in-vehicle camera 311 may be in a stopped (power off) state.
[0066] 優先するカメラ動作がある場合は (ステップ S616 : Yes)、カメラ動作が終了するま で待機する。優先するカメラ動作がない場合は (ステップ S616 :No)、 ROM402に 記録されて ヽる制御テーブル 560を参照して、発話された語句に対応する搭乗者の 搭乗位置に車載用カメラ 311の撮影方向を変更し (ステップ S617)、搭乗者を撮影 する (ステップ S618)。一方、テキスト情報 541に含まれる語句が発話されない場合 は (ステップ S614 : No)、ステップ S613に戻り、音声の監視を継続する。  [0066] If there is a priority camera operation (step S616: Yes), the process waits until the camera operation ends. If there is no priority camera operation (step S616: No), refer to the control table 560 recorded in ROM 402 and the shooting direction of the in-vehicle camera 311 at the boarding position of the passenger corresponding to the spoken phrase (Step S617) and photograph the passenger (step S618). On the other hand, when the phrase included in the text information 541 is not uttered (step S614: No), the process returns to step S613, and the voice monitoring is continued.
[0067] なお、テキスト情報 541に含まれる語句が発話された場合 (ステップ S614参照)に 必ず搭乗者を撮影するのではなぐテキスト情報 541に含まれる語句に加えて、特定 のキーワード (撮影指示キーワード)が発話された場合にのみ、撮影をおこなってもよ い。撮影指示キーワードとは、たとえば、「撮影」や「撮って」など、直接的に撮影を指 示する語句の他、「〇〇さんを見て」「〇〇さんに向けて」など、間接的に撮影を指示 する語句である。撮影指示キーワードは、あら力じめ決められていてもよいし、ユーザ によって登録できるようにしてもよ!、。 [0067] It should be noted that when a phrase included in the text information 541 is uttered (see step S614), a specific keyword (shooting instruction keyword) is added in addition to the phrase included in the text information 541 that does not necessarily photograph the passenger. You may shoot only when) is spoken. Shooting instruction keywords are, for example, “shooting” and “take a picture”. In addition to the words shown, it is a word that indirectly instructs the photographer, such as “Look at Mr. OO” and “Towards Mr. OO”. Shooting instruction keywords may be determined in advance or registered by the user!
[0068] また、たとえば、ステップ S617で搭乗者に撮影方向を向けた後、撮影指示キーヮ ードが含まれる発話がされた時点で、撮影を開始してもよい。この場合、被撮影者を 特定するキーワードが発せられた後、所定時間経過するまで撮影指示キーワードが 発話されない場合は、タイムアウトとして車載用カメラ 311の撮影方向を、初期状態に 戻すようにしてもよい。  [0068] Alternatively, for example, after the shooting direction is directed to the passenger in step S617, shooting may be started when an utterance including a shooting instruction key word is made. In this case, if a shooting instruction keyword is not spoken until a predetermined time elapses after a keyword for specifying a subject is issued, the shooting direction of the in-vehicle camera 311 may be returned to the initial state as a timeout. .
[0069] また、逆に、撮影指示キーワードが含まれる発話がされた時点で撮影を開始し、撮 影中にテキスト情報 541に含まれる語句が発話された場合は、発話された語句に対 応する搭乗者の方へ向けるようにしてもよい。さらに、撮影指示キーワードが含まれる 発話がされた時点では発言者に対して撮影方向を向けておき、その後、テキスト情 報 541に含まれる語句が発話された時点で、発話された語句に対応する搭乗者の方 へ撮影方向を向け、撮影を開始するようにしてもよい。  [0069] Conversely, shooting starts when an utterance that includes a shooting instruction keyword is made, and if a phrase included in text information 541 is uttered during shooting, the phrase that corresponds to the spoken phrase is handled. It may be directed toward the passenger. Furthermore, when the utterance containing the shooting instruction keyword is made, the shooting direction is directed toward the speaker, and then, when the phrase included in the text information 541 is spoken, the phrase corresponding to the spoken phrase is dealt with. The shooting direction may be directed toward the passenger to start shooting.
[0070] また、撮影対象となる搭乗者は 1人には限らず、たとえば、「全員」「男性」など、複 数の搭乗者に共通する属性が発話された場合には、該当する搭乗者を順次撮影す る。また、発話の中に、テキスト情報 541に含まれる語句が複数発話された場合も、そ れぞれの語句に対応する搭乗者を順次撮影する。また、各搭乗者の撮影優先順位 を決めておいてもよい。  [0070] In addition, the number of passengers to be photographed is not limited to one. For example, when an attribute common to a plurality of passengers such as "all" or "male" is spoken, the corresponding passenger Shoot sequentially. In addition, when a plurality of words included in the text information 541 are uttered in the utterance, the passengers corresponding to each word are photographed sequentially. In addition, the shooting priority order of each passenger may be determined.
[0071] ナビゲーシヨン装置 300は、撮影終了の指示があった場合や (ステップ S619 : Yes )、撮影を開始して力も所定時間が経過した場合は (ステップ S620 : Yes)、搭乗者の 撮影を終了する (ステップ S621)。ここで、撮影終了の指示とは、たとえば、「ストップ」 や「終了」など、直接的に撮影の終了を指示する語句の他、特定のキーワードの発話 やボタン操作などである。また、撮影終了の指示がなく(ステップ S619 :No)、撮影を 開始してから所定時間が経過していない間は (ステップ S620 :No)、ステップ S618 に戻り、搭乗者の撮影を継続する。  [0071] The navigation device 300, when instructed to end the shooting (step S619: Yes) or when the shooting has started and the force has elapsed for a predetermined time (step S620: Yes), End (step S621). Here, the shooting end instruction includes, for example, words such as “stop” and “end” that directly indicate the end of shooting, as well as utterances of specific keywords and button operations. If there is no instruction to end the shooting (Step S619: No) and the predetermined time has not elapsed since the start of shooting (Step S620: No), the process returns to Step S618 and the passenger's shooting is continued.
[0072] ステップ S621で撮影を終了した後は、車載用カメラ 311の撮影方向を元の位置に 戻す (ステップ S622)。車両の走行が終了するまでは(ステップ S623 : No)、ステツ プ S613に戻り、以降の処理を継続する。そして、車両の走行が終了すると (ステップ S623 :Yes)、今回搭乗者データベースを消去して(ステップ S624)、本フローチヤ ートの処理を終了する。 [0072] After shooting is completed in step S621, the shooting direction of the in-vehicle camera 311 is returned to the original position (step S622). Until the vehicle finishes running (Step S623: No), Return to step S613 to continue the following processing. When the vehicle finishes traveling (step S623: Yes), the current passenger database is deleted (step S624), and the process of this flow chart is terminated.
[0073] なお、上述した説明では、搭乗予定者情報データベース 510に搭乗予定者情報を 蓄積することとしたが、たとえば、搭乗予定者情報データベース 510を作成せず、毎 回の走行ごとに搭乗者情報を入力させて今回搭乗者データベース 530を生成する のみであってもよい。すなわち、毎回の走行ごとに図 8に示したような登録画面を表 示させ、そのとき搭乗している搭乗者全員に搭乗者情報を入力させ、搭乗位置情報 とともに今回搭乗者データベース 530に格納する。今回搭乗者データベース 530は 、走行が終了する度に消去されるため、搭乗者は搭乗するごとに情報の入力が必要 となるが、ナビゲーシヨン装置 300では、搭乗者情報を保持する必要がない。  [0073] In the above description, the planned passenger information is stored in the planned passenger information database 510. For example, the planned passenger information database 510 is not created, and the It is also possible to simply generate the passenger database 530 this time by inputting information. In other words, the registration screen as shown in FIG. 8 is displayed for each run, and all the passengers on board at that time enter the passenger information and store it in the current passenger database 530 together with the boarding position information. . Since the passenger database 530 is deleted every time traveling is completed, the passenger needs to input information every time he / she gets on the board, but the navigation device 300 does not need to hold the passenger information.
[0074] また、本実施例では、車両内の様子を撮影することとした力 たとえば、テレビ会議 システムなどでも同様に適用することができる。この場合、本実施例における搭乗予 定者は会議出席予定者であり、搭乗位置は会議室における着席位置となる。  [0074] Further, in the present embodiment, the power of photographing the state in the vehicle can be similarly applied to, for example, a video conference system. In this case, the boarding person in the present embodiment is a meeting attendee and the boarding position is the seating position in the meeting room.
[0075] 以上説明したように、ナビゲーシヨン装置 300によれば、車両内の音声から撮影対 象を特定し、撮影対象に向けて車載用カメラ 311の撮影方向を制御する。これにより 、搭乗者の発話の内容を考慮して、車載用カメラ 311の撮影方向を変更することがで きる。たとえば、後部座席に座っている子供に声をかけて、車載用カメラ 311の撮影 方向を子供に向け、その映像をディスプレイ 414に出力させることができる。これによ り、運転中においても、後方を振り返ることなぐ後部座席の子供の様子を確認するこ とがでさる。  As described above, according to the navigation device 300, the shooting target is specified from the sound in the vehicle, and the shooting direction of the in-vehicle camera 311 is controlled toward the shooting target. Accordingly, the shooting direction of the in-vehicle camera 311 can be changed in consideration of the content of the passenger's utterance. For example, a child sitting in the back seat can be called out, the shooting direction of the in-vehicle camera 311 can be directed to the child, and the video can be output to the display 414. This makes it possible to confirm the child in the rear seat without looking back while driving.
[0076] また、ナビゲーシヨン装置 300は、車載用カメラ 311によって撮影された映像や車 載用マイク 312から収音された音声から、搭乗者の位置を抽出する。これにより、搭 乗者に乗車位置の入力をおこなわせることなぐ搭乗者の乗車位置を特定することが できる。さらに、搭乗予定者データベースに搭乗予定者情報を蓄積しているため、そ の都度搭乗者に関する情報を入力させることなぐ搭乗者に煩雑な処理をおこなわ せることなく撮影をおこなうことができる。  In addition, the navigation device 300 extracts the position of the occupant from the video captured by the in-vehicle camera 311 and the sound collected from the in-vehicle microphone 312. As a result, the boarding position of the passenger can be specified without having the passenger input the boarding position. Furthermore, since the boarding person information is stored in the boarding person database, it is possible to take a picture without performing complicated processing for the passenger without having to input information about the boarder each time.
[0077] なお、本実施の形態で説明したカメラ制御方法は、あら力じめ用意されたプログラム をパーソナルコンピュータやワークステーションなどのコンピュータで実行することによ り実現することができる。このプログラムは、ハードディスク、フレキシブルディスク、 C D— ROM、 MO、 DVDなどのコンピュータで読み取り可能な記録媒体に記録され、 コンピュータによって記録媒体力も読み出されることによって実行される。またこのプ ログラムは、インターネットなどのネットワークを介して配布することが可能な伝送媒体 であってもよい。 It should be noted that the camera control method described in the present embodiment is a program prepared in advance. Can be realized by executing it on a computer such as a personal computer or a workstation. This program is recorded on a computer-readable recording medium such as a hard disk, a flexible disk, a CD-ROM, an MO, and a DVD, and is executed by reading the recording medium force by the computer. The program may be a transmission medium that can be distributed through a network such as the Internet.

Claims

請求の範囲 The scope of the claims
[1] カメラ周辺の音声を取得する取得手段と、  [1] Acquisition means for acquiring audio around the camera;
前記取得手段によって取得された音声から、前記カメラによる撮影対象を特定する 語句 (以下、「特定語句」 ヽぅ)を判別する判別手段と、  Discriminating means for discriminating a phrase (hereinafter referred to as “specific phrase” ヽ ぅ) for identifying a subject to be photographed by the camera from the voice acquired by the acquiring means;
前記判別手段によって判別された特定語句に基づいて、前記カメラの撮影方向を 制御する制御手段と、  Control means for controlling the shooting direction of the camera based on the specific phrase determined by the determination means;
を備えることを特徴とするカメラ制御装置。  A camera control device comprising:
[2] 前記撮影対象の候補に関する情報の入力を受け付ける入力手段と、  [2] An input means for receiving input of information on the candidate to be photographed;
前記撮影対象の候補の位置を検出する検出手段と、を備え、  Detecting means for detecting the position of the candidate for photographing,
前記判別手段は、  The discrimination means includes
前記入力手段に入力された前記撮影対象の候補に関する情報と略一致する語句 を前記特定語句として判別し、  A phrase that substantially matches the information about the candidate to be photographed input to the input means is determined as the specific phrase,
前記制御手段は、  The control means includes
前記判別手段によって前記撮影対象の候補に関する情報と略一致する語句が判 別された場合、前記検出手段によって検出された前記撮影対象の候補の位置に前 記カメラの撮影方向を向けることを特徴とする請求項 1に記載のカメラ制御装置。  In the case where a word or phrase that substantially matches the information related to the candidate for the photographing target is determined by the determining means, the photographing direction of the camera is directed to the position of the candidate for the photographing target detected by the detecting means. The camera control device according to claim 1.
[3] 前記入力手段は、 [3] The input means includes
前記撮影対象の候補に関する情報として、前記撮影対象の候補の名称情報また は前記撮影対象の候補の属性情報のうち、少なくともいずれかが入力されることを特 徴とする請求項 2に記載のカメラ制御装置。  3. The camera according to claim 2, wherein at least one of name information of the candidate for photographing target or attribute information of the candidate for photographing target is input as the information regarding the candidate for photographing target. Control device.
[4] 前記入力手段は、 [4] The input means includes
前記撮影対象の候補の画像が入力され、  A candidate image of the shooting target is input,
前記検出手段は、  The detection means includes
前記画像と前記カメラによって撮影された映像とを照合して前記撮影対象の候補の 位置を検出することを特徴とする請求項 2に記載のカメラ制御装置。  3. The camera control device according to claim 2, wherein the position of the candidate to be photographed is detected by comparing the image and a video photographed by the camera.
[5] 前記入力手段は、 [5] The input means includes
前記撮影対象の候補の音声が入力され、  The voice of the candidate for shooting is input,
前記検出手段は、 前記音声と前記取得手段によって取得された音声とを照合して前記撮影対象の候 補の位置を検出することを特徴とする請求項 2に記載のカメラ制御装置。 The detection means includes The camera control apparatus according to claim 2, wherein the position of the candidate to be imaged is detected by comparing the sound and the sound acquired by the acquisition unit.
[6] 前記カメラは、 [6]
車両内部に設置されており、  It is installed inside the vehicle,
前記取得手段は、  The acquisition means includes
前記車両の搭乗者の発話を取得し、  Obtaining the utterance of the passenger of the vehicle,
前記判別手段は、  The discrimination means includes
前記撮影対象となる前記搭乗者に関する情報を特定語句として判別し、 前記制御手段は、  Discriminating information about the passenger to be photographed as a specific phrase, the control means,
前記特定語句によって特定される前記搭乗者の乗車位置に前記カメラの撮影方向 を向けることを特徴とする請求項 1〜5のいずれか一つに記載のカメラ制御装置。  6. The camera control device according to claim 1, wherein a shooting direction of the camera is directed to a boarding position of the passenger specified by the specific phrase.
[7] カメラ周辺の音声を取得する取得工程と、 [7] An acquisition process for acquiring audio around the camera;
前記取得工程によって取得された音声から、前記カメラによる撮影対象を特定する 語句 (以下、特定語句と!ゝぅ)を判別する判別工程と、  A discrimination step of discriminating a phrase (hereinafter referred to as a specific phrase and! ゝ ぅ) that identifies a subject to be photographed by the camera from the voice acquired in the acquisition step;
前記判別工程によって判別された特定語句に基づいて、前記カメラの撮影方向を 制御する制御工程と、  A control step of controlling the shooting direction of the camera based on the specific phrase determined by the determination step;
を含んだことを特徴とするカメラ制御方法。  A camera control method comprising:
[8] 請求項 7に記載のカメラ制御方法をコンピュータに実行させることを特徴とするカメ ラ制御プログラム。 [8] A camera control program that causes a computer to execute the camera control method according to claim 7.
[9] 請求項 8に記載のカメラ制御プログラムを記録したコンピュータに読み取り可能な記 録媒体。  [9] A computer-readable recording medium in which the camera control program according to claim 8 is recorded.
PCT/JP2007/062151 2006-06-16 2007-06-15 Camera control apparatus, camera control method, camera control program, and recording medium WO2007145331A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2008521277A JP4660592B2 (en) 2006-06-16 2007-06-15 Camera control apparatus, camera control method, camera control program, and recording medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2006167403 2006-06-16
JP2006-167403 2006-06-16

Publications (1)

Publication Number Publication Date
WO2007145331A1 true WO2007145331A1 (en) 2007-12-21

Family

ID=38831844

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2007/062151 WO2007145331A1 (en) 2006-06-16 2007-06-15 Camera control apparatus, camera control method, camera control program, and recording medium

Country Status (2)

Country Link
JP (1) JP4660592B2 (en)
WO (1) WO2007145331A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011013605A1 (en) * 2009-07-27 2011-02-03 三洋電機株式会社 Presentation system
JP2013131919A (en) * 2011-12-21 2013-07-04 Canon Inc Imaging apparatus, control method thereof, and program
JP2017188715A (en) * 2016-04-01 2017-10-12 富士通フロンテック株式会社 Video display system and video display method
WO2019073558A1 (en) * 2017-10-11 2019-04-18 三菱電機株式会社 Air conditioning controller
KR20200120977A (en) * 2019-04-05 2020-10-23 현대자동차주식회사 Vehicle status management apparatus and mehtod
WO2023120244A1 (en) * 2021-12-24 2023-06-29 ソニーグループ株式会社 Transmission device, transmission method, and program

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0983856A (en) * 1995-09-07 1997-03-28 Nippon Telegr & Teleph Corp <Ntt> Intelligent camera equipment
JP2004023676A (en) * 2002-06-20 2004-01-22 Hitachi Kokusai Electric Inc Cctv set and state output method for cctv set
JP2004297271A (en) * 2003-03-26 2004-10-21 Kenwood Corp In-vehicle imaging system
JP2005184386A (en) * 2003-12-18 2005-07-07 Sony Corp Sound collecting/video recording device

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3733258B2 (en) * 1999-02-23 2006-01-11 キヤノン株式会社 Image photographing device, image photographing method, storage medium, video conference system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0983856A (en) * 1995-09-07 1997-03-28 Nippon Telegr & Teleph Corp <Ntt> Intelligent camera equipment
JP2004023676A (en) * 2002-06-20 2004-01-22 Hitachi Kokusai Electric Inc Cctv set and state output method for cctv set
JP2004297271A (en) * 2003-03-26 2004-10-21 Kenwood Corp In-vehicle imaging system
JP2005184386A (en) * 2003-12-18 2005-07-07 Sony Corp Sound collecting/video recording device

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011013605A1 (en) * 2009-07-27 2011-02-03 三洋電機株式会社 Presentation system
JP2013131919A (en) * 2011-12-21 2013-07-04 Canon Inc Imaging apparatus, control method thereof, and program
JP2017188715A (en) * 2016-04-01 2017-10-12 富士通フロンテック株式会社 Video display system and video display method
WO2019073558A1 (en) * 2017-10-11 2019-04-18 三菱電機株式会社 Air conditioning controller
JPWO2019073558A1 (en) * 2017-10-11 2020-01-23 三菱電機株式会社 Air conditioning controller
CN111183322A (en) * 2017-10-11 2020-05-19 三菱电机株式会社 Controller for air conditioning
KR20200120977A (en) * 2019-04-05 2020-10-23 현대자동차주식회사 Vehicle status management apparatus and mehtod
KR102645054B1 (en) * 2019-04-05 2024-03-08 현대자동차주식회사 Vehicle status management apparatus and mehtod
WO2023120244A1 (en) * 2021-12-24 2023-06-29 ソニーグループ株式会社 Transmission device, transmission method, and program

Also Published As

Publication number Publication date
JPWO2007145331A1 (en) 2009-11-12
JP4660592B2 (en) 2011-03-30

Similar Documents

Publication Publication Date Title
US10853629B2 (en) Method for identifying a user entering an autonomous vehicle
JP6604151B2 (en) Speech recognition control system
JP4660592B2 (en) Camera control apparatus, camera control method, camera control program, and recording medium
JP6466385B2 (en) Service providing apparatus, service providing method, and service providing program
US11450316B2 (en) Agent device, agent presenting method, and storage medium
US20200152203A1 (en) Agent device, agent presentation method, and storage medium
JP4556586B2 (en) Driving assistance device
KR20130046759A (en) Apparatus and method for recogniting driver command in a vehicle
JP2004354930A (en) Speech recognition system
JP6785889B2 (en) Service provider
JP2009113725A (en) Device, method and program for controlling instrument, and recording medium
US11841980B2 (en) Information output device and method of controlling information output based on vehicle occupancy
JP2009098217A (en) Speech recognition device, navigation device with speech recognition device, speech recognition method, speech recognition program and recording medium
US20220020374A1 (en) Method, device, and program for customizing and activating a personal virtual assistant system for motor vehicles
JP2005010035A (en) Navigation apparatus for vehicle
US20210061102A1 (en) Operation restriction control device and operation restriction control method
JP2018180424A (en) Speech recognition apparatus and speech recognition method
WO2022172393A1 (en) Voice recognition device and voice recognition method
JP7386076B2 (en) On-vehicle device and response output control method
JP2009086132A (en) Speech recognition device, navigation device provided with speech recognition device, electronic equipment provided with speech recognition device, speech recognition method, speech recognition program and recording medium
US20210302179A1 (en) Server device, control device, program, vehicle, and operation method of information processing system
WO2024029192A1 (en) Speech command reception device and speech command reception method
WO2023047654A1 (en) Vehicular recording control device and recording method
JP2023095046A (en) Information processing system
JP2006030908A (en) Voice recognition device for vehicle and moving body

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 07745405

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2008521277

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 07745405

Country of ref document: EP

Kind code of ref document: A1