WO2010001778A1 - 撮像装置、画像表示装置および電子カメラ - Google Patents
撮像装置、画像表示装置および電子カメラ Download PDFInfo
- Publication number
- WO2010001778A1 WO2010001778A1 PCT/JP2009/061490 JP2009061490W WO2010001778A1 WO 2010001778 A1 WO2010001778 A1 WO 2010001778A1 JP 2009061490 W JP2009061490 W JP 2009061490W WO 2010001778 A1 WO2010001778 A1 WO 2010001778A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- accuracy
- positioning
- image file
- unit
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 61
- 238000005259 measurement Methods 0.000 claims abstract description 5
- 238000001514 detection method Methods 0.000 claims description 11
- 230000008859 change Effects 0.000 claims description 8
- 238000000034 method Methods 0.000 description 92
- 230000008569 process Effects 0.000 description 68
- 238000012545 processing Methods 0.000 description 54
- 238000006243 chemical reaction Methods 0.000 description 24
- 238000010586 diagram Methods 0.000 description 18
- 230000004048 modification Effects 0.000 description 11
- 238000012986 modification Methods 0.000 description 11
- 230000000694 effects Effects 0.000 description 10
- 230000006870 function Effects 0.000 description 10
- 230000015556 catabolic process Effects 0.000 description 5
- 238000006731 degradation reaction Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000005286 illumination Methods 0.000 description 4
- 230000015654 memory Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000011514 reflex Effects 0.000 description 3
- 238000013500 data storage Methods 0.000 description 2
- 238000010790 dilution Methods 0.000 description 2
- 239000012895 dilution Substances 0.000 description 2
- 101100456831 Caenorhabditis elegans sams-5 gene Proteins 0.000 description 1
- 125000002066 L-histidyl group Chemical group [H]N1C([H])=NC(C([H])([H])[C@](C(=O)[*])([H])N([H])[H])=C1[H] 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000003936 working memory Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/64—Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
- G06T1/20—Processor architectures; Processor configuration, e.g. pipelining
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
- G06T1/60—Memory management
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/162—Segmentation; Edge detection involving graph-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/77—Determining position or orientation of objects or cameras using statistical methods
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/633—Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
- H04N5/772—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera the recording apparatus and the television camera being placed in the same enclosure
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/907—Television signal recording using static stores, e.g. storage tubes or semiconductor memories
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/7921—Processing of colour television signals in connection with recording for more than one processing mode
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/8042—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/28—Indexing scheme for image data processing or generation, in general involving image processing hardware
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20072—Graph-based image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/61—Scene description
Definitions
- the present invention relates to an imaging device, an image display device, and an electronic camera including the image display device.
- the mobile terminal sends accuracy limit data indicating a desired accuracy limit for the self-location data to the position processing system, the position processing system providing the position data of the mobile terminal limited to the desired accuracy.
- a data providing method is known (for example, see Patent Document 1).
- Patent Document 2 a camera that records positioning data and information on the accuracy of the positioning data together with the photographed image on a photographic film is known as a prior art (see, for example, Patent Document 2).
- the camera described in Patent Document 2 can detect a location where the user has taken an image or the like later using the positioning data, but the location where the positioning data is low can be detected. May be mistakenly recognized.
- the imaging device records the data of the photographing position together with the photographed image, the photographing unit that photographs the image, the positioning unit that measures the photographing position when the image is photographed by the photographing unit.
- a control unit that determines whether or not, and a recording unit that records only the photographed image or the photographed image and the photographing position data according to the determination by the control unit.
- the control unit determines whether to record the data of the photographing position together with the photographed image based on the photographing position.
- the accuracy of setting whether or not to record the shooting position data with respect to a specific place having a predetermined area and the recording accuracy of the shooting position data when recording is possible.
- a storage unit is provided, and the control unit collates the shooting position with a specific location in the accuracy storage unit, determines whether or not to record the shooting position data together with the shot image, and determines the recording accuracy of the shooting position data when recording. preferable.
- a person storage unit storing features for a specific person, and a specific person is recognized from a photographed image with reference to the person storage unit
- the control unit converts the recording accuracy of the shooting position data to a low accuracy when a specific person is recognized in the captured image by the person recognition unit.
- the positioning unit outputs the positioning accuracy of the shooting position data, and the control unit is a recording accuracy whose positioning accuracy is determined by the control unit. If it is lower, it is preferable to repeat the positioning by the positioning unit until the positioning accuracy satisfies the recording accuracy.
- the control unit determines whether or not to record the shooting position data together with the shot image based on the shot image.
- a person storage unit storing features for a specific person and a person recognition for recognizing a specific person from a photographed image with reference to the person storage unit
- the control unit determines the recording accuracy depending on whether or not a specific person is recognized in the captured image by the person recognition unit.
- the imaging device includes an imaging unit that captures an image, a positioning unit that measures the imaging position when the imaging unit captures an image, and data of an imaging position measured by the positioning unit. And a recording unit for recording the captured image, the shooting position data, and the recording accuracy determined by the control unit.
- the control unit determines the recording accuracy of the data at the photographing position based on the photographing position.
- the control unit determines the recording accuracy of the data of the photographing position based on the photographed image.
- the display mode of the captured image recorded in the recording unit is changed according to the recording accuracy recorded in the recording unit. It is preferable to further include a display control unit.
- an image display device searches for an image file having a predetermined positioning accuracy or more from a plurality of image files having positioning data and positioning accuracy information in the positioning data.
- a search unit; and an image file display control unit that displays an image of the image file searched by the image file search unit on a display.
- the image file display control unit arranges and searches in order from an image file having a high positioning accuracy to an image file having a low positioning accuracy. It is preferable to display an image of the processed image file.
- the image file display control unit makes the image of the image file with high positioning accuracy larger than the image of the image file with low positioning accuracy. Display is preferred.
- the image display device includes a current position detection unit that detects a current position of the image display device, and the image file search unit includes a plurality of image file search units. Search the image file for positioning data that coincides with the current position detected by the current position detection unit, or having positioning data within a predetermined distance from the current position, and select a predetermined one from the searched image files. It is preferable to search for an image file having a positioning accuracy higher than that.
- the image display device includes a current position detection unit that detects a current position of the image display device, and the image file display control unit determines the position of positioning data from the current position. It is preferable to display the images of the image file side by side according to the value obtained by multiplying the distance up to the value representing the positioning accuracy.
- the image file search unit includes positioning data that matches the current position detected by the current position detection device from the plurality of image files.
- the image file display control unit preferably displays the image of the image file together with a map around the position of the positioning data in the image file.
- the image display device includes a map scale changing unit that changes a map scale, and the image file search unit has the map scale changed by the map scale changing unit.
- the search is performed by changing the predetermined positioning accuracy based on the changed map scale, and the image file display control unit reduces the image of the image file searched by the image file search unit by the map scale change unit. Is preferably displayed together with the changed map.
- the image file display control unit preferably displays the image of the image file together with the radar chart.
- an image input unit for inputting an image selected from images displayed by the image file display control unit, and an input by the image input unit It is preferable that a map display control unit that displays a map around the position of the positioning data in the image file of the obtained image is provided.
- the map display control unit determines the scale of the map to be displayed based on the positioning accuracy in the image file of the image input by the image input unit. It is preferable to do this.
- the image display device includes a positioning accuracy input unit that inputs positioning accuracy, and the image file search unit is input by the measurement accuracy input unit. It is preferable to search using the determined positioning accuracy as a predetermined measurement accuracy.
- an electronic camera includes the image display device according to any one of the twelfth to twenty-third aspects.
- the present invention since whether or not to record the shooting position of the image and the recording accuracy of the shooting position are determined, it is possible to prevent a situation in which a shooting location that is not to be disclosed accidentally is disclosed together with the image. be able to. Also, only image files with a predetermined positioning accuracy or higher can be displayed.
- photography position data of 1st Embodiment The flowchart which shows the required precision determination processing of the position data of 1st Embodiment It is a figure which shows person database A5, B5, C5. It is a figure which shows precision conversion database A8, B8, C3. It is a figure which shows request
- FIG. It is a figure which shows precision conversion database A8, B8, C3 used in 2nd Embodiment. It is an external view of the electronic camera in the 3rd Embodiment of this invention.
- FIG. 1 is a diagram showing the configuration of an embodiment of the invention.
- the camera A and the camera B are in a “linked relationship” with each other, for example, with the same owner.
- the cameras accuracy conversion database, required accuracy table, person database, etc.
- Some photographers have a single-lens reflex camera and a compact camera, and use them depending on the subject and conditions. In such a case, for example, a single-lens reflex camera is set as camera A and a compact camera is set as camera B, and data regarding the same accuracy conversion is shared.
- the center C includes a database that can be used by the authentication of the same ID by the camera A and the camera B having a “linked relationship” with each other.
- the camera A and the camera B do not have to provide a database regarding the position data conversion.
- the camera A and the camera B may be the same camera, or may be different types of cameras, such as a single-lens reflex camera and the other a compact camera.
- the photographing processing units A1 and B1 include a photographing lens, an image sensor, an image processing device, and the like (not shown), photograph a subject image, and perform various processes.
- the recording processing units A2 and B2 record the subject image captured by the imaging processing units A1 and B1 on a recording device such as a memory card. Although details will be described later, the position data of the shooting location is recorded together with the shot image.
- the positioning processing units A3 and B3 measure the shooting location in synchronization with the shooting by the shooting processing units A1 and B1, and detect the latitude X and longitude Y of the shooting location.
- the positioning method includes a well-known GPS positioning method, WiFi positioning method, base station positioning method using a mobile phone, and the like, and the shooting position is detected by at least one of these positioning methods.
- the face detection (person recognition) processing units A4 and B4 determine whether or not a person registered in advance is included in the images photographed by the photographing processing units A1 and B1.
- FIG. 4 shows feature quantity data relating to a specific person stored in the person databases A5, B5, and C5.
- the owners of camera A and camera B use the face image of the person they want to recognize as a template image, store the feature data in the person databases A5, B5, and C5 in association with the person names, and register them in advance.
- the face detection (person recognition) processing units A4 and B4 collate the feature amount data of the person registered in the person databases A5, B5, and C5 with the images photographed by the photographing processing units A1 and B1, It is determined whether or not a person registered in advance is included in the photographed image by the face recognition method.
- the user authentication information sections A6 and B6 store ID information for authenticating whether or not the user is a contract user whose data has been registered in the database in advance when using various data stored in the center C. .
- the user authentication processing unit C1 of the center C collates the ID information of the registered user registered in advance with the ID information of the user sent from the camera A or the camera B, and is stored in advance when the two match.
- Various data of contracted users is provided to camera A or camera B.
- the transmission / reception processing units A7, B7, and C2 perform transmission and reception of various data with the camera A, the camera B, and the center C.
- the display processing units A10 and B10 read the image recorded on the recording device such as a memory card by the recording processing units A2 and B2, and perform display processing on the rear monitor (not shown) of the camera.
- FIG. 5 is a diagram showing the accuracy conversion databases A8, B8, and C3.
- the accuracy conversion databases A8, B8, and C3 are for converting the position data of the shooting location to the accuracy corresponding to the registered region when the shooting location detected by the positioning processing units A3 and B3 is in the pre-registered region. Database.
- the area is an area designated in advance by a user with latitude X and longitude Y and radius R centered on the latitude and longitude.
- the recording accuracy of the position information at the time of non-recognition of the person is the required accuracy of the position data recorded together with the captured image when a pre-registered person is not captured in the captured image.
- the accuracy Low or the position data recording prohibition Off is set.
- the recording accuracy of the position data at the time of person recognition is the case where the face detection (person recognition) processing units A4 and B4 recognize persons registered in the person databases A5, B5, and C5 (see FIG. 4) in the photographed images.
- the required accuracy of the position data recorded together with the photographed image is set by the user as high accuracy high or low accuracy low, or position data recording prohibition Off.
- the recording frequency of the position history is the recording frequency of the position data when the movement trajectory of the user holding the cameras A and B is detected and recorded in the log file. Depending on the user, the high frequency High or low frequency Low, or Off that is not recorded is set.
- the user can use the map interface and other areas around his / her home, the company's area around the company, and the travel destination area.
- the recording accuracy and the recording frequency of the position history can be registered in the accuracy conversion databases A8, B8, and C3.
- the transmission / reception processing unit A7 is performed by registering or updating data regarding the area and accuracy in the accuracy conversion database A8 of one camera A.
- the same data is automatically registered or updated in the accuracy conversion database B8 of the other camera B via B7.
- the shooting location is not in the registration area of the accuracy conversion databases A8, B8, and C3, the recording accuracy of the position data when the person is not recognized, the recording accuracy of the position data when the person is recognized, and the recording frequency of the position history
- the default values stored in advance in the cameras A and B are set.
- FIG. 6 is a diagram showing the required accuracy tables A9, B9, and C4.
- the required accuracy tables A9, B9, and C4 are tables in which position data recording accuracy High, Low, Off and position history recording frequencies High, Low, Off are set, and can be arbitrarily set by the user.
- the recording accuracy of the position data for example, as shown in FIG. 6, in accordance with the accuracy degradation index DOP (Dilution ⁇ of Precision) provided together with the position data in GPS positioning, the position data having a DOP of 6 or less is set to high accuracy High.
- Position data greater than 6 is defined as low accuracy Low.
- the recording frequency of the position history for example, as shown in FIG. 6, a case where recording is performed every 10 seconds is set to high accuracy High, and a case where recording is performed every 60 seconds is set to low accuracy Low. If the position data and position history are not recorded, it is set to Off.
- FIG. 2 and FIG. 3 are flowcharts showing a conversion process of position data accuracy according to an embodiment.
- Camera A and camera B execute this processing every predetermined time (for example, 10 seconds) while the power is on.
- the accuracy conversion process of the camera A will be described here, the same applies to the camera B.
- step 1 of FIG. 2 the current position is measured by the positioning processing unit A3, and latitude, longitude, and DOP are detected as position data.
- step 2 it is determined whether or not photographing by the photographing processing unit A1 has been performed. If photographing has not been performed, the processing ends. Note that shooting includes still image shooting, moving image shooting, and image shooting with sound.
- step 3 determines the required accuracy of the position data recorded together with the photographed image with reference to the accuracy conversion database A8.
- This required accuracy determination processing is performed by executing the subroutine of FIG.
- a human face is detected from the captured image.
- a known method may be used, but here it is sufficient that a human face can be recognized, and it is not necessary to determine which person the recognized face is.
- step 12 it is determined whether or not one or more human faces have been detected. If no human face is shown in the photographed image, the process proceeds to step 13 to request recording accuracy of position data when the person is not recognized. To do. That is, referring to the accuracy conversion database A8, it is determined whether or not the positioning position (latitude, longitude) of the shooting location is in a pre-registered area, and the position at the time of non-recognition of the person according to the shooting position area The data recording accuracy is the required accuracy. If the shooting position is not within the registration area, a default value stored in advance in the camera A is set as the recording accuracy of the position data when the person is not recognized.
- step 14 to perform person recognition processing. That is, referring to the person database A5, it is determined whether or not the face in the photographed image matches the template image of the person's face registered in advance. If the face in the photographed image does not match the face of the registered person, that is, if no registered person is recognized in the photographed image, the process proceeds to step 13 and the person is not recognized according to the area of the photographing position as described above. Require time recording accuracy of position data.
- the process proceeds to step 16 and position data at the time of person recognition is detected.
- the recording accuracy is required. That is, with reference to the accuracy conversion database A8, it is determined whether or not the positioning position (latitude, longitude) of the shooting location is within a pre-registered area, and the position data at the time of person recognition according to the shooting position area Is the required accuracy. If the shooting position is not within the registration area, a default value stored in advance in the camera is set as the recording accuracy of the position data at the time of person recognition.
- step 4 it is determined whether or not the required accuracy is Off which does not record the position data. If the required accuracy is Off, the process ends. If the required accuracy is not Off, the process proceeds to Step 5. In step 5, it is determined whether or not the required recording accuracy of the position data in accordance with the area determination of the shooting position and the presence / absence of person recognition is high accuracy high. If high accuracy high is required, the process proceeds to step 6. .
- step 6 it is determined whether or not the positioning accuracy of the current position (positioning accuracy in step 1) is lower than the required accuracy. If the positioning accuracy is lower than the required accuracy, the process proceeds to step 8, and the current position is repositioned by the positioning processing unit A3. For example, when the degradation accuracy DOP of the positioning accuracy is a low accuracy Low greater than 6, although the required accuracy according to the area determination of the shooting position and the presence / absence of person recognition is a high accuracy high, the current location Perform re-positioning. On the other hand, if the positioning accuracy of the current position is greater than the required accuracy, the process proceeds to step 7. For example, when the required accuracy is high accuracy high and the positioning accuracy degradation index DOP is 6 or less and high accuracy high, the process proceeds to step 7 and the positioning result position data is recorded as it is together with the captured image.
- step 9 converts the position data of the positioning result to low-accuracy.
- the latitude and longitude of the positioning result is “ddmm.mmmm, N / S (north latitude / south latitude), dddmm.mmmm, E / W (east longitude / west longitude)”
- the decimal part is forced to 0.
- the address is deleted from the section of “1-6-3 Nishioi, Shinagawa-ku, Tokyo” and is renamed “Shinagawa-ku, Tokyo”.
- the station number and the telephone number are deleted from “+ 81-3-3773-1111” to be “+ 81-3”.
- “140-8601” is set to “140”.
- step 9 the conversion of the position data to low precision in step 9 is irrelevant whether the position data of the positioning result is high precision (DOP ⁇ 6) or low precision (DOP> 6).
- DOP ⁇ 6 high precision
- DOP> 6 low precision
- the latitude and longitude include numerical values after the decimal point. In this case, the low-precision latitude and longitude are forced to zero after the decimal point. Convert to precision data.
- the converted position data is recorded together with the photographed image in step 10, and the process is terminated.
- information indicating recording accuracy is recorded together with the captured image and position data as one image file.
- the recording accuracy for example, information obtained by encoding the aforementioned High, Low, and Off is recorded. In the case of High and Low, the DOP value may be recorded.
- FIG. 7 is a diagram illustrating a flowchart of display processing executed by the display processing unit A10 in the camera A.
- the CPU in camera A performs the processing of FIG. 7 by executing a predetermined program.
- a predetermined program In the present embodiment, in the normal display mode, only images with high recording accuracy are selected from image files recorded on a recording device such as a memory card, and thumbnails are displayed on the rear monitor of camera A.
- a code indicating one of recording accuracy High, Low, and Off is recorded in a predetermined area.
- step S21 the designated image file is read from the memory card.
- the latest image file is read.
- step S22 it is determined whether the recording accuracy is high. If it is High, the process proceeds to Step S23. If it is not High (Low or Off), the process proceeds to Step S25.
- step S23 thumbnail image data is read from the read image file and displayed on the rear monitor (not shown) of camera A.
- step S24 it is determined whether or not a predetermined number of thumbnail images are displayed on the rear monitor. If the predetermined number has not been displayed yet, the process proceeds to step S25. If the predetermined number is displayed, the process is terminated. In step S25, the next image file is set, and the process returns to step S21 to repeat the process.
- processing in FIG. 7 has been described as processing by the camera A, the processing is also performed by the camera B in the same manner. Further, the same processing may be performed by the display processing unit C6 of the center C. Furthermore, the processing of FIG. 7 may be executed on a personal computer or another image display device. When processing is performed by the center C or personal computer, the image file captured and recorded by the camera may be stored in a storage device (database) of the center C or personal computer via a memory card or communication.
- database storage device
- the display mode of the image can be changed according to the recording accuracy by displaying or not displaying the image or changing the size of the image.
- the required recording accuracy of the position data may be determined according to the image shooting conditions. For example, when the shooting angle of view is wide, a wide range is shot, so the position data of the shooting location does not need to be highly accurate, and in that case, low accuracy is required. On the other hand, when shooting with telephoto, the shooting target is narrowed down, and the position data of the shooting location needs to be highly accurate. In that case, high accuracy is required.
- the recording resolution of the photographed image may be changed according to the recording accuracy of the photographing position data. For example, when the recording accuracy of the shooting position data is high, the recording resolution of the shot image is increased.
- the shooting position when the image is shot is measured, and it is determined whether to record the shooting position data together with the shot image based on the shooting position.
- it includes a database that sets whether or not recording of shooting position data for a specific area of a predetermined area and the recording accuracy of shooting position data when recording is possible, and compares the shooting position with a specific area of the database, Whether to record the shooting position data together with the shot image and the recording accuracy of the shooting position data when recording are determined. Whether to record the shooting position of the image and whether to record the recording accuracy. Is automatically determined, and the operability is improved, and it is possible to prevent a situation where a shooting location that is not to be disclosed accidentally is disclosed together with the image.
- a database storing features for a specific person is provided, and the specific person is recognized from the captured image with reference to the database, and the specific position is recognized in the captured image. Since the data recording accuracy is converted to a low accuracy, it is possible to prevent a situation in which an accurate address and work place of a person preset by the photographer are disclosed together with the photographed image.
- the positioning accuracy is lower than the required accuracy, since the positioning means repeats the positioning until the positioning accuracy satisfies the required accuracy, it is possible to reliably record the shooting position data with the required accuracy.
- FIG. 8 is a diagram showing the accuracy conversion databases A8, B8, and C3 used in the second embodiment.
- the required recording accuracy is set to High when the person is not recognized, the required recording accuracy is Off (not recorded) for the person A and the required recording accuracy is Low for the people B and C at the time of person recognition.
- an example of selecting High is shown.
- the person A is the owner of the camera, and the persons B and C are family members and close friends. That is, it is set so that the recording of the position data itself is turned off when the camera owner is photographed, and the recording accuracy is converted to Low in an image where a family member or close friend is photographed.
- the “other” person is a case where a person is shown in the image but a person who is not a related person is shown.
- the recording accuracy is not changed by a combination of the recognition of the person and the place where the image was taken, but it is determined whether or not the position data is recorded only by the recognized person.
- the accuracy can be changed. Thereby, when a specific person is photographed, the recording of the position data can be turned off or the recording accuracy can be lowered so that the location where the photograph was taken is not identified or is difficult to identify.
- the recognition target may be another subject instead of a person.
- the recognition target may be another subject instead of a person.
- it may be a home building or a car owned by the user. That is, instead of the person databases A5, B5, and C5 in which feature amount data relating to a specific person is stored, a database in which feature amount data relating to other recognition targets is stored may be provided.
- FIG. 9 is a diagram showing the electronic camera 1.
- FIG. 9A is a diagram of the electronic camera 1 viewed from an oblique front
- FIG. 9B is a diagram of the electronic camera 1 viewed from an oblique rear.
- a lens 121a of a photographing optical system see FIG. 10, reference numeral 121
- an illumination device 110 that illuminates the subject are provided on the front surface of the electronic camera 1.
- the electronic camera 1 can be connected to a GPS (Global Positioning System) device 2 and can acquire positioning data and information related to the accuracy of the positioning data from the GPS device 2.
- a release switch 103 a is provided on the upper surface of the electronic camera 1.
- a liquid crystal monitor 104 and operation buttons 103b to 103g are provided on the back of the electronic camera 1.
- FIG. 10 is a block diagram illustrating the configuration of the electronic camera 1.
- an electronic camera 1 includes a control circuit 101, a memory 102, an operation unit 103, a display monitor 104, a speaker 105, an external interface (I / F) circuit 106, and a memory card interface (I / F). 107, a power source 108, a photometric device 109, an illumination device 110, a map data storage device 111, and a GPS interface (I / F) circuit 112.
- the electronic camera 1 is connected to the GPS device 2 via the GPS interface circuit 112.
- a memory card 150 is mounted on the memory card interface 107.
- the control circuit 101 performs a predetermined calculation using signals output from each part in the electronic camera based on the control program.
- the control circuit 101 controls the photographing operation of the electronic camera 1 by sending a control signal to each part in the electronic camera.
- the control program is stored in a ROM (ROMRead Only Memory) (not shown) in the control circuit 101.
- the control circuit 101 creates an image file from information on imaging conditions acquired from each part in the electronic camera, positioning data acquired from a GPS device and information on the accuracy of the positioning data, image data of a captured image, etc., and the memory 150 Remember me.
- This image file is created in accordance with the Exif (Exchangeable image file format) image file specification.
- the image file 3 is composed of main image data and a plurality of tags in which auxiliary information of the main image data is described.
- a tag 31 that describes whether the positioning position is north latitude or south latitude
- a tag 32 that describes the latitude of the positioning position, and whether the positioning position is east longitude or west longitude
- a tag 34 that describes the longitude of the measured position
- a tag 35 that describes the reliability of positioning, that is, the accuracy of positioning.
- the accuracy of positioning is represented by a DOP (Dilution of Precision) value.
- DOP Deution of Precision
- the positioning data 31 to 34 are data on the current position of the user and the current position of the device used for shooting, usually when the user shots the image.
- the DOP value is called an accuracy degradation coefficient and is an index indicating the degree of accuracy degradation due to the positioning satellite positioning.
- An ideal satellite arrangement when performing positioning by positioning satellites is an arrangement in which the zenith satellite and three neighboring satellites dispersed by 120 ° form a regular triangular pyramid shape.
- the DOP value of this arrangement is 1. Depending on how many times the precision is inferior to that, it is expressed as an index such as 2, 3, 4,. That is, the positioning accuracy decreases as the DOP value increases.
- the DOP value is calculated according to how small the volume of the triangular pyramid formed by the four positioning satellites is compared to the case of the DOP value 1.
- the operation unit 103 includes a release button 103a, operation buttons 103b to 103g, and the like, and sends an operation signal corresponding to the operated button to the control circuit 101.
- the memory card interface 107 writes an image file to the memory card 150 or reads an image file from the memory card 150 according to an instruction from the control circuit 101.
- the memory card 150 is an external recording medium that can be attached to and detached from the memory card interface 107.
- the display monitor 104 displays information such as an image and text of an image file according to an instruction from the control circuit 101. On the display monitor 104, one image is displayed in a large size, or a plurality of reduced images (thumbnails) are displayed.
- the speaker 105 outputs sound in response to an instruction from the control circuit 101.
- the external interface circuit 106 transmits and receives commands and data to and from an external device (personal computer, printer, etc.) via a cable (not shown) according to an instruction from the control circuit 101.
- the power supply 108 has a battery, a power supply circuit, and the like, and supplies power to each part of the electronic camera 1.
- the photometric device 109 detects the luminance of the subject by the photometric sensor and sends the luminance information to the control circuit 101.
- the control circuit 101 calculates exposure setting values such as a shutter speed and an aperture value based on the luminance information.
- the illumination device 110 illuminates the subject with a predetermined amount of light, for example, flashing when a light emission instruction is issued from the control circuit 101 during photographing. The illumination light is emitted forward of the electronic camera.
- the map data storage device 111 stores map data for displaying a map on the display monitor 104.
- the GPS interface circuit 112 is an interface for connecting the electronic camera 1 to the GPS device 2.
- the imaging unit 120 includes an imaging optical system 121, an imaging element 122, and an imaging control circuit 123, and performs imaging of a subject in accordance with an instruction from the control circuit 101.
- the imaging optical system 121 forms a subject image on the imaging surface of the imaging element 122.
- As the imaging device 122 a CCD (Charge Coupled Device) imaging device, a CMOS (Complementary Metal Oxide Semiconductor) imaging device, or the like is used.
- the imaging control circuit 123 drives and controls the imaging element 122 according to an instruction from the control circuit 101 and performs predetermined signal processing on the image signal output from the imaging element 122.
- the image data after the signal processing is recorded in the memory card 150 as the above-mentioned Exif image file stipulated image file.
- the GPS device 2 measures the position of the GPS device 2 based on the propagation time of the radio wave transmitted from the positioning satellite and the position of the positioning satellite. Since the radio wave is transmitted from the positioning satellite in synchronization with the accurate clock, the GPS device 2 can calculate the propagation time from the radio wave reception time.
- the radio wave transmitted from the positioning satellite includes orbit data of the positioning satellite, and the GPS device 2 can calculate the position of the positioning satellite from the orbit data. Further, the GPS device 2 calculates the above DOP value.
- FIG. 12 is a diagram for explaining a display screen of the display monitor 104 that displays an image of an image file stored in the memory card 150.
- images 41a to 41d of image files whose DOP values are equal to or less than a predetermined value are displayed side by side as reduced images. Therefore, the image of the image file whose DOP value is larger than the predetermined value is not displayed.
- the positions of the positioning data of these images 41a to 41d are substantially the same positions as the positions where the GPS device 2 connected to the electronic camera 1 is positioning, that is, the current position of the electronic camera 1.
- the images 41a to 41d are displayed from left to right and from the upper level to the lower level in the order from the smallest DOP value to the larger one, that is, from the highest positioning accuracy to the lowest. . Accordingly, the user can recognize that the positioning accuracy of the image 41a is the highest (the DOP value is the smallest), and the positioning accuracy deteriorates (the DOP value increases) as the images 41b, 41c, and 41d progress. .
- DOP thumbnail display refers to displaying an image of an image file as a reduced image based on the DOP value of the image file.
- step S501 the GPS device 2 measures the current position.
- step S502 the DOP value when the current position is measured is acquired from the GPS device 2.
- step S503 an image file in which the position of positioning data is the same as the current position is searched from image files stored in the memory card 150.
- the same as the current position means that the current position and the position of the positioning data are the same within a predetermined allowable range.
- step S504 an image file having a DOP value equal to or lower than the DOP value acquired from the GPS device 2 is searched from the image files searched in step S503.
- step S505 a reduced image of the image file searched in step S504 is created.
- step S506 the reduced images of the image file are displayed side by side from left to right and from the top to the bottom, in the order from the image of the image file with the small DOP value to the image of the image file with the large DOP value.
- An image file having a predetermined positioning accuracy or higher is searched from a plurality of image files having positioning data 31 to 34 and positioning accuracy information 35 in the positioning data, and an image 41a of the searched image file is searched. -41d are displayed on the display monitor 104. Thereby, only an image file having a predetermined positioning accuracy or higher can be displayed.
- a GPS device 2 for detecting the current position of the electronic camera 1 is provided, and an image file having positioning data substantially matching the current position detected by the GPS device 2 is searched from a plurality of image files.
- Image files having a predetermined positioning accuracy or higher are searched from the searched image files, and the images 41a to 41d of the searched image files are displayed on the display monitor 104.
- the above third embodiment can be modified as follows. (1) In the third embodiment described above, an image file having positioning data that substantially matches the current position of the electronic camera 1 is searched, and an image file having a predetermined positioning accuracy or higher is searched from the searched image files.
- the images 41a to 41d of the searched image file are displayed on the display monitor 104.
- an image file having positioning data within a predetermined distance from the current position of the electronic camera 1 is searched, an image file having a predetermined positioning accuracy or more is further searched from the searched image files, and the searched image file is searched.
- the image of the file may be displayed on the display monitor 104. This is convenient when an image file with high positioning accuracy around the current position is required. For example, it is convenient when shooting around the current position at the same place as when shooting before, or when checking how much the current scenery has changed compared to the old scenery.
- the images of the image file are displayed side by side in the order from the image of the image file with high positioning accuracy to the image of the image file with low positioning accuracy.
- the order of arranging the images in the image file is determined according to the DOP value, that is, the value representing the positioning accuracy, multiplied by the distance from the current position to the position of the positioning data, and the images in the image file are displayed side by side. May be. Thereby, it is possible to determine the order in which the images are arranged by comprehensively determining the positioning accuracy and the distance from the current position.
- the image of the image file to be displayed may be an image in which the position of the positioning data is substantially the same as the current position of the electronic camera 1 or an image in which the position of the positioning data is within a predetermined distance from the current position of the electronic camera 1. Good. In addition, only an image with a positioning accuracy higher than a predetermined positioning accuracy may be displayed.
- FIG. 14 A modification of the image file image display method according to the third embodiment of the present invention will be described with reference to the flowchart of FIG.
- the processing in FIG. 14 is executed in the control circuit 101 by a program that starts when the user operates the operation buttons 103b to 103g to select the DOP thumbnail display function.
- step S601 the GPS device 2 measures the current position.
- step S602 the DOP value when the current position is measured is acquired from the GPS device 2.
- step S603 an image file in which the position of the positioning data is within a predetermined distance from the current position is searched from image files stored in the memory card 150.
- step S604 an image file having a DOP value equal to or lower than the DOP value acquired from the GPS device 2 is searched from the image files searched in step S603.
- step S605 for each image file searched in step S604, a value (U) obtained by multiplying the distance from the current position to the position of the positioning data by the DOP value is calculated.
- step S606 a reduced image of the searched image file is created.
- step S607 the reduced images of the image files are displayed side by side in the order from the image file with a small U value to the image file with a large U value.
- the electronic camera 1 according to the fourth embodiment of the present invention displays an image on the display monitor 104 while changing the size of the image.
- the order of image sizes is determined based on the DOP value of the image file.
- the image is displayed together with a map around the position of the positioning data. Since the configuration of the electronic camera 1 according to the fourth embodiment is not different from the configuration of the electronic camera 1 according to the third embodiment, description of the configuration of the electronic camera 1 according to the fourth embodiment is omitted.
- FIG. 15 is a diagram for explaining a display screen of the display monitor 104 displaying an image of the image file stored in the memory card 150.
- images 42a to 42c of image files having a DOP value equal to or smaller than a predetermined value are displayed as reduced images together with a map 50A.
- the map 50 ⁇ / b> A is a map around the current position 51 of the electronic camera 1.
- the position of the positioning data in the image files of the images 42a to 42c is almost the same position as the current position 51 of the electronic camera 1. Accordingly, the map 50A is also a map around the position of the positioning data in the image files of the images 42a to 42c.
- the sizes of the images 42a to 42c become smaller in the order from the smallest DOP value to the larger one, that is, in the order from the higher positioning accuracy to the lower one. Therefore, the positioning accuracy of the largest image 42a is the highest (the DOP value is the smallest), and the positioning accuracy of the smallest image 42c is the worst (the DOP value is large).
- the image display processing of the image file in the fourth embodiment of the present invention will be described with reference to the flowchart of FIG.
- the processing in FIG. 16 is executed in the control circuit 101 by a program that starts when the user operates the operation buttons 103b to 103g to select the DOP thumbnail display function.
- the processes common to the processes in FIG. 13 are denoted by the same reference numerals, and differences from the fifth process will be mainly described.
- step S504 the process proceeds to step S801.
- step S801 the reduced image of the image file searched in step S504 is created by gradually reducing the size in the order from the smallest DOP value to the largest one.
- step S802 a map around the current position is displayed on the display monitor 104 at a predetermined scale.
- step S803 the reduced images are displayed away from the current position as the DOP value increases so that the images do not overlap.
- the image 42a (42b) of the image file with high positioning accuracy is displayed larger than the image 42b (42c) of the image file with low positioning accuracy. Thereby, an image with high positioning accuracy can be preferentially selected.
- an image of an image file having positioning data that substantially matches the current position of the electronic camera 1 is displayed on the map.
- an image of an image file having positioning data within a predetermined distance from the current position of the electronic camera 1 may be displayed.
- images 43a to 43c may be displayed as shown in FIG.
- the position of the positioning data in the image 43a is denoted by reference numeral 44a
- the position of the positioning data in the image 43b is denoted by reference numeral 44b
- the position of the positioning data in the image 43c is denoted by reference numeral 44c.
- the positions 44a to 44c of these positioning data are within a predetermined distance 52 from the current position 51 of the electronic camera 1.
- FIG. 18 A modification of the image file image display process in the fourth embodiment of the present invention will be described with reference to the flowchart of FIG.
- the processing in FIG. 18 is executed in the control circuit 101 by a program that starts when the user operates the operation buttons 103b to 103g to select the DOP thumbnail display function. Processes common to those in FIG. 14 are denoted by common reference numerals, and different portions from the processes in FIG. 14 will be mainly described.
- step S1001 the reduced image of the image file searched in step S604 is created by gradually reducing the size in the order from the smallest DOP value to the largest one.
- step S1002 a map around the current position is displayed on the display monitor 104 at a predetermined scale.
- step S1003 a reduced image is displayed near the position of the positioning data.
- the size of the plurality of images displayed on the display monitor 104 together with the map is not limited to the third embodiment. For example, you may make it make the magnitude
- the electronic camera 1 according to the fifth embodiment of the present invention determines a DOP value serving as a reference for searching for an image file to be displayed on the display monitor 104 according to the scale of the map displayed on the display monitor 104. Since the configuration of the electronic camera 1 of the fifth embodiment is not different from the configuration of the electronic camera 1 of the third embodiment, description of the configuration of the electronic camera 1 of the fifth embodiment is omitted.
- FIG. 19 is a diagram for explaining a display screen of the display monitor 104 that displays an image of an image file stored in the memory card 150.
- FIG. 19A is a diagram for explaining a display screen in which an image is displayed on a map of a predetermined scale.
- FIG. 19B shows a map scale as compared with the map of FIG. It is a figure for demonstrating the display screen which displayed the image on the map when it changes to the detail side.
- images 43a to 43c of image files having a DOP value equal to or smaller than a predetermined value are displayed as reduced images together with the map 50A.
- the sizes of the images 43a to 43c are gradually reduced in the order from the smallest DOP value to the largest one.
- the DOP value that is a criterion for displaying the image of the image file is changed from a predetermined value to a smaller value.
- the image 43a and the image 43b are displayed, and the image 43c with poor positioning accuracy is not displayed.
- the closer the scale of the map displayed on the display monitor 104 is to the detailed side the smaller the DOP value that is a criterion for displaying the image of the image file, and the closer the scale of the map displayed on the display monitor 104 is to the wide area side, The DOP value that is a criterion for displaying the image of the image file is increased.
- FIG. 20 An image file image display process according to the fifth embodiment of the present invention will be described with reference to the flowchart of FIG.
- the processing in FIG. 20 is executed in the control circuit 101 by a program that starts when the user operates the operation buttons 103b to 103g to select the DOP thumbnail display function.
- Processes common to those in FIG. 18 are denoted by common reference numerals, and different parts from the processes in FIG. 18 will be mainly described.
- the user can set the scale of the map displayed on the display monitor 104 by operating the operation buttons 103b to 103g. Thereby, the user can change the scale of the map.
- step S601 the process proceeds to step S1201.
- step S1201 the scale of the map set in the electronic camera 1 is detected.
- step S1202 a reference DOP value is determined from the scale of the map.
- the reference DOP value becomes smaller as the scale of the map is closer to the detailed side, and the reference DOP value becomes larger as the scale of the map is larger.
- step S603 the process proceeds to step S1203.
- step S1203 an image file having a DOP value equal to or lower than the reference DOP value is searched. Then, the process proceeds to step S1001 and proceeds to step S1204. In step S1204, the map is displayed on the display monitor 104 at the set scale. Then, the process proceeds to step S1003, and the process proceeds to step S1205. In step S1205, it is determined whether or not the DOP thumbnail display is ended by the user's operation of the operation buttons 103b to 103g. When the DOP thumbnail display is ended, an affirmative determination is made in step S1205, and the image display process is ended. If the DOP thumbnail display has not ended, a negative determination is made in step S1205, and the process returns to step S501.
- the following operational effects can be obtained in addition to the operational effects of the third embodiment.
- the reference DOP value which is a criterion for determining whether images can be displayed
- the DOP value of the image file is equal to or less than the changed reference DOP value.
- the above fifth embodiment can be modified as follows.
- the size of the plurality of images displayed together with the map is not limited to the fifth embodiment. For example, you may make it make the magnitude
- an electronic camera according to a sixth embodiment of the present invention will be described with reference to the drawings.
- the user can set a DOP value that is a reference for searching for an image file to be displayed. Since the configuration of the electronic camera 1 according to the sixth embodiment is not different from the configuration of the electronic camera 1 according to the third embodiment, description of the configuration of the electronic camera 1 according to the sixth embodiment is omitted.
- FIG. 21 is a diagram for explaining a display screen of the display monitor 104 that displays an image of an image file stored in the memory card 150 on a predetermined map.
- a reference DOP value display column 60 indicating a DOP value input by the user operating the operation buttons 103b to 103g is displayed together with the map 50B and the images 43a to 43c.
- a scale of the DOP value and an inverted triangle mark 61 moving on the scale are displayed.
- the mark 61 moves on the scale of the DOP value input by the user.
- the sizes of the images 43a to 43c are gradually reduced in the order from the smallest DOP value to the largest one.
- FIG. 21A is a display screen when the user inputs “3” as the DOP value. Images 43a and 43b whose DOP value of the image file is 3 or less are displayed on the display screen. FIG. 21B is a display screen when the user inputs “7” as the DOP value. In addition to the images 43a and 43b whose image file DOP value is 3 or less, an image 43c whose DOP value is greater than 3 and 7 or less is also displayed on the display screen.
- the image display processing of the image file in the sixth embodiment of the present invention will be described with reference to the flowchart of FIG.
- the processing in FIG. 22 is executed in the control circuit 101 by a program that starts when the user operates the operation buttons 103b to 103g to select the DOP thumbnail display function.
- the processes common to those in FIG. 20 are denoted by the same reference numerals, and differences from the processes in FIG. 20 will be mainly described.
- step S1401 the reference DOP value is set to 2. Then, the process proceeds to step S601 and step S603. After step S1003, the process proceeds to step S1402.
- step S1402 the user operates the operation buttons 103b to 103g to input a DOP value, and determines whether or not the reference DOP value has been changed. If the reference DOP value has been changed, an affirmative determination is made in step S1402 and the process proceeds to step S1403. If the reference DOP value has not been changed, a negative determination is made in step S1402 and the process proceeds to step S1205. In step S1403, the reference DOP value is changed to the DOP value input by the user. Then, the process returns to step S601.
- the sixth embodiment described above can be modified as follows.
- the size of the plurality of reduced images displayed together with the map is not limited to the fifth embodiment.
- a plurality of reduced images displayed together with the map may all have the same size.
- the size of the image corresponding to the DOP value may be determined in advance, and the size of the image to be displayed may be changed according to the DOP value.
- an electronic camera according to a seventh embodiment of the present invention will be described with reference to the drawings.
- the electronic camera 1 according to the seventh embodiment of the present invention when one image is selected from a plurality of images displayed side by side on the display monitor 104, the position of the positioning data of the selected image is displayed. A surrounding map is displayed on the display monitor 104. In addition, the selected image is displayed on the map. Since the configuration of the electronic camera 1 of the seventh embodiment is not different from the configuration of the electronic camera 1 of the third embodiment, the description of the configuration of the electronic camera 1 of the seventh embodiment is omitted.
- FIG. 23A and FIG. 24A are diagrams for explaining a display screen on which images 41a to 41d of an image file having a DOP value equal to or less than a predetermined value are displayed side by side as reduced images.
- the image 45a is selected by the user among the plurality of images 45a to 45d
- the image 45c is selected from the plurality of images 45a to 45d by the user.
- the image can be selected by the user operating the operation buttons 103b to 103g.
- FIG. 23 (b) is a diagram for explaining a display screen displayed after the image 45a is selected in FIG. 23 (a).
- a map 50B around the positioning position 46a of the selected image 45a and the selected image 45a are displayed on the display screen.
- the scale of the map 50B is determined based on the positioning accuracy of the positioning data of the image 45a. When the positioning accuracy is high, that is, when the DOP value is small, a detailed map is displayed, and when the positioning accuracy is poor, that is, when the DOP value is large, a wide area map is displayed.
- FIG. 24 (b) is a diagram for explaining a display screen displayed after the image 45c is selected in FIG. 24 (a).
- a map 50A around the positioning position 46c of the selected image 45c and the selected image 45c are displayed. Since the positioning accuracy of the image 45c is worse than that of the image 45a, the scale of the map 50A is on the wide area side compared to the map 50B. The size of the images 45a and 45c displayed on the maps 50A and 50B is the same.
- the image display processing of the image file in the seventh embodiment of the present invention will be described with reference to the flowchart of FIG.
- the processing in FIG. 25 is executed in the control circuit 101 by a program that starts when the user operates the operation buttons 103b to 103g to select the DOP thumbnail display function.
- the user can select a desired image from images displayed side by side on the display monitor 104 by operating the operation buttons 103b to 103g.
- step S1701 a reduced image is created from the image file stored in the memory card 150.
- step S1702 the reduced images are displayed side by side on the display monitor 104.
- step S1703 it is determined whether a reduced image has been selected. If a reduced image is selected, an affirmative determination is made in step S1703 and the process proceeds to step S1704. If a reduced image is not selected, step S1703 is repeated.
- step S1704 the position and DOP value of the positioning data of the selected reduced image are detected.
- step S1705 the scale of the map is determined based on the detected DOP value. As described above, when the DOP value is small, the map scale is set to the detailed side, and when the DOP value is large, the map scale is set to the wide area side.
- step S1706 a map around the position of the positioning data is displayed at the determined scale.
- step S1707 the reduced image selected in step S1703 is displayed on the map.
- step S1708 it is determined whether or not an operation for returning to the previous screen has been performed by the user's operation of the operation buttons 103b to 103g. If an operation to return to the previous screen is performed, an affirmative determination is made in step S1708, and the process returns to step S1702. When the operation for returning to the previous screen is not performed, a negative determination is made in step S1708, and the process proceeds to step S1709.
- step S1709 it is determined whether or not the DOP thumbnail display is finished by the user's operation of the operation buttons 103b to 103g. When the DOP thumbnail display is ended, an affirmative determination is made in step S1709, and the process ends. If the DOP thumbnail display has not ended, a negative determination is made in step S1709, and the process returns to step S1708.
- the following operational effects can be obtained in addition to the operational effects of the third embodiment.
- (1) When one image is selected from the images displayed side by side on the display monitor 104, a map around the position of the positioning data in the image file of the selected image is displayed. Thereby, the position where the image of the image file is taken can be confirmed on the map, which is convenient.
- the scale of the map displayed when one image is selected from the images displayed side by side on the display monitor 104 is determined based on the positioning accuracy in the image file of the selected image. Thereby, although it is actually an image taken at a position outside the map display range, it can be prevented from being displayed as if it was taken within the map display range due to poor positioning accuracy. .
- the seventh embodiment described above can be modified as follows. (1) When images are arranged and displayed on the display monitor 104 for selection by the user, an image of an image file having a predetermined positioning accuracy or higher may be displayed. This is because, when an image with extremely poor positioning accuracy is displayed together with a map, the scale of the map determined based on the positioning accuracy is likely to be on the wide area side, so the position of positioning data is not well understood from the map.
- the size of the image corresponding to the DOP value may be determined in advance, and the size of the image displayed together with the map may be changed according to the DOP value. This is convenient because the positioning accuracy can be recognized just by looking at the size of the image. Alternatively, the order of arrangement based on the positioning accuracy of the image file may be determined and the images may be displayed side by side, and the size of the image displayed together with the map may be determined in that order.
- an image is displayed on a map.
- the images 47a to 47e may be displayed so as to overlap the radar chart 70 as shown in FIG.
- an image closer to the center 71 of the radar chart 70 is an image having a smaller DOP value
- an image farther from the center 71 is an image having a larger DOP value.
- the positioning accuracy of the images 47a to 47e can be grasped visually, which is convenient.
- the sizes of the images 47a to 47e may be changed according to the order of the DOP values or according to the DOP values.
- the current position 81 of the electronic camera 1 is displayed as shown in FIG. It may be displayed so that the map is not displayed. Also in this case, it is possible to determine an image in which the position of the positioning data is close to or far from the current position 81 of the electronic camera 1.
- the DOP value when the GPS device 2 measures the current position is the DOP value that serves as a reference when searching for an image file.
- the DOP value used as a reference when searching for the image file is larger than the DOP value of the positioning accuracy that causes the point that is significantly distant from the point where the image was actually taken to be mistakenly recognized.
- the present invention is not limited to the embodiment.
- a numerical value for determining whether or not to display an image is not limited to a DOP value as long as the numerical value can represent positioning accuracy.
- the GPS device 2 of the electronic camera 1 is attached.
- the GPS device 2 may be built in the electronic camera 1.
- the present invention is not limited to the electronic camera 1 provided with the display monitor 104 as long as it is an image display device.
- a mobile phone having a GPS function may be used.
- Processing such as image display in the above-described embodiment may be performed by a personal computer.
- the processing program can be provided through a recording medium such as a CD-ROM or a data signal such as the Internet.
- the processing program is supplied as a computer-readable computer program product in various forms such as a recording medium and a data signal (carrier wave).
- the display may be changed according to the positioning accuracy DOP value as in the third to seventh embodiments.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Probability & Statistics with Applications (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Studio Devices (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
しかしながら、上述した従来の位置データの提供方法では、撮影の都度、位置データの精度限度を指示しなければならず、この指示操作を怠ると、自宅周辺の画像とともに、自宅周辺の正確な位置データが公開されることになる。
本発明の第2の態様によると、第1の態様の撮像装置において、制御部は、撮影位置に基づいて、撮影画像とともに撮影位置のデータを記録するか否かを決定するのが好ましい。
本発明の第3の態様によると、第2の態様の撮像装置において、所定の広さの特定場所に対する撮影位置データの記録可否と記録可の場合の撮影位置データの記録精度とを設定した精度記憶部を備え、制御部は、撮影位置を精度記憶部の特定場所と照合し、撮影画像とともに撮影位置データを記録するか否かと、記録する場合の撮影位置データの記録精度を決定するのが好ましい。
本発明の第4の態様によると、第2または第3の態様の撮像装置において、特定人物に対する特徴を記憶した人物記憶部と、人物記憶部を参照して撮影画像の中から特定人物を認識する人物認識部とを備え、制御部は、人物認識部により撮影画像の中に特定人物が認識された場合には、撮影位置データの記録精度を低精度に変換するのが好ましい。
本発明の第5の態様によると、第3または第4の態様の撮像装置において、測位部は撮影位置データの測位精度を出力し、制御部は、測位精度が制御部により決定された記録精度より低い場合には、測位精度が記録精度を満たすまで測位部による測位をやり直すのが好ましい。
本発明の第6の態様によると、第1の態様の撮像装置において、制御部は、撮影画像に基づいて、撮影画像とともに撮影位置のデータを記録するか否かを決定するのが好ましい。
本発明の第7の態様によると、第6の態様の撮像装置において、特定人物に対する特徴を記憶した人物記憶部と、人物記憶部を参照して撮影画像の中から特定人物を認識する人物認識部とを備え、制御部は、人物認識部により撮影画像の中に特定人物が認識されたかどうかによって記録精度を決定するのが好ましい。
本発明の第8の態様によると、撮像装置は、画像を撮影する撮影部と、撮影部により画像を撮影したときの撮影位置を測位する測位部と、測位部で測位された撮影位置のデータの記録精度を決定する制御部と、撮影画像と撮影位置データと制御部により決定された記録精度を記録する記録部とを備える。
本発明の第9の態様によると、第8の態様の撮像装置において、制御部は、撮影位置に基づいて撮影位置のデータの記録精度を決定するのが好ましい。
本発明の第10の態様によると、第8の態様の撮像装置において、制御部は、撮影画像に基づいて撮影位置のデータの記録精度を決定するのが好ましい。
本発明の第11の態様によると、第3から10のいずれかの態様の撮像装置において、記録部に記録された記録精度に応じて記録部に記録された撮影画像の表示の態様を変更する表示制御部をさらに備えるのが好ましい。
本発明の第12の態様によると、画像表示装置は、測位データとその測位データにおける測位精度の情報とを有する複数の画像ファイルの中から、所定の測位精度以上の画像ファイルを検索する画像ファイル検索部と、画像ファイル検索部によって検索された画像ファイルの画像を表示器に表示する画像ファイル表示制御部とを備える。
本発明の第13の態様によると、第12の態様の画像表示装置において、画像ファイル表示制御部は、測位精度の高い画像ファイルの画像から測位精度の低い画像ファイルの画像へ至る順序で並べて検索された画像ファイルの画像を表示するのが好ましい。
本発明の第14の態様によると、第12または13の態様の画像表示装置において、画像ファイル表示制御部は、測位精度の高い画像ファイルの画像を測位精度の低い画像ファイルの画像に比べて大きく表示するのが好ましい。
本発明の第15の態様によると、第12~14のいずれか1の態様の画像表示装置において、画像表示装置の現在位置を検出する現在位置検出部を備え、画像ファイル検索部は、複数の画像ファイルの中から、現在位置検出部によって検出された現在位置と一致する測位データ、または現在位置から所定距離内の測位データを有する画像ファイルを検索し、その検索された画像ファイルの中から所定の測位精度以上の画像ファイルを検索するがのが好ましい。
本発明の第16の態様によると、第12の態様の画像表示装置において、画像表示装置の現在位置を検出する現在位置検出部を備え、画像ファイル表示制御部は、現在位置から測位データの位置までの距離を測位精度を表す値に乗算した値に応じて、画像ファイルの画像を並べて表示するのが好ましい。
本発明の第17の態様によると、第16の態様の画像表示装置において、画像ファイル検索部は、複数の画像ファイルの中から、現在位置検出装置によって検出された現在位置と一致する測位データ、または現在位置から所定距離内の測位データを有する画像ファイルを検索し、その検索された画像ファイルの中から所定の測位精度以上の画像ファイルを検索するがのが好ましい。
本発明の第18の態様によると、
第12~176のいずれか1の態様の画像表示装置において、画像ファイル表示制御部は、画像ファイルの画像を、画像ファイルにおける測位データの位置周辺の地図とともに表示するのが好ましい。
本発明の第19の態様によると、第18の態様の画像表示装置において、地図の縮尺を変更する地図縮尺変更部を備え、画像ファイル検索部は、地図縮尺変更部によって地図の縮尺が変更されると、変更された地図の縮尺に基づいて所定の測位精度を変更して検索し、画像ファイル表示制御部は、画像ファイル検索部によって検索された画像ファイルの画像を、地図縮尺変更部によって縮尺が変更された地図とともに表示するのが好ましい。
本発明の第20の態様によると、第12~17のいずれか1の態様の画像表示装置において、画像ファイル表示制御部は、画像ファイルの画像を、レーダーチャートとともに表示するのが好ましい。
本発明の第21の態様によると、第13の態様の画像表示装置において、画像ファイル表示制御部によって表示された画像の中から選択された画像を入力する画像入力部と、画像入力部によって入力された画像の画像ファイルにおける測位データの位置周辺の地図を表示する地図表示制御部とを備えるのが好ましい。
本発明の第22の態様によると、第21の態様の画像表示装置において、地図表示制御部は、画像入力部によって入力された画像の画像ファイルにおける測位精度に基づいて表示する地図の縮尺を決定するのが好ましい。
本発明の第23の態様によると、第12~22のいずれか1の態様の画像表示装置において、測位精度を入力する測位精度入力部を備え、画像ファイル検索部は、測定精度入力部によって入力された測位精度を所定の測定精度として検索を行うのが好ましい。
本発明の第24の態様によると、電子カメラは、第12~23のいずれか1の態様の画像表示装置を備える。
図1は発明の一実施の形態の構成を示す図である。カメラAとカメラBは、例えば所有者が同一であるなど、互いに“紐付けられている関係”にある。つまり、撮影した画像に撮影場所の位置データを付加する場合に、位置データを変換するために参照するデータベース(精度変換データベース、要求精度テーブル、人物データベースなど)がまったく同一であるカメラどうしである。撮影者によっては一眼レフカメラとコンパクトカメラを所持し、撮影対象や撮影条件に応じてそれらを使い分けることがある。そのような場合に例えば一眼レフカメラをカメラAとし、コンパクトカメラをカメラBとして同一の精度変換に関するデータを共有する。
第1の実施の形態では、図5に示すような精度変換データベースA8、B8、C3を使用する例を説明した。すなわち、画像における人物の認識と撮影場所がどの領域内かによるかの組み合わせによって、位置データを記録するか否かを判断したり、記録精度を変換する例を示した。第2の実施の形態では、画像データに基づいてのみ、位置データを記録するか否かを判断したり、記録精度を変換する例を説明する。
以下、図を参照して本発明の第3の実施の形態による電子カメラを説明する。
(1)測位データ31~34とその測位データにおける測位精度の情報35とを有する、複数の画像ファイルの中から、所定の測位精度以上の画像ファイルを検索し、検索された画像ファイルの画像41a~41dを表示モニタ104に表示するようにした。これにより、所定の測位精度以上の画像ファイルのみを表示することができる。
(1)以上の第3の実施の形態では、電子カメラ1の現在位置とほぼ一致する測位データを有する画像ファイルを検索し、その検索された画像ファイルの中から所定の測位精度以上の画像ファイルをさらに検索して、検索された画像ファイルの画像41a~41dを表示モニタ104に表示するようにした。しかし、電子カメラ1の現在位置から所定距離内の測位データを有する画像ファイルを検索し、その検索された画像ファイルの中から所定の測位精度以上の画像ファイルをさらに検索して、検索された画像ファイルの画像を表示モニタ104に表示するようにしてもよい。これにより、現在位置周辺の測位精度の高い画像ファイルを必要とするとき便利である。たとえば、現在位置の周辺で、以前撮影したときと同じ場所で撮影を行いたい場合や、昔の景色と比べて今の景色がどの程度景色が変わったかを調べたい場合などに便利である。
以下、図を参照して本発明の第4の実施の形態による電子カメラを説明する。本発明の第4の実施の形態による電子カメラ1は、画像の大きさを変えて画像を表示モニタ104に表示する。画像の大きさの順位は画像ファイルのDOP値に基づいて決定される。また、画像は測位データの位置周辺の地図とともに表示される。第4の実施の形態の電子カメラ1の構成は、第3の実施の形態の電子カメラ1の構成と変わらないので、第4の実施の形態の電子カメラ1の構成の説明は省略する。
(1)測位精度の高い画像ファイルの画像42a(42b)を測位精度の低い画像ファイルの画像42b(42c)に比べて大きく表示するようにした。これにより、測位精度の高い画像を優先的に選択することができる。
(1)以上の第4の実施の形態では、電子カメラ1の現在位置とほぼ一致する測位データを有する画像ファイルの画像を地図上に表示した。しかし、電子カメラ1の現在位置から所定距離内の測位データを有する画像ファイルの画像を表示するようにしてもよい。たとえば、図17に示すように画像43a~43cを表示するようにしてもよい。ここで、画像43aの測位データの位置は符号44aであり、画像43bの測位データの位置は符号44bであり、画像43cの測位データの位置は符号44cである。これらの測位データの位置44a~44cは電子カメラ1の現在位置51から所定距離内52である。
以下、図を参照して本発明の第5の実施の形態による電子カメラを説明する。本発明の第5の実施の形態による電子カメラ1は、表示モニタ104に表示される地図の縮尺によって、表示モニタ104に表示する画像ファイルを検索する基準となるDOP値を決定する。第5の実施の形態の電子カメラ1の構成は、第3の実施の形態の電子カメラ1の構成と変わらないので、第5の実施の形態の電子カメラ1の構成の説明は省略する。
地図の縮尺が変更されると、変更された地図の縮尺に基づいて画像表示可否の判断基準となる基準DOP値を変更し、画像ファイルのDOP値が、変更された基準DOP値以下である画像を、縮尺が変更された地図とともに表示するようにした。これにより、地図の縮尺が詳細側に変更され、地図の表示範囲は狭くなったことが原因で、実際は地図の表示範囲から外れた位置で撮影された画像であるにもかかわらず、地図の表示範囲内で撮影されたように表示されるのを防止することができる。
地図とともに表示する複数の画像の大きさは第5の実施の形態に限定されない。たとえば、地図とともに表示する複数の画像の大きさを全て同じにするようにしてもよい。また、DOP値に対応する画像の大きさを予め決めておき、DOP値に応じて表示する画像の大きさを変えるようにしてもよい。
以下、図を参照して本発明の第6の実施の形態による電子カメラを説明する。本発明の第6の実施の形態による電子カメラ1は、表示する画像ファイルを検索する基準となるDOP値をユーザが設定できる。第6の実施の形態の電子カメラ1の構成は、第3の実施の形態の電子カメラ1の構成と変わらないので、第6の実施の形態の電子カメラ1の構成の説明は省略する。
地図とともに表示する複数の縮小画像の大きさは第5の実施の形態に限定されない。たとえば、地図とともに表示する複数の縮小画像の大きさを全て同じにするようにしてもよい。また、DOP値に対応する画像の大きさを予め決めておき、DOP値に応じて表示する画像の大きさを変えるようにしてもよい。
以下、図を参照して本発明の第7の実施の形態による電子カメラを説明する。本発明の第7の実施の形態による電子カメラ1では、表示モニタ104に並べて表示されている複数の画像の中から1つの画像がユーザによって選択されると、選択された画像の測位データの位置周辺の地図が表示モニタ104に表示される。また、その地図の上に選択された画像が表示される。第7の実施の形態の電子カメラ1の構成は、第3の実施の形態の電子カメラ1の構成と変わらないので、第7の実施の形態の電子カメラ1の構成の説明は省略する。
(1)表示モニタ104に並べて表示された画像の中から1つの画像が選択されると、選択された画像の画像ファイルにおける測位データの位置周辺の地図を表示するようにした。これにより、画像ファイルの画像を撮影した位置などを地図上で確認することができ、便利である。
(1)ユーザに選択させるために画像を並べて表示モニタ104に表示するとき、所定の測位精度以上の画像ファイルの画像を表示するようにしてもよい。測位精度が著しく悪い画像を地図とともに表示した場合、測位精度に基づいて決定される地図の縮尺は著しく広域側になると思われるので、測位データの位置が地図から見てよくわからないからである。
(1)以上の第2~5の実施の形態では、地図の上に画像を表示した。しかし、図26(a)に示すようにレーダーチャート70に重ねて画像47a~47eを表示するようにしてもよい。ここで、レーダーチャート70の中心71に近い画像ほどDOP値が小さい画像であり、中心71から離れている画像ほどDOP値が大きい画像である。画像47a~47eの測位精度が視覚的に把握でき、便利である。なお、DOP値の順序に応じて、またはDOP値に応じて画像47a~47eの大きさを変えるようにしてもよい。
日本国特許出願2008年第172343号(2008年7月1日出願)
日本国特許出願2008年第182672号(2008年7月14日出願)
Claims (24)
- 画像を撮影する撮影部と、
前記撮影部により画像を撮影したときの撮影位置を測位する測位部と、
前記撮影画像とともに前記撮影位置のデータを記録するか否か決定する制御部と、
前記制御部による決定に応じて、前記撮影画像のみ、または前記撮影画像と前記撮影位置データを記録する記録部とを備える撮像装置。 - 請求項1に記載の撮像装置において、
前記制御部は、前記撮影位置に基づいて、前記撮影画像とともに前記撮影位置のデータを記録するか否かを決定する撮像装置。 - 請求項2に記載の撮像装置において、
所定の広さの特定場所に対する撮影位置データの記録可否と記録可の場合の撮影位置データの記録精度とを設定した精度記憶部を備え、
前記制御部は、前記撮影位置を前記精度記憶部の前記特定場所と照合し、前記撮影画像とともに前記撮影位置データを記録するか否かと、記録する場合の前記撮影位置データの記録精度を決定する撮像装置。 - 請求項2または請求項3に記載の撮像装置において、
特定人物に対する特徴を記憶した人物記憶部と、
前記人物記憶部を参照して前記撮影画像の中から前記特定人物を認識する人物認識部とを備え、
前記制御部は、前記人物認識部により前記撮影画像の中に前記特定人物が認識された場合には、前記撮影位置データの記録精度を低精度に変換する撮像装置。 - 請求項3または請求項4に記載の撮像装置において、
前記測位部は前記撮影位置データの測位精度を出力し、
前記制御部は、前記測位精度が前記制御部により決定された記録精度より低い場合には、前記測位精度が前記記録精度を満たすまで前記測位部による測位をやり直す撮像装置。 - 請求項1に記載の撮像装置において、
前記制御部は、前記撮影画像に基づいて、前記撮影画像とともに前記撮影位置のデータを記録するか否かを決定する撮像装置。 - 請求項6に記載の撮像装置において、
特定人物に対する特徴を記憶した人物記憶部と、
前記人物記憶部を参照して前記撮影画像の中から前記特定人物を認識する人物認識部とを備え、
前記制御部は、前記人物認識部により前記撮影画像の中に前記特定人物が認識されたかどうかによって前記記録精度を決定する撮像装置。 - 画像を撮影する撮影部と、
前記撮影部により画像を撮影したときの撮影位置を測位する測位部と、
前記測位部で測位された前記撮影位置のデータの記録精度を決定する制御部と、
前記撮影画像と前記撮影位置データと前記制御部により決定された記録精度を記録する記録部とを備える撮像装置。 - 請求項8に記載の撮像装置において、
前記制御部は、前記撮影位置に基づいて前記撮影位置のデータの記録精度を決定する撮像装置。 - 請求項8に記載の撮像装置において、
前記制御部は、前記撮影画像に基づいて前記撮影位置のデータの記録精度を決定する撮像装置。 - 請求項3から10のいずれかに記載の撮像装置において、
前記記録部に記録された前記記録精度に応じて前記記録部に記録された前記撮影画像の表示の態様を変更する表示制御部をさらに備える撮像装置。 - 測位データとその測位データにおける測位精度の情報とを有する複数の画像ファイルの中から、所定の測位精度以上の画像ファイルを検索する画像ファイル検索部と、
前記画像ファイル検索部によって検索された画像ファイルの画像を表示器に表示する画像ファイル表示制御部とを備える画像表示装置。 - 請求項12に記載の画像表示装置において、
前記画像ファイル表示制御部は、前記測位精度の高い画像ファイルの画像から前記測位精度の低い画像ファイルの画像へ至る順序で並べて前記検索された画像ファイルの画像を表示する画像表示装置。 - 請求項12または13に記載の画像表示装置において、
前記画像ファイル表示制御部は、測位精度の高い画像ファイルの画像を測位精度の低い画像ファイルの画像に比べて大きく表示する画像表示装置。 - 請求項12~14のいずれか1項に記載の画像表示装置において、
前記画像表示装置の現在位置を検出する現在位置検出部を備え、
前記画像ファイル検索部は、前記複数の画像ファイルの中から、前記現在位置検出部によって検出された現在位置と一致する測位データ、または前記現在位置から所定距離内の測位データを有する画像ファイルを検索し、その検索された画像ファイルの中から前記所定の測位精度以上の画像ファイルを検索するが画像表示装置。 - 請求項12に記載の画像表示装置において、
前記画像表示装置の現在位置を検出する現在位置検出部を備え、
前記画像ファイル表示制御部は、前記現在位置から測位データの位置までの距離を測位精度を表す値に乗算した値に応じて、前記画像ファイルの画像を並べて表示する画像表示装置。 - 請求項16に記載の画像表示装置において、
前記画像ファイル検索部は、前記複数の画像ファイルの中から、前記現在位置検出装置によって検出された現在位置と一致する測位データ、または前記現在位置から所定距離内の測位データを有する画像ファイルを検索し、その検索された画像ファイルの中から前記所定の測位精度以上の画像ファイルを検索するが画像表示装置。 - 請求項12~176のいずれか1項に記載の画像表示装置において、
前記画像ファイル表示制御部は、前記画像ファイルの画像を、前記画像ファイルにおける測位データの位置周辺の地図とともに表示する画像表示装置。 - 請求項18に記載の画像表示装置において、
前記地図の縮尺を変更する地図縮尺変更部を備え、
前記画像ファイル検索部は、前記地図縮尺変更部によって前記地図の縮尺が変更されると、変更された地図の縮尺に基づいて前記所定の測位精度を変更して検索し、
前記画像ファイル表示制御部は、前記画像ファイル検索部によって検索された画像ファイルの画像を、前記地図縮尺変更部によって縮尺が変更された地図とともに表示する画像表示装置。 - 請求項12~17のいずれか1項に記載の画像表示装置において、
前記画像ファイル表示制御部は、前記画像ファイルの画像を、レーダーチャートとともに表示する画像表示装置。 - 請求項13に記載の画像表示装置において、
前記画像ファイル表示制御部によって表示された画像の中から選択された画像を入力する画像入力部と、
前記画像入力部によって入力された画像の画像ファイルにおける測位データの位置周辺の地図を表示する地図表示制御部とを備える画像表示装置。 - 請求項21に記載の画像表示装置において、
前記地図表示制御部は、前記画像入力部によって入力された画像の画像ファイルにおける測位精度に基づいて前記表示する地図の縮尺を決定する画像表示装置。 - 請求項12~22のいずれか1項に記載の画像表示装置において、
前記測位精度を入力する測位精度入力部を備え、
前記画像ファイル検索部は、前記測定精度入力部によって入力された測位精度を前記所定の測定精度として検索を行う画像表示装置。 - 請求項12~23のいずれか1項に記載の画像表示装置を備える電子カメラ。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008172343A JP5109836B2 (ja) | 2008-07-01 | 2008-07-01 | 撮像装置 |
CN200980125950.4A CN102084648B (zh) | 2008-07-01 | 2009-06-24 | 成像装置、图像显示装置和电子照相机 |
US12/999,766 US20110085057A1 (en) | 2008-07-01 | 2009-06-24 | Imaging device, image display device, and electronic camera |
KR1020107028267A KR101600115B1 (ko) | 2008-07-01 | 2009-06-24 | 촬상 장치, 화상 표시 장치와 전자 카메라 |
EP09773356A EP2299701A4 (en) | 2008-07-01 | 2009-06-24 | IMAGING DEVICE, IMAGE DISPLAY DEVICE, AND ELECTRONIC CAMERA |
US14/816,734 US20150341588A1 (en) | 2008-07-01 | 2015-08-03 | Imaging device, image display device, and electronic camera that determines whether to record the position at which an image is photographed and the accuracy of the photographic position to be recorded |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008-172343 | 2008-07-01 | ||
JP2008172343A JP5109836B2 (ja) | 2008-07-01 | 2008-07-01 | 撮像装置 |
JP2008182672A JP5115375B2 (ja) | 2008-07-14 | 2008-07-14 | 画像表示装置および電子カメラ |
JP2008-182672 | 2008-07-14 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/999,766 A-371-Of-International US20110085057A1 (en) | 2008-07-01 | 2009-06-24 | Imaging device, image display device, and electronic camera |
US14/816,734 Continuation US20150341588A1 (en) | 2008-07-01 | 2015-08-03 | Imaging device, image display device, and electronic camera that determines whether to record the position at which an image is photographed and the accuracy of the photographic position to be recorded |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010001778A1 true WO2010001778A1 (ja) | 2010-01-07 |
Family
ID=41465876
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2009/061490 WO2010001778A1 (ja) | 2008-07-01 | 2009-06-24 | 撮像装置、画像表示装置および電子カメラ |
Country Status (6)
Country | Link |
---|---|
US (2) | US20110085057A1 (ja) |
EP (1) | EP2299701A4 (ja) |
JP (1) | JP5109836B2 (ja) |
KR (1) | KR101600115B1 (ja) |
CN (2) | CN102084648B (ja) |
WO (1) | WO2010001778A1 (ja) |
Families Citing this family (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8098899B2 (en) * | 2005-11-14 | 2012-01-17 | Fujifilm Corporation | Landmark search system for digital camera, map data, and method of sorting image data |
JP4548529B2 (ja) * | 2008-08-18 | 2010-09-22 | ソニー株式会社 | 情報記録装置、撮像装置、情報記録方法およびプログラム |
JP2010081427A (ja) * | 2008-09-26 | 2010-04-08 | Casio Computer Co Ltd | 撮像装置、および、プログラム |
JP2010081428A (ja) * | 2008-09-26 | 2010-04-08 | Casio Computer Co Ltd | 撮像装置、および、プログラム |
JP5310237B2 (ja) * | 2009-05-01 | 2013-10-09 | 株式会社ニコン | 撮像装置 |
US9135277B2 (en) | 2009-08-07 | 2015-09-15 | Google Inc. | Architecture for responding to a visual query |
US9087059B2 (en) * | 2009-08-07 | 2015-07-21 | Google Inc. | User interface for presenting search results for multiple regions of a visual query |
US8670597B2 (en) | 2009-08-07 | 2014-03-11 | Google Inc. | Facial recognition with social network aiding |
KR20110040248A (ko) * | 2009-10-13 | 2011-04-20 | 삼성전자주식회사 | 디지털 영상 처리기에서 소비 전력 저감 장치 및 방법 |
US20110128288A1 (en) * | 2009-12-02 | 2011-06-02 | David Petrou | Region of Interest Selector for Visual Queries |
US8977639B2 (en) * | 2009-12-02 | 2015-03-10 | Google Inc. | Actionable search results for visual queries |
US8811742B2 (en) | 2009-12-02 | 2014-08-19 | Google Inc. | Identifying matching canonical documents consistent with visual query structural information |
US9183224B2 (en) | 2009-12-02 | 2015-11-10 | Google Inc. | Identifying matching canonical documents in response to a visual query |
US9176986B2 (en) | 2009-12-02 | 2015-11-03 | Google Inc. | Generating a combination of a visual query and matching canonical document |
US8805079B2 (en) | 2009-12-02 | 2014-08-12 | Google Inc. | Identifying matching canonical documents in response to a visual query and in accordance with geographic information |
US9405772B2 (en) * | 2009-12-02 | 2016-08-02 | Google Inc. | Actionable search results for street view visual queries |
US9852156B2 (en) | 2009-12-03 | 2017-12-26 | Google Inc. | Hybrid use of location sensor data and visual query to return local listings for visual query |
US8477215B2 (en) * | 2010-12-15 | 2013-07-02 | Eka Technologies, Inc. | Wireless data module for imaging systems |
JP5677073B2 (ja) * | 2010-12-15 | 2015-02-25 | キヤノン株式会社 | 画像制御装置及び画像制御方法、情報処理装置及び情報処理方法、プログラム並びに記憶媒体 |
JP2012142825A (ja) * | 2011-01-05 | 2012-07-26 | Sony Corp | 情報処理装置、情報表示方法及びコンピュータプログラム |
JP5698551B2 (ja) | 2011-02-04 | 2015-04-08 | キヤノン株式会社 | 情報処理装置、及びその制御方法 |
JP2012199756A (ja) * | 2011-03-22 | 2012-10-18 | Eastman Kodak Co | 携帯機器 |
JP5723211B2 (ja) * | 2011-05-11 | 2015-05-27 | キヤノン株式会社 | 画像生成装置及びその制御方法、プログラム、及び記録媒体 |
US9424765B2 (en) * | 2011-09-20 | 2016-08-23 | Sony Corporation | Image processing apparatus, image processing method, and program |
CN103577789B (zh) * | 2012-07-26 | 2018-02-13 | 中兴通讯股份有限公司 | 检测方法和装置 |
US8935246B2 (en) | 2012-08-08 | 2015-01-13 | Google Inc. | Identifying textual terms in response to a visual query |
KR102065408B1 (ko) | 2013-08-16 | 2020-01-13 | 엘지전자 주식회사 | 이동 단말기 |
WO2016110967A1 (ja) * | 2015-01-07 | 2016-07-14 | 日立マクセル株式会社 | 情報処理装置及び情報処理方法、及びそれに用いる処理プログラム |
KR101658790B1 (ko) * | 2015-07-21 | 2016-09-23 | 박상운 | 경작지 인증 단말기 |
EP3169069A1 (en) * | 2015-11-10 | 2017-05-17 | FEI Company | Systems and methods for imaging device interfaces |
JP6713153B1 (ja) * | 2019-11-15 | 2020-06-24 | 株式会社Patic Trust | 情報処理装置、情報処理方法、プログラム及びカメラシステム |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH09127594A (ja) | 1995-10-27 | 1997-05-16 | Konica Corp | カメラ |
JPH11295803A (ja) * | 1998-04-15 | 1999-10-29 | Canon Inc | 画像記録装置及びカメラ |
JP2001320759A (ja) | 2000-03-25 | 2001-11-16 | Hewlett Packard Co <Hp> | 移動端末についての位置データを提供する方法 |
JP2006074475A (ja) * | 2004-09-02 | 2006-03-16 | Canon Inc | 撮像装置 |
JP2008172343A (ja) | 2007-01-09 | 2008-07-24 | Funai Electric Co Ltd | ホワイトバランス調整システム、ホワイトバランス調整方法およびpdp表示装置 |
JP2008182672A (ja) | 2007-01-24 | 2008-08-07 | Toshiba Corp | ドキュメント処理装置を介して電子ドキュメントにアクセスするシステムおよび方法 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003207556A (ja) * | 2002-01-10 | 2003-07-25 | Hitachi Ltd | 端末位置情報システムにおける端末およびサーバ装置 |
US7146179B2 (en) * | 2002-03-26 | 2006-12-05 | Parulski Kenneth A | Portable imaging device employing geographic information to facilitate image access and viewing |
CN1719874A (zh) * | 2004-07-07 | 2006-01-11 | 上海乐金广电电子有限公司 | 监视用相机的保密屏蔽设定方法 |
US20060182433A1 (en) * | 2005-02-15 | 2006-08-17 | Nikon Corporation | Electronic camera |
JP2007088754A (ja) * | 2005-09-21 | 2007-04-05 | Olympus Imaging Corp | コンテンツデータ処理装置及びコンテンツデータ処理プログラム |
JP2007266928A (ja) * | 2006-03-28 | 2007-10-11 | Casio Comput Co Ltd | 携帯機器及びプログラム |
US9020940B2 (en) * | 2007-08-07 | 2015-04-28 | Nokia Corporation | Method and device for modifying meta data of media objects |
JP4840403B2 (ja) * | 2008-04-30 | 2011-12-21 | ソニー株式会社 | 情報記録装置、撮像装置、情報記録方法およびプログラム |
-
2008
- 2008-07-01 JP JP2008172343A patent/JP5109836B2/ja not_active Expired - Fee Related
-
2009
- 2009-06-24 WO PCT/JP2009/061490 patent/WO2010001778A1/ja active Application Filing
- 2009-06-24 CN CN200980125950.4A patent/CN102084648B/zh not_active Expired - Fee Related
- 2009-06-24 CN CN2012105550047A patent/CN103067658A/zh active Pending
- 2009-06-24 KR KR1020107028267A patent/KR101600115B1/ko active IP Right Grant
- 2009-06-24 EP EP09773356A patent/EP2299701A4/en not_active Withdrawn
- 2009-06-24 US US12/999,766 patent/US20110085057A1/en not_active Abandoned
-
2015
- 2015-08-03 US US14/816,734 patent/US20150341588A1/en not_active Abandoned
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH09127594A (ja) | 1995-10-27 | 1997-05-16 | Konica Corp | カメラ |
JPH11295803A (ja) * | 1998-04-15 | 1999-10-29 | Canon Inc | 画像記録装置及びカメラ |
JP2001320759A (ja) | 2000-03-25 | 2001-11-16 | Hewlett Packard Co <Hp> | 移動端末についての位置データを提供する方法 |
JP2006074475A (ja) * | 2004-09-02 | 2006-03-16 | Canon Inc | 撮像装置 |
JP2008172343A (ja) | 2007-01-09 | 2008-07-24 | Funai Electric Co Ltd | ホワイトバランス調整システム、ホワイトバランス調整方法およびpdp表示装置 |
JP2008182672A (ja) | 2007-01-24 | 2008-08-07 | Toshiba Corp | ドキュメント処理装置を介して電子ドキュメントにアクセスするシステムおよび方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2299701A4 |
Also Published As
Publication number | Publication date |
---|---|
JP5109836B2 (ja) | 2012-12-26 |
US20150341588A1 (en) | 2015-11-26 |
CN102084648B (zh) | 2014-01-08 |
CN103067658A (zh) | 2013-04-24 |
EP2299701A1 (en) | 2011-03-23 |
US20110085057A1 (en) | 2011-04-14 |
EP2299701A4 (en) | 2012-02-08 |
KR101600115B1 (ko) | 2016-03-04 |
CN102084648A (zh) | 2011-06-01 |
KR20110046393A (ko) | 2011-05-04 |
JP2010016462A (ja) | 2010-01-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2010001778A1 (ja) | 撮像装置、画像表示装置および電子カメラ | |
JP5194650B2 (ja) | 電子カメラ | |
JP4366601B2 (ja) | タイムシフト画像配信システム、タイムシフト画像配信方法、タイムシフト画像要求装置および画像サーバ | |
JP5425341B2 (ja) | 撮影装置及びプログラム | |
KR100897436B1 (ko) | 지리정보 확인시스템의 제어방법 및 이동단말기 | |
JP2006020166A (ja) | 地図表示システム及びデジタルカメラ | |
KR20090019184A (ko) | 전자지도에 포함된 이미지 파일을 이용한 이미지 재생장치, 이의 재생 방법 및 상기 방법을 실행하기 위한프로그램을 기록한 기록매체. | |
KR101998903B1 (ko) | 서버, 클라이언트 단말, 시스템 및 프로그램 | |
JP2001216309A (ja) | 対象物特定装置及びカメラ | |
JP5186984B2 (ja) | 電子カメラ | |
CN108492334A (zh) | 一种基于定位定向数据实现商用相机照片地理标定的方法 | |
JP5007631B2 (ja) | 電子カメラ | |
JP2012019374A (ja) | 電子アルバム作成サーバー、情報処理装置、電子アルバム作成システム、及び、電子アルバム作成サーバーの制御方法 | |
JP5511001B2 (ja) | 撮像装置及び撮影案内システム並びにプログラム | |
KR20090132485A (ko) | 지리정보 확인시스템의 제어방법 및 이동단말기 | |
KR100956114B1 (ko) | 촬상장치를 이용한 지역 정보 제공 장치 및 방법 | |
JP5115375B2 (ja) | 画像表示装置および電子カメラ | |
JP2009111827A (ja) | 撮影装置及び画像ファイル提供システム | |
JP2009116795A (ja) | 情報検索装置、電子カメラ、情報検索方法 | |
JP5034880B2 (ja) | 電子カメラ、画像表示装置 | |
JP2006350592A (ja) | 音楽情報提供装置 | |
JP2007020054A (ja) | 画像管理方法及び装置 | |
JP2006178804A (ja) | 被写体オブジェクト情報提供方法および被写体オブジェクト情報提供サーバ | |
JP2006350879A (ja) | 情報提供システム | |
JP5533201B2 (ja) | 撮像画像表示装置、撮像画像表示システム、制御方法、撮像画像表示方法およびプログラム。 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200980125950.4 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09773356 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20107028267 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12999766 Country of ref document: US |
|
REEP | Request for entry into the european phase |
Ref document number: 2009773356 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009773356 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |