WO2014132816A1 - 画像処理装置および方法、並びにプログラム - Google Patents
画像処理装置および方法、並びにプログラム Download PDFInfo
- Publication number
- WO2014132816A1 WO2014132816A1 PCT/JP2014/053481 JP2014053481W WO2014132816A1 WO 2014132816 A1 WO2014132816 A1 WO 2014132816A1 JP 2014053481 W JP2014053481 W JP 2014053481W WO 2014132816 A1 WO2014132816 A1 WO 2014132816A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- unit
- additional information
- processing apparatus
- imaging
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 104
- 238000003384 imaging method Methods 0.000 claims description 136
- 230000008569 process Effects 0.000 claims description 92
- 238000000605 extraction Methods 0.000 claims description 42
- 238000006243 chemical reaction Methods 0.000 claims description 26
- 238000003672 processing method Methods 0.000 claims description 2
- 238000005516 engineering process Methods 0.000 abstract description 17
- 238000001514 detection method Methods 0.000 description 51
- 238000007906 compression Methods 0.000 description 32
- 230000006835 compression Effects 0.000 description 31
- 230000008921 facial expression Effects 0.000 description 15
- 101000610557 Homo sapiens U4/U6 small nuclear ribonucleoprotein Prp31 Proteins 0.000 description 10
- 101001109965 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) 60S ribosomal protein L7-A Proteins 0.000 description 10
- 101001109960 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) 60S ribosomal protein L7-B Proteins 0.000 description 10
- 102100040118 U4/U6 small nuclear ribonucleoprotein Prp31 Human genes 0.000 description 10
- 239000002131 composite material Substances 0.000 description 10
- 230000008929 regeneration Effects 0.000 description 7
- 238000011069 regeneration method Methods 0.000 description 7
- 230000015572 biosynthetic process Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 5
- 238000003786 synthesis reaction Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 239000012141 concentrate Substances 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000036651 mood Effects 0.000 description 2
- 230000002194 synthesizing effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 230000006837 decompression Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011946 reduction process Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000005549 size reduction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/60—Editing figures and text; Combining figures or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/175—Static expression
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/45—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/951—Computational photography systems, e.g. light-field imaging systems by using two or more images to influence resolution, frame rate or aspect ratio
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2628—Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
Definitions
- the present technology relates to an image processing apparatus, method, and program, and more particularly, to an image processing apparatus, method, and program capable of displaying an image more effectively and improving convenience.
- an imaging apparatus equipped with two cameras is known. Some of such imaging devices can capture images simultaneously with two cameras.
- the sub camera when a main camera and a sub camera capture an image and the photographer's voice is collected by a sub microphone, the sub camera can There is also one that synthesizes and records the obtained images (for example, see Patent Document 1).
- the present technology has been made in view of such a situation, and can display an image more effectively and improve convenience.
- An image processing apparatus includes a first imaging unit that images a first direction, a second imaging unit that images a second direction different from the first direction, and the first An additional information generation unit that generates additional information based on the first image obtained by imaging by the imaging unit, and the additional information in the image data of the second image obtained by imaging by the second imaging unit. And an addition processing unit for adding.
- the additional information generation unit can generate the additional information by compressing the information amount of the first image.
- the additional information generation unit can generate the additional information by compressing the size of the first image.
- the first imaging unit causes the direction of the user operating the image processing apparatus to be imaged as the first direction, and the image processing apparatus synthesizes an image obtained from the additional information with the second image.
- a synthesizing unit can be further provided.
- the first imaging unit causes the direction of the user operating the image processing device to capture an image as the first direction
- the additional information generation unit performs face recognition processing on the first image.
- the photographer information of the second image can be generated as the additional information.
- the image processing apparatus includes an extraction unit that detects the second image captured by a specific photographer from the plurality of second images based on the additional information, and the first unit detected by the extraction unit.
- a reproduction unit that displays only two images can be further provided.
- the image processing apparatus may further include a classification unit that records the second image in a recording area determined by the photographer information as the additional information.
- the image processing apparatus is provided on the surface on the user side that operates the image processing apparatus provided with the first imaging unit, and is captured by the second imaging unit at the time of imaging the second image.
- a display unit for displaying an image is further provided, and the additional information generation unit detects the direction of the user's line of sight based on the first image, so that the user on the second image is gazing. It is possible to generate information indicating the gaze area as the additional information.
- the image processing apparatus may further include a reproducing unit that controls display of the second image based on the additional information.
- the reproduction unit can display a plurality of the second images based on the additional information.
- the image processing apparatus further includes a field angle conversion unit that performs a field angle conversion process on the second image, and the additional information generation unit includes a region including a main subject on the first image.
- the cropped image obtained by cutting out can be used as the additional information, and the additional processing unit can add the additional information to the image data of the second image subjected to the angle-of-view conversion processing.
- the cropped image as the additional information and the second image that has been subjected to the angle-of-view conversion process are combined side by side to compose the first image captured by the second imaging unit.
- a synthesis unit that generates a synthesized image having the same aspect ratio as the second image can be further provided.
- the first imaging unit causes the direction of the user operating the image processing apparatus to capture an image as the first direction
- the additional information generation unit causes the user to smile based on the first image.
- the degree of smile of the user is generated as the additional information
- the second imaging unit can capture the second direction when the smile of the user is detected.
- the image processing apparatus may further include a classification unit for recording the second image in a recording area determined by the degree of smile as the additional information.
- the first imaging unit causes the direction of the user operating the image processing apparatus to capture an image as the first direction
- the additional information generation unit causes the user to smile based on the first image. It is possible to further provide a classification unit for detecting and generating the degree of smile of the user as the additional information and recording the second image in a recording area determined by the degree of smile as the additional information.
- An image processing method or program images a first direction by a first imaging unit, images a second direction different from the first direction by a second imaging unit, and Additional information is generated based on the first image obtained by imaging by the first imaging unit, and the additional information is added to the image data of the second image obtained by imaging by the second imaging unit. Includes steps.
- the first imaging unit captures a first direction
- the second imaging unit captures a second direction different from the first direction
- the first imaging unit The additional information is generated based on the first image obtained by the imaging by the second imaging unit, and the additional information is added to the image data of the second image obtained by the imaging by the second imaging unit.
- an image can be displayed more effectively and convenience can be improved.
- the present technology relates to a device having a plurality of imaging units capable of imaging different directions, such as a multi-function mobile phone and a digital camera, and when the main imaging unit captures an image, the image is captured at the same time. Information obtained from other images is generated as additional information. In the present technology, an image can be displayed more effectively by generating such additional information. Further, by embedding additional information in the image data, it is possible to perform display control with a higher degree of freedom at the time of image reproduction without requiring a troublesome operation at the time of image capture, thereby improving convenience. be able to.
- FIG. 1 is a diagram illustrating a configuration example of an external appearance of an image processing apparatus such as a multi-function mobile phone to which the present technology is applied.
- the appearance of the image processing apparatus 11 indicated by an arrow A ⁇ b> 11 indicates the appearance on the front side, that is, the side that can be seen by the user when the user operates the image processing apparatus 11.
- the appearance of the image processing apparatus 11 indicated by the arrow A12 indicates an appearance when the image processing apparatus 11 indicated by the side, that is, the arrow A11 is viewed from the left side or the right side in the drawing.
- the appearance of the image processing apparatus 11 indicated by the arrow A13 indicates the appearance of the back side, that is, the side opposite to the front side.
- a display unit 21 on which an image or the like is displayed is provided on the surface side of the image processing apparatus 11, and a touch panel for a user to perform an input operation is superimposed on the display unit 21.
- a touch panel for a user to perform an input operation is superimposed on the display unit 21.
- an in-image imaging unit 22 that captures a subject in the direction of the surface side of the image processing apparatus 11 is also provided. Therefore, in most cases, an image captured by the in-image capturing unit 22 is an image of a user who is operating the image processing apparatus 11.
- the image captured by the in-image capturing unit 22 will be referred to as an in-image, and the description will be continued assuming that the user's image is captured as the in-image.
- an out-image capturing unit 23 that captures a subject in the direction of the back side of the image processing device 11 is provided on the back surface of the image processing device 11.
- the out-image capturing unit 23 is provided so as to capture a direction opposite to the capturing direction of the in-image capturing unit 22, but the in-image capturing unit 22 and the out-image capturing unit 23 are mutually in the capturing direction. Need only be arranged differently.
- the image captured by the out-image capturing unit 23 is usually an image of a landscape or a person that the user is viewing, that is, an image of a subject that the user wants to capture.
- an image captured by the out image capturing unit 23 is referred to as an out image.
- the user U11 points the out-image capturing unit 23 of the image processing device 11 toward the target subject H11. To instruct imaging. Then, the image processing apparatus 11 captures the in-image and the out-image at the same time or almost at the same time in accordance with the user U11's imaging instruction.
- an image of the subject H11 is obtained as an out image
- an image of the user U11 is obtained as an in image
- the image processing apparatus 11 performs a process such as encoding on the in-image obtained in this way as necessary, and uses the resulting image as additional information. Then, the image processing apparatus 11 adds the obtained additional information to the image data of the out image.
- the additional information is not synthesized on the out image, but is embedded as additional information in the image data of the out image.
- the additional information is embedded as so-called Exif data, that is, metadata about the out image, with respect to the image data of the out image conforming to, for example, Exif (Exchangeable image file format).
- the user can select whether to display only the out images or to display the images obtained from the additional information combined with the out images. For example, when an image obtained from the additional information is combined with the out image and displayed, the out image shown on the right side in the figure is displayed on the display unit 21.
- the image IP11 obtained from the additional information is superimposed on the out image in which the subject H11 is shown.
- the image IP11 is an image of the face of the user U11
- the user who appreciates the out image is angry about who captured the out image
- the facial expression of the user U11 at the time of the out image capture that is, laughing. You can know if you are excited or not.
- the image obtained from the in-image is embedded in the image data as additional information of the out-image, and additional information is used as needed when reproducing the out-image, so that the photographer at the time of capturing the out-image You can check the situation.
- an out-image captured during travel can be more enjoyed as a memory when viewed later.
- the image processing apparatus 11 since the situation at the time of capturing the out image can be confirmed as an image, for example, if the out image is captured as reference information when investigating an accident, the photographer, the imaging environment, the imaging Information such as location can be easily left behind.
- FIG. 3 is a block diagram showing a more detailed configuration example of the image processing apparatus 11.
- the same reference numerals are given to the portions corresponding to those in FIG. 1, and the description thereof will be omitted as appropriate.
- 3 includes a display unit 21, an in-image capturing unit 22, an out-image capturing unit 23, an operation input unit 51, a control unit 52, a signal processing unit 53, and a recording unit 54.
- the operation input unit 51 includes, for example, a touch panel provided so as to be superimposed on the display unit 21, and supplies a signal corresponding to a user operation to the control unit 52.
- the control unit 52 controls the overall operation of the image processing apparatus 11 according to the signal supplied from the operation input unit 51. For example, the control unit 52 instructs the in-image capturing unit 22 and the out-image capturing unit 23 to capture an image, and instructs the signal processing unit 53 to reproduce and record the out-image.
- the signal processing unit 53 generates additional information based on the in-image supplied from the in-image capturing unit 22, adds the additional information to the out image supplied from the out-image capturing unit 23, and adds the additional information.
- the out image is supplied to the recording unit 54 to be recorded.
- the signal processing unit 53 reads out images from the recording unit 54, supplies them to the display unit 21, and displays them.
- the recording unit 54 records the out image supplied from the signal processing unit 53 and supplies the recorded out image to the signal processing unit 53 as necessary.
- the signal processing unit 53 in FIG. 3 is configured as shown in FIG. 4 in more detail.
- the signal processing unit 53 shown in FIG. 4 includes a compression unit 81, an addition processing unit 82, an extraction unit 83, a decoding unit 84, and a reproduction unit 85.
- the compression unit 81 generates additional information by compressing the information amount and size of the in-image supplied from the in-image capturing unit 22 and supplies the additional information to the addition processing unit 82.
- the compression unit 81 includes an information amount compression unit 91, a size compression unit 92, and an encoding unit 93.
- the information amount compression unit 91 extracts the contour of the subject on the in-image by binarizing the in-image. That is, the in-image is converted into a binary image by the binarization process, and thereby the information amount of the in-image is compressed.
- the size compression unit 92 compresses the size of the binary image by reducing the size of the binary image obtained by the information amount compression unit 91.
- the encoding unit 93 encodes the binary image size-compressed by the size compression unit 92 and uses it as additional information.
- the encoding process by the encoding unit 93 is also a process of compressing the information amount of the in-image (binary image).
- the addition processing unit 82 adds the additional information supplied from the compression unit 81 to the image data of the out image supplied from the out image capturing unit 23, and supplies the additional information to the recording unit 54.
- the extraction unit 83 reads out images from the recording unit 54 and extracts additional information from the out images.
- the extraction unit 83 supplies the out image to the reproduction unit 85 and also supplies the additional information extracted from the out image to the decoding unit 84.
- the decoding unit 84 decodes the additional information supplied from the extraction unit 83 and supplies the binary image obtained as a result to the reproduction unit 85.
- the decoding unit 84 includes an expansion unit 94, and the expansion unit 94 expands the size of the binary image as necessary.
- the reproduction unit 85 supplies the out image supplied from the extraction unit 83 to the display unit 21 for display.
- the reproduction unit 85 includes a combining unit 95, and the combining unit 95 combines the binary image supplied from the decoding unit 84 with the out image. Therefore, the reproducing unit 85 can cause the display unit 21 to display an out image obtained by combining the binary images as necessary.
- step S ⁇ b> 11 the out-image capturing unit 23 captures an out-image according to an instruction from the control unit 52 and supplies it to the additional processing unit 82.
- a landscape image or the like is captured as an out image.
- step S ⁇ b> 12 the in-image capturing unit 22 captures an in-image according to an instruction from the control unit 52 and supplies the in-image to the compression unit 81.
- an image of the face of the user who is operating the image processing apparatus 11 is captured as an in-image.
- the information amount compression unit 91 converts the in-image supplied from the in-image capturing unit 22 into a binary image.
- the information amount compression unit 91 converts the in image, which is a color image, into a monochrome image with 256 gradations, and binarizes the monochrome image to generate a binary image.
- a user's face image obtained by capturing an out image is obtained as a binary image. Since this binary image is an image of the user's face when the out image is captured, the binary image is an image that allows the user to know the user's facial expression, mood, and the like when the out image is captured.
- step S14 the size compression unit 92 compresses the size of the binary image obtained by the information amount compression unit 91. For example, a binary image having the same size as the out image is reduced to an image having a size of 160 pixels ⁇ 120 pixels, that is, a thumbnail size.
- step S15 the encoding unit 93 encodes the binary image size-compressed by the size compression unit 92 and sets it as additional information.
- the encoding unit 93 encodes the binary image RP11 by the run-length method, and uses the run-length code obtained as a result as additional information.
- Each square of the binary image RP11 represents one pixel.
- a white square represents a pixel having a pixel value “1”
- a black square represents a square having a pixel value “0”.
- FIG. 6 coordinates indicating the position in the horizontal direction in the diagram of the binary image RP11 are shown.
- the coordinates of the pixel located at the left end in the diagram of the binary image RP11 are “1”.
- the coordinate of the pixel located at the right end is “11”. That is, the coordinates of each pixel indicate how many pixels are located from the left end of the binary image RP11.
- the binary image RP11 is encoded for each pixel column arranged in the horizontal direction in the drawing.
- a pixel column to be coded is represented by a run-length code composed of the start point position of a pixel column composed of pixels having pixel values “1” and the length of the pixel column.
- the pixel whose coordinates are “1” has the pixel value “1”.
- the pixel with the coordinate “2” adjacent to the pixel value is “0”. Therefore, when a pixel column having a pixel value “1” starting from a pixel having coordinates “1” is expressed by “start point, length” of the pixel column, “1, 1” is obtained.
- a pixel row composed of pixels having a pixel value “1” starting from a pixel with coordinates “4” is expressed as “4, 3” as “start point, length”.
- a pixel row composed of pixels having a pixel value “1” starting from a pixel whose coordinates are “10” is represented by “10, 2”.
- the encoding unit 93 obtains codes “1, 1, 4, 3, 10, 2” obtained by arranging these three “starting point and length” codes.
- the run length code of the pixel row of interest is used.
- the encoding unit 93 uses the run-length code obtained for each pixel column of the binary image RP11 in this way as additional information of the binary image RP11.
- the compression unit 81 supplies the additional information obtained by the encoding by the encoding unit 93 to the additional processing unit 82.
- step S ⁇ b> 16 the addition processing unit 82 adds the additional information supplied from the compression unit 81 to the image data of the out image supplied from the out image capturing unit 23, and the recording unit. 54. That is, the additional information is embedded in the image data of the out image as metadata of the out image.
- step S17 the recording unit 54 records the out-image supplied from the addition processing unit 82, and the imaging process ends.
- the image processing apparatus 11 compresses the information of the in-image captured at the same time as the out-image to obtain additional information, and adds the additional information to the out-image.
- additional information it is possible to display the image more effectively using the additional information at the time of reproducing the out image, and it is possible to improve convenience.
- the image processing apparatus 11 since additional information is added to the out image, it is possible to specify whether to display only the out image during playback or to display the out image using the additional information. Therefore, the out image can be displayed more effectively and the convenience can be improved.
- the user does not need to perform an operation of instructing the synthesis of the image obtained from the in-image to the out-image when capturing the out-image, and can concentrate on the imaging work. Thereby, the convenience for the user can be further improved.
- the user can operate the operation input unit 51 to reproduce the recorded out image. At that time, for example, the user can select whether to display only the out image or to display the image obtained from the additional information on the out image by combining the operation input unit 51.
- step S41 the extraction unit 83 reads out images specified by the user from the recording unit 54 in accordance with instructions from the control unit 52.
- step S42 the extraction unit 83 supplies the read out image to the reproduction unit 85, reads additional information from the image data of the out image, and supplies the additional information to the decoding unit 84.
- step S43 the decoding unit 84 decodes the additional information supplied from the extraction unit 83. For example, when the additional information is a run length code, the decoding unit 84 restores a binary image from the run length code.
- step S44 the decompression unit 94 decompresses the size of the binary image obtained by decoding. For example, a 160 ⁇ 120 pixel thumbnail-sized binary image is enlarged to an appropriate-sized binary image.
- the decoding unit 84 supplies the binary image expanded by the expansion unit 94 to the reproduction unit 85 as an image obtained from the additional information.
- step S45 the reproducing unit 85 determines whether or not to reproduce the binary image.
- a signal to that effect is sent from the control unit 52 to the signal processing unit 53.
- the reproduction unit 85 determines to reproduce the binary image when the signal supplied from the control unit 52 to the signal processing unit 53 is a signal for displaying the binary image combined with the out image.
- step S46 the synthesis unit 95 synthesizes the binary image supplied from the decoding unit 84 with the out image supplied from the extraction unit 83.
- the position on the out image where the binary image is combined may be a predetermined position or a position designated by the user, or a background without a subject (foreground) specified by image recognition or the like. It may be the position of the region.
- the reproduction unit 85 supplies the out image obtained by combining the binary images to the display unit 21 and controls the display of the out image.
- step S47 the display unit 21 displays the out image obtained by combining the binary images supplied from the reproduction unit 85, and the reproduction process ends.
- step S45 If it is determined in step S45 that the binary image is not reproduced, the reproducing unit 85 supplies the out image supplied from the extracting unit 83 to the display unit 21 as it is, and the process proceeds to step S48. In this case, an out image in which a binary image is not synthesized is supplied to the display unit 21.
- step S48 the display unit 21 displays the out image supplied from the playback unit 85, and the playback process ends.
- the image processing apparatus 11 displays the out image as it is in accordance with a user instruction, or causes the binary image obtained from the additional information added to the out image to be combined with the out image and displayed.
- the additional information added to the out image in this way, it is possible to display only the out image according to the mood of the user at that time, or to display the out image in which the binary image is synthesized. it can. Thereby, the out image can be displayed more effectively and the convenience can be improved.
- a user who views the out image can know the facial expression of the user when the out image is captured, and can more enjoy viewing the out image. .
- in-image information As an example of compressing in-image information, the case of performing in-monochrome monochromeization, binarization, size reduction, and encoding has been described. However, any of these compression processes is described. May be combined as appropriate. For example, only the in-image may be converted into additional information by performing monochrome conversion, or the in-image may be converted into additional information by performing monochrome conversion and encoding. Further, compression may be performed on a partial area of the in-image, and the resulting image or code (code) may be used as additional information.
- the signal processing unit 53 is configured as shown in FIG. 8, for example.
- parts corresponding to those in FIG. 4 are denoted by the same reference numerals, and description thereof is omitted as appropriate.
- the signal processing unit 53 in FIG. 8 has a configuration in which a face recognition unit 121 is further provided in the signal processing unit 53 in FIG.
- the face recognizing unit 121 performs face recognition on the in-image supplied from the in-image capturing unit 22 using the registration information recorded in advance in the recording unit 54, and uses the recognition result as additional information to the additional processing unit 82. Supply.
- the feature amount of the user's face image registered in advance, the user's face discriminator, and the like and the user name of the user are associated and recorded as registration information.
- the face recognition unit 121 uses the registration information and the in-image to identify the user on the in-image by face recognition, and uses the identification result (face recognition result) as additional information.
- the addition processing unit 82 adds the face recognition result supplied from the face recognition unit 121 and the binary image supplied from the compression unit 81 as additional information to the out image from the out image capturing unit 23, and the recording unit 54.
- or step S75 is the same as the process of FIG.5 S11 thru
- step S ⁇ b> 72 the in-image obtained by imaging is supplied from the in-image imaging unit 22 to the compression unit 81 and the face recognition unit 121.
- step S ⁇ b> 76 the face recognition unit 121 performs face recognition on the in-image supplied from the in-image imaging unit 22 using the registration information recorded in the recording unit 54, and adds the recognition result as additional information.
- the user name of the user identified as a result of face recognition that is, the photographer information is supplied to the additional processing unit 82 as additional information.
- step S77 the addition processing unit 82 uses the photographer information supplied from the face recognition unit 121 and the encoded binary image supplied from the compression unit 81 as additional information from the out-image capturing unit 23. And is supplied to the recording unit 54.
- step S78 the recording unit 54 records the out-image supplied from the addition processing unit 82, and the imaging process ends.
- the image processing apparatus 11 adds the binary image obtained from the in-image and the photographer information as additional information to the out-image.
- out images can be displayed more effectively and the convenience can be improved.
- out images can be effectively displayed, such as displaying a list of out images captured by a specific photographer.
- the user can operate the operation input unit 51 to display a list of out images captured by a specific user. At that time, for example, the user can select whether to display only the out image or to display the image obtained from the additional information on the out image by combining the operation input unit 51.
- step S101 the extraction unit 83 reads all the out images recorded in the recording unit 54.
- step S102 the extraction unit 83 reads photographer information as additional information added to each read out image, and detects an out image captured by a specific photographer.
- a specific photographer is a photographer input by the user operating the operation input unit 51.
- the control unit 52 reads registration information from the recording unit 54 through the signal processing unit 53 as necessary, and displays information indicating the photographer designated by the user as a signal processing unit. 53. Then, based on the information supplied from the control unit 52, the extraction unit 83 detects an out image captured by a specific photographer.
- step S103 the extraction unit 83 supplies only the detected out-image of the specific photographer to the reproduction unit 85 and encodes 2 as additional information from the image data of the out-image of the specific photographer.
- the value image is read out and supplied to the decoding unit 84.
- a run-length code is supplied to the decoding unit 84 as an encoded binary image.
- step S104 to step S107 is performed thereafter. Since these processing are the same as the processing from step S43 to step S46 in FIG. .
- steps S104 to S107 for all out images taken by a specific photographer, encoded binary images as additional information read from the out images are decoded, and 2 for the out images.
- the value image is synthesized.
- step S108 the display unit 21 displays a list of out images obtained by combining the binary images supplied from the reproduction unit 85, and the reproduction process ends.
- the out image SM11-1 to the out image SM11-6 shown in FIG. In this example, six out images SM11-1 to SM11-6 are displayed side by side on the display unit 21. Then, the binary images IP21-1 to IP21-6 as additional information are combined and displayed on the out images SM11-1 to SM11-6.
- out images SM11-1 to SM11-6 are displayed as a list
- the user operates the operation input unit 51 to display only one of the out images on the display unit 21, These out images can be displayed as a slide show.
- a user's face image is recorded in advance as registration information, and the face image is synthesized and displayed on the out image. May be.
- step S106 when it is determined in step S106 that the binary image is not to be reproduced, the reproducing unit 85 supplies the out image supplied from the extracting unit 83 to the display unit 21 as it is, and the processing is performed. The process proceeds to step S109.
- step S109 the display unit 21 displays a list of the out images supplied from the reproduction unit 85, and the reproduction process ends.
- out images SM11-1 to SM11-6 are displayed as a list
- the user operates the operation input unit 51 to display only one of the out images on the display unit 21, These out images can be displayed as a slide show.
- the image processing apparatus 11 uses the photographer information added as additional information to the out image, detects the out image of the specific photographer, and displays the detected out images as a list. In addition, when displaying the out images as a list, the image processing apparatus 11 synthesizes and displays a binary image on the out images in accordance with a user instruction.
- the photographer information as additional information added to the out image, it is possible to display the out image more effectively, such as displaying only the out image of a specific photographer. Can be improved.
- encoded binary images as additional information when out images are displayed as a list, only out images are displayed, or out images obtained by combining binary images are displayed. Can do.
- ⁇ Third Embodiment> ⁇ Configuration example of signal processor>
- a case where a binary image encoded as additional information and photographer information is generated has been described.
- information indicating a gaze area on an out-image may be generated as additional information.
- the gaze region is a region on the out image that the user who is the photographer is gazing at the time of capturing the out image.
- the signal processing unit 53 is configured as shown in FIG. 13, for example.
- the same reference numerals are given to the portions corresponding to those in FIG. 4, and description thereof will be omitted as appropriate.
- the signal processing unit 53 in FIG. 13 includes a gaze area detection unit 151, an additional processing unit 82, an extraction unit 83, a cost calculation unit 152, and a reproduction unit 85. That is, the signal processing unit 53 in FIG. 13 has a configuration in which a gaze area detection unit 151 and a cost calculation unit 152 are provided instead of the compression unit 81 and the decoding unit 84 in the signal processing unit 53 in FIG.
- the gaze area detection unit 151 detects a gaze area that is an area on the out image that the user is gazing based on the in-image supplied from the in-image imaging unit 22, and performs an additional process using the detection result as additional information. To the unit 82.
- the addition processing unit 82 adds information indicating the detection result of the gaze area supplied from the gaze area detection unit 151 as additional information to the image data of the out image supplied from the out image imaging unit 23, and adds the information to the recording unit 54. Supply.
- the cost calculation unit 152 gazes the area when trying to display a plurality of out images simultaneously or sequentially. Is calculated as a cost and supplied to the reproducing unit 85.
- the reproduction unit 85 controls the display of the out image by the display unit 21 based on the cost supplied from the cost calculation unit 152, and displays a plurality of out images in a superimposed manner so that the gaze areas of the respective out images do not overlap as much as possible.
- step S131 and step S132 is the same as the process of step S11 of FIG. 5, and step S12, the description is abbreviate
- step S131 when the out image is captured, the image captured by the out image capturing unit 23 is supplied as a preview image of the out image to the display unit 21 via the reproduction unit 85 of the signal processing unit 53 and displayed.
- the user operates the operation input unit 51 in a state where the preview image of the out image is displayed on the display unit 21 and instructs to capture the out image.
- step S ⁇ b> 133 the gaze area detection unit 151 detects the user's gaze direction at the time of imaging the out image based on the in image supplied from the in image imaging unit 22, so that the user's gaze area on the out image is detected. Is detected.
- the gaze area detection unit 151 detects a human face area from the in-image, and determines an eye area on the detected face area. Then, the gaze area detection unit 151 obtains the user's line-of-sight direction by performing pattern matching using the determined eye area and a learning data group prepared in advance. For example, in the pattern matching, the line-of-sight direction associated with the learning data having the highest score obtained by matching with the eye region is set as the line-of-sight direction to be obtained.
- the gaze area detection unit 151 determines the relative positional relationship between the actual user face position and the in-image imaging unit 22 and the display unit 21 based on the user face position and the face size on the in-image.
- the gaze area is detected from the positional relationship and the line-of-sight direction. That is, since the preview image of the out image is displayed on the display unit 21 when the in-image is captured, the region of the out image displayed in the region on the display unit 21 that the user is gazing at the time of capturing the in-image is displayed. , It should be the area that the user is paying attention to.
- the gaze area detection unit 151 detects the gaze area on the out image based on the in-image, as a detection result, for example, the coordinates (rectangular coordinates) of the rectangular area on the out image indicating the gaze area are added as additional information. To the unit 82.
- step S134 the addition processing unit 82 adds the additional information supplied from the gaze area detection unit 151 to the image data of the out image supplied from the out image capturing unit 23, and supplies the additional information to the recording unit 54.
- step S135 the recording unit 54 records the out-image supplied from the addition processing unit 82, and the imaging process ends.
- the image processing apparatus 11 uses the information indicating the gaze area of the out image detected based on the in image captured simultaneously with the out image as additional information, and adds the additional information to the out image.
- the out image by adding additional information to the out image, it is possible to display the image more effectively by using the additional information at the time of reproducing the out image, and it is possible to improve convenience. For example, when a plurality of out images are displayed simultaneously, it is possible to display the gaze regions on the out images so as not to overlap as much as possible.
- the information indicating the gaze area generated as the additional information is information that can be obtained only when the out image is captured, and the out image can be effectively presented if used when the out image is displayed.
- the user can operate the operation input unit 51 to display the out image as a slide show. At that time, for example, the user can select whether to display the out images one by one or to superimpose a plurality of out images simultaneously by operating the operation input unit 51.
- step S161 and step S162 are the same as the process of step S41 of FIG. 7, and step S42, the description is abbreviate
- step S161 for example, a plurality of out images such as out images in a folder designated by the user are read.
- step S ⁇ b> 162 information indicating the gaze area as additional information read from each of the plurality of out images is supplied from the extraction unit 83 to the cost calculation unit 152.
- step S ⁇ b> 163 the cost calculation unit 152 calculates the overlapping degree of the subject as a cost based on the information indicating the gaze area as the additional information supplied from the extraction unit 83, and supplies it to the reproduction unit 85.
- the cost calculation unit 152 obtains the importance of each area of the out-image with the gaze area as a saliency (important point). For example, the importance of the gaze area is increased, and the importance of the area other than the gaze area is decreased. At this time, the closer to the center of the gaze area, the higher the importance may be.
- the cost calculation unit 152 calculates the cost of the overlapping degree of each subject for how to superimpose a plurality of out images displayed at the same time. 85. For example, in a case where a region with higher importance is overlapped with other out-images and cannot be seen, cost calculation is performed so that the cost becomes higher.
- step S164 the reproduction unit 85 determines whether or not to display a plurality of out images in a superimposed manner.
- step S165 the synthesis unit 95 performs a plurality of outs supplied from the extraction unit 83 based on the cost supplied from the cost calculation unit 152. Combine images.
- the synthesizing unit 95 synthesizes a plurality of out images in a superimposing manner determined by the lowest cost supplied from the cost calculating unit 152. Thereby, a plurality of out images are superimposed so that main subjects of each out image do not overlap as much as possible, that is, the gaze area is not hidden behind other out images.
- the playback unit 85 supplies the plurality of synthesized out images to the display unit 21 and instructs the slide unit display of the plurality of out images.
- step S166 the display unit 21 displays a plurality of synthesized out images supplied from the reproduction unit 85, thereby displaying the out images as a slide show in a state where the plurality of out images are simultaneously displayed, and reproduction processing. Ends.
- an out image shown in FIG. 16 is displayed on the display unit 21.
- three out images P11-1 to P11-3 are displayed on the display unit 21, and some of these out images overlap each other.
- three out images are superimposed so that the main subject of each out image is not hidden behind other out images.
- the out images may be displayed one by one in order, or after a plurality of out images are superimposed and displayed at the same time. Next, a plurality of different out images may be displayed simultaneously.
- step S ⁇ b> 164 when it is determined in step S ⁇ b> 164 that the out image is not superimposed and displayed, the reproducing unit 85 sequentially supplies the out image supplied from the extraction unit 83 to the display unit 21. The process proceeds to step S167.
- step S167 the display unit 21 displays the out images supplied from the reproduction unit 85 in order to perform a slide show display of the out images, and the reproduction process ends.
- the image processing apparatus 11 superimposes a plurality of out images on the gaze area so that other out images do not overlap as much as possible based on the information indicating the gaze area added as additional information to the out image. indicate.
- the out image can be displayed more effectively by using the additional information that can be obtained only when the out image is captured.
- the information indicating the gaze area as the additional information is used for displaying the out image in an overlapping manner.
- the information indicating the gaze area is used for another display method of the out image. You may be made to do. For example, when the out images are sequentially displayed one by one, the gaze area of the out image may be enlarged (zoomed) based on the information indicating the gaze area.
- ⁇ Fourth embodiment> ⁇ Configuration example of signal processor> Further, a partial area of the in-image may be used as additional information, and an image that is used as additional information when the out-image is reproduced may be combined with the out-image.
- the angle-of-view conversion process is performed on the out image OT11, and the image T11 obtained as a result is the final out image.
- the image T11 is an image obtained by converting the aspect ratio of the out image OT11, and all the subjects on the out image OT11 are shown in the image T11. That is, the image T11 is an image obtained by reducing the out image OT11 in the horizontal direction in the drawing.
- processing such as face recognition is performed on the in-image IN11, a main subject region is detected from the in-image IN11, and a region including the main subject region of the in-image IN11 is cut out to obtain additional information.
- image C11 In particular, when the main subject is a human face, if the image transformation process is performed on the main subject area, the human face will be distorted, but here the main subject area is cut out. An image without distortion can be obtained.
- out-side aspect parameter and in-side aspect parameter are such that when the image T11 and the image C11 are arranged in the horizontal direction in the drawing to form one image, the aspect ratio of the obtained image is that of the original out image OT11. It is determined to be the same as the aspect ratio.
- the image C11 is added to the image data of the image T11 as additional information.
- the user when reproducing the image T11, the user can select whether to reproduce the image T11 as it is or to display the image T11 with the image C11 as additional information combined.
- the image processing device 11 when display of an image obtained by combining the image C11 with the image T11 is instructed, the image processing device 11 combines the image T11 and the image C11 in the horizontal direction in the drawing to generate a combined image P21. Then, the image processing device 11 displays the obtained composite image P21.
- the composite image P21 is an image including the state of the photographer at the time of capturing the out image OT11 while leaving all the contents of the original out image OT11, and displays the image more effectively. can do.
- the signal processing unit 53 is configured as shown in FIG. 18, for example.
- parts corresponding to those in FIG. 4 are denoted by the same reference numerals, and description thereof will be omitted as appropriate.
- 18 includes an angle-of-view conversion unit 181, an addition processing unit 82, a main area detection unit 182, a crop processing unit 183, an extraction unit 83, and a reproduction unit 85.
- the angle-of-view conversion unit 181 performs angle-of-view conversion processing on the out image supplied from the out image capturing unit 23 so that the out image has a target aspect ratio based on the designated out-side aspect parameter. And supplied to the additional processing unit 82.
- the main area detection unit 182 detects a main subject area (hereinafter referred to as a main area) from the in-image supplied from the in-image imaging unit 22 and supplies the detection result and the in-image to the crop processing unit 183. To do.
- a main area a main subject area
- the crop processing unit 183 cuts out an area including the main area of the in-image based on the detection result of the main area supplied from the main area detection unit 182 and the specified in-side aspect parameter, and adds the crop image as a crop image. This is supplied to the processing unit 82. In other words, the crop image obtained by the crop processing for the in-image is supplied to the additional processing unit 82 as additional information.
- the addition processing unit 82 adds the crop image supplied from the crop processing unit 183 to the out image supplied from the view angle conversion unit 181 as additional information, and supplies the additional information to the recording unit 54.
- the extraction unit 83 extracts additional information from the out image read from the recording unit 54, and supplies the out image and the additional information to the reproduction unit 85.
- step S191 and step S192 is the same as the process of step S11 and step S12 of FIG. 5, the description is abbreviate
- step S193 the angle-of-view conversion unit 181 performs angle-of-view conversion processing on the out image supplied from the out-image capturing unit 23 based on the designated out-side aspect parameter, and supplies it to the addition processing unit 82. .
- the angle of view conversion unit 181 converts the angle of view of the out image by seam carving processing.
- the seam carving process until the aspect ratio of the out image reaches the aspect ratio indicated by the out-side aspect parameter, a vertical path that minimizes the cost is obtained, and the process of thinning out pixels on the path is repeated. Done. That is, pixel thinning is performed so that the change in the pattern from the original image is minimized.
- an image T11 shown in FIG. 17 is obtained as a final out image and supplied to the additional processing unit 82.
- the image reduction process may be performed as the angle-of-view conversion process.
- step S194 the main area detection unit 182 detects the main area from the in-image supplied from the in-image capturing unit 22, and supplies the detection result and the in-image to the crop processing unit 183.
- face recognition processing is performed on the in-image, and a human face area in the in-image is set as a main area.
- step S195 the crop processing unit 183 applies the in-image supplied from the main region detection unit 182 to the in-image supplied from the main region detection unit 182 based on the detection result of the main region supplied from the main region detection unit 182 and the designated in-side aspect parameter. Crop processing is performed on the image.
- the crop processing unit 183 includes the main region in the in-image and cuts out the region having the aspect ratio indicated by the in-side aspect parameter as a crop image and supplies the crop image to the addition processing unit 82.
- the image C11 shown in FIG. 17 is obtained as a crop image.
- the crop image obtained in this way is used as additional information of the out image.
- step S196 the addition processing unit 82 adds the crop image supplied from the crop processing unit 183 as additional information to the image data of the out image supplied from the angle-of-view conversion unit 181 and supplies the additional information to the recording unit 54.
- step S197 the recording unit 54 records the out-image supplied from the addition processing unit 82, and the imaging process ends.
- the image processing apparatus 11 uses the crop image obtained by cutting out the region including the main region of the in-image as additional information and adds it to the out-image.
- the image processing apparatus 11 uses the crop image obtained by cutting out the region including the main region of the in-image as additional information and adds it to the out-image.
- the crop image can be displayed more effectively and the convenience can be improved.
- the user can display the out image by operating the operation input unit 51.
- the user can select, for example, by operating the operation input unit 51 to display only the out image or to display the out image combined with the crop image as additional information.
- step S221 the extraction unit 83 reads out-images designated by the user from the recording unit 54 in accordance with instructions from the control unit 52.
- step S222 the extraction unit 83 reads a crop image as additional information from the image data of the out image, and supplies the crop image and the out image to the reproduction unit 85.
- step S223 the reproducing unit 85 determines whether to combine the cropped image with the out image.
- the combining unit 95 combines the crop image with the out image supplied from the extraction unit 83 in step S224. Specifically, the composition unit 95 arranges the out image and the crop image, and generates a single composite image by performing weighted addition processing of the out image and the crop image for an area near the boundary between the images. To do. Thereby, for example, a composite image P21 shown in FIG. 17 is obtained.
- the weight used in the weighted addition process is determined so that the weight of the out image increases, for example, as the position is on the out image side.
- the size of the out image and the cropped image is increased by an amount necessary for the weighted addition process, and the finally obtained composite image is obtained.
- the aspect ratio is the same as the original out image.
- the reproduction unit 85 supplies the composite image to the display unit 21.
- step S225 the display unit 21 displays the composite image supplied from the playback unit 85, and the playback process ends.
- step S223 If it is determined in step S223 that the crop image is not synthesized, the playback unit 85 supplies the out image supplied from the extraction unit 83 to the display unit 21 as it is, and the process proceeds to step S226.
- step S226 the display unit 21 displays the out image supplied from the playback unit 85, and the playback process ends.
- the image processing apparatus 11 displays the out image as it is, or combines and displays the crop image added as additional information to the out image in accordance with a user instruction.
- a desired image of the out image or the composite image can be displayed. That is, the out image can be displayed more effectively, and the out image can be more enjoyed.
- the recording unit 54 is provided with a folder FD11 for the photographer A and a folder FD12 for the photographer B in advance.
- the photographer information added as additional information to the out image P41 and the out image P44 is information indicating the photographer A, and the photographer information added as additional information to the out image P42 and the out image P43 is photographed. It is assumed that the information indicates the person B. In such a case, the out image P41 and the out image P44 are recorded in the folder FD11, and the out image P42 and the out image P43 are recorded in the folder FD12.
- the signal processing unit 53 is configured as shown in FIG. 22, for example.
- the same reference numerals are given to the portions corresponding to those in FIG. 8, and description thereof will be omitted as appropriate.
- the signal processing unit 53 in FIG. 22 has a configuration in which a classification unit 211 is further provided in the signal processing unit 53 in FIG.
- the classification unit 211 supplies the out image from the addition processing unit 82 to the recording unit 54, and out of the folders in the recording unit 54, the out image is stored in a folder determined by photographer information as additional information added to the out image.
- the out image is recorded so that is stored.
- step S257 since the process of step S251 thru
- step S257 the image data of the out image to which the photographer information and the encoded binary image are added as additional information is supplied from the addition processing unit 82 to the classification unit 211.
- step S258 the classification unit 211 causes the recording unit 54 to record the out image according to the additional information of the out image supplied from the addition processing unit 82, and the imaging process ends.
- each folder provided in the recording unit 54 is associated with photographer information, and the classification unit 211 associates the same photographer information with the photographer information added to the out image. Identify the folder that is being used. Then, the classification unit 211 supplies the out image to the recording unit 54 so that the out image is stored in the specified folder, and records it. That is, the out image is recorded in the recording area of the recording unit 54 that is the area of the specified folder.
- the image processing apparatus 11 adds the binary image obtained from the in-image and the photographer information as additional information to the out-image, and records the out-image in a folder determined by the photographer information.
- out images can be effectively displayed, such as displaying a list of out images captured by a specific photographer.
- out image is recorded in a folder for each photographer in accordance with the photographer information, so that the out image of a specific photographer can be quickly identified.
- the user can operate the operation input unit 51 to display a list of out images captured by a specific user. At that time, for example, the user can select whether to display only the out image or to display the image obtained from the additional information on the out image by combining the operation input unit 51.
- step S281 the extraction unit 83 reads all the out images recorded in the specific folder of the recording unit 54.
- the extraction unit 83 reads out images stored in a folder of an arbitrary user (photographer) designated by the user through an operation on the operation input unit 51.
- the extraction unit 83 may read out images stored in a folder of a user (photographer) who is logged in to the image processing apparatus 11. Further, the in-image capturing unit 22 captures an in-image, and the face recognition unit 121 performs face recognition, and stores it in the folder of the user identified as a result of the face recognition, that is, the user operating the image processing apparatus 11. The out image that has been displayed may be read out.
- step S282 to step S288 When the out image is read out in this way, the processing from step S282 to step S288 is performed thereafter, and the reproduction processing ends. However, these processing are the same as the processing from step S103 to step S109 in FIG. The description is omitted.
- the image processing apparatus 11 reads out images from the folder provided for each photographer indicated by the photographer information added as additional information to the out image, and displays the out images as a list. In addition, when displaying the out images as a list, the image processing apparatus 11 synthesizes and displays a binary image on the out images in accordance with a user instruction.
- the photographer information as additional information added to the out image, it is possible to display the out image more effectively, such as displaying only the out image of a specific photographer. Further, if the out image is stored in a folder determined by the photographer information as additional information, the out image of the specific photographer can be identified more easily and quickly, and convenience can be improved.
- ⁇ Sixth embodiment> ⁇ Configuration example of signal processor> Furthermore, the smile of the user who operates the image processing apparatus 11 is detected, and the out image and the in image are captured at the timing when the smile is detected, and the degree of smile of the user is added to the out image as additional information. May be.
- the signal processing unit 53 is configured as shown in FIG. 25, for example. Note that, in FIG. 25, the same reference numerals are given to the portions corresponding to those in FIG. 8, and description thereof will be omitted as appropriate.
- the signal processing unit 53 in FIG. 25 has a configuration in which a smile detection unit 241 is further provided in the signal processing unit 53 in FIG.
- the smile detection unit 241 detects the smile of the user who operates the image processing apparatus 11 from the in-image supplied from the in-image imaging unit 22, that is, the user who is the photographer of the out-image, and the imaging timing is determined according to the detection result. Information to that effect is supplied to the control unit 52.
- the control unit 52 controls the out-image capturing unit 23 and the in-image capturing unit 22 in accordance with the information indicating that the imaging timing has been supplied from the smile detection unit 241, and captures the out-image and the in-image.
- the smile detection unit 241 generates the degree of smile of the user detected from the in-image as additional information and supplies it to the addition processing unit 82.
- step S311 the in-image capturing unit 22 captures an in-image in accordance with an instruction from the control unit 52 and supplies the in-image to the compression unit 81, the face recognition unit 121, and the smile detection unit 241.
- step S312 the smile detection unit 241 detects the smile of the user from the in-image supplied from the in-image capturing unit 22.
- the smile detection unit 241 holds learning data for a number of predetermined smiles.
- the learning data is image data of a person's face with a facial expression such as a smile or a big laughter regardless of a specific smile.
- the smile detection unit 241 calculates the score value by comparing the feature value obtained from each learning data with the feature value obtained from the in-image, thereby obtaining the degree of smile of the user on the in-image. At this time, the degree of similarity with the in-image is set as the degree of smile of the user on the in-image. Further, for example, when the degree of smile is greater than or equal to a specific value, it is assumed that a smile is detected from the in-image.
- the smile detection method may be any method as long as a smile can be detected from the in-image.
- the smile detection unit 241 may hold learning data for each smile level for a plurality of smile levels set in advance.
- the learning data is an average smile of a person having a specific smile level, that is, image data of an average face of a person with a specific expression such as a smile or a big laugh.
- the smile detection unit 241 compares the feature amount obtained from each learning data with the feature amount obtained from the in-image, and identifies learning data most similar to the in-image from among the learning data. The degree of smile of the user on the in-image is obtained. At this time, if there is learning data in which the degree of similarity with the in-image is equal to or greater than a specific value, a smile is detected from the in-image, and the degree of smile of the learning data with the highest degree of similarity is determined on the in-image. The degree of smile of the user.
- step S313 the smile detection unit 241 determines whether a smile is detected from the in-image.
- step S313 If it is determined in step S313 that no smile has been detected, the process returns to step S311 and the above-described process is repeated.
- step S313 when it is determined in step S313 that a smile has been detected, the smile detection unit 241 supplies information indicating that the imaging timing has come to the control unit 52, and the process proceeds to step S314.
- the smile detection unit 241 supplies the degree of smile of the user on the in-image obtained by the smile detection to the addition processing unit 82 as additional information of the out image.
- the smile level which is the additional information of the out image, is information indicating the smile (expression) of the photographer who captures the out image.
- the smile detection unit 241 may calculate the degree of smile using the shape of the mouth of the user's face detected from the in-image, the size of the tooth region, or the like.
- step S314 to step S319 When a smile is detected from the in-image, the processing from step S314 to step S319 is performed thereafter. Since these processing are the same as the processing from step S71 to step S76 in FIG. 9, the description thereof is omitted. However, in step S314 and step S315, the out image and the in image are transmitted to the out image capturing unit 23 and the in image capturing unit 22 based on the information indicating that the control unit 52 has reached the image capturing timing supplied from the smile detecting unit 241. To image. That is, the out image and the in image are captured at the timing when the photographer's smile is detected.
- step S320 the additional processing unit 82 uses the photographer information from the face recognition unit 121, the encoded binary image from the compression unit 81, and the degree of smile from the smile detection unit 241 as additional information, as an out image. This is added to the image data of the out image from the imaging unit 23. Then, the addition processing unit 82 supplies the image data of the out image added with the additional information to the recording unit 54.
- step S321 the recording unit 54 records the out image supplied from the addition processing unit 82, and the imaging process ends.
- the image processing apparatus 11 adds the binary image obtained from the in-image, the photographer information, and the degree of smile to the out-image as additional information.
- the out image can be displayed more effectively and the convenience can be improved.
- the degree of smile is used as additional information, it is possible to selectively display an out image in which the photographer's facial expression is a specific facial expression, that is, a specific degree of smile at the time of imaging.
- the smile level is used as additional information, an out-image having a specific smile level can be detected easily and quickly, and convenience can be improved.
- the user can operate the operation input unit 51 to display the out images having a specific smile level as a slide show or as a list. At that time, for example, the user can select whether to display only the out image or to display the image obtained from the additional information on the out image by combining the operation input unit 51.
- step S351 the extraction unit 83 reads all the out images recorded in the recording unit 54.
- step S352 the extraction unit 83 reads the degree of smile as additional information added to each read out image, and detects an out image having a specific smile level.
- the specific smile level is designated in advance by the user.
- step S353 When an out-image having a specific smile level is detected, the processing from step S353 to step S357 is performed thereafter, which is the same as the processing from step S103 to step S107 in FIG. Is omitted. However, in step S353, the additional information is read from the out image having a specific smile level, and the out images are supplied to the reproduction unit 85. In addition, the reproduction unit 85 supplies the out image obtained by combining the binary images to the display unit 21 in order.
- step S358 the display unit 21 sequentially displays the out images obtained by combining the binary images supplied from the reproduction unit 85, and the reproduction process ends.
- step S356 when it is determined in step S356 that the binary image is not reproduced, the reproducing unit 85 supplies the out image supplied from the extracting unit 83 to the display unit 21 as it is, and the process proceeds to step S359.
- step S359 the display unit 21 sequentially displays the out images supplied from the reproduction unit 85, and the reproduction process ends.
- the image processing apparatus 11 detects and displays an out image having a specific smile level by using the smile level added as additional information to the out image. Further, when displaying the out image, the image processing apparatus 11 synthesizes and displays the binary image on the out image in accordance with a user instruction.
- the out image can be displayed more effectively, such as displaying only the out image captured when the photographer has a specific facial expression. Can be displayed.
- the recording unit 54 is provided with a folder FD21 with a smile degree A and a folder FD22 with a smile degree B in advance.
- the smile level added as additional information to the out image P51 and the out image P53 is information indicating the smile level B
- the smile level added as additional information to the out image P52 is the smile level A. It is assumed that the information is information. In such a case, the out image P52 is recorded in the folder FD21, and the out image P51 and the out image P53 are recorded in the folder FD22.
- the signal processing unit 53 is configured as shown in FIG. 29, for example.
- the same reference numerals are given to the portions corresponding to those in FIG. 25, and the description thereof will be omitted as appropriate.
- the signal processing unit 53 in FIG. 29 has a configuration in which a classification unit 271 is further provided in the signal processing unit 53 in FIG.
- the classification unit 271 supplies the out image from the addition processing unit 82 to the recording unit 54, and out of the folders in the recording unit 54, the out image is stored in a folder determined by the degree of smile as additional information added to the out image.
- the out image is recorded so that is stored.
- step S390 image data of the photographer information, the encoded binary image, and the out image to which the degree of smile is added as additional information is supplied from the addition processing unit 82 to the classification unit 271.
- step S391 the classification unit 271 causes the recording unit 54 to record the out image according to the additional information of the out image supplied from the addition processing unit 82, and the imaging process ends.
- each folder provided in the recording unit 54 is associated with a smile degree, and the classification unit 271 associates the same smile degree with the smile degree added to the out image. Identify the folder that is being used. Then, the classification unit 271 supplies the out image to the recording unit 54 so that the out image is stored in the specified folder, and records the out image. That is, the out image is recorded in the recording area of the recording unit 54 that is the area of the specified folder.
- the image processing apparatus 11 adds the binary image obtained from the in-image, the photographer information, and the smile level as additional information to the out-image, and places the out-image in a folder determined by the smile level. Record.
- the out image can be displayed more effectively and the convenience can be improved.
- an out image can be effectively displayed, such as displaying an out image captured with a specific facial expression.
- the out image is recorded in a folder for each photographer's facial expression according to the degree of smile, so that it becomes possible to quickly identify an out image with a specific smile level, Convenience can be improved.
- a folder may be provided for each photographer for each smile level.
- the out image is stored in a folder determined by the photographer information of the out image and the degree of smile.
- the user can operate the operation input unit 51 to display or display a list of the out images captured by the user with a specific facial expression. At that time, for example, the user can select whether to display only the out image or to display the image obtained from the additional information on the out image by combining the operation input unit 51.
- step S431 the extraction unit 83 reads all the out images recorded in the specific folder of the recording unit 54.
- the extraction unit 83 reads out images stored in a folder with an arbitrary smile level designated by the user through an operation on the operation input unit 51.
- step S432 to step S438 When the out image is read in this way, the processing from step S432 to step S438 is performed thereafter, and the reproduction processing ends. However, these processing are the same as the processing from step S353 to step S359 in FIG. The description is omitted.
- the image processing apparatus 11 reads the out image from the folder provided for each photographer's facial expression indicated by the degree of smile added as additional information to the out image, and displays the out image. Further, when displaying the out image, the image processing apparatus 11 synthesizes and displays the binary image on the out image in accordance with a user instruction.
- the out image is displayed more effectively, such as displaying only the out image captured when the photographer has a specific facial expression. can do.
- the out image is stored in a folder determined by the degree of smile as additional information, the out image with a specific smile level can be identified more easily and quickly, and convenience can be improved. .
- the signal processing unit 53 is configured as shown in FIG. 29, for example.
- the smile detection unit 241 does not supply the control unit 52 with the information indicating that the imaging timing has been reached, and the control unit 52 responds to the operation on the operation input unit 51 with the out-image capturing unit. 23 and the in-image capturing unit 22 are instructed to capture the out-image and the in-image.
- or step S466 is the same as the process of step S251 thru
- step S467 the smile detection unit 241 detects the user's smile from the in-image supplied from the in-image capturing unit 22, and the additional processing unit uses the degree of the user's smile on the in-image obtained as a result as additional information. 82. For example, in step S467, processing similar to that in step S312 in FIG. 26 is performed.
- step S468 the addition processing unit 82 uses the photographer information from the face recognition unit 121, the encoded binary image from the compression unit 81, and the degree of smile from the smile detection unit 241 as additional information, as an out image. This is added to the image data of the out image from the imaging unit 23. Then, the addition processing unit 82 supplies the image data of the out image to which the additional information is added to the classification unit 271.
- step S469 the classification unit 271 causes the recording unit 54 to record the out image according to the additional information of the out image supplied from the addition processing unit 82, and the imaging process ends.
- the classification unit 271 specifies a folder that is associated with the same smile degree as the smile degree added to the out image, and records the out image so that the out image is stored in the folder.
- the data is supplied to the unit 54 and recorded. That is, in step S469, processing similar to that in step S391 in FIG. 30 is performed.
- the image processing apparatus 11 adds the binary image obtained from the in-image, the photographer information, and the smile level as additional information to the out-image, and places the out-image in a folder determined by the smile level. Record.
- the out image can be displayed more effectively and the convenience can be improved.
- an out image can be effectively displayed, such as displaying an out image captured with a specific facial expression.
- the out image is recorded, the out image is recorded in a folder for each photographer's facial expression according to the degree of smile, so that the out image having a specific smile level can be quickly identified.
- the image processing apparatus 11 displays the out image captured by the user with a specific facial expression in accordance with the operation of the operation input unit 51 by the user. In such a case, the image processing apparatus 11 performs the reproduction process described with reference to FIG. That is, the out image stored in the folder of the smile level designated by the user is displayed.
- the series of processes described above can be executed by hardware or can be executed by software.
- a program constituting the software is installed in the computer.
- the computer includes, for example, a general-purpose personal computer capable of executing various functions by installing a computer incorporated in dedicated hardware and various programs.
- FIG. 33 is a block diagram illustrating a configuration example of hardware of a computer that executes the above-described series of processing by a program.
- a CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- An input / output interface 505 is further connected to the bus 504.
- An input unit 506, an output unit 507, a recording unit 508, a communication unit 509, and a drive 510 are connected to the input / output interface 505.
- the input unit 506 includes a keyboard, a mouse, a microphone, an image sensor, and the like.
- the output unit 507 includes a display, a speaker, and the like.
- the recording unit 508 includes a hard disk, a nonvolatile memory, and the like.
- the communication unit 509 includes a network interface or the like.
- the drive 510 drives a removable medium 511 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
- the CPU 501 loads the program recorded in the recording unit 508 to the RAM 503 via the input / output interface 505 and the bus 504 and executes the program, for example. Is performed.
- the program executed by the computer (CPU 501) can be provided by being recorded on the removable medium 511 as a package medium, for example.
- the program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the program can be installed in the recording unit 508 via the input / output interface 505 by attaching the removable medium 511 to the drive 510. Further, the program can be received by the communication unit 509 via a wired or wireless transmission medium and installed in the recording unit 508. In addition, the program can be installed in the ROM 502 or the recording unit 508 in advance.
- the program executed by the computer may be a program that is processed in time series in the order described in this specification, or in parallel or at a necessary timing such as when a call is made. It may be a program for processing.
- the present technology can take a cloud computing configuration in which one function is shared by a plurality of devices via a network and is jointly processed.
- each step described in the above flowchart can be executed by one device or can be shared by a plurality of devices.
- the plurality of processes included in the one step can be executed by being shared by a plurality of apparatuses in addition to being executed by one apparatus.
- the present technology can be configured as follows.
- a first imaging unit that images in a first direction
- a second imaging unit that images a second direction different from the first direction
- An additional information generating unit that generates additional information based on a first image obtained by imaging by the first imaging unit
- An image processing apparatus comprising: an additional processing unit that adds the additional information to image data of a second image obtained by imaging by the second imaging unit.
- the additional information generation unit generates the additional information by compressing an information amount of the first image.
- the image processing apparatus according to [1] or [2], wherein the additional information generation unit generates the additional information by compressing a size of the first image.
- the first imaging unit captures an image of a direction of a user operating the image processing apparatus as the first direction, The image processing apparatus according to any one of [1] to [3], further including a combining unit that combines an image obtained from the additional information with the second image.
- the first imaging unit captures an image of a direction of a user operating the image processing apparatus as the first direction, The image processing apparatus according to [1], wherein the additional information generation unit generates photographer information of the second image as the additional information by performing face recognition processing on the first image.
- the additional information generation unit detects information indicating a gaze area of the second image on the second image as the additional information by detecting the user's line-of-sight direction based on the first image.
- the image processing apparatus according to [1].
- the image processing apparatus further including a reproducing unit that controls display of the second image based on the additional information.
- the image processing device according to [9], wherein the reproduction unit displays a plurality of the second images based on the additional information.
- An angle-of-view conversion unit that performs an angle-of-view conversion process on the second image;
- the additional information generation unit uses, as the additional information, a crop image obtained by cutting out an area including a main subject on the first image,
- the image processing apparatus according to [1], wherein the addition processing unit adds the additional information to image data of the second image on which the angle-of-view conversion processing has been performed.
- the crop image as the additional information and the second image that has been subjected to the angle-of-view conversion process are combined and combined to form the same aspect as the second image captured by the second imaging unit.
- the first imaging unit captures an image of a direction of a user operating the image processing apparatus as the first direction
- the additional information generation unit detects the smile of the user based on the first image and generates the degree of smile of the user as the additional information.
- the image processing apparatus further including a classification unit that records the second image in a recording area determined by the degree of smile as the additional information.
- the first imaging unit captures an image of a direction of a user operating the image processing apparatus as the first direction
- the additional information generation unit detects the smile of the user based on the first image and generates the degree of smile of the user as the additional information.
- the image processing apparatus according to [1], further including a classification unit that records the second image in a recording area determined by the degree of smile as the additional information.
- 11 image processing device 21 display unit, 22 in-image imaging unit, 23 out-image imaging unit, 53 signal processing unit, 81 compression unit, 82 additional processing unit, 84 decoding unit, 85 playback unit, 121 face recognition unit, 151 gaze Area detection unit, 181 angle of view conversion unit, 182 main area detection unit, 183 crop processing unit
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Computing Systems (AREA)
- Studio Devices (AREA)
- Image Processing (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
〈本技術の概要について〉
本技術は、例えば多機能型携帯電話機やデジタルカメラなど、互いに異なる方向を撮像可能な複数の撮像部を有する機器に関し、主となる撮像部で画像を撮像したときに、その画像と同時に撮像された他の画像から得られる情報を付加情報として生成するものである。本技術では、このような付加情報を生成することで、より効果的に画像を表示することができるようになる。また、付加情報を画像データに埋め込むことで、画像の撮像時に面倒な操作を必要とせずに、画像の再生時により自由度の高い表示制御を行なうことができるようになり、利便性を向上させることができる。
次に、画像処理装置11のより詳細な構成について説明する。
また、図3の信号処理部53は、より詳細には図4に示すように構成される。
ところで、ユーザが操作入力部51に対する操作を行なって、アウト画像に付加情報を付加するモードを選択し、所望の被写体の撮像を指示すると、画像処理装置11は撮像処理を開始して、付加情報が埋め込まれたアウト画像を記録させる。以下、図5のフローチャートを参照して、画像処理装置11による撮像処理について説明する。
記録部54にアウト画像が記録されると、ユーザは操作入力部51を操作して、記録されたアウト画像を再生させることができる。その際、ユーザは例えば操作入力部51を操作することで、アウト画像のみを表示させるか、またはアウト画像に付加情報から得られる画像を合成表示させるかを選択することができる。
〈信号処理部の構成例〉
また、以上においては、付加情報としてイン画像から得られた2値画像を生成する例について説明したが、2値画像以外にも他の情報が付加情報として生成されるようにしてもよい。例えば、付加情報としてイン画像から得られた撮影者情報と、2値画像とが生成されるようにしてもよい。
次に、図9のフローチャートを参照して、信号処理部53が図8に示す構成とされる場合に行なわれる撮像処理について説明する。
また、記録部54にアウト画像が記録されると、ユーザは操作入力部51を操作して、特定のユーザが撮像したアウト画像の一覧を表示させることができる。その際、ユーザは例えば操作入力部51を操作することで、アウト画像のみを表示させるか、またはアウト画像に付加情報から得られる画像を合成表示させるかを選択することができる。
〈信号処理部の構成例〉
また、以上においては、付加情報として符号化された2値画像と撮影者情報が生成される場合について説明したが、付加情報としてアウト画像上の注視領域を示す情報が生成されるようにしてもよい。ここで、注視領域とは、アウト画像の撮像時に撮影者であるユーザが注視していたアウト画像上の領域である。
次に、図14のフローチャートを参照して、信号処理部53が図13に示した構成とされる場合に行なわれる撮像処理について説明する。
また、記録部54にアウト画像が記録されると、ユーザは操作入力部51を操作してアウト画像をスライドショー表示させることができる。その際、ユーザは例えば操作入力部51を操作することで、アウト画像を1枚ずつ表示させるか、または複数のアウト画像を重ねて同時に表示させるかを選択することができる。
〈信号処理部の構成例〉
また、イン画像の一部の領域が付加情報とされて、アウト画像の再生時に付加情報とされた画像がアウト画像に合成されるようにしてもよい。
次に、図19のフローチャートを参照して、信号処理部53が図18の構成とされる場合に画像処理装置11により行なわれる撮像処理について説明する。
また、記録部54にアウト画像が記録されると、ユーザは操作入力部51を操作してアウト画像を表示させることができる。その際、ユーザは例えば操作入力部51を操作することで、アウト画像のみを表示させるか、またはアウト画像に付加情報としてのクロップ画像を合成して表示させるかを選択することができる。
〈信号処理部の構成例〉
また、上述した第2の実施の形態では、アウト画像に付加情報として撮影者情報が付加される例について説明したが、付加情報としての撮影者情報に応じてアウト画像の記録領域が定められるようにしてもよい。
次に、図23のフローチャートを参照して、信号処理部53が図22に示す構成とされる場合に行なわれる撮像処理について説明する。
また、記録部54にアウト画像が記録されると、ユーザは操作入力部51を操作して、特定のユーザが撮像したアウト画像の一覧を表示させることができる。その際、ユーザは例えば操作入力部51を操作することで、アウト画像のみを表示させるか、またはアウト画像に付加情報から得られる画像を合成表示させるかを選択することができる。
〈信号処理部の構成例〉
さらに、画像処理装置11を操作するユーザの笑顔を検出し、笑顔が検出されたタイミングでアウト画像とイン画像を撮像するとともに、ユーザの笑顔の度合いが付加情報としてアウト画像に付加されるようにしてもよい。
次に、図26のフローチャートを参照して、信号処理部53が図25に示す構成とされる場合に行なわれる撮像処理について説明する。この場合、ユーザにより操作入力部51が操作され、アウト画像の撮像開始が指示されると撮像処理が開始される。
また、記録部54にアウト画像が記録されると、ユーザは操作入力部51を操作して、特定の笑顔の度合いのアウト画像をスライドショー表示させたり、一覧表示させたりすることができる。その際、ユーザは例えば操作入力部51を操作することで、アウト画像のみを表示させるか、またはアウト画像に付加情報から得られる画像を合成表示させるかを選択することができる。
〈信号処理部の構成例〉
さらに、以上においてはアウト画像の付加情報として笑顔の度合いが付加される例について説明したが、付加情報としての笑顔の度合いに応じてアウト画像の記録領域が定められるようにしてもよい。
次に、図30のフローチャートを参照して、信号処理部53が図29に示す構成とされる場合に行なわれる撮像処理について説明する。
また、記録部54にアウト画像が記録されると、ユーザは操作入力部51を操作して、特定の表情でユーザが撮像したアウト画像を表示させたり、一覧表示させたりすることができる。その際、ユーザは例えば操作入力部51を操作することで、アウト画像のみを表示させるか、またはアウト画像に付加情報から得られる画像を合成表示させるかを選択することができる。
〈撮像処理の説明〉
また、上述した第7の実施の形態では、イン画像から笑顔が検出されたときにアウト画像が撮像される例について説明したが、アウト画像とイン画像がユーザの指示に応じて撮像され、笑顔の度合いに応じたフォルダにアウト画像が記録されるようにしてもよい。
第1の方向を撮像する第1の撮像部と、
前記第1の方向とは異なる第2の方向を撮像する第2の撮像部と、
前記第1の撮像部による撮像で得られた第1の画像に基づいて付加情報を生成する付加情報生成部と、
前記第2の撮像部による撮像で得られた第2の画像の画像データに前記付加情報を付加する付加処理部と
を備える画像処理装置。
[2]
前記付加情報生成部は、前記第1の画像の情報量を圧縮することで、前記付加情報を生成する
[1]に記載の画像処理装置。
[3]
前記付加情報生成部は、前記第1の画像のサイズを圧縮することで、前記付加情報を生成する
[1]または[2]に記載の画像処理装置。
[4]
前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報から得られる画像を前記第2の画像に合成する合成部をさらに備える
[1]乃至[3]の何れかに記載の画像処理装置。
[5]
前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報生成部は、前記第1の画像に対する顔認識処理を行なうことで、前記第2の画像の撮影者情報を前記付加情報として生成する
[1]に記載の画像処理装置。
[6]
前記付加情報に基づいて、複数の前記第2の画像から特定の撮影者により撮像された前記第2の画像を検出する抽出部と、
前記抽出部により検出された前記第2の画像のみを表示させる再生部と
をさらに備える[5]に記載の画像処理装置。
[7]
前記付加情報としての前記撮影者情報により定まる記録領域に前記第2の画像を記録させる分類部をさらに備える
[5]に記載の画像処理装置。
[8]
前記第1の撮像部が設けられた、前記画像処理装置を操作するユーザ側の面に設けられ、前記第2の画像の撮像時に前記第2の撮像部により取り込まれた画像を表示する表示部をさらに備え、
前記付加情報生成部は、前記第1の画像に基づいて前記ユーザの視線方向を検出することで、前記第2の画像上の前記ユーザが注視していた注視領域を示す情報を前記付加情報として生成する
[1]に記載の画像処理装置。
[9]
前記付加情報に基づいて、前記第2の画像の表示を制御する再生部をさらに備える
[8]に記載の画像処理装置。
[10]
前記再生部は、前記付加情報に基づいて複数の前記第2の画像を重ねて表示させる
[9]に記載の画像処理装置。
[11]
前記第2の画像に対して画角変換処理を施す画角変換部をさらに備え、
前記付加情報生成部は、前記第1の画像上の主要な被写体を含む領域を切り出して得られたクロップ画像を前記付加情報とし、
前記付加処理部は、前記画角変換処理が施された前記第2の画像の画像データに前記付加情報を付加する
[1]に記載の画像処理装置。
[12]
前記付加情報としての前記クロップ画像と、前記画角変換処理が施された前記第2の画像とを並べて合成することで、前記第2の撮像部により撮像された前記第2の画像と同じアスペクト比の合成画像を生成する合成部をさらに備える
[11]に記載の画像処理装置。
[13]
前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報生成部は、前記第1の画像に基づいて前記ユーザの笑顔を検出するとともに前記ユーザの笑顔の度合いを前記付加情報として生成し、
前記第2の撮像部は、前記ユーザの笑顔が検出されたとき前記第2の方向を撮像する
[1]に記載の画像処理装置。
[14]
前記付加情報としての前記笑顔の度合いにより定まる記録領域に前記第2の画像を記録させる分類部をさらに備える
[13]に記載の画像処理装置。
[15]
前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報生成部は、前記第1の画像に基づいて前記ユーザの笑顔を検出するとともに前記ユーザの笑顔の度合いを前記付加情報として生成し、
前記付加情報としての前記笑顔の度合いにより定まる記録領域に前記第2の画像を記録させる分類部をさらに備える
[1]に記載の画像処理装置。
Claims (17)
- 第1の方向を撮像する第1の撮像部と、
前記第1の方向とは異なる第2の方向を撮像する第2の撮像部と、
前記第1の撮像部による撮像で得られた第1の画像に基づいて付加情報を生成する付加情報生成部と、
前記第2の撮像部による撮像で得られた第2の画像の画像データに前記付加情報を付加する付加処理部と
を備える画像処理装置。 - 前記付加情報生成部は、前記第1の画像の情報量を圧縮することで、前記付加情報を生成する
請求項1に記載の画像処理装置。 - 前記付加情報生成部は、前記第1の画像のサイズを圧縮することで、前記付加情報を生成する
請求項2に記載の画像処理装置。 - 前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報から得られる画像を前記第2の画像に合成する合成部をさらに備える
請求項3に記載の画像処理装置。 - 前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報生成部は、前記第1の画像に対する顔認識処理を行なうことで、前記第2の画像の撮影者情報を前記付加情報として生成する
請求項1に記載の画像処理装置。 - 前記付加情報に基づいて、複数の前記第2の画像から特定の撮影者により撮像された前記第2の画像を検出する抽出部と、
前記抽出部により検出された前記第2の画像のみを表示させる再生部と
をさらに備える請求項5に記載の画像処理装置。 - 前記付加情報としての前記撮影者情報により定まる記録領域に前記第2の画像を記録させる分類部をさらに備える
請求項5に記載の画像処理装置。 - 前記第1の撮像部が設けられた、前記画像処理装置を操作するユーザ側の面に設けられ、前記第2の画像の撮像時に前記第2の撮像部により取り込まれた画像を表示する表示部をさらに備え、
前記付加情報生成部は、前記第1の画像に基づいて前記ユーザの視線方向を検出することで、前記第2の画像上の前記ユーザが注視していた注視領域を示す情報を前記付加情報として生成する
請求項1に記載の画像処理装置。 - 前記付加情報に基づいて、前記第2の画像の表示を制御する再生部をさらに備える
請求項8に記載の画像処理装置。 - 前記再生部は、前記付加情報に基づいて複数の前記第2の画像を重ねて表示させる
請求項9に記載の画像処理装置。 - 前記第2の画像に対して画角変換処理を施す画角変換部をさらに備え、
前記付加情報生成部は、前記第1の画像上の主要な被写体を含む領域を切り出して得られたクロップ画像を前記付加情報とし、
前記付加処理部は、前記画角変換処理が施された前記第2の画像の画像データに前記付加情報を付加する
請求項1に記載の画像処理装置。 - 前記付加情報としての前記クロップ画像と、前記画角変換処理が施された前記第2の画像とを並べて合成することで、前記第2の撮像部により撮像された前記第2の画像と同じアスペクト比の合成画像を生成する合成部をさらに備える
請求項11に記載の画像処理装置。 - 前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報生成部は、前記第1の画像に基づいて前記ユーザの笑顔を検出するとともに前記ユーザの笑顔の度合いを前記付加情報として生成し、
前記第2の撮像部は、前記ユーザの笑顔が検出されたとき前記第2の方向を撮像する
請求項1に記載の画像処理装置。 - 前記付加情報としての前記笑顔の度合いにより定まる記録領域に前記第2の画像を記録させる分類部をさらに備える
請求項13に記載の画像処理装置。 - 前記第1の撮像部は、前記画像処理装置を操作するユーザの方向を前記第1の方向として撮像し、
前記付加情報生成部は、前記第1の画像に基づいて前記ユーザの笑顔を検出するとともに前記ユーザの笑顔の度合いを前記付加情報として生成し、
前記付加情報としての前記笑顔の度合いにより定まる記録領域に前記第2の画像を記録させる分類部をさらに備える
請求項1に記載の画像処理装置。 - 第1の撮像部により第1の方向を撮像し、
第2の撮像部により前記第1の方向とは異なる第2の方向を撮像し、
前記第1の撮像部による撮像で得られた第1の画像に基づいて付加情報を生成し、
前記第2の撮像部による撮像で得られた第2の画像の画像データに前記付加情報を付加する
ステップを含む画像処理方法。 - 第1の撮像部により第1の方向を撮像し、
第2の撮像部により前記第1の方向とは異なる第2の方向を撮像し、
前記第1の撮像部による撮像で得られた第1の画像に基づいて付加情報を生成し、
前記第2の撮像部による撮像で得られた第2の画像の画像データに前記付加情報を付加する
ステップを含む処理をコンピュータに実行させるプログラム。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/762,250 US9727993B2 (en) | 2013-02-27 | 2014-02-14 | Image processing apparatus, image processing method, and program |
AU2014222020A AU2014222020A1 (en) | 2013-02-27 | 2014-02-14 | Image processing device, method, and program |
CN201480009561.6A CN105027552B (zh) | 2013-02-27 | 2014-02-14 | 图像处理设备和图像处理方法 |
EP14757727.4A EP2963910A4 (en) | 2013-02-27 | 2014-02-14 | IMAGE PROCESSING DEVICE, METHOD, AND PROGRAM |
US15/434,848 US20170163887A1 (en) | 2013-02-27 | 2017-02-16 | Image processing apparatus, image processing method, and program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-037867 | 2013-02-27 | ||
JP2013037867 | 2013-02-27 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/762,250 A-371-Of-International US9727993B2 (en) | 2013-02-27 | 2014-02-14 | Image processing apparatus, image processing method, and program |
US15/434,848 Continuation US20170163887A1 (en) | 2013-02-27 | 2017-02-16 | Image processing apparatus, image processing method, and program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014132816A1 true WO2014132816A1 (ja) | 2014-09-04 |
Family
ID=51428088
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/053481 WO2014132816A1 (ja) | 2013-02-27 | 2014-02-14 | 画像処理装置および方法、並びにプログラム |
Country Status (5)
Country | Link |
---|---|
US (2) | US9727993B2 (ja) |
EP (1) | EP2963910A4 (ja) |
CN (1) | CN105027552B (ja) |
AU (1) | AU2014222020A1 (ja) |
WO (1) | WO2014132816A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018128958A (ja) * | 2017-02-10 | 2018-08-16 | 富士通株式会社 | 端末装置、情報処理方法、情報処理プログラムおよび情報処理装置 |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9990524B2 (en) * | 2016-06-16 | 2018-06-05 | Hand Held Products, Inc. | Eye gaze detection controlled indicia scanning system and method |
JP2023058322A (ja) * | 2021-10-13 | 2023-04-25 | キヤノン株式会社 | 映像処理装置及びその制御方法及びプログラム |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6246779B1 (en) | 1997-12-12 | 2001-06-12 | Kabushiki Kaisha Toshiba | Gaze position detection apparatus and method |
JP2005094741A (ja) * | 2003-08-14 | 2005-04-07 | Fuji Photo Film Co Ltd | 撮像装置及び画像合成方法 |
JP2006033611A (ja) * | 2004-07-20 | 2006-02-02 | Canon Inc | 処理装置及び処理方法 |
JP2007110262A (ja) * | 2005-10-11 | 2007-04-26 | Nikon Corp | 画像鑑賞用ソフトウェア及び画像蓄積装置 |
US7595809B2 (en) | 2005-05-11 | 2009-09-29 | Hewlett-Packard Development Company, L.P. | Method and system for determining an occlusion cost for concurrently presenting one or more images in a shared viewing region |
JP2010003201A (ja) * | 2008-06-23 | 2010-01-07 | Olympus Imaging Corp | 画像処理装置、画像処理方法、およびプログラム |
JP2010062853A (ja) * | 2008-09-03 | 2010-03-18 | Fujifilm Corp | 情報処理装置、方法およびプログラム |
JP2011250340A (ja) | 2010-05-31 | 2011-12-08 | Sanyo Electric Co Ltd | 撮像装置およびその制御方法 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3948387B2 (ja) | 2002-10-24 | 2007-07-25 | 松下電器産業株式会社 | ディジタルカメラおよびディジタルカメラ付き携帯電話装置 |
JP2004266376A (ja) | 2003-02-25 | 2004-09-24 | Matsushita Electric Ind Co Ltd | 映像合成装置 |
KR100836616B1 (ko) * | 2006-11-14 | 2008-06-10 | (주)케이티에프테크놀로지스 | 영상 합성 기능을 가지는 휴대용 단말기 및 휴대용단말기의 영상 합성 방법 |
US7991285B2 (en) * | 2008-01-08 | 2011-08-02 | Sony Ericsson Mobile Communications Ab | Using a captured background image for taking a photograph |
JP5740826B2 (ja) * | 2010-03-29 | 2015-07-01 | セイコーエプソン株式会社 | 画像表示装置、画像情報処理装置及び画像情報処理方法 |
US20120179960A1 (en) * | 2011-01-07 | 2012-07-12 | Cok Ronald S | Image layout adjustment aparatus |
JP5814566B2 (ja) * | 2011-02-28 | 2015-11-17 | オリンパス株式会社 | 撮像装置、撮像方法及び撮像装置の制御プログラム |
US8891926B2 (en) * | 2011-12-19 | 2014-11-18 | Corning Cable Systems Llc | Methods of reducing and/or avoiding fiber ordering in a connectorized multi-fiber, fiber optic cable system, and related fiber optic cables and assemblies |
JP2014160982A (ja) * | 2013-02-20 | 2014-09-04 | Sony Corp | 画像処理装置および撮影制御方法、並びにプログラム |
-
2014
- 2014-02-14 CN CN201480009561.6A patent/CN105027552B/zh active Active
- 2014-02-14 EP EP14757727.4A patent/EP2963910A4/en not_active Ceased
- 2014-02-14 US US14/762,250 patent/US9727993B2/en active Active
- 2014-02-14 WO PCT/JP2014/053481 patent/WO2014132816A1/ja active Application Filing
- 2014-02-14 AU AU2014222020A patent/AU2014222020A1/en not_active Abandoned
-
2017
- 2017-02-16 US US15/434,848 patent/US20170163887A1/en not_active Abandoned
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6246779B1 (en) | 1997-12-12 | 2001-06-12 | Kabushiki Kaisha Toshiba | Gaze position detection apparatus and method |
JP2005094741A (ja) * | 2003-08-14 | 2005-04-07 | Fuji Photo Film Co Ltd | 撮像装置及び画像合成方法 |
JP2006033611A (ja) * | 2004-07-20 | 2006-02-02 | Canon Inc | 処理装置及び処理方法 |
US7595809B2 (en) | 2005-05-11 | 2009-09-29 | Hewlett-Packard Development Company, L.P. | Method and system for determining an occlusion cost for concurrently presenting one or more images in a shared viewing region |
JP2007110262A (ja) * | 2005-10-11 | 2007-04-26 | Nikon Corp | 画像鑑賞用ソフトウェア及び画像蓄積装置 |
JP2010003201A (ja) * | 2008-06-23 | 2010-01-07 | Olympus Imaging Corp | 画像処理装置、画像処理方法、およびプログラム |
JP2010062853A (ja) * | 2008-09-03 | 2010-03-18 | Fujifilm Corp | 情報処理装置、方法およびプログラム |
JP2011250340A (ja) | 2010-05-31 | 2011-12-08 | Sanyo Electric Co Ltd | 撮像装置およびその制御方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2963910A4 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018128958A (ja) * | 2017-02-10 | 2018-08-16 | 富士通株式会社 | 端末装置、情報処理方法、情報処理プログラムおよび情報処理装置 |
Also Published As
Publication number | Publication date |
---|---|
CN105027552B (zh) | 2019-02-19 |
EP2963910A4 (en) | 2016-12-07 |
US20150348301A1 (en) | 2015-12-03 |
AU2014222020A1 (en) | 2015-08-06 |
EP2963910A1 (en) | 2016-01-06 |
US20170163887A1 (en) | 2017-06-08 |
US9727993B2 (en) | 2017-08-08 |
CN105027552A (zh) | 2015-11-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2339536B1 (en) | Image processing system, image processing apparatus, image processing method, and program | |
KR102650850B1 (ko) | 영상 음향 처리 장치 및 방법, 및 프로그램이 저장된 컴퓨터 판독 가능한 기록 매체 | |
JP7441926B2 (ja) | 映像コーディングを行うコンピュータプログラム | |
JP2020095611A (ja) | 情報処理装置、システム、情報処理装置の制御方法、及び、プログラム | |
WO2014132816A1 (ja) | 画像処理装置および方法、並びにプログラム | |
CN114998935A (zh) | 图像处理方法、装置、计算机设备及存储介质 | |
JP2004056488A (ja) | 画像処理方法、画像処理装置および画像通信装置 | |
CN112381749A (zh) | 一种图像处理方法、图像处理装置和电子设备 | |
JP2015156540A (ja) | 画像処理装置、画像処理方法および画像処理プログラム | |
JP6340675B1 (ja) | オブジェクト抽出装置、オブジェクト認識システム及びメタデータ作成システム | |
CN113647093A (zh) | 图像处理装置、3d模型生成方法和程序 | |
JP2017188787A (ja) | 撮像装置、画像合成方法、および画像合成プログラム | |
JP2010097449A (ja) | 画像合成装置、及び画像合成方法、画像合成プログラム | |
CN116962743A (zh) | 视频图像编码、抠图方法和装置及直播系统 | |
CN111034187A (zh) | 动态图像的生成方法、装置、可移动平台和存储介质 | |
JP2011029947A (ja) | 画像合成装置および画像合成方法 | |
JP2016129281A (ja) | 画像処理装置 | |
JP7191514B2 (ja) | 画像処理装置、画像処理方法、およびプログラム | |
JP5694060B2 (ja) | 画像処理装置、画像処理方法、プログラム、撮像装置及びテレビジョン受像機 | |
JP4223940B2 (ja) | 画像生成装置、携帯端末、画像生成方法、画像表示方法及びプログラム | |
JP2014220546A (ja) | 画像記録装置及びその方法 | |
JP5234086B2 (ja) | 領域特定方法、領域特定装置、プログラム、サーバ及びシステム | |
JP2012119798A (ja) | 画像生成方法、画像生成装置及びプログラム | |
KR101897692B1 (ko) | 프레임 분리 방식을 이용한 영상 합성 서버 | |
JP2011077934A (ja) | 画像処理装置、画像処理方法及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201480009561.6 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14757727 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14762250 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2014222020 Country of ref document: AU Date of ref document: 20140214 Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014757727 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
NENP | Non-entry into the national phase |
Ref country code: JP |