WO2017183543A1 - 情報処理装置、情報処理方法、及びプログラム - Google Patents
情報処理装置、情報処理方法、及びプログラム Download PDFInfo
- Publication number
- WO2017183543A1 WO2017183543A1 PCT/JP2017/015038 JP2017015038W WO2017183543A1 WO 2017183543 A1 WO2017183543 A1 WO 2017183543A1 JP 2017015038 W JP2017015038 W JP 2017015038W WO 2017183543 A1 WO2017183543 A1 WO 2017183543A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- unit
- information
- region
- image
- image processing
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/197—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/60—Editing figures and text; Combining figures or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
- G06T7/0014—Biomedical image inspection using an image reference approach
- G06T7/0016—Biomedical image inspection using an image reference approach involving temporal comparison
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/80—Recognising image objects characterised by unique random patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20192—Edge enhancement; Edge preservation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20201—Motion blur correction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30041—Eye; Retina; Ophthalmic
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30101—Blood vessel; Artery; Vein; Vascular
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/95—Pattern authentication; Markers therefor; Forgery detection
Definitions
- the present disclosure relates to an information processing apparatus, an information processing method, and a program.
- iris authentication is a technique for performing identity verification using iris information acquired from an annular iris around the pupil of the eye.
- iris authentication is used for authentication when entering a room, logging on to a computer, and the like, and in recent years, it is also used for authentication when using a mobile terminal such as a smartphone.
- an image disclosed to a third party includes an iris
- the third party may acquire iris information from the image, and unauthorized authentication by the third party may be performed in iris authentication.
- the present disclosure proposes an information processing apparatus, an information processing method, and a program capable of suppressing unauthorized authentication due to acquisition of iris information from an image.
- image processing is performed on a processing area corresponding to the living body area so that a living body information that can be acquired from the living body area and an area specifying unit that specifies a living body area for biometric authentication from image data are changed.
- An information processing apparatus is provided.
- the biometric area for biometric authentication is specified from the image data, and image processing is performed on the processing area corresponding to the biometric area so that biometric information that can be acquired from the biometric area changes.
- An information processing method is provided.
- the processing area corresponding to the biometric area is changed so that the function of specifying the biometric area for biometric authentication from the image data and the biometric information that can be acquired from the biometric area are changed.
- a program for realizing a function of performing image processing is provided.
- FIG. 6 is an explanatory diagram illustrating an operation example when restoring an input image from an output image in the embodiment. It is a block diagram which shows the structural example of the information processing apparatus 4 which concerns on the modification 1.
- FIG. It is a block diagram showing an example of composition of information processor 5 which applied this art to fingerprint authentication. It is explanatory drawing which shows the example of a feature point deformation
- transformation process. 7 is an explanatory diagram illustrating an operation example of the information processing apparatus 5.
- FIG. It is a block diagram showing an example of composition of information processor 6 which applied this art to vein authentication. It is explanatory drawing which shows the example of a vein deformation
- FIG. 7 is an explanatory diagram illustrating an operation example of the information processing apparatus 6.
- FIG. It is a block diagram which shows the structural example of the information processing system 1000 which concerns on the modification 3.
- FIG. It is a flowchart figure which shows the operation example of the modification. It is a flowchart figure which shows the operation example of the modification. It is explanatory drawing which shows the hardware structural example.
- Iris authentication is a technique for performing identity verification using iris information (an example of biometric information) acquired from an annular iris around the pupil of the eye.
- iris information an example of biometric information
- camera technology not only a device dedicated to iris authentication but also a camera of a mobile terminal such as a smartphone can shoot with image quality capable of acquiring iris information for iris authentication.
- iris information for iris authentication is acquired by the third party, and unauthorized authentication is performed by the third party in iris authentication. There is a fear.
- the present embodiment has been created with the above circumstances in mind. According to the present embodiment, it is possible to suppress unauthorized authentication due to acquisition of iris information from an image.
- a configuration example for realizing the above-described effect will be described.
- FIG. 1 is an explanatory diagram illustrating a configuration example of the information processing apparatus 1 according to the first embodiment of the present disclosure.
- the information processing apparatus 1 according to the present embodiment includes a control unit 10, a communication unit 11, an imaging unit 12, a storage unit 13, an operation unit 14, and a display unit 15.
- the control unit 10 controls each component of the information processing apparatus 1.
- the control unit 10 has a communication control function for controlling communication by the communication unit 11 described later and a display control function for controlling display by the display unit 15.
- the control unit 10 according to the present embodiment also functions as an area specifying unit 120, a determination unit 140, an image processing unit 160, and an iris registration authentication unit 180.
- the region specifying unit 120 specifies an iris region (an example of a biological region) from an input image (image data) acquired by the imaging unit 12.
- the area specifying unit 120 specifies a black eye part (pupil) and a white eye part in an eye area detected by a face recognition technique or the like, and an area specified by the boundary between the pupil and the white eye is set as an iris area. You may specify.
- the information of the iris region specified by the region specifying unit 120 is provided to the determination unit 140, the image processing unit 160, and the iris registration authentication unit 180.
- the image data that is the target of the iris region specification by the region specification unit 120 is not limited to the image acquired by the imaging unit 12, and may be an image acquired from an external device via the communication unit 11, for example.
- the image may be an image stored in advance in the storage unit 13.
- the determination unit 140 determines whether iris information used for iris authentication can be acquired from the iris region specified by the region specifying unit 120. For example, the determination unit 140 may determine whether the iris information can be acquired from the iris region based on the size of the iris region or the color included in the iris region.
- the size of the iris region used for determination by the determination unit 140 may be the horizontal distance between the left end and the right end of the iris region, or the outer diameter of a circle obtained by approximating the iris region to a circle. It may be the area of the iris region. Further, the determination unit 140 may determine that the iris information can be acquired from the iris region, for example, when the size of the iris region is a predetermined value or more.
- the determination unit 140 may determine that the iris information can be acquired from the iris region when the average value of the colors included in the iris region is included in the predetermined color range.
- the predetermined color range may be, for example, a range indicating blue or green.
- the determination unit 140 includes the iris region. It may be determined that iris information can be acquired from
- the control unit 10 may control the display unit 15 so that a warning is displayed.
- the user grasps that the iris information can be acquired from the image (image data), and for example, prevents the image from being transmitted to the outside, or performs image processing by the image processing unit 160 described later. It is possible to set as follows.
- the determination unit 140 determines the iris from the iris region specified from the image data by the region specifying unit 120. It may be determined whether or not information can be acquired. According to this configuration, since it is determined whether or not the iris information can be acquired from the image data before the image data is transmitted to the outside, it is possible to further suppress unauthorized authentication.
- the image processing unit 160 performs image processing on the processing region corresponding to the iris region so that the iris information that can be acquired from the iris region specified by the region specifying unit 120 changes, and generates an output image.
- the processing area may be, for example, the iris area itself, a rectangular area circumscribing the iris area, or an area including the iris area and the pupil area.
- the image processing unit 160 extracts the iris region specified by the region specifying unit 120 as the processing region from the input image, performs image processing, combines the image processing result with the input image, and generates an output image. May be.
- the iris information that can be acquired from the output image is different from the iris information that can be acquired from the input image, so that even if the output image is provided to a third party, for example, unauthorized authentication by a third party is performed. It is possible to suppress.
- the image processing unit 160 may perform image processing on the processing region when the determination unit 140 determines that the iris information can be acquired from the iris region.
- image processing described later may be performed when iris information can be acquired from an iris region, and the processing amount when iris information cannot be acquired can be suppressed.
- the image processing by the image processing unit 160 is not performed, and thus the influence (for example, image degradation) due to the image processing does not occur.
- the image processing unit 160 has a function as the blurring unit 162 as illustrated in FIG. 1, and the image processing performed by the image processing unit 160 includes blurring processing by the blurring unit 162.
- the blurring process performed by the blurring unit 162 may be a normalized blur (Box blur) process in which a 3 ⁇ 3 filter as shown in the following expression is applied to a 3 ⁇ 3 pixel range around each pixel. Good.
- the blurring process by the blurring unit 162 may be a Gaussian blur process in which a 3 ⁇ 3 filter as shown in the following equation is applied to a range of 3 ⁇ 3 pixels around each pixel. .
- FIG. 2 is an explanatory diagram schematically showing image processing by the image processing unit 160 according to the present embodiment.
- the iris region E12 is specified by the region specifying unit 120.
- the iris region E12 is specified as a region around the pupil region E14, for example.
- the image processing unit 160 extracts an iris region E12 from the input image E10 shown in FIG. 2, performs blurring processing (an example of image processing), combines the result of blurring processing with the input image E10, and outputs an output image E20. Generate. As shown in FIG. 2, the iris region E22 in the output image E20 has a blurring effect and the pattern is lost or unclear compared to the iris region E12 in the input image E10.
- the output image E20 is provided to a third party. Even so, fraud authentication by a third party can be suppressed.
- the blurring process is performed on the processing area (the iris area E12 in the example of FIG. 2), the blurring process is performed on the area other than the iris area E22 as in the output image E20 of FIG. There is little influence and the feeling of incongruity is small as compared with the case where the entire image is processed.
- FIG. 2 shows an example in which the input image is an eye image
- the input image is not limited to an eye image
- the output image generated by the image processing unit 160 may be stored in, for example, the storage unit 13 or may be transmitted to an external device via the communication unit 11.
- the iris registration authentication unit 180 shown in FIG. 1 performs iris information registration processing related to iris authentication, and authentication processing for collating the registered iris information with the iris information for authentication.
- the iris registration authentication unit 180 may perform registration processing or authentication processing in accordance with a user input via the operation unit 14 described later.
- the iris registration authentication unit 180 acquires the iris information from the iris region specified by the region specifying unit 120 and stores the iris information in the storage unit 13 to perform the registration process. In addition, the iris registration authentication unit 180 acquires the iris information from the iris region specified by the region specifying unit 120 and collates with the iris information stored in the storage unit 13 to perform the authentication process.
- the communication unit 11 performs data transmission / reception (communication) with an external device.
- the communication unit 11 may transmit the output image generated by the image processing unit 160 to an external device.
- the imaging unit 12 is a camera module that acquires an image.
- the imaging unit 12 acquires an image by imaging a real space using an imaging element such as a CCD (Charge Coupled Device) or a CMOS (Complementary Metal Oxide Semiconductor).
- the image acquired by the imaging unit 12 is provided to the region specifying unit 120 as an input image, for example.
- the storage unit 13 stores programs and parameters for the functions of the information processing apparatus 1 to function.
- the storage unit 13 may store registered iris information.
- the storage unit 13 may store an output image generated by the image processing unit 160 performing image processing on the iris region of the input image.
- the operation unit 14 receives user input and provides it to the control unit 10.
- the user may select the mode of the information processing apparatus 1 by operating the operation unit 14, for example, registration / authentication mode for performing iris information registration or iris authentication, or shooting and recording an image.
- One of the shooting modes for performing the above may be selected.
- the user may operate the operation unit 14 to switch on / off setting of an image processing function for the iris region (hereinafter, also referred to as iris processing setting) by the image processing unit 160.
- the user may operate the operation unit 14 to perform an operation input for transmitting image data captured by the imaging unit 12 or image data stored in the storage unit 13 to the outside.
- the operation unit 14 may be realized by a touch panel, for example, or may be realized by a mouse, a keyboard, a button, a switch, a lever, a dial, or the like.
- the display unit 15 is a display device such as a display, for example, and is controlled by the control unit 10 to display various screens.
- the display unit 15 may display a mode selection screen for allowing the user to select either the registration / authentication mode or the shooting mode.
- the display unit 15 prompts the iris processing setting to be turned on when the determination unit 140 determines that the iris information can be acquired from the iris region of the input image and the iris processing setting is off, A warning screen for switching processing settings may be displayed.
- the configuration example of the information processing apparatus 1 according to the present embodiment has been specifically described above.
- the configuration of the information processing apparatus 1 illustrated in FIG. 1 is an example, and the present embodiment is not limited to the example.
- the example in which the information processing apparatus 1 includes the image processing function and the iris registration authentication function has been described above.
- the present technology is not limited to the example, and the image processing function and the iris registration authentication function are different devices. It may be provided.
- the information processing apparatus 1 may be a portable terminal carried by a user such as a smartphone or a PDA (Personal Digital Assistant), or may be a PC (Personal Computer), a digital home appliance, a game machine, or other types.
- the apparatus may be used.
- each function of the control unit 10 according to the present embodiment may be included in another information processing apparatus connected via the communication unit 11.
- FIG. 3 is a flowchart illustrating an operation example of the information processing apparatus 1 according to the present embodiment.
- the display unit 15 is controlled by the control unit 10, and the registration / authentication mode or the shooting mode is set to the user.
- a mode selection screen for selection is displayed (S104).
- step S106 iris information registration processing or iris authentication processing by the iris registration authentication unit 180 is performed.
- step S108 the region specifying unit 120 specifies the iris region from the image acquired by the imaging unit 12.
- the determination unit 140 determines whether or not iris information can be acquired from the iris region identified in step S108 (S110).
- the control unit 10 When it is determined that the iris information can be acquired and the iris processing setting is off (YES in S110 and NO in S112), the control unit 10 prompts to turn on the iris processing setting and switches the iris processing setting. Is displayed on the display unit 15 (S114). After the user operation is performed on the warning / setting screen, the process proceeds to step S116.
- step S116 when it is determined that the iris information cannot be acquired (NO in S110), or when it is determined that the iris information can be acquired and the iris processing setting is on (YES in S110 and YES in S112). The process proceeds to step S116.
- step S116 for example, when shooting is performed by a user operation, an input image is acquired by the imaging unit 12, and the region specifying unit 120 specifies an iris region from the acquired input image (S118). Subsequently, the determination unit 140 determines whether or not iris information can be acquired from the iris region specified in step S118 (S120). If it is determined that iris information cannot be acquired (NO in S120), or if it is determined that iris information can be acquired but the iris processing setting is off (YES in S120 and NO in S122), the process ends. .
- the blurring unit 162 of the image processing unit 160 extracts the iris region (an example of the processing region) extracted from the input image. ) Is subjected to a blurring process (S124). Furthermore, the image processing unit 160 synthesizes the result of blurring the iris region with the iris region of the input image, and generates an output image (S126).
- the operation example of the information processing apparatus 1 according to the present embodiment has been specifically described above.
- the operation of the information processing apparatus 1 illustrated in FIG. 3 is an example, and the present embodiment is not limited to the example.
- the same processing as steps S118 to S126 shown in FIG. 3 may be performed using an image acquired from the external device by the communication unit 11 as an input image.
- the image processing unit 160 changes the iris information that can be acquired from the iris region.
- the present technology is not limited to such an example.
- the image processing unit 160 may perform image processing that changes the iris information that can be acquired from the iris region, regardless of the setting by the user. Good.
- the information processing apparatus 1 may not include the determination unit 140.
- the image processing unit 160 may always perform image processing on the iris region specified by the region specifying unit 120.
- Second embodiment >> The first embodiment of the present disclosure has been described above. Subsequently, a second embodiment of the present disclosure will be described. Image processing performed in the second embodiment described below includes processing for changing the value of a randomly selected pixel and edge enhancement processing in addition to the blurring processing described in the first embodiment.
- FIG. 4 is a block diagram illustrating a configuration example of the information processing apparatus 2 according to the second embodiment of the present disclosure.
- the information processing apparatus 2 according to the present embodiment is different from the functional structure of the control unit 10 shown in FIG. And different. 4 that are substantially the same as the components illustrated in FIG. 1 are denoted by the same reference numerals, and thus the description thereof is omitted.
- the function as the image processing part 260 which the control part 20 which concerns on this embodiment has is demonstrated.
- the image processing unit 260 corresponds to the iris region so that the iris information that can be acquired from the iris region specified by the region specifying unit 120 changes. Image processing is performed on the processing area to generate an output image.
- the image processing unit 260 according to the present embodiment has functions as a random deformation unit 264 and an edge enhancement unit 266 in addition to the function as the blurring unit 162.
- functions of the random deformation unit 264 and the edge enhancement unit 266 included in the image processing unit 260 will be described.
- the image processing performed by the image processing unit 260 according to the present embodiment includes the processing by the random transformation unit 264 and the edge enhancement unit 266 described below in addition to the blurring processing by the blurring unit 162 described with reference to FIG. It further includes processing by.
- the random deformation unit 264 changes the value (for example, so-called RGB value) of a pixel selected at random from the pixels of the processing region (for example, iris region).
- the value of the pixel after the change may be a value obtained by adding a predetermined value to the value of the pixel before the change, or may be a value selected at random. Further, the process of changing the pixel value by the random deformation unit 264 may be performed after the blurring process by the blurring unit 162, for example.
- the edge emphasizing unit 266 performs edge emphasis processing for emphasizing an edge (contour) on the processing region.
- the edge enhancement processing by the edge enhancement unit 266 may be, for example, processing that applies an edge enhancement filter. Further, the edge enhancement process by the edge enhancement unit 266 may be performed after the process of changing the pixel value by the random deformation unit 264, for example.
- FIG. 5 is a flowchart illustrating an operation example of the information processing apparatus 2 according to the present embodiment.
- steps S202 to S224 shown in FIG. 5 has been described with reference to FIG. Since it is the same as the processing of steps S102 to S124, the description thereof is omitted.
- the random deformation unit 264 of the image processing unit 260 performs a random deformation process for changing the value of a randomly selected pixel among the pixels in the iris region (an example of the processing region).
- the edge enhancement unit 266 of the image processing unit 260 performs edge enhancement processing for enhancing the edge (contour) on the iris region (S228). Further, the image processing unit 260 synthesizes the result obtained by performing the processing of steps S224 to S228 on the iris region with the iris region of the input image, and generates an output image (S230).
- the information processing apparatus further includes a restoration unit that restores an input image from an output image obtained by performing image processing on the processing region.
- the expression that the input image is restored from the output image is not limited to acquiring an image that completely matches the input image, but includes estimating the input image from the output image.
- FIG. 6 is a block diagram illustrating a configuration example of the information processing apparatus 3 according to the third embodiment of the present disclosure. As shown in FIG. 6, the information processing apparatus 3 according to the present embodiment is different from the functional configuration of the control unit 10 shown in FIG. And different. Note that, among the components illustrated in FIG. 6, components that are substantially the same as the components illustrated in FIG. 1 are denoted by the same reference numerals, and description thereof is omitted. Hereinafter, functions of the image processing unit 360 and the restoration unit 370 included in the control unit 30 according to the present embodiment will be described.
- the image processing unit 360 changes the iris information that can be acquired from the iris region. Image processing is performed on the processing region corresponding to the iris region to generate an output image. Further, the image processing unit 360 according to the present embodiment may cause the storage unit 13 to store parameters for image processing performed on the processing region. Note that the processing area to which the image processing unit 360 according to the present embodiment performs image processing may be, for example, the iris area itself.
- FIG. 7 is an explanatory diagram schematically illustrating an example of the image processing unit 360 according to the present embodiment.
- F (x, y) input to the filter unit 362 illustrated in FIG. 7 indicates an iris region (an example of a processing region) of the input image.
- the transfer function H (u, v) of the filter unit 362 shown in FIG. 7 is, for example, a Fourier transform of the function h (x, y) shown below.
- r represents a radius (a distance from the filter center), and ⁇ 2 represents a dispersion.
- the adding unit 364 adds the noise function ⁇ (x, y) to the output of the filter unit 362, and outputs g (x, y).
- the noise function ⁇ (x, y) represents, for example, white gaussian noise with an average of 0.
- the image processing unit 360 combines g (x, y) with the input image to generate an output image.
- the image processing by the image processing unit 360 is not limited to the above example.
- the image processing by the image processing unit 360 may include, for example, a random deformation process and an edge enhancement process, like the image processing unit 260 described with reference to FIG.
- the restoration unit 370 restores an input image (image data) from an output image obtained by performing image processing on the processing area.
- the restoration processing by the restoration unit 370 may be performed by various methods in accordance with the image processing performed by the image processing unit 360. Further, the restoration unit 370 may acquire the parameters of the image processing performed by the image processing unit 360 from the storage unit 13 and restore the input image based on the parameters.
- the restoration unit 370 may be a filter having a transfer function such as the following equation.
- H (u, v) is a transfer function of the filter unit 362 shown in FIG.
- the restoration unit 370 may be a filter having a transfer function such as the following equation.
- H (u, v) is a transfer function of the filter unit 362 shown in FIG. H * (u, v) is a complex conjugate of H (u, v).
- S ⁇ (u, v) and S f (u, v) are the power spectral densities of ⁇ (x, y) and f (x, y) described with reference to FIG.
- the Fourier transforms of (x, y) and f (x, y) are F (u, v) and N (u, v), they are expressed by the following equations, respectively.
- the restoration unit 370 generates a restored image obtained by restoring the input image, for example, by combining the output result obtained by applying the above-described filter based on the transfer function to the processing region with the output image.
- the processing area to which the restoration unit 370 performs the restoration process may be an area corresponding to the iris area specified by the area specifying unit 120 or an area specified based on the parameters included in the storage unit 13. There may be.
- the user can save or publish the output image that has been subjected to image processing by the image processing unit 360, and further obtain a restored image obtained by restoring the input image from the output image.
- the image processing by the image processing unit 360 and the restoration processing by the restoration unit 370 be concealed (the processing method is not disclosed to a third party). Since the image processing by the image processing unit 360 and the restoration processing by the restoration unit 370 are concealed, it becomes difficult for a third party to restore the input image, and unauthorized authentication can be suppressed.
- FIG. 8 is an explanatory diagram showing an operation example when restoring an input image from an output image in the present embodiment.
- an image (output image) to be restored is selected from images stored in the storage unit 13 by a user operation (S302).
- the area specifying unit 120 specifies an iris area from the image selected in step S302 (S304).
- the restoration unit 370 performs restoration processing on the iris region based on the image processing parameters stored in the storage unit 13 (S306), and combines the restoration result with the iris region of the output image to generate the restored image.
- Generate S308.
- the operation example when restoring the input image from the output image has been specifically described.
- the operation illustrated in FIG. 8 is an example, and the present embodiment is not limited to the example.
- the restoration process may be performed on an image acquired from an external device by the communication unit 11.
- the information processing apparatus includes a restoration unit, so that a restored image obtained by restoring the input image can be obtained from the output image on which image processing has been performed.
- FIG. 9 is a block diagram illustrating a configuration example of the information processing apparatus 4 according to the first modification.
- the information processing apparatus 4 according to this modification is different in part in the functional configuration of the control unit 40 from the functional configuration of the control unit 20 according to the second embodiment shown in FIG. 4.
- the function as the image process part 460 which the control part 40 which concerns on this modification has is demonstrated.
- the image processing unit 460 changes the iris information that can be acquired from the iris region. Image processing is performed on the processing region corresponding to the iris region to generate an output image. As illustrated in FIG. 9, the image processing unit 460 according to the present modification has a function as a replacement unit 464 in addition to functions as a blur unit 162 and an edge enhancement unit 266. Hereinafter, a function of the image processing unit 460 as the replacement unit 464 will be described.
- the image processing performed by the image processing unit 460 according to the present modification includes the replacement unit 464 described below in addition to the blurring processing by the blurring unit 162 and the edge enhancement processing by the edge enhancement unit 266 described with reference to FIG. It further includes a replacement process.
- the replacement unit 464 performs replacement processing for replacing the iris region (an example of the processing region) on the processing region.
- the replacement process may be a process of replacing the iris region with one iris pattern selected from a plurality of iris patterns (an example of a biological pattern) stored in advance in the storage unit 13.
- the iris pattern may be generated in a simulated manner simulating the iris, or may be generated based on the iris of another person who has been imaged.
- the iris pattern may be an iris image or iris pattern information.
- the replacement process performed by the replacement unit 464 may be a process of replacing the values of all pixels included in the iris region with the values of the pixels of the iris pattern.
- the replacement process performed by the replacement unit 464 may be a process of extracting a pattern from the iris region and replacing a pixel corresponding to the extracted pattern with a pixel value of the iris pattern.
- the replacement unit 464 may select an iris pattern at random from a plurality of iris patterns stored in advance in the storage unit 13 to replace the iris region, or select an iris pattern similar to the iris region to select the iris region. May be replaced.
- the configuration example according to this modification has been described above.
- the operation of the information processing apparatus 4 according to the present modification is that the above-described replacement process is performed instead of the random deformation process (S226) between the blurring process (S224) and the edge enhancement process (S228). Since the operation of the information processing apparatus 2 described with reference to FIG.
- the modification 1 has been described above. According to this modification, since the iris region is replaced with another iris pattern by the replacement process, it is possible to change the obtainable iris information and suppress unauthorized authentication.
- the image processing unit 460 performs the blurring process and the edge enhancement process before and after the replacement process has been described.
- the present invention is not limited to such an example.
- the image processing unit 460 may not perform the blurring process or the edge enhancement process.
- the image processing unit 460 may perform other image processing such as random deformation processing.
- FIG. 10 is a block diagram illustrating a configuration example of the information processing apparatus 5 in which the present technology is applied to fingerprint authentication.
- the information processing device 5 is different from the functional configuration of the control unit 20 according to the second embodiment shown in FIG. Different from device 2. 10 that are substantially the same as the components illustrated in FIG. 4 are denoted by the same reference numerals, and thus the description thereof is omitted.
- specification part 520, the feature point detection part 530, the determination part 540, and the image process part 560 which the control part 50 has is demonstrated.
- the area specifying unit 520 specifies a living body area from the input image (image data) acquired by the imaging unit 12 in the same manner as the area specifying unit 120 described with reference to FIG. However, the region specifying unit 520 is different from the region specifying unit 120 in that the finger region for fingerprint authentication is specified as the living body region instead of the iris region. Note that the area specifying unit 520 may specify the entire finger as a finger area for fingerprint authentication, or may specify only the fingertip (which will have a fingerprint). Since the fingerprint is not included when the finger area is the area on the back side of the hand, the area specifying unit 520 may specify only the finger area on the palm side as the finger area for fingerprint authentication. Since the finger region can be specified by a well-known image recognition technique, detailed description thereof is omitted here.
- Information on the finger area specified by the area specifying unit 520 is provided to the feature point detecting unit 530, the determining unit 540, and the image processing unit 560.
- the image data that is the target for specifying the finger region by the region specifying unit 520 is not limited to the image acquired by the imaging unit 12, and may be an image acquired from an external device via the communication unit 11, for example.
- the image may be an image stored in advance in the storage unit 13.
- the feature point detection unit 530 detects feature points from the finger region specified by the region specification unit 520.
- the feature point detected by the feature point detection unit 530 may be a feature point used for fingerprint authentication, and may be, for example, a center point, a branch point, an end point, or a junction point of a fingerprint pattern.
- the feature point detection unit 530 provides information on the detected feature points (for example, the number, position, type, orientation, and the like of the detected feature points) to the determination unit 540 and the image processing unit 560.
- Determination unit 540 determines whether or not fingerprint information (an example of biometric information) used for fingerprint authentication can be acquired from the finger region specified by region specifying unit 520. For example, the determination unit 540 may determine whether or not the fingerprint information can be acquired from the finger region based on the information on the feature points provided from the feature point detection unit 530. For example, the determination unit 540 may determine whether fingerprint information can be acquired from the finger region by determining whether a sufficient number of feature points has been detected.
- fingerprint information an example of biometric information
- the image processing unit 560 changes the biological information that can be acquired from the biological region.
- Image processing is performed on the processing region corresponding to the living body region to generate an output image.
- the image processing unit 560 is configured so that the fingerprint information (an example of biometric information) that can be acquired from the finger region (an example of the biometric region) specified by the region specifying unit 520 changes. Is subjected to image processing to generate an output image.
- the image processing unit 560 included in the information processing apparatus 5 has a function as a feature point transformation unit 564 in addition to functions as a blurring unit 162 and an edge enhancement unit 266.
- the function of the feature point transformation unit 564 that the image processing unit 560 has will be described.
- the image processing performed by the image processing unit 560 includes feature points by the feature point deforming unit 564 described below in addition to the blurring processing by the blurring unit 162 and the edge enhancement processing by the edge enhancement unit 266 described with reference to FIG. Further included is a deformation process.
- the feature point deforming unit 564 performs image processing in which information about feature points for fingerprint authentication detected from the finger region is changed with respect to the finger region (an example of a processing region) (hereinafter referred to as feature point deformation processing). Apply.
- the feature point deforming unit 564 may perform a feature point deforming process based on movement (change of coordinates in the finger region), deletion, or combination of the above feature points detected from the finger region by the feature point detecting unit 530. .
- the feature point deforming unit 564 does not need to perform the feature point deformation process on all detected feature points, and performs the feature point deformation process on some of the detected feature points. You may go.
- the feature point deforming unit 564 sets a feature region around the feature point detected from the finger region by the feature point detection unit 530, moves the feature region according to the movement of the feature point, and moves the feature region around Alternatively, a pseudo fingerprint pattern (stripe) may be generated. Further, the feature point deforming unit 564 sets a feature region around the feature point detected from the finger region by the feature point detecting unit 530, and in a pseudo manner around the feature region and the feature region according to the deletion of the feature point. A fingerprint pattern may be generated. The feature point deforming unit 564 may perform processing such that the fingerprint pattern originally included in the finger area is lost.
- FIGS. 11 and 12 are explanatory diagrams illustrating an example of feature point deformation processing by the feature point deforming unit 564.
- a position P11 and a position P12 indicate the positions of the detected feature points, respectively, and feature regions R11 and R12 are set around the positions P11 and P12, respectively.
- Schematic diagram M10 shown in FIG. 11 schematically shows the movement of the feature points by feature point deforming unit 564.
- the feature point deforming unit 564 moves the feature point at the position P11 to the position P21, and moves the feature point at the position P12 to the position P22.
- a finger region F20 shown in FIG. 11 is an explanatory diagram for explaining the movement of the feature region according to the movement of the feature point by the feature point deforming unit 564 and the generation of a pseudo fingerprint pattern.
- the above-described feature regions R11 and R12 move according to the movement of the feature points at the positions P11 and P12 in the finger region F10, and the feature regions R21 and R22 around the moved positions P21 and P22.
- a pseudo fingerprint pattern stripe is generated by the feature point deforming unit 564 around the feature regions R21 and R22.
- the finger area F10 shown in FIG. 12 is the same as the finger area F10 shown in FIG.
- a schematic diagram M20 illustrated in FIG. 12 schematically illustrates movement and deletion of feature points by the feature point deforming unit 564.
- the feature point deforming unit 564 deletes the feature point at the position P11 and moves the feature point at the position P12 to the position P32.
- the finger region F30 shown in FIG. 12 is an explanatory diagram for explaining the movement of the feature region according to the movement of the feature point by the feature point deforming unit 564 and the generation of a pseudo fingerprint pattern.
- the feature region R11 is also deleted in accordance with the deletion of the feature point at the position P11 in the finger region F10 described above.
- the feature region R12 moves in accordance with the movement of the feature point at the position P12 in the finger region F30 described above, and is indicated as the feature region R32 around the moved position P32.
- a pseudo fingerprint pattern stripe
- the feature point deformation process by the feature point deforming unit 564 has been described above. 11 and 12, an example in which feature points are moved or feature points are moved and deleted has been described as the feature point transformation processing. However, feature points may be deleted even when only feature points are deleted. Included in the transformation process.
- FIG. 13 is an explanatory diagram illustrating an operation example of the information processing apparatus 5 illustrated in FIG. 10.
- the camera is activated by a user operation (S502).
- the region specifying unit 520 specifies a finger region from the image acquired by the imaging unit 12 (S504)
- the feature point detecting unit 530 detects a feature point from the specified finger region (S506).
- the determination unit 540 determines whether fingerprint information can be acquired from the finger region specified in step S506 based on the feature point detected in step S506 (S510).
- the control unit 50 When it is determined that the fingerprint information can be acquired and the fingerprint processing setting is OFF (YES in S510 and NO in S512), the control unit 50 prompts to turn on the fingerprint processing setting and switches the fingerprint processing setting.
- the warning / setting screen is displayed on the display unit 15 (S514). After the user operation is performed on the warning / setting screen, the process proceeds to step S516.
- step S516 when it is determined that fingerprint information cannot be acquired (NO in S510), or when it is determined that fingerprint information can be acquired and the fingerprint processing setting is on (YES in S510 and YES in S512). The process proceeds to step S516.
- step S516 for example, when shooting is performed by a user operation, an input image is acquired by the imaging unit 12, and the region specifying unit 520 specifies a finger region from the acquired input image (S518). Subsequently, the determination unit 540 determines whether or not fingerprint information can be acquired from the finger region specified in step S518 (S520). If it is determined that fingerprint information cannot be acquired (NO in S520), or if it is determined that fingerprint information can be acquired but the fingerprint processing setting is OFF (YES in S520 and NO in S522), the process ends. .
- the blurring unit 562 of the image processing unit 560 causes the finger region extracted from the input image (an example of a processing region) ) Is subjected to a blurring process (S524).
- the feature point deforming unit 564 of the image processing unit 560 performs a feature point deforming process on the finger region (an example of the processing region) such that information on the feature points detected from the finger region is changed (S526).
- the edge enhancement unit 266 of the image processing unit 560 performs edge enhancement processing for enhancing the edge (contour) on the finger region (S528). Further, the image processing unit 560 generates an output image by synthesizing the finger region of the input image with the result of performing the processing of steps S524 to S528 on the finger region (S530).
- the information processing apparatus 5 described above it is possible to suppress unauthorized authentication even when fingerprint authentication is used as biometric authentication by performing image processing that changes fingerprint information that can be acquired. .
- the image processing unit 560 performs the blurring process, the feature point deformation process, and the edge enhancement process has been described.
- the present invention is not limited to such an example.
- the image processing by the image processing unit 560 may include the random deformation process described with reference to FIG. 4 or the replacement process described with reference to FIG.
- the image processing unit 560 performs the replacement process, a plurality of fingerprint patterns are stored in the storage unit 13, and the image processing unit 560 defines the processing area with one fingerprint pattern selected from the plurality of fingerprint patterns. It may be replaced.
- FIG. 14 is a block diagram illustrating a configuration example of the information processing apparatus 6 in which the present technology is applied to vein authentication.
- the information processing apparatus 6 is different from the functional configuration of the control unit 20 according to the second embodiment shown in FIG. Different from device 2.
- components substantially the same as those shown in FIG. 4 are denoted by the same reference numerals, and description thereof is omitted.
- specification part 620, the vein detection part 630, the determination part 640, and the image process part 660 which the control part 60 has is demonstrated.
- the area specifying unit 620 specifies a finger area (an example of a living body area) from the input image (image data) acquired by the imaging unit 12, similarly to the area specifying unit 520 described with reference to FIG. However, the area specifying unit 620 may specify the entire finger as the finger area in order to specify the finger area for vein authentication. Since the finger region can be specified by a well-known image recognition technique, detailed description thereof is omitted here.
- Information on the finger region specified by the region specifying unit 620 is provided to the vein detecting unit 630, the determining unit 640, and the image processing unit 660.
- the image data that is the target for specifying the finger area by the area specifying unit 620 is not limited to the image acquired by the imaging unit 12, and may be an image acquired from an external device via the communication unit 11, for example.
- the image may be an image stored in advance in the storage unit 13.
- the vein detection unit 630 detects a vein from the finger region specified by the region specifying unit 620. Since the vein detection can be realized by a well-known image recognition technique, a detailed description thereof is omitted here.
- the vein detection unit 630 provides information about the detected veins (for example, the number, position, type, orientation, and length of the detected veins) to the determination unit 640 and the image processing unit 660.
- the determining unit 640 determines whether or not the vein information used for vein authentication can be acquired from the finger region specified by the region specifying unit 620. For example, the determination unit 640 may determine whether or not information can be acquired from the finger region based on information regarding the vein provided from the vein detection unit 630. For example, the determination unit 640 may determine whether vein information can be acquired from the finger region by determining whether a sufficient number of veins have been detected.
- the image processing unit 660 changes the biological information that can be acquired from the biological region.
- Image processing is performed on the processing region corresponding to the living body region to generate an output image.
- the image processing unit 660 changes the processing area corresponding to the finger area so that the vein information (an example of the biological information) that can be acquired from the finger area (an example of the biological area) specified by the area specifying unit 620 changes. Is subjected to image processing to generate an output image.
- the image processing unit 660 included in the information processing apparatus 6 has a function as a vein deformation unit 664 in addition to functions as a blurring unit 162 and an edge enhancement unit 266.
- the function of the vein deformation unit 664 included in the image processing unit 660 will be described. Note that the image processing performed by the image processing unit 660 is not limited to the blurring processing by the blurring unit 162 and the edge enhancement processing by the edge enhancement unit 266 described with reference to FIG. Further included.
- the vein deformation unit 664 performs image processing (hereinafter referred to as “venous deformation processing”) on the finger region (an example of the processing region) to change information regarding the vein detected from the finger region.
- the vein deforming unit 664 may perform the vein deforming process by moving the vein detected from the finger region (coordinate change in the finger region), deleting, changing the length, or a combination of the above, detected by the vein detecting unit 630. Good.
- the vein deforming unit 664 does not need to perform vein deformation processing on all detected veins, and may perform vein deformation processing on some of the detected veins.
- FIG. 15 is an explanatory diagram illustrating an example of vein deformation processing by the vein deforming unit 664.
- the vein deforming unit 664 may change the length of the vein V40 like a finger region F41 shown in FIG. Further, the vein deforming unit 664 may move the vein V40 like a finger region F42 shown in FIG. The vein deforming unit 664 may move the vein V40 after inverting or rotating the vein V40 when moving the vein V40. Further, the vein deforming unit 664 may delete the vein V40 as in the finger region F43 illustrated in FIG.
- vein deformation process by the vein deforming unit 664 has been described above.
- vein movement coordinate change in the finger region
- deletion or length change is performed as the vein deformation process, but the above combinations are also included in the vein deformation process.
- FIG. 16 is an explanatory diagram illustrating an operation example of the information processing apparatus 6 illustrated in FIG. 14.
- the camera imaging unit 12
- the region specifying unit 620 specifies a finger region from the image acquired by the imaging unit 12 (S604)
- the vein detecting unit 630 detects a vein from the specified finger region (S606).
- the determination unit 640 determines whether vein information can be acquired from the finger region specified in step S606 based on the vein detected in step S606 (S610).
- the control unit 60 When it is determined that the vein information can be acquired and the vein processing setting is off (YES in S610 and NO in S612), the control unit 60 prompts to turn on the vein processing setting and switches the vein processing setting.
- the warning / setting screen is displayed on the display unit 15 (S614). After the user operation is performed on the warning / setting screen, the process proceeds to step S616.
- step S616 when it is determined that vein information cannot be acquired (NO in S610), or when it is determined that vein information can be acquired and the vein processing setting is on (YES in S610 and YES in S612). The process proceeds to step S616.
- step S616 for example, when shooting is performed by a user operation, an input image is acquired by the imaging unit 12, and the region specifying unit 620 specifies a finger region from the acquired input image (S618). Subsequently, the determination unit 640 determines whether or not vein information can be acquired from the finger area specified in step S618 (S620). If it is determined that vein information cannot be acquired (NO in S620), or if it is determined that vein information can be acquired but the vein processing setting is OFF (YES in S620 and NO in S622), the process ends. .
- the blurring unit 662 of the image processing unit 660 uses the finger region extracted from the input image (an example of the processing region) ) Is subjected to a blurring process (S624). Subsequently, the vein deforming unit 664 of the image processing unit 660 performs a vein deforming process on the finger region (an example of the processing region) such that information related to the vein detected from the finger region is changed (S626).
- the edge enhancement unit 266 of the image processing unit 660 performs edge enhancement processing for enhancing the edge (contour) on the finger region (S628). Further, the image processing unit 660 generates an output image by synthesizing the finger region of the input image with the result of performing the processing of steps S624 to S628 on the finger region (S630).
- the present technology can also be applied to biometric authentication other than iris authentication.
- the combination of the above-described biological region or biological information and image processing is an example, and the present invention is not limited to such an example.
- blurring processing, edge enhancement processing, biological pattern replacement processing, and processing for changing information about feature points are effective for changing biological information regardless of the type of target biological region.
- the random deformation process is particularly effective for changing iris information.
- deletion and movement of feature points are particularly effective for changing fingerprint information or vein information.
- the region specifying unit may be able to specify a plurality of types of biological regions, and the subsequent processing may be performed so as to change the biological information corresponding to the biological region that could be specified by the region specifying unit.
- the region specifying unit attempts to specify all of the iris region, fingerprint region, and vein region from the input image.
- the subsequent feature point detection unit, vein detection unit, determination unit, and image processing unit apply biometric information corresponding to the biological region specified by the region specifying unit to obtain biometric information that can be acquired. It may be changed.
- image processing that changes both fingerprint information and vein information may be performed on an input image in which a finger is shown.
- image process that changes all of the iris information, fingerprint information, and vein information may be performed on an input image in which both fingers and eyes are shown.
- FIG. 17 is a block diagram illustrating a configuration example of the information processing system 1000 according to the present modification.
- the information processing system 1000 according to this modification includes a user terminal 7, a server 8, and a communication network 9.
- the components shown in FIG. 17 components substantially the same as those shown in FIG. 1 are denoted by the same reference numerals, and description thereof is omitted.
- the user terminal 7 is an information processing apparatus including a communication unit 11, an imaging unit 12, a storage unit 13, an operation unit 14, a display unit 15, and a control unit 70, for example, a smartphone or a tablet terminal It may be.
- the functions of the communication unit 11, the imaging unit 12, the storage unit 13, the operation unit 14, and the display unit 15 illustrated in FIG. 17 are the same as those of the communication unit 11, the imaging unit 12, the storage unit 13, and the operation unit described with reference to FIG. 14 and the function of the display unit 15 are omitted.
- the control unit 70 controls each configuration of the user terminal 7.
- the control unit 70 has a communication control function for controlling communication by the communication unit 11 and a display control function for controlling display by the display unit 15.
- the control unit 70 according to the present modification has an encryption function for encrypting an input image acquired by the imaging unit 12 and a decryption function for decrypting information received by the communication unit 11.
- the server 8 is an information processing apparatus including a communication unit 81 and a control unit 80 as shown in FIG.
- the function of the communication unit 81 shown in FIG. 17 is the same as the function of the communication unit 11 described with reference to FIG.
- the control unit 80 controls each component of the server 8.
- the control unit 80 has a communication control function for controlling communication by the communication unit 81, an encryption function for encrypting an image, a decryption function for decrypting information received by the communication unit 11, and the like.
- the server 8 may provide services such as SNS (Social Networking Service), social media, and the like, and has a function of releasing information such as images to other devices via the communication network 9. You may have.
- the control unit 80 has functions as a region specifying unit 820, a determination unit 840, and an image processing unit 860.
- the area specifying unit 820 specifies a biometric area for biometric authentication in the same manner as the area specifying unit 120, the area specifying unit 520, and the area specifying unit 620 described above.
- the function of the area specifying unit 820 may be the same as the function of any of the area specifying unit 120, the area specifying unit 520, and the area specifying unit 620, or may be a combination of the above.
- the determination unit 840 determines whether or not biological information can be acquired from the biological region specified by the region specifying unit 820 in the same manner as the determination unit 140, the determination unit 540, and the determination unit 640 described above. Note that the function of the determination unit 840 may be the same as any of the functions of the determination unit 140, the determination unit 540, and the determination unit 640, or may be a combination of the above.
- the image processing unit 860 stores biological information that can be acquired from the biological region. Image processing is performed on the processing region corresponding to the living body region so as to change.
- the function of the image processing unit 860 may be the same as any of the functions of the image processing unit 160, the image processing unit 260, the image processing unit 360, the image processing unit 460, the image processing unit 560, and the image processing unit 660. It may be a combination of the above. That is, the image processing performed by the image processing unit 860 may include one or more of the above-described blurring processing, edge processing, random deformation processing, replacement processing, feature point deformation processing, and vein deformation processing.
- control unit 80 may have functions as a feature point detection unit and a vein detection unit.
- the communication network 9 is a wired or wireless transmission path for information transmitted from a device connected to the communication network 9.
- the communication network 9 may include a public line network such as the Internet, a telephone line network, a satellite communication network, various LANs including the Ethernet (registered trademark), a WAN (Wide Area Network), and the like.
- the communication network 9 may include a dedicated line network such as an IP-VPN (Internet Protocol-Virtual Private Network).
- FIGS. FIG. 18 and FIG. 19 are flowcharts each showing an operation example of this modification.
- the server 8 has a function of providing services such as SNS (Social Networking Service) and social media, and publishing an image to other devices.
- SNS Social Networking Service
- the control unit 70 of the user terminal 7 encrypts the input image (S704), and the encrypted image is transmitted from the user terminal 7 to the server 8 (S706).
- the control unit 80 of the server 8 Upon receiving the encrypted image, the control unit 80 of the server 8 performs a decryption process and acquires an input image (S708). Subsequently, the area specifying unit 820 of the server 8 specifies a living body area from the input image (S710). Subsequently, the determination unit 840 of the server 8 determines whether or not biological information can be acquired from the biological region specified in step S710 (S712).
- the image processing unit 860 of the server 8 displays an image in the processing region corresponding to the biological region so that the biological information that can be acquired from the biological region changes. Processing is performed (S714). Further, the server 8 publishes the output image obtained by performing the image processing in step S714 to other devices via the communication network 9 (S716).
- the server 8 publishes the input image to another device via the communication network 9 without performing image processing on the input image ( S716).
- the server 8 may not have a function of publishing an image to another device, and the user terminal 7 has an image publishing function after the server 8 performs image processing.
- the operation example in the case of transmitting an image to another server (for example, an SNS server not shown) is shown.
- the processing in steps S802 to S810 shown in FIG. 19 is the same as the processing in steps S702 to S710 described with reference to FIG.
- step S812 the determination unit 840 of the server 8 determines whether or not biological information can be acquired from the biological region identified in step S810.
- the image processing unit 860 of the server 8 displays an image in the processing region corresponding to the biological region so that the biological information that can be acquired from the biological region changes. Processing is performed (S814). Further, the server 8 encrypts the output image obtained by performing the image processing in step S814 (S816).
- the server 8 encrypts the input image without performing image processing on the input image (S816).
- step S816 the image encrypted in step S816 is transmitted from the server 8 to the user terminal 7 (S818), and decrypted by the user terminal 7 (S820). Subsequently, the user terminal 7 transmits the image obtained by decoding in step S820 to an SNS server (not shown) (S822).
- step S812 when it is determined in step S812 that biometric information cannot be acquired, information indicating that image processing is not necessary is transmitted from the server 8 to the user terminal 7, and the user terminal 7 receives the input image obtained in step S802. May be transmitted to an SNS server (not shown).
- Hardware configuration example >> In the above, each embodiment and modification of this indication were explained. Information processing such as the above-described region identification processing, determination processing, image processing, iris registration authentication processing, restoration processing, and the like is realized by cooperation of software and hardware of the information processing apparatuses 1 to 3.
- Information processing such as the above-described region identification processing, determination processing, image processing, iris registration authentication processing, restoration processing, and the like is realized by cooperation of software and hardware of the information processing apparatuses 1 to 3.
- a hardware configuration example of the information processing device 2000 will be described as a hardware configuration example of the information processing devices 1 to 6, the user terminal 7, and the server 8 that are information processing devices according to the present embodiment.
- FIG. 9 is an explanatory diagram illustrating an example of a hardware configuration of the information processing apparatus 2000.
- the information processing apparatus 2000 includes a CPU (Central Processing Unit) 2001, a ROM (Read Only Memory) 2002, a RAM (Random Access Memory) 2003, an input device 2004, an output device 2005, and the like.
- the CPU 2001 functions as an arithmetic processing device and a control device, and controls the overall operation in the information processing device 2000 according to various programs. Further, the CPU 2001 may be a microprocessor.
- the ROM 2002 stores programs used by the CPU 2001, calculation parameters, and the like.
- the RAM 2003 temporarily stores programs used in the execution of the CPU 2001, parameters that change as appropriate during the execution, and the like. These are connected to each other by a host bus composed of a CPU bus or the like. Mainly, the functions of the control unit 10, the control unit 20, the control unit 30, the control unit 40, the control unit 50, the control unit 60, the control unit 70, and the control unit 80 are achieved by the cooperation of the CPU 2001, the ROM 2002, the RAM 2003, and the software. Realized.
- the input device 2004 includes an input means for a user to input information, such as a mouse, a keyboard, a touch panel, a button, a microphone, a switch, and a lever, and an input control circuit that generates an input signal based on the input by the user and outputs the input signal to the CPU 2001 Etc.
- a user of the information processing apparatus 2000 can input various data or instruct a processing operation to the information processing apparatus 2000 by operating the input device 2004.
- the output device 2005 includes, for example, a display device such as a liquid crystal display (LCD) device, an OLED device, a see-through display, and a lamp. Furthermore, the output device 2005 includes an audio output device such as a speaker and headphones. For example, the display device displays a captured image, a generated image, and the like. On the other hand, the audio output device converts audio data or the like into audio and outputs it.
- the output device 2005 corresponds to, for example, the display unit 15 described with reference to FIG.
- the storage device 2006 is a device for storing data.
- the storage device 2006 may include a storage medium, a recording device that records data on the storage medium, a reading device that reads data from the storage medium, a deletion device that deletes data recorded on the storage medium, and the like.
- the storage device 2006 stores programs executed by the CPU 2001 and various data.
- the storage device 2006 corresponds to the storage unit 13 described with reference to FIG.
- the imaging device 2007 includes an imaging optical system such as a photographing lens and a zoom lens that collects light, and a signal conversion element such as a CCD (Charge Coupled Device) or a CMOS (Complementary Metal Oxide Semiconductor).
- the imaging optical system collects light emitted from the subject and forms a subject image in the signal conversion unit, and the signal conversion element converts the formed subject image into an electrical image signal.
- the imaging device 2007 corresponds to the imaging unit 12 described with reference to FIG.
- the communication device 2008 is a communication interface configured by a communication device for connecting to a communication network, for example.
- the communication device 2008 may include a wireless LAN (Local Area Network) compatible communication device, an LTE (Long Term Evolution) compatible communication device, a wire communication device that performs wired communication, or a Bluetooth (registered trademark) communication device.
- the communication device 2008 corresponds to, for example, the communication unit 11 described with reference to FIG. 1 and the communication unit 81 described with reference to FIG.
- each step in the above-described embodiment does not necessarily have to be processed in time series in the order described as a flowchart.
- each step in the processing of the above embodiment may be processed in an order different from the order described as the flowchart diagram or may be processed in parallel.
- a computer program for causing hardware such as the CPU 2001, the ROM 2002, and the RAM 2003 to perform the same functions as the configurations of the information processing apparatuses 1 to 6, the user terminal 7, and the server 8 described above. Can also be provided.
- a recording medium on which the computer program is recorded is also provided.
- An area specifying unit for specifying a biometric area for biometric authentication from image data An image processing unit that performs image processing on a processing area corresponding to the biological area so that biological information that can be acquired from the biological area changes;
- An information processing apparatus comprising: (2) The information processing apparatus according to (1), wherein the image processing performed by the image processing unit includes blurring processing. (3) The information processing apparatus according to (2), wherein the image processing performed by the image processing unit further includes an edge enhancement process. (4) The information processing apparatus according to (3), wherein the image processing performed by the image processing unit further includes a process of changing a value of a randomly selected pixel.
- the information processing apparatus includes a replacement process for replacing the processing region.
- the replacement process is a process of replacing the processing area with one biological pattern selected from a plurality of biological patterns stored in advance.
- the image processing performed by the image processing unit includes a feature point deformation process that changes information about a feature point detected from the living body region. apparatus.
- the feature point transformation process includes at least one of movement and deletion of feature points detected from the living body region.
- the information processing apparatus includes vein deformation processing for changing information related to a vein detected from the living body region.
- the vein deformation process includes at least one of movement, deletion, and length change of the vein.
- the information processing apparatus further includes a determination unit that determines whether the biological information can be acquired from the biological region, The image processing unit performs the image processing on the processing region when the determination unit determines that the biological information can be acquired from the biological region, any one of (1) to (10) The information processing apparatus according to one item.
- the information processing apparatus includes: A determination unit for determining whether or not biological information can be acquired from the biological region specified by the region specifying unit; The information processing apparatus according to any one of (1) to (10), wherein a warning screen is displayed when the determination unit determines that the biological information can be acquired from the biological region. (13)
- the information processing apparatus further includes a communication unit that communicates with the outside, The determination unit determines whether or not the biological information can be acquired from a biological region identified from the image data when an operation input for transmitting image data to the outside is performed.
- the information processing apparatus according to 12).
- the determination unit according to any one of (11) to (13), wherein the determination unit determines whether the biological information can be acquired from the biological region based on a size of the biological region. Information processing device.
- the determination unit determines whether the biological information can be acquired from the biological region based on a color included in the biological region.
- Information processing device (16) The information processing apparatus according to any one of (1) to (15), further including a restoration unit that restores the image data from an image on which the image processing has been performed on the processing region. apparatus. (17) The information processing apparatus according to (16), wherein the restoration unit restores the image data based on the image processing parameters. (18) The information processing apparatus according to any one of (1) to (17), wherein the biological information includes at least one of iris information, fingerprint information, and vein information.
- An information processing method including: (20) On the computer, A function for identifying a biometric area for biometric authentication from image data; A function of performing image processing on a processing area corresponding to the biological area so that biological information that can be acquired from the biological area changes; A program to realize
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Human Computer Interaction (AREA)
- Ophthalmology & Optometry (AREA)
- Collating Specific Patterns (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Description
<<1.第一の実施形態>>
<1-1.概要>
<1-2.構成例>
<1-3.動作例>
<1-4.効果>
<1-5.補足>
<<2.第二の実施形態>>
<2-1.構成例>
<2-2.動作例>
<2-3.効果>
<<3.第三の実施形態>>
<3-1.構成例>
<3-2.動作例>
<3-3.効果>
<3-4.補足>
<<4.変形例>>
<4-1.変形例1>
<4-2.変形例2>
<4-3.変形例3>
<<5.ハードウェア構成例>>
<<6.むすび>>
<1-1.背景>
本開示の第一の実施形態に係る情報処理装置について説明する前に、まず、本実施形態による情報処理装置の創作に至った背景を説明する。
図1は、本開示の第一の実施形態に係る情報処理装置1の構成例を示す説明図である。図1に示すように、本実施形態に係る情報処理装置1は、制御部10、通信部11、撮像部12、記憶部13、操作部14、及び表示部15を備える。
続いて、本実施形態による情報処理装置1の動作例について、図3を参照して説明する。図3は、本実施形態による情報処理装置1の動作例を示すフローチャートである。
以上説明したように、本開示の第一の実施形態によれば、取得可能な虹彩情報を変化させる画像処理を施すことで、不正認証を抑制することが可能である。
なお、上記では、判定部140により虹彩領域から虹彩情報が取得可能と判定され、かつ虹彩処理設定がオンの場合に、画像処理部160が、虹彩領域から取得可能な虹彩情報を変化させる画像処理を施す例を説明したが、本技術は係る例に限定されない。
以上、本開示の第一の実施形態を説明した。続いて、本開示の第二の実施形態を説明する。以下に説明する第二の実施形態で施される画像処理は、第一の実施形態で説明したぼかし処理に加え、ランダムに選択されるピクセルの値を変更する処理、及びエッジ強調処理を含む。
図4は、本開示の第二の実施形態に係る情報処理装置2の構成例を示すブロック図である。図4に示すように、本実施形態に係る情報処理装置2は、制御部20の機能構成が図1に示した制御部10の機能構成と一部異なる点で、図1の情報処理装置1と異なる。なお、図4に示す各構成のうち、図1に示した各構成と実質的に同様の構成については、同一の符号を付してあるため、説明を省略する。以下では、本実施形態に係る制御部20が有する画像処理部260としての機能について説明する。
以上、本開示の第二の実施形態に係る情報処理装置2の構成例を説明した。続いて、本実施形態による情報処理装置2の動作例について、図5を参照して説明する。図5は、本実施形態による情報処理装置2の動作例を示すフローチャートである。
以上説明したように、本開示の第二の実施形態によれば、取得可能な虹彩情報を変化させる画像処理を施すことで、不正認証を抑制することが可能である。また、ぼかし処理とランダム変形処理を施した後にエッジ強調処理を施すことで、取得可能な虹彩情報を変化させつつ、ユーザに与える違和感を小さくすることが可能となる。
以上、本開示の第一の実施形態、及び第二の実施形態を説明した。続いて、本開示の第三の実施形態として、情報処理装置が、処理領域に画像処理が施された出力画像から入力画像を復元する復元部をさらに備える例を説明する。なお、本実施形態において、出力画像から入力画像を復元する、という表現は、入力画像と完全に一致する画像を取得することに限定されず、出力画像から入力画像を推定することを含む。
図6は、本開示の第三の実施形態に係る情報処理装置3の構成例を示すブロック図である。図6に示すように、本実施形態に係る情報処理装置3は、制御部30の機能構成が図1に示した制御部10の機能構成と一部異なる点で、図1の情報処理装置1と異なる。なお、図6に示す各構成のうち、図1に示した各構成と実質的に同様の構成については、同一の符号を付してあるため、説明を省略する。以下では、本実施形態に係る制御部30が有する画像処理部360、及び復元部370としての機能について説明する。
以上、本開示の第三の実施形態に係る情報処理装置3の構成例を説明した。続いて、本実施形態による情報処理装置3の動作例について説明する。なお、本実施形態において、入力画像に画像処理を施す際の動作例は、例えばステップS124の代わりに図7を参照して説明した画像処理が行われる点を除いて、図3を参照して説明した第一の実施形態の動作例と同様であるため、説明を省略する。
以上説明したように、本開示の第三の実施形態によれば、取得可能な虹彩情報を変化させる画像処理を施すことで、不正認証を抑制することが可能である。また、本実施形態による情報処理装置は、復元部を備えることで、画像処理が施された出力画像から、入力画像を復元した復元画像を得ることも可能である。
なお、上記では、情報処理装置3が画像処理部と復元部の両方の機能を有する例を説明したが、本技術は係る例に限定されない。例えば、画像処理部と復元部の機能は、異なる装置に備えられてもよい。また、係る場合、画像処理部を備える装置から復元部を備える装置に、画像処理のパラメータが提供されてもよい。
以上、本開示の各実施形態を説明した。以下では、本開示の各実施形態の幾つかの変形例を説明する。なお、以下に説明する各変形例は、単独で本開示の各実施形態に適用されてもよいし、組み合わせで本開示の各実施形態に適用されてもよい。また、各変形例は、本開示の各実施形態で説明した構成に代えて適用されてもよいし、本開示の各実施形態で説明した構成に対して追加的に適用されてもよい。
上記実施形態では、画像処理部が処理領域に施す画像処理の例として、ぼかし処理、エッジ強調処理、ランダム変形処理について説明したが、本技術は係る例に限定されない。画像処理部は処理領域に他の画像処理を施してもよい。以下では、変形例1として、画像処理部が他の画像処理を施す例を説明する。
上記実施形態では、虹彩認証用の虹彩領域から取得可能な虹彩情報が変化するように画像処理を施す例について説明したが、本技術は虹彩認証以外の生体認証にも広く適用することが可能である。つまり、上述した虹彩認証は生体認証の一例であり、虹彩領域は生体認証用の生体領域の一例であると解される。以下では、変形例2として、他の生体認証への適用例について説明する。
まず、本技術を指紋認証へ適用する例について説明する。図10は、本技術を指紋認証へ適用した情報処理装置5の構成例を示すブロック図である。図10に示すように、情報処理装置5は、制御部50の機能構成が図4に示した第2の実施形態に係る制御部20の機能構成と一部異なる点で、図4の情報処理装置2と異なる。なお、図10に示す各構成のうち、図4に示した各構成と実質的に同様の構成については、同一の符号を付してあるため、説明を省略する。以下では、制御部50が有する領域特定部520、特徴点検出部530、判定部540、及び画像処理部560としての機能について説明する。
続いて、本技術を静脈認証(生体認証の一例)へ適用する例について説明する。図14は、本技術を静脈認証へ適用した情報処理装置6の構成例を示すブロック図である。図14に示すように、情報処理装置6は、制御部60の機能構成が図4に示した第2の実施形態に係る制御部20の機能構成と一部異なる点で、図4の情報処理装置2と異なる。なお、図14に示す各構成のうち、図4に示した各構成と実質的に同様の構成については、同一の符号を付してあるため、説明を省略する。以下では、制御部60が有する領域特定部620、静脈検出部630、判定部640、及び画像処理部660としての機能について説明する。
上記実施形態では、画像の取得(撮像)を行う情報処理装置が、生体領域の特定、生体領域への画像処理等を行う例を説明したが、本技術は係る例に限定されない。例えば、画像の取得を行う装置と、生体領域の特定、生体領域への画像処理等を行う装置は異なる装置であってもよく、以下では変形例3として、係る例について説明する。
以上、本開示の各実施形態と変形例を説明した。上述した領域特定処理、判定処理、画像処理、虹彩登録認証処理、復元処理等の情報処理は、ソフトウェアと、情報処理装置1~3のハードウェアとの協働により実現される。以下では、本実施形態に係る情報処理装置である情報処理装置1~6、ユーザ端末7、サーバ8のハードウェア構成例として、情報処理装置2000のハードウェア構成例について説明する。
以上、添付図面を参照しながら本開示の好適な実施形態について詳細に説明したが、本開示の技術的範囲はかかる例に限定されない。本開示の技術分野における通常の知識を有する者であれば、特許請求の範囲に記載された技術的思想の範疇内において、各種の変更例または修正例に想到し得ることは明らかであり、これらについても、当然に本開示の技術的範囲に属するものと了解される。
(1)
画像データから生体認証用の生体領域を特定する領域特定部と、
前記生体領域から取得可能な生体情報が変化するように、前記生体領域に応じた処理領域に画像処理を施す画像処理部と、
を備える情報処理装置。
(2)
前記画像処理部が施す前記画像処理は、ぼかし処理を含む、前記(1)に記載の情報処理装置。
(3)
前記画像処理部が施す前記画像処理は、エッジ強調処理をさらに含む、前記(2)に記載の情報処理装置。
(4)
前記画像処理部が施す前記画像処理は、ランダムに選択されるピクセルの値を変更する処理をさらに含む、前記(3)に記載の情報処理装置。
(5)
前記画像処理部が施す前記画像処理は、前記処理領域を置き換える置換処理を含む、前記(1)~(4)のいずれか一項に記載の情報処理装置。
(6)
前記置換処理は、予め記憶された複数の生体パターンから選択された1の生体パターンで前記処理領域を置き換える処理である、前記(5)に記載の情報処理装置。
(7)
前記画像処理部が施す前記画像処理は、前記生体領域から検出される特徴点に関する情報を変化させる特徴点変形処理を含む、前記(1)~(6)のいずれか一項に記載の情報処理装置。
(8)
前記特徴点変形処理は、前記生体領域から検出された特徴点の移動、削除のうち、少なくともいずれか一つを含む、前記(7)に記載の情報処理装置。
(9)
前記画像処理部が施す前記画像処理は、前記生体領域から検出される静脈に関する情報を変化させる静脈変形処理を含む、前記(1)~(8)のいずれか一項に記載の情報処理装置。
(10)
前記静脈変形処理は、前記静脈の移動、削除、長さの変更のうち、少なくともいずれか一つを含む、前記(9)に記載の情報処理装置。
(11)
前記情報処理装置は、前記生体領域から前記生体情報を取得可能であるか否かを判定する判定部をさらに備え、
前記画像処理部は、前記判定部により前記生体領域から前記生体情報を取得可能であると判定された場合に、前記処理領域に前記画像処理を施す、前記(1)~(10)のいずれか一項に記載の情報処理装置。
(12)
前記情報処理装置は、
前記領域特定部により特定された前記生体領域から生体情報を取得可能であるか否かを判定する判定部をさらに備え、
前記判定部により前記生体領域から前記生体情報を取得可能であると判定された場合に、警告画面を表示させる、前記(1)~(10)のいずれか一項に記載の情報処理装置。
(13)
前記情報処理装置は、外部と通信を行う通信部をさらに備え、
前記判定部は、画像データを外部に送信させるための操作入力が行われた場合に、前記画像データから特定された生体領域から前記生体情報を取得可能であるか否かを判定する、前記(12)に記載の情報処理装置。
(14)
前記判定部は、前記生体領域の大きさに基づいて、前記生体領域から前記生体情報を取得可能であるか否かを判定する、前記(11)~(13)のいずれか一項に記載の情報処理装置。
(15)
前記判定部は、前記生体領域に含まれる色に基づいて、前記生体領域から前記生体情報を取得可能であるか否かを判定する、前記(11)~(14)のいずれか一項に記載の情報処理装置。
(16)
前記情報処理装置は、前記処理領域に前記画像処理が施された画像から、前記画像データを復元する復元部をさらに備える、前記(1)~(15)のいずれか一項に記載の情報処理装置。
(17)
前記復元部は、前記画像処理のパラメータに基づいて、前記画像データを復元する、前記(16)に記載の情報処理装置。
(18)
前記生体情報は、虹彩情報、指紋情報、静脈情報のうち、少なくともいずれか一つを含む、前記(1)~(17)のいずれか一項に記載の情報処理装置。
(19)
画像データから生体認証用の生体領域を特定することと、
前記生体領域から取得可能な生体情報が変化するように、前記生体領域に応じた処理領域に画像処理を施すことと、
を含む情報処理方法。
(20)
コンピュータに、
画像データから生体認証用の生体領域を特定する機能と、
前記生体領域から取得可能な生体情報が変化するように、前記生体領域に応じた処理領域に画像処理を施す機能と、
を実現させるための、プログラム。
7 ユーザ端末
8 サーバ
9 通信網
10、20、30、40、50、60、70、80 制御部
11 通信部
12 撮像部
13 記憶部
14 操作部
15 表示部
120、520、620、820 領域特定部
140、540、640、840 判定部
160、260、360、460、560、660、860 画像処理部
162、562、662 ぼかし部
180 虹彩登録認証部
264 ランダム変形部
266 エッジ強調部
362 フィルタ部
364 加算部
370 復元部
464 置換部
530 特徴点検出部
564 特徴点変形部
630 静脈検出部
664 静脈変形部
Claims (20)
- 画像データから生体認証用の生体領域を特定する領域特定部と、
前記生体領域から取得可能な生体情報が変化するように、前記生体領域に応じた処理領域に画像処理を施す画像処理部と、
を備える情報処理装置。 - 前記画像処理部が施す前記画像処理は、ぼかし処理を含む、請求項1に記載の情報処理装置。
- 前記画像処理部が施す前記画像処理は、エッジ強調処理をさらに含む、請求項2に記載の情報処理装置。
- 前記画像処理部が施す前記画像処理は、ランダムに選択されるピクセルの値を変更する処理をさらに含む、請求項3に記載の情報処理装置。
- 前記画像処理部が施す前記画像処理は、前記処理領域を置き換える置換処理を含む、請求項1に記載の情報処理装置。
- 前記置換処理は、予め記憶された複数の生体パターンから選択された1の生体パターンで前記処理領域を置き換える処理である、請求項5に記載の情報処理装置。
- 前記画像処理部が施す前記画像処理は、前記生体領域から検出される特徴点に関する情報を変化させる特徴点変形処理を含む、請求項1に記載の情報処理装置。
- 前記特徴点変形処理は、前記生体領域から検出された特徴点の移動、削除のうち、少なくともいずれか一つを含む、請求項7に記載の情報処理装置。
- 前記画像処理部が施す前記画像処理は、前記生体領域から検出される静脈に関する情報を変化させる静脈変形処理を含む、請求項1に記載の情報処理装置。
- 前記静脈変形処理は、前記静脈の移動、削除、長さの変更のうち、少なくともいずれか一つを含む、請求項9に記載の情報処理装置。
- 前記情報処理装置は、前記生体領域から前記生体情報を取得可能であるか否かを判定する判定部をさらに備え、
前記画像処理部は、前記判定部により前記生体領域から前記生体情報を取得可能であると判定された場合に、前記処理領域に前記画像処理を施す、請求項1に記載の情報処理装置。 - 前記情報処理装置は、
前記領域特定部により特定された前記生体領域から生体情報を取得可能であるか否かを判定する判定部をさらに備え、
前記判定部により前記生体領域から前記生体情報を取得可能であると判定された場合に、警告画面を表示させる、請求項1に記載の情報処理装置。 - 前記情報処理装置は、外部と通信を行う通信部をさらに備え、
前記判定部は、画像データを外部に送信させるための操作入力が行われた場合に、前記画像データから特定された生体領域から前記生体情報を取得可能であるか否かを判定する、請求項12に記載の情報処理装置。 - 前記判定部は、前記生体領域の大きさに基づいて、前記生体領域から前記生体情報を取得可能であるか否かを判定する、請求項11に記載の情報処理装置。
- 前記判定部は、前記生体領域に含まれる色に基づいて、前記生体領域から前記生体情報を取得可能であるか否かを判定する、請求項11に記載の情報処理装置。
- 前記情報処理装置は、前記処理領域に前記画像処理が施された画像から、前記画像データを復元する復元部をさらに備える、請求項1に記載の情報処理装置。
- 前記復元部は、前記画像処理のパラメータに基づいて、前記画像データを復元する、請求項16に記載の情報処理装置。
- 前記生体情報は、虹彩情報、指紋情報、静脈情報のうち、少なくともいずれか一つを含む、請求項1に記載の情報処理装置。
- 画像データから生体認証用の生体領域を特定することと、
前記生体領域から取得可能な生体情報が変化するように、前記生体領域に応じた処理領域に画像処理を施すことと、
を含む情報処理方法。 - コンピュータに、
画像データから生体認証用の生体領域を特定する機能と、
前記生体領域から取得可能な生体情報が変化するように、前記生体領域に応じた処理領域に画像処理を施す機能と、
を実現させるための、プログラム。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201780023238.8A CN108885695B (zh) | 2016-04-21 | 2017-04-12 | 信息处理设备、信息处理方法和程序 |
JP2018513135A JP7075703B2 (ja) | 2016-04-21 | 2017-04-12 | 情報処理装置、情報処理方法、及びプログラム |
CN202210133338.9A CN114267081A (zh) | 2016-04-21 | 2017-04-12 | 信息处理设备和程序 |
CN202210119717.2A CN114241593A (zh) | 2016-04-21 | 2017-04-12 | 信息处理设备和程序 |
EP17785882.6A EP3447724A1 (en) | 2016-04-21 | 2017-04-12 | Information processing device, information processing method, and program |
US16/094,887 US11810399B2 (en) | 2016-04-21 | 2017-04-12 | Information processing device, information processing method, and program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016-084969 | 2016-04-21 | ||
JP2016084969 | 2016-04-21 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017183543A1 true WO2017183543A1 (ja) | 2017-10-26 |
Family
ID=60116061
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2017/015038 WO2017183543A1 (ja) | 2016-04-21 | 2017-04-12 | 情報処理装置、情報処理方法、及びプログラム |
Country Status (5)
Country | Link |
---|---|
US (1) | US11810399B2 (ja) |
EP (1) | EP3447724A1 (ja) |
JP (1) | JP7075703B2 (ja) |
CN (3) | CN114267081A (ja) |
WO (1) | WO2017183543A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018191075A (ja) * | 2017-04-28 | 2018-11-29 | キヤノンマーケティングジャパン株式会社 | 画像処理装置、画像処理装置の制御方法、およびプログラム |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS5831478A (ja) * | 1981-08-19 | 1983-02-24 | Hitachi Ltd | 指紋照合装置 |
WO2007060980A1 (ja) * | 2005-11-25 | 2007-05-31 | Nikon Corporation | 電子カメラおよび画像処理装置 |
JP2008046677A (ja) * | 2006-08-10 | 2008-02-28 | Fujitsu Ltd | 生体情報読取装置、生体情報読取方法および生体情報読取プログラム |
WO2010058514A1 (ja) * | 2008-11-19 | 2010-05-27 | 日本電気株式会社 | 認証システム、装置、認証方法及びプログラムが格納された記憶媒体 |
Family Cites Families (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3768225B2 (ja) * | 2003-04-02 | 2006-04-19 | 松下電器産業株式会社 | 画像処理方法、画像処理装置、撮影装置および画像出力装置 |
DE102004017970A1 (de) | 2004-04-12 | 2005-10-27 | Leica Microsystems (Schweiz) Ag | Stativ, insbesondere für Operationsmikroskope, mit einem Kraftspeicherelement |
JP4664147B2 (ja) | 2005-07-29 | 2011-04-06 | 株式会社山武 | 虹彩認証装置 |
JP2008123206A (ja) * | 2006-11-10 | 2008-05-29 | Sony Corp | 登録装置、照合装置、登録方法、照合方法及びプログラム |
KR100905675B1 (ko) * | 2007-08-13 | 2009-07-03 | 한국전자통신연구원 | 지문인식 장치 및 방법 |
US9036871B2 (en) * | 2007-09-01 | 2015-05-19 | Eyelock, Inc. | Mobility identity platform |
US8411910B2 (en) * | 2008-04-17 | 2013-04-02 | Biometricore, Inc. | Computationally efficient feature extraction and matching iris recognition |
JP2010020670A (ja) | 2008-07-14 | 2010-01-28 | Panasonic Corp | 認証装置 |
US9633261B2 (en) * | 2008-08-22 | 2017-04-25 | International Business Machines Corporation | Salting system and method for cancelable iris biometric |
US8290219B2 (en) * | 2008-08-22 | 2012-10-16 | International Business Machines Corporation | Registration-free transforms for cancelable iris biometrics |
WO2010125653A1 (ja) * | 2009-04-28 | 2010-11-04 | 富士通株式会社 | 生体認証装置、生体認証方法および生体認証プログラム |
JP2010287023A (ja) | 2009-06-11 | 2010-12-24 | Seiko Epson Corp | 情報表示装置 |
US8712157B2 (en) * | 2011-04-19 | 2014-04-29 | Xerox Corporation | Image quality assessment |
US20130259322A1 (en) * | 2012-03-31 | 2013-10-03 | Xiao Lin | System And Method For Iris Image Analysis |
US8842888B2 (en) * | 2012-06-15 | 2014-09-23 | Aoptix Technologies, Inc. | User interface for combined biometric mobile device |
KR102206877B1 (ko) | 2014-02-21 | 2021-01-26 | 삼성전자주식회사 | 생체 정보디스플레이 방법 및 장치 |
US9239957B2 (en) * | 2014-04-04 | 2016-01-19 | Fotonation Limited | Image processing method and apparatus |
CN104021350B (zh) | 2014-05-13 | 2016-07-06 | 小米科技有限责任公司 | 隐私信息隐藏方法及装置 |
EP3065024A1 (en) * | 2015-03-06 | 2016-09-07 | Universidad de Sevilla | Wearable camera apparatus with selectable processing of image content |
CN104778417A (zh) | 2015-03-20 | 2015-07-15 | 惠州Tcl移动通信有限公司 | 防止个人指纹信息通过照片泄露的方法、移动终端 |
CN104809744B (zh) | 2015-04-29 | 2017-09-22 | 小米科技有限责任公司 | 图片处理方法和装置 |
-
2017
- 2017-04-12 US US16/094,887 patent/US11810399B2/en active Active
- 2017-04-12 JP JP2018513135A patent/JP7075703B2/ja active Active
- 2017-04-12 EP EP17785882.6A patent/EP3447724A1/en active Pending
- 2017-04-12 CN CN202210133338.9A patent/CN114267081A/zh active Pending
- 2017-04-12 CN CN201780023238.8A patent/CN108885695B/zh active Active
- 2017-04-12 CN CN202210119717.2A patent/CN114241593A/zh active Pending
- 2017-04-12 WO PCT/JP2017/015038 patent/WO2017183543A1/ja active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS5831478A (ja) * | 1981-08-19 | 1983-02-24 | Hitachi Ltd | 指紋照合装置 |
WO2007060980A1 (ja) * | 2005-11-25 | 2007-05-31 | Nikon Corporation | 電子カメラおよび画像処理装置 |
JP2008046677A (ja) * | 2006-08-10 | 2008-02-28 | Fujitsu Ltd | 生体情報読取装置、生体情報読取方法および生体情報読取プログラム |
WO2010058514A1 (ja) * | 2008-11-19 | 2010-05-27 | 日本電気株式会社 | 認証システム、装置、認証方法及びプログラムが格納された記憶媒体 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018191075A (ja) * | 2017-04-28 | 2018-11-29 | キヤノンマーケティングジャパン株式会社 | 画像処理装置、画像処理装置の制御方法、およびプログラム |
Also Published As
Publication number | Publication date |
---|---|
EP3447724A4 (en) | 2019-02-27 |
JP7075703B2 (ja) | 2022-05-26 |
US11810399B2 (en) | 2023-11-07 |
US20190130179A1 (en) | 2019-05-02 |
CN108885695A (zh) | 2018-11-23 |
EP3447724A1 (en) | 2019-02-27 |
CN114267081A (zh) | 2022-04-01 |
CN114241593A (zh) | 2022-03-25 |
JPWO2017183543A1 (ja) | 2019-02-28 |
CN108885695B (zh) | 2022-03-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110321790B (zh) | 一种对抗样本的检测方法及电子设备 | |
CN108197586B (zh) | 脸部识别方法和装置 | |
US11074466B2 (en) | Anti-counterfeiting processing method and related products | |
JP7266672B2 (ja) | 画像処理方法および画像処理装置、ならびにデバイス | |
CN107408205B (zh) | 用红外成像区分前景和背景 | |
JP7026225B2 (ja) | 生体検出方法、装置及びシステム、電子機器並びに記憶媒体 | |
WO2021031609A1 (zh) | 活体检测方法及装置、电子设备和存储介质 | |
WO2019052329A1 (zh) | 人脸识别方法及相关产品 | |
CN109711546B (zh) | 神经网络训练方法及装置、电子设备和存储介质 | |
EP3481058A1 (en) | Electronic apparatus and communication method thereof | |
EP3513326B1 (en) | Methods, systems, and media for detecting stereoscopic videos by generating fingerprints for multiple portions of a video frame | |
CN111243105B (zh) | 增强现实处理方法及装置、存储介质和电子设备 | |
CN110532957B (zh) | 人脸识别方法及装置、电子设备和存储介质 | |
CN110111281A (zh) | 图像处理方法及装置、电子设备和存储介质 | |
WO2013056431A1 (en) | Methods and apparatuses for gesture recognition | |
CN114332975A (zh) | 识别利用模拟覆盖物部分覆盖的对象 | |
CN110225331B (zh) | 选择性地将色彩施加到图像 | |
CN105426904B (zh) | 照片处理方法、装置和设备 | |
JP6736534B2 (ja) | 通信中継装置、システム、方法及びプログラム | |
US20140055551A1 (en) | Image processing method and apparatus for personal protection in video call | |
WO2017183543A1 (ja) | 情報処理装置、情報処理方法、及びプログラム | |
WO2018092378A1 (ja) | 情報処理装置、情報処理方法、及びプログラム | |
CN111656776B (zh) | 图像处理器、图像处理方法和计算机可读记录介质 | |
EP3217644B1 (en) | Information processing device | |
WO2021136979A1 (en) | Image processing method and apparatus, electronic device, and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 2018513135 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2017785882 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2017785882 Country of ref document: EP Effective date: 20181121 |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17785882 Country of ref document: EP Kind code of ref document: A1 |