WO2014103732A1 - 画像処理装置および画像処理方法、並びにプログラム - Google Patents
画像処理装置および画像処理方法、並びにプログラム Download PDFInfo
- Publication number
- WO2014103732A1 WO2014103732A1 PCT/JP2013/083297 JP2013083297W WO2014103732A1 WO 2014103732 A1 WO2014103732 A1 WO 2014103732A1 JP 2013083297 W JP2013083297 W JP 2013083297W WO 2014103732 A1 WO2014103732 A1 WO 2014103732A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- unit
- line
- user
- tracking
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 98
- 238000003672 processing method Methods 0.000 title claims abstract description 7
- 238000001514 detection method Methods 0.000 claims abstract description 92
- 238000000034 method Methods 0.000 claims abstract description 58
- 238000003384 imaging method Methods 0.000 claims description 109
- 230000000007 visual effect Effects 0.000 claims description 5
- 238000004148 unit process Methods 0.000 claims description 2
- 230000003287 optical effect Effects 0.000 abstract description 19
- 210000001508 eye Anatomy 0.000 abstract description 9
- 230000033001 locomotion Effects 0.000 description 19
- 238000005516 engineering process Methods 0.000 description 17
- 238000004891 communication Methods 0.000 description 12
- 230000006870 function Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 5
- 210000001747 pupil Anatomy 0.000 description 5
- 238000005401 electroluminescence Methods 0.000 description 4
- 230000008929 regeneration Effects 0.000 description 4
- 238000011069 regeneration method Methods 0.000 description 4
- 239000004065 semiconductor Substances 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 230000004424 eye movement Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 210000005252 bulbus oculi Anatomy 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 125000002066 L-histidyl group Chemical group [H]N1C([H])=NC(C([H])([H])[C@](C(=O)[*])([H])N([H])[H])=C1[H] 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000004434 saccadic eye movement Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/53—Constructional details of electronic viewfinders, e.g. rotatable or detachable
- H04N23/531—Constructional details of electronic viewfinders, e.g. rotatable or detachable being rotatable or detachable
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/633—Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
- H04N23/635—Region indicators; Field of view indicators
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
Definitions
- the present technology relates to an image processing device, an image processing method, and a program, and more particularly, to an image processing device, an image processing method, and a program that can more appropriately capture or display an image of an object that is watched by a user.
- Assist technology has been proposed to assist users in capturing or displaying images.
- the predetermined object to be controlled may be different from the subject intended by the user, and the image is processed in a state where image processing based on the control target not intended by the user is performed.
- the image was sometimes captured or displayed.
- the present technology has been made in view of such a situation, and in particular, by detecting the direction of the user's line of sight, the subject intended by the user is accurately captured as a control target in the image, and the captured control is performed.
- the captured control is performed.
- An image processing apparatus is based on a gaze direction detection unit that detects a gaze direction of a user with respect to an image, a gaze direction detected by the gaze direction detection unit, and the image. And a tracking object detection unit that detects a tracking object in the image that is tracked by the user's line of sight based on a time-series gaze area estimated by the estimation unit.
- a tracking unit that searches for and tracks the tracking object detected by the tracking object detection unit, and an image control unit that controls an image of the tracking object tracked by the tracking unit.
- the tracking object detection unit is configured to detect a face image as a tracking object in the image that is being tracked by the user's line of sight based on the time-series gaze region estimated by the estimation unit. can do.
- the tracking object detection unit is configured to detect a specific object as a tracking object in the image that is tracked by the user's line of sight based on the time-series gaze region estimated by the estimation unit. can do.
- the image processing device may be a glasses-type imaging device, and further includes an imaging unit that captures an image having a field of view that can be viewed by the user when the glasses-type imaging device is put on.
- the gaze direction detection unit detects the gaze direction of the user with respect to an image having a field of view that can be seen by the user when the spectacle-type imaging device is put on, and the image
- the control unit can be controlled so that the image of the tracking object tracked by the tracking unit can be optimally captured by the imaging unit.
- the image processing device may be an image pickup device that is used for image pickup by the user and further includes an image pickup unit that picks up the image and a viewfinder that looks into the image picked up by the image pickup unit.
- the line-of-sight detection unit is configured to cause the line-of-sight direction detection unit to detect the line-of-sight direction of the user by looking through a viewfinder with respect to the image captured by the imaging unit.
- the image of the tracking object tracked by the tracking unit can be controlled so as to be optimally captured by the imaging unit.
- a reproduction unit that reproduces an image, and an image processing unit that processes the image reproduced by the reproduction unit and displays the image on the display unit can be further included.
- the image processing unit can optimally display the image of the tracking object tracked by the tracking unit.
- the image reproduced by the reproduction unit can be processed and displayed on the display unit.
- the line-of-sight direction detection unit may be the glasses-type device, and detects a user's line-of-sight direction with respect to an image having a field of view that can be seen by the user when the glasses-type device is put on. Can be.
- An image processing method performs a gaze direction detection process for detecting a gaze direction of a user with respect to an image, and the image is based on the gaze direction detected by the gaze direction detection process and the image.
- a tracking object for detecting a tracking object in the image that is tracked by the user's line of sight based on a time-series gaze area estimated by the estimation process.
- a program includes a gaze direction detection step for detecting a gaze direction of a user with respect to an image, a gaze direction detected by the processing of the gaze direction detection step, and the image. And a tracking object for detecting a tracking object in the image, which is tracked by the user's line of sight based on an estimation step for estimating a gaze area in the image and a time-series gaze area estimated by the processing of the estimation step A detection step; a tracking step for searching and tracking a tracking object detected by the processing of the tracking object detection step; and an image control step for controlling an image of the tracking object tracked by the processing of the tracking step.
- a gaze direction detection step for detecting a gaze direction of a user with respect to an image
- a gaze direction detected by the processing of the gaze direction detection step and the image.
- a tracking object for detecting a tracking object in the image, which is tracked by the user's line of sight based on an estimation step for estimating a gaze area in the image and a time-series gaze area estimated by the
- the gaze direction of the user with respect to the image is detected, the gaze area in the image is estimated based on the detected gaze direction and the image, and the estimated time-series gaze Based on the area, a tracking object in the image that is tracked by the user's line of sight is detected, the detected tracking object is searched and tracked, and the image of the tracked tracking object is controlled. .
- the image processing apparatus of the present technology may be an independent apparatus or a block that performs image processing.
- an object that is intended by the user is appropriately selected, and the image is processed and processed as a control target in the image, thereby capturing and displaying an image that appropriately reflects the user's intention. Can be realized.
- FIG. 11 is a diagram illustrating a configuration example of a general-purpose personal computer.
- First embodiment an example using a spectacle-type imaging device
- Modified example an example of using a normal imaging device
- Second embodiment an example of using a playback device
- FIG. 1 is a diagram illustrating an external configuration example of an imaging apparatus to which the present technology is applied.
- FIG. 2 is a functional block diagram illustrating functions for realizing an imaging apparatus to which the present technology is applied.
- the imaging apparatus of FIG. 1 is a spectacle-type imaging apparatus, and when a user wears spectacles, the user captures an image of a field of view that can be seen through the spectacles worn. At this time, the imaging apparatus of FIG. 1 estimates the gaze area in the user's image by detecting the user's gaze direction, and the subject that the user is gazing from the estimated gaze area information.
- the control target is specified, and an image is picked up while focusing on the subject to be specified as the control target.
- the imaging apparatus 11 of FIG. 1 includes transmission type displays 31-1 and 31-2, line-of-sight detection units 32-1 and 32-2, a control unit 33, an imaging mechanism unit 34, and an eyeglass-type frame 35. It is configured. Note that the transmissive displays 31-1 and 31-2 and the line-of-sight detection units 32-1 and 32-2 are simply referred to as the transmissive display 31 and the line-of-sight detection unit 32 unless it is necessary to distinguish between them. Other configurations are also referred to in the same manner.
- the transmissive display 31 has a configuration provided in place of the spectacle lens in the normal spectacle-type frame 35, and is configured by, for example, an organic EL (Electro-Luminescence), and the entire operation is controlled by the control unit 33. Has been.
- the transmissive display 31 allows the user to use the spectacle-type imaging device 11 in the same manner as when wearing spectacles, so that the user can transmit light that passes through the transmissive display 31, that is, the transmissive display 31. You can see the scenery that passes through.
- the transmissive display 31 displays a gaze frame indicating a gaze area where the user gazes based on the gaze direction detected by the gaze detection unit 32.
- the transmissive display 31 includes a target object that is a subject that is being watched by a specified user based on a gaze area that is detected in time series, and a region where an object that is specified as a control target exists. Display the target frame. By displaying the target frame indicating the area where the object identified as the subject to be controlled exists in this way, the user can select which of the viewable views corresponding to the viewable view. It is possible to recognize whether the area is a subject specified as a control target. That is, the transmissive display 31 displays a gaze frame or a target frame in principle, but other regions remain transparent.
- the line-of-sight detection unit 32 is an imaging device that can image the position of the pupil of the user's eyes when the user puts the spectacle-type frame 35 on to use glasses, for example, a CCD (Charge-Coupled Device) or Consists of CMOS (Complementary Metal Oxide Semiconductor).
- the line-of-sight detection unit 32 supplies, for example, a line-of-sight image in which the movement of the pupil position necessary for recognizing the line-of-sight direction is captured to the control unit 33.
- the imaging mechanism unit 34 is provided between the left and right transmissive displays 31-1 and 31-2 in the eyeglass-type frame 35 so as to capture a direction facing the wearing user.
- the imaging mechanism unit 34 supplies the captured image to the control unit 33.
- the angle of view of the image captured by the imaging mechanism unit 34 is the same as the visual field seen through the transmissive displays 32-1 and 32-2 when the user wears the glasses-type imaging device 11. Accordingly, when the user wears the imaging device 11, the user can monitor the image captured by the imaging mechanism unit 34 by viewing the visual field itself through the transmissive displays 31-1 and 31-2. .
- control unit 33 includes a line-of-sight recognition unit 51, a gaze region estimation unit 52, a gaze region specification unit 53, a control target detection unit 54, a control target tracking unit 55, and an optical system block control unit. 56.
- the imaging mechanism unit 34 includes an aperture, an optical system block 71 composed of various lenses, and an imaging unit 72 composed of CCD or CMOS.
- the line-of-sight recognition unit 51 acquires a line-of-sight image indicating the position of the user's pupil supplied from the line-of-sight detection units 32-1 and 32-2, recognizes the direction of the line of sight, and gazes at information on the line of sight. This is supplied to the region estimation unit 52.
- the line-of-sight recognition unit 51 includes a line-of-sight movement time-series pattern analysis unit 61. Based on the time-series line-of-sight image, the movement pattern in the line-of-sight direction is analyzed in time series, and the line-of-sight as an analysis result Information on the movement pattern of the direction and information on the line-of-sight direction are supplied to the gaze region estimation unit 52.
- fixation is a pattern indicating a state in which the eyeball is moving and still.
- slidable following movement is a pattern in which the eyeball follows the movement and moves slowly while keeping a close eye while the visual object is moving.
- impulsive eye movement is a line-of-sight pattern that quickly switches the line of sight.
- the pattern consisting of fixation and sliding follow-up movement is considered to be a time-series pattern in which the user is gazing at the subject he is interested in with his / her own eyes. Therefore, when the time-series pattern in the gaze direction is fixation or sliding follow-up motion, the gaze area estimation unit 52 considers that the object is gaze at that time, and the gaze direction at that time Based on this, the gaze area is estimated.
- the gaze area estimation unit 52 is based on the information on the line-of-sight direction supplied from the line-of-sight recognition unit 51, and the imaging mechanism unit 34, the gaze area is estimated from the line-of-sight image captured by the imaging unit 72 configured by CCD, CMOS, and the like, and information on the gaze area that is the estimation result is supplied to the gaze area specifying unit 53.
- the gaze area specifying unit 53 specifies the gaze area based on the gaze area information that is the estimation result supplied from the gaze area estimation unit 52, and information indicating the position of the identified gaze area is the control target detection unit 54. , And the frame display control unit 57.
- the control target detection unit 54 detects a region where the subject to be controlled exists in the image based on the information specified as the gaze region, and performs control target tracking on the image including the region where the subject to be controlled exists. Supplied to the unit 55.
- the control target tracking unit 55 tracks information on the area where the subject to be controlled exists in the image by searching in time series from the image captured by the imaging unit 72 of the imaging mechanism unit 34.
- the control target tracking unit 55 supplies information indicating the position of the region of the object in the image to be controlled being tracked to the optical system block control unit 56 and the frame display control unit 57.
- the optical system block control unit 56 controls the optical system block 71 provided in the imaging mechanism unit 34 so that the subject in the image to be controlled is in focus, so that the subject in the image to be controlled becomes an object. Control is performed so that an image is captured by the imaging unit 72 in a focused state.
- the frame display control unit 57 controls the transmissive display 31 based on the information on the gaze area supplied from the gaze area specifying unit 53 and the image captured by the imaging unit 72, and transmits corresponding to the gaze area.
- a rectangular gaze frame consisting of a dotted line is displayed at a position on the mold display.
- the frame display control unit 57 controls the transmissive display 31 based on the information on the position of the subject in the image to be controlled by the control target tracking unit 55 and the image captured by the imaging unit 72,
- a rectangular target frame consisting of a solid line is displayed at a position on the transmissive display 31 corresponding to the position where the subject to be controlled exists.
- the storage unit 58 includes, for example, a semiconductor memory, and stores the image data captured by the imaging unit 72 of the imaging mechanism unit 34 while compressing the data in a predetermined compression format. Further, the storage unit 58 reads out the stored image data and displays it on a display unit (not shown) or the like as necessary.
- step S ⁇ b> 11 the imaging mechanism unit 34 controls the imaging unit 72 to capture an image composed of light incident through the optical system block 71 and sequentially stores the captured image in the storage unit 58. This is supplied to the gaze area estimation unit 52, the control target tracking unit 55, and the frame display control unit 57. At this time, the captured image is compressed by a predetermined compression method and stored in the storage unit 58.
- step S ⁇ b> 12 the frame display control unit 57 tracks the control target based on whether or not the control target tracking unit 55 has supplied information on the area in the image where the subject image to be controlled exists. It is determined whether or not. If it is determined in step S12 that the image of the subject to be controlled is not tracked, the process proceeds to step S13.
- step S13 the line-of-sight detection units 32-1 and 32-2 capture line-of-sight images indicating the positions of the eyes of the left and right eyes of the user wearing the eyeglass-type frame 35, respectively, and sequentially capture the captured line-of-sight images. This is supplied to the line-of-sight recognition unit 51.
- step S14 the line-of-sight recognition unit 51 controls the line-of-sight movement time-series pattern analysis unit 61 to analyze the movement pattern in the line-of-sight direction based on the time-series line-of-sight image.
- the gaze movement pattern analysis result is supplied to the gaze area estimation unit 52.
- step S15 the gaze region estimation unit 52 determines whether the line-of-sight movement pattern is fixation or sliding follow-up motion, that is, whether the user is gazing at a specific subject.
- step S15 for example, when it is recognized as a fixation or slidable tracking motion, it is assumed that the user is gazing at a specific subject, and the process proceeds to step S16.
- step S ⁇ b> 16 the gaze area estimation unit 52 estimates the center position of the gaze area in the image captured by the imaging unit 72 of the imaging mechanism unit 34 based on the information on the gaze direction, and the gaze that is the estimation result. Information on the center position of the area is supplied to the gaze area specifying unit 53.
- the gaze area specifying unit 53 specifies the rectangular gaze area in the image captured by the imaging unit 72 based on the center position of the gaze area supplied from the gaze area estimation unit 52, and is controlled This is supplied to the detection unit 54 and the frame display control unit 57.
- the frame display control unit 57 surrounds an area corresponding to the specified gaze area on the transmissive display 31 within a range in which the user can see through the transmissive display 31. In this way, the gaze frame made up of dotted lines is displayed for a predetermined time.
- the control target detection unit 54 detects a subject to be controlled, that is, a subject to which the user sends a line of sight with interest, based on information on a predetermined number of gaze areas that are closest.
- the images P0 to P4 are images captured by the imaging unit 72 in time series, and at the same time, are scenes that the user wears the glasses-type frame 35 and is visually recognized through the transmissive display 31.
- C0 to C4 in each of the gaze areas E0 to E4 are center positions of the gaze area estimated based on the line-of-sight direction. Accordingly, the gaze areas E0 to E4 are rectangular areas that are identified with the positions C0 to C4 in the line-of-sight direction as the center.
- the gaze areas E0 to E4 are displayed on the transmissive display 31 for a predetermined time, so that they are visually recognized as gaze frames corresponding to the scenery seen through the transmissive display 31 in the user's field of view. Is. Therefore, hereinafter, the dotted frames surrounding the gaze regions E0 to E4 are also referred to as gaze frames E0 to E4. That is, in the images P0 to P4 in the upper part of FIG. 4, an image in which two persons are running in the right direction in the figure is shown. The figure of the person running and the background of the person are the scenery that can be visually recognized by the light transmitted through the transmissive display 31. Further, this viewable scene becomes the image itself captured by the imaging unit 72.
- the gaze frames E0 to E4 made up of dotted lines are displayed on the transmissive display 31 based on the information on the user's gaze direction, so that the user can superimpose the gaze frames E0 to E4 on the scenery. It can be visually recognized as a region where the subject is present.
- control target detection unit 54 extracts the image of the area indicated by the gaze frames E0 to E4 from the images P0 to P4 as shown in the middle part of FIG. 4, and as shown in the lower part of FIG. A region where a subject to be controlled exists, that is, a target region T is detected by overlapping the same portions. At this time, the control target detection unit 54 superimposes the same portion of the image in the area indicated by the gaze frames E0 to E4, and as a subject to be controlled, for example, a specific object such as a face image or a vehicle image By detecting this, an area where the subject to be controlled exists is detected as the target area.
- step S19 the control target detection unit 54 determines whether or not a target area where a subject to be controlled exists has been detected. More specifically, the control target detection unit 54 determines that the area of the region where the gaze frames extracted from the most recent predetermined number of images are overlapped is greater than a predetermined ratio with respect to the size set as the target region T. If the area is smaller than the large area, the area set by the gaze frame in the most recent images is fixed, that is, it is assumed that the user's gaze direction is determined by the subject to be controlled, and the target area is detected It is regarded as being done.
- step S19 for example, when the target area is detected, that is, when it is considered that the user's line-of-sight direction is determined to be the subject to be controlled, the process proceeds to step S20.
- step S20 the control target detection unit 54 supplies the detected target region image information to the control target tracking unit 55.
- the control target tracking unit 55 acquires the image information of the target area and searches for the image of the target area from the image supplied from the imaging unit 72. Then, the control target tracking unit 55 performs tracking by searching for an image of the target region, and information indicating the region in the image in which the image of the target region that is the tracking result has been searched is obtained as an optical system block control unit 56, and This is supplied to the frame display control unit 57.
- step S ⁇ b> 21 the frame display control unit 57 surrounds the area where the image of the tracked target frame is searched on the transmissive display 31 based on the information of the area where the image of the tracked target frame is searched.
- a target frame consisting of a solid line is displayed for a predetermined time.
- step S ⁇ b> 22 the optical system block control unit 56 controls the optical system block 71 of the imaging mechanism unit 34 so that the control target in the captured image is focused near the center in the tracked image. Control to adjust the focus.
- the optical system block 71 is controlled to be focused on the center position T ′ of the target frame T by tracking the target frame T. It is possible to capture an image in which the subject is focused.
- step S23 it is determined whether or not an operation unit (not shown) is operated to instruct the end of the operation. If the end is not instructed, the process returns to step S11, and the subsequent processes are repeated. Then, in step S23, when the end is instructed, the process ends.
- an operation unit not shown
- step S12 determines whether the control target is being tracked. If it is determined in step S12 that the control target is being tracked, the processes in steps S13 to S18 are skipped, and the process proceeds to step S19.
- the fact that the control target is being tracked means that the subject that is the control target that the user is interested in and has turned the line of sight is already detected and the focus can be adjusted. Since the process for specifying the gaze area is not necessary, the processes in steps S13 to S18 are skipped.
- step S15 if the line-of-sight movement pattern is not fixation or sliding follow-up motion, or if the control object is not detected in step S19, it is assumed that the control object cannot be searched, and the process proceeds to step S23. Proceed to
- the user can focus and take an image only by directing the direction of the line of sight to the subject to be controlled by himself / herself.
- the operation is performed even if a configuration for performing an operation or the like for specifying a subject to be controlled is provided. It is difficult to provide an operation unit or the like at an easy position. Further, providing the operation unit with priority on operability may impair the design. Therefore, by providing a focus position that can be changed in the line-of-sight direction without providing a special operation unit in this way, focusing that matches the user's intention is achieved without reducing operability and design. An image can be taken at the position.
- control target that the user is interested in is specified from the line-of-sight direction and the optical system block is controlled so that the position where the specified subject to be controlled exists is in focus is described.
- other imaging control may be performed based on the searched area where the subject to be controlled exists, for example, the optical system block is set so as to zoom up the area where the subject to be controlled exists. You may control.
- the hand-held imaging device 101 is substantially the same as the display unit 131 made up of an LCD, an organic EL, etc. It is composed of a main body 133 having functions and an imaging mechanism unit 134 similar to the imaging mechanism unit 34.
- the right side in the drawing is a front external view with the imaging mechanism unit 134 as the front, and the left portion in the drawing is a rear external view including the back on the right side in the drawing.
- the display unit 131 displays an image in which an image captured by the imaging mechanism unit 134 is combined with an image of a gaze frame displayed on the transmissive display 31 and an image of the target frame.
- an image similar to the image displayed on the display unit 131 is displayed on the viewing window 131 ′ as a small image that can be viewed by the user.
- the display unit 131 can be rotated in the direction of the arrow about the position indicated by the dotted line X. When the display unit 131 is housed on the main body 133 side, the display of the display unit 131 is stopped. It becomes a state.
- the viewing window 131 ′ is a so-called view finder, and is provided with a small LCD or organic EL, and displays an image similar to the image displayed on the display unit 131. Therefore, the user can view the same image as the image displayed on the display unit 131 by looking into the viewing window 131 ′ with the right eye or the left eye.
- the viewing window 131 ′ is provided with a line-of-sight detection unit 132 that captures a line-of-sight image indicating the position of the pupil for recognizing the user's line-of-sight direction from the inside.
- the gaze area is estimated based on the line-of-sight direction and the gaze frame is displayed as in the imaging device 11 in FIGS. Then, a subject to be controlled is specified and a target frame is displayed. Further, an optical system block (not shown) is controlled so as to focus on a subject existing at a position in the image corresponding to the target frame region.
- the basic functions of the image pickup apparatus 101 in FIG. 5 are the same as those of the image pickup apparatus 11 in FIGS. Also, the tracking imaging process by the imaging apparatus 101 in FIG. 5 is substantially the same as the process described with reference to the flowchart in FIG. 3, and thus the description thereof is omitted.
- Second Embodiment> ⁇ Configuration example of playback system>
- the subject to be controlled is specified by simply changing the line of sight to the subject desired by the user, and the target frame indicating the specified subject is displayed.
- an example has been described in which an image can be captured in a state where the optical block is controlled to focus on the vicinity of the center of the target frame in the captured image.
- the object that the user is interested in can be tracked based on the line-of-sight direction, it becomes a control object that the user is interested in from the reproduced image in the reproduction process by the same technique. It is also possible to search for a subject, track the searched subject to be controlled, perform appropriate image processing, and display the subject.
- FIG. 6 shows a playback processing system in which a zoomed-up image is generated and displayed as the center of a target area where a subject to be controlled that is of interest to the user exists during playback processing.
- FIG. 7 is a functional block diagram showing functions for realizing each device constituting the reproduction processing system of FIG.
- the reproduction processing system in FIG. 6 includes a glasses-type line-of-sight detection device 201, a reproduction device 202, and a display unit 203.
- the spectacle-type line-of-sight detection device 201 is similar in appearance to the spectacle-type imaging device 11 of FIG. 1, but further has a function of transmitting information on the target area to the playback device 202.
- the playback device 202 searches for the target region in the image to be played based on the information on the target region transmitted from the eyeglass-type line-of-sight detection device 201, and obtains an image based on the center position of the searched target region.
- An image zoomed up by the magnification is generated and displayed on the display unit 203.
- the transmissive displays 231-1 and 231-2, the line-of-sight detectors 232-1 and 232-2, the control unit 233, and the imaging mechanism unit 234 are respectively transmissive displays 31-1 and 31-2 of the imaging device 11.
- the line-of-sight detection units 32-1 and 32-2, the control unit 33, and the imaging mechanism unit 34 are basically provided with the same functions, and thus description thereof is omitted.
- a display position detection unit 251 that detects the display position of the display unit 203 in an image captured by the imaging unit 72, and a communication unit 252 that transmits information on the target area to the playback device 202 are newly added. Is provided.
- the image captured by the imaging mechanism unit 234 is not used except for specifying the target area where the subject to be controlled exists, and the captured image is basically Shall not be recorded.
- the playback device 202 plays back an image for which playback has been instructed and displays it on the display unit 203. More specifically, the playback device 202 includes a storage unit 271, a playback unit 272, an image processing unit 273, and a communication unit 274.
- the storage unit 271 stores image data to be reproduced and supplies the image data to the reproduction unit 272 as necessary.
- the reproduction unit 272 reproduces an image based on the image data read from the storage unit 271 and supplies information on the reproduced image to the image processing unit 273.
- the image processing unit 273 is based on the display position information of the display unit 203 from the glasses-type line-of-sight detection device 201 via the communication unit 274 and the center position of the target region among the images supplied from the reproduction unit 272. Then, an image enlarged at a predetermined magnification is generated on the basis of the center position of the target region and displayed on the display unit 203.
- step S71 the reproduction unit 272 reads out image data designated by an operation unit (not shown) from among the image data stored in the storage unit 271, reproduces the image, and supplies the image data to the image processing unit 273. To do.
- the image processing unit 273 causes the display unit 203 to display the image supplied from the reproduction unit 272 as it is.
- the display position detection unit 251 detects the position of the display surface of the display unit 203 from the image captured by the imaging unit 72 in step S53. To the communication unit 252. Further, in step S ⁇ b> 54, the communication unit 252 transmits information on the display position of the display unit 203 and information on the target area supplied from the control target tracking unit 55 to the playback device 202.
- step S72 the communication unit 274 determines whether or not the information on the target area has been transmitted from the eyeglass-type line-of-sight detection device 201. If the target area information is not transmitted from the eyeglass-type line-of-sight detection device 201 in step S72, the process proceeds to step S75, where it is instructed to end the process, or whether the reproduction has not ended. Determined. If the end of the process is not instructed in step S75 and the reproduction has not ended, the process returns to step S72. In step S75, if the end of the process is instructed or the reproduction is ended, the process ends.
- step S72 for example, when information on the target area is transmitted by the process of step S54, the process proceeds to step S73.
- step S ⁇ b> 73 the communication unit 274 acquires the display position of the display unit 203 imaged by the imaging unit 72 and the information on the target area supplied from the eyeglass-type line-of-sight detection device 201, and sends the information to the image processing unit 273. Supply.
- step S74 the image processing unit 273 uses the display position of the display unit 203 in the image picked up by the image pickup unit 72 and the information on the target area in the image displayed on the display unit 203.
- the position to be the target area specified as the line-of-sight direction is specified.
- the image processing unit 273 zooms up the reproduction image at a predetermined magnification around the area specified as the target area in the image displayed on the display unit 203 and causes the display unit 203 to display the image.
- the user wears the eyeglass-type gaze detection device 201, views the reproduced image, and directs his gaze toward the subject he is interested in.
- a nearby area is estimated as a gaze area, and an area of a subject that is interested from the information of the image set as the gaze area is set as a target area.
- the up image can be viewed as a reproduced image.
- the example in which the image of the target area is zoomed up has been described.
- other image processing may be performed on the target area. For example, the resolution of the image of the target area may be reduced. You may make it raise, and conversely make it low.
- the example using the eyeglass-type line-of-sight detection device 201 has been described.
- an imaging device that specifies the position of the user's pupil is located at a position such as the upper portion of the display unit 203, for example.
- the display unit 203 may be provided so as to face the user who views the image, and the user's line-of-sight direction may be specified from the captured image to implement the same processing.
- the display of the gaze frame and the target frame may be displayed on the display unit 203 by being combined with the reproduced image instead of being displayed on the transmissive display 231.
- the above-described series of processing can be executed by hardware, but can also be executed by software.
- a program constituting the software may execute various functions by installing a computer incorporated in dedicated hardware or various programs. For example, it is installed from a recording medium in a general-purpose personal computer or the like.
- FIG. 9 shows a configuration example of a general-purpose personal computer.
- This personal computer incorporates a CPU (Central Processing Unit) 1001.
- An input / output interface 1005 is connected to the CPU 1001 via a bus 1004.
- a ROM (Read Only Memory) 1002 and a RAM (Random Access Memory) 1003 are connected to the bus 1004.
- the input / output interface 1005 includes an input unit 1006 including an input device such as a keyboard and a mouse for a user to input an operation command, an output unit 1007 for outputting a processing operation screen and an image of the processing result to a display device, programs, and various types.
- a storage unit 1008 including a hard disk drive for storing data, a LAN (Local Area Network) adapter, and the like are connected to a communication unit 1009 that executes communication processing via a network represented by the Internet.
- magnetic disks including flexible disks
- optical disks including CD-ROM (Compact Disc-Read Only Memory), DVD (Digital Versatile Disc)), magneto-optical disks (including MD (Mini Disc)), or semiconductors
- a drive 1010 for reading / writing data from / to a removable medium 1011 such as a memory is connected.
- the CPU 1001 is read from a program stored in the ROM 1002 or a removable medium 1011 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, installed in the storage unit 1008, and loaded from the storage unit 1008 to the RAM 1003. Various processes are executed according to the program.
- the RAM 1003 also appropriately stores data necessary for the CPU 1001 to execute various processes.
- the CPU 1001 loads the program stored in the storage unit 1008 to the RAM 1003 via the input / output interface 1005 and the bus 1004 and executes the program, for example. Is performed.
- the program executed by the computer (CPU 1001) can be provided by being recorded on the removable medium 1011 as a package medium, for example.
- the program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the program can be installed in the storage unit 1008 via the input / output interface 1005 by attaching the removable medium 1011 to the drive 1010. Further, the program can be received by the communication unit 1009 via a wired or wireless transmission medium and installed in the storage unit 1008. In addition, the program can be installed in advance in the ROM 1002 or the storage unit 1008.
- the program executed by the computer may be a program that is processed in time series in the order described in this specification, or in parallel or at a necessary timing such as when a call is made. It may be a program for processing.
- the system means a set of a plurality of components (devices, modules (parts), etc.), and it does not matter whether all the components are in the same housing. Accordingly, a plurality of devices housed in separate housings and connected via a network and a single device housing a plurality of modules in one housing are all systems. .
- the present technology can take a cloud computing configuration in which one function is shared by a plurality of devices via a network and is jointly processed.
- each step described in the above flowchart can be executed by one device or can be shared by a plurality of devices.
- the plurality of processes included in the one step can be executed by being shared by a plurality of apparatuses in addition to being executed by one apparatus.
- this technique can also take the following structures.
- a gaze direction detection unit that detects a gaze direction of the user with respect to the image;
- An estimation unit that estimates a gaze area in the image based on the gaze direction detected by the gaze direction detection unit and the image;
- a tracking object detection unit that detects a tracking object in the image that is tracked by the user's line of sight based on a time-series gaze region estimated by the estimation unit;
- An image processing apparatus comprising: an image control unit that controls an image of the tracking object tracked by the tracking unit.
- the tracking object detection unit detects a face image as a tracking object in the image that is tracked by the user's line of sight based on the time-series gaze region estimated by the estimation unit.
- the image processing apparatus according to (1) detects a specific object as a tracking object in the image that is tracked by the user's line of sight based on the time-series gaze region estimated by the estimation unit.
- the image processing apparatus according to (1) or (2) detects a specific object as a tracking object in the image that is tracked by the user's line of sight based on the time-series gaze region estimated by the estimation unit.
- the image processing apparatus is a glasses-type imaging apparatus, An imaging unit that captures an image having a field of view that can be viewed by the user when the glasses-type imaging device is put on;
- the line-of-sight direction detection unit detects a user's line-of-sight direction with respect to an image having a visual field that can be viewed by the user when the spectacle-type imaging device is put on,
- the image processing apparatus according to any one of (1) to (3), wherein the image control unit performs control so that an image of the tracking object tracked by the tracking unit can be optimally captured by the imaging unit.
- the image processing apparatus is an image pickup apparatus that is used for image pickup by the user.
- An imaging unit that captures the image
- a viewfinder that looks into an image captured by the imaging unit
- the line-of-sight direction detection unit detects the user's line-of-sight direction through the viewfinder for the image captured by the imaging unit
- the image processing apparatus according to any one of (1) to (3), wherein the image control unit performs control so that an image of the tracking object tracked by the tracking unit can be optimally captured by the imaging unit.
- the line-of-sight direction detection unit detects the line-of-sight direction of the user when the user views the image displayed on the display unit, The image according to (1), wherein the image processing unit processes the image reproduced by the reproduction unit so that the image of the tracking object tracked by the tracking unit can be optimally displayed and displays the image on the display unit. Processing equipment.
- the line-of-sight direction detection unit includes the glasses-type device, and the user's line-of-sight with respect to an image having a field of view through which the user can see the display unit when the glasses-type device is put on.
- Gaze direction detection processing for detecting the user's gaze direction with respect to the image, Based on the line-of-sight direction detected by the line-of-sight direction detection process and the image, an estimation process for estimating a gaze area in the image, A tracking object detection process for detecting a tracking object in the image that is being tracked by the user's line of sight based on a time-series gaze region estimated by the estimation process, Tracking processing to search and track the tracking object detected by the tracking object detection processing, An image processing method including a step of performing image control processing for controlling an image of the tracking object tracked by the tracking processing.
- a line-of-sight direction detecting step for detecting a user's line-of-sight direction with respect to the image;
- An estimation step for estimating a gaze region in the image based on the gaze direction detected by the processing of the gaze direction detection step and the image;
- a tracking object detection step for detecting a tracking object in the image that is being tracked by the user's line of sight based on a time-series gaze region estimated by the processing of the estimation step;
- An image control step for controlling an image of the tracked object tracked by the tracking step processing.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- Studio Devices (AREA)
- Eye Examination Apparatus (AREA)
- User Interface Of Digital Computer (AREA)
- Image Analysis (AREA)
- Focusing (AREA)
- Automatic Focus Adjustment (AREA)
- Position Input By Displaying (AREA)
Abstract
Description
1. 第1の実施の形態(眼鏡型の撮像装置を用いた一例)
2. 変形例(通常の撮像装置を用いたときの一例)
3. 第2の実施の形態(再生装置を用いたときの一例)
<撮像装置の構成例>
図1は、本技術を適用した撮像装置の外観構成例を示す図である。また、図2は、本技術を適用した撮像装置を実現するための機能を示す機能ブロック図である。図1の撮像装置は、眼鏡型の撮像装置であり、使用者は、眼鏡を掛けるようにして装着すると、装着した眼鏡越しに見える視界の画像を撮像する。この際、図1の撮像装置は、使用者の視線方向を検出することで、使用者の画像内における注視領域を推定し、推定した注視領域の情報から、使用者が注視している被写体を制御対象として特定し、特定した制御対象となる被写体をトラッキングしながら焦点を合わせて画像を撮像する。
次に、図3のフローチャートを参照して、図1,図2の撮像装置11によるトラッキング撮像処理について説明する。
<撮像装置のその他の構成例>
以上においては、眼鏡型の撮像装置を用いた例について説明してきたが、当然のことながら、一般的に使用されている手持ちの撮像装置に対して応用するようにしても同様の効果を奏するものである。
<再生システムの構成例>
以上においては、撮像装置によりトラッキング撮像処理をすることで、使用者が所望とする被写体に視線方向を変えるだけで制御対象となる被写体が特定されて、特定された被写体を示す対象枠が表示されると共に、撮像される画像内における対象枠の中心付近に合焦するように光学ブロックが制御された状態で画像を撮像することができる例について説明してきた。しかしながら、視線方向に基づいて、使用者が興味を示す対象物をトラッキングすることができるので、同様の手法により、再生処理において、再生されている画像内から使用者が興味を示す制御対象となる被写体を検索し、検索された制御対象となる被写体をトラッキングして、適切な画像処理を施して表示するようにしてもよい。
次に、図8のフローチャートを参照して、図6,図7の再生処理システムによるトラッキング再生処理について説明する。尚、図8のフローチャートにおけるステップS41乃至S52,S55の処理は、図3のフローチャートにおけるステップS11乃至S23の処理と同様であるので、その説明は省略するものとする。また、基本的には、再生装置202に対して、図示せぬ操作部等により何らかの画像データの再生指示がなされており、ステップS71の処理により、再生装置202により画像データの再生がなされ、表示部203に再生された画像が表示される処理が最初に実行される。
(1) 画像に対する使用者の視線方向を検出する視線方向検出部と、
前記視線方向検出部により検出された視線方向と、前記画像とに基づいて、前記画像内における注視領域を推定する推定部と、
前記推定部により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体を検出する追跡物体検出部と、
前記追跡物体検出部により検出された追跡物体を検索してトラッキングするトラッキング部と、
前記トラッキング部によりトラッキングされた前記追跡物体の画像を制御する画像制御部と
を含む画像処理装置。
(2) 前記追跡物体検出部は、前記推定部により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体として、顔画像を検出する
(1)に記載の画像処理装置。
(3) 前記追跡物体検出部は、前記推定部により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体として、特定物体を検出する
(1)または(2)に記載の画像処理装置。
(4) 前記画像処理装置は、眼鏡型の撮像装置であり、
前記眼鏡型の撮像装置を掛けたとき、前記使用者が見ることができる視野からなる画像を撮像する撮像部をさらに含み、
前記視線方向検出部は、前記眼鏡型の撮像装置を掛けたときに、前記使用者が見ることができる視野からなる画像に対する使用者の視線方向を検出し、
前記画像制御部は、前記トラッキング部によりトラッキングされた前記追跡物体の画像を前記撮像部で最適に撮像できるように制御する
(1)乃至(3)のいずれかに記載の画像処理装置。
(5) 前記画像処理装置は、前記使用者により手持ちで撮像に使用される撮像装置であり、
前記画像を撮像する撮像部と、
前記撮像部により撮像される画像を覗くビューファインダとをさらに含み、
前記視線方向検出部は、前記撮像部により撮像された画像に対する、ビューファインダを覗く、前記使用者の視線方向を検出し、
前記画像制御部は、前記トラッキング部によりトラッキングされた前記追跡物体の画像を前記撮像部で最適に撮像できるように制御する
(1)乃至(3)のいずれかに記載の画像処理装置。
(6) 画像を再生する再生部と、
前記再生部により再生された画像を処理して表示部に表示させる画像処理部とをさらに含み、
前記視線方向検出部は、前記表示部に表示された画像を前記使用者が視聴するときの、前記使用者の視線方向を検出し、
前記画像処理部は、前記トラッキング部によりトラッキングされた前記追跡物体の画像を最適に表示できるように前記再生部により再生された画像を処理して前記表示部に表示させる
(1)に記載の画像処理装置。
(7) 前記視線方向検出部は、前記眼鏡型の装置からなり、前記眼鏡型の装置を掛けたときに、前記使用者が前記表示部を見ることができる視野からなる画像に対する使用者の視線方向を検出する
(6)に記載の画像処理装置。
(8) 画像に対する使用者の視線方向を検出する視線方向検出処理をし、
前記視線方向検出処理により検出された視線方向と、前記画像とに基づいて、前記画像内における注視領域を推定する推定処理をし、
前記推定処理により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体を検出する追跡物体検出処理をし、
前記追跡物体検出処理により検出された追跡物体を検索してトラッキングするトラッキング処理をし、
前記トラッキング処理によりトラッキングされた前記追跡物体の画像を制御する画像制御処理をする
ステップを含む画像処理方法。
(9) 画像に対する使用者の視線方向を検出する視線方向検出ステップと、
前記視線方向検出ステップの処理により検出された視線方向と、前記画像とに基づいて、前記画像内における注視領域を推定する推定ステップと、
前記推定ステップの処理により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体を検出する追跡物体検出ステップと、
前記追跡物体検出ステップの処理により検出された追跡物体を検索してトラッキングするトラッキングステップと、
前記トラッキングステップの処理によりトラッキングされた前記追跡物体の画像を制御する画像制御ステップと
を含む処理をコンピュータに実行させるためのプログラム。
Claims (9)
- 画像に対する使用者の視線方向を検出する視線方向検出部と、
前記視線方向検出部により検出された視線方向と、前記画像とに基づいて、前記画像内における注視領域を推定する推定部と、
前記推定部により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体を検出する追跡物体検出部と、
前記追跡物体検出部により検出された追跡物体を検索してトラッキングするトラッキング部と、
前記トラッキング部によりトラッキングされた前記追跡物体の画像を制御する画像制御部と
を含む画像処理装置。 - 前記追跡物体検出部は、前記推定部により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体として、顔画像を検出する
請求項1に記載の画像処理装置。 - 前記追跡物体検出部は、前記推定部により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体として、特定物体を検出する
請求項1に記載の画像処理装置。 - 前記画像処理装置は、眼鏡型の撮像装置であり、
前記眼鏡型の撮像装置を掛けたとき、前記使用者が見ることができる視野からなる画像を撮像する撮像部をさらに含み、
前記視線方向検出部は、前記眼鏡型の撮像装置を掛けたときに、前記使用者が見ることができる視野からなる画像に対する使用者の視線方向を検出し、
前記画像制御部は、前記トラッキング部によりトラッキングされた前記追跡物体の画像を前記撮像部で最適に撮像できるように制御する
請求項1に記載の画像処理装置。 - 前記画像処理装置は、前記使用者により手持ちで撮像に使用される撮像装置であり、
前記画像を撮像する撮像部と、
前記撮像部により撮像される画像を覗くビューファインダとをさらに含み、
前記視線方向検出部は、前記撮像部により撮像された画像に対する、ビューファインダを覗く、前記使用者の視線方向を検出し、
前記画像制御部は、前記トラッキング部によりトラッキングされた前記追跡物体の画像を前記撮像部で最適に撮像できるように制御する
請求項1に記載の画像処理装置。 - 画像を再生する再生部と、
前記再生部により再生された画像を処理して表示部に表示させる画像処理部とをさらに含み、
前記視線方向検出部は、前記表示部に表示された画像を前記使用者が視聴するときの、前記使用者の視線方向を検出し、
前記画像処理部は、前記トラッキング部によりトラッキングされた前記追跡物体の画像を最適に表示できるように前記再生部により再生された画像を処理して前記表示部に表示させる
請求項1に記載の画像処理装置。 - 前記視線方向検出部は、前記眼鏡型の装置からなり、前記眼鏡型の装置を掛けたときに、前記使用者が前記表示部を見ることができる視野からなる画像に対する使用者の視線方向を検出する
請求項6に記載の画像処理装置。 - 画像に対する使用者の視線方向を検出する視線方向検出処理をし、
前記視線方向検出処理により検出された視線方向と、前記画像とに基づいて、前記画像内における注視領域を推定する推定処理をし、
前記推定処理により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体を検出する追跡物体検出処理をし、
前記追跡物体検出処理により検出された追跡物体を検索してトラッキングするトラッキング処理をし、
前記トラッキング処理によりトラッキングされた前記追跡物体の画像を制御する画像制御処理をする
ステップを含む画像処理方法。 - 画像に対する使用者の視線方向を検出する視線方向検出ステップと、
前記視線方向検出ステップの処理により検出された視線方向と、前記画像とに基づいて、前記画像内における注視領域を推定する推定ステップと、
前記推定ステップの処理により推定された時系列の注視領域に基づいて、前記使用者の視線により追跡している、前記画像内の追跡物体を検出する追跡物体検出ステップと、
前記追跡物体検出ステップの処理により検出された追跡物体を検索してトラッキングするトラッキングステップと、
前記トラッキングステップの処理によりトラッキングされた前記追跡物体の画像を制御する画像制御ステップと
を含む処理をコンピュータに実行させるためのプログラム。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201380067240.7A CN104871525A (zh) | 2012-12-26 | 2013-12-12 | 图像处理装置、以及图像处理方法及程序 |
US14/652,519 US20150331486A1 (en) | 2012-12-26 | 2013-12-12 | Image processing device, image processing method and program |
JP2014554311A JPWO2014103732A1 (ja) | 2012-12-26 | 2013-12-12 | 画像処理装置および画像処理方法、並びにプログラム |
EP13869165.4A EP2940985A4 (en) | 2012-12-26 | 2013-12-12 | IMAGE PROCESSING DEVICE, AND IMAGE PROCESSING METHOD AND PROGRAM |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012-282612 | 2012-12-26 | ||
JP2012282612 | 2012-12-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014103732A1 true WO2014103732A1 (ja) | 2014-07-03 |
Family
ID=51020821
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2013/083297 WO2014103732A1 (ja) | 2012-12-26 | 2013-12-12 | 画像処理装置および画像処理方法、並びにプログラム |
Country Status (5)
Country | Link |
---|---|
US (1) | US20150331486A1 (ja) |
EP (1) | EP2940985A4 (ja) |
JP (1) | JPWO2014103732A1 (ja) |
CN (1) | CN104871525A (ja) |
WO (1) | WO2014103732A1 (ja) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2016208380A (ja) * | 2015-04-27 | 2016-12-08 | ソニーセミコンダクタソリューションズ株式会社 | 画像処理装置、撮像装置、画像処理方法およびプログラム |
JP2018507432A (ja) * | 2014-12-30 | 2018-03-15 | ノキア テクノロジーズ オサケユイチア | 個人的コンテンツの表示法 |
JP2018110399A (ja) * | 2016-12-30 | 2018-07-12 | アクシス アーベー | 方法、装置、およびカメラ |
JP2021124669A (ja) * | 2020-02-07 | 2021-08-30 | キヤノン株式会社 | 電子機器 |
US11361540B2 (en) | 2020-02-27 | 2022-06-14 | Samsung Electronics Co., Ltd. | Method and apparatus for predicting object of interest of user |
JP2022106849A (ja) * | 2017-02-08 | 2022-07-20 | キヤノン株式会社 | 画像処理装置、撮像装置および制御方法 |
JP2022546081A (ja) * | 2019-08-29 | 2022-11-02 | 株式会社ソニー・インタラクティブエンタテインメント | パーソナルデバイスによって支援されるtvストリーミング及びレンダリングコンテンツの中心窩最適化 |
JP7518684B2 (ja) | 2020-07-16 | 2024-07-18 | キヤノン株式会社 | 撮像制御装置、その制御方法、プログラム及び記録媒体 |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106464959B (zh) * | 2014-06-10 | 2019-07-26 | 株式会社索思未来 | 半导体集成电路和具备该半导体集成电路的显示装置及控制方法 |
EP3293960A4 (en) * | 2015-04-28 | 2018-12-05 | Sony Corporation | Information processing device, information processing method, and program |
US9898865B2 (en) * | 2015-06-22 | 2018-02-20 | Microsoft Technology Licensing, Llc | System and method for spawning drawing surfaces |
CN105892634A (zh) * | 2015-11-18 | 2016-08-24 | 乐视致新电子科技(天津)有限公司 | 防眩晕方法及虚拟现实显示输出设备 |
CN106557167B (zh) * | 2016-11-23 | 2020-08-04 | 上海擎感智能科技有限公司 | 智能眼镜及其控制设备的方法、系统及控制器 |
TWI632395B (zh) * | 2016-12-27 | 2018-08-11 | 廣達電腦股份有限公司 | 虛擬實境裝置以及其畫面產生方法 |
JP6953247B2 (ja) * | 2017-09-08 | 2021-10-27 | ラピスセミコンダクタ株式会社 | ゴーグル型表示装置、視線検出方法及び視線検出システム |
US10943115B2 (en) * | 2018-07-24 | 2021-03-09 | Apical Ltd. | Processing image data to perform object detection |
US10996751B2 (en) * | 2018-12-21 | 2021-05-04 | Tobii Ab | Training of a gaze tracking model |
CN112215120B (zh) * | 2020-09-30 | 2022-11-22 | 山东理工大学 | 视觉搜索区域的确定方法、装置及驾驶模拟器 |
KR20220058194A (ko) | 2020-10-30 | 2022-05-09 | 삼성전자주식회사 | 디스플레이를 포함하는 웨어러블 전자 장치, 그 디스플레이를 제어하는 방법, 및 그 웨어러블 전자 장치 및 케이스를 포함하는 시스템 |
US20240045496A1 (en) * | 2022-08-03 | 2024-02-08 | Sony Interactive Entertainment Inc. | Improving accuracy of interactions for gaze-enabled ar objects when in motion |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH01241511A (ja) * | 1988-03-23 | 1989-09-26 | Canon Inc | 視線検出手段を備えるカメラ |
JP2011053587A (ja) * | 2009-09-04 | 2011-03-17 | Sharp Corp | 画像処理装置 |
JP2012029209A (ja) * | 2010-07-27 | 2012-02-09 | Hitachi Ltd | 音処理システム |
JP2012137600A (ja) | 2010-12-27 | 2012-07-19 | Sony Corp | 撮像システム、撮像装置、およびプログラム |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5333029A (en) * | 1990-10-12 | 1994-07-26 | Nikon Corporation | Camera capable of detecting eye-gaze |
JP3725460B2 (ja) * | 2000-10-06 | 2005-12-14 | 株式会社ソニー・コンピュータエンタテインメント | 画像処理装置、画像処理方法、記録媒体、コンピュータプログラム、半導体デバイス |
US7396129B2 (en) * | 2004-11-22 | 2008-07-08 | Carestream Health, Inc. | Diagnostic system having gaze tracking |
US7697827B2 (en) * | 2005-10-17 | 2010-04-13 | Konicek Jeffrey C | User-friendlier interfaces for a camera |
US7855743B2 (en) * | 2006-09-08 | 2010-12-21 | Sony Corporation | Image capturing and displaying apparatus and image capturing and displaying method |
JP4961914B2 (ja) * | 2006-09-08 | 2012-06-27 | ソニー株式会社 | 撮像表示装置、撮像表示方法 |
US7860382B2 (en) * | 2006-10-02 | 2010-12-28 | Sony Ericsson Mobile Communications Ab | Selecting autofocus area in an image |
WO2008072374A1 (ja) * | 2006-12-11 | 2008-06-19 | Nikon Corporation | 電子カメラ |
WO2010102037A2 (en) * | 2009-03-03 | 2010-09-10 | The Ohio State University | Gaze tracking measurement and training system and method |
WO2011008793A1 (en) * | 2009-07-13 | 2011-01-20 | Emsense Corporation | Systems and methods for generating bio-sensory metrics |
US9197736B2 (en) * | 2009-12-31 | 2015-11-24 | Digimarc Corporation | Intuitive computing methods and systems |
US8384534B2 (en) * | 2010-01-14 | 2013-02-26 | Toyota Motor Engineering & Manufacturing North America, Inc. | Combining driver and environment sensing for vehicular safety systems |
JP2011150594A (ja) * | 2010-01-22 | 2011-08-04 | Sony Corp | 画像処理装置および方法、並びにプログラム |
US20110273369A1 (en) * | 2010-05-10 | 2011-11-10 | Canon Kabushiki Kaisha | Adjustment of imaging property in view-dependent rendering |
US8379981B1 (en) * | 2011-08-26 | 2013-02-19 | Toyota Motor Engineering & Manufacturing North America, Inc. | Segmenting spatiotemporal data based on user gaze data |
US9323325B2 (en) * | 2011-08-30 | 2016-04-26 | Microsoft Technology Licensing, Llc | Enhancing an object of interest in a see-through, mixed reality display device |
US20130258089A1 (en) * | 2011-11-03 | 2013-10-03 | Intel Corporation | Eye Gaze Based Image Capture |
US20130155309A1 (en) * | 2011-12-20 | 2013-06-20 | Research In Motion Limited | Method and Apparatus for Controlling Camera Functions Using Facial Recognition and Eye Tracking |
EP2847648A4 (en) * | 2012-05-09 | 2016-03-02 | Intel Corp | EYE PERSONALIZED SELECTIVE STRESSING OF PARTS OF AN INDICATION |
US8854447B2 (en) * | 2012-12-21 | 2014-10-07 | United Video Properties, Inc. | Systems and methods for automatically adjusting audio based on gaze point |
-
2013
- 2013-12-12 EP EP13869165.4A patent/EP2940985A4/en not_active Withdrawn
- 2013-12-12 JP JP2014554311A patent/JPWO2014103732A1/ja not_active Abandoned
- 2013-12-12 WO PCT/JP2013/083297 patent/WO2014103732A1/ja active Application Filing
- 2013-12-12 CN CN201380067240.7A patent/CN104871525A/zh active Pending
- 2013-12-12 US US14/652,519 patent/US20150331486A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH01241511A (ja) * | 1988-03-23 | 1989-09-26 | Canon Inc | 視線検出手段を備えるカメラ |
JP2011053587A (ja) * | 2009-09-04 | 2011-03-17 | Sharp Corp | 画像処理装置 |
JP2012029209A (ja) * | 2010-07-27 | 2012-02-09 | Hitachi Ltd | 音処理システム |
JP2012137600A (ja) | 2010-12-27 | 2012-07-19 | Sony Corp | 撮像システム、撮像装置、およびプログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP2940985A4 |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018507432A (ja) * | 2014-12-30 | 2018-03-15 | ノキア テクノロジーズ オサケユイチア | 個人的コンテンツの表示法 |
JP2016208380A (ja) * | 2015-04-27 | 2016-12-08 | ソニーセミコンダクタソリューションズ株式会社 | 画像処理装置、撮像装置、画像処理方法およびプログラム |
US11089289B2 (en) | 2015-04-27 | 2021-08-10 | Sony Semiconductor Solutions Corporation | Image processing device, imaging device, image processing method and program |
JP2018110399A (ja) * | 2016-12-30 | 2018-07-12 | アクシス アーベー | 方法、装置、およびカメラ |
JP7353015B2 (ja) | 2016-12-30 | 2023-09-29 | アクシス アーベー | 方法、装置、およびカメラ |
JP2022106849A (ja) * | 2017-02-08 | 2022-07-20 | キヤノン株式会社 | 画像処理装置、撮像装置および制御方法 |
JP7346654B2 (ja) | 2017-02-08 | 2023-09-19 | キヤノン株式会社 | 画像処理装置、撮像装置、制御方法、プログラム、及び記憶媒体 |
JP2022546081A (ja) * | 2019-08-29 | 2022-11-02 | 株式会社ソニー・インタラクティブエンタテインメント | パーソナルデバイスによって支援されるtvストリーミング及びレンダリングコンテンツの中心窩最適化 |
JP7339435B2 (ja) | 2019-08-29 | 2023-09-05 | 株式会社ソニー・インタラクティブエンタテインメント | パーソナルデバイスによって支援されるtvストリーミング及びレンダリングコンテンツの中心窩最適化 |
JP2021124669A (ja) * | 2020-02-07 | 2021-08-30 | キヤノン株式会社 | 電子機器 |
JP7500211B2 (ja) | 2020-02-07 | 2024-06-17 | キヤノン株式会社 | 電子機器 |
US11361540B2 (en) | 2020-02-27 | 2022-06-14 | Samsung Electronics Co., Ltd. | Method and apparatus for predicting object of interest of user |
JP7518684B2 (ja) | 2020-07-16 | 2024-07-18 | キヤノン株式会社 | 撮像制御装置、その制御方法、プログラム及び記録媒体 |
Also Published As
Publication number | Publication date |
---|---|
EP2940985A1 (en) | 2015-11-04 |
US20150331486A1 (en) | 2015-11-19 |
JPWO2014103732A1 (ja) | 2017-01-12 |
EP2940985A4 (en) | 2016-08-17 |
CN104871525A (zh) | 2015-08-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014103732A1 (ja) | 画像処理装置および画像処理方法、並びにプログラム | |
US11860511B2 (en) | Image pickup device and method of tracking subject thereof | |
US10666856B1 (en) | Gaze-directed photography via augmented reality feedback | |
US9245389B2 (en) | Information processing apparatus and recording medium | |
US8249305B2 (en) | Information processing apparatus, information processing method, program, and recording medium | |
US11172158B2 (en) | System and method for augmented video production workflow | |
US20150036999A1 (en) | Viewer Attention Controlled Video Playback | |
EP2720464B1 (en) | Generating image information | |
US20120026364A1 (en) | Image pickup apparatus | |
JP7467114B2 (ja) | 撮像装置およびその制御方法 | |
JP2013162333A (ja) | 画像処理装置、画像処理方法、プログラムおよび記録媒体 | |
US20160261908A1 (en) | Media streaming system and control method thereof | |
US20230328355A1 (en) | Information processing apparatus, information processing method, and program | |
KR101810671B1 (ko) | 전방향 영상의 방향 정보를 생성하는 방법 및 이러한 방법을 수행하는 장치 | |
US20210132682A1 (en) | User signal processing method and device for performing method | |
JP6087615B2 (ja) | 画像処理装置およびその制御方法、撮像装置、および表示装置 | |
CN108108025A (zh) | 基于头戴设备的信息指引方法、装置、头戴设备和介质 | |
WO2022137928A1 (ja) | 画像処理装置、画像処理方法、プログラム | |
JP2012124767A (ja) | 撮像機器 | |
WO2018074262A1 (ja) | 通信装置および通信方法、並びにプログラム | |
US20220309992A1 (en) | Head-mounted display, display control system, information processor, and method for controlling head-mounted display | |
US20230345122A1 (en) | Imaging apparatus, imaging method, and imaging program | |
US20160172004A1 (en) | Video capturing apparatus | |
JP5659856B2 (ja) | 撮像装置、撮像方法、及びプログラム | |
JP2023124598A (ja) | 主被写体判定装置、撮像装置、主被写体判定方法、及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13869165 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2014554311 Country of ref document: JP Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2013869165 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2013869165 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14652519 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |