Utility model content
Based on this, it is necessary to for the imaging sensor price for being integrated with microlens array currently used for image refocusing
Valuableness, it is at high cost, and it is multifocal be easy to the problem of reducing whole clarity away under conditions of, a kind of control that image focuses again is provided
System processed.
In order to realize the purpose of this utility model, the utility model adopts the following technical solution:
A kind of control system that image focuses again, comprising:
Pattern establishes module, is set as the feature dot pattern of projection target object;
Image capture module is set as under different capture angles and different focal length, at least according to the capture of feature dot pattern
Two images;
Offset obtains module, is set as obtaining the picture point group for corresponding to same characteristic point in different images, obtains picture point group
Position offset between middle picture point;
Depth map generation module is set as obtaining the depth of the characteristic point according to picture point group, position offset and focal length
Information is spent, depth map is generated;
Focus module is set as selected target image, is focused again according to depth map to the target area on image.
In one embodiment, the pattern establishes module and includes:
First map unit is arranged to set up the first mapping relations of target object Yu feature dot pattern;
Projecting unit, is set as according to the first mapping relations, goes out feature dot pattern to scene simulation using target object.
In one embodiment, in the feature dot pattern, characteristic point is arranged by preset rules.
In one embodiment, the projecting unit includes projector.
In one embodiment, described image trapping module includes multiple cameras with different apertures or a tool
There is the camera of multiple apertures.
In one embodiment, the offset acquisition module includes:
First selected unit, is set as a selected at least characteristic point, obtains corresponding picture point group according to characteristic point;
First acquisition unit is set as obtaining the position coordinates of picture point Zu Zhongge picture point dot center;
Second acquisition unit is set as obtaining the position offset in picture point group between picture point according to the position coordinates.
In one embodiment, the offset obtains module further include:
Second map unit, the second mapping relations being arranged to set up between characteristic point and picture point.
In one embodiment, the focus module includes:
Second selected unit is set as being compared the corresponding focal length of different images, by the corresponding image of maximum focal length
It is chosen to be target image;
Focusing unit is set as focusing the target area on image again according to the depth map.
A kind of control system that image focuses again, comprising:
Pattern establishes module, is set as the feature dot pattern of projection target object;
Image capture module is set as under different capture angles and different focal length, at least according to the capture of feature dot pattern
Two images;
Offset obtains module, is set as obtaining the picture point group for corresponding to same characteristic point in different images, obtains picture point group
Position offset between middle picture point;
Depth map generation module is set as obtaining the depth of the characteristic point according to picture point group, position offset and focal length
Information is spent, depth map is generated;
Focus module is set as selected target image, is focused again according to depth map to the target area on image;
Wherein, the pattern establishes module and includes:
First map unit is arranged to set up the first mapping relations of target object Yu feature dot pattern;
Projecting unit, is set as according to the first mapping relations, goes out feature dot pattern to scene simulation using target object;
Described image trapping module includes multiple cameras with different apertures or a camera shooting with multiple apertures
Head;
The offset obtains module
First selected unit, is set as a selected at least characteristic point, obtains corresponding picture point group according to characteristic point;
First acquisition unit is set as obtaining the position coordinates of picture point Zu Zhongge picture point dot center;
Second acquisition unit is set as obtaining the position offset in picture point group between picture point according to the position coordinates;
Second map unit, the second mapping relations being arranged to set up between characteristic point and picture point;
The focus module includes:
Second selected unit is set as being compared the corresponding focal length of different images, by the corresponding image of maximum focal length
It is chosen to be target image;
Focusing unit is set as focusing the target area on image again according to the depth map.
Above-mentioned control system, including pattern establish module, image capture module, offset and obtain module, depth map generation
Module and focus module establish the feature dot pattern of module projects target object by pattern;Image capture module is in difference
It captures under angle and different focal length, according to feature dot pattern capture at least two images;Then offset obtains module and obtains not
With the position offset between the picture point group picture point for corresponding to same characteristic point in image, depth map generation module is according to picture point group, position
It sets offset and focal length obtains the higher depth information of accuracy, generate depth map;Focus module is according to depth map to target
Region is focused.System can be realized the vernier focusing to target area on image as a result, improve the clear of image entirety
Degree, improves the experience of user.
Specific embodiment
The utility model is more fully retouched below with reference to relevant drawings for the ease of understanding the utility model,
It states.The alternative embodiment of the utility model is given in attached drawing.But the utility model can come in many different forms in fact
It is existing, however it is not limited to embodiment described herein.On the contrary, purpose of providing these embodiments is makes public affairs to the utility model
The understanding for opening content is more thorough and comprehensive.
Unless otherwise defined, all technical and scientific terms used herein are led with the technology for belonging to the utility model
The normally understood meaning of the technical staff in domain is identical.Terminology used in the description of the utility model herein only be
The purpose of description specific embodiment, it is not intended that in limitation the utility model.In the utility model embodiment and appended power
The "an" of singular used in sharp claim, " described " and "the" are also intended to including most forms, unless context
It clearly expresses other meaning.It is also understood that term "and/or" used herein refers to and includes one or more related
Connection is listed any or all of project and may be combined.Meanwhile the ordinal numbers such as " first " used herein, " second " limit
Fixed element herein is not offered as the sequencing that priority, order or method between element execute, and is only for
Distinguish the different elements with same names.
Referring to Fig. 1, Fig. 1 is the structure chart for the control system that image focuses again in an embodiment.
The control system of the present embodiment includes: that pattern establishes module 101, image capture module 102, offset acquisition module
103, depth map generation module 104 and focus module 105.Wherein, image capture apparatus 102 establishes 101 phase of device with pattern
Matching, offset acquisition device 103 connect image capture apparatus 102, and depth map generation device 104 is separately connected offset acquisition
Device 103 and focusing device 105, meanwhile, focusing device 105 is also connected with image capture apparatus 102 with selected target image.Specifically
Ground:
Pattern establishes module 101, is set as the feature dot pattern of projection target object.
In the present embodiment, target object refers to target reference object, i.e., the image focused again is finally needed to clap
Main body is taken the photograph, the usual target object is object with different depths, such as three-dimension object;Feature dot pattern refers to and object
The corresponding projection dot pattern of body spatial point, can correspond to different depth information.
Wherein, feature dot pattern can be arranged by preset rules, such as can be the sparse dot pattern with regular intervals,
Including but not limited to the matrix dot of ranks aturegularaintervals, the matrix dot at line discipline interval, column aturegularaintervals matrix dot (please assist
Referring to fig. 2, Fig. 2 is by taking the matrix dot of ranks aturegularaintervals as an example, and wherein small circle a represents characteristic point);Certainly, feature dot pattern
It is also possible to the discrete point of random arrangement;The distance of two neighboring characteristic point can be identical, can not also be identical, here, right
The specific location of characteristic point and define not further limited.When projection dot pattern is the sparse dot pattern with regular intervals
When, it is possible to reduce projection cost, meanwhile, it is also convenient for the picture point group of the selected corresponding same characteristic point of subsequent step.It needs to illustrate
It is that a set point being located on target object at any time only occupies a unique position in characteristic point.
Wherein, it includes but is not limited to the projecting subassemblies such as projector that pattern, which establishes module 101, can utilize mesh by projector
Mark object goes out feature dot pattern to scene simulation, or goes out feature to scene simulation using target object by other projecting subassemblies
Dot pattern.
In one embodiment, it includes the first map unit and projecting unit that pattern, which establishes module 101,.
First map unit is arranged to set up the first mapping relations of target object Yu feature dot pattern.
Only occupy a unique position in characteristic point due to being located at a set point on target object at any time
It sets, therefore can establish the first mapping relations of target object Yu feature dot pattern, such as pass through perspective projection transformation or orthogonal
Projective transformation etc. obtains the first mapping relations.
Projecting unit, is set as according to the first mapping relations, goes out feature dot pattern to scene simulation using target object.
After establishing the first mapping relations of target object and feature dot pattern, projection group is controlled according to the first mapping relations
Part projects feature dot pattern to target object.In one embodiment, projecting unit includes projector.
Image capture module 102 is set as under different capture angles and different focal length, according to feature dot pattern capture to
Few two images.
It in the present embodiment, can be using multiple cameras that there are different apertures and be located at different shooting angles to same
The capture that scene (i.e. feature dot pattern) carries out multiple images (please be assisted referring to Fig. 3, Fig. 3 is with the camera of two different apertures
For, wherein the first camera 10 is set as having biggish f-number, and second camera 20 is set as having smaller aperture
Value);Multiple images can also be carried out to Same Scene in multiple shooting angle using a camera with multiple apertures
Capture.Wherein, camera can be using in the electronic device, such as applies in mobile phone, tablet computer, vehicle-mounted computer, wearable
Equipment, digital camera etc., which have, takes pictures, in any electronic equipment of camera function.
In one embodiment, image capture module 102 include but is not limited to multiple cameras with different apertures or
One camera with multiple apertures.When being captured by different cameras to Same Scene, the center of different cameras
Point is in the same plane;When one camera captures Same Scene, central point is kept in camera moving process
Motion track is in the same plane.
It, can be different to capturing to Same Scene according to feature dot pattern under different capture angles and different focal length
Image, since capture angle is different, thus on Same Scene different images there is offset in the position of picture point;Since focal length is different,
There are the image of focusing and the image of defocus between Same Scene different images, but also image point position shifts.Wherein, defocus
Image correspond to large aperture camera, corresponding focal length is larger, and image resolution ratio is higher, has relatively fuzzy picture point (referring to figure
4, image T1 are 10 captured image of the first camera;Wherein, picture point A1, B1, P1 respectively corresponds characteristic point A, B, P of Fig. 3, circle
The point at circle middle part is the dot center of picture point);The image of focusing corresponds to small aperture camera, and corresponding focal length is smaller, has relatively clear
(referring to fig. 4, image T2 is 20 captured image of second camera to clear picture point;Wherein, picture point A2, B2, P2 respectively corresponds Fig. 3
Characteristic point A, B, P).Wherein, since characteristic point A, B, P correspond to different depth informations, picture point A1, B1, P1 can divide
Other places are not ipsilateral in the focal plane of 10 place camera system of the first camera;Picture point A2, B2, P2 can be respectively at second and take the photograph
As first 20 place camera system focal plane it is not ipsilateral.
Offset obtains module 103, is set as obtaining the picture point group for corresponding to same characteristic point in different images, obtains picture point
Position offset in group between picture point.
In the present embodiment, position offset refers to that the position between the picture point group for corresponding to same characteristic point in different images is deposited
Offset, in one embodiment, position offset refers to the size of positional shift or side between picture point Zu Ge picture point dot center
To position offset is vector shift.Since a certain characteristic point of any moment only occupies a unique position in space, thus
Each picture point can only be corresponding with only one characteristic point in each image, and thus the position of same picture point Zu Jian picture point dot center is inclined
Shifting amount is unique.
In the present embodiment, a certain characteristic point of any moment only occupies a unique position, thus every width figure in space
As upper each picture point can only be corresponding with only one characteristic point, i.e., picture point group and characteristic point have unique mapping relations.
It includes but is not limited to image analysis apparatus that offset, which obtains module 103,.In one embodiment, offset obtains mould
Block 103 includes the first selected unit, first acquisition unit and second acquisition unit.
First selected unit, is set as a selected at least characteristic point, obtains corresponding picture point group according to characteristic point.
When a selected wherein characteristic point, then the corresponding picture point group of this feature point can be obtained according to mapping relations.Alternatively,
A picture point can also be selected, obtains character pair point according to mapping relations, to obtain corresponding same characteristic point according to characteristic point
Other picture points.
First acquisition unit is set as obtaining the position coordinates of picture point Zu Zhongge picture point dot center.
After determining the corresponding picture point group of same characteristic point, the position of each picture point of picture point group is determined, and obtain each picture point point
The position coordinates at center.Since the corresponding capture focal length of image is different, the fog-level of different images is different, that is, together
The image fog-level of the corresponding picture point group of one characteristic point is different, and the detection of picture point dot center, with image fog-level and
Ambient lighting is unrelated, therefore under different image-forming conditions, by picture point dot center position coordinates calculating position offset, can guarantee
The reliability of measurement accuracy improves the accuracy of Depth Information Acquistion.
Second acquisition unit is set as obtaining the position offset in picture point group between picture point according to the position coordinates.
After filtering out the corresponding picture point group of same characteristic point, position can be obtained according to picture point dot center position coordinates
Offset.
In said units, plane where can choosing capture image is X/Y plane, and two-dimensional coordinate is established on X/Y plane
System, the origin position of two-dimensional coordinate system is not further in this application to be limited.It is by two images of captured image
Example, position offset can by the first image it is Chong Die with the second image later be mapped on X/Y plane, and acquisition two images in
Vector distance between each picture point dot center position coordinates of corresponding same characteristic point.For example, a certain characteristic point A is chosen, according to
Mapping relations select picture point group (auxiliary is referring to Fig. 3-Fig. 5) corresponding with characteristic point A: A1 (being located at image T1) and A2 (are located at
Image T2), picture point A1 X/Y plane coordinate information be A1 (X11, Y11), picture point A2 X/Y plane coordinate information be (X21,
Y21), position offset D1 can be obtained according to picture point A1 and picture point A2;B1 (being located at image T1) and B2 (being located at image T2),
Picture point B1 is B1 (X12, Y12) in the coordinate information of X/Y plane, and picture point B2 is (X22, Y22), root in the coordinate information of X/Y plane
Position offset D2 can be obtained according to picture point B1 and picture point B2.
It is directed to each characteristic point as a result, the position of the picture point dot center of same characteristic point is corresponded in available two images
Set the vector distance between coordinate;According to determining multiple characteristic points, can to obtaining the corresponding each picture point of each characteristic point,
And the corresponding position offset of each picture point group.
In one embodiment, in order to improve the matching degree between characteristic point and picture point group, offset obtains module 103 and also wraps
Include the second map unit.
Second map unit, the second mapping relations being arranged to set up between characteristic point and picture point.
Second between characteristic point and picture point can be preset by the modes such as experiment, K-theoretic operation K or the mode combined
Mapping relations, then according to the second mapping relations matching characteristic point and corresponding picture point group.It in one embodiment, can be preparatory
The mapping table between characteristic point and picture point is established, and fits the second mapping between characteristic point and picture point according to mapping table
Relationship.Wherein, the mapping relations between fit characteristic point and picture point can determine the position of characteristic point by the way that function model is arranged
It is bent to draw fitting by computer geometric techniques in two-dimensional coordinate system for the function for setting the position coordinates satisfaction of coordinate and picture point
Line, so that it is determined that the function that the position coordinates of characteristic point meet with the position coordinate value of corresponding picture point.
Depth map generation module 104 is set as obtaining the depth of characteristic point according to picture point group, position offset and focal length
Information generates depth map.Depth map generation module 104 includes but is not limited to image analysis apparatus.
In the present embodiment, position coordinates of the depth map generation module 104 according to picture point group dot center, picture point group dot center
Position offset and the camera utilized focal length obtain characteristic point depth information.Image is captured with two cameras
For, wherein the central point of the first camera, the central point of second camera are generally aligned in the same plane, and can be set and be determined two
Distance and the first camera and second camera between the camera site (shooting angle) of camera and camera central point
Focal length.Based on principle of triangulation, available characteristic point and the distance between plane Z where two camera central points,
In, distance Z is the depth information of characteristic point.Specifically, (first takes the photograph the distance between two camera central points of distance Z=*
As head or the focal length of second camera)/position offset.Wherein, since position offset is vector, the weight of depth information
Build the interior outside of the focal plane of camera system where can extend to camera.
Optionally, this programme can be applicable to include three or three or more cameras electronic equipment.With three
It is illustrated for camera, may be constructed the combination of camera two-by-two, two available spies of camera in each combination
The depth information of point is levied, thus available three groups of depth informations, it can be using the mean depth of three groups of depth informations as spy
Levy the actual depth of point.The accuracy of Depth Information Acquistion is improved, and then realizes the exact focus to shooting object.
In the present embodiment, depth map generation module 104 can first obtain the corresponding depth map of sparse dot pattern.If needed
The depth information of all characteristic points is obtained, it is deep in fixed characteristic point to can use Surface interpolating algorithm or approximate algorithm etc.
Other depth values for not determining characteristic point are calculated between degree information.
Focus module 105 is set as selected target image, is gathered again according to depth map to the target area on image
It is burnt.Focus module 105 includes but is not limited to image processing apparatus.
In the present embodiment, the image for needing to carry out refocusing is first selected, then determines a target area from target image
Domain is focused again.Wherein, target area refers to interested region on target image, such as the face area in portrait,
Or other regions with special marking.Target area can carry out the selection of size according to actual needs.
In one embodiment, target image chooses the preferable image of defocus effects, i.e. the selection biggish camera of aperture is caught
The image obtained, the image have relatively fuzzy point, and higher image resolution ratio, contrast is higher when being focused again, meet again
Burnt effect is more prominent.
For example, focus module 105 includes the second selected unit and focusing unit.
Second selected unit is set as being compared the corresponding focal length of different images, by the corresponding image of maximum focal length
It is chosen to be target image.Second selected unit can be image Chooser.
Focusing unit is set as focusing the target area on image again according to depth map.Focus pack can be with
It is image processor.
In one embodiment, multiple target areas can be selected, are followed successively by each target area matching of target image not
Same depth information;Target area is focused again according to matched depth information.It is met again one by one to target area
Burnt mode, for the mode that entire target image once focuses, focusing accuracy is greatly enhanced, and clarity is more preferable,
And the effect background blurring for the later period is significantly improved.
Control system provided in this embodiment, including pattern establish module, image capture module, offset obtain module,
Depth map generation module and focus module establish the feature dot pattern of module projects target object by pattern;Image capture
Module is under different capture angles and different focal length, according to feature dot pattern capture at least two images;Then offset obtains
Module obtains the position offset corresponded between the picture point group picture point of same characteristic point in different images, depth map generation module according to
Picture point group, position offset and focal length obtain the higher depth information of accuracy, generate depth map;Focus module is according to depth
Figure is focused target area.System can be realized the vernier focusing to target area on image as a result, and it is whole to improve image
Clarity, improve the experience of user.
Modules in above-mentioned control system can be realized fully or partially through software, hardware and combinations thereof.It is above-mentioned
Each module can be embedded in the form of hardware or independently of in the processor in computer equipment, can also be stored in meter in a software form
It calculates in the memory in machine equipment, executes the corresponding operation of the above modules in order to which processor calls.
Each technical characteristic of embodiment described above can be combined arbitrarily, for simplicity of description, not to above-mentioned reality
It applies all possible combination of each technical characteristic in example to be all described, as long as however, the combination of these technical characteristics is not deposited
In contradiction, all should be considered as described in this specification.
Above-described embodiments merely represent several embodiments of the utility model, the description thereof is more specific and detailed,
But it cannot be understood as the limitations to utility model patent range.It should be pointed out that for the common skill of this field
For art personnel, without departing from the concept of the premise utility, various modifications and improvements can be made, these are belonged to
The protection scope of the utility model.Therefore, the scope of protection shall be subject to the appended claims for the utility model patent.