CN107360412A - 3D image creation method, photographing apparatus, and readable storage medium - Google Patents
3D image creation method, photographing apparatus, and readable storage medium Download PDFInfo
- Publication number
- CN107360412A CN107360412A CN201710719601.1A CN201710719601A CN107360412A CN 107360412 A CN107360412 A CN 107360412A CN 201710719601 A CN201710719601 A CN 201710719601A CN 107360412 A CN107360412 A CN 107360412A
- Authority
- CN
- China
- Prior art keywords
- image
- target area
- rendering
- focusing position
- focal plane
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 238000003860 storage Methods 0.000 title claims abstract description 15
- 238000009877 rendering Methods 0.000 claims description 41
- 238000003384 imaging method Methods 0.000 claims description 12
- 238000004364 calculation method Methods 0.000 claims description 7
- 230000004308 accommodation Effects 0.000 claims description 5
- 238000004590 computer program Methods 0.000 claims description 5
- 238000005192 partition Methods 0.000 claims description 4
- 238000009826 distribution Methods 0.000 claims description 3
- 238000000605 extraction Methods 0.000 description 8
- 230000000694 effects Effects 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 230000006870 function Effects 0.000 description 4
- 239000000284 extract Substances 0.000 description 3
- 210000003128 head Anatomy 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 238000013316 zoning Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 210000003733 optic disk Anatomy 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000010422 painting Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/122—Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Processing (AREA)
Abstract
A3D image creating method, a photographing apparatus, and a readable storage medium, the method including: controlling a focusing plane to move to an end point position at preset distance intervals from a nearest focusing position in a shooting scene, wherein the end point position is a position where the focusing plane is located when a focal length is equal to an image distance, and acquiring an image shot at the focusing position; dividing the image into a plurality of area images according to the same division rule, and calculating the contrast of the area images; and extracting target area images with the contrast quantity larger than a threshold value in the images, and splicing the target area images into a complete image to form a 3D image of the shooting scene. The embodiment of the invention creates the 3D image through the shooting equipment with the single lens without using the double lenses, and has simple structure and low cost.
Description
Technical field
The present invention relates to capture apparatus field, more particularly to a kind of 3D rendering creation method, capture apparatus and readable deposit
Storage media.
Background technology
3D technology nearly one or two years achieves huge success in film, has also increasingly won liking for consumer.With
The technology development of digital camera, 3D technology also becomes a kind of style of shooting of camera, thus just generates 3D cameras.3D phases
Machine, refer to the camera that relief painting picture or animation can be appreciated with bore hole.The birth of 3D digital cameras, also imply that people can be with
Professional glasses need not be used, with the naked eye can appreciates the effect of stereo-picture.
Similar eyes form the three-dimensional profile of object from object from the point of view of the difference of left and right two.Existing 3D cameras one
As assemble two camera lenses, can be to reproduce stereopsis by the camera lens of left and right two.But existing camera technique can not possibly connect
The function of human eye is bordering on, two lens shootings have regular hour deviation, cause the deviation of left and right image, and 3D effect is undesirable,
And need to handle the image of left and right two simultaneously, its complex disposal process, computationally intensive, cost is high.
The content of the invention
In view of above-mentioned condition, it is necessary to the problem of left and right image deviates be present for 3D rendering of the prior art shooting,
3D rendering creation method, capture apparatus and readable storage medium storing program for executing are provided.
A kind of 3D rendering creation method, including:
Focal plane is controlled to move position to terminal apart from interval from the nearest focusing position in photographed scene with default
Put, the final position is position of the focal length with image distance when equal where the focal plane, and obtains the focusing position and clap
The image taken the photograph;
Described image is divided into multiple area images according to identical division rule, calculates the contrast of the area image
Amount;
Extract described image medium contrast amount and be more than the target area image of threshold value, and the target area image is spliced into
One complete image, to form the 3D rendering of the photographed scene.
Above-mentioned 3D rendering creation method, wherein, the step that the target area image is spliced into a complete image
Suddenly include:
The target area image is added to the corresponding region of iconic model, to be spliced into a complete image.
Above-mentioned 3D rendering creation method, wherein, it is described that described image is divided into multiple areas according to identical division rule
The step of area image, includes:
Described image is divided according to identical area size, region quantity and zone number, to form multiple area images,
Each corresponding zone number of the area image;
The corresponding region that the target area image is added to iconic model, to be spliced into a complete image
Step includes:
With by described image model partition being multiple filling regions according to the division rule;
The target area image is added to the filling region that same area is numbered in described image model.
Above-mentioned 3D rendering creation method, wherein, it is described to control focal plane from the nearest focusing position in photographed scene
With it is default be moved to final position apart from interval when, the 3D rendering creation method also includes step:
The positional information of the focusing position is obtained, and stores the corresponding relation of described image and focusing position information;
The step that the target area image is added to the filling region that same area is numbered in described image model
Suddenly include:
According to the focusing position information, the corresponding target area is sequentially added to identical in described image model
The filling region of zone number.
Above-mentioned 3D rendering creation method, wherein, it is described to control focal plane from the nearest focusing position in photographed scene
Included with default the step of being moved to final position apart from interval:
Photocentre is adjusted to the distance of imaging sensor with the default amplitude of accommodation, so that the focal plane is from photographed scene
In nearest focusing position be moved to final position apart from interval with default.
Above-mentioned 3D rendering creation method, wherein, the contrast amount C of area image calculation formula is:
C=∑ δ (i, j)2Pδ(i,j);
Wherein, δ (i, j)=| i-j |, represent the gray scale difference between adjacent two pixels;Pδ(i, i), it is adjacent pixel
Between gray scale difference be δ pixel distribution probability.
A kind of capture apparatus, including:
Mobile module, for control focal plane from the nearest focusing position in photographed scene with default apart from interval
Final position is moved to, the final position is position of the focal length with image distance when equal where focal plane;
Acquisition module, for obtaining the image of focusing position shooting;
Division module, for described image to be divided into multiple area images according to identical division rule;
Computing module, for calculating the contrast amount of the area image;
Concatenation module, the target area image of threshold value is more than for extracting described image medium contrast amount, and by the target
Area image is spliced into a complete image, to form the 3D rendering of the photographed scene.
Above-mentioned capture apparatus, wherein, concatenation module is specifically used for:
The target area image is added to the corresponding region of iconic model, it is described to be spliced into a complete image
Iconic model is divided into multiple regions according to the division rule.
A kind of readable storage medium storing program for executing, is stored thereon with computer program, is realized when described program is executed by processor above-mentioned
The step of method.
A kind of capture apparatus, including memory, processor and storage can be run on a memory and on a processor
Computer program, the step of realizing the above method during computing device described program.
Each image is divided into by the present embodiment by adjusting the image of the position acquisition different focus position of focal plane
Multiple area images, extraction contrast amount exceed the area image of threshold value, i.e., the area image at the clear position in image.Will extraction
Area image be superimposed to corresponding position in an iconic model, you can obtain three-dimensional image.The present embodiment passes through single mirror
The capture apparatus of head creates 3D rendering, compared with prior art, without using the i.e. reproducible stereopsis of two camera lenses, is not present
The deviation of left and right image, realize good 3D image effects.Moreover, compared with prior art, single-lens image processing process
Simply, amount of calculation is small, and low to the structural requirement of picture pick-up device, can greatly save production cost.
Brief description of the drawings
Fig. 1 is the 3D rendering creation method in first embodiment of the invention;
Fig. 2 is the 3D rendering creation method in second embodiment of the invention;
Fig. 3 is the structured flowchart of the capture apparatus in third embodiment of the invention.
Embodiment
Embodiments of the invention are described below in detail, the example of the embodiment is shown in the drawings, wherein from beginning to end
Same or similar label represents same or similar element or the element with same or like function.Below with reference to attached
The embodiment of figure description is exemplary, is only used for explaining the present invention, and is not considered as limiting the invention.
With reference to following description and accompanying drawing, it will be clear that these and other aspects of embodiments of the invention.In these descriptions
In accompanying drawing, some particular implementations in embodiments of the invention are specifically disclosed, to represent to implement the implementation of the present invention
Some modes of the principle of example, but it is to be understood that the scope of embodiments of the invention is not limited.On the contrary, the present invention
Embodiment includes all changes, modification and the equivalent fallen into the range of the spirit and intension of attached claims.
Fig. 1 is referred to as the 3D rendering creation method in first embodiment of the invention, including step:S11~S13.
Step S11, control focal plane are moved from the nearest focusing position in photographed scene with default apart from interval
Position to terminal, and obtain the image of the focusing position shooting.The final position is that focusing is flat when focal length is equal with image distance
Position where face.
Using single-lens, for example, single-lens camera, mobile phone or flat board etc., which are taken pictures, to be set for capture apparatus in the present embodiment
It is standby.In shooting image, scenery blur-free imaging on egative film is made by focusing.Focusing position is the position where focal plane
Put, by taking camera as an example, during diverse location between camera lens of the focal plane of camera, the difference for obtaining photographed scene is clear flat
The image in face.For example, for the people in photographed scene, the different parts of face are different from the distance of camera, when camera
Focal plane is in the object of the photographed scene focusing position nearest from camera lens, i.e. focal plane is in nose areas, the figure of shooting
Nose areas is apparent from as in and other positions are more fuzzy.Mobile focussing plane, can obtain the portion from camera lens different distance
The picture rich in detail of position.
Based on image-forming principle,Wherein μ is object distance, and υ is image distance, and f is focal length.In the lens system of camera,
When focal length f is constant, when image distance υ increases to equal with focal length from minimum is gradual, object distance μ gradually increases to infinity.
In above-mentioned steps, the photocentre of capture apparatus is adjusted to the distance of imaging sensor with the default amplitude of accommodation, so that
The focal plane is moved to final position with default from the nearest focusing position in photographed scene apart from interval.It is specific real
Shi Shi, by the axial location of certain group eyeglass in moving lens, the photocentre of capture apparatus is adjusted to the distance of imaging sensor.It is logical
The length for changing the distance between convex lens and imaging sensor L is crossed, regulation photocentre sends out image distance to the distance of imaging sensor
Changing.When it is implemented, the position of convex lens can be controlled regulation by magazine controlling organization, such as stepper motor.
Step S12, described image is divided into multiple area images according to identical division rule, calculates the administrative division map
The contrast amount of picture.
In above-mentioned steps, according to the actual requirements by the image of each focal position of collection according to identical division rule
It is divided into multiple regions.I.e. each image is divided into quantity, size, the multiple area images of shape identical.The number of zoning
Amount is configured according to the performance and 3D rendering quality requirement of camera, and general pattern divides thinner, and obtained 3D rendering is more certainly
So, profile is more clear.For cpu performance is preferable, the faster camera of operation, each image can be divided into multiple administrative division maps
The size of picture is not more than the size that human eye can be differentiated, and such as each area image may be sized to 0.1mm*0.1mm.
Contrast amount is relevant with the definition of image, and the bigger area image definition of contrast amount is higher.The contrast amount of image
And the contrast of image, the contrast amount C of each area image calculation formula are:
C=∑ δ (i, j)2Pδ(i,j);
Wherein, δ (i, j)=| i-j |, represent the gray scale difference between adjacent two pixels;Pδ(i, i), it is adjacent pixel
Between gray scale difference be δ pixel distribution probability.
Step S13, extraction described image medium contrast amount are more than the target area image of threshold value, and by the target area figure
As being spliced into a complete image, to form the 3D rendering of the photographed scene.
When it is implemented, obtain clearly area image of the scenery from camera lens different distance position by changing image distance.Carry
Take each image medium contrast amount to exceed the area image of threshold value, and be superimposed to the correspondence position of an iconic model, you can worked as
The 3D rendering that preceding photographed is presented in the form of 3D.
When it is implemented, the size of iconic model is identical with the size of the image of shooting, and according to default division rule
It is divided into multiple regions.The region of iconic model division and the area image of image correspond.By each administrative division map of extraction
The region as corresponding to filling to iconic model.
Further, the controlling switch that 2D image takings and 3D rendering can be set to shoot in the camera, passes through the control
Switch carries out the switching between 2D and 3D screening-modes.Under 3D screening-modes, the depth of field for controlling camera is minimum, when depth of field foot
When enough shallow, the phenomenon of focus can be presented on focal plane, picture is clear, contrast is big, is imaged sharp keen, obtained 3D rendering third dimension
It is stronger.The aperture of the depth of field and camera, penetrate it is relevant away from, lens focus.When therefore carrying out 3D shootings, the aperture setting maximum of camera,
And use telephoto lens.
Each image is divided into by the present embodiment by adjusting the image of the position acquisition different focus position of focal plane
Multiple area images, extraction contrast amount exceed the area image of threshold value, i.e., the area image at the clear position in image.Will extraction
Area image be superimposed to corresponding position in an iconic model, you can obtain three-dimensional image.By to single-lens equipment
The image of shooting is handled, the image of different focus position is divided, calculate contrast amount, extraction target area image,
Splicing target area image can obtain preferable 3D rendering, and its image processing process is simple, and amount of calculation is small, and cost is low.
The present embodiment creates 3D rendering by single-lens capture apparatus, compared with prior art, without using two mirrors
The i.e. reproducible stereopsis of head, in the absence of the deviation of left and right image, realizes good 3D image effects.Moreover, and prior art
Compare, single-lens image processing process is simple, and amount of calculation is small, and low to the structural requirement of picture pick-up device, can greatly save life
Produce cost.
Referring to Fig. 2, be the 3D rendering creation method in second embodiment of the invention, including step S21~S25.
Step S21, control focal plane are moved from the nearest focusing position in photographed scene with default apart from interval
Position to terminal, the final position are position of the focal length with image distance when equal where focal plane, and obtain focusing position bat
The image taken the photograph.
Step S22, obtains the positional information of the focusing position, and stores the correspondence of described image and focusing position information
Relation.
Further, the photocentre of capture apparatus is adjusted to the distance of imaging sensor with the default amplitude of accommodation, so that institute
State focal plane and be moved to final position apart from interval from the nearest focusing position in photographed scene with default.The terminal position
It is set to position of the focal length with image distance when equal where focal plane, i.e. focal plane at infinity.Camera lens passes through in certain journey
The distance of convex lens and imaging sensor is finely tuned on degree, the object distance of imaging is gradually increased into infinity by minimum, you can to incite somebody to action
Focal plane is moved to infinity by some object point nearest from camera lens.Equipment is when finely tuning the position of certain group convex lens, root
According to the distance between convex lens and imaging sensor, current image distance can be calculated, then can obtain current focusing position.Shooting is set
It is standby to store each image and the corresponding relation of image distance.
Step S23, described image is divided according to identical area size, region quantity and zone number, it is multiple to be formed
Area image, each corresponding zone number of the area image.
In above-mentioned steps, the image of each focal position is divided into the area image of identical quantity, size, and often
A zone number corresponding to one area image.Each area image marked region can be numbered during specific implementation.
Step S24, calculates the contrast amount of the area image, and extracts the target that described image medium contrast amount is more than threshold value
Area image.The threshold value of contrast value can determine that definition requires high settable larger according to the definition of system requirements
Threshold value.By by the area image contrast amount in each image compared with threshold value, it is determined that clearly image.
Step S25, with default iconic model is divided into multiple filling regions according to the division rule, according to described right
Burnt position, the corresponding target area is sequentially added to the filling region that same area is numbered in described image model, with
Form 3D rendering.
Each corresponding focusing position information of image, i.e. the area image of each image zooming-out is also to the position that should focus
Confidence ceases.When it is implemented, according to each area image and the corresponding relation of focusing position information, each area image is searched
Positional information, according to focusing position, from as far as near or from being closely sequentially added to the corresponding target area to remote order
The filling region that same area is numbered in described image model.Clearly target area can accurately be determined by the comparison of contrast amount
Area image, so that clearly 3D rendering is presented, by region being numbered the quick splicing of achievable 3D rendering, improve system fortune
Line efficiency.
According to the corresponding relation of image distance and object distance, the focusing position of capture apparatus can be indicated by image distance, each
The corresponding image distance value of image of focusing position shooting.The position of focal plane is adjusted by adjusting image distance, in photographed scene
Nearest focusing position when can also be image distance minimum value, the position where focal plane;Final position in photographed scene
As, when image distance increases to equal with focal length, the position where focal plane.
Further, original image can be deleted after forming region image, to discharge the memory capacity of capture apparatus, improved
The speed of service of system.
The present embodiment obtains the clearly figure of the object of different distance in photographed scene by adjusting the position of focal plane
Picture, the area image that each image medium contrast amount is exceeded to threshold value are filled into iconic model, obtain 3D rendering.That is, by obtaining
The clearly area image at the position of the different distance of object is taken, clearly area image will be superimposed to form complete 3D
Image.For the human body in photographed scene, equipment shooting first image for focal plane in photographed scene from camera lens
The picture rich in detail of nearest focusing position, i.e. nose areas, extract the area image that first image medium contrast amount is more than threshold value
The as area image of nose areas.Second image is image when focal plane is moved to face, eyes, extraction the
The area image that two image medium contrast amounts are more than threshold value is face, the area image of eyes.By that analogy, obtain shooting field
The picture rich in detail at the position of each different distance of human body in scape.The picture rich in detail extracted is compiled according to the region of zoning
Number be added to image template in corresponding region, you can obtain the three-dimensional image of human body.
3D rendering is created by single-lens capture apparatus, compared with prior art, can be again without using two camera lenses
Existing stereopsis, in the absence of the deviation of left and right image, realizes good 3D image effects.Moreover, compared with prior art, Dan Jing
The image processing process of head is simple, and amount of calculation is small, and low to the structural requirement of picture pick-up device, can greatly save production cost.
Referring to Fig. 3, be the capture apparatus in third embodiment of the invention, including:
Mobile module 100, for controlling focal plane from the nearest focusing position in photographed scene with default distance
Interval is moved to final position, and the final position is position of the focal length with image distance when equal where focal plane;
Acquisition module 200, for obtaining the image of focusing position shooting;
Division module 300, for described image to be divided into multiple area images according to identical division rule;
Computing module 400, for calculating the contrast amount of the area image;
Concatenation module 500, the target area image of threshold value is more than for extracting described image medium contrast amount, and by the mesh
Mark area image is spliced into a complete image, to form the 3D rendering of the photographed scene.
Further, the concatenation module 500 is specifically used for:
The target area image is added to the corresponding region of iconic model, it is described to be spliced into a complete image
Iconic model is divided into multiple regions according to the division rule.
Further, the division module 300 is specifically used for, and is compiled according to identical area size, region quantity and region
Number division described image, to form multiple area images, each corresponding zone number of the area image;
The concatenation module specific 500 is used for:
With by described image model partition being multiple filling regions according to the division rule;
The target area image is added to the filling region that same area is numbered in described image model.
Further, acquisition module is additionally operable to obtain the positional information of each focusing position, and stores each described
The corresponding relation of image and focusing position information;
The step that the target area image is added to the filling region that same area is numbered in described image model
Suddenly include:
According to the focusing position information, the corresponding target area is sequentially added to identical in described image model
The filling region of zone number.
Further, the acquisition module 100 be specifically used for the default amplitude of accommodation adjust capture apparatus photocentre to
The distance of imaging sensor so that the focal plane from the nearest focusing position in photographed scene with default apart from interval
It is moved to target location.
The device of the present embodiment, it can be used for the technical scheme for performing any shown embodiments of the method for Fig. 1 to Fig. 2, in fact
Existing principle is similar with technique effect, and here is omitted.
Present invention also offers a kind of readable storage medium storing program for executing, computer program is stored thereon with, it is characterised in that the journey
The step of method of any one in above-described embodiment 1 to 2 is realized when sequence is executed by processor.
Present invention also offers a kind of capture apparatus, including memory, processor and storage are on a memory and can be
The computer program run on processor, any one in above-described embodiment 1 to 2 is realized during the computing device described program
Method the step of.
Expression or logic and/or step described otherwise above herein in flow charts, for example, being considered use
In the order list for the executable instruction for realizing logic function, may be embodied in any computer-readable medium, for
Instruction execution system, device or equipment (such as computer based system including the system of processor or other can be held from instruction
The system of row system, device or equipment instruction fetch and execute instruction) use, or combine these instruction execution systems, device or set
It is standby and use.For the purpose of this specification, " computer-readable medium " can any can be included, store, communicate, propagate or pass
Defeated program is for instruction execution system, device or equipment or the dress used with reference to these instruction execution systems, device or equipment
Put.
The more specifically example (non-exhaustive list) of computer-readable medium includes following:Connected up with one or more
Electrical connection section (electronic installation), portable computer diskette box (magnetic device), random access memory (RAM), read-only storage
(ROM), erasable edit read-only storage (EPROM or flash memory), fiber device, and portable optic disk is read-only deposits
Reservoir (CDROM).In addition, computer-readable medium, which can even is that, to print the paper of described program thereon or other are suitable
Medium, because can then enter edlin, interpretation or if necessary with it for example by carrying out optical scanner to paper or other media
His suitable method is handled electronically to obtain described program, is then stored in computer storage.
It should be appreciated that each several part of the present invention can be realized with hardware, software, firmware or combinations thereof.Above-mentioned
In embodiment, software that multiple steps or method can be performed in memory and by suitable instruction execution system with storage
Or firmware is realized.If, and in another embodiment, can be with well known in the art for example, realized with hardware
Any one of row technology or their combination are realized:With the logic gates for realizing logic function to data-signal
Discrete logic, have suitable combinational logic gate circuit application specific integrated circuit, programmable gate array (PGA), scene
Programmable gate array (FPGA) etc..
In the description of this specification, reference term " one embodiment ", " some embodiments ", " example ", " specifically show
The description of example " or " some examples " etc. means specific features, structure, material or the spy for combining the embodiment or example description
Point is contained at least one embodiment or example of the present invention.In this manual, to the schematic representation of above-mentioned term not
Necessarily refer to identical embodiment or example.Moreover, specific features, structure, material or the feature of description can be any
One or more embodiments or example in combine in an appropriate manner.
Embodiment described above only expresses the several embodiments of the present invention, and its description is more specific and detailed, but simultaneously
Therefore the limitation to the scope of the claims of the present invention can not be interpreted as.It should be pointed out that for one of ordinary skill in the art
For, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to the guarantor of the present invention
Protect scope.Therefore, the protection domain of patent of the present invention should be determined by the appended claims.
Claims (10)
- A kind of 1. 3D rendering creation method, it is characterised in that including:Control focal plane is moved to final position with default from the nearest focusing position in photographed scene apart from interval, institute It is position of the focal length with image distance when equal where the focal plane to state final position, and obtains the figure of the focusing position shooting Picture;Described image is divided into multiple area images according to identical division rule, calculates the contrast amount of the area image;The target area image that described image medium contrast amount is more than threshold value is extracted, and it is complete that the target area image is spliced into one Whole image, to form the 3D rendering of the photographed scene.
- 2. 3D rendering creation method as claimed in claim 1, it is characterised in that described to be spliced into the target area image The step of one complete image, includes:The target area image is added to the corresponding region of iconic model, to be spliced into a complete image.
- 3. 3D rendering creation method as claimed in claim 2, it is characterised in that described to divide described image according to identical The step of regular partition is multiple area images includes:Described image is divided according to identical area size, region quantity and zone number, to form multiple area images, each The corresponding zone number of the area image;The corresponding region that the target area image is added to iconic model, the step of to be spliced into a complete image Including:With by described image model partition being multiple filling regions according to the division rule;The target area image is added to the filling region that same area is numbered in described image model.
- 4. 3D rendering creation method as claimed in claim 3, it is characterised in that the control focal plane is from photographed scene Nearest focusing position with it is default be moved to final position apart from interval when, the 3D rendering creation method also include step Suddenly:The positional information of the focusing position is obtained, and stores the corresponding relation of described image and focusing position information;The described the step of target area image is added to the filling region that same area is numbered in described image model, wraps Include:According to the focusing position information, the corresponding target area is sequentially added to same area in described image model The filling region of numbering.
- 5. 3D rendering creation method as claimed in claim 1, it is characterised in that the control focal plane is from photographed scene Nearest focusing position included with default the step of being moved to final position apart from interval:Photocentre is adjusted to the distance of imaging sensor with the default amplitude of accommodation, so that the focal plane is from photographed scene Nearest focusing position is moved to final position with default apart from interval.
- 6. 3D rendering creation method as claimed in claim 1, it is characterised in that the contrast amount C of area image calculating Formula is:C=∑ δ (i, j)2Pδ(i,j);Wherein, δ (i, j)=| i-j |, represent the gray scale difference between adjacent two pixels;Pδ(i, i), the ash between adjacent pixel The pixel distribution probability that degree difference is δ.
- A kind of 7. capture apparatus, it is characterised in that including:Mobile module, for controlling focal plane to be moved from the nearest focusing position in photographed scene with default apart from interval Position to terminal, the final position are position of the focal length with image distance when equal where the focal plane;Acquisition module, for obtaining the image of the focusing position shooting;Division module, for described image to be divided into multiple area images according to identical division rule;Computing module, for calculating the contrast amount of the area image;Concatenation module, the target area image of threshold value is more than for extracting described image medium contrast amount, and by the target area Image mosaic is a complete image, to form the 3D rendering of the photographed scene.
- 8. capture apparatus as claimed in claim 6, it is characterised in that concatenation module is specifically used for:The target area image is added to the corresponding region of iconic model, to be spliced into a complete image, described image Model is divided into multiple regions according to the division rule.
- 9. a kind of readable storage medium storing program for executing, is stored thereon with computer program, it is characterised in that when described program is executed by processor The step of realizing the method as described in any one in claim 1-6.
- 10. a kind of capture apparatus, including memory, processor and storage are on a memory and the meter that can run on a processor Calculation machine program, it is characterised in that realized during the computing device described program as described in any one in claim 1-6 Method the step of.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710719601.1A CN107360412A (en) | 2017-08-21 | 2017-08-21 | 3D image creation method, photographing apparatus, and readable storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710719601.1A CN107360412A (en) | 2017-08-21 | 2017-08-21 | 3D image creation method, photographing apparatus, and readable storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107360412A true CN107360412A (en) | 2017-11-17 |
Family
ID=60289737
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710719601.1A Pending CN107360412A (en) | 2017-08-21 | 2017-08-21 | 3D image creation method, photographing apparatus, and readable storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107360412A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109104602A (en) * | 2018-09-26 | 2018-12-28 | 成都工业学院 | A kind of stereo scene filming apparatus |
CN109332900A (en) * | 2018-10-22 | 2019-02-15 | 西安云鑫电子科技有限公司 | Away rust by laser method and device, storage medium, electric terminal |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101615289A (en) * | 2009-08-05 | 2009-12-30 | 北京优纳科技有限公司 | The three-dimensional acquisition of biopsy tissues and fusion method of multilayer images |
US20120051625A1 (en) * | 2010-08-23 | 2012-03-01 | Texas Instruments Incorporated | Method and Apparatus for 2D to 3D Conversion Using Scene Classification and Face Detection |
CN103049933A (en) * | 2011-10-17 | 2013-04-17 | 联咏科技股份有限公司 | Image processing device and method thereof |
CN104867125A (en) * | 2015-06-04 | 2015-08-26 | 北京京东尚科信息技术有限公司 | Image obtaining method and image obtaining device |
CN104967778A (en) * | 2015-06-16 | 2015-10-07 | 广东欧珀移动通信有限公司 | Focusing reminding method and terminal |
CN105676432A (en) * | 2016-03-16 | 2016-06-15 | 凯迈(洛阳)测控有限公司 | Optical compensation type long-wave infrared continuous zooming optical system |
CN205581385U (en) * | 2016-03-16 | 2016-09-14 | 凯迈(洛阳)测控有限公司 | Optical compensation formula long wave infrared continuous zoom optical system |
CN106998459A (en) * | 2017-03-15 | 2017-08-01 | 河南师范大学 | A kind of single camera stereoscopic image generation method of continuous vari-focus technology |
-
2017
- 2017-08-21 CN CN201710719601.1A patent/CN107360412A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101615289A (en) * | 2009-08-05 | 2009-12-30 | 北京优纳科技有限公司 | The three-dimensional acquisition of biopsy tissues and fusion method of multilayer images |
US20120051625A1 (en) * | 2010-08-23 | 2012-03-01 | Texas Instruments Incorporated | Method and Apparatus for 2D to 3D Conversion Using Scene Classification and Face Detection |
CN103049933A (en) * | 2011-10-17 | 2013-04-17 | 联咏科技股份有限公司 | Image processing device and method thereof |
CN104867125A (en) * | 2015-06-04 | 2015-08-26 | 北京京东尚科信息技术有限公司 | Image obtaining method and image obtaining device |
CN104967778A (en) * | 2015-06-16 | 2015-10-07 | 广东欧珀移动通信有限公司 | Focusing reminding method and terminal |
CN105676432A (en) * | 2016-03-16 | 2016-06-15 | 凯迈(洛阳)测控有限公司 | Optical compensation type long-wave infrared continuous zooming optical system |
CN205581385U (en) * | 2016-03-16 | 2016-09-14 | 凯迈(洛阳)测控有限公司 | Optical compensation formula long wave infrared continuous zoom optical system |
CN106998459A (en) * | 2017-03-15 | 2017-08-01 | 河南师范大学 | A kind of single camera stereoscopic image generation method of continuous vari-focus technology |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109104602A (en) * | 2018-09-26 | 2018-12-28 | 成都工业学院 | A kind of stereo scene filming apparatus |
CN109332900A (en) * | 2018-10-22 | 2019-02-15 | 西安云鑫电子科技有限公司 | Away rust by laser method and device, storage medium, electric terminal |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107087107B (en) | Image processing apparatus and method based on dual camera | |
CN108076286B (en) | Image blurring method and device, mobile terminal and storage medium | |
US8885091B2 (en) | Imaging device and distance information detecting method | |
KR102032882B1 (en) | Autofocus method, device and electronic apparatus | |
US8830357B2 (en) | Image processing device and image processing method including a blurring process | |
CN104680501B (en) | The method and device of image mosaic | |
CN107945105B (en) | Background blurring processing method, device and equipment | |
CN106454090B (en) | Atomatic focusing method and system based on depth camera | |
CN107358593B (en) | Image forming method and apparatus | |
US8830381B2 (en) | Imaging device and method to provide bokeh effect in captured image by determining distance and focus of captured objects in secondary image sequence | |
RU2565855C1 (en) | Image capturing device, method of controlling said device and programme | |
CN106292162A (en) | Stereographic device and corresponding control methods | |
WO2019105261A1 (en) | Background blurring method and apparatus, and device | |
CN110677621B (en) | Camera calling method and device, storage medium and electronic equipment | |
CN106998459A (en) | A kind of single camera stereoscopic image generation method of continuous vari-focus technology | |
US8810665B2 (en) | Imaging device and method to detect distance information for blocks in secondary images by changing block size | |
CN104184935A (en) | Image shooting device and method | |
US8072487B2 (en) | Picture processing apparatus, picture recording apparatus, method and program thereof | |
KR101549929B1 (en) | Method and apparatus of generating depth map | |
CN111757149B (en) | Video editing method, device, equipment and storage medium | |
CN107809576A (en) | A kind of multi-cam module | |
CN106296574A (en) | 3-d photographs generates method and apparatus | |
JP2019121860A (en) | Image processing apparatus and control method therefor | |
CN107360412A (en) | 3D image creation method, photographing apparatus, and readable storage medium | |
CN114363522A (en) | Photographing method and related device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20171117 |