CN106357980A - Image virtualization processing method and device as well as mobile terminal - Google Patents
Image virtualization processing method and device as well as mobile terminal Download PDFInfo
- Publication number
- CN106357980A CN106357980A CN201610913393.4A CN201610913393A CN106357980A CN 106357980 A CN106357980 A CN 106357980A CN 201610913393 A CN201610913393 A CN 201610913393A CN 106357980 A CN106357980 A CN 106357980A
- Authority
- CN
- China
- Prior art keywords
- depth
- image
- virtualization
- focus area
- difference
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2622—Signal amplitude transition in the zone between image portions, e.g. soft edges
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
Abstract
The invention discloses an image virtualization processing method and an image virtualization processing device as well as a mobile terminal. The method comprises the following steps: acquiring field depth information of an image region outside a focal point region in a pre-viewed picture; calculating a difference value between reference field depth information of a current camera lens module and the field depth information; and if the difference value is greater than a pre-set threshold value, carrying out virtualization processing on the image region outside the focal point region. Therefore, an image is automatically controlled to be virtualized according to a field depth value, so that an image virtualization process is simplified and the quality of the image is promoted.
Description
Technical field
The present invention relates to technical field of image processing, more particularly, to a kind of image virtualization processing method, device and movement are eventually
End.
Background technology
Generally, in order to project the theme of shooting so that the focus of shooting image, on the theme shooting, can blur bat
Take the photograph the background area in picture.
However, the operation at present shooting picture being blurred, it usually needs user's Non-follow control, or need corresponding
Repair in figure application program, choose the region in shooting picture, carry out virtualization process etc., virtualization effect is poor, and blur efficiency
Relatively low.
Content of the invention
The purpose of the present invention is intended at least solve to a certain extent one of above-mentioned technical problem.
For this reason, the first of the present invention purpose is to propose a kind of image virtualization processing method, the method can be according to the depth of field
Value automatically controls image virtualization, simplifies image virtualization process, improves picture quality.
Second object of the present invention is to propose a kind of image virtualization processing meanss.
Third object of the present invention is to propose a kind of mobile terminal.
Fourth object of the present invention is to propose another kind of mobile terminal.
For reaching above-mentioned purpose, embodiment proposes a kind of image virtualization processing method according to a first aspect of the present invention, including
Following steps: the depth of view information of the image-region outside focus area in acquisition preview screen;Calculate the ginseng of current lens module
Examine the difference of depth of view information and described depth of view information, if described difference is more than predetermined threshold value, outside described focus area
Image-region carry out virtualization process.
The image virtualization processing method of the embodiment of the present invention, the image-region outside focus area in acquisition preview screen
Depth of view information, calculates the reference depth of view information of current lens module and the difference of depth of view information, if difference is more than predetermined threshold value,
Then the image-region outside focusing region carries out virtualization process.Thus, image virtualization is automatically controlled according to depth of field value, simplify
Image blurs process, improves picture quality.
Second aspect present invention embodiment provides a kind of image virtualization processing meanss, comprising: acquisition module, for obtaining
The depth of view information of the image-region outside focus area in preview screen;Processing module, for calculating the ginseng of current lens module
Examine the difference of depth of view information and described depth of view information, when described difference is more than predetermined threshold value, outside described focus area
Image-region carries out virtualization process.
The image virtualization processing meanss of the embodiment of the present invention, the image-region outside focus area in acquisition preview screen
Depth of view information, calculates the reference depth of view information of current lens module and the difference of depth of view information, if difference is more than predetermined threshold value,
Then the image-region outside focusing region carries out virtualization process.Thus, image virtualization is automatically controlled according to depth of field value, simplify
Image blurs process, improves picture quality.
Third aspect present invention embodiment provides a kind of mobile terminal, including the image of second aspect present invention embodiment
Virtualization processing meanss.
The mobile terminal of the embodiment of the present invention, the depth of field letter of the image-region outside focus area in acquisition preview screen
Breath, calculates the reference depth of view information of current lens module and the difference of depth of view information, if difference is more than predetermined threshold value, focuses
Image-region outside point region carries out virtualization process.Thus, image virtualization is automatically controlled according to depth of field value, simplify image empty
Change process, improves picture quality.
Fourth aspect present invention embodiment provides another kind of mobile terminal, comprising: housing, processor, memorizer, circuit
Plate and power circuit, wherein, described circuit board is placed in the interior volume that described housing surrounds, described processor and described storage
Device is arranged on described circuit board;Described power circuit, for being each circuit of described mobile terminal or device is powered;Described
Memorizer is used for storing executable program code;Described processor passes through to read the executable program generation of storage in described memorizer
Code running program corresponding with described executable program code, for executing following steps: obtain focus in preview screen
The depth of view information of the image-region outside region;Calculate the reference depth of view information of current lens module and the difference of described depth of view information
Value, if described difference is more than predetermined threshold value, carries out virtualization process to the image-region outside described focus area.
The mobile terminal of the embodiment of the present invention, the depth of field letter of the image-region outside focus area in acquisition preview screen
Breath, calculates the reference depth of view information of current lens module and the difference of depth of view information, if difference is more than predetermined threshold value, focuses
Image-region outside point region carries out virtualization process.Thus, image virtualization is automatically controlled according to depth of field value, simplify image empty
Change process, improves picture quality.
The additional aspect of the present invention and advantage will be set forth in part in the description, and partly will become from the following description
Obtain substantially, or recognized by the practice of the present invention.
Brief description
The above-mentioned and/or additional aspect of the present invention and advantage will become from reference to the description to embodiment for the accompanying drawings below
Substantially and easy to understand, wherein:
Fig. 1 is the flow chart that image according to an embodiment of the invention blurs processing method;
Fig. 2 is the flow chart that image in accordance with another embodiment of the present invention blurs processing method;
Fig. 3 is the structural representation that image according to an embodiment of the invention blurs processing meanss;
Fig. 4 is the structural representation that image in accordance with another embodiment of the present invention blurs processing meanss;
Fig. 5 is the structural representation of the image virtualization processing meanss according to another embodiment of the present invention;
Fig. 6 is the structural representation of mobile terminal according to an embodiment of the invention;And
Fig. 7 is the structural representation of mobile terminal in accordance with another embodiment of the present invention.
Specific embodiment
Embodiments of the invention are described below in detail, the example of described embodiment is shown in the drawings, wherein from start to finish
The element that same or similar label represents same or similar element or has same or like function.Below with reference to attached
The embodiment of figure description is exemplary it is intended to be used for explaining the present invention, and is not considered as limiting the invention.
Below with reference to the accompanying drawings image virtualization processing method, device and the mobile terminal of the embodiment of the present invention are described.
Generally, the background area in order to project the target object of shooting, outside the target object place focus area that will shoot
It is conventional means that domain carries out virtualization, and it is related to the depth of field to blur effect.
Wherein, after the target object shooting is focused on, one section before and after the focus area that target object is located
The spatial depth scope of the blur-free imaging that human eye is allowed is the depth of field, and the image-region outside the depth of field can be blurred, wherein, in Jiao
Before point region, the scope of blur-free imaging is foreground depth of field, and after focus area, the scope of blur-free imaging is the background depth of field.
And in actual applications, in follow shot application scenarios, sometimes using the larger depth of field preferably, such as shoot wind
Jing Shi, the picture quality shot with the larger depth of field preferably, sometimes with the less depth of field preferably, when such as shooting portrait, is used
The picture quality that the less depth of field is shot is preferable.
Because the aperture of the size follow shot of the depth of field, lens focus, object distance are relevant, aperture is bigger, and that is, f-number is less, scape
Deep less, aperture is less, and that is, f-number is bigger, and the depth of field is bigger, and lens focus are longer, and the depth of field is less etc..
As such, it will be understood that mobile terminal is if it is intended to take the effect of the little depth of field it is necessary to use large aperture, focal length
Away from, near target object if it is desired to shoot the big depth of field effect it is necessary to using little aperture, short focus, away from mesh
Mark object is.
But for corresponding mobile terminal, because its aperture and camera lens are all fixing, it is only capable of when shooting changing quilt
Shoot the distance between object and camera lens, the closer to target object, its field depth is more shallow, could preferably realize blurring, dash forward
Go out to shoot object, when the target object distance shooting is further away from photographic head, then be difficult to blur effect.
Thus, in correlation technique, the virtualization of usual Non-follow control shooting image, or in conventional image procossing application journey
In sequence, realize the virtualization of image, image virtualization operating efficiency is low, and it is bad to blur effect.
In order to solve the above problems, we propose a kind of image virtualization processing method, can be automatically controlled according to depth of field value
Image blurs, and simplification figure picture blurs process, lifts picture quality, specific as follows:
Fig. 1 is the flow chart that image according to an embodiment of the invention blurs processing method.As shown in figure 1, this image
Virtualization processing method includes:
S110, the depth of view information of the image-region outside focus area in acquisition preview screen.
Specifically, in order to realize the depth of view information according to image, realize the virtualization to shooting image and process, obtain preview and draw
The depth of view information of the image-region outside focus area in face, wherein, depth of view information includes image-region before focus area
Background depth of view information of image-region after foreground depth of field information and focus area etc..
It should be noted that according to the difference of concrete application scene, obtaining the figure outside the focus area in preview screen
As the mode of the depth of view information in region is different,
The first example, can obtain the relevant parameter of shooting, to be calculated in preview screen according to the formula shooting photographic head
The depth of view information of the image-region outside focus area.
In this example, can obtain shooting photographic head allows that disperse circular diameter, f-number, focal length, focal distance etc. are joined
Number, thus according to formula: foreground depth of field=(f-number * allow disperse circular diameter * focal distance square)/(focal length square+
F-number * allows disperse circular diameter * focal distance) calculate foreground depth of field, and according to the formula background depth of field=(f-number * holds
Permitted disperse circular diameter * focal distance square)/(focal length square-f-number * allows disperse circular diameter * focal distance) calculate
Go out the background depth of field.
Second example is it may be determined that the depth of field value of each pixel outside focus area generate depth of field map, root
Obtain the foreground depth of field information of the image-region before focus area according to depth of field map, and, focus is obtained according to depth of field map
The background depth of view information of image-region after region.
S120, calculates the reference depth of view information of current lens module and the difference of depth of view information, if difference is more than preset
Threshold value, then the image-region outside focusing region carry out virtualization process.
Specifically, the reference depth of view information of current camera module and the difference of depth of view information are calculated, thus relatively more current
The depth of field of the image-region outside focus area, if in the range of the blur-free imaging of camera module.Wherein, camera module
May include one or more photographic head.
Thus, if the scape of the image-region outside focus area in the reference depth of view information of camera lens module and preview screen
Deeply convince that the difference of breath is more than predetermined threshold value, then show that this image-region is not at the scope of the blur-free imaging of camera lens module, thus
Virtualization process is carried out to the image-region outside this focus area, with the image of prominent focus area.
Wherein, the reference depth of view information of above-mentioned camera lens module corresponds to camera lens module under present filming scene, clear one-tenth
The depth of field value of picture, including foreground depth of field value and background depth of field value.
In addition, the above-mentioned image-region comparing difference with predetermined threshold value to judge outside current focus area, if
Exceed the field depth of the current blur-free imaging of camera lens module, and the difference according to concrete application scene, the figure after focus area
Depth of view information as region can be with positive number or negative number representation.
Thus, if the foreground depth of field value of the image-region before current focus area, after current focus area
The background depth of field value of image-region is the absolute value representation of depth of field value, then predetermined threshold value is 0.
Thus by judging whether foreground depth of field value is more than with reference to the difference of the foreground depth of field value in the depth of field with camera lens module
0, judge whether the image-region before current focus area exceedes the field depth of the current blur-free imaging of camera lens module, by sentencing
Whether brokeback scape depth of field value is more than 0 with camera lens module with reference to the difference of the background depth of field value in the depth of field, judges current focus area
Whether image-region afterwards exceedes the field depth of the current blur-free imaging of camera lens module.
Or, if the foreground depth of field value of the image-region before current focus area is represented with the positive number more than 0, with work as
The background depth of field value of the image-region after front focus region is the negative number representation less than 0.
Thus by judging whether foreground depth of field value is more than with reference to the difference of the foreground depth of field value in the depth of field with camera lens module
0, judge whether the image-region before current focus area exceedes the field depth of the current blur-free imaging of camera lens module, by sentencing
Whether brokeback scape depth of field value is more than 0 with camera lens module with reference to the absolute value of the difference of background depth of field value in the depth of field, judges current
Whether the image-region after focus area exceedes the field depth of the current blur-free imaging of camera lens module.
In sum, the image virtualization processing method of the embodiment of the present invention, obtains outside focus area in preview screen
The depth of view information of image-region, calculates the reference depth of view information of current lens module and the difference of depth of view information, if difference is big
In predetermined threshold value, then the image-region outside focusing region carries out virtualization process.Thus, image is automatically controlled according to depth of field value
Virtualization, simplifies image virtualization process, improves picture quality.
Based on above example it should be appreciated that because camera module includes one or more photographic head, and have
There is the camera module of different number of photographic head, the depth of view information of image-region outside focus area in acquisition preview screen
Specific embodiment is different.
Thus, below to include the camera module of two photographic head, obtain the image outside focus area in preview screen
The depth of view information implementation process in region, the image virtualization processing method of the description embodiment of the present invention.
Wherein, including the camera module of two post-positioned pick-up heads, obtain the image district outside focus area in preview screen
The depth of view information in domain be by determine focus area outside image-region depth of field map realize, specific as follows:
Fig. 2 is the flow chart that image in accordance with another embodiment of the present invention blurs processing method, as shown in Fig. 2 this figure
As virtualization processing method includes:
S210, the scape of the image-region outside focus area is determined according to the preview image data that dual camera obtains respectively
Deep map.
S220, determines the foreground depth of field before focus area according to depth of field map, and/or, the background after focus area
The depth of field.
Specifically, when being shot by dual camera, subject is imaged through dual camera and generates preview image data, should
Coordinate figure, pixel value of each pixel etc. is may include in preview image data.
Thus, the image-region outside focus area being determined according to the preview image data that dual camera obtains respectively
Depth of field map.
Wherein, the set-up mode of dual camera can be horizontally disposed, vertically setting etc..
It should be noted that according to the difference of concrete application scene, can be true according to the different parameters in preview image data
The depth of field map of the image focusing outside region, is illustrated below:
Due to the position of two photographic head and differ, thus, two post-positioned pick-up heads relatively with the target object shooting
There is certain differential seat angle and range difference, the preview image data that therefore the two obtains there is also certain phase contrast.
For example, for a point on photographic subjects object, in the preview image data of photographic head 1, a point is corresponding
Pixel point coordinates is (30,50), and in the preview image data of photographic head 2, a point corresponding pixel point coordinates is (30,48),
The phase contrast of a point corresponding pixel in two preview image data is 50-48=2.
In this example, the pass of depth of view information and phase contrast can be set up previously according to experimental data or photographic head parameter
System, and then, can be searched according to phase contrast in the preview image data that two photographic head obtain for picture point each in preview image
Corresponding depth of view information.
For example, the corresponding phase contrast 2 for above-mentioned a point, if inquire corresponding according to default corresponding relation
The depth of field is 5 meters, then in preview screen, the corresponding depth of view information of a point is 5 meters.Thus, can get each picture in current preview picture
The depth of view information of vegetarian refreshments, i.e. the depth of field map of the image-region outside acquisition focus area.
And then, after the depth of field map of the image-region outside obtaining focus area, can further determine that focus area it
The foreground depth of field information of front image-region, and the background depth of view information after focus area.
It should be noted that according to the difference of concrete application scene, the prospect of the image-region outside acquisition focus area
Depth of view information is different with the mode of background depth of view information, is illustrated below:
The first example, in recognizable preview screen, foreground area before focus area, and by positioned at foreground area
The depth of view information of pixel constitutes the depth of view information of foreground area.And, the background in identification preview screen, after focus area
Region, and it is made up of the depth of view information of background area the depth of view information of the pixel positioned at background area.
Second example, the phase contrast of each pixel in the preview image data that can be obtained respectively according to dual camera,
Inquire about the depth of view information of each pixel.And then it may be determined that in current preview picture, it is located at each pixel of focus area
Depth of view information, obtains the focus depth of view information of focusing area in present preview image.
And then, in this example, focus depth of field meansigma methodss are obtained according to focus depth of view information, according to default regulation coefficient
The focus depth of field meansigma methodss of focusing depth of view information are adjusted, and obtain the foreground depth of field letter of foreground area in present preview image
Breath, so using the region depth of view information outside foreground depth of field information as background area background depth of view information.
Such as, focus depth of field meansigma methodss can be obtained the field depth upper limit plus default regulation coefficient, by the focus depth of field
Meansigma methodss deduct default regulation coefficient and obtain field depth lower limit.Between field depth lower limit and the field depth upper limit
Depth of view information is the foreground depth of field information of foreground area in current preview picture, and then by the region outside foreground depth of field information
Depth of view information is as the background depth of view information of background area.
Again such as, focus depth of field meansigma methodss can be multiplied by default regulation coefficient and obtain the field depth upper limit, by focus scape
Deep meansigma methodss obtain field depth lower limit divided by default regulation coefficient.Between field depth lower limit and the field depth upper limit
Depth of view information be the foreground depth of field information of foreground area in present preview image, and then by the area outside foreground depth of field information
Domain depth of view information is as the background depth of view information of background area.
S230, the background depth of field after calculating focus area and first after current lens module focus are with reference to the depth of field
Whether the first difference is more than default first threshold.
Specifically, after obtaining the background depth of field after focus area, calculate the background depth of field after focus area and work as
Whether the first difference of the first reference information after front camera lens module focus is more than the first predetermined threshold value, to judge current focus
Whether the background area after region is in the range of camera lens module blur-free imaging.
Wherein, the first reference information after camera lens module focus, is under present filming scene, camera lens module blur-free imaging
The background depth of field.
S240, if the first difference is more than first threshold, the rear in focusing region carries out virtualization process.
Specifically, if the first difference is more than first threshold, show the back of the body in current shooting picture, outside focus area
Scene area, can not be in the background field depth of blur-free imaging in camera lens module, thus focusing regional background region is blurred
Process.
S250, the foreground depth of field before calculating focus area and second before current lens module focus are with reference to the depth of field
Whether the second difference is more than default Second Threshold.
Specifically, after obtaining the foreground depth of field before focus area, calculate the foreground depth of field before focus area and work as
Whether the second difference of the second reference information before front camera lens module focus is more than the second predetermined threshold value, to judge current focus
Whether the foreground area before region is in the range of camera lens module blur-free imaging.
Wherein, the second reference information before camera lens module focus, is under present filming scene, camera lens module blur-free imaging
Foreground depth of field.
S260, if the second difference is more than Second Threshold, the front in focusing region carries out virtualization process.
Specifically, if the second difference be more than Second Threshold, show in current shooting picture, outside focus area before
Scene area, can not be in the background field depth of blur-free imaging in camera lens module, thus focusing region fore carries out virtualization process.
In addition, in one embodiment of the invention, so that the virtualization to preview image processes the sense that more has levels, can
Reference depth of view information for current lens module and the size of the difference of depth of view information, determine intensity of virtualization etc..
Specifically, virtualization strategy can be pre-set, this virtualization strategy includes difference and virtualization corresponding with difference
Coefficient, wherein, virtualization coefficient is relevant with virtualization intensity, and virtualization coefficient is bigger, and virtualization intensity is higher.For example, in this virtualization
In strategy, difference 2 corresponds to virtualization coefficient 2, difference 3 corresponds to virtualization coefficient 4 etc..
And then, according to the reference depth of view information of calculated current lens module, poor with the depth of field outside focus area
Value, inquiry default virtualization strategy obtain virtualization coefficient corresponding with difference, and then according to blur coefficient focusing region outside
Image-region carry out virtualization process.
Such as, calculate and can obtain calculating first after the background depth of field and the current lens module focus after focus area
First difference of the reference depth of field is more than first threshold, and this first difference is 2, then inquire about default virtualization strategy acquisition corresponding
Virtualization coefficient is 2, and then according to this virtualization coefficient corresponding virtualization intensity, background area is blurred.
Again such as, calculate the second reference depth of field before foreground depth of field and the current lens module focus before focus area
The second difference be more than default Second Threshold, and the second difference be 3, then inquire about default virtualization strategy obtain corresponding virtualization
Coefficient is 4, and then according to this virtualization coefficient corresponding virtualization intensity, foreground area is blurred.
In sum, the image virtualization processing method of the embodiment of the present invention, includes dual camera in current lens module, then
The depth of field map of the image-region outside focus area is determined according to the preview image data that dual camera obtains respectively, and according to
Depth of field map determines the foreground depth of field before focus area, and/or, the background depth of field after focus area, thus, can be according to meter
Calculate the reference depth of view information of current lens module and the difference of depth of view information, and when difference is more than predetermined threshold value, to corresponding
Image-region carries out virtualization process.Thus, image virtualization is automatically controlled according to depth of field value, simplify image virtualization process, lifting
Picture quality.
In order to realize above-described embodiment, the invention allows for a kind of image virtualization processing meanss, Fig. 3 is according to the present invention
The image of one embodiment blurs the structural representation of processing meanss, as shown in figure 3, this image virtualization processing meanss include obtaining
Module 110 and processing module 120.
Wherein, acquisition module 110, for obtaining the depth of view information of the image-region outside focus area in preview screen.
Processing module 120, for calculating the reference depth of view information of current lens module and the difference of depth of view information, in difference
During more than predetermined threshold value, the image-region outside focusing region carries out virtualization process.
It should be noted that the aforementioned explanation that image is blurred with processing method embodiment, it is also applied for the present invention real
The image applying example blurs the explanation of processing meanss, and image of the present invention blurs the details not disclosed in processing meanss embodiment,
Will not be described here.
In sum, the image virtualization processing meanss of the embodiment of the present invention, obtain outside focus area in preview screen
The depth of view information of image-region, calculates the reference depth of view information of current lens module and the difference of depth of view information, if difference is big
In predetermined threshold value, then the image-region outside focusing region carries out virtualization process.Thus, image is automatically controlled according to depth of field value
Virtualization, simplifies image virtualization process, improves picture quality.
Fig. 4 is the structural representation that image in accordance with another embodiment of the present invention blurs processing meanss, as shown in figure 4,
On the basis of as shown in Figure 3, acquisition module 110 includes first acquisition unit 111 and determining unit 112, and processing module 120 is wrapped
Include the first computing unit 121, first processing units 122, the second computing unit 123 and second processing unit 124.
Wherein, first acquisition unit 111, the preview image data for being obtained respectively according to dual camera determines focal area
The depth of field map of the image-region outside domain.
Determining unit 112, for determining the foreground depth of field before focus area according to depth of field map, and/or, focus area
The background depth of field afterwards.
First computing unit 121, after calculating the background depth of field and the current lens module focus after focus area
First whether be more than default first threshold with reference to the first difference of the depth of field.
First processing units 122, for when the first difference is more than first threshold, the rear in focusing region is blurred
Process.
Second computing unit 123, before calculating foreground depth of field and the current lens module focus before focus area
Second whether be more than default Second Threshold with reference to the second difference of the depth of field.
Second processing unit 124, for when the second difference is more than Second Threshold, the front in focusing region is blurred
Process.
In addition, in one embodiment of the invention, so that the virtualization to preview image processes the sense that more has levels, can
Reference depth of view information for current lens module and the size of the difference of depth of view information, determine intensity of virtualization etc..
Fig. 5 is the structural representation of the image virtualization processing meanss according to another embodiment of the present invention, as shown in figure 5,
On the basis of as shown in Figure 3, processing module 120 includes second acquisition unit 125 and the 3rd processing unit 126.
Wherein, second 125 are obtained, for virtualization coefficient corresponding with difference is obtained according to default virtualization strategy.
3rd processing unit 126, for carrying out virtualization process according to the image-region outside virtualization coefficient focusing region.
It should be noted that the aforementioned explanation that image is blurred with processing method embodiment, it is also applied for the present invention real
The image applying example blurs the explanation of processing meanss, and image of the present invention blurs the details not disclosed in processing meanss embodiment,
Will not be described here.
In sum, the image virtualization processing meanss of the embodiment of the present invention, include dual camera in current lens module, then
The depth of field map of the image-region outside focus area is determined according to the preview image data that dual camera obtains respectively, and according to
Depth of field map determines the foreground depth of field before focus area, and/or, the background depth of field after focus area, thus, can be according to meter
Calculate the reference depth of view information of current lens module and the difference of depth of view information, and when difference is more than predetermined threshold value, to corresponding
Image-region carries out virtualization process.Thus, image virtualization is automatically controlled according to depth of field value, simplify image virtualization process, lifting
Picture quality.
In order to realize above-described embodiment, the invention allows for a kind of mobile terminal.Fig. 6 is according to one enforcement of the present invention
The structural representation of the mobile terminal of example, as shown in fig. 6, this mobile terminal includes image virtualization processing meanss 100.
It should be noted that the image of aforementioned combination Fig. 3-Fig. 5 description blurs the description of processing meanss, it is also applied for this
The image virtualization processing meanss 100 of bright embodiment, will not be described here.
In sum, the mobile terminal of the embodiment of the present invention, obtains the image-region outside focus area in preview screen
Depth of view information, calculate the difference of reference depth of view information and the depth of view information of current lens module, if difference is more than presets threshold
Value, then the image-region outside focusing region carries out virtualization process.Thus, image virtualization, letter are automatically controlled according to depth of field value
Change image virtualization process, improve picture quality.
In order to realize above-described embodiment, the invention allows for another kind of mobile terminal.Fig. 7 be according to the present invention another
The structural representation of the mobile terminal of embodiment.As shown in Figure 7 this mobile terminal 1 000 include housing 1100, processor 1111,
Memorizer 1112, circuit board 1110 and power circuit 1120, wherein, circuit board 1110 is placed in the space that housing 1100 surrounds
Portion, processor 1111 and memorizer 1112 are arranged on circuit boards;Power circuit 1120, for each circuit for mobile terminal
Or device powers;Memorizer 1112 is used for storing executable program code;Processor 1111 passes through to read in memorizer 1112 to be deposited
The executable program code of storage running program corresponding with executable program code, for executing following steps:
The depth of view information of the image-region outside focus area in acquisition preview screen;
Calculate the reference depth of view information of current lens module and the difference of depth of view information, if difference is more than predetermined threshold value,
Then the image-region outside focusing region carries out virtualization process.
It should be noted that the aforementioned explanation that image is blurred with processing method embodiment, it is also applied for the present invention real
Apply the explanation of the mobile terminal of example, the details not disclosed in mobile terminal embodiment of the present invention, will not be described here.
In sum, the mobile terminal of the embodiment of the present invention, obtains the image-region outside focus area in preview screen
Depth of view information, calculate the difference of reference depth of view information and the depth of view information of current lens module, if difference is more than presets threshold
Value, then the image-region outside focusing region carries out virtualization process.Thus, image virtualization, letter are automatically controlled according to depth of field value
Change image virtualization process, improve picture quality.
Additionally, term " first ", " second " are only used for describing purpose, and it is not intended that indicating or hint relative importance
Or the implicit quantity indicating indicated technical characteristic.Thus, define " first ", the feature of " second " can express or
Implicitly include at least one this feature.In describing the invention, " multiple " are meant that at least two, such as two, three
Individual etc., unless otherwise expressly limited specifically.
In the description of this specification, reference term " embodiment ", " some embodiments ", " example ", " specifically show
The description of example " or " some examples " etc. means specific features, structure, material or the spy describing with reference to this embodiment or example
Point is contained at least one embodiment or the example of the present invention.In this manual, to the schematic representation of above-mentioned term not
Identical embodiment or example must be directed to.And, the specific features of description, structure, material or feature can be in office
Combine in an appropriate manner in one or more embodiments or example.Additionally, in the case of not conflicting, the skill of this area
The feature of the different embodiments described in this specification or example and different embodiment or example can be tied by art personnel
Close and combine.
Although embodiments of the invention have been shown and described above it is to be understood that above-described embodiment is example
Property it is impossible to be interpreted as limitation of the present invention, those of ordinary skill in the art within the scope of the invention can be to above-mentioned
Embodiment is changed, changes, replacing and modification.
Claims (12)
1. a kind of image virtualization processing method is it is characterised in that comprise the following steps:
The depth of view information of the image-region outside focus area in acquisition preview screen;
Calculate the reference depth of view information of current lens module and the difference of described depth of view information, if described difference is more than default threshold
Value, then carry out virtualization process to the image-region outside described focus area.
If 2. the method for claim 1 is it is characterised in that current lens module includes dual camera, described acquisition
The depth of view information of the image-region outside focus area in preview screen, comprising:
The scape of the image-region outside described focus area is determined according to the preview image data that described dual camera obtains respectively
Deep map;
Foreground depth of field before described focus area is determined according to described depth of field map, and/or, the back of the body after described focus area
The scape depth of field.
3. the method for claim 1 is it is characterised in that the reference depth of view information of described calculating current lens module and institute
State the difference of depth of view information, if described difference is more than predetermined threshold value, the image-region outside described focus area is carried out
Virtualization is processed, comprising:
Calculate first of the first reference depth of field after the background depth of field and the current lens module focus after described focus area
Whether difference is more than default first threshold;
If described first difference is more than described first threshold, virtualization process is carried out to the rear of described focus area.
4. the method for claim 1 is it is characterised in that the reference depth of view information of described calculating current lens module and institute
State the difference of depth of view information, if described difference is more than predetermined threshold value, the image-region outside described focus area is carried out
Virtualization is processed, comprising:
Calculate second of the second reference depth of field before foreground depth of field and the current lens module focus before described focus area
Whether difference is more than default Second Threshold;
If described second difference is more than described Second Threshold, virtualization process is carried out to the front of described focus area.
5. described method as arbitrary in claim 1-4 is it is characterised in that carry out to the image-region outside described focus area
Virtualization is processed, comprising:
Virtualization coefficient corresponding with described difference is obtained according to default virtualization strategy;
Virtualization process is carried out to the image-region outside described focus area according to described virtualization coefficient.
6. a kind of image virtualization processing meanss are it is characterised in that include:
Acquisition module, for obtaining the depth of view information of the image-region outside focus area in preview screen;
Processing module, for calculating the reference depth of view information of current lens module and the difference of described depth of view information, in described difference
When value is more than predetermined threshold value, virtualization process is carried out to the image-region outside described focus area.
If 7. device as claimed in claim 6 is it is characterised in that current lens module includes dual camera, described acquisition
Module includes:
First acquisition unit, for according to the preview image data that described dual camera obtains respectively determine described focus area it
The depth of field map of outer image-region;
Determining unit, for determining the foreground depth of field before described focus area according to described depth of field map, and/or, described Jiao
The background depth of field after point region.
8. device as claimed in claim 6 is it is characterised in that described processing module includes:
First computing unit, for calculating the after the background depth of field after described focus area and current lens module focus
Whether the first difference of the one reference depth of field is more than default first threshold;
First processing units, for when described first difference is more than described first threshold, entering to the rear of described focus area
Row virtualization is processed.
9. device as claimed in claim 6 is it is characterised in that described processing module includes:
Second computing unit, for calculating the before foreground depth of field before described focus area and current lens module focus
Whether the second difference of the two reference depth of field is more than default Second Threshold;
Second processing unit, for when described second difference is more than described Second Threshold, entering to the front of described focus area
Row virtualization is processed.
10. described device as arbitrary in claim 6-9 is it is characterised in that described processing module includes:
Second acquisition unit, for obtaining virtualization coefficient corresponding with described difference according to default virtualization strategy;
3rd processing unit, for carrying out at virtualization to the image-region outside described focus area according to described virtualization coefficient
Reason.
A kind of 11. mobile terminals are it is characterised in that include described image virtualization processing meanss as arbitrary in claim 6-10.
A kind of 12. mobile terminals are it is characterised in that include: housing, processor, memorizer, circuit board and power circuit, wherein,
Described circuit board is placed in the interior volume that described housing surrounds, and described processor and described memorizer are arranged on described circuit board
On;Described power circuit, for being each circuit of described mobile terminal or device is powered;Described memorizer is used for storing can be held
Line program code;Described processor is run by the executable program code of storage in the described memorizer of reading and is held with described
The corresponding program of line program code, for execution following steps:
The depth of view information of the image-region outside focus area in acquisition preview screen;
Calculate the reference depth of view information of current lens module and the difference of described depth of view information, if described difference is more than default threshold
Value, then carry out virtualization process to the image-region outside described focus area.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610913393.4A CN106357980A (en) | 2016-10-19 | 2016-10-19 | Image virtualization processing method and device as well as mobile terminal |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610913393.4A CN106357980A (en) | 2016-10-19 | 2016-10-19 | Image virtualization processing method and device as well as mobile terminal |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106357980A true CN106357980A (en) | 2017-01-25 |
Family
ID=57865163
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610913393.4A Pending CN106357980A (en) | 2016-10-19 | 2016-10-19 | Image virtualization processing method and device as well as mobile terminal |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106357980A (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106960413A (en) * | 2017-03-24 | 2017-07-18 | 深圳市金立通信设备有限公司 | A kind of method and terminal of image virtualization |
CN107085825A (en) * | 2017-05-27 | 2017-08-22 | 成都通甲优博科技有限责任公司 | Image weakening method, device and electronic equipment |
CN107454328A (en) * | 2017-08-24 | 2017-12-08 | 广东欧珀移动通信有限公司 | Image processing method, device, computer-readable recording medium and computer equipment |
CN107820019A (en) * | 2017-11-30 | 2018-03-20 | 广东欧珀移动通信有限公司 | Blur image acquiring method, device and equipment |
CN108076286A (en) * | 2017-11-30 | 2018-05-25 | 广东欧珀移动通信有限公司 | Image weakening method, device, mobile terminal and storage medium |
CN108900763A (en) * | 2018-05-30 | 2018-11-27 | Oppo(重庆)智能科技有限公司 | Filming apparatus, electronic equipment and image acquiring method |
CN108900738A (en) * | 2018-05-31 | 2018-11-27 | Oppo(重庆)智能科技有限公司 | Filming apparatus and electronic equipment |
CN110910304A (en) * | 2019-11-08 | 2020-03-24 | 北京达佳互联信息技术有限公司 | Image processing method, image processing apparatus, electronic device, and medium |
CN111246092A (en) * | 2020-01-16 | 2020-06-05 | Oppo广东移动通信有限公司 | Image processing method, image processing device, storage medium and electronic equipment |
CN111866369A (en) * | 2020-05-28 | 2020-10-30 | 北京迈格威科技有限公司 | Image processing method and device |
US10997696B2 (en) | 2017-11-30 | 2021-05-04 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Image processing method, apparatus and device |
CN114143442A (en) * | 2020-09-03 | 2022-03-04 | 武汉Tcl集团工业研究院有限公司 | Image blurring method, computer device, computer-readable storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104333700A (en) * | 2014-11-28 | 2015-02-04 | 广东欧珀移动通信有限公司 | Image blurring method and image blurring device |
CN104463775A (en) * | 2014-10-31 | 2015-03-25 | 小米科技有限责任公司 | Device and method for achieving depth-of-field effect of image |
CN105100615A (en) * | 2015-07-24 | 2015-11-25 | 青岛海信移动通信技术股份有限公司 | Image preview method, apparatus and terminal |
CN105303514A (en) * | 2014-06-17 | 2016-02-03 | 腾讯科技(深圳)有限公司 | Image processing method and apparatus |
CN105979165A (en) * | 2016-06-02 | 2016-09-28 | 广东欧珀移动通信有限公司 | Blurred photos generation method, blurred photos generation device and mobile terminal |
-
2016
- 2016-10-19 CN CN201610913393.4A patent/CN106357980A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105303514A (en) * | 2014-06-17 | 2016-02-03 | 腾讯科技(深圳)有限公司 | Image processing method and apparatus |
CN104463775A (en) * | 2014-10-31 | 2015-03-25 | 小米科技有限责任公司 | Device and method for achieving depth-of-field effect of image |
CN104333700A (en) * | 2014-11-28 | 2015-02-04 | 广东欧珀移动通信有限公司 | Image blurring method and image blurring device |
CN105100615A (en) * | 2015-07-24 | 2015-11-25 | 青岛海信移动通信技术股份有限公司 | Image preview method, apparatus and terminal |
CN105979165A (en) * | 2016-06-02 | 2016-09-28 | 广东欧珀移动通信有限公司 | Blurred photos generation method, blurred photos generation device and mobile terminal |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106960413A (en) * | 2017-03-24 | 2017-07-18 | 深圳市金立通信设备有限公司 | A kind of method and terminal of image virtualization |
CN107085825A (en) * | 2017-05-27 | 2017-08-22 | 成都通甲优博科技有限责任公司 | Image weakening method, device and electronic equipment |
CN107454328B (en) * | 2017-08-24 | 2019-08-23 | Oppo广东移动通信有限公司 | Image processing method, device, computer readable storage medium and computer equipment |
CN107454328A (en) * | 2017-08-24 | 2017-12-08 | 广东欧珀移动通信有限公司 | Image processing method, device, computer-readable recording medium and computer equipment |
CN108076286B (en) * | 2017-11-30 | 2019-12-27 | Oppo广东移动通信有限公司 | Image blurring method and device, mobile terminal and storage medium |
WO2019105214A1 (en) * | 2017-11-30 | 2019-06-06 | Oppo广东移动通信有限公司 | Image blurring method and apparatus, mobile terminal and storage medium |
CN108076286A (en) * | 2017-11-30 | 2018-05-25 | 广东欧珀移动通信有限公司 | Image weakening method, device, mobile terminal and storage medium |
CN107820019A (en) * | 2017-11-30 | 2018-03-20 | 广东欧珀移动通信有限公司 | Blur image acquiring method, device and equipment |
US10997696B2 (en) | 2017-11-30 | 2021-05-04 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Image processing method, apparatus and device |
CN108900763A (en) * | 2018-05-30 | 2018-11-27 | Oppo(重庆)智能科技有限公司 | Filming apparatus, electronic equipment and image acquiring method |
CN108900763B (en) * | 2018-05-30 | 2022-03-22 | Oppo(重庆)智能科技有限公司 | Shooting device, electronic equipment and image acquisition method |
CN108900738A (en) * | 2018-05-31 | 2018-11-27 | Oppo(重庆)智能科技有限公司 | Filming apparatus and electronic equipment |
CN108900738B (en) * | 2018-05-31 | 2021-01-15 | Oppo(重庆)智能科技有限公司 | Imaging device and electronic apparatus |
CN110910304A (en) * | 2019-11-08 | 2020-03-24 | 北京达佳互联信息技术有限公司 | Image processing method, image processing apparatus, electronic device, and medium |
CN110910304B (en) * | 2019-11-08 | 2023-12-22 | 北京达佳互联信息技术有限公司 | Image processing method, device, electronic equipment and medium |
CN111246092B (en) * | 2020-01-16 | 2021-07-20 | Oppo广东移动通信有限公司 | Image processing method, image processing device, storage medium and electronic equipment |
CN111246092A (en) * | 2020-01-16 | 2020-06-05 | Oppo广东移动通信有限公司 | Image processing method, image processing device, storage medium and electronic equipment |
CN111866369A (en) * | 2020-05-28 | 2020-10-30 | 北京迈格威科技有限公司 | Image processing method and device |
CN114143442A (en) * | 2020-09-03 | 2022-03-04 | 武汉Tcl集团工业研究院有限公司 | Image blurring method, computer device, computer-readable storage medium |
CN114143442B (en) * | 2020-09-03 | 2023-08-01 | 武汉Tcl集团工业研究院有限公司 | Image blurring method, computer device, and computer-readable storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106357980A (en) | Image virtualization processing method and device as well as mobile terminal | |
US10997696B2 (en) | Image processing method, apparatus and device | |
CN108076286B (en) | Image blurring method and device, mobile terminal and storage medium | |
CN107613199B (en) | Blurred photo generation method and device and mobile terminal | |
CN105979165A (en) | Blurred photos generation method, blurred photos generation device and mobile terminal | |
US9313419B2 (en) | Image processing apparatus and image pickup apparatus where image processing is applied using an acquired depth map | |
CN110677621B (en) | Camera calling method and device, storage medium and electronic equipment | |
CN103945118B (en) | Image weakening method, device and electronic equipment | |
US8885091B2 (en) | Imaging device and distance information detecting method | |
CN108737718B (en) | Shooting method and device and intelligent equipment | |
CN106060379B (en) | The control method of picture pick-up device and picture pick-up device | |
CN107358593B (en) | Image forming method and apparatus | |
CN108234858B (en) | Image blurring processing method and device, storage medium and electronic equipment | |
US8830381B2 (en) | Imaging device and method to provide bokeh effect in captured image by determining distance and focus of captured objects in secondary image sequence | |
WO2012132486A1 (en) | Imaging device, imaging method, program, and program storage medium | |
CN106707674A (en) | Automatic focusing method of projection equipment and the projection equipment | |
CN113940057B (en) | System and method for controlling exposure settings based on motion characteristics associated with an image sensor | |
CN105657238B (en) | Track focusing method and device | |
US8810665B2 (en) | Imaging device and method to detect distance information for blocks in secondary images by changing block size | |
CN111246093B (en) | Image processing method, image processing device, storage medium and electronic equipment | |
CN111246114B (en) | Photographing processing method and device, terminal equipment and storage medium | |
CN110231018B (en) | Structured light distance measuring method and device and computer readable storage medium | |
CN111917991B (en) | Image quality control method, device, equipment and storage medium | |
CN104580901A (en) | Shooting parameter determining method and terminal | |
JP5899745B2 (en) | Image processing apparatus, image processing method, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170125 |
|
RJ01 | Rejection of invention patent application after publication |