Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete
Site preparation description, it is clear that described embodiments are some of the embodiments of the present invention, instead of all the embodiments.Based on this hair
Embodiment in bright, every other implementation obtained by those of ordinary skill in the art without making creative efforts
Example, shall fall within the protection scope of the present invention.
First embodiment
It is the flow chart of image pickup method provided in an embodiment of the present invention referring to Fig. 1, Fig. 1, which is used to include two
The mobile terminal of a camera, as shown in Figure 1, comprising the following steps:
Step 101 obtains the first image and the second image that described two cameras shoot same target.
In the embodiment of the present invention, same target is shot using two cameras, due to the position of two cameras
Difference makes shooting angle different, so as to obtain the first image and the second image of different angle.For example, two cameras are excellent
Choosing is in same plane, and such as the placement position of two eyes of people, the image of shooting be will be such as human eye, shoots area
Shooting angle will will appear and intersect in other words in domain, and then shoot to same target.For the space-time for avoiding reference object
Difference, the first image and the second image are preferably that two cameras carry out same target in synchronization to shoot acquisition.
Same barrier present in step 102, identification the first image and second image is in first figure
Display area as in.
It is blocked if two cameras have barrier in shooting process, the first image and second image have
It is likely to occur that there are the situations of same barrier, the display area of the same barrier in the first image can be identified at this time
Come.Specifically, contour curve of the barrier in the first image can be identified, other identification methods can also be used, herein
Without limitation.
Step 103, in the first image, remove the first image data of the display area.
Determine first image of the same barrier after the display area in the first image, i.e., to the display area
Data are purged, and the same obstacle for being present in the first image and the second image is avoided under the premise of keeping the first image original appearance
Object appears in the first image.
Step 104 determines the second image data in second image with the display area co-located region.
Due to the first image and the second image be from different perspectives to same target shot as a result, thus can be compared with
The first image and the second image are placed under the same coordinate system with same position basic point for what is coincide, so that the first image and second
The display area of image is substantially overlapping in position.According to same barrier, the location parameter of display area can in the first image
Determine the display area and corresponding second image data in the second image with same position parameter.Because of the second picture number
According to identical in display area location parameter as the first image data, so the second image data can more reflect in the first image
Reference object is present in the image original appearance that the same barrier of the first image and the second image blocks.
Step 105 carries out image data filling to the first image using second image data.
Specifically, utilizing the second image identical with the display area location parameter of the first image data in the second image
Data carry out image data filling to the first image, the display area removed in the first image may make to refill accordingly
Image data, with guarantee the first image integrality.
The filled the first image of image data is passed through in step 106, output.
When the first image complete image data filling after, can using the first image as the shooting result to reference object into
Row output, specifically can be in a manner of preview or other modes export, herein without limitation.
In the embodiment of the present invention, above-mentioned mobile terminal any can have the mobile terminal of two cameras, such as: hand
Machine, tablet computer (Tablet Personal Computer), laptop computer (Laptop Computer), individual digital help
Manage (Personal Digital Assistant, abbreviation PDA), mobile Internet access device (Mobile Internet Device,
) or wearable device (Wearable Device) etc. MID.
The image pickup method of the embodiment of the present invention, for the mobile terminal including two cameras, comprising: obtain described two
The first image and the second image that camera shoots same target;Exist in identification the first image and second image
Display area of the same barrier in the first image;In the first image, the of the display area is removed
One image data;Determine the second image data in second image with the display area co-located region;Using institute
It states the second image data and image data filling is carried out to the first image;Image data filled described first is passed through in output
Image.User shoots the image come in this way can avoid blocking for barrier in time, without carrying out post-processing to image
Reference object is more completely presented in the picture, simplifies operation sequence.
Second embodiment
Referring to fig. 2, Fig. 2 is the flow chart of image pickup method provided in an embodiment of the present invention, which is used to include two
The mobile terminal of a camera, as shown in Figure 2, comprising the following steps:
Step 201 obtains the first image and the second image that described two cameras shoot same target.
In the embodiment of the present invention, same target is shot using two cameras, due to the position of two cameras
Difference makes shooting angle different, so as to obtain the first image and the second image of different angle.For example, two cameras are excellent
Choosing is in same plane, and such as the placement position of two eyes of people, the image of shooting be will be such as human eye, shoots area
Shooting angle will will appear and intersect in other words in domain, and then shoot to same target.For the space-time for avoiding reference object
Difference, the first image and the second image are preferably that two cameras carry out same target in synchronization to shoot acquisition.
Step 202 judges the first image and second image with the presence or absence of same barrier, if so then execute step
Rapid 203, if it is not, then terminating process.
On the one hand, the difference of shooting angle will cause the object appeared in the first image and not necessarily appear in second
In image, the object for being only present in two cross one another shooting areas of camera just can be simultaneously in the first image and the second figure
Show as in;On the other hand, existing simultaneously the object in the first image and the second image is all not barrier.It is general and
Speech, such as cobweb or protective fence small objects are more difficult in shooting process avoids, occupy small percentage in photographed screen.It hides
Gear reference object and the object for occupying photographed screen preset ratio are the barriers that the embodiment of the present invention defines.More than being thus based on
Two aspect reasons need to judge in the first image and the second image with the presence or absence of same barrier.
Same barrier present in step 203, identification the first image and second image is in first figure
Display area as in.
It, can be aobvious in the first image by the same barrier when the first image and the second image are there are when same barrier
Show that region recognition comes out.Specifically, contour curve of the barrier in the first image can be identified, other knowledges can also be used
Other mode, herein without limitation.
Step 204, in the first image, remove the first image data of the display area.
Determine first image of the same barrier after the display area in the first image, i.e., to the display area
Data are purged, and the same obstacle for being present in the first image and the second image is avoided under the premise of keeping the first image original appearance
Object appears in the first image.
Step 205 determines the second image data in second image with the display area co-located region.
Due to the first image and the second image be from different perspectives to same target shot as a result, thus can be compared with
The first image and the second image are placed under the same coordinate system with same position basic point for what is coincide, so that the first image and second
The display area of image is substantially overlapping in position.According to same barrier, the location parameter of display area can in the first image
Determine the display area and the second image data in the second image with same position parameter.Because of the second image data and the
One image data is identical in display area location parameter, and second image data can more reflect the shooting pair in the first image
The image original appearance that same barrier as being present in the first image and the second image blocks.
Step 206 carries out image data filling to the first image using second image data.
Specifically, utilizing the second image identical with the display area location parameter of the first image data in the second image
Data carry out image data filling to the first image, the display area removed in the first image may make to refill accordingly
Image data, with guarantee the first image integrality.
The filled the first image of image data is passed through in step 207, output.
When the first image complete image data filling after, can using the first image as the shooting result to reference object into
Row output, specifically can be in a manner of preview or other modes export, herein without limitation.
Optionally, the step of judgement the first image and second image whether there is same barrier, packet
It includes: the contour curve of each object in identification the first image and second image;According to the first image and described
The contour curve of each object in second image judges the first image and second image with the presence or absence of the default ruler of satisfaction
Very little same object;If there is the same object for meeting pre-set dimension in the first image and second image, according to institute
Display position of the same object for meeting pre-set dimension in the first image and second image is stated, judges the satisfaction
Whether the same object of pre-set dimension is barrier.
In the embodiment of the present invention, in identifying the first image and the second image after the contour curve of each object, lead to
It crosses the judgement of level-one grade and gradually determines the same barrier for being present in the first image and the second image.
Specifically, carrying out contour curve identification to each object present in the first image and the second image respectively, specifically
Gray scale difference value can be greater than pre- by mode by calculating the gray scale difference value in image between each pixel pixel adjacent thereto
The pixel for determining difference threshold determines edge pixel point, and the curve of these edge pixel points composition is to form contour of object, when
So specific contour curve recognition methods is without being limited thereto.When the profile for identifying each object in the first image and the second image is bent
Line judges the first image and described then according to the contour curve of each object in the first image and second image
Two images whether there is the same object for meeting pre-set dimension;If so, being existed according to the same object for meeting pre-set dimension
Display position in the first image and second image judges whether the same object for meeting pre-set dimension is barrier
Hinder object;If it is not, then terminating process.
Optionally, the contour curve of each object according to the first image and in second image, judges institute
State the step of the first image and second image are with the presence or absence of the same object for meeting pre-set dimension, comprising: by described first
The contour curve of each object is compared with the contour curve of each object in second image in image;If described first
There are the objects of same profile curve in image and second image, then are judged as the first image and second image
There are same objects;By the size of the contour curve of same object existing for the first image and second image and preset
Size threshold is compared;If the size of the contour curve of same object existing for the first image and second image is small
In being equal to the pre-set dimension threshold value, then it is pre- to be judged as that same object existing for the first image and second image meets
If size.
Specifically, after identifying the contour curve of each object in the first image and the second image, then by the first figure
The contour curve of each object is compared one by one in picture and the second image, is determined together according to the matching degree of the two contour curve
The same object being present in the first image and the second image, for example, being shot from different perspectives to same flower, although clapping
It takes the photograph angle and there is influence to flower-shape to a certain extent, the image that different filming apparatus obtain can have any different, but be all based on flower
The contour curve matching degree and similarity of flower are quite high present in the deformation of profile itself, the first image and the second image
's.Specifically, can detect that the matching degree of the contour curve of the first objects in images and the contour curve of the second objects in images,
If the matching degree reaches preset threshold, it is judged as that there are same objects for the first image and second image.
In addition, it is more difficult to the barrier for blocking reference object avoided is generally such as cobweb or protective fence small objects,
It can be seen that these barriers dimensionally have general character, thus existed simultaneously at this time according to the judgement of contour of object curve in the first image
Whether meet pre-set dimension with the same object in the second image, then can exclude a part and be unlikely to belong to barrier but to belong to together
In the presence of the same object with the first image and the second image.Specifically, will be present in same in the first image and the second image
The size of the contour curve of one object is compared with pre-set dimension threshold value;If the size of the contour curve of the same object is less than
Equal to the pre-set dimension threshold value, then judge that the same object meets pre-set dimension.Wherein, the setting one of pre-set dimension threshold value
As for length or width to be less than photographed screen long or wide 5%, for example, the main part of reference object is to exist together certainly
With the same object of the first image and the second image, but the main part of reference object necessarily occupies most of region of image,
It can not be able to satisfy the pre-set dimension of barrier, then need to be excluded determining except the range of barrier.
Optionally, it is described according to the same object for meeting pre-set dimension in the first image and second image
In display position, the step of whether same object for meeting pre-set dimension is barrier judged, comprising: obtain it is described full
First display position of the same object of sufficient pre-set dimension in the first image;It obtains and described meets the same of pre-set dimension
Second display position of the object in second image;It calculates between first display position and second display position
Relative distance;According to the relative distance, judge whether the same object for meeting pre-set dimension is barrier.
Since the position of barrier must be present in camera and reference object between camera and reference object
Between object due to the influence of two camera different shooting angles, the display position on the first image and the second image must
Can so there be a biggish alternate position spike, and it is smaller in the presence of then being influenced by shooting angle with the object on reference object, it will not go out
Existing visibility point is poor.As shown in figure 3, main camera A and pair camera B are generally aligned in the same plane, shooting angle is same in cross-pair
One background F is shot.Barrier C where main camera A and secondary camera B between plane and background F, is shown in master
Shadow region D in first image of camera A shooting, the shadow region E being shown in the second image of secondary camera B shooting.It is aobvious
So, there can be biggish relative distance between shadow region D and shadow region E.Thus, according to the same object point for meeting pre-set dimension
It not can determine in the first display position in the first image and the relative distance between the second display position in the second image
Whether the same object is barrier.
Optionally, described according to the relative distance, judge whether the same object for meeting pre-set dimension is obstacle
The step of object, comprising: be compared the relative distance with pre-determined distance threshold value;If the relative distance is more than or equal to described
Pre-determined distance threshold value is then judged as that the same object for meeting pre-set dimension is barrier.
As noted previously, as the difference of two camera shooting angle, meets the same object of pre-set dimension in the first figure
It the first display position as in and is certainly existed between the second display position in the second image and meets pre-determined distance condition
Relative distance.Specifically, design parameter that pre-determined distance threshold value can shoot same target according to two cameras (such as shoot
The shooting angle etc. of focal length and two cameras) it is determined.When meeting the same object of pre-set dimension in the first image
In the first display position and between the second display position in the second image relative distance be more than or equal to pre-determined distance threshold value
When, then judge that this meets the same object of pre-set dimension for barrier, although thus will meet pre-set dimension be not at two
Small objects between a camera and reference object exclude the ranks of barrier, avoid to have with reference object sheet
Small objects are handled as barrier, improve the accuracy of barrier judgment.Optionally, described to use second image
The step of data carry out image data filling to the first image, comprising: obtain the corresponding pixel of second image data
Block;The block of pixels is filled into the display area of the first image.Specifically, since the second image data is
It is identical as the display area location parameter of the first image data in two images, it can largely reappear in the first image by obstacle
The shooting original appearance that object blocks, thus the corresponding block of pixels of the second image data is filled into the first image the first image data
In display area, can better quality the accuracy that reference object is shown of the first image of guarantee.
The image pickup method of the embodiment of the present invention, for the mobile terminal including two cameras, comprising: obtain described two
The first image and the second image that camera shoots same target;Judge whether the first image and second image deposit
In same barrier;If so, same barrier present in identification the first image and second image is described the
Display area in one image;In the first image, the first image data of the display area is removed;Determine described
In two images with the second image data of the display area co-located region;Using second image data to described
One image carries out image data filling;The filled the first image of image data is passed through in output.User, which shoots, in this way comes
Image can avoid blocking for barrier in time, without to image carry out post-processing, can in the picture by reference object compared with
Completely to present, operation sequence is simplified.
3rd embodiment
Referring to fig. 4, Fig. 4 is the structure chart that the present invention implements the mobile terminal provided, is able to achieve first embodiment to second
The details of the operating method of mobile terminal in embodiment, and reach identical effect.Mobile terminal includes two cameras, such as
Shown in Fig. 4, mobile terminal 400 further includes obtaining module 401, identification module 402, removing module 403, determining module 404, filling
Module 405 and output module 406, wherein obtain module 401 connect with identification module 402, and identification module 402 also respectively with
It removes module 403 and determining module 404 connects, filling module 405 connects with removing module 403 and determining module 404 respectively
It connects, filling module 405 is also connect with output module 406, in which:
The acquisition module 401, the first image and second that same target is shot for obtaining described two cameras
Image.
The identification module 402, it is described for identification to obtain the first image and second figure that module 401 obtains
Display area of the same barrier in the first image as present in.
The removing module 403, in the first image, removing the described aobvious of the identification of identification module 402
Show first image data in region.
The determining module 404, for determining the display in second image with the identification module 402 identification
Second image data of region co-located region.
The filling module 405, second image data for being determined using the determining module 404 is to described clear
Except the first image that module 403 is removed carries out image data filling.
The output module 406 passes through 405 image data of filling module filled described first for exporting
Figure.
Optionally, as shown in figure 5, the mobile terminal 400 further include:
Judgment module 407, for judging the first image and second image with the presence or absence of same barrier.
The identification module 402, if judging that the first image and second image are deposited for the judgment module 407
In same barrier, then identify same barrier present in the first image and second image in the first image
In display area.
Optionally, as shown in fig. 6, the judgment module 407 includes:
Identify submodule 4071, the profile of each object is bent in the first image and second image for identification
Line.
First judging submodule 4072, the first image and described for being identified according to the identification submodule 4071
The contour curve of each object in second image judges the first image and second image with the presence or absence of the default ruler of satisfaction
Very little same object.
Second judgment submodule 4073, if judging the first image and described for first judging submodule 4072
There is the same object for meeting pre-set dimension in the second image, then according to the same object for meeting pre-set dimension described first
Display position in image and second image judges whether the same object for meeting pre-set dimension is barrier.
Optionally, as described in Figure 7, first judging submodule 4072 includes:
Comparing unit 40721, for will be in the contour curve of object each in the first image and second image
The contour curve of each object is compared.
First judging unit 40722, if judging the first image and second figure for the comparing unit 40721
There are the objects of same profile curve as in, then are judged as that there are same objects for the first image and second image.
Comparing unit 40723, for first judging unit 40722 to be judged the first image and second figure
As the size of the contour curve of existing same object is compared with pre-set dimension threshold value.
Second judgment unit 40724, if judging the first image and second figure for the comparing unit 40723
As existing same object contour curve size be less than or equal to the pre-set dimension threshold value, then be judged as the first image
Meet pre-set dimension with same object existing for second image.
Optionally, as described in Figure 8, the second judgment submodule 4073 includes:
First acquisition unit 40731, for obtaining the same object for meeting pre-set dimension in the first image
The first display position.
Second acquisition unit 40732, for obtaining the same object for meeting pre-set dimension in second image
The second display position;
Computing unit 40733, for calculating the first display position that the first acquisition unit 40731 obtains and described the
The relative distance between the second display position that two acquiring units 40732 obtain.
Third judging unit 40734, the relative distance for being obtained according to the computing unit 40733, judges institute
It states and meets whether the same object of pre-set dimension is barrier.
Optionally, as shown in figure 9, the third judging unit 40734 includes:
Comparing subunit 407341, for the relative distance to be compared with pre-determined distance threshold value.
It is described to judge that the relative distance is more than or equal to for the comparing subunit 407341 for judgment sub-unit 407342
Pre-determined distance threshold value is then judged as that the same object for meeting pre-set dimension is barrier.
Optionally, as shown in Figure 10, the filling module 405 includes:
Acquisition submodule 4051, for obtaining the corresponding block of pixels of second image data.
Submodule 4052 is filled, the block of pixels for obtaining the acquisition submodule 4051 is filled into described first
In the display area of image.
Mobile terminal 400 can be realized each process that mobile terminal is realized in the embodiment of the method for Fig. 1 to Fig. 3, to keep away
Exempt to repeat, which is not described herein again.
The mobile terminal 400 of the embodiment of the present invention, including two cameras, the mobile terminal further include: module is obtained,
The first image and the second image that same target is shot for obtaining described two cameras;Identification module, for identification institute
It states and obtains same barrier present in the first image and second image that module obtains in the first image
Display area;Module is removed, is used in the first image, the display area of the identification module identification is removed
First image data;Determining module, for determining the display area in second image with identification module identification
Second image data of co-located region;Module is filled, second picture number for determining using the determining module
Image data filling is carried out according to the first image removed to the removing module;Output module, for exporting described in process
Fill the filled the first image of module image data.User shoots the image come in this way can avoid barrier in time
It blocks, without carrying out post-processing to image, reference object more completely can be presented in the picture, simplify operation journey
Sequence.
Fourth embodiment
It is the structure chart that the present invention implements the mobile terminal provided referring to Figure 11, Figure 11, is able to achieve first embodiment and the
The details of image pickup method in two embodiments, and reach identical effect.As shown in figure 11, mobile terminal 1100 includes: at least
One processor 1101, memory 1102, at least one network interface 1104 and user interface 1103.In mobile terminal 1100
Various components are coupled by bus system 1105.It is understood that bus system 1105 is for realizing between these components
Connection communication.Bus system 1105 further includes that power bus, control bus and status signal are total in addition to including data/address bus
Line.But for the sake of clear explanation, various buses are all designated as bus system 1105 in Figure 11, mobile terminal 1100 also wraps
Two picture shooting assemblies 1106 are included, in addition, two picture shooting assemblies 1106 pass through the various components of bus system 1105 and mobile terminal
Connection.
Wherein, user interface 1103 may include display, keyboard or pointing device (for example, mouse, trace ball
(Track Ball), touch-sensitive plate or touch screen etc..
It is appreciated that the memory 1102 in the embodiment of the present invention can be volatile memory or non-volatile memories
Device, or may include both volatile and non-volatile memories.Wherein, nonvolatile memory can be read-only memory
(Read-Only Memory, ROM), programmable read only memory (Programmable ROM, PROM), erasable programmable are only
Read memory (Erasable PROM, EPROM), electrically erasable programmable read-only memory (Electrically EPROM,
) or flash memory EEPROM.Volatile memory can be random access memory (Random Access Memory, RAM), use
Make External Cache.By exemplary but be not restricted explanation, the RAM of many forms is available, such as static random-access
Memory (Static RAM, SRAM), dynamic random access memory (Dynamic RAM, DRAM), synchronous dynamic random-access
Memory (Synchronous DRAM, SDRAM), double data speed synchronous dynamic RAM (Double Data
Rate SDRAM, DDRSDRAM), it is enhanced Synchronous Dynamic Random Access Memory (Enhanced SDRAM, ESDRAM), synchronous
Connect dynamic random access memory (Synchlink DRAM, SLDRAM) and direct rambus random access memory
(Direct Rambus RAM, DRRAM).The memory 1102 of system and method described herein is intended to include but is not limited to this
A little and any other suitable type memory.
In some embodiments, memory 1102 stores following element, executable modules or data structures, or
Their subset of person or their superset: operating system 11021 and application program 11022.
Wherein, operating system 11021 include various system programs, such as ccf layer, core library layer, driving layer etc., are used for
Realize various basic businesses and the hardware based task of processing.Application program 11022 includes various application programs, such as matchmaker
Body player (Media Player), browser (Browser) etc., for realizing various applied business.Realize that the present invention is implemented
The program of example method may be embodied in application program 11022.
In embodiments of the present invention, by the program or instruction of calling memory 1102 to store, specifically, can be application
The program or instruction stored in program 11022, processor 1101 are used for: obtaining what described two cameras shot same target
First image and the second image;Identify same barrier present in the first image and second image described first
Display area in image;In the first image, the first image data of the display area is removed;Determine described second
In image with the second image data of the display area co-located region;Using second image data to described first
Image carries out image data filling;The filled the first image of image data is passed through in output.
The method that the embodiments of the present invention disclose can be applied in processor 1101, or real by processor 1101
It is existing.Processor 1101 may be a kind of IC chip, the processing capacity with signal.During realization, the above method
Each step can be completed by the instruction of the integrated logic circuit of the hardware in processor 1101 or software form.Above-mentioned
Processor 1101 can be general processor, digital signal processor (Digital Signal Processor, DSP), dedicated
Integrated circuit (Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field
Programmable Gate Array, FPGA) either other programmable logic device, discrete gate or transistor logic,
Discrete hardware components.It may be implemented or execute disclosed each method, step and the logic diagram in the embodiment of the present invention.It is general
Processor can be microprocessor or the processor is also possible to any conventional processor etc..In conjunction with institute of the embodiment of the present invention
The step of disclosed method, can be embodied directly in hardware decoding processor and execute completion, or with the hardware in decoding processor
And software module combination executes completion.Software module can be located at random access memory, and flash memory, read-only memory may be programmed read-only
In the storage medium of this fields such as memory or electrically erasable programmable memory, register maturation.The storage medium is located at
The step of memory 1102, processor 1101 reads the information in memory 1102, completes the above method in conjunction with its hardware.
It is understood that embodiments described herein can with hardware, software, firmware, middleware, microcode or its
Combination is to realize.For hardware realization, processing unit be may be implemented in one or more specific integrated circuit (Application
Specific Integrated Circuits, ASIC), digital signal processor (Digital Signal Processing,
DSP), digital signal processing appts (DSP Device, DSPD), programmable logic device (Programmable Logic
Device, PLD), field programmable gate array (Field-Programmable Gate Array, FPGA), general processor,
In controller, microcontroller, microprocessor, other electronic units for executing herein described function or combinations thereof.
For software implementations, it can be realized herein by executing the module (such as process, function etc.) of function described herein
The technology.Software code is storable in memory and is executed by processor.Memory can in the processor or
It is realized outside processor.
Optionally, processor 1101 is also used to: judging the first image and second image with the presence or absence of same barrier
Hinder object;If the first image and second image there are same barrier, execute the identification the first image and
The step of display area of the same barrier present in second image in the first image.
Optionally, processor 1101 is also used to: the wheel of each object in identification the first image and second image
Wide curve;According to the contour curve of each object in the first image and second image, judge the first image and
Second image whether there is the same object for meeting pre-set dimension;If the first image and second image exist full
The same object of sufficient pre-set dimension, then according to the same object for meeting pre-set dimension in the first image and described second
Display position in image judges whether the same object for meeting pre-set dimension is barrier.
Optionally, the processor 1101 is also used to: by the contour curve of object each in the first image with it is described
The contour curve of each object is compared in second image;If in the first image and second image, there are identical wheels
The object of wide curve is then judged as that there are same objects for the first image and second image;By the first image and
The size of the contour curve of same object existing for second image is compared with pre-set dimension threshold value;If first figure
The size of the contour curve of same object existing for picture and second image is less than or equal to the pre-set dimension threshold value, then judges
Meet pre-set dimension for same object existing for the first image and second image.
Optionally, the processor 1101 is also used to: obtaining the same object for meeting pre-set dimension described first
The first display position in image;Obtain second display of the same object for meeting pre-set dimension in second image
Position;Calculate the relative distance between first display position and second display position;According to the relative distance, sentence
Whether the same object for meeting pre-set dimension that breaks is barrier.
Optionally, the processor 1101 is also used to: the relative distance is compared with pre-determined distance threshold value;If institute
Relative distance is stated more than or equal to the pre-determined distance threshold value, then is judged as that the same object for meeting pre-set dimension is obstacle
Object.
Optionally, the processor 1101 is also used to: obtaining the corresponding block of pixels of second image data;By the picture
Plain block is filled into the display area of the first image.
Mobile terminal 1100 can be realized each process that mobile terminal is realized in previous embodiment, to avoid repeating, this
In repeat no more.
The mobile terminal 1100 of the embodiment of the present invention obtains the first image that described two cameras shoot same target
With the second image;Identify same barrier present in the first image and second image in the first image
Display area;In the first image, the first image data of the display area is removed;Determine in second image with
Second image data of the display area co-located region;The first image is carried out using second image data
Image data filling;The filled the first image of image data is passed through in output.In this way user shoot come image can and
When avoid blocking for barrier, without to image carry out post-processing, can more completely be in by reference object in the picture
It is existing, simplify operation sequence.
5th embodiment
Please refer to Figure 12, Figure 12 is the structure chart that the present invention implements the mobile terminal provided, be able to achieve first embodiment and
The details of image pickup method in second embodiment, and reach identical effect.As shown in figure 12, mobile terminal 1200 includes radio frequency
(Radio Frequency, RF) circuit 1210, memory 1220, input unit 1230, display unit 1240, processor 1250,
Voicefrequency circuit 1260, communication module 1270, power supply 1280 and two picture shooting assemblies 1290.
Wherein, input unit 1230 can be used for receiving the number or character information of user's input, and generate with movement eventually
The related signal input of the user setting and function control at end 1200.Specifically, in the embodiment of the present invention, the input unit
1230 may include touch panel 1231.Touch panel 1231, also referred to as touch screen collect the touching of user on it or nearby
It touches operation (for example user uses the operations of any suitable object or attachment on touch panel 1231 such as finger, stylus), and
Corresponding attachment device is driven according to preset formula.Optionally, touch panel 1231 may include touch detecting apparatus and
Two parts of touch controller.Wherein, the touch orientation of touch detecting apparatus detection user, and detect touch operation bring letter
Number, transmit a signal to touch controller;Touch controller receives touch information from touch detecting apparatus, and is converted into
Contact coordinate, then give the processor 1250, and order that processor 1250 is sent can be received and executed.Furthermore, it is possible to
Touch panel 1231 is realized using multiple types such as resistance-type, condenser type, infrared ray and surface acoustic waves.In addition to touch panel
1231, input unit 1230 can also include other input equipments 1232, other input equipments 1232 can include but is not limited to
One of physical keyboard, function key (such as volume control button, switch key etc.), trace ball, mouse, operating stick etc. are more
Kind.
Wherein, display unit 1240 can be used for showing information input by user or be supplied to the information and movement of user
The various menu interfaces of terminal 1200.Display unit 1240 may include display panel 1241, optionally, using LCD or can have
The forms such as machine light emitting diode (Organic Light-Emitting Diode, OLED) configure display panel 1241.
It should be noted that touch panel 1231 can cover display panel 1241, touch display screen is formed, when the touch display screen
After detecting touch operation on it or nearby, processor 1250 is sent to determine the type of touch event, is followed by subsequent processing
Device 1250 provides corresponding visual output according to the type of touch event in touch display screen.
Touch display screen includes Application Program Interface viewing area and common control viewing area.The Application Program Interface viewing area
And arrangement mode of the common control viewing area does not limit, can be arranged above and below, left-right situs etc. can distinguish two it is aobvious
Show the arrangement mode in area.The Application Program Interface viewing area is displayed for the interface of application program.Each interface can be with
The interface elements such as the icon comprising at least one application program and/or widget desktop control.The Application Program Interface viewing area
Or the empty interface not comprising any content.This commonly uses control viewing area for showing the higher control of utilization rate, for example,
Application icons such as button, interface number, scroll bar, phone directory icon etc. are set.
Wherein processor 1250 is the control centre of mobile terminal 1200, utilizes various interfaces and connection whole mobile phone
Various pieces, by running or execute the software program and/or module that are stored in first memory 1221, and call and deposit
The data in second memory 1222 are stored up, the various functions and processing data of mobile terminal 1200 are executed, thus eventually to movement
End 1200 carries out integral monitoring.Optionally, processor 1250 may include one or more processing units.
In embodiments of the present invention, by call store the first memory 1221 in software program and/or module and/
Or the data in the second memory 1222, processor 1250 are used for: obtaining what described two cameras shot same target
First image and the second image;Identify same barrier present in the first image and second image described first
Display area in image;In the first image, the first image data of the display area is removed;Determine described second
In image with the second image data of the display area co-located region;Using second image data to described first
Image carries out image data filling;The filled the first image of image data is passed through in output.
Optionally, processor 1250 is also used to: judging the first image and second image with the presence or absence of same barrier
Hinder object;If the first image and second image there are same barrier, execute the identification the first image and
The step of display area of the same barrier present in second image in the first image.
Optionally, processor 1250 is also used to: the wheel of each object in identification the first image and second image
Wide curve;According to the contour curve of each object in the first image and second image, judge the first image and
Second image whether there is the same object for meeting pre-set dimension;If the first image and second image exist full
The same object of sufficient pre-set dimension, then according to the same object for meeting pre-set dimension in the first image and described second
Display position in image judges whether the same object for meeting pre-set dimension is barrier.
Optionally, processor 1250 is also used to: by the contour curve of object each in the first image and described second
The contour curve of each object is compared in image;If there are same profile songs in the first image and second image
The object of line is then judged as that there are same objects for the first image and second image;By the first image and described
The size of the contour curve of same object existing for second image is compared with pre-set dimension threshold value;If the first image and
The size of the contour curve of same object existing for second image is less than or equal to the pre-set dimension threshold value, then is judged as institute
It states same object existing for the first image and second image and meets pre-set dimension.
Optionally, processor 1250 is also used to: obtaining the same object for meeting pre-set dimension in the first image
In the first display position;Obtain second display position of the same object for meeting pre-set dimension in second image
It sets;Calculate the relative distance between first display position and second display position;According to the relative distance, judgement
Whether the same object for meeting pre-set dimension is barrier.
Optionally, processor 1250 is also used to: the relative distance is compared with pre-determined distance threshold value;If the phase
It adjusts the distance and is more than or equal to the pre-determined distance threshold value, be then judged as that the same object for meeting pre-set dimension is barrier.
Optionally, processor 1250 is also used to: obtaining the corresponding block of pixels of second image data;By the block of pixels
It is filled into the display area of the first image.
Mobile terminal 1200 can be realized each process that mobile terminal is realized in previous embodiment, and can reach identical
Technical effect, to avoid repeating, which is not described herein again.
In this way, obtaining what described two cameras shot same target in the mobile terminal 1200 of the embodiment of the present invention
First image and the second image;Identify same barrier present in the first image and second image described first
Display area in image;In the first image, the first image data of the display area is removed;Determine described second
In image with the second image data of the display area co-located region;Using second image data to described first
Image carries out image data filling;The filled the first image of image data is passed through in output.User shoots in this way
Image can avoid blocking for barrier in time, without to image carry out post-processing, can in the picture by reference object more
It is complete to present, simplify operation sequence.
Those of ordinary skill in the art may be aware that list described in conjunction with the examples disclosed in the embodiments of the present disclosure
Member and algorithm steps can be realized with the combination of electronic hardware or computer software and electronic hardware.These functions are actually
It is implemented in hardware or software, the specific application and design constraint depending on technical solution.Professional technician
Each specific application can be used different methods to achieve the described function, but this realization is it is not considered that exceed
The scope of the present invention.
It is apparent to those skilled in the art that for convenience and simplicity of description, the system of foregoing description,
The specific work process of device and unit, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In embodiment provided herein, it should be understood that disclosed device and method can pass through others
Mode is realized.For example, the apparatus embodiments described above are merely exemplary, for example, the division of the unit, only
A kind of logical function partition, there may be another division manner in actual implementation, for example, multiple units or components can combine or
Person is desirably integrated into another system, or some features can be ignored or not executed.Another point, shown or discussed is mutual
Between coupling, direct-coupling or communication connection can be through some interfaces, the INDIRECT COUPLING or communication link of device or unit
It connects, can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit
The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple
In network unit.Some or all of unit therein can be selected to realize the embodiment of the present invention according to the actual needs
Purpose.
It, can also be in addition, the functional units in various embodiments of the present invention may be integrated into one processing unit
It is that each unit physically exists alone, can also be integrated in one unit with two or more units.
It, can be with if the function is realized in the form of SFU software functional unit and when sold or used as an independent product
It is stored in a computer readable storage medium.Based on this understanding, technical solution of the present invention is substantially in other words
The part of the part that contributes to existing technology or the technical solution can be embodied in the form of software products, the meter
Calculation machine software product is stored in a storage medium, including some instructions are used so that a computer equipment (can be a
People's computer, server or network equipment etc.) it performs all or part of the steps of the method described in the various embodiments of the present invention.
And storage medium above-mentioned includes: that USB flash disk, mobile hard disk, ROM, RAM, magnetic or disk etc. are various can store program code
Medium.
The above description is merely a specific embodiment, but scope of protection of the present invention is not limited thereto, any
Those familiar with the art in the technical scope disclosed by the present invention, can easily think of the change or the replacement, and should all contain
Lid is within protection scope of the present invention.Therefore, protection scope of the present invention should be subject to the protection scope in claims.