CN110248111A - Control method, apparatus, electronic equipment and the computer readable storage medium of shooting - Google Patents
Control method, apparatus, electronic equipment and the computer readable storage medium of shooting Download PDFInfo
- Publication number
- CN110248111A CN110248111A CN201910585686.8A CN201910585686A CN110248111A CN 110248111 A CN110248111 A CN 110248111A CN 201910585686 A CN201910585686 A CN 201910585686A CN 110248111 A CN110248111 A CN 110248111A
- Authority
- CN
- China
- Prior art keywords
- image
- camera
- processing units
- processing unit
- exposure time
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/73—Circuitry for compensating brightness variation in the scene by influencing the exposure time
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
Abstract
The invention relates to method, apparatus, electronic equipment and the computer readable storage mediums of a kind of control shooting.The above method, comprising: when first processing units receive the image capture instruction of the second processing unit transmission, according to image capture instruction control the first camera the first image of acquisition;When first processing units receive the synchronization signal of second camera transmission, the first exposure time of the first camera and the second exposure time of second camera are obtained;According to the first exposure time and the second exposure time computation delay duration;When the duration that first processing units receive synchronization signal reaches delay duration, synchronization signal is forwarded to the first camera, synchronization signal is used to indicate the first camera and starts to expose and acquire the first image;The first image is handled by first processing units.Method, apparatus, electronic equipment and the computer readable storage medium of above-mentioned control shooting, synchronous effect are good, it is ensured that the picture material of two cameras acquisition is consistent.
Description
Technical field
This application involves field of computer technology, more particularly to a kind of method, apparatus of control shooting, electronic equipment and
Computer readable storage medium.
Background technique
With the fast development of image technology on intelligent terminal, on more and more intelligent terminals installation there are two and two
Above camera is cooperated using multiple cameras, to collect visual effect better image.In order to guarantee two camera shootings
The picture that finally acquires of head is consistent, needs to control and synchronizes between two cameras.And in traditional mode, two camera shootings
It would generally be connected by hardware signal line between head, and be synchronized by the signal that every frame starts exposure.When two cameras
Between exposure time difference it is larger when, it is big to will lead to the picture material difference acquired between two cameras for poor synchronization.
Summary of the invention
The embodiment of the present application provides method, apparatus, electronic equipment and the computer readable storage medium of a kind of control shooting,
Synchronous effect is good, it is ensured that the picture material of two cameras acquisition is consistent.
A method of control shooting, comprising:
When first processing units receive the image capture instruction of the second processing unit transmission, acquired according to described image
Instruction the first camera of control acquires the first image, and described image acquisition instructions receive data for described the second processing unit and obtain
Transmission when taking request, the data acquisition request are used to indicate the second processing unit control second camera and acquire the second figure
Picture;
When first processing units receive the synchronization signal that the second camera is sent, first camera is obtained
The first exposure time and second camera the second exposure time, the synchronization signal is that the second camera acquires every frame
In the signal for starting to send at the time of exposure when the second image;
According to first exposure time and the second exposure time computation delay duration;
When the duration that the first processing units receive the synchronization signal reaches the delay duration, Xiang Suoshu
One camera forwards the synchronization signal, and the synchronization signal is used to indicate first camera and starts to expose and acquire first
Image;
The first image is handled by the first processing units, and the first image is sent to by treated
Described the second processing unit.
A kind of device of control shooting, comprising:
Image capture module, for receiving the image capture instruction of the second processing unit transmission when first processing units
When, the first camera is controlled according to described image acquisition instructions and acquires the first image, and described image acquisition instructions are described second
Transmission when processing unit receives data acquisition request, the data acquisition request are used to indicate the second processing unit control
Second camera acquires the second image;
Signal receiving module, when for receiving the synchronization signal of the second camera transmission when first processing units,
The first exposure time of first camera and the second exposure time of second camera are obtained, the synchronization signal is described
In the signal for starting to send at the time of exposure when second camera acquires every the second image of frame;
Computing module, for according to first exposure time and the second exposure time computation delay duration;
Signal forwarding module, the duration for receiving the synchronization signal when the first processing units reach described and prolong
When Shi Shichang, the first camera of Xiang Suoshu forwards the synchronization signal, and the synchronization signal is used to indicate first camera
Start to expose and acquire the first image;
Processing module, for being handled by the first processing units the first image, and by treated
First image is sent to described the second processing unit.
A kind of electronic equipment, including first processing units, the second processing unit and camera module, first processing are single
Member is connected with described the second processing unit and camera module respectively;The camera module includes that the first camera and second is taken the photograph
As head, the first processing units connect first camera by control line, and described the second processing unit passes through control line
The second camera is connected, the first processing units are connect with described the second processing unit, and the first processing units are also
It is connect respectively with first camera and second camera by signal wire;
Described the second processing unit, for being controlled according to the data acquisition request when receiving data acquisition request
The second camera acquires the second image, and sends image capture instruction to the first processing units;
The first processing units, for when receiving the image capture instruction that described the second processing unit is sent, root
The first camera, which is controlled, according to described image acquisition instructions acquires the first image;
The second camera is sent out at the time of starting exposure to first processing units when for acquiring every the second image of frame
Send synchronization signal;
The first processing units are also used to receive the synchronous letter that the second camera is sent when first processing units
Number when, obtain the first exposure time of first camera and the second exposure time of second camera;
The first processing units, when being also used to according to first exposure time and the second exposure time computation delay
It is long;
The first processing units, the duration for being also used to receive the synchronization signal when the first processing units reach
When the delay duration, the first camera of Xiang Suoshu forwards the synchronization signal;
First camera exposes and acquires the first image for starting according to synchronization signal;
The first processing units are also used to handle the first image by the first processing units, and
By treated, the first image is sent to described the second processing unit.
A kind of computer readable storage medium, is stored thereon with computer program, and the computer program is held by processor
Method as described above is realized when row.
Method, apparatus, electronic equipment and the computer readable storage medium of above-mentioned control shooting, when first processing units connect
When receiving the synchronization signal of second camera transmission, according to the exposure time computation delay duration of two cameras, at first
When the duration that reason unit receives synchronization signal reaches delay duration, synchronization signal is forwarded to the first camera, is taken the photograph according to first
As the time point of the exposure time of head and second camera dynamic adjustment forwarding synchronization signal, to be dynamically adapted the first camera shooting
On the head opportunity synchronous with second camera, synchronous effect is good, when the exposure time difference between two cameras is larger, still
It can guarantee that the picture material of two cameras acquisition is consistent.
Detailed description of the invention
Fig. 1 is the application scenario diagram that the method for shooting is controlled in one embodiment;
Fig. 2 is the application scenario diagram that the method for shooting is controlled in another embodiment;
Fig. 3 is the block diagram of electronic equipment in one embodiment;
Fig. 4 is the flow diagram that the method for shooting is controlled in one embodiment;
Fig. 5 is the flow diagram handled in one embodiment the first image;
Fig. 6 is to obtain the flow diagram for referring to speckle image in one embodiment according to the temperature of laser;
Fig. 7 is the flow diagram for selecting data transmission channel in one embodiment according to the security level of application program;
Fig. 8 is the block diagram that the device of shooting is controlled in one embodiment;
Fig. 9 is the block diagram of processing module in one embodiment.
Specific embodiment
It is with reference to the accompanying drawings and embodiments, right in order to which the objects, technical solutions and advantages of the application are more clearly understood
The application is further elaborated.It should be appreciated that specific embodiment described herein is only to explain the application, not
For limiting the application.
It is appreciated that term " first " used in this application, " second " etc. can be used to describe various elements herein,
But these elements should not be limited by these terms.These terms are only used to distinguish the first element from the other element.Citing comes
It says, in the case where not departing from scope of the present application, the first client can be known as the second client, and similarly, can incite somebody to action
Second client is known as the first client.The first client and the second client both client, but it is not same visitor
Family end.
Fig. 1 is the application scenario diagram that the method for shooting is controlled in one embodiment.As shown in Figure 1, the application scenarios can wrap
Include the first camera 110, second camera 120, first processing units 130 and the second processing unit 140.First camera 110
It can be Laser video camera head, second camera 120 can image for RGB (Red/Green/Blue, red green blue color mode)
Head.First processing units 130 can be MCU (Microcontroller Unit, micro-control unit) module etc., the second processing unit
140 can be CPU (Central Processing Unit, central processing unit) module etc..First processing units 130 pass through control
Line connects the first camera 110, and the second processing unit 140 connects second camera 120 by control line.First processing units
130 connect 140 with the second processing unit.First processing units 130 also by signal wire respectively with the first camera 110 and second
Camera 120 connects.
When the second processing unit 140 receives data acquisition request, control line traffic control can be passed through according to data acquisition request
Second camera 120 processed acquires the second image, and sends image capture instruction to first processing units 130.When the first processing is single
When member 130 receives the image capture instruction of the transmission of the second processing unit 140, control line traffic control can be passed through according to image capture instruction
It makes the first camera and acquires the first image.It, can be logical at the time of starting exposure when second camera 120 acquires every the second image of frame
It crosses signal wire and sends synchronization signal to first processing units 130.It is sent out when first processing units 130 receive second camera 120
When the synchronization signal sent, can obtain the first camera 110 the first exposure time and second camera 120 second exposure when
It is long, and according to the first exposure time and the second exposure time computation delay duration.When first processing units 130 receive synchronous letter
Number duration when reaching delay duration, synchronization signal can be forwarded to the first camera 110 by signal wire.First camera 110
After receiving synchronization signal, it can start to expose and acquire the first image, it can be by the first image transmitting of acquisition to the first processing
Unit 130.First processing units 130 can be handled the first image, and the first image is sent at second by treated
Manage unit 140.
Fig. 2 is the application scenario diagram that the method for shooting is controlled in another embodiment.As shown in Fig. 2, electronic equipment 200 can
Including camera module 210, the second processing unit 220, first processing units 230.Above-mentioned the second processing unit 220 can be CPU
Module.Above-mentioned first processing units 230 can be MCU module etc..Wherein, first processing units 230 are connected to the second processing unit
Between 220 and camera module 210, above-mentioned first processing units 230 can control Laser video camera head 212 in camera module 210,
Floodlight 214 and color-changing lamp 218, above-mentioned the second processing unit 220 can control RGB camera 216 in camera module 210.
It include Laser video camera head 212, floodlight 214, RGB camera 216 and color-changing lamp 218 in camera module 210.On
Stating Laser video camera head 212 can be infrared camera, for obtaining infrared image.Above-mentioned floodlight 214 is that can emit infrared light
Area source;Above-mentioned color-changing lamp 218 is that can emit the point light source of laser and be with figuratum point light source.Wherein, when floodlight 214
When emitting area source, Laser video camera head 212 can obtain infrared image according to the light being reflected back.When 218 launch point light of color-changing lamp
When source, Laser video camera head 212 can obtain speckle image according to the light being reflected back.Above-mentioned speckle image is that color-changing lamp 218 emits
Reflected with figuratum point light source after the pattern image that deformation occurs.
First processing units 230 can be separately connected RGB camera 216 and Laser video camera head 212 by signal wire.Work as RGB
When camera 216 acquires every frame image, synchronization signal can be sent to first processing units 230.First processing units 230 receive
After the synchronization signal that RGB camera 216 is sent, the exposure of the exposure time and RGB camera 216 of Laser video camera head 212 can be obtained
Light time is long, and according to the exposure time computation delay duration of the exposure time of Laser video camera head 212 and RGB camera 216.When
It, can be by signal wire to Laser video camera head 212 when the duration that one processing unit 230 receives synchronization signal reaches the delay duration
Forward synchronization signal.Laser video camera head 212 receives synchronization signal, can be started to be exposed and acquire according to synchronization signal infrared
Image or speckle image etc..
The second processing unit 220 may include in TEE (Trusted execution environment, credible operation ring
Border) CPU core that runs under environment and under REE (Rich Execution Environment, natural running environment) environment
The CPU core of operation.Wherein, TEE environment and REE environment are ARM module (Advanced RISC Machines, advanced essence
Simple instruction set processor) operational mode.Wherein, the security level of TEE environment is higher, has in the second processing unit 220 and only
There is a CPU core that can operate in simultaneously under TEE environment.Under normal conditions, the higher operation of 200 medium security level of electronic equipment
Behavior needs execute in the CPU core under TEE environment, and the lower operation behavior of security level can be in the CPU under REE environment
It is executed in core.
First processing units 230 include PWM (Pulse Width Modulation, pulse width modulation) module 232,
SPI/I2C (Serial Peripheral Interface/Inter-Integrated Circuit, Serial Peripheral Interface (SPI)/bis-
To two-wire system synchronous serial interface) interface 234, RAM (Random Access Memory, random access memory) module 236
With depth engine 238.Above-mentioned PWM module 232 can emit pulse to camera module, control floodlight 214 or color-changing lamp 218 is opened
It opens, so that Laser video camera head 212 can collect infrared image or speckle image.Above-mentioned SPI/I2C interface 234 is for receiving second
The image capture instruction that processing unit 220 is sent.Above-mentioned depth engine 238 can handle speckle image to obtain depth parallax
Figure.
When the second processing unit 220 receives the data acquisition request of application program, for example, when application program need into
When the unlock of pedestrian's face, face payment, image can be sent to first processing units 230 by the CPU core operated under TEE environment
Acquisition instructions.After first processing units 230 receive image capture instruction, impulse wave control can be emitted by PWM module 232
Floodlight 214 is opened and acquires infrared image, control camera module 210 by Laser video camera head 212 in camera module 210
Middle color-changing lamp 218 is opened and acquires speckle image by Laser video camera head 212.Camera module 210 can will be collected infrared
Image and speckle image are sent to first processing units 230.First processing units 230 can be to the infrared image received at
Reason obtains infrared disparity map;The speckle image received is handled to obtain speckle disparity map or depth parallax figure.Wherein,
One processing unit 230 carries out processing to above-mentioned infrared image and speckle image and refers to and carry out school to infrared image or speckle image
Just, influence of the inside and outside parameter to image in camera module 210 is removed.Wherein, first processing units 230 can be set to different
The image of mode, different mode output is different.When first processing units 230 are set as speckle chart-pattern, first processing units
230 pairs of speckle image processings obtain speckle disparity map, and target speckle image can be obtained according to above-mentioned speckle disparity map;At first
When reason unit 230 is set as depth chart-pattern, first processing units 230 obtain depth parallax figure to speckle image processing, according to
Depth image can be obtained in above-mentioned depth parallax figure, and above-mentioned depth image refers to the image with depth information.First processing units
230 can be sent to above-mentioned infrared disparity map and speckle disparity map the second processing unit 220, and first processing units 230 can also incite somebody to action
Above-mentioned infrared disparity map and depth parallax figure are sent to the second processing unit 220.The second processing unit 220 can be according to above-mentioned infrared
Disparity map obtains Infrared Targets image, obtains depth image according to above-mentioned depth parallax figure.Further, the second processing unit
220 can carry out recognition of face, face matching, In vivo detection and acquisition according to Infrared Targets image, depth image detects
Face depth information.
Communicate between first processing units 230 and the second processing unit 220 is by fixed safe interface, to ensure
Transmit the safety of data.As shown in Figure 1, the data that the second processing unit 220 is sent to first processing units 230 are to pass through
SECURE SPI/I2C 240, the data that first processing units 230 are sent to the second processing unit 220 are to pass through SECURE
MIPI (Mobile Industry Processor Interface, mobile industry processor interface) 250.
In one embodiment, first processing units 230 can also be obtained according to above-mentioned infrared disparity map Infrared Targets image,
Above-mentioned depth parallax figure, which calculates, obtains depth image, then above-mentioned Infrared Targets image, depth image are sent to second processing list
Member 220.
Fig. 3 is the block diagram of electronic equipment in one embodiment.As shown in figure 3, the electronic equipment includes passing through system bus
Processor, memory, display screen and the input unit of connection.Wherein, memory may include non-volatile memory medium and processing
Device.The non-volatile memory medium of electronic equipment is stored with operating system and computer program, and the computer program is by processor
Method when execution to realize a kind of control provided in the embodiment of the present application shooting.The processor is calculated and is controlled for providing
Ability supports the operation of entire electronic equipment.Built-in storage in electronic equipment is the computer in non-volatile memory medium
The operation of program provides environment.The display screen of electronic equipment can be liquid crystal display or electric ink display screen etc., input
Device can be the touch layer covered on display screen, be also possible to the key being arranged on electronic equipment casing, trace ball or touch-control
Plate is also possible to external keyboard, Trackpad or mouse etc..The electronic equipment can be mobile phone, tablet computer or a number
Word assistant or wearable device etc..It will be understood by those skilled in the art that structure shown in Fig. 3, only with the application side
The block diagram of the relevant part-structure of case, does not constitute the restriction for the electronic equipment being applied thereon to application scheme, specifically
Electronic equipment may include perhaps combining certain components or with different than more or fewer components as shown in the figure
Component layout.
As shown in figure 4, in one embodiment, providing a kind of method of control shooting, comprising the following steps:
Step 410, when first processing units receive the image capture instruction of the second processing unit transmission, according to image
Acquisition instructions control the first camera and acquire the first image, and image capture instruction receives data acquisition for the second processing unit and asks
It is sent when asking, data acquisition request is used to indicate the second processing unit control second camera and acquires the second image.
When application program needs to obtain human face data in electronic equipment, the first camera can control to open, and acquires the
One image, wherein human face data may include but be not limited under the scenes such as face unlock, face payment need to carry out face verification
Data and face depth information etc..First camera can be Laser video camera head, and Laser video camera head can collect different wave length
Black light image.First image may include but be not limited to infrared image, speckle image etc., and speckle image is referred to scattered
The infrared image of spot image.
When application program needs to obtain human face data, data acquisition request can be sent to the second processing unit.At second
After managing unit reception data acquisition request, image capture instruction can be sent to first processing units, wherein first processing units can
To be MCU module, the second processing unit can be CPU module.Optionally, the second processing unit can first detection data acquisition request
In whether include visible images acquisition instruction, if can illustrate that application program is obtaining comprising visible images acquisition instruction
While human face data, need to obtain the visible images comprising face simultaneously.If in data acquisition request including visible light figure
As acquisition instruction, the second processing unit can control second camera according to visible images acquisition instruction and acquire the second image,
In, second camera can be RGB camera, and the second image then can be the RGB image comprising face.
After first processing units receive image capture instruction, the acquisition of the first camera can be controlled according to image capture instruction
First image, wherein the first image may include infrared image, speckle image etc..First processing units are controllable to open camera
Floodlight in mould group simultaneously acquires infrared image by Laser video camera head, can open the lasers such as the color-changing lamp in camera module
And speckle image etc. is acquired by Laser video camera head.Floodlight can be a kind of point light source of to all the winds uniform irradiation, floodlight
The light of lamp transmitting can be infrared light, and Laser video camera head can acquire face and obtain infrared image.The laser that laser issues can be by
Lens and DOE (diffractive optical elements, diffractive-optical element) carry out diffraction and generate with speckle particle
Pattern, by the pattern projection with speckle particle to target object, by target object each point, difference is produced at a distance from electronic equipment
The offset of raw speckle pattern, Laser video camera head are acquired target object to obtain speckle image.
Step 420, when first processing units receive the synchronization signal of second camera transmission, the first camera is obtained
The first exposure time and second camera the second exposure time, synchronization signal is that second camera acquires every the second image of frame
When at the time of starting exposure the signal that sends.
First processing units can connect the first camera by control line, pass through control the first camera of line traffic control acquisition the
One image.The second processing unit can connect second camera by control line, pass through control line traffic control second camera acquisition the
Two images.First processing units can be connect with the second processing unit.First processing units can also be by signal wire respectively with first
Camera and second camera connection, wherein signal wire can be synchronous signal line.
Second camera, can be at the time of starting exposure to the first processing for being connected to signal wire when acquiring every frame image
Unit sends synchronization signal, which can be the beginning flag SOF (Start of Frame) of frame, can be used for every frame figure
As starting to expose.When first processing units receive the synchronization signal of second camera transmission, the of the first camera can be obtained
Second exposure time of one exposure time and second camera, exposure time may refer to photosensitive duration, and exposure time is longer, into
Light can be more.Usually, the second exposure time difference of the first exposure time of the first camera and second camera compared with
Greatly, the first exposure time of the first camera is smaller than the second exposure time of second camera, but not limited to this, it may also deposit
The case where the first exposure time of the first camera is greater than the second exposure time of second camera etc..
Step 430, according to the first exposure time and the second exposure time computation delay duration.
First processing units can be according to the first exposure time of the first camera and the second exposure time of second camera
Computation delay duration, the delay duration are referred to extending the time span that the first camera starts exposure, be taken the photograph by delaying first
At the time of starting exposure as head, to can guarantee that the first camera is synchronous with second camera.
In one embodiment, it is same in exposure process that the first camera and second camera can be preset in electronic equipment
At the time of step, wherein may refer to the duration that the first camera has exposed at the time of synchronous in exposure process and account for the first exposure
The ratio that the duration that the ratio and second camera of duration have exposed accounts for the second exposure time is identical.For example, settable first takes the photograph
It is consistent as head and second camera terminate to expose simultaneously, or at the time of expose half, or in exposure arrival 3/4 when
It carves consistent etc..First processing units can be synchronous in exposure process according to the first exposure time, the second exposure time and setting
At the time of computation delay duration.
Step 440, when the duration that first processing units receive synchronization signal reaches delay duration, to the first camera
Synchronization signal is forwarded, synchronization signal is used to indicate the first camera and starts to expose and acquire the first image.
After delay duration is calculated in first processing units, delay duration can be reached in the duration for receiving synchronization signal
When, synchronization signal is forwarded to the first camera.After first camera receives synchronization signal, start to be exposed, to can guarantee
It is consistent at the time of first camera and second camera synchronous in exposure process.For example, electronic equipment can preparatory equipment
Consistent at the time of exposing half, then delay duration is calculated in first processing units, and in the duration for receiving synchronization signal
When reaching delay duration, synchronization signal is forwarded to the first camera, can be when the first camera be exposed to half, second takes the photograph
As head is also exposed to half, the two is consistent.
Step 450, the first image is handled by first processing units, and the first image is sent to by treated
The second processing unit.
First image of acquisition can be sent to first processing units by the first camera, and first processing units can be to the first figure
As being handled.First processing units can be set to different modes, and different mode can acquire the first different images, and to
One image carries out different processing etc..When first processing units are infrared mode, first processing units are controllable to open floodlight
Lamp, and infrared image is acquired by the first camera, infrared image can be handled to obtain infrared disparity map.When the first processing
When unit is speckle chart-pattern, first processing units are controllable to open color-changing lamp, and acquires speckle image by the first camera,
Speckle image can be handled to obtain speckle disparity map.When first processing units are depth chart-pattern, first processing units
Speckle image can be handled to obtain depth parallax figure.
In one embodiment, first processing units can be corrected processing to the first image, be corrected processing and refer to
The first image picture material as caused by inside and outside parameter of the first camera and second camera etc. is corrected to deviate, such as due to
Offset of picture material caused by placement position between Laser video camera head deflection angle, Laser video camera head and RGB camera etc. etc..
After being corrected processing to the first image, the disparity map of the first image can be obtained, handled for example, being corrected to infrared image
To infrared disparity map, speckle image is corrected, speckle disparity map or depth parallax figure etc. can be obtained.First image is carried out
Correction process can prevent the image finally presented on the screen of the electronic device from the case where ghost image occur.
First processing units handle the first image, can will treated that the first image is sent to second processing list
Member.The second processing unit can the first image obtains target image, such as Infrared Targets image, target speckle pattern according to treated
Picture and target depth figure etc..The second processing unit can be handled target image according to the demand of application program.
For example, then the second processing unit can then be carried out according to target image etc. when application program needs to carry out face verification
Face datection, wherein Face datection may include recognition of face, face matching and In vivo detection.Recognition of face refers to identification target
It whether there is face in image, face matching, which refers to, matches face in target image with the face prestored, In vivo detection
Refer to whether face has bioactivity etc. in detection target image.If application program needs to obtain the depth information of face,
The target depth figure of generation can be uploaded to application program, application program can carry out at U.S. face according to the target depth figure received
Reason, three-dimensional modeling etc..
In the present embodiment, when first processing units receive the synchronization signal of second camera transmission, according to two
The exposure time computation delay duration of camera, when the duration that first processing units receive synchronization signal reaches delay duration
When, synchronization signal is forwarded to the first camera, is forwarded according to the dynamic adjustment of the exposure time of the first camera and second camera
At the time point of synchronization signal, to be dynamically adapted the first camera opportunity synchronous with second camera, synchronous effect is good,
When exposure time difference between two cameras is larger, it can still guarantee that the picture material of two cameras acquisition is consistent.
In one embodiment, step 430 is according to the first exposure time and the second exposure time computation delay duration, packet
It includes: calculating the exposure time difference of the first exposure time and the second exposure time, and the time difference will be exposed divided by 2, obtain delay duration.
Settable first camera of electronic equipment is consistent at the time of exposing half with second camera, when the first camera
When being exposed to half, second camera is also exposed to half.When first processing units receive second camera transmission
After synchronization signal, the exposure time difference of the first exposure time and the second exposure time can be calculated, and the time difference will be exposed divided by 2, is obtained
Delay duration.Delay duration T3=T1-T2/ 2, wherein T1Indicate the first exposure time, T2Indicate the second exposure time.For example, the
First exposure time of one camera is 3ms (millisecond), and the second exposure time of second camera is 30ms, then can first calculate the
The exposure time difference of one exposure time and the second exposure time is 17ms, and will expose the time difference divided by 2, and obtaining delay duration is
13.5ms。
Optionally, after the exposure time difference that first processing units calculate the first exposure time and the second exposure time, can first by
The exposure time difference is compared with time threshold, judges to expose whether the time difference is greater than time threshold, if more than time threshold, then can incite somebody to action
The time difference is exposed divided by 2, obtains delay duration, and reach delay duration in the duration that first processing units receive synchronization signal
When, then synchronization signal is forwarded to the first camera.If exposing the time difference is less than or equal to time threshold, first processing units can be straight
Connect to the first camera forward synchronization signal, do not extend the first camera start exposure at the time of.Time threshold can be according to reality
Demand is set, such as 1ms, 2ms etc., and the acquisition picture material of the first camera of guarantee and second camera is tolerable
It distinguishes within error, mitigates the calculating pressure of first processing units.
In one embodiment, in order to guarantee that second camera also exposes when the first camera is exposed to half
To half, first processing units can also calculate separately the first intermediate exposures moment and the second exposure time of the first exposure time
The second intermediate exposures moment, wherein at the time of the intermediate exposures moment refers to being exposed to half.First processing units can determine
The difference at one intermediate exposures moment and the second intermediate exposures moment, and using the difference as delay duration.Delay duration T3=T1/
2-T2/ 2, wherein T1Indicate the first exposure time, T2Indicate the second exposure time.For example, when the first exposure of the first camera
The second exposure time of a length of 3ms, second camera are 30ms, then when can first calculate the first intermediate exposures of the first exposure time
Carving is 1.5ms, and the second intermediate exposures moment of the second exposure time is 15ms, then can calculate the first intermediate exposures moment and second
The difference at intermediate exposures moment is 13.5ms, can be using difference 13.5ms as delay duration.It is to be appreciated that it can also be used
His algorithm guarantees the synchronization between the first camera and second camera, is not limited in above-mentioned several ways.
It in the present embodiment, can be according to the first camera letter synchronous with the dynamic adjustment forwarding of the exposure time of second camera
Number time point guarantee the first camera and the to be dynamically adapted the first camera opportunity synchronous with second camera
Two cameras are consistent at the time of exposing half, and synchronous effect is good.
As shown in figure 5, in one embodiment, step 450 is handled the first image by first processing units, and
By treated, the first image is sent to the second processing unit, comprising the following steps:
Step 502, the reference speckle image for obtaining storage has reference depth information with reference to speckle image.
In camera coordinate system, using vertical imaging plane and across the straight line at mirror surface center as Z axis, if object is imaging
The coordinate of machine coordinate system is (X, Y, Z), then Z value therein is depth information of the object in the video camera imaging plane.If
Application program needs to obtain the depth information of face, then needs to acquire the depth map comprising face depth information.First processing is single
Member is controllable to open color-changing lamp, and acquires speckle image by the first camera.Ginseng can be previously stored in first processing units
Speckle pattern is examined, reference depth information can be had with reference to speckle pattern, can be obtained according to the speckle image of acquisition and with reference to speckle image
The depth information for each pixel for including in speckle image.
Step 504, it will be matched with reference to speckle image with speckle image, and obtain matching result.
First processing units can be successively centered on each pixel for including in the speckle image of acquisition, and selection one pre-
If size block of pixels, such as 31pixel (pixel) * 31pixel size, the block of pixels with selection is searched on reference speckle image
The block to match.The block of pixels and refer to what speckle image matched that first processing units can be selected from the speckle image of acquisition
In block, speckle image is found and with reference to two points on same laser optical path respectively in speckle image, same laser optical path
On two points speckle information it is consistent, two points on same laser optical path can regard as corresponding pixel.With reference to
In speckle image, the depth information of the point on each laser optical path is all known.First processing units can calculate target and dissipate
Offset of the spot image with reference speckle image between two corresponding pixels on same laser optical path, and according to inclined
The depth information for each pixel for including in the speckle pattern of acquisition is calculated in shifting amount.
In one embodiment, first processing units carry out the meter of offset by the speckle image of acquisition and with reference to speckle pattern
It calculates, the depth information for each pixel for including in speckle image is calculated according to offset, calculation formula can be such as formula (1)
It is shown:
Wherein, ZDIndicate the depth information of pixel namely the depth value of pixel;L is Laser video camera head and laser
The distance between;F is the focal length of lens in Laser video camera head, Z0Reference planes are set apart from electronics when for reference to speckle image acquisition
The depth value of standby Laser video camera head, P are the offset in the speckle image and reference speckle image of acquisition between corresponding pixel points
Amount.P can be obtained by target speckle pattern and the amount of pixels with reference to pixel offset in speckle pattern multiplied by the actual range of a pixel
It arrives.When the distance between target object and Laser video camera head are greater than the distance between reference planes and Laser video camera head, P is negative
Value, when the distance between target object and Laser video camera head are less than the distance between reference planes and the first collector, P is positive
Value.
Step 506, depth parallax figure is generated according to reference depth information and matching result, and depth parallax figure is sent to
The second processing unit is handled to obtain depth map by the second processing unit to depth parallax figure.
The depth information for each pixel for including in the speckle image that first processing units are acquired, can be to acquisition
Speckle image is corrected processing, corrects the speckle image of acquisition due to the first camera and the inside and outside parameter of second camera etc.
Caused by picture material deviate.First processing units can be according to each pixel in the speckle image and speckle image after correction
The depth value of point generates depth parallax figure, and depth parallax figure is sent to the second processing unit.The second processing unit can basis
Depth parallax figure obtains depth map, may include the depth information of each pixel in depth map.The second processing unit can be by depth
Figure is uploaded to application program, and application program can carry out U.S. face, three-dimensional modeling etc. according to the depth information of face in depth map.Second
Processing unit can also carry out In vivo detection according to the depth information of face in depth map, and it is two-dimensional flat for can preventing the face of acquisition
Dough figurine face etc..
In the present embodiment, the depth information for the image that can be precisely acquired by first processing units, data processing
It is high-efficient, and improve the accuracy of image procossing.
As shown in fig. 6, in one embodiment, before the reference speckle image that step 502 obtains storage, further include with
Lower step:
Step 602, it every the temperature of acquisition time section acquisition laser, and obtains corresponding with temperature with reference to speckle pattern
Picture.
Electronic equipment can be provided with temperature sensor by laser, wherein laser refers to color-changing lamp etc., and passes through
The temperature of temperature sensor acquisition laser.The second processing unit can obtain swashing for temperature sensor acquisition every acquisition time section
The temperature of light device, wherein acquisition time section can be set according to actual needs, such as 3 seconds, 4 seconds etc., but not limited to this.By
In deformation when the temperature of laser changes, may be caused to camera module, influences the first camera and second and take the photograph
As the inside and outside parameter of head.Influence under different temperatures to camera is different, therefore, at different temperature, can correspond to different
With reference to speckle image
The second processing unit can obtain corresponding with temperature with reference to speckle image, and refer to speckle according to corresponding with temperature
Image handles the speckle image acquired at such a temperature, obtains depth map.Optionally, the second processing unit can be set in advance
Fixed multiple and different temperature range, such as 0 DEG C of (degrees celsius)~30 DEG C, 30 DEG C~60 DEG C, 60 DEG C~90 DEG C etc., but be not limited to
This, different temperatures section can correspond to different reference speckle images.After the second processing unit temperature collection, it may be determined that the temperature institute
The temperature range at place, and obtain corresponding with the temperature range with reference to speckle image.
Step 604, the reference speckle image that stores is not in the reference speckle image and first processing units obtained when this
When consistent, first processing units are written in the reference speckle pattern that this is obtained.
After the second processing unit obtains reference speckle image corresponding with the temperature of acquisition, this reference obtained can determine whether
Whether the reference speckle image stored in speckle image and first processing units is consistent, with reference to can carry image in speckle image
Mark, image identification by one of number, wordline and character etc. or a variety of can form.The second processing unit can be from first
Manage unit in read storage reference speckle image image identification, and by this obtain reference speckle image image identification
It is compared with the image identification read from first processing units.If two image identifications are inconsistent, it can illustrate this acquisition
Reference speckle image and first processing units in the reference speckle image that stores it is inconsistent, then the second processing unit can be by this
First processing units are written in the reference speckle image of acquisition.First processing units can store the reference speckle image being newly written, and
The reference speckle image stored before deleting.
In the present embodiment, it can be obtained according to the temperature of laser corresponding with temperature with reference to speckle image, reduction temperature
Influence to the depth map finally exported, the depth information made are more accurate.
As shown in fig. 7, in one embodiment, the method for above-mentioned control shooting is further comprising the steps of:
Step 702, when the second processing unit receives the data acquisition request of application program, the safety of application program is obtained
Rank.
In one embodiment, the second processing unit may include two kinds of operational modes in electronic equipment, wherein the first operation
Mode can be TEE, and TEE is credible running environment, and security level is high;Second operational mode can be REE, and REE is nature fortune
The safety level of row environment, REE is lower.After the second processing unit receives the data acquisition request of application program transmission, it can lead to
It crosses the first operational mode and sends image capture instruction to first processing units.When the second processing unit be monokaryon when, can directly by
Above-mentioned monokaryon is switched to the first operational mode by the second operational mode;It, can be by a kernel when the second processing unit is multicore
First operational mode is switched to by the second operational mode, other kernels still operate in the second operational mode, and by operating in
Kernel under first operational mode sends image capture instruction to first processing units.First figure of the first processing units to acquisition
It, can the first image be sent to the kernel of the operation in the first operating mode by treated as after being handled, it is ensured that the
One processing unit is run under credible running environment always, improves safety.
When the application program of electronic equipment sends data acquisition request to the second processing unit, the second processing unit can be obtained
The application type of application program is taken, and obtains security level corresponding with application type.Application type may include but be not limited to solve
Lock application, payment application, camera applications, U.S. face application etc..The security level of different application type can be different, for example, payment is answered
With that can be height with the corresponding security level of unlock application, camera applications, U.S. face can be low using corresponding security level, but not
It is limited to this.
Step 704, data transmission channel corresponding with security level is determined.
The second processing unit can determine data transmission channel corresponding with the security level of application program, data transmission channel
It may include but be not limited to exit passageway and Common passageway, wherein exit passageway can correspond to the higher application program of security level, general
Circulation passage can correspond to the lower application program of security level.For example, payment application can correspond to exit passageway, U.S. face application can be corresponded to
Common passageway.In exit passageway, the data of transmission can be encrypted, avoid leaking data or be stolen.
Step 706, depth map is sent to application program by corresponding data transmission channel.
The second processing unit can be sent depth map by data transmission channel corresponding with the security level of application program
To application program, depth map is sent to the higher application program of security level by exit passageway, depth map can be encrypted,
Depth map is sent to the lower application program of security level by common road, it can speeding up data transmission speed.Optionally, in addition to
Application program is sent outside depth map, can also be by data transmission channel corresponding with the security level of application program to application program
Other data are sent, for example, the verification result etc. of face verification is carried out, but not limited to this.
In one embodiment, the second processing unit can according to the security level of application program, to application program send with
Security level corresponds to the depth map of precision, and precision is higher, and corresponding depth image is more clear, and the depth information for including is more.Peace
The complete high application program of rank can correspond to depth map with high accuracy, and the low application program of security level can correspond to the low depth of precision
Figure.Optionally, the second processing unit can adjust the precision of images of image data by adjusting image resolution ratio, and resolution ratio is higher,
The precision of images is higher, and resolution ratio is lower, and the precision of images is lower.The number of the point of control color-changing lamp diffraction, image essence can also be passed through
Degree is higher, and the point of diffraction can be more, and the precision of images is lower, and the point of diffraction can be fewer.It is to be appreciated that other modes can also be used
Control the precision of images, however it is not limited to above-mentioned several ways.It, can be with according to the precision of the security level percentage regulation figure of application program
Improve the safety of data.
In the present embodiment, corresponding data channel is chosen according to the security level of application program to transmit data, improved
The safety of data transmission.
In one embodiment, a kind of method of control shooting is provided, comprising the following steps:
Step (1), when first processing units receive the image capture instruction of the second processing unit transmission, according to image
Acquisition instructions control the first camera and acquire the first image, and image capture instruction receives data acquisition for the second processing unit and asks
It is sent when asking, data acquisition request is used to indicate the second processing unit control second camera and acquires the second image.
In one embodiment, first processing units connect the first camera by control line, and the second processing unit passes through
Control line connects second camera, and first processing units are connect with the second processing unit, and first processing units also pass through signal wire
It is connect respectively with the first camera and second camera.
Step (2) obtains the first camera when first processing units receive the synchronization signal of second camera transmission
The first exposure time and second camera the second exposure time, synchronization signal is that second camera acquires every the second image of frame
When at the time of starting exposure the signal that sends.
Step (3), according to the first exposure time and the second exposure time computation delay duration.
In one embodiment, step (3), comprising: the exposure time difference of the first exposure time and the second exposure time is calculated,
And the time difference will be exposed divided by 2, obtain delay duration.
In one embodiment, step (3), comprising: calculate separately the first exposure time the first intermediate exposures moment and
The second intermediate exposures moment of the second exposure time;Determine the difference at the first intermediate exposures moment and the second intermediate exposures moment,
And take the difference as delay duration.
Step (4), when the duration that first processing units receive synchronization signal reaches delay duration, to the first camera
Synchronization signal is forwarded, synchronization signal is used to indicate the first camera and starts to expose and acquire the first image.
Step (5) is handled the first image by first processing units, and the first image is sent to by treated
The second processing unit.
In one embodiment, step (5), comprising: the reference speckle image for obtaining storage, with reference to speckle image with ginseng
Examine depth information;It will be matched with reference to speckle image with speckle image, and obtain matching result;According to reference depth information and
Depth parallax figure is generated with result, and depth parallax figure is sent to the second processing unit, by the second processing unit to depth
Disparity map is handled to obtain depth map.
In one embodiment, before the reference speckle image that step obtains storage, further includes: every acquisition time section
The temperature of laser is acquired, and is obtained corresponding with temperature with reference to speckle image;When the reference speckle image and that this is obtained
When the reference speckle image stored in one processing unit is inconsistent, the first processing of reference speckle image write-in that this is obtained is single
Member.
In one embodiment, the method for above-mentioned control shooting, further includes: when the second processing unit receives application program
Data acquisition request, obtain the security level of application program;Determine data transmission channel corresponding with security level;By depth
Figure is sent to application program by corresponding data transmission channel.
In the present embodiment, when first processing units receive the synchronization signal of second camera transmission, according to two
The exposure time computation delay duration of camera, when the duration that first processing units receive synchronization signal reaches delay duration
When, synchronization signal is forwarded to the first camera, is forwarded according to the dynamic adjustment of the exposure time of the first camera and second camera
At the time point of synchronization signal, to be dynamically adapted the first camera opportunity synchronous with second camera, synchronous effect is good,
When exposure time difference between two cameras is larger, it can still guarantee that the picture material of two cameras acquisition is consistent.
It should be understood that although each step in above-mentioned each flow diagram is successively shown according to the instruction of arrow
Show, but these steps are not that the inevitable sequence according to arrow instruction successively executes.Unless expressly state otherwise herein, this
There is no stringent sequences to limit for the execution of a little steps, these steps can execute in other order.Moreover, above-mentioned each stream
At least part step in journey schematic diagram may include that perhaps these sub-steps of multiple stages or stage be simultaneously for multiple sub-steps
It is not necessarily and executes completion in synchronization, but can execute at different times, the execution in these sub-steps or stage
Sequence, which is also not necessarily, successively to be carried out, but can be at least the one of the sub-step or stage of other steps or other steps
Part executes in turn or alternately.
In one embodiment, a kind of electronic equipment, including first processing units, the second processing unit and camera are provided
Mould group, first processing units can be connected with the second processing unit and camera module respectively.Camera module may include first taking the photograph
As head and second camera, first processing units can connect the first camera by control line, and the second processing unit can pass through control
Line processed connects second camera.First processing units are connect with the second processing unit, and first processing units also pass through signal wire point
It is not connect with the first camera and second camera.
The second processing unit, for being imaged according to data acquisition request control second when receiving data acquisition request
Head the second image of acquisition, and image capture instruction is sent to first processing units.
First processing units, for being adopted according to image when receiving the image capture instruction of the second processing unit transmission
Collection instruction the first camera of control acquires the first image.
Second camera is sent together at the time of starting exposure to first processing units when for acquiring every the second image of frame
Walk signal.
First processing units are also used to obtain when first processing units receive the synchronization signal of second camera transmission
Take the first exposure time of the first camera and the second exposure time of second camera.
First processing units are also used to according to the first exposure time and the second exposure time computation delay duration.
First processing units are also used to when the duration that first processing units receive synchronization signal reaches delay duration,
The synchronization signal is forwarded to the first camera.
First camera exposes and acquires the first image for starting according to synchronization signal.
First processing units are also used to handle the first image by first processing units, and will be after processing
The first image be sent to the second processing unit.
In the present embodiment, when first processing units receive the synchronization signal of second camera transmission, according to two
The exposure time computation delay duration of camera, when the duration that first processing units receive synchronization signal reaches delay duration
When, synchronization signal is forwarded to the first camera, is forwarded according to the dynamic adjustment of the exposure time of the first camera and second camera
At the time point of synchronization signal, to be dynamically adapted the first camera opportunity synchronous with second camera, synchronous effect is good,
When exposure time difference between two cameras is larger, it can still guarantee that the picture material of two cameras acquisition is consistent.
In one embodiment, first processing units are also used to calculate the exposure of the first exposure time and the second exposure time
Light time is poor, and will expose the time difference divided by 2, obtains delay duration.
In one embodiment, first processing units are also used to calculate separately the first intermediate exposures of the first exposure time
The second intermediate exposures moment at moment and the second exposure time determines the first intermediate exposures moment and the second intermediate exposures moment
Difference, and take the difference as delay duration.
It in the present embodiment, can be according to the first camera letter synchronous with the dynamic adjustment forwarding of the exposure time of second camera
Number time point guarantee the first camera and the to be dynamically adapted the first camera opportunity synchronous with second camera
Two cameras are consistent at the time of exposing half, and synchronous effect is good.
In one embodiment, first processing units are also used to obtain the reference speckle image of storage, and will be with reference to scattered
Spot image is matched with the speckle image, obtains matching result, has reference depth information with reference to speckle image.
First processing units, are also used to generate depth parallax figure according to reference depth information and matching result, and by depth
Disparity map is sent to the second processing unit.
The second processing unit is also used to be handled to obtain depth map to depth parallax figure.
In the present embodiment, the depth information for the image that can be precisely acquired by first processing units, data processing
It is high-efficient, and improve the accuracy of image procossing.
In one embodiment, the second processing unit is also used to acquire the temperature of laser every acquisition time section, and obtains
It takes corresponding with temperature with reference to speckle image.
The second processing unit is also used to when the reference stored in the reference speckle image of this acquisition and first processing units
When speckle image is inconsistent, first processing units are written in the reference speckle image that this is obtained.
In the present embodiment, it can be obtained according to the temperature of laser corresponding with temperature with reference to speckle image, reduction temperature
Influence to the depth map finally exported, the depth information made are more accurate.
In one embodiment, the second processing unit is also used to obtain when the data acquisition request for receiving application program
The security level of application program.
The second processing unit, is also used to determine corresponding with security level data transmission channel, and depth map is passed through pair
The data transmission channel answered is sent to application program.
In the present embodiment, corresponding data channel is chosen according to the security level of application program to transmit data, improved
The safety of data transmission.
As shown in figure 8, in one embodiment, providing a kind of device 800 of control shooting, including image capture module
810, signal receiving module 820, computing module 830, signal forwarding module 840, processing module 850.
Image capture module 810, the Image Acquisition for receiving the second processing unit transmission when first processing units refer to
When enabling, the first camera is controlled according to image capture instruction and acquires the first image, image capture instruction connects for the second processing unit
Transmission when receiving data acquisition request, data acquisition request are used to indicate the second processing unit control second camera acquisition second
Image.
Signal receiving module 820 obtains when for receiving the synchronization signal of second camera transmission when first processing units
The first exposure time of the first camera and the second exposure time of second camera are taken, synchronization signal is second camera acquisition
In the signal for starting to send at the time of exposure when every the second image of frame.
Computing module 830, for according to the first exposure time and the second exposure time computation delay duration.
Signal forwarding module 840, when the duration for receiving synchronization signal when first processing units reaches delay duration,
The synchronization signal is forwarded to the first camera, synchronization signal is used to indicate the first camera and starts to expose and acquire the first figure
Picture.
Processing module 850, for being handled by first processing units the first image, and will treated the first figure
As being sent to the second processing unit.
In the present embodiment, when first processing units receive the synchronization signal of second camera transmission, according to two
The exposure time computation delay duration of camera, when the duration that first processing units receive synchronization signal reaches delay duration
When, synchronization signal is forwarded to the first camera, is forwarded according to the dynamic adjustment of the exposure time of the first camera and second camera
At the time point of synchronization signal, to be dynamically adapted the first camera opportunity synchronous with second camera, synchronous effect is good,
When exposure time difference between two cameras is larger, it can still guarantee that the picture material of two cameras acquisition is consistent.
In one embodiment, computing module 830 are also used to calculate the exposure of the first exposure time and the second exposure time
The time difference, and the time difference will be exposed divided by 2, obtain delay duration.
In one embodiment, computing module 830 calculate separately the first intermediate exposures moment and of the first exposure time
The second intermediate exposures moment of two exposure times determines the difference at the first intermediate exposures moment and the second intermediate exposures moment, and
Take the difference as delay duration.
It in the present embodiment, can be according to the first camera letter synchronous with the dynamic adjustment forwarding of the exposure time of second camera
Number time point guarantee the first camera and the to be dynamically adapted the first camera opportunity synchronous with second camera
Two cameras are consistent at the time of exposing half, and synchronous effect is good.
As shown in figure 9, in one embodiment, processing module 850, including image acquisition unit 852, matching unit 854
And generation unit 856.
Image acquisition unit 852 is believed with reference to speckle image with reference depth for obtaining the reference speckle image of storage
Breath.
Matching unit 854 matches with speckle image for that will refer to speckle image, obtains matching result.
Generation unit 856, for generating depth parallax figure according to reference depth information and matching result, and by depth parallax
Figure is sent to the second processing unit, is handled to obtain depth map to depth parallax figure by the second processing unit.
In the present embodiment, the depth information for the image that can be precisely acquired by first processing units, data processing
It is high-efficient, and improve the accuracy of image procossing.
In one embodiment, above-mentioned processing module 850, in addition to include image acquisition unit 852, matching unit 854 and
Generation unit 856 further includes temperature collecting cell and writing unit.
Temperature collecting cell for the temperature every acquisition time section acquisition laser, and obtains ginseng corresponding with temperature
Examine speckle image.
Writing unit, for when the reference speckle pattern stored in the reference speckle image and first processing units that this is obtained
As it is inconsistent when, by this obtain reference speckle image write-in first processing units.
In the present embodiment, it can be obtained according to the temperature of laser corresponding with temperature with reference to speckle image, reduction temperature
Influence to the depth map finally exported, the depth information made are more accurate.
In one embodiment, the device 800 of above-mentioned control shooting, in addition to including image capture module 810, signal reception
Module 820, computing module 830, signal forwarding module 840 and processing module 850 further include that rank obtains module, channel determines
Module and sending module.
Rank obtains module, for receiving the data acquisition request of application program when the second processing unit, obtains application
The security level of program.
Channel determining module, for determining data transmission channel corresponding with security level.
Sending module, for depth map to be sent to application program by corresponding data transmission channel.
In the present embodiment, corresponding data channel is chosen according to the security level of application program to transmit data, improved
The safety of data transmission.
In one embodiment, a kind of computer readable storage medium is provided, computer program is stored thereon with, the calculating
Machine program realizes above-mentioned control shooting method when being executed by processor.
In one embodiment, a kind of computer program product comprising computer program is provided, when it sets in computer
When standby upper operation, so that computer equipment realizes above-mentioned control shooting method when executing.
Those of ordinary skill in the art will appreciate that realizing all or part of the process in above-described embodiment method, being can be with
Relevant hardware is instructed to complete by computer program, the program can be stored in a non-volatile computer and can be read
In storage medium, the program is when being executed, it may include such as the process of the embodiment of above-mentioned each method.Wherein, the storage is situated between
Matter can be magnetic disk, CD, read-only memory (Read-Only Memory, ROM) etc..
It may include as used herein non-volatile to any reference of memory, storage, database or other media
And/or volatile memory.Suitable nonvolatile memory may include read-only memory (ROM), programming ROM (PROM),
Electrically programmable ROM (EPROM), electrically erasable ROM (EEPROM) or flash memory.Volatile memory may include arbitrary access
Memory (RAM), it is used as external cache.By way of illustration and not limitation, RAM is available in many forms, such as
It is static RAM (SRAM), dynamic ram (DRAM), synchronous dram (SDRAM), double data rate sdram (DDR SDRAM), enhanced
SDRAM (ESDRAM), synchronization link (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM
(RDRAM), direct memory bus dynamic ram (DRDRAM) and memory bus dynamic ram (RDRAM).
Each technical characteristic of embodiment described above can be combined arbitrarily, for simplicity of description, not to above-mentioned reality
It applies all possible combination of each technical characteristic in example to be all described, as long as however, the combination of these technical characteristics is not deposited
In contradiction, all should be considered as described in this specification.
The several embodiments of the application above described embodiment only expresses, the description thereof is more specific and detailed, but simultaneously
It cannot therefore be construed as limiting the scope of the patent.It should be pointed out that coming for those of ordinary skill in the art
It says, without departing from the concept of this application, various modifications and improvements can be made, these belong to the protection of the application
Range.Therefore, the scope of protection shall be subject to the appended claims for the application patent.
Claims (15)
1. a kind of method of control shooting characterized by comprising
When first processing units receive the image capture instruction of the second processing unit transmission, according to described image acquisition instructions
It controls the first camera and acquires the first image, described image acquisition instructions receive data acquisition for described the second processing unit and ask
It is sent when asking, the data acquisition request is used to indicate the second processing unit control second camera and acquires the second image;
When first processing units receive the synchronization signal that the second camera is sent, the of first camera is obtained
Second exposure time of one exposure time and second camera, the synchronization signal are that the second camera acquires every frame second
In the signal for starting to send at the time of exposure when image;
According to first exposure time and the second exposure time computation delay duration;
When the duration that the first processing units receive the synchronization signal reaches the delay duration, Xiang Suoshu first is taken the photograph
As the head forwarding synchronization signal, the synchronization signal is used to indicate first camera and starts to expose and acquire the first figure
Picture;
The first image is handled by the first processing units, and will treated that the first image is sent to is described
The second processing unit;
It is described that the first image is handled by the first processing units, and the first image is sent to by treated
Described the second processing unit, comprising:
Obtain the mode of the first processing units;
The first image is handled according to the mode, obtains the corresponding disparity map of the mode, and by the parallax
Figure is sent to described the second processing unit, is handled by described the second processing unit the disparity map to obtain target figure
Picture.
2. the method according to claim 1, wherein the first processing units connect described by control line
One camera, described the second processing unit connect the second camera by control line, the first processing units with it is described
The second processing unit connection, the first processing units also pass through signal wire respectively with first camera and second camera
Connection.
3. the method according to claim 1, wherein it is described according to first exposure time and second expose when
Long computation delay duration, comprising:
The exposure time difference of first exposure time and the second exposure time is calculated, and the exposure time difference is prolonged divided by 2
Shi Shichang.
4. the method according to claim 1, wherein it is described according to first exposure time and second expose when
Long computation delay duration, comprising:
Calculate separately first exposure time the first intermediate exposures moment and the second exposure time the second intermediate exposures when
It carves;
When determining the difference at the first intermediate exposures moment and the second intermediate exposures moment, and taking the difference as delay
It is long.
5. the method according to claim 1, wherein the first image includes speckle image;
It is described that the first image is handled according to the mode, the corresponding disparity map of the mode is obtained, and will be described
Disparity map is sent to described the second processing unit, is handled to obtain target to the disparity map by described the second processing unit
Image, comprising:
When the mode is depth chart-pattern, the reference speckle image of storage is obtained, it is described to have reference with reference to speckle image
Depth information;
It is matched described with reference to speckle image with the speckle image, obtains matching result;
Depth parallax figure is generated according to the reference depth information and matching result, and the depth parallax figure is sent to described
The second processing unit is handled to obtain depth map by described the second processing unit to the depth parallax figure.
6. according to the method described in claim 5, it is characterized in that, obtaining described with reference to before speckle image, the method
Further include:
Every the temperature of acquisition time section acquisition laser, and obtain corresponding with the temperature with reference to speckle image;
It, will be described when the reference speckle image stored in the reference speckle image and first processing units that this is obtained is inconsistent
The first processing units are written in this reference speckle image obtained.
7. according to the method described in claim 5, it is characterized in that, the method also includes:
When the second processing unit receives the data acquisition request of application program, the security level of the application program is obtained;
Determine data transmission channel corresponding with the security level;
The depth map is sent to the application program by the corresponding data transmission channel.
8. a kind of device of control shooting characterized by comprising
Image capture module, when for receiving the image capture instruction of the second processing unit transmission when first processing units, root
The first camera is controlled according to described image acquisition instructions and acquires the first image, and described image acquisition instructions are the second processing list
Transmission when member receives data acquisition request, the data acquisition request are used to indicate the second processing unit control second and take the photograph
As head acquires the second image;
Signal receiving module, for obtaining when first processing units receive the synchronization signal that the second camera is sent
First exposure time of first camera and the second exposure time of second camera, the synchronization signal are described second
In the signal for starting to send at the time of exposure when camera acquires every the second image of frame;
Computing module, for according to first exposure time and the second exposure time computation delay duration;
Signal forwarding module, when the duration for receiving the synchronization signal when the first processing units reaches the delay
When long, the first camera of Xiang Suoshu forwards the synchronization signal, and the synchronization signal is used to indicate first camera and starts
It exposes and acquires the first image;
Processing module, for being handled by the first processing units the first image, and will treated first
Image is sent to described the second processing unit;
The processing module, is specifically used for: obtaining the mode of the first processing units;
The first image is handled according to the mode, obtains the corresponding disparity map of the mode, and by the parallax
Figure is sent to described the second processing unit, is handled by described the second processing unit the disparity map to obtain target figure
Picture.
9. a kind of electronic equipment, which is characterized in that described including first processing units, the second processing unit and camera module
First processing units are connected with described the second processing unit and camera module respectively;The camera module includes the first camera shooting
Head and second camera, the first processing units connect first camera, described the second processing unit by control line
The second camera is connected by control line, the first processing units are connect with described the second processing unit, and described first
Processing unit also passes through signal wire and connect respectively with first camera and second camera;
Described the second processing unit, for when receiving data acquisition request, according to data acquisition request control
Second camera acquires the second image, and sends image capture instruction to the first processing units;
The first processing units, for when receiving the image capture instruction that described the second processing unit is sent, according to institute
It states image capture instruction and controls the first camera the first image of acquisition;
The second camera is sent together at the time of starting exposure to first processing units when for acquiring every the second image of frame
Walk signal;
The first processing units are also used to receive the synchronization signal that the second camera is sent when first processing units
When, obtain the first exposure time of first camera and the second exposure time of second camera;
The first processing units are also used to according to first exposure time and the second exposure time computation delay duration;
The first processing units are also used to reach described when the duration that the first processing units receive the synchronization signal
When delay duration, the first camera of Xiang Suoshu forwards the synchronization signal;
First camera exposes and acquires the first image for starting according to synchronization signal;
The first processing units are also used to handle the first image by the first processing units, and will place
The first image after reason is sent to described the second processing unit.
10. electronic equipment according to claim 9, which is characterized in that the first processing units are also used to calculate described
The exposure time difference of first exposure time and the second exposure time, and the exposure time difference is obtained into delay duration divided by 2.
11. electronic equipment according to claim 9, which is characterized in that the first processing units are also used to calculate separately
First intermediate exposures moment of first exposure time and the second intermediate exposures moment of the second exposure time determine described
The difference at one intermediate exposures moment and the second intermediate exposures moment, and take the difference as delay duration.
12. electronic equipment according to claim 9, which is characterized in that the first image includes speckle image, and described
One processing unit is also used to obtain the reference speckle image of storage when the mode is depth chart-pattern, and by the reference
Speckle image is matched with the speckle image, obtains matching result, described to have reference depth information with reference to speckle image;
The first processing units are also used to generate depth parallax figure according to the reference depth information and matching result, and will
The depth parallax figure is sent to described the second processing unit;
Described the second processing unit is also used to be handled to obtain depth map to the depth parallax figure.
13. electronic equipment according to claim 12, which is characterized in that described the second processing unit is also used to every adopting
Collect the temperature of period acquisition laser, and obtains corresponding with the temperature with reference to speckle image;
Described the second processing unit is also used to when the reference stored in the reference speckle image of this acquisition and first processing units
When speckle image is inconsistent, the first processing units are written into the reference speckle image that this is obtained.
14. electronic equipment according to claim 12, which is characterized in that described the second processing unit is also used to when reception
To the data acquisition request of application program, the security level of the application program is obtained;
Described the second processing unit, is also used to determine corresponding with security level data transmission channel, and by the depth
Figure is sent to the application program by the corresponding data transmission channel.
15. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the computer program
The method as described in claim 1 to 7 is any is realized when being executed by processor.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910585686.8A CN110248111B (en) | 2018-04-28 | 2018-04-28 | Method and device for controlling shooting, electronic equipment and computer-readable storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810404282.XA CN108419017B (en) | 2018-04-28 | 2018-04-28 | Control method, apparatus, electronic equipment and the computer readable storage medium of shooting |
CN201910585686.8A CN110248111B (en) | 2018-04-28 | 2018-04-28 | Method and device for controlling shooting, electronic equipment and computer-readable storage medium |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810404282.XA Division CN108419017B (en) | 2018-04-28 | 2018-04-28 | Control method, apparatus, electronic equipment and the computer readable storage medium of shooting |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110248111A true CN110248111A (en) | 2019-09-17 |
CN110248111B CN110248111B (en) | 2020-08-28 |
Family
ID=63137390
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910585686.8A Active CN110248111B (en) | 2018-04-28 | 2018-04-28 | Method and device for controlling shooting, electronic equipment and computer-readable storage medium |
CN201810404282.XA Active CN108419017B (en) | 2018-04-28 | 2018-04-28 | Control method, apparatus, electronic equipment and the computer readable storage medium of shooting |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810404282.XA Active CN108419017B (en) | 2018-04-28 | 2018-04-28 | Control method, apparatus, electronic equipment and the computer readable storage medium of shooting |
Country Status (1)
Country | Link |
---|---|
CN (2) | CN110248111B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111230290A (en) * | 2020-01-17 | 2020-06-05 | 北京工业大学 | System and method for synchronizing ultrafast laser and ICCD camera by photoelectric signal |
CN111292488A (en) * | 2020-02-13 | 2020-06-16 | 展讯通信(上海)有限公司 | Image data processing method, device and storage medium |
CN113132551A (en) * | 2019-12-30 | 2021-07-16 | 浙江舜宇智能光学技术有限公司 | Synchronous control method and device of multi-camera system and electronic equipment |
CN114071022A (en) * | 2020-08-07 | 2022-02-18 | 北京图森未来科技有限公司 | Control method, device and equipment of image acquisition equipment and storage medium |
WO2024060763A1 (en) * | 2022-09-22 | 2024-03-28 | 恒玄科技(上海)股份有限公司 | Wireless smart wearable device and image acquisition method thereof |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3627827B1 (en) | 2018-04-28 | 2024-05-01 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for controlling photographing, electronic device, and computer readable storage medium |
CN111107248B (en) * | 2018-10-25 | 2022-02-08 | 北京图森智途科技有限公司 | Multi-channel video acquisition synchronization system and method and acquisition controller |
CN109635539B (en) * | 2018-10-30 | 2022-10-14 | 荣耀终端有限公司 | Face recognition method and electronic equipment |
WO2020181494A1 (en) * | 2019-03-12 | 2020-09-17 | 深圳市大疆创新科技有限公司 | Parameter synchronization method, image capture apparatus, and movable platform |
CN110460824B (en) * | 2019-07-03 | 2022-10-11 | 青岛小鸟看看科技有限公司 | Frame synchronization method of image data and camera |
CN114071120A (en) * | 2020-08-03 | 2022-02-18 | 炬才微电子(深圳)有限公司 | Camera testing system, method, storage medium and electronic equipment |
CN114371722B (en) * | 2021-12-03 | 2024-05-10 | 深圳供电局有限公司 | Data acquisition method, device, unmanned aerial vehicle and storage medium |
CN114863510B (en) * | 2022-03-25 | 2023-08-01 | 荣耀终端有限公司 | Face recognition method and device |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1710935A (en) * | 2004-06-17 | 2005-12-21 | 株式会社日立制作所 | Imaging apparatus |
CN103338334A (en) * | 2013-07-17 | 2013-10-02 | 中测新图(北京)遥感技术有限责任公司 | System and method for controlling multi-cameral digital aerial photographic camera synchronous exposure |
CN105657243A (en) * | 2015-11-08 | 2016-06-08 | 乐视移动智能信息技术(北京)有限公司 | Anti-jitter delay photographing method and device |
WO2016168781A1 (en) * | 2015-04-17 | 2016-10-20 | The Lightco Inc. | Methods and apparatus for syncronizing readout of multiple image sensors |
US20170202441A1 (en) * | 2014-11-21 | 2017-07-20 | Olympus Corporation | Image pickup system |
CN106973231A (en) * | 2017-04-19 | 2017-07-21 | 宇龙计算机通信科技(深圳)有限公司 | Picture synthetic method and system |
CN107231533A (en) * | 2017-06-12 | 2017-10-03 | 深圳市瑞立视多媒体科技有限公司 | A kind of synchronous exposure method, device and terminal device |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9626803B2 (en) * | 2014-12-12 | 2017-04-18 | Qualcomm Incorporated | Method and apparatus for image processing in augmented reality systems |
CN107040726B (en) * | 2017-04-19 | 2020-04-07 | 宇龙计算机通信科技(深圳)有限公司 | Double-camera synchronous exposure method and system |
CN107395998A (en) * | 2017-08-24 | 2017-11-24 | 维沃移动通信有限公司 | A kind of image capturing method and mobile terminal |
-
2018
- 2018-04-28 CN CN201910585686.8A patent/CN110248111B/en active Active
- 2018-04-28 CN CN201810404282.XA patent/CN108419017B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1710935A (en) * | 2004-06-17 | 2005-12-21 | 株式会社日立制作所 | Imaging apparatus |
CN103338334A (en) * | 2013-07-17 | 2013-10-02 | 中测新图(北京)遥感技术有限责任公司 | System and method for controlling multi-cameral digital aerial photographic camera synchronous exposure |
US20170202441A1 (en) * | 2014-11-21 | 2017-07-20 | Olympus Corporation | Image pickup system |
WO2016168781A1 (en) * | 2015-04-17 | 2016-10-20 | The Lightco Inc. | Methods and apparatus for syncronizing readout of multiple image sensors |
CN105657243A (en) * | 2015-11-08 | 2016-06-08 | 乐视移动智能信息技术(北京)有限公司 | Anti-jitter delay photographing method and device |
CN106973231A (en) * | 2017-04-19 | 2017-07-21 | 宇龙计算机通信科技(深圳)有限公司 | Picture synthetic method and system |
CN107231533A (en) * | 2017-06-12 | 2017-10-03 | 深圳市瑞立视多媒体科技有限公司 | A kind of synchronous exposure method, device and terminal device |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113132551A (en) * | 2019-12-30 | 2021-07-16 | 浙江舜宇智能光学技术有限公司 | Synchronous control method and device of multi-camera system and electronic equipment |
CN113132551B (en) * | 2019-12-30 | 2023-08-08 | 浙江舜宇智能光学技术有限公司 | Synchronous control method and synchronous control device for multi-camera system and electronic equipment |
CN111230290A (en) * | 2020-01-17 | 2020-06-05 | 北京工业大学 | System and method for synchronizing ultrafast laser and ICCD camera by photoelectric signal |
CN111230290B (en) * | 2020-01-17 | 2021-07-30 | 北京工业大学 | System and method for synchronizing ultrafast laser and ICCD camera by photoelectric signal |
CN111292488A (en) * | 2020-02-13 | 2020-06-16 | 展讯通信(上海)有限公司 | Image data processing method, device and storage medium |
CN114071022A (en) * | 2020-08-07 | 2022-02-18 | 北京图森未来科技有限公司 | Control method, device and equipment of image acquisition equipment and storage medium |
US11706533B2 (en) | 2020-08-07 | 2023-07-18 | Beijing Tusen Weilai Technology Co., Ltd. | Control method for image acquisition device, control device therefor, and storage medium |
CN114071022B (en) * | 2020-08-07 | 2023-12-15 | 北京图森未来科技有限公司 | Control method, device and equipment of image acquisition equipment and storage medium |
WO2024060763A1 (en) * | 2022-09-22 | 2024-03-28 | 恒玄科技(上海)股份有限公司 | Wireless smart wearable device and image acquisition method thereof |
Also Published As
Publication number | Publication date |
---|---|
CN108419017B (en) | 2019-07-30 |
CN110248111B (en) | 2020-08-28 |
CN108419017A (en) | 2018-08-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108419017B (en) | Control method, apparatus, electronic equipment and the computer readable storage medium of shooting | |
CN108668078B (en) | Image processing method, device, computer readable storage medium and electronic equipment | |
EP3627827B1 (en) | Method for controlling photographing, electronic device, and computer readable storage medium | |
CN108764052A (en) | Image processing method, device, computer readable storage medium and electronic equipment | |
CN109040591B (en) | Image processing method, image processing device, computer-readable storage medium and electronic equipment | |
CN108549867A (en) | Image processing method, device, computer readable storage medium and electronic equipment | |
EP3614659B1 (en) | Image processing method, electronic apparatus, and computer-readable storage medium | |
JP2021520154A (en) | Image processing methods, computer-readable storage media, and electronic devices | |
CN108805024A (en) | Image processing method, device, computer readable storage medium and electronic equipment | |
CN108650472A (en) | Control method, apparatus, electronic equipment and the computer readable storage medium of shooting | |
CN108804895A (en) | Image processing method, device, computer readable storage medium and electronic equipment | |
CN108696682B (en) | Data processing method, device, electronic equipment and computer readable storage medium | |
WO2019206020A1 (en) | Image processing method, apparatus, computer-readable storage medium, and electronic device | |
WO2019196683A1 (en) | Method and device for image processing, computer-readable storage medium, and electronic device | |
CN109151303B (en) | Image processing method and device, electronic equipment and computer readable storage medium | |
CN108573170A (en) | Information processing method and device, electronic equipment, computer readable storage medium | |
CN108833887A (en) | Data processing method, device, electronic equipment and computer readable storage medium | |
CN108921903A (en) | Camera calibration method, device, computer readable storage medium and electronic equipment | |
CN108830141A (en) | Image processing method, device, computer readable storage medium and electronic equipment | |
CN109146906A (en) | Image processing method and device, electronic equipment, computer readable storage medium | |
CN109190533A (en) | Image processing method and device, electronic equipment, computer readable storage medium | |
CN104155839B (en) | System and method for providing 3 d image | |
CN108924421A (en) | Image processing method, device, computer readable storage medium and electronic equipment | |
CN108981782A (en) | A method of it is realized using mobile phone and calculates relevance imaging | |
CN108810516B (en) | Data processing method and device, electronic equipment and computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |