CN110278413A - Image processing method, device, server and storage medium - Google Patents
Image processing method, device, server and storage medium Download PDFInfo
- Publication number
- CN110278413A CN110278413A CN201910579248.0A CN201910579248A CN110278413A CN 110278413 A CN110278413 A CN 110278413A CN 201910579248 A CN201910579248 A CN 201910579248A CN 110278413 A CN110278413 A CN 110278413A
- Authority
- CN
- China
- Prior art keywords
- target
- camera
- reference object
- image
- shooting
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/292—Multi-camera tracking
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/65—Control of camera operation in relation to power supply
- H04N23/651—Control of camera operation in relation to power supply for reducing power consumption by affecting camera operations, e.g. sleep mode, hibernation mode or power off of selective parts of the camera
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/265—Mixing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
Abstract
This application discloses a kind of image processing method, device, server and storage mediums, are related to technical field of image processing.This method is applied to server, server is connect with multiple cameras, different positions is arranged in for shooting different regions in multiple camera, and the region of the every two adjacent camera shooting in multiple camera connects or partly overlaps, which is stored with target reference object.This method comprises: from the image that each camera in multiple cameras takes, extract multiple target images including target reference object, obtain the shooting time of each target image in multiple target images, by the sequencing of the shooting time of each target image, multiple target images are sequentially spliced, the monitor video of the target reference object is obtained.The application passes through distributed camera and carries out track up to target reference object, and obtains the monitor video of the target reference object, to promote the monitoring effect of target reference object.
Description
Technical field
This application involves technical field of image processing, more particularly, to a kind of image processing method, device, server
And storage medium.
Background technique
With the development of society and the progress of science and technology, more and more places start to arrange monitoring system.Currently, passing through
In the application scenarios that monitoring system is monitored, can only often some region fixed be supervised by monitoring used camera
Control, monitoring effect are bad.
Summary of the invention
In view of the above problems, present applicant proposes a kind of image processing method, device, server and storage medium, with
It solves the above problems.
In a first aspect, the embodiment of the present application provides a kind of image processing method, be applied to server, the server with
Different positions is arranged in for shooting different regions in the connection of multiple cameras, the multiple camera, and the multiple takes the photograph
As the region that the every two adjacent camera in head is shot connects or partly overlaps, the server is stored with target shooting pair
As, which comprises from the image that each camera in the multiple camera takes, extracting includes the target
Multiple target images of reference object;Obtain the shooting time of each target image in the multiple target image;By described
The sequencing of the shooting time of each target image sequentially splices the multiple target image, obtains the target shooting
The monitor video of object.
Second aspect, the embodiment of the present application provide a kind of image processing method, are applied to distributed system, including service
Different positions is arranged in for shooting difference in device and the multiple cameras connecting with the server, the multiple camera
Region, and in the multiple camera every two adjacent camera shooting region connect or partly overlap, it is described more
Each camera in a camera is stored with target reference object, which comprises every in the multiple camera
A camera extracts multiple target images including the target reference object from the image taken;The multiple camera shooting
Head obtains the shooting time of each target image in the multiple target image;Each camera is by the multiple target
The shooting time of each target image in image and the multiple target image is uploaded to the server;The server is pressed
The sequencing of the shooting time of each target image, the multiple target image is sequentially spliced, and obtains the target
The monitor video of reference object.
The third aspect, the embodiment of the present application provide a kind of image processing apparatus, be applied to server, the server with
Different positions is arranged in for shooting different regions in the connection of multiple cameras, the multiple camera, and the multiple takes the photograph
As the region that the every two adjacent camera in head is shot connects or partly overlaps, the server is stored with target shooting pair
As described device includes: image zooming-out module, the image for taking from each camera in the multiple camera
In, extract multiple target images including the target reference object;Shooting time obtains module, for obtaining the multiple mesh
The shooting time of each target image in logo image;Image mosaic module, when for pressing the shooting of each target image
Between sequencing, the multiple target image is sequentially spliced, the monitor video of the target reference object is obtained.
Fourth aspect, the embodiment of the present application provide a kind of server, including memory and processor, the memory coupling
It is connected to the processor, the memory store instruction, the processor executes when executed by the processor
The above method.
5th aspect, the embodiment of the present application provides a kind of computer-readable storage medium, described computer-readable
Program code is stored in storage medium, said program code can be called by processor and execute the above method.
Image processing method, device, server and storage medium provided by the embodiments of the present application are applied to server,
Server is connect with multiple cameras, and different positions is arranged in for shooting different regions in multiple camera, and this is more
The region of every two adjacent camera shooting in a camera connects or partly overlaps, which is stored with target shooting pair
As.From the image that each camera in multiple cameras takes, multiple target figures including target reference object are extracted
Picture obtains the shooting time of each target image in multiple target images, by each target image shooting time it is successive
Sequentially, multiple target images are sequentially spliced, obtains the monitor video of the target reference object, to pass through distributed camera
Track up is carried out to target reference object, and obtains the monitor video of the target reference object, to promote target reference object
Monitoring effect.
Detailed description of the invention
In order to more clearly explain the technical solutions in the embodiments of the present application, make required in being described below to embodiment
Attached drawing is briefly described, it should be apparent that, the drawings in the following description are only some examples of the present application, for
For those skilled in the art, without creative efforts, it can also be obtained according to these attached drawings other attached
Figure.
Fig. 1 shows the schematic diagram of distributed system provided by the embodiments of the present application;
Fig. 2 shows the flow diagrams for the image processing method that the application one embodiment provides;
Fig. 3 shows the flow diagram of the image processing method of another embodiment of the application offer;
Fig. 4 shows distributed system provided by the embodiments of the present application and carries out track up to multiple target reference objects
Schematic diagram;
Fig. 5 shows the flow diagram of the image processing method of the application further embodiment offer;
Fig. 6 shows a kind of schematic diagram of setting position of multiple cameras provided by the embodiments of the present application;
Fig. 7 shows the schematic diagram of another setting position of multiple cameras provided by the embodiments of the present application;
Fig. 8 shows the flow diagram of the step S303 of the image processing method shown in fig. 5 of the application;
Fig. 9 shows the flow diagram of the image processing method of another embodiment of the application offer;
Figure 10 shows a kind of flow diagram of the step S402 of the image processing method shown in Fig. 9 of the application;
Figure 11 shows another flow diagram of the step S402 of the image processing method shown in fig. 6 of the application;
Figure 12 shows the flow diagram of the application image processing method that further embodiment provides again;
Figure 13 shows the module frame chart of image processing apparatus provided by the embodiments of the present application;
Figure 14 shows the embodiment of the present application for executing the server of the image processing method according to the embodiment of the present application
Block diagram;
Figure 15 shows realizing at according to the image of the embodiment of the present application for saving or carrying for the embodiment of the present application
The storage unit of the program code of reason method.
Specific embodiment
In order to make those skilled in the art more fully understand application scheme, below in conjunction in the embodiment of the present application
Attached drawing, the technical scheme in the embodiment of the application is clearly and completely described.
In recent years, with the progress of the development of society and science and technology, more and more places start to arrange monitoring system, and
In the most application scenarios being monitored by monitoring system, used camera is all often single camera, can only
Some region fixed is monitored, monitoring effect is bad.To solve the above problems, panoramic video monitoring system is gradually emerging
It rises, wherein panoramic video monitoring system includes multiple cameras for being mounted on fixed position, for shooting multiple and different regions
Image simultaneously forms panoramic picture, but range acquired in panoramic picture is excessive, can not be monitored for some object, can not
Reach ideal monitoring effect.
In view of the above-mentioned problems, inventor has found by long-term research and proposes at image provided by the embodiments of the present application
Method, apparatus, server and storage medium are managed, track up is carried out to target reference object by distributed camera, and
The monitor video of the target reference object is obtained, to promote the monitoring effect of target reference object.Wherein, specific image procossing
Method is described in detail in subsequent embodiment.
It will be described below for the distributed system suitable for image processing method provided by the embodiments of the present application.
Referring to Fig. 1, Fig. 1 shows the schematic diagram of distributed system provided by the embodiments of the present application, wherein the distribution
System includes server 100 and multiple cameras 200 (quantity of camera 200 shown in Fig. 1 is 4), wherein server 100
It is connected respectively with each camera 200 in multiple cameras 200, for carrying out data interaction with each camera 200 respectively,
For example, server 100, which receives the image of the transmission of camera 200, server 100, sends instruction etc. to camera 200, do not do herein
It is specific to limit.In addition, the server 100 can be Cloud Server, or traditional server, the camera 200 can be with
For gun-type camera, hemisphere camera, high-definition intelligent spherical shape camera, pen holder type camera, veneer camera, flying saucer camera shooting
Head, Mobile phone type camera etc., and the camera lens of the camera can use wide-angle lens, standard lens, telephoto lens, varifocal mirror
Head, pin hole mirror head etc., do not do specific restriction herein.
In some embodiments, different positions is arranged in for shooting different regions in multiple cameras 200, and more
The region that every two adjacent camera 200 in a camera 200 is shot connects or partly overlaps.It is understood that every
A camera 200 can correspond to according to the difference of its field angle and setting position and shoot different regions, by the way that every two is arranged
The shooting area of adjacent camera 200 connects or partly overlaps, and the region that distributed system can be made to be shot is covered entirely
Lid.Wherein, multiple cameras 200 can be spaced along a length direction is arranged side by side, for shooting the length direction region
Image, multiple cameras 200 can also be spaced setting along a circumferential direction, for shooting the figure in the annular region
Picture, certainly, multiple cameras 200 can also include other set-up modes, herein not as restriction.
In some embodiments, which is stored with target reference object.Wherein, target reference object can be with
It is to be uploaded to server 100 after electronic equipment receives external input, and deposited target reference object by server 100
Storage.As a kind of mode, which can be personage, animal and other objects etc., wherein when target shooting pair
When as personage, which may include man, woman, children, youth, old man etc., when target reference object is
When animal, which may include cat, dog, rabbit lamp, and when target reference object is other objects, which is clapped
Taking the photograph object may include car, lorry, car etc., it is not limited here.
Referring to Fig. 2, the flow diagram of the image processing method provided Fig. 2 shows the application one embodiment.Institute
Image processing method is stated for carrying out track up to target reference object by distributed camera, and the target is obtained and shoots
The monitor video of object, to promote the monitoring effect of target reference object.In the particular embodiment, described image processing method
Applied to image processing apparatus 300 as shown in fig. 13 that and the server 100 configured with described image processing unit 300 (is schemed
14).It will illustrate the detailed process of the present embodiment by taking server as an example below, it will of course be understood that, the present embodiment is applied
Server can be Cloud Server, or traditional server, it is not limited here.The server and multiple cameras connect
It connects, and different positions is arranged in for shooting different regions, and the every two in multiple camera in multiple camera
The region of adjacent camera shooting connects or partly overlaps, which is stored with target reference object.It will be directed to Fig. 2 institute below
The process shown is explained in detail, and shown image processing method can specifically include following steps:
Step S101: from the image that each camera in the multiple camera takes, extracting includes the mesh
Mark multiple target images of reference object.
As a kind of mode, each camera in multiple cameras can be in open state, and each camera
In real time the shooting area covered shoot and the image taken is uploaded to server.Alternatively,
Each camera in multiple cameras can receive external command, and response external instruction is in the open state or closes shape
State, wherein camera in the open state can carry out shooting to the shooting area covered and will be on the image that taken
Reach server, wherein external command may include the server being connect with multiple cameras send automatically command information, with
Command information that the electronic equipment of multiple cameras connection is sent based on user's operation, user directly trigger life in multiple cameras
At command information etc., it is not limited here.
In the present embodiment, server receives the image of each camera shooting overhead pass in multiple cameras taken, and
The image taken to each camera identifies, from the image that each camera takes, extracting includes that target is shot
Multiple target images of object.In some embodiments, server can read pre-stored target shooting pair from local
As, wherein the target reference object that server is read may include the description file of target reference object, also may include target
The image file etc. of reference object, wherein description file can be used for believing by feature of the text information to target reference object
Breath is described, and image file can be used for being described by characteristic information of the image information to target reference object.
In some embodiments, server is receiving taking for each camera shooting overhead pass in multiple cameras
After image, all images received can be compared with target reference object, include to be extracted from all images
Multiple target images of target reference object, it is to be understood that wrapped in each target image in multiple target image
Including target reference object, wherein the target reference object for including in each target image can be complete target reference object,
It can also be with the part of target reference object, it is not limited here.As a kind of mode, server is in receiving multiple cameras
Each camera shooting overhead pass the image taken after, can by all images received carry out feature information extraction, and will
The characteristic information of the characteristic information and target reference object that extract is matched, and by the characteristic information with target reference object
Image where matched characteristic information is determined as target image.For example, face can be carried out to by all images received
Information extraction, and the face information of the face information extracted and target reference object is matched, and will be shot with target
Image where the matched face information of the face information of object is determined as target image.
Step S102: the shooting time of each target image in the multiple target image is obtained.
As a kind of mode, each camera in multiple cameras can will be shot when uploading the image taken
To image and take the shooting time of the image and be uploaded to server together, correspondingly, server is from the institute received
Have after determining multiple target images in image, the corresponding shooting time of multiple target image can be searched respectively, to obtain
Take the shooting time of each target image in multiple target images.In some embodiments, each of multiple cameras
Camera by the image taken and can take when uploading the image taken and the shooting time for taking the image
The shooting time of the image is associated, for example, the one-to-one mapping relations of image and shooting time are established, correspondingly, clothes
Device be engaged in after determining multiple target images in all images received, it can be based on the shooting time of image and image
Incidence relation searches the corresponding shooting time of multiple target image, to obtain each target image in multiple target images
Shooting time.
Alternatively, server obtains in all images from each camera shooting overhead pass in multiple cameras
To after multiple target images, the corresponding camera of each target image in multiple target images can be obtained respectively, is then sent out
Send command information to the corresponding camera of each target image, which is used to indicate the bat of cam feedback target image
Take the photograph the time, wherein the command information carries the identification information of each target image, correspondingly, the response of each camera receives
The command information arrived, and target image is determined based on the identification information extracted from command information, it is searched in the local of camera
The shooting time of the target image is uploaded to server by the shooting time of the target image, should so that server is available
The shooting time of each target image in multiple target images.
Step S103: by the sequencing of the shooting time of each target image, the multiple target image is suitable
Secondary splicing obtains the monitor video of the target reference object.
In some embodiments, server is obtaining each target in multiple target images and multiple target images
After the shooting time of image, multiple target images can be arranged by the sequencing of the shooting time of each target image
Sequence, it is to be understood that the shooting time of target image of the shooting time for the forward target image that sorts earlier than sequence rearward,
Then multiple target images are spliced by the sequence of multiple target images, to generate monitor video, wherein the monitor video
It include target reference object by each frame of playback progress, so as to promote the monitoring effect of target reference object.
In some embodiments, server can regard the monitoring after the monitor video for obtaining target reference object
Frequency is sent to target terminal corresponding with the target reference object, so that the corresponding user's timely learning target of the target terminal
The case where reference object.For example, the target reference object can be old man or child, the target terminal can correspond to the old man or
The guardian of child, avoids the generation of fortuitous event at can be with the old man or child of timely learning family so as to guardian the case where.
In some embodiments, server, can be automatically to the prison after the monitor video for obtaining target reference object
Control video is analyzed, and whether abnormal conditions occurs with the target reference object judged in the monitor video, when judging result table
When levying the target reference object and abnormal conditions occur, it is whole to target corresponding with the target reference object that warning message can be sent
End, so that the corresponding user of target terminal makes corresponding processing in time.Wherein, abnormal conditions may include falling down, crouching long not
Rise etc., it is not limited here.In addition, server can also send the monitor video when sending warning message to target terminal
Or the corresponding video clip of the abnormal conditions is shot to target terminal so that the corresponding user of target terminal understands target in time
The truth of object.
The image processing method that the application one embodiment provides, takes from each camera in multiple cameras
In image, multiple target images including target reference object are extracted, each target image in multiple target images is obtained
Shooting time is sequentially spliced multiple target images by the sequencing of the shooting time of each target image, obtains the target
The monitor video of reference object to carry out track up to target reference object by distributed camera, and obtains the mesh
The monitor video of reference object is marked, to promote the monitoring effect of target reference object.
Referring to Fig. 3, Fig. 3 shows the flow diagram of the image processing method of another embodiment of the application offer.
This method is applied to above-mentioned server, which connect with multiple cameras, and different positions is arranged in multiple camera
It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple camera connects or part weight
Folded, which is stored with target reference object.Wherein, target reference object includes first object reference object and the second target
Reference object will be explained in detail for process shown in Fig. 3 below, shown image processing method can specifically include with
Lower step:
Step S201: from the image that each camera in the multiple camera takes, extracting respectively includes institute
State multiple first object images of first object reference object and multiple second targets including the second target reference object
Image.
Step S202: obtaining the shooting time of each first object image in the multiple first object image respectively,
With the shooting time of the second target image of each of the multiple second target image.
Step S203: by the sequencing of the shooting time of each first object image, by the multiple first mesh
Logo image is sequentially spliced, and obtains the monitor video of the first object reference object, and press each second target image
Shooting time sequencing, the multiple second target image is sequentially spliced, the second target reference object is obtained
Monitor video.
In some embodiments, which can be used for carrying out tracking bat to multiple target reference objects respectively
It takes the photograph, and generates the monitor video of each target reference object in multiple target reference objects respectively.As shown in figure 4, Fig. 4 is shown
Distributed system provided by the embodiments of the present application carries out the schematic diagram of track up to multiple target reference objects, wherein point
Cloth system can be grouped each target reference object of multiple target reference objects, for example, multiple cameras are clapped
What is taken the photograph includes that first object reference object (target 1) is added to a grouping, by multiple cameras take including second
Target reference object (target 2) is added to another grouping, by multiple cameras take include third target reference object
(target 3) is added to another and is grouped, and details are not described herein.
It in the present embodiment, include first object reference object and the second target reference object with multiple target reference objects
For be illustrated, track up can be carried out to first object reference object and the second target reference object respectively, and to the
One target reference object and the second target reference object are grouped, respectively generate first object reference object monitor video and
The monitor video of second target reference object.For example, the first object reference object is old man, which is
Child, then can carry out track up to old man and child respectively, and the image including old man is added to a grouping, will include
The image of child is added to a grouping, and generates the monitor video of old man and the monitor video of child respectively.
Wherein, the specific descriptions of step S201- step S203 please refer to step S101- step S103, and details are not described herein.
The image processing method that another embodiment of the application provides, takes from each camera in multiple cameras
Image in, respectively extract include first object reference object multiple first object images and including the second target reference object
Multiple second target images, obtain the shooting time of each first object image in multiple first object images respectively, and
The shooting time of the second target image of each of multiple second target images, by the shooting time of each first object image
Sequencing sequentially splices multiple first object images, obtains the monitor video and each the of first object reference object
The sequencing of the shooting time of two target images sequentially splices multiple second target images, obtains the shooting pair of the second target
The monitor video of elephant.Compared to image processing method shown in Fig. 2, the present embodiment can also be right respectively by distributed camera
Multiple target reference objects carry out track up, and obtain the monitor video of multiple target reference object respectively, more to be promoted
The monitoring effect of a target reference object.
Referring to Fig. 5, Fig. 5 shows the flow diagram of the image processing method of the application further embodiment offer.
This method is applied to above-mentioned server, which connect with multiple cameras, and different positions is arranged in multiple camera
It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple camera connects or part weight
Folded, which is stored with target reference object.The multiple camera includes the first camera and images with described first
At least one second camera that head is disposed adjacent will be explained in detail, shown image for process shown in fig. 5 below
Processing method can specifically include following steps:
Step S301: when the target reference object is located in the shooting area of first camera, described in control
First camera is in the open state and at least one described second camera of control is in close state.
In the present embodiment, multiple camera include the first camera and be disposed adjacent with first camera to
A few second camera.Wherein, the setting position of first camera and at least one second camera can be adjacent, and the
The region of one camera and the shooting of at least one second camera connects or partly overlaps.As shown in fig. 6, Fig. 6 shows this Shen
Please embodiment provide multiple cameras a kind of setting position view, in Fig. 6, multiple camera 200 include first
Camera 200A and two second camera 200B being disposed adjacent with the first camera, it is possible to understand that, this two second
Camera 200B is disposed adjacent with the first camera 200A respectively, and one of second camera 200B is set to the first camera shooting
The left side of head, another second camera 200B are set to the right side of the first camera.As shown in fig. 7, Fig. 7 shows the application
Another setting position view in multiple cameras that embodiment provides, in Fig. 7, multiple camera 200 includes the
An one camera 200A and second camera 200B being disposed adjacent with the first camera, it is possible to understand that, this first is taken the photograph
As the marginal position of distributed system is arranged in head 200A, and the side of first camera 200A is disposed adjacent one second and takes the photograph
As head 200B.
In some embodiments, multiple cameras can be controlled in advance and be in open state, pass through multiple cameras
In each camera image taking is carried out to its corresponding shooting area, whether and identifying in the image taken includes target
Reference object, wherein when recognition result, which characterizes the target object, to be located in the shooting area of the first camera, that is to say, that
When recognizing target reference object from the image that the first camera takes, and other cameras are shot from multiple cameras
To image in do not recognize target reference object when, can control the first camera it is in the open state and control at least one
A second camera is in close state, i.e., when the quantity of second camera is one, controls at a second camera
It controls two second cameras when the quantity of second camera is two in closed state and is in close state, to reduce
The power consumption of at least one second camera.
Step S302: the target reference object is shot by first camera, and monitors the target
The behavioral data of reference object.
It should be understood that the zone of action of the target reference object is the corresponding shooting area of the first camera at this time, because
This, can shoot target reference object by the first camera, and monitor the behavioral data of the target reference object.Make
For a kind of mode, the behavioral data of target object may include movement speed, moving direction, behavior act etc., not limit herein
It is fixed.
Step S303: when the behavioral data characterizes the shooting that the target reference object will leave first camera
When region enters target area, target second corresponding with the target area is determined from least one described second camera
Camera.
Wherein, the shooting area of the first camera will be left when the behavioral data that monitoring obtains characterizes the target reference object
When domain enters target area, target second camera can be determined from least one second camera based on the target area.
In some embodiments, when behavior data characterization target reference object moves towards target area from the shooting area of the first camera
When domain, it is believed that the shooting area that the target reference object will leave the first camera enters target area;Work as the behavior
When data characterization target reference object is located at the marginal position object-oriented region of the shooting area of the first camera, it is believed that
The shooting area that the target reference object will leave the first camera enters target area etc., and details are not described herein.
Referring to Fig. 8, Fig. 8 shows the process signal of the step S303 of the image processing method shown in fig. 5 of the application
Figure.It will be explained in detail below for process shown in Fig. 8, the method can specifically include following steps:
Step S3031: the moving direction of the target reference object is obtained based on the behavioral data.
In some embodiments, the movement of the target reference object can be obtained from the behavioral data that monitoring obtains
Direction.It is understood that the target can be obtained according to behavioral datas such as walking postures, the direction of travel of target reference object
The moving direction of reference object, details are not described herein.
Step S3032: determining the target area based on the moving direction, from least one described second camera
Determine target second camera corresponding with the target area.
For example, as shown in fig. 6, when target reference object is located in the shooting area of the first camera 200A to walk to left lateral
When, it is believed that the shooting area that the target reference object will leave the first camera 200A, which enters, is set to the first camera
The left side of 200A target shooting area corresponding with the second camera 200B that the first camera 200A is disposed adjacent, can be by position
It is taken the photograph in the second camera 200B that the left side of the first camera 200A is disposed adjacent with the first camera 200A as target second
As head.
In another example as shown in fig. 6, when target reference object is located in the shooting area of the first camera 200A to walk to right lateral
When, it is believed that the shooting area that the target reference object will leave the first camera 200A, which enters, is set to the first camera
The right side of 200A target shooting area corresponding with the second camera 200B that the first camera 200A is disposed adjacent, can be by position
It is taken the photograph in the second camera 200B that the right side of the first camera 200A is disposed adjacent with the first camera 200A as target second
As head.
Step S304: the target second camera is controlled by closed state and is switched to open state.
In the present embodiment, after determining target second camera at least one second camera, it can control the mesh
It marks second camera and open state is switched to by closed state, for the shooting area that will enter the target second camera
Target reference object carry out track up.
Step S305: when the target reference object enters the target area, pass through the target second camera
The target reference object is shot, and monitors the behavioral data of the target reference object.
Wherein, after the shooting area that target reference object leaves the first camera enters target area, it can control this
First camera is in off state by open state switching, also be can control first camera and is continued to keep it turned on,
This is without limitation.
In some embodiments, when the shooting area that target reference object leaves the first camera enters target area
When, the zone of action of the target reference object is therefore the corresponding shooting area of target second camera can pass through mesh at this time
Mark second camera shoots target reference object, and monitors the behavioral data of the target reference object.
Step S306: extracting the target signature information of the target reference object, and the target signature information includes biology
At least one of characteristic information and behavior characteristic information.
In some embodiments, server can extract target shooting pair from the target reference object being locally stored
The target signature information of elephant, the target signature information include generating at least one of characteristic information and behavior characteristic information,
In, generating characteristic information may include face characteristic information, physical characteristic information, height characteristic information etc., behavior characteristic information
It may include walking postures, the speed of travel, behavior act etc., it is not limited here.
Step S307: it from the image that each camera in the multiple camera takes, extracts and the target
The matched characteristic information of characteristic information.
In some embodiments, after the target signature information for extracting target reference object, the target can be based on
Characteristic information extracts and the matched characteristic information of the target signature information from the image of each camera shooting overhead pass.For example, when should
When target signature information is the face information of target reference object, it can be extracted from all images of each camera shooting overhead pass
With the matched face information of face information of the target reference object.In another example when the target signature information is target shooting pair
When the walking postures of elephant, the walking appearance with the target reference object can be extracted from all images of each camera shooting overhead pass
The matched walking postures of gesture.
Step S308: object corresponding with the matched characteristic information of the target signature information is determined as the target and is clapped
Object is taken the photograph, and extracts multiple target images including the target reference object.
In some embodiments, it is extracted in the image from each camera shooting overhead pass matched with target signature information
After characteristic information, object corresponding with the matched characteristic information of target signature information can be determined in the picture, it can will be with mesh
The corresponding object of the mark matched characteristic information of characteristic information is determined as target reference object, and will include the target reference object
Image be determined as target image, therefore, can from it is each camera shooting overhead pass all images in extract including the target clap
Take the photograph multiple target images of object.
Step S309: the shooting time of each target image in the multiple target image is obtained.
Step S310: by the sequencing of the shooting time of each target image, the multiple target image is suitable
Secondary splicing obtains the monitor video of the target reference object.
Wherein, the specific descriptions of step S309- step S310 please refer to step S104- step S105, and details are not described herein.
The image processing method that the application further embodiment provides, when target reference object is located at the bat of the first camera
When taking the photograph in region, the first camera of control is in the open state and controls at least one second camera and is in close state, and leads to
It crosses the first camera to shoot target reference object, and monitors the behavioral data of the target reference object, when behavior number
When the shooting area for leaving the first camera being entered target area according to characterization target reference object, from least one the second camera shooting
Target second camera corresponding with target area is determined in head, is controlled target second camera by closed state and is switched to unlatching
State shoots target reference object by target second camera when target reference object enters target area, and
Monitor the behavioral data of the target reference object.Extract the target signature information of target reference object, the target signature information packet
At least one of biological information and behavior characteristic information are included, the figure taken from each camera in multiple cameras
As in, extraction and the matched characteristic information of target signature information will be corresponding with the matched characteristic information of target signature information right
As being determined as target reference object, and multiple target images including target reference object are extracted, obtained in multiple target images
Each target image shooting time, it is by the sequencing of the shooting time of each target image, multiple target images are suitable
Secondary splicing obtains the monitor video of target reference object.Compared to image processing method shown in Fig. 2, the present embodiment also according to
Region where target reference object controls corresponding camera and opens or closes, to reduce the power consumption of camera.In addition, this reality
The target signature information that example is applied also according to target reference object carries out recognition and tracking to target reference object, to promote target shooting
The recognition success rate of object.
Referring to Fig. 9, Fig. 9 shows the flow diagram of the image processing method of another embodiment of the application offer.
This method is applied to above-mentioned server, which connect with multiple cameras, and different positions is arranged in multiple camera
It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple camera connects or part weight
Folded, which is stored with target reference object.It will be explained in detail below for process shown in Fig. 9, shown image
Processing method can specifically include following steps:
Step S401: the historical behavior data of the target reference object are obtained.
In some embodiments, server, can be to image when receiving the image of each camera shooting overhead pass every time
In object behavioral data carry out analysis record and store, history of forming behavioral data.In the present embodiment, server can be from
Locally read the historical behavior data of the target reference object, wherein historical behavior data may include historical act region, go through
History activity time etc., it is not limited here.
Step S402: the historical behavior data are based on, shooting control is carried out to the multiple camera.
In some embodiments, server can be based on after the historical behavior data for obtaining the target reference object
The historical behavior data of target reference object carry out shooting control to multiple cameras.For example, target reference object can be based on
Historical behavior data control that multiple cameras are in open state, the multiple cameras of control are in closed state, control
A part of camera in multiple cameras is in the open state and another part camera is in close state, and does not do herein
It limits.
Referring to Fig. 10, Figure 10 shows a kind of stream of the step S402 of the image processing method shown in Fig. 9 of the application
Journey schematic diagram.It will be explained in detail below for process shown in Fig. 10, the method can specifically include following steps:
Step S4021A: historical act region is extracted from the historical behavior data, wherein the historical act region
It is greater than the region of preset duration for duration where characterizing the target reference object.
As an implementation, which may include historical act region, wherein the historical act area
Domain is greater than the region of preset duration for the duration where characterizing the target reference object.Specifically, which can be preparatory
Preset duration is obtained and is stored with, the preset duration is for acting on target reference object in the judgement of the duration where each region
Foundation, therefore, in this present embodiment, the available target reference object of server is in the duration where each region, by this
Target reference object where each region duration and preset duration be compared, wherein when comparison result characterizes the target
Reference object is when the duration where some region is greater than preset duration, it is believed that the region is the history of target reference object
It zone of action can when comparison result characterizes the target reference object when the duration where some region is no more than preset duration
It is the historical act region of target reference object to think the region not.
For example, when the distributed system be applied to family when, home area respectively include parlor, bedroom, toilet and
Four, kitchen region, can obtain duration of the target reference object in parlor, the duration in bedroom, the duration in toilet respectively
And the duration in kitchen, and by judge respectively the duration in parlor, the duration in bedroom, toilet duration and
Whether the duration in kitchen is greater than the mode of preset duration, determines to go through from parlor, bedroom, toilet and four, kitchen region
History zone of action.Wherein, the duration can be total duration or the average duration as unit of one day, it is not limited here.
Step S4022A: multiple target camera shootings corresponding with the historical act region are searched from the multiple camera
Head.
It in the present embodiment, can be based on the historical act regional search and the history after obtaining historical act region
The corresponding multiple target cameras in zone of action.In some embodiments, the covering model in the available historical act region
The shooting for enclosing, and searching the shooting area of the included camera of the coverage area, and the camera that server is included being covered
The corresponding multiple cameras in region are determined as multiple target cameras.
Step S4023A: it is in the open state to control the multiple target camera, and controls in the multiple camera
Other cameras in addition to the multiple target camera are in close state.
In some embodiments, the duration due to historical act area attribute target reference object in the region is greater than pre-
If duration, that is to say, that therefore the main activities region of target reference object, can correspond in the historical act region
The camera in historical act region is opened, and closes the camera outside historical act region, to realize to target reference object
Track up while, moreover it is possible to reduce the power consumption of distributed system.Correspondingly, in the present embodiment, can control and history
The corresponding multiple target cameras in zone of action are in the open state, and control in multiple cameras except multiple target camera
Except other cameras be in close state.
Figure 11 is please referred to, Figure 11 shows another of the step S402 of the image processing method shown in Fig. 9 of the application
Flow diagram.It will be explained in detail below for process shown in Figure 11, the method can specifically include following step
It is rapid:
Step S4021B: the historical act time is extracted from the historical behavior data, wherein the historical act time
The time being located at for characterizing the target reference object in the shooting area of the multiple camera.
As an implementation, which may include the historical act time, wherein when the historical act
Between for characterizing the time that the target reference object is located in the shooting area of multiple cameras.For example, multiple camera
Shooting area is interior, then, which is located at the indoor time for characterizing the target reference object, that is,
It says, the historical act time is for characterizing the time that the target reference object is in.
Step S4022B: the multiple camera is controlled in the historical act time corresponding duration in opening state
State, and the multiple camera of control are in close state outside the historical act time corresponding duration.
In some embodiments, since historical act time representation target reference object is in the shooting of multiple camera
Time in region, that is to say, that within the time, target reference object is located in the shooting area of multiple camera,
Outside the time, target reference object is located at outside the shooting area of multiple camera.Therefore, in the present embodiment, can control
Multiple cameras are in the open state in historical act time corresponding duration, so that the tracking to target reference object is clapped
It takes the photograph, and the multiple cameras of control are in close state outside historical act time corresponding duration, to reduce distributed system
The power consumption of system.
Step S403: extracting the target signature information of the target reference object, and the target signature information includes biology
At least one of characteristic information and behavior characteristic information.
Step S404: it from the image that each camera in the multiple camera takes, extracts and the target
The matched characteristic information of characteristic information.
Step S405: object corresponding with the matched characteristic information of the target signature information is determined as the target and is clapped
Object is taken the photograph, and extracts multiple target images including the target reference object.
Step S406: the shooting time of each target image in the multiple target image is obtained.
Step S407: by the sequencing of the shooting time of each target image, the multiple target image is suitable
Secondary splicing obtains the monitor video of the target reference object.
Wherein, the specific descriptions of step S403- step S407 please refer to step S306- step S310, and details are not described herein.
The image processing method that another embodiment of the application provides obtains the historical behavior data of target reference object,
Based on historical behavior data, shooting control is carried out to multiple cameras.Extract the target signature information of target reference object, the mesh
Marking characteristic information includes at least one of biological information and behavior characteristic information, from each camera shooting in multiple cameras
In the image that head takes, extraction and the matched characteristic information of target signature information will be with the matched features of target signature information
The corresponding object of information is determined as target reference object, and extracts multiple target images including target reference object, obtains more
The shooting time of each target image in a target image will be more by the sequencing of the shooting time of each target image
A target image sequentially splices, and obtains the monitor video of target reference object.Compared to image processing method shown in Fig. 2, originally
Embodiment controls corresponding camera also according to the historical behavior data of target reference object and opens or closes, to reduce camera
Power consumption.In addition, the present embodiment also according to target reference object target signature information to target reference object identified with
Track, to promote the recognition success rate of target reference object.
Figure 12 is please referred to, Figure 12 shows the timing diagram of the application image processing method that further embodiment provides again.
This method is applied to above-mentioned distributed system, multiple to take the photograph including server and the multiple cameras connecting with the server
It is used to shoot different regions as different positions is arranged in head, and the every two adjacent camera shooting in multiple cameras
Region connect or partly overlap, each camera in multiple cameras is stored with target reference object.It will be directed to below
Process shown in Figure 12 is explained in detail, and shown image processing method can specifically include following steps:
Step S501: for each camera in the multiple camera from the image taken, extracting includes the mesh
Mark multiple target images of reference object.
Step S502: when the multiple camera obtains the shooting of each target image in the multiple target image
Between.
Step S503: each camera is by each mesh in the multiple target image and the multiple target image
The shooting time of logo image is uploaded to the server.
Step S504: the server, will be the multiple by the sequencing of the shooting time of each target image
Target image sequentially splices, and obtains the monitor video of the target reference object.
In some embodiments, each camera in multiple camera is stored with target reference object.Wherein, mesh
Mark reference object, which can be electronic equipment and receive, is sent to each camera after external input, and by each camera by target
Reference object is stored locally.In the present embodiment, camera is according to the target reference object being locally stored directly to bat
The multiple images for taking the photograph acquisition are judged, and the target figure including target reference object is obtained from the multiple images that shooting obtains
Target image is uploaded to server by picture, to reduce the data volume of the image uploaded, transfer efficiency in promotion.
Wherein, the specific descriptions of step S501- step S504 please refer to step S101- step S103, and details are not described herein.
The application image processing method that further embodiment provides again, each camera in multiple cameras is from shooting
To image in, extract include target reference object multiple target images, multiple cameras obtain in multiple target images
The shooting time of each target image, each camera is by each target image in multiple target images and multiple target images
Shooting time be uploaded to server, server by the shooting time of each target image sequencing, by multiple target figures
As sequentially splicing, the monitor video of target reference object is obtained, to carry out by distributed camera to target reference object
Track up, and the monitor video of the target reference object is obtained, to promote the monitoring effect of target reference object.
Figure 13 is please referred to, Figure 13 shows the module frame chart of image processing apparatus 300 provided by the embodiments of the present application.The figure
Picture processing unit 300 is applied to above-mentioned server, and server is connect with multiple cameras, and different positions is arranged in multiple cameras
It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple cameras connects or part weight
Folded, server is stored with target reference object.It will be illustrated below for block diagram shown in Figure 13, shown image processing apparatus
300 include: image zooming-out module 310, shooting time acquisition module 320 and image mosaic module 330, in which:
Image zooming-out module 310, for extracting from the image that each camera in the multiple camera takes
Multiple target images including the target reference object.Further, described image extraction module 310 includes: target signature
Information extraction submodule, feature information extraction submodule and image zooming-out submodule, in which:
Target signature information extracting sub-module, for extracting the target signature information of the target reference object, the mesh
Marking characteristic information includes at least one of biological information and behavior characteristic information.
Feature information extraction submodule, for from the image that each camera in the multiple camera takes,
It extracts and the matched characteristic information of the target signature information.
Image zooming-out submodule, for object corresponding with the matched characteristic information of the target signature information to be determined as
The target reference object, and extract multiple target images including the target reference object.
Shooting time obtains module 320, when for obtaining the shooting of each target image in the multiple target image
Between.
Image mosaic module 330 will be the multiple for the sequencing of the shooting time by each target image
Target image sequentially splices, and obtains the monitor video of the target reference object.
Further, described image processing unit 300 further include: first state control module, the first behavior data monitoring
Module, target camera determining module, the second status control module and the second behavioral data monitoring modular, in which:
First state control module, for being located in the shooting area of first camera when the target reference object
When, control first camera is in the open state and at least one described second camera of control is in close state.
First behavior data monitoring module, for being clapped by first camera the target reference object
It takes the photograph, and monitors the behavioral data of the target reference object.
Target camera determining module, for described the will to be left when the behavioral data characterizes the target reference object
When the shooting area of one camera enters target area, the determining and target area from least one described second camera
Corresponding target second camera.Further, the target camera determining module include: moving direction acquisition submodule and
Target camera determines submodule, in which:
Moving direction acquisition submodule, for obtaining the mobile side of the target reference object based on the behavioral data
To.
Target camera determines submodule, for determining the target area based on the moving direction, from it is described at least
Target second camera corresponding with the target area is determined in one second camera.
Second status control module is switched to open state by closed state for controlling the target second camera.
Second behavioral data monitoring modular, for passing through institute when the target reference object enters the target area
It states target second camera to shoot the target reference object, and monitors the behavioral data of the target reference object.
Further, described image processing unit 300 further include: behavioral data obtains module and shooting control module,
In:
Behavioral data obtains module, for obtaining the historical behavior data of the target reference object.
Control module is shot, for being based on the historical behavior data, shooting control is carried out to the multiple camera.Into
One step, the shooting control module include: that historical act extracted region submodule, target camera search submodule and the
Three condition control submodule, in which:
Historical act extracted region submodule, for extracting historical act region from the historical behavior data, wherein
Duration where the historical act region is used to characterize the target reference object is greater than the region of preset duration.
Target camera searches submodule, corresponding with the historical act region for searching from the multiple camera
Multiple target cameras.
Third state control submodule, it is in the open state for controlling the multiple target camera, and described in control
Other cameras in multiple cameras in addition to the multiple target camera are in close state.
Further, described, the shooting control module further include: historical act time extracting sub-module and the 4th state
Control submodule, in which:
Historical act time extracting sub-module, for extracting the historical act time from the historical behavior data, wherein
The historical act time is for characterizing the time that the target reference object is located in the shooting area of the multiple camera.
4th mode control word module, for controlling the multiple camera in the historical act time corresponding duration
It is interior in the open state, and control the multiple camera and be in closing shape outside the historical act time corresponding duration
State.
It is apparent to those skilled in the art that for convenience and simplicity of description, foregoing description device and
The specific work process of module, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In several embodiments provided herein, the mutual coupling of module can be electrical property, mechanical or other
The coupling of form.
It, can also be in addition, can integrate in a processing module in each functional module in each embodiment of the application
It is that modules physically exist alone, can also be integrated in two or more modules in a module.Above-mentioned integrated mould
Block both can take the form of hardware realization, can also be realized in the form of software function module.
Figure 14 is please referred to, it illustrates a kind of structural block diagrams of server 100 provided by the embodiments of the present application.The server
100 can be Cloud Server, be also possible to traditional server.Server 100 in the application may include it is one or more such as
Lower component: processor 110, memory 120 and one or more application program, wherein one or more application programs can be by
It is stored in memory 120 and is configured as being executed by one or more processors 110, one or more programs are configured to hold
Row method as described in preceding method embodiment.
Wherein, processor 110 may include one or more processing core.Processor 110 utilizes various interfaces and route
The various pieces in entire server 100 are connected, by running or executing the instruction being stored in memory 120, program, code
Collection or instruction set, and the data being stored in memory 120 are called, the various functions and processing data of execute server 100.
Optionally, processor 110 can be compiled using Digital Signal Processing (Digital Signal Processing, DSP), scene
Journey gate array (Field-Programmable Gate Array, FPGA), programmable logic array (Programmable
Logic Array, PLA) at least one of example, in hardware realize.Processor 110 can integrating central processor (Central
Processing Unit, CPU), in graphics processor (Graphics Processing Unit, GPU) and modem etc.
One or more of combinations.Wherein, the main processing operation system of CPU, user interface and application program etc.;GPU is for being responsible for
Show the rendering and drafting of content;Modem is for handling wireless communication.It is understood that above-mentioned modem
It can not be integrated into processor 110, be realized separately through one piece of communication chip.
Memory 120 may include random access memory (Random Access Memory, RAM), also may include read-only
Memory (Read-Only Memory).Memory 120 can be used for store instruction, program, code, code set or instruction set.It deposits
Reservoir 120 may include storing program area and storage data area, wherein the finger that storing program area can store for realizing operating system
Enable, for realizing at least one function instruction (such as touch function, sound-playing function, image player function etc.), be used for
Realize the instruction etc. of following each embodiments of the method.Storage data area can also store the number that terminal 100 is created in use
According to (such as phone directory, audio, video data, chat record data) etc..
Figure 15 is please referred to, it illustrates a kind of structural frames of computer readable storage medium provided by the embodiments of the present application
Figure.Program code is stored in the computer-readable medium 400, said program code can be called by processor and execute the above method
Method described in embodiment.
Computer readable storage medium 400 can be such as flash memory, EEPROM (electrically erasable programmable read-only memory),
The electronic memory of EPROM, hard disk or ROM etc.Optionally, computer readable storage medium 400 includes non-volatile meter
Calculation machine readable medium (non-transitory computer-readable storage medium).Computer-readable storage
Medium 400 has the memory space for the program code 410 for executing any method and step in the above method.These program codes can
With from reading or be written in one or more computer program product in this one or more computer program product.
Program code 410 can for example be compressed in a suitable form.
In conclusion image processing method provided by the embodiments of the present application, device, server and storage medium, application
In server, server is connect with multiple cameras, and different positions is arranged in for shooting different areas in multiple camera
Domain, and the region of the every two adjacent camera shooting in multiple camera connects or partly overlaps, which is stored with
Target reference object.From the image that each camera in multiple cameras takes, extracting includes target reference object
Multiple target images obtain the shooting time of each target image in multiple target images, by the shooting of each target image
The sequencing of time sequentially splices multiple target images, obtains the monitor video of the target reference object, thus by dividing
Cloth camera carries out track up to target reference object, and obtains the monitor video of the target reference object, to promote mesh
Mark the monitoring effect of reference object.
Finally, it should be noted that above embodiments are only to illustrate the technical solution of the application, rather than its limitations;Although
The application is described in detail with reference to the foregoing embodiments, those skilled in the art are when understanding: it still can be with
It modifies the technical solutions described in the foregoing embodiments or equivalent replacement of some of the technical features;And
These are modified or replaceed, do not drive corresponding technical solution essence be detached from each embodiment technical solution of the application spirit and
Range.
Claims (12)
1. a kind of image processing method, which is characterized in that be applied to server, the server is connect with multiple cameras, institute
It states multiple cameras and different positions is set for shooting different regions, and the every two in the multiple camera is adjacent
Camera shooting region connect or partly overlap, the server is stored with target reference object, which comprises
From the image that each camera in the multiple camera takes, extracting includes the more of the target reference object
A target image;
Obtain the shooting time of each target image in the multiple target image;
By the sequencing of the shooting time of each target image, the multiple target image is sequentially spliced, obtains institute
State the monitor video of target reference object.
2. the method according to claim 1, wherein each camera from the multiple camera is clapped
In the image taken the photograph, multiple target images including the target reference object are extracted, comprising:
The target signature information of the target reference object is extracted, the target signature information includes biological information and behavior
At least one of characteristic information;
From the image that each camera in the multiple camera takes, extract matched with the target signature information
Characteristic information;
Object corresponding with the matched characteristic information of the target signature information is determined as the target reference object, and is extracted
Multiple target images including the target reference object.
3. the method according to claim 1, wherein the target reference object includes first object reference object
In the image taken with the second target reference object, each camera from the multiple camera, extraction includes
Multiple target images of the target reference object, comprising:
From the image that each camera in the multiple camera takes, extracting respectively includes that the first object is shot
Multiple first object images of object and multiple second target images including the second target reference object;
The shooting time for obtaining each target image in the multiple target image, comprising:
The shooting time and the multiple second of each first object image in the multiple first object image is obtained respectively
The shooting time of the second target image of each of target image;
The sequencing of the shooting time by each target image, the multiple target image is sequentially spliced, is obtained
Obtain the monitor video of the target reference object, comprising:
By the sequencing of the shooting time of each first object image, the multiple first object image is sequentially spelled
It connects, obtains the monitor video of the first object reference object, and by the shooting time of each second target image
Sequencing sequentially splices the multiple second target image, obtains the monitor video of the second target reference object.
4. method according to claim 1-3, which is characterized in that the multiple camera includes the first camera
And at least one second camera being disposed adjacent with first camera, it is described from each of the multiple camera
In the image that camera takes, before extraction includes multiple target images of the target reference object, further includes:
When the target reference object is located in the shooting area of first camera, controls first camera and be in
At least one second camera described in open state and control is in close state;
The target reference object is shot by first camera, and monitors the behavior of the target reference object
Data;
The shooting area for leaving first camera is entered into target when the behavioral data characterizes the target reference object
When region, target second camera corresponding with the target area is determined from least one described second camera;
It controls the target second camera and open state is switched to by closed state;
When the target reference object enters the target area, the target is shot by the target second camera
Object is shot, and monitors the behavioral data of the target reference object.
5. according to the method described in claim 4, it is characterized in that, described when the behavioral data characterizes the target shooting pair
When as the shooting area for leaving first camera being entered target area, determined from least one described second camera
Target second camera corresponding with the target area, comprising:
The moving direction of the target reference object is obtained based on the behavioral data;
The target area is determined based on the moving direction, the determining and target from least one described second camera
The corresponding target second camera in region.
6. method according to claim 1-3, which is characterized in that described from each of the multiple camera
In the image that camera takes, before extraction includes multiple target images of the target reference object, further includes:
Obtain the historical behavior data of the target reference object;
Based on the historical behavior data, shooting control is carried out to the multiple camera.
7. according to the method described in claim 6, it is characterized in that, described be based on the historical behavior data, to the multiple
Camera carries out shooting control, comprising:
Historical act region is extracted from the historical behavior data, wherein the historical act region is for characterizing the mesh
Duration is greater than the region of preset duration where marking reference object;
Multiple target cameras corresponding with the historical act region are searched from the multiple camera;
It is in the open state to control the multiple target camera, and controls in the multiple camera except the multiple target is taken the photograph
As other cameras except head are in close state.
8. according to the method described in claim 6, it is characterized in that, described be based on the historical behavior data, to the multiple
Camera carries out shooting control, comprising:
The historical act time is extracted from the historical behavior data, wherein the historical act time is for characterizing the mesh
Mark reference object is located at the time in the shooting area of the multiple camera;
It is in the open state in the historical act time corresponding duration to control the multiple camera, and described in control
Multiple cameras are in close state outside the historical act time corresponding duration.
9. a kind of image processing method, which is characterized in that be applied to distributed system, including server and with the server
Different positions is arranged in for shooting different regions in multiple cameras of connection, the multiple camera, and the multiple
The region of every two adjacent camera shooting in camera connects or partly overlaps, and each of the multiple camera is taken the photograph
As head is stored with target reference object, which comprises
For each camera in the multiple camera from the image taken, extracting includes the more of the target reference object
A target image;
The multiple camera obtains the shooting time of each target image in the multiple target image;
Each camera is by the shooting of each target image in the multiple target image and the multiple target image
Time is uploaded to the server;
The server is sequentially spelled the multiple target image by the sequencing of the shooting time of each target image
It connects, obtains the monitor video of the target reference object.
10. a kind of image processing apparatus, which is characterized in that be applied to server, the server is connect with multiple cameras, institute
It states multiple cameras and different positions is set for shooting different regions, and the every two in the multiple camera is adjacent
Camera shooting region connect or partly overlap, the server is stored with target reference object, and described device includes:
Image zooming-out module includes institute for extracting from the image that each camera in the multiple camera takes
State multiple target images of target reference object;
Shooting time obtains module, for obtaining the shooting time of each target image in the multiple target image;
Image mosaic module, for the sequencing of the shooting time by each target image, by the multiple target figure
As sequentially splicing, the monitor video of the target reference object is obtained.
11. a kind of server, which is characterized in that including memory and processor, the memory is couple to the processor, institute
Memory store instruction is stated, the processor is executed as claim 1-8 is any when executed by the processor
Method described in.
12. a kind of computer-readable storage medium, which is characterized in that be stored with journey in the computer-readable storage medium
Sequence code, said program code can be called by processor and execute the method according to claim 1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910579248.0A CN110278413A (en) | 2019-06-28 | 2019-06-28 | Image processing method, device, server and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910579248.0A CN110278413A (en) | 2019-06-28 | 2019-06-28 | Image processing method, device, server and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110278413A true CN110278413A (en) | 2019-09-24 |
Family
ID=67962588
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910579248.0A Pending CN110278413A (en) | 2019-06-28 | 2019-06-28 | Image processing method, device, server and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110278413A (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111405203A (en) * | 2020-03-30 | 2020-07-10 | 杭州海康威视数字技术股份有限公司 | Method and device for determining picture switching, electronic equipment and storage medium |
CN112601021A (en) * | 2020-12-14 | 2021-04-02 | 中标慧安信息技术股份有限公司 | Method and system for processing monitoring video of network camera |
CN112788229A (en) * | 2019-11-11 | 2021-05-11 | 株式会社数据价值 | Indoor self-shooting support camera system based on Internet of things |
CN112887666A (en) * | 2021-01-06 | 2021-06-01 | 浙江大华技术股份有限公司 | Video processing method and device, network camera, server and storage medium |
CN113114950A (en) * | 2021-05-28 | 2021-07-13 | 苏州智奇胜软件科技有限公司 | IoT camera control method and control system |
CN113438450A (en) * | 2021-06-11 | 2021-09-24 | 深圳市大工创新技术有限公司 | Dynamic target tracking monitoring method, monitoring system, electronic device and storage medium |
CN114697501A (en) * | 2022-03-23 | 2022-07-01 | 南京云创大数据科技股份有限公司 | Monitoring camera image processing method and system based on time |
CN116582653A (en) * | 2023-07-14 | 2023-08-11 | 广东天亿马信息产业股份有限公司 | Intelligent video monitoring method and system based on multi-camera data fusion |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101105892A (en) * | 2007-07-30 | 2008-01-16 | 深圳市融合视讯科技有限公司 | Vehicle traffic accident automatic detection method |
US20090052739A1 (en) * | 2007-08-23 | 2009-02-26 | Hitachi Kokusai Electric Inc. | Human pursuit system, human pursuit apparatus and human pursuit program |
CN104660998A (en) * | 2015-02-16 | 2015-05-27 | 苏州阔地网络科技有限公司 | Relay tracking method and system |
CN104680555A (en) * | 2015-02-13 | 2015-06-03 | 电子科技大学 | Border-crossing detection method and border-crossing monitoring system based on video monitoring |
CN104902223A (en) * | 2015-04-17 | 2015-09-09 | 国家电网公司 | Fault monitoring system of power line |
CN105245850A (en) * | 2015-10-27 | 2016-01-13 | 太原市公安局 | Method, device and system for tracking target across surveillance cameras |
CN105530465A (en) * | 2014-10-22 | 2016-04-27 | 北京航天长峰科技工业集团有限公司 | Security surveillance video searching and locating method |
CN105788166A (en) * | 2016-01-29 | 2016-07-20 | 维沃移动通信有限公司 | Monitoring and alarm method and apparatus |
CN108234961A (en) * | 2018-02-13 | 2018-06-29 | 欧阳昌君 | A kind of multichannel video camera coding and video flowing drainage method and system |
CN108540754A (en) * | 2017-03-01 | 2018-09-14 | 中国电信股份有限公司 | Methods, devices and systems for more video-splicings in video monitoring |
CN109729287A (en) * | 2018-12-06 | 2019-05-07 | 浙江大华技术股份有限公司 | A kind of method, apparatus and calculating equipment, storage medium of perimeter region monitoring |
-
2019
- 2019-06-28 CN CN201910579248.0A patent/CN110278413A/en active Pending
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101105892A (en) * | 2007-07-30 | 2008-01-16 | 深圳市融合视讯科技有限公司 | Vehicle traffic accident automatic detection method |
US20090052739A1 (en) * | 2007-08-23 | 2009-02-26 | Hitachi Kokusai Electric Inc. | Human pursuit system, human pursuit apparatus and human pursuit program |
CN105530465A (en) * | 2014-10-22 | 2016-04-27 | 北京航天长峰科技工业集团有限公司 | Security surveillance video searching and locating method |
CN104680555A (en) * | 2015-02-13 | 2015-06-03 | 电子科技大学 | Border-crossing detection method and border-crossing monitoring system based on video monitoring |
CN104660998A (en) * | 2015-02-16 | 2015-05-27 | 苏州阔地网络科技有限公司 | Relay tracking method and system |
CN104902223A (en) * | 2015-04-17 | 2015-09-09 | 国家电网公司 | Fault monitoring system of power line |
CN105245850A (en) * | 2015-10-27 | 2016-01-13 | 太原市公安局 | Method, device and system for tracking target across surveillance cameras |
CN105788166A (en) * | 2016-01-29 | 2016-07-20 | 维沃移动通信有限公司 | Monitoring and alarm method and apparatus |
CN108540754A (en) * | 2017-03-01 | 2018-09-14 | 中国电信股份有限公司 | Methods, devices and systems for more video-splicings in video monitoring |
CN108234961A (en) * | 2018-02-13 | 2018-06-29 | 欧阳昌君 | A kind of multichannel video camera coding and video flowing drainage method and system |
CN109729287A (en) * | 2018-12-06 | 2019-05-07 | 浙江大华技术股份有限公司 | A kind of method, apparatus and calculating equipment, storage medium of perimeter region monitoring |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112788229A (en) * | 2019-11-11 | 2021-05-11 | 株式会社数据价值 | Indoor self-shooting support camera system based on Internet of things |
CN111405203A (en) * | 2020-03-30 | 2020-07-10 | 杭州海康威视数字技术股份有限公司 | Method and device for determining picture switching, electronic equipment and storage medium |
CN112601021A (en) * | 2020-12-14 | 2021-04-02 | 中标慧安信息技术股份有限公司 | Method and system for processing monitoring video of network camera |
CN112601021B (en) * | 2020-12-14 | 2021-08-31 | 中标慧安信息技术股份有限公司 | Method and system for processing monitoring video of network camera |
CN112887666A (en) * | 2021-01-06 | 2021-06-01 | 浙江大华技术股份有限公司 | Video processing method and device, network camera, server and storage medium |
CN113114950A (en) * | 2021-05-28 | 2021-07-13 | 苏州智奇胜软件科技有限公司 | IoT camera control method and control system |
CN113438450A (en) * | 2021-06-11 | 2021-09-24 | 深圳市大工创新技术有限公司 | Dynamic target tracking monitoring method, monitoring system, electronic device and storage medium |
CN113438450B (en) * | 2021-06-11 | 2022-05-17 | 深圳市大工创新技术有限公司 | Dynamic target tracking and monitoring method |
CN114697501A (en) * | 2022-03-23 | 2022-07-01 | 南京云创大数据科技股份有限公司 | Monitoring camera image processing method and system based on time |
CN114697501B (en) * | 2022-03-23 | 2023-08-11 | 南京云创大数据科技股份有限公司 | Time-based monitoring camera image processing method and system |
CN116582653A (en) * | 2023-07-14 | 2023-08-11 | 广东天亿马信息产业股份有限公司 | Intelligent video monitoring method and system based on multi-camera data fusion |
CN116582653B (en) * | 2023-07-14 | 2023-10-27 | 广东天亿马信息产业股份有限公司 | Intelligent video monitoring method and system based on multi-camera data fusion |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110278413A (en) | Image processing method, device, server and storage medium | |
US20130123987A1 (en) | Robotic system, robot control method and robot control program | |
CN110267008B (en) | Image processing method, image processing apparatus, server, and storage medium | |
US10334158B2 (en) | Autonomous media capturing | |
CN103902046B (en) | Intelligent prompt method and terminal | |
CN107786549B (en) | Adding method, device, system and the computer-readable medium of audio file | |
CN110267007A (en) | Image processing method, device, server and storage medium | |
CN110278414A (en) | Image processing method, device, server and storage medium | |
CN105126355A (en) | Child companion robot and child companioning system | |
CN108875667B (en) | Target identification method and device, terminal equipment and storage medium | |
CN110177258A (en) | Image processing method, device, server and storage medium | |
US10971152B2 (en) | Imaging control method and apparatus, control device, and imaging device | |
CN109300476A (en) | Active chat device | |
CN109995991A (en) | A kind of image pickup method, robot and mobile terminal | |
CN103618944A (en) | Method for controlling video and user terminal | |
CN108197213A (en) | Action performs method, apparatus, storage medium and electronic device | |
CN111340848A (en) | Object tracking method, system, device and medium for target area | |
CN107168182A (en) | A kind of system and method for Indoor Robot VR applications | |
CN110266953B (en) | Image processing method, image processing apparatus, server, and storage medium | |
CN106325520A (en) | Photographing method and device and mobile terminal | |
CN110267010B (en) | Image processing method, image processing apparatus, server, and storage medium | |
CN103929460A (en) | Method for obtaining state information of contact and mobile device | |
CN110191324A (en) | Image processing method, device, server and storage medium | |
CN110267009A (en) | Image processing method, device, server and storage medium | |
CN110427502A (en) | Display methods, device, electronic equipment and the storage medium of virtual content |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190924 |
|
RJ01 | Rejection of invention patent application after publication |