CN110278413A - Image processing method, device, server and storage medium - Google Patents

Image processing method, device, server and storage medium Download PDF

Info

Publication number
CN110278413A
CN110278413A CN201910579248.0A CN201910579248A CN110278413A CN 110278413 A CN110278413 A CN 110278413A CN 201910579248 A CN201910579248 A CN 201910579248A CN 110278413 A CN110278413 A CN 110278413A
Authority
CN
China
Prior art keywords
target
camera
reference object
image
shooting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910579248.0A
Other languages
Chinese (zh)
Inventor
杜鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority to CN201910579248.0A priority Critical patent/CN110278413A/en
Publication of CN110278413A publication Critical patent/CN110278413A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/292Multi-camera tracking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/65Control of camera operation in relation to power supply
    • H04N23/651Control of camera operation in relation to power supply for reducing power consumption by affecting camera operations, e.g. sleep mode, hibernation mode or power off of selective parts of the camera
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/265Mixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Abstract

This application discloses a kind of image processing method, device, server and storage mediums, are related to technical field of image processing.This method is applied to server, server is connect with multiple cameras, different positions is arranged in for shooting different regions in multiple camera, and the region of the every two adjacent camera shooting in multiple camera connects or partly overlaps, which is stored with target reference object.This method comprises: from the image that each camera in multiple cameras takes, extract multiple target images including target reference object, obtain the shooting time of each target image in multiple target images, by the sequencing of the shooting time of each target image, multiple target images are sequentially spliced, the monitor video of the target reference object is obtained.The application passes through distributed camera and carries out track up to target reference object, and obtains the monitor video of the target reference object, to promote the monitoring effect of target reference object.

Description

Image processing method, device, server and storage medium
Technical field
This application involves technical field of image processing, more particularly, to a kind of image processing method, device, server And storage medium.
Background technique
With the development of society and the progress of science and technology, more and more places start to arrange monitoring system.Currently, passing through In the application scenarios that monitoring system is monitored, can only often some region fixed be supervised by monitoring used camera Control, monitoring effect are bad.
Summary of the invention
In view of the above problems, present applicant proposes a kind of image processing method, device, server and storage medium, with It solves the above problems.
In a first aspect, the embodiment of the present application provides a kind of image processing method, be applied to server, the server with Different positions is arranged in for shooting different regions in the connection of multiple cameras, the multiple camera, and the multiple takes the photograph As the region that the every two adjacent camera in head is shot connects or partly overlaps, the server is stored with target shooting pair As, which comprises from the image that each camera in the multiple camera takes, extracting includes the target Multiple target images of reference object;Obtain the shooting time of each target image in the multiple target image;By described The sequencing of the shooting time of each target image sequentially splices the multiple target image, obtains the target shooting The monitor video of object.
Second aspect, the embodiment of the present application provide a kind of image processing method, are applied to distributed system, including service Different positions is arranged in for shooting difference in device and the multiple cameras connecting with the server, the multiple camera Region, and in the multiple camera every two adjacent camera shooting region connect or partly overlap, it is described more Each camera in a camera is stored with target reference object, which comprises every in the multiple camera A camera extracts multiple target images including the target reference object from the image taken;The multiple camera shooting Head obtains the shooting time of each target image in the multiple target image;Each camera is by the multiple target The shooting time of each target image in image and the multiple target image is uploaded to the server;The server is pressed The sequencing of the shooting time of each target image, the multiple target image is sequentially spliced, and obtains the target The monitor video of reference object.
The third aspect, the embodiment of the present application provide a kind of image processing apparatus, be applied to server, the server with Different positions is arranged in for shooting different regions in the connection of multiple cameras, the multiple camera, and the multiple takes the photograph As the region that the every two adjacent camera in head is shot connects or partly overlaps, the server is stored with target shooting pair As described device includes: image zooming-out module, the image for taking from each camera in the multiple camera In, extract multiple target images including the target reference object;Shooting time obtains module, for obtaining the multiple mesh The shooting time of each target image in logo image;Image mosaic module, when for pressing the shooting of each target image Between sequencing, the multiple target image is sequentially spliced, the monitor video of the target reference object is obtained.
Fourth aspect, the embodiment of the present application provide a kind of server, including memory and processor, the memory coupling It is connected to the processor, the memory store instruction, the processor executes when executed by the processor The above method.
5th aspect, the embodiment of the present application provides a kind of computer-readable storage medium, described computer-readable Program code is stored in storage medium, said program code can be called by processor and execute the above method.
Image processing method, device, server and storage medium provided by the embodiments of the present application are applied to server, Server is connect with multiple cameras, and different positions is arranged in for shooting different regions in multiple camera, and this is more The region of every two adjacent camera shooting in a camera connects or partly overlaps, which is stored with target shooting pair As.From the image that each camera in multiple cameras takes, multiple target figures including target reference object are extracted Picture obtains the shooting time of each target image in multiple target images, by each target image shooting time it is successive Sequentially, multiple target images are sequentially spliced, obtains the monitor video of the target reference object, to pass through distributed camera Track up is carried out to target reference object, and obtains the monitor video of the target reference object, to promote target reference object Monitoring effect.
Detailed description of the invention
In order to more clearly explain the technical solutions in the embodiments of the present application, make required in being described below to embodiment Attached drawing is briefly described, it should be apparent that, the drawings in the following description are only some examples of the present application, for For those skilled in the art, without creative efforts, it can also be obtained according to these attached drawings other attached Figure.
Fig. 1 shows the schematic diagram of distributed system provided by the embodiments of the present application;
Fig. 2 shows the flow diagrams for the image processing method that the application one embodiment provides;
Fig. 3 shows the flow diagram of the image processing method of another embodiment of the application offer;
Fig. 4 shows distributed system provided by the embodiments of the present application and carries out track up to multiple target reference objects Schematic diagram;
Fig. 5 shows the flow diagram of the image processing method of the application further embodiment offer;
Fig. 6 shows a kind of schematic diagram of setting position of multiple cameras provided by the embodiments of the present application;
Fig. 7 shows the schematic diagram of another setting position of multiple cameras provided by the embodiments of the present application;
Fig. 8 shows the flow diagram of the step S303 of the image processing method shown in fig. 5 of the application;
Fig. 9 shows the flow diagram of the image processing method of another embodiment of the application offer;
Figure 10 shows a kind of flow diagram of the step S402 of the image processing method shown in Fig. 9 of the application;
Figure 11 shows another flow diagram of the step S402 of the image processing method shown in fig. 6 of the application;
Figure 12 shows the flow diagram of the application image processing method that further embodiment provides again;
Figure 13 shows the module frame chart of image processing apparatus provided by the embodiments of the present application;
Figure 14 shows the embodiment of the present application for executing the server of the image processing method according to the embodiment of the present application Block diagram;
Figure 15 shows realizing at according to the image of the embodiment of the present application for saving or carrying for the embodiment of the present application The storage unit of the program code of reason method.
Specific embodiment
In order to make those skilled in the art more fully understand application scheme, below in conjunction in the embodiment of the present application Attached drawing, the technical scheme in the embodiment of the application is clearly and completely described.
In recent years, with the progress of the development of society and science and technology, more and more places start to arrange monitoring system, and In the most application scenarios being monitored by monitoring system, used camera is all often single camera, can only Some region fixed is monitored, monitoring effect is bad.To solve the above problems, panoramic video monitoring system is gradually emerging It rises, wherein panoramic video monitoring system includes multiple cameras for being mounted on fixed position, for shooting multiple and different regions Image simultaneously forms panoramic picture, but range acquired in panoramic picture is excessive, can not be monitored for some object, can not Reach ideal monitoring effect.
In view of the above-mentioned problems, inventor has found by long-term research and proposes at image provided by the embodiments of the present application Method, apparatus, server and storage medium are managed, track up is carried out to target reference object by distributed camera, and The monitor video of the target reference object is obtained, to promote the monitoring effect of target reference object.Wherein, specific image procossing Method is described in detail in subsequent embodiment.
It will be described below for the distributed system suitable for image processing method provided by the embodiments of the present application.
Referring to Fig. 1, Fig. 1 shows the schematic diagram of distributed system provided by the embodiments of the present application, wherein the distribution System includes server 100 and multiple cameras 200 (quantity of camera 200 shown in Fig. 1 is 4), wherein server 100 It is connected respectively with each camera 200 in multiple cameras 200, for carrying out data interaction with each camera 200 respectively, For example, server 100, which receives the image of the transmission of camera 200, server 100, sends instruction etc. to camera 200, do not do herein It is specific to limit.In addition, the server 100 can be Cloud Server, or traditional server, the camera 200 can be with For gun-type camera, hemisphere camera, high-definition intelligent spherical shape camera, pen holder type camera, veneer camera, flying saucer camera shooting Head, Mobile phone type camera etc., and the camera lens of the camera can use wide-angle lens, standard lens, telephoto lens, varifocal mirror Head, pin hole mirror head etc., do not do specific restriction herein.
In some embodiments, different positions is arranged in for shooting different regions in multiple cameras 200, and more The region that every two adjacent camera 200 in a camera 200 is shot connects or partly overlaps.It is understood that every A camera 200 can correspond to according to the difference of its field angle and setting position and shoot different regions, by the way that every two is arranged The shooting area of adjacent camera 200 connects or partly overlaps, and the region that distributed system can be made to be shot is covered entirely Lid.Wherein, multiple cameras 200 can be spaced along a length direction is arranged side by side, for shooting the length direction region Image, multiple cameras 200 can also be spaced setting along a circumferential direction, for shooting the figure in the annular region Picture, certainly, multiple cameras 200 can also include other set-up modes, herein not as restriction.
In some embodiments, which is stored with target reference object.Wherein, target reference object can be with It is to be uploaded to server 100 after electronic equipment receives external input, and deposited target reference object by server 100 Storage.As a kind of mode, which can be personage, animal and other objects etc., wherein when target shooting pair When as personage, which may include man, woman, children, youth, old man etc., when target reference object is When animal, which may include cat, dog, rabbit lamp, and when target reference object is other objects, which is clapped Taking the photograph object may include car, lorry, car etc., it is not limited here.
Referring to Fig. 2, the flow diagram of the image processing method provided Fig. 2 shows the application one embodiment.Institute Image processing method is stated for carrying out track up to target reference object by distributed camera, and the target is obtained and shoots The monitor video of object, to promote the monitoring effect of target reference object.In the particular embodiment, described image processing method Applied to image processing apparatus 300 as shown in fig. 13 that and the server 100 configured with described image processing unit 300 (is schemed 14).It will illustrate the detailed process of the present embodiment by taking server as an example below, it will of course be understood that, the present embodiment is applied Server can be Cloud Server, or traditional server, it is not limited here.The server and multiple cameras connect It connects, and different positions is arranged in for shooting different regions, and the every two in multiple camera in multiple camera The region of adjacent camera shooting connects or partly overlaps, which is stored with target reference object.It will be directed to Fig. 2 institute below The process shown is explained in detail, and shown image processing method can specifically include following steps:
Step S101: from the image that each camera in the multiple camera takes, extracting includes the mesh Mark multiple target images of reference object.
As a kind of mode, each camera in multiple cameras can be in open state, and each camera In real time the shooting area covered shoot and the image taken is uploaded to server.Alternatively, Each camera in multiple cameras can receive external command, and response external instruction is in the open state or closes shape State, wherein camera in the open state can carry out shooting to the shooting area covered and will be on the image that taken Reach server, wherein external command may include the server being connect with multiple cameras send automatically command information, with Command information that the electronic equipment of multiple cameras connection is sent based on user's operation, user directly trigger life in multiple cameras At command information etc., it is not limited here.
In the present embodiment, server receives the image of each camera shooting overhead pass in multiple cameras taken, and The image taken to each camera identifies, from the image that each camera takes, extracting includes that target is shot Multiple target images of object.In some embodiments, server can read pre-stored target shooting pair from local As, wherein the target reference object that server is read may include the description file of target reference object, also may include target The image file etc. of reference object, wherein description file can be used for believing by feature of the text information to target reference object Breath is described, and image file can be used for being described by characteristic information of the image information to target reference object.
In some embodiments, server is receiving taking for each camera shooting overhead pass in multiple cameras After image, all images received can be compared with target reference object, include to be extracted from all images Multiple target images of target reference object, it is to be understood that wrapped in each target image in multiple target image Including target reference object, wherein the target reference object for including in each target image can be complete target reference object, It can also be with the part of target reference object, it is not limited here.As a kind of mode, server is in receiving multiple cameras Each camera shooting overhead pass the image taken after, can by all images received carry out feature information extraction, and will The characteristic information of the characteristic information and target reference object that extract is matched, and by the characteristic information with target reference object Image where matched characteristic information is determined as target image.For example, face can be carried out to by all images received Information extraction, and the face information of the face information extracted and target reference object is matched, and will be shot with target Image where the matched face information of the face information of object is determined as target image.
Step S102: the shooting time of each target image in the multiple target image is obtained.
As a kind of mode, each camera in multiple cameras can will be shot when uploading the image taken To image and take the shooting time of the image and be uploaded to server together, correspondingly, server is from the institute received Have after determining multiple target images in image, the corresponding shooting time of multiple target image can be searched respectively, to obtain Take the shooting time of each target image in multiple target images.In some embodiments, each of multiple cameras Camera by the image taken and can take when uploading the image taken and the shooting time for taking the image The shooting time of the image is associated, for example, the one-to-one mapping relations of image and shooting time are established, correspondingly, clothes Device be engaged in after determining multiple target images in all images received, it can be based on the shooting time of image and image Incidence relation searches the corresponding shooting time of multiple target image, to obtain each target image in multiple target images Shooting time.
Alternatively, server obtains in all images from each camera shooting overhead pass in multiple cameras To after multiple target images, the corresponding camera of each target image in multiple target images can be obtained respectively, is then sent out Send command information to the corresponding camera of each target image, which is used to indicate the bat of cam feedback target image Take the photograph the time, wherein the command information carries the identification information of each target image, correspondingly, the response of each camera receives The command information arrived, and target image is determined based on the identification information extracted from command information, it is searched in the local of camera The shooting time of the target image is uploaded to server by the shooting time of the target image, should so that server is available The shooting time of each target image in multiple target images.
Step S103: by the sequencing of the shooting time of each target image, the multiple target image is suitable Secondary splicing obtains the monitor video of the target reference object.
In some embodiments, server is obtaining each target in multiple target images and multiple target images After the shooting time of image, multiple target images can be arranged by the sequencing of the shooting time of each target image Sequence, it is to be understood that the shooting time of target image of the shooting time for the forward target image that sorts earlier than sequence rearward, Then multiple target images are spliced by the sequence of multiple target images, to generate monitor video, wherein the monitor video It include target reference object by each frame of playback progress, so as to promote the monitoring effect of target reference object.
In some embodiments, server can regard the monitoring after the monitor video for obtaining target reference object Frequency is sent to target terminal corresponding with the target reference object, so that the corresponding user's timely learning target of the target terminal The case where reference object.For example, the target reference object can be old man or child, the target terminal can correspond to the old man or The guardian of child, avoids the generation of fortuitous event at can be with the old man or child of timely learning family so as to guardian the case where.
In some embodiments, server, can be automatically to the prison after the monitor video for obtaining target reference object Control video is analyzed, and whether abnormal conditions occurs with the target reference object judged in the monitor video, when judging result table When levying the target reference object and abnormal conditions occur, it is whole to target corresponding with the target reference object that warning message can be sent End, so that the corresponding user of target terminal makes corresponding processing in time.Wherein, abnormal conditions may include falling down, crouching long not Rise etc., it is not limited here.In addition, server can also send the monitor video when sending warning message to target terminal Or the corresponding video clip of the abnormal conditions is shot to target terminal so that the corresponding user of target terminal understands target in time The truth of object.
The image processing method that the application one embodiment provides, takes from each camera in multiple cameras In image, multiple target images including target reference object are extracted, each target image in multiple target images is obtained Shooting time is sequentially spliced multiple target images by the sequencing of the shooting time of each target image, obtains the target The monitor video of reference object to carry out track up to target reference object by distributed camera, and obtains the mesh The monitor video of reference object is marked, to promote the monitoring effect of target reference object.
Referring to Fig. 3, Fig. 3 shows the flow diagram of the image processing method of another embodiment of the application offer. This method is applied to above-mentioned server, which connect with multiple cameras, and different positions is arranged in multiple camera It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple camera connects or part weight Folded, which is stored with target reference object.Wherein, target reference object includes first object reference object and the second target Reference object will be explained in detail for process shown in Fig. 3 below, shown image processing method can specifically include with Lower step:
Step S201: from the image that each camera in the multiple camera takes, extracting respectively includes institute State multiple first object images of first object reference object and multiple second targets including the second target reference object Image.
Step S202: obtaining the shooting time of each first object image in the multiple first object image respectively, With the shooting time of the second target image of each of the multiple second target image.
Step S203: by the sequencing of the shooting time of each first object image, by the multiple first mesh Logo image is sequentially spliced, and obtains the monitor video of the first object reference object, and press each second target image Shooting time sequencing, the multiple second target image is sequentially spliced, the second target reference object is obtained Monitor video.
In some embodiments, which can be used for carrying out tracking bat to multiple target reference objects respectively It takes the photograph, and generates the monitor video of each target reference object in multiple target reference objects respectively.As shown in figure 4, Fig. 4 is shown Distributed system provided by the embodiments of the present application carries out the schematic diagram of track up to multiple target reference objects, wherein point Cloth system can be grouped each target reference object of multiple target reference objects, for example, multiple cameras are clapped What is taken the photograph includes that first object reference object (target 1) is added to a grouping, by multiple cameras take including second Target reference object (target 2) is added to another grouping, by multiple cameras take include third target reference object (target 3) is added to another and is grouped, and details are not described herein.
It in the present embodiment, include first object reference object and the second target reference object with multiple target reference objects For be illustrated, track up can be carried out to first object reference object and the second target reference object respectively, and to the One target reference object and the second target reference object are grouped, respectively generate first object reference object monitor video and The monitor video of second target reference object.For example, the first object reference object is old man, which is Child, then can carry out track up to old man and child respectively, and the image including old man is added to a grouping, will include The image of child is added to a grouping, and generates the monitor video of old man and the monitor video of child respectively.
Wherein, the specific descriptions of step S201- step S203 please refer to step S101- step S103, and details are not described herein.
The image processing method that another embodiment of the application provides, takes from each camera in multiple cameras Image in, respectively extract include first object reference object multiple first object images and including the second target reference object Multiple second target images, obtain the shooting time of each first object image in multiple first object images respectively, and The shooting time of the second target image of each of multiple second target images, by the shooting time of each first object image Sequencing sequentially splices multiple first object images, obtains the monitor video and each the of first object reference object The sequencing of the shooting time of two target images sequentially splices multiple second target images, obtains the shooting pair of the second target The monitor video of elephant.Compared to image processing method shown in Fig. 2, the present embodiment can also be right respectively by distributed camera Multiple target reference objects carry out track up, and obtain the monitor video of multiple target reference object respectively, more to be promoted The monitoring effect of a target reference object.
Referring to Fig. 5, Fig. 5 shows the flow diagram of the image processing method of the application further embodiment offer. This method is applied to above-mentioned server, which connect with multiple cameras, and different positions is arranged in multiple camera It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple camera connects or part weight Folded, which is stored with target reference object.The multiple camera includes the first camera and images with described first At least one second camera that head is disposed adjacent will be explained in detail, shown image for process shown in fig. 5 below Processing method can specifically include following steps:
Step S301: when the target reference object is located in the shooting area of first camera, described in control First camera is in the open state and at least one described second camera of control is in close state.
In the present embodiment, multiple camera include the first camera and be disposed adjacent with first camera to A few second camera.Wherein, the setting position of first camera and at least one second camera can be adjacent, and the The region of one camera and the shooting of at least one second camera connects or partly overlaps.As shown in fig. 6, Fig. 6 shows this Shen Please embodiment provide multiple cameras a kind of setting position view, in Fig. 6, multiple camera 200 include first Camera 200A and two second camera 200B being disposed adjacent with the first camera, it is possible to understand that, this two second Camera 200B is disposed adjacent with the first camera 200A respectively, and one of second camera 200B is set to the first camera shooting The left side of head, another second camera 200B are set to the right side of the first camera.As shown in fig. 7, Fig. 7 shows the application Another setting position view in multiple cameras that embodiment provides, in Fig. 7, multiple camera 200 includes the An one camera 200A and second camera 200B being disposed adjacent with the first camera, it is possible to understand that, this first is taken the photograph As the marginal position of distributed system is arranged in head 200A, and the side of first camera 200A is disposed adjacent one second and takes the photograph As head 200B.
In some embodiments, multiple cameras can be controlled in advance and be in open state, pass through multiple cameras In each camera image taking is carried out to its corresponding shooting area, whether and identifying in the image taken includes target Reference object, wherein when recognition result, which characterizes the target object, to be located in the shooting area of the first camera, that is to say, that When recognizing target reference object from the image that the first camera takes, and other cameras are shot from multiple cameras To image in do not recognize target reference object when, can control the first camera it is in the open state and control at least one A second camera is in close state, i.e., when the quantity of second camera is one, controls at a second camera It controls two second cameras when the quantity of second camera is two in closed state and is in close state, to reduce The power consumption of at least one second camera.
Step S302: the target reference object is shot by first camera, and monitors the target The behavioral data of reference object.
It should be understood that the zone of action of the target reference object is the corresponding shooting area of the first camera at this time, because This, can shoot target reference object by the first camera, and monitor the behavioral data of the target reference object.Make For a kind of mode, the behavioral data of target object may include movement speed, moving direction, behavior act etc., not limit herein It is fixed.
Step S303: when the behavioral data characterizes the shooting that the target reference object will leave first camera When region enters target area, target second corresponding with the target area is determined from least one described second camera Camera.
Wherein, the shooting area of the first camera will be left when the behavioral data that monitoring obtains characterizes the target reference object When domain enters target area, target second camera can be determined from least one second camera based on the target area. In some embodiments, when behavior data characterization target reference object moves towards target area from the shooting area of the first camera When domain, it is believed that the shooting area that the target reference object will leave the first camera enters target area;Work as the behavior When data characterization target reference object is located at the marginal position object-oriented region of the shooting area of the first camera, it is believed that The shooting area that the target reference object will leave the first camera enters target area etc., and details are not described herein.
Referring to Fig. 8, Fig. 8 shows the process signal of the step S303 of the image processing method shown in fig. 5 of the application Figure.It will be explained in detail below for process shown in Fig. 8, the method can specifically include following steps:
Step S3031: the moving direction of the target reference object is obtained based on the behavioral data.
In some embodiments, the movement of the target reference object can be obtained from the behavioral data that monitoring obtains Direction.It is understood that the target can be obtained according to behavioral datas such as walking postures, the direction of travel of target reference object The moving direction of reference object, details are not described herein.
Step S3032: determining the target area based on the moving direction, from least one described second camera Determine target second camera corresponding with the target area.
For example, as shown in fig. 6, when target reference object is located in the shooting area of the first camera 200A to walk to left lateral When, it is believed that the shooting area that the target reference object will leave the first camera 200A, which enters, is set to the first camera The left side of 200A target shooting area corresponding with the second camera 200B that the first camera 200A is disposed adjacent, can be by position It is taken the photograph in the second camera 200B that the left side of the first camera 200A is disposed adjacent with the first camera 200A as target second As head.
In another example as shown in fig. 6, when target reference object is located in the shooting area of the first camera 200A to walk to right lateral When, it is believed that the shooting area that the target reference object will leave the first camera 200A, which enters, is set to the first camera The right side of 200A target shooting area corresponding with the second camera 200B that the first camera 200A is disposed adjacent, can be by position It is taken the photograph in the second camera 200B that the right side of the first camera 200A is disposed adjacent with the first camera 200A as target second As head.
Step S304: the target second camera is controlled by closed state and is switched to open state.
In the present embodiment, after determining target second camera at least one second camera, it can control the mesh It marks second camera and open state is switched to by closed state, for the shooting area that will enter the target second camera Target reference object carry out track up.
Step S305: when the target reference object enters the target area, pass through the target second camera The target reference object is shot, and monitors the behavioral data of the target reference object.
Wherein, after the shooting area that target reference object leaves the first camera enters target area, it can control this First camera is in off state by open state switching, also be can control first camera and is continued to keep it turned on, This is without limitation.
In some embodiments, when the shooting area that target reference object leaves the first camera enters target area When, the zone of action of the target reference object is therefore the corresponding shooting area of target second camera can pass through mesh at this time Mark second camera shoots target reference object, and monitors the behavioral data of the target reference object.
Step S306: extracting the target signature information of the target reference object, and the target signature information includes biology At least one of characteristic information and behavior characteristic information.
In some embodiments, server can extract target shooting pair from the target reference object being locally stored The target signature information of elephant, the target signature information include generating at least one of characteristic information and behavior characteristic information, In, generating characteristic information may include face characteristic information, physical characteristic information, height characteristic information etc., behavior characteristic information It may include walking postures, the speed of travel, behavior act etc., it is not limited here.
Step S307: it from the image that each camera in the multiple camera takes, extracts and the target The matched characteristic information of characteristic information.
In some embodiments, after the target signature information for extracting target reference object, the target can be based on Characteristic information extracts and the matched characteristic information of the target signature information from the image of each camera shooting overhead pass.For example, when should When target signature information is the face information of target reference object, it can be extracted from all images of each camera shooting overhead pass With the matched face information of face information of the target reference object.In another example when the target signature information is target shooting pair When the walking postures of elephant, the walking appearance with the target reference object can be extracted from all images of each camera shooting overhead pass The matched walking postures of gesture.
Step S308: object corresponding with the matched characteristic information of the target signature information is determined as the target and is clapped Object is taken the photograph, and extracts multiple target images including the target reference object.
In some embodiments, it is extracted in the image from each camera shooting overhead pass matched with target signature information After characteristic information, object corresponding with the matched characteristic information of target signature information can be determined in the picture, it can will be with mesh The corresponding object of the mark matched characteristic information of characteristic information is determined as target reference object, and will include the target reference object Image be determined as target image, therefore, can from it is each camera shooting overhead pass all images in extract including the target clap Take the photograph multiple target images of object.
Step S309: the shooting time of each target image in the multiple target image is obtained.
Step S310: by the sequencing of the shooting time of each target image, the multiple target image is suitable Secondary splicing obtains the monitor video of the target reference object.
Wherein, the specific descriptions of step S309- step S310 please refer to step S104- step S105, and details are not described herein.
The image processing method that the application further embodiment provides, when target reference object is located at the bat of the first camera When taking the photograph in region, the first camera of control is in the open state and controls at least one second camera and is in close state, and leads to It crosses the first camera to shoot target reference object, and monitors the behavioral data of the target reference object, when behavior number When the shooting area for leaving the first camera being entered target area according to characterization target reference object, from least one the second camera shooting Target second camera corresponding with target area is determined in head, is controlled target second camera by closed state and is switched to unlatching State shoots target reference object by target second camera when target reference object enters target area, and Monitor the behavioral data of the target reference object.Extract the target signature information of target reference object, the target signature information packet At least one of biological information and behavior characteristic information are included, the figure taken from each camera in multiple cameras As in, extraction and the matched characteristic information of target signature information will be corresponding with the matched characteristic information of target signature information right As being determined as target reference object, and multiple target images including target reference object are extracted, obtained in multiple target images Each target image shooting time, it is by the sequencing of the shooting time of each target image, multiple target images are suitable Secondary splicing obtains the monitor video of target reference object.Compared to image processing method shown in Fig. 2, the present embodiment also according to Region where target reference object controls corresponding camera and opens or closes, to reduce the power consumption of camera.In addition, this reality The target signature information that example is applied also according to target reference object carries out recognition and tracking to target reference object, to promote target shooting The recognition success rate of object.
Referring to Fig. 9, Fig. 9 shows the flow diagram of the image processing method of another embodiment of the application offer. This method is applied to above-mentioned server, which connect with multiple cameras, and different positions is arranged in multiple camera It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple camera connects or part weight Folded, which is stored with target reference object.It will be explained in detail below for process shown in Fig. 9, shown image Processing method can specifically include following steps:
Step S401: the historical behavior data of the target reference object are obtained.
In some embodiments, server, can be to image when receiving the image of each camera shooting overhead pass every time In object behavioral data carry out analysis record and store, history of forming behavioral data.In the present embodiment, server can be from Locally read the historical behavior data of the target reference object, wherein historical behavior data may include historical act region, go through History activity time etc., it is not limited here.
Step S402: the historical behavior data are based on, shooting control is carried out to the multiple camera.
In some embodiments, server can be based on after the historical behavior data for obtaining the target reference object The historical behavior data of target reference object carry out shooting control to multiple cameras.For example, target reference object can be based on Historical behavior data control that multiple cameras are in open state, the multiple cameras of control are in closed state, control A part of camera in multiple cameras is in the open state and another part camera is in close state, and does not do herein It limits.
Referring to Fig. 10, Figure 10 shows a kind of stream of the step S402 of the image processing method shown in Fig. 9 of the application Journey schematic diagram.It will be explained in detail below for process shown in Fig. 10, the method can specifically include following steps:
Step S4021A: historical act region is extracted from the historical behavior data, wherein the historical act region It is greater than the region of preset duration for duration where characterizing the target reference object.
As an implementation, which may include historical act region, wherein the historical act area Domain is greater than the region of preset duration for the duration where characterizing the target reference object.Specifically, which can be preparatory Preset duration is obtained and is stored with, the preset duration is for acting on target reference object in the judgement of the duration where each region Foundation, therefore, in this present embodiment, the available target reference object of server is in the duration where each region, by this Target reference object where each region duration and preset duration be compared, wherein when comparison result characterizes the target Reference object is when the duration where some region is greater than preset duration, it is believed that the region is the history of target reference object It zone of action can when comparison result characterizes the target reference object when the duration where some region is no more than preset duration It is the historical act region of target reference object to think the region not.
For example, when the distributed system be applied to family when, home area respectively include parlor, bedroom, toilet and Four, kitchen region, can obtain duration of the target reference object in parlor, the duration in bedroom, the duration in toilet respectively And the duration in kitchen, and by judge respectively the duration in parlor, the duration in bedroom, toilet duration and Whether the duration in kitchen is greater than the mode of preset duration, determines to go through from parlor, bedroom, toilet and four, kitchen region History zone of action.Wherein, the duration can be total duration or the average duration as unit of one day, it is not limited here.
Step S4022A: multiple target camera shootings corresponding with the historical act region are searched from the multiple camera Head.
It in the present embodiment, can be based on the historical act regional search and the history after obtaining historical act region The corresponding multiple target cameras in zone of action.In some embodiments, the covering model in the available historical act region The shooting for enclosing, and searching the shooting area of the included camera of the coverage area, and the camera that server is included being covered The corresponding multiple cameras in region are determined as multiple target cameras.
Step S4023A: it is in the open state to control the multiple target camera, and controls in the multiple camera Other cameras in addition to the multiple target camera are in close state.
In some embodiments, the duration due to historical act area attribute target reference object in the region is greater than pre- If duration, that is to say, that therefore the main activities region of target reference object, can correspond in the historical act region The camera in historical act region is opened, and closes the camera outside historical act region, to realize to target reference object Track up while, moreover it is possible to reduce the power consumption of distributed system.Correspondingly, in the present embodiment, can control and history The corresponding multiple target cameras in zone of action are in the open state, and control in multiple cameras except multiple target camera Except other cameras be in close state.
Figure 11 is please referred to, Figure 11 shows another of the step S402 of the image processing method shown in Fig. 9 of the application Flow diagram.It will be explained in detail below for process shown in Figure 11, the method can specifically include following step It is rapid:
Step S4021B: the historical act time is extracted from the historical behavior data, wherein the historical act time The time being located at for characterizing the target reference object in the shooting area of the multiple camera.
As an implementation, which may include the historical act time, wherein when the historical act Between for characterizing the time that the target reference object is located in the shooting area of multiple cameras.For example, multiple camera Shooting area is interior, then, which is located at the indoor time for characterizing the target reference object, that is, It says, the historical act time is for characterizing the time that the target reference object is in.
Step S4022B: the multiple camera is controlled in the historical act time corresponding duration in opening state State, and the multiple camera of control are in close state outside the historical act time corresponding duration.
In some embodiments, since historical act time representation target reference object is in the shooting of multiple camera Time in region, that is to say, that within the time, target reference object is located in the shooting area of multiple camera, Outside the time, target reference object is located at outside the shooting area of multiple camera.Therefore, in the present embodiment, can control Multiple cameras are in the open state in historical act time corresponding duration, so that the tracking to target reference object is clapped It takes the photograph, and the multiple cameras of control are in close state outside historical act time corresponding duration, to reduce distributed system The power consumption of system.
Step S403: extracting the target signature information of the target reference object, and the target signature information includes biology At least one of characteristic information and behavior characteristic information.
Step S404: it from the image that each camera in the multiple camera takes, extracts and the target The matched characteristic information of characteristic information.
Step S405: object corresponding with the matched characteristic information of the target signature information is determined as the target and is clapped Object is taken the photograph, and extracts multiple target images including the target reference object.
Step S406: the shooting time of each target image in the multiple target image is obtained.
Step S407: by the sequencing of the shooting time of each target image, the multiple target image is suitable Secondary splicing obtains the monitor video of the target reference object.
Wherein, the specific descriptions of step S403- step S407 please refer to step S306- step S310, and details are not described herein.
The image processing method that another embodiment of the application provides obtains the historical behavior data of target reference object, Based on historical behavior data, shooting control is carried out to multiple cameras.Extract the target signature information of target reference object, the mesh Marking characteristic information includes at least one of biological information and behavior characteristic information, from each camera shooting in multiple cameras In the image that head takes, extraction and the matched characteristic information of target signature information will be with the matched features of target signature information The corresponding object of information is determined as target reference object, and extracts multiple target images including target reference object, obtains more The shooting time of each target image in a target image will be more by the sequencing of the shooting time of each target image A target image sequentially splices, and obtains the monitor video of target reference object.Compared to image processing method shown in Fig. 2, originally Embodiment controls corresponding camera also according to the historical behavior data of target reference object and opens or closes, to reduce camera Power consumption.In addition, the present embodiment also according to target reference object target signature information to target reference object identified with Track, to promote the recognition success rate of target reference object.
Figure 12 is please referred to, Figure 12 shows the timing diagram of the application image processing method that further embodiment provides again. This method is applied to above-mentioned distributed system, multiple to take the photograph including server and the multiple cameras connecting with the server It is used to shoot different regions as different positions is arranged in head, and the every two adjacent camera shooting in multiple cameras Region connect or partly overlap, each camera in multiple cameras is stored with target reference object.It will be directed to below Process shown in Figure 12 is explained in detail, and shown image processing method can specifically include following steps:
Step S501: for each camera in the multiple camera from the image taken, extracting includes the mesh Mark multiple target images of reference object.
Step S502: when the multiple camera obtains the shooting of each target image in the multiple target image Between.
Step S503: each camera is by each mesh in the multiple target image and the multiple target image The shooting time of logo image is uploaded to the server.
Step S504: the server, will be the multiple by the sequencing of the shooting time of each target image Target image sequentially splices, and obtains the monitor video of the target reference object.
In some embodiments, each camera in multiple camera is stored with target reference object.Wherein, mesh Mark reference object, which can be electronic equipment and receive, is sent to each camera after external input, and by each camera by target Reference object is stored locally.In the present embodiment, camera is according to the target reference object being locally stored directly to bat The multiple images for taking the photograph acquisition are judged, and the target figure including target reference object is obtained from the multiple images that shooting obtains Target image is uploaded to server by picture, to reduce the data volume of the image uploaded, transfer efficiency in promotion.
Wherein, the specific descriptions of step S501- step S504 please refer to step S101- step S103, and details are not described herein.
The application image processing method that further embodiment provides again, each camera in multiple cameras is from shooting To image in, extract include target reference object multiple target images, multiple cameras obtain in multiple target images The shooting time of each target image, each camera is by each target image in multiple target images and multiple target images Shooting time be uploaded to server, server by the shooting time of each target image sequencing, by multiple target figures As sequentially splicing, the monitor video of target reference object is obtained, to carry out by distributed camera to target reference object Track up, and the monitor video of the target reference object is obtained, to promote the monitoring effect of target reference object.
Figure 13 is please referred to, Figure 13 shows the module frame chart of image processing apparatus 300 provided by the embodiments of the present application.The figure Picture processing unit 300 is applied to above-mentioned server, and server is connect with multiple cameras, and different positions is arranged in multiple cameras It sets for shooting different regions, and the region of the every two adjacent camera shooting in multiple cameras connects or part weight Folded, server is stored with target reference object.It will be illustrated below for block diagram shown in Figure 13, shown image processing apparatus 300 include: image zooming-out module 310, shooting time acquisition module 320 and image mosaic module 330, in which:
Image zooming-out module 310, for extracting from the image that each camera in the multiple camera takes Multiple target images including the target reference object.Further, described image extraction module 310 includes: target signature Information extraction submodule, feature information extraction submodule and image zooming-out submodule, in which:
Target signature information extracting sub-module, for extracting the target signature information of the target reference object, the mesh Marking characteristic information includes at least one of biological information and behavior characteristic information.
Feature information extraction submodule, for from the image that each camera in the multiple camera takes, It extracts and the matched characteristic information of the target signature information.
Image zooming-out submodule, for object corresponding with the matched characteristic information of the target signature information to be determined as The target reference object, and extract multiple target images including the target reference object.
Shooting time obtains module 320, when for obtaining the shooting of each target image in the multiple target image Between.
Image mosaic module 330 will be the multiple for the sequencing of the shooting time by each target image Target image sequentially splices, and obtains the monitor video of the target reference object.
Further, described image processing unit 300 further include: first state control module, the first behavior data monitoring Module, target camera determining module, the second status control module and the second behavioral data monitoring modular, in which:
First state control module, for being located in the shooting area of first camera when the target reference object When, control first camera is in the open state and at least one described second camera of control is in close state.
First behavior data monitoring module, for being clapped by first camera the target reference object It takes the photograph, and monitors the behavioral data of the target reference object.
Target camera determining module, for described the will to be left when the behavioral data characterizes the target reference object When the shooting area of one camera enters target area, the determining and target area from least one described second camera Corresponding target second camera.Further, the target camera determining module include: moving direction acquisition submodule and Target camera determines submodule, in which:
Moving direction acquisition submodule, for obtaining the mobile side of the target reference object based on the behavioral data To.
Target camera determines submodule, for determining the target area based on the moving direction, from it is described at least Target second camera corresponding with the target area is determined in one second camera.
Second status control module is switched to open state by closed state for controlling the target second camera.
Second behavioral data monitoring modular, for passing through institute when the target reference object enters the target area It states target second camera to shoot the target reference object, and monitors the behavioral data of the target reference object.
Further, described image processing unit 300 further include: behavioral data obtains module and shooting control module, In:
Behavioral data obtains module, for obtaining the historical behavior data of the target reference object.
Control module is shot, for being based on the historical behavior data, shooting control is carried out to the multiple camera.Into One step, the shooting control module include: that historical act extracted region submodule, target camera search submodule and the Three condition control submodule, in which:
Historical act extracted region submodule, for extracting historical act region from the historical behavior data, wherein Duration where the historical act region is used to characterize the target reference object is greater than the region of preset duration.
Target camera searches submodule, corresponding with the historical act region for searching from the multiple camera Multiple target cameras.
Third state control submodule, it is in the open state for controlling the multiple target camera, and described in control Other cameras in multiple cameras in addition to the multiple target camera are in close state.
Further, described, the shooting control module further include: historical act time extracting sub-module and the 4th state Control submodule, in which:
Historical act time extracting sub-module, for extracting the historical act time from the historical behavior data, wherein The historical act time is for characterizing the time that the target reference object is located in the shooting area of the multiple camera.
4th mode control word module, for controlling the multiple camera in the historical act time corresponding duration It is interior in the open state, and control the multiple camera and be in closing shape outside the historical act time corresponding duration State.
It is apparent to those skilled in the art that for convenience and simplicity of description, foregoing description device and The specific work process of module, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In several embodiments provided herein, the mutual coupling of module can be electrical property, mechanical or other The coupling of form.
It, can also be in addition, can integrate in a processing module in each functional module in each embodiment of the application It is that modules physically exist alone, can also be integrated in two or more modules in a module.Above-mentioned integrated mould Block both can take the form of hardware realization, can also be realized in the form of software function module.
Figure 14 is please referred to, it illustrates a kind of structural block diagrams of server 100 provided by the embodiments of the present application.The server 100 can be Cloud Server, be also possible to traditional server.Server 100 in the application may include it is one or more such as Lower component: processor 110, memory 120 and one or more application program, wherein one or more application programs can be by It is stored in memory 120 and is configured as being executed by one or more processors 110, one or more programs are configured to hold Row method as described in preceding method embodiment.
Wherein, processor 110 may include one or more processing core.Processor 110 utilizes various interfaces and route The various pieces in entire server 100 are connected, by running or executing the instruction being stored in memory 120, program, code Collection or instruction set, and the data being stored in memory 120 are called, the various functions and processing data of execute server 100. Optionally, processor 110 can be compiled using Digital Signal Processing (Digital Signal Processing, DSP), scene Journey gate array (Field-Programmable Gate Array, FPGA), programmable logic array (Programmable Logic Array, PLA) at least one of example, in hardware realize.Processor 110 can integrating central processor (Central Processing Unit, CPU), in graphics processor (Graphics Processing Unit, GPU) and modem etc. One or more of combinations.Wherein, the main processing operation system of CPU, user interface and application program etc.;GPU is for being responsible for Show the rendering and drafting of content;Modem is for handling wireless communication.It is understood that above-mentioned modem It can not be integrated into processor 110, be realized separately through one piece of communication chip.
Memory 120 may include random access memory (Random Access Memory, RAM), also may include read-only Memory (Read-Only Memory).Memory 120 can be used for store instruction, program, code, code set or instruction set.It deposits Reservoir 120 may include storing program area and storage data area, wherein the finger that storing program area can store for realizing operating system Enable, for realizing at least one function instruction (such as touch function, sound-playing function, image player function etc.), be used for Realize the instruction etc. of following each embodiments of the method.Storage data area can also store the number that terminal 100 is created in use According to (such as phone directory, audio, video data, chat record data) etc..
Figure 15 is please referred to, it illustrates a kind of structural frames of computer readable storage medium provided by the embodiments of the present application Figure.Program code is stored in the computer-readable medium 400, said program code can be called by processor and execute the above method Method described in embodiment.
Computer readable storage medium 400 can be such as flash memory, EEPROM (electrically erasable programmable read-only memory), The electronic memory of EPROM, hard disk or ROM etc.Optionally, computer readable storage medium 400 includes non-volatile meter Calculation machine readable medium (non-transitory computer-readable storage medium).Computer-readable storage Medium 400 has the memory space for the program code 410 for executing any method and step in the above method.These program codes can With from reading or be written in one or more computer program product in this one or more computer program product. Program code 410 can for example be compressed in a suitable form.
In conclusion image processing method provided by the embodiments of the present application, device, server and storage medium, application In server, server is connect with multiple cameras, and different positions is arranged in for shooting different areas in multiple camera Domain, and the region of the every two adjacent camera shooting in multiple camera connects or partly overlaps, which is stored with Target reference object.From the image that each camera in multiple cameras takes, extracting includes target reference object Multiple target images obtain the shooting time of each target image in multiple target images, by the shooting of each target image The sequencing of time sequentially splices multiple target images, obtains the monitor video of the target reference object, thus by dividing Cloth camera carries out track up to target reference object, and obtains the monitor video of the target reference object, to promote mesh Mark the monitoring effect of reference object.
Finally, it should be noted that above embodiments are only to illustrate the technical solution of the application, rather than its limitations;Although The application is described in detail with reference to the foregoing embodiments, those skilled in the art are when understanding: it still can be with It modifies the technical solutions described in the foregoing embodiments or equivalent replacement of some of the technical features;And These are modified or replaceed, do not drive corresponding technical solution essence be detached from each embodiment technical solution of the application spirit and Range.

Claims (12)

1. a kind of image processing method, which is characterized in that be applied to server, the server is connect with multiple cameras, institute It states multiple cameras and different positions is set for shooting different regions, and the every two in the multiple camera is adjacent Camera shooting region connect or partly overlap, the server is stored with target reference object, which comprises
From the image that each camera in the multiple camera takes, extracting includes the more of the target reference object A target image;
Obtain the shooting time of each target image in the multiple target image;
By the sequencing of the shooting time of each target image, the multiple target image is sequentially spliced, obtains institute State the monitor video of target reference object.
2. the method according to claim 1, wherein each camera from the multiple camera is clapped In the image taken the photograph, multiple target images including the target reference object are extracted, comprising:
The target signature information of the target reference object is extracted, the target signature information includes biological information and behavior At least one of characteristic information;
From the image that each camera in the multiple camera takes, extract matched with the target signature information Characteristic information;
Object corresponding with the matched characteristic information of the target signature information is determined as the target reference object, and is extracted Multiple target images including the target reference object.
3. the method according to claim 1, wherein the target reference object includes first object reference object In the image taken with the second target reference object, each camera from the multiple camera, extraction includes Multiple target images of the target reference object, comprising:
From the image that each camera in the multiple camera takes, extracting respectively includes that the first object is shot Multiple first object images of object and multiple second target images including the second target reference object;
The shooting time for obtaining each target image in the multiple target image, comprising:
The shooting time and the multiple second of each first object image in the multiple first object image is obtained respectively The shooting time of the second target image of each of target image;
The sequencing of the shooting time by each target image, the multiple target image is sequentially spliced, is obtained Obtain the monitor video of the target reference object, comprising:
By the sequencing of the shooting time of each first object image, the multiple first object image is sequentially spelled It connects, obtains the monitor video of the first object reference object, and by the shooting time of each second target image Sequencing sequentially splices the multiple second target image, obtains the monitor video of the second target reference object.
4. method according to claim 1-3, which is characterized in that the multiple camera includes the first camera And at least one second camera being disposed adjacent with first camera, it is described from each of the multiple camera In the image that camera takes, before extraction includes multiple target images of the target reference object, further includes:
When the target reference object is located in the shooting area of first camera, controls first camera and be in At least one second camera described in open state and control is in close state;
The target reference object is shot by first camera, and monitors the behavior of the target reference object Data;
The shooting area for leaving first camera is entered into target when the behavioral data characterizes the target reference object When region, target second camera corresponding with the target area is determined from least one described second camera;
It controls the target second camera and open state is switched to by closed state;
When the target reference object enters the target area, the target is shot by the target second camera Object is shot, and monitors the behavioral data of the target reference object.
5. according to the method described in claim 4, it is characterized in that, described when the behavioral data characterizes the target shooting pair When as the shooting area for leaving first camera being entered target area, determined from least one described second camera Target second camera corresponding with the target area, comprising:
The moving direction of the target reference object is obtained based on the behavioral data;
The target area is determined based on the moving direction, the determining and target from least one described second camera The corresponding target second camera in region.
6. method according to claim 1-3, which is characterized in that described from each of the multiple camera In the image that camera takes, before extraction includes multiple target images of the target reference object, further includes:
Obtain the historical behavior data of the target reference object;
Based on the historical behavior data, shooting control is carried out to the multiple camera.
7. according to the method described in claim 6, it is characterized in that, described be based on the historical behavior data, to the multiple Camera carries out shooting control, comprising:
Historical act region is extracted from the historical behavior data, wherein the historical act region is for characterizing the mesh Duration is greater than the region of preset duration where marking reference object;
Multiple target cameras corresponding with the historical act region are searched from the multiple camera;
It is in the open state to control the multiple target camera, and controls in the multiple camera except the multiple target is taken the photograph As other cameras except head are in close state.
8. according to the method described in claim 6, it is characterized in that, described be based on the historical behavior data, to the multiple Camera carries out shooting control, comprising:
The historical act time is extracted from the historical behavior data, wherein the historical act time is for characterizing the mesh Mark reference object is located at the time in the shooting area of the multiple camera;
It is in the open state in the historical act time corresponding duration to control the multiple camera, and described in control Multiple cameras are in close state outside the historical act time corresponding duration.
9. a kind of image processing method, which is characterized in that be applied to distributed system, including server and with the server Different positions is arranged in for shooting different regions in multiple cameras of connection, the multiple camera, and the multiple The region of every two adjacent camera shooting in camera connects or partly overlaps, and each of the multiple camera is taken the photograph As head is stored with target reference object, which comprises
For each camera in the multiple camera from the image taken, extracting includes the more of the target reference object A target image;
The multiple camera obtains the shooting time of each target image in the multiple target image;
Each camera is by the shooting of each target image in the multiple target image and the multiple target image Time is uploaded to the server;
The server is sequentially spelled the multiple target image by the sequencing of the shooting time of each target image It connects, obtains the monitor video of the target reference object.
10. a kind of image processing apparatus, which is characterized in that be applied to server, the server is connect with multiple cameras, institute It states multiple cameras and different positions is set for shooting different regions, and the every two in the multiple camera is adjacent Camera shooting region connect or partly overlap, the server is stored with target reference object, and described device includes:
Image zooming-out module includes institute for extracting from the image that each camera in the multiple camera takes State multiple target images of target reference object;
Shooting time obtains module, for obtaining the shooting time of each target image in the multiple target image;
Image mosaic module, for the sequencing of the shooting time by each target image, by the multiple target figure As sequentially splicing, the monitor video of the target reference object is obtained.
11. a kind of server, which is characterized in that including memory and processor, the memory is couple to the processor, institute Memory store instruction is stated, the processor is executed as claim 1-8 is any when executed by the processor Method described in.
12. a kind of computer-readable storage medium, which is characterized in that be stored with journey in the computer-readable storage medium Sequence code, said program code can be called by processor and execute the method according to claim 1.
CN201910579248.0A 2019-06-28 2019-06-28 Image processing method, device, server and storage medium Pending CN110278413A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910579248.0A CN110278413A (en) 2019-06-28 2019-06-28 Image processing method, device, server and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910579248.0A CN110278413A (en) 2019-06-28 2019-06-28 Image processing method, device, server and storage medium

Publications (1)

Publication Number Publication Date
CN110278413A true CN110278413A (en) 2019-09-24

Family

ID=67962588

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910579248.0A Pending CN110278413A (en) 2019-06-28 2019-06-28 Image processing method, device, server and storage medium

Country Status (1)

Country Link
CN (1) CN110278413A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111405203A (en) * 2020-03-30 2020-07-10 杭州海康威视数字技术股份有限公司 Method and device for determining picture switching, electronic equipment and storage medium
CN112601021A (en) * 2020-12-14 2021-04-02 中标慧安信息技术股份有限公司 Method and system for processing monitoring video of network camera
CN112788229A (en) * 2019-11-11 2021-05-11 株式会社数据价值 Indoor self-shooting support camera system based on Internet of things
CN112887666A (en) * 2021-01-06 2021-06-01 浙江大华技术股份有限公司 Video processing method and device, network camera, server and storage medium
CN113114950A (en) * 2021-05-28 2021-07-13 苏州智奇胜软件科技有限公司 IoT camera control method and control system
CN113438450A (en) * 2021-06-11 2021-09-24 深圳市大工创新技术有限公司 Dynamic target tracking monitoring method, monitoring system, electronic device and storage medium
CN114697501A (en) * 2022-03-23 2022-07-01 南京云创大数据科技股份有限公司 Monitoring camera image processing method and system based on time
CN116582653A (en) * 2023-07-14 2023-08-11 广东天亿马信息产业股份有限公司 Intelligent video monitoring method and system based on multi-camera data fusion

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101105892A (en) * 2007-07-30 2008-01-16 深圳市融合视讯科技有限公司 Vehicle traffic accident automatic detection method
US20090052739A1 (en) * 2007-08-23 2009-02-26 Hitachi Kokusai Electric Inc. Human pursuit system, human pursuit apparatus and human pursuit program
CN104660998A (en) * 2015-02-16 2015-05-27 苏州阔地网络科技有限公司 Relay tracking method and system
CN104680555A (en) * 2015-02-13 2015-06-03 电子科技大学 Border-crossing detection method and border-crossing monitoring system based on video monitoring
CN104902223A (en) * 2015-04-17 2015-09-09 国家电网公司 Fault monitoring system of power line
CN105245850A (en) * 2015-10-27 2016-01-13 太原市公安局 Method, device and system for tracking target across surveillance cameras
CN105530465A (en) * 2014-10-22 2016-04-27 北京航天长峰科技工业集团有限公司 Security surveillance video searching and locating method
CN105788166A (en) * 2016-01-29 2016-07-20 维沃移动通信有限公司 Monitoring and alarm method and apparatus
CN108234961A (en) * 2018-02-13 2018-06-29 欧阳昌君 A kind of multichannel video camera coding and video flowing drainage method and system
CN108540754A (en) * 2017-03-01 2018-09-14 中国电信股份有限公司 Methods, devices and systems for more video-splicings in video monitoring
CN109729287A (en) * 2018-12-06 2019-05-07 浙江大华技术股份有限公司 A kind of method, apparatus and calculating equipment, storage medium of perimeter region monitoring

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101105892A (en) * 2007-07-30 2008-01-16 深圳市融合视讯科技有限公司 Vehicle traffic accident automatic detection method
US20090052739A1 (en) * 2007-08-23 2009-02-26 Hitachi Kokusai Electric Inc. Human pursuit system, human pursuit apparatus and human pursuit program
CN105530465A (en) * 2014-10-22 2016-04-27 北京航天长峰科技工业集团有限公司 Security surveillance video searching and locating method
CN104680555A (en) * 2015-02-13 2015-06-03 电子科技大学 Border-crossing detection method and border-crossing monitoring system based on video monitoring
CN104660998A (en) * 2015-02-16 2015-05-27 苏州阔地网络科技有限公司 Relay tracking method and system
CN104902223A (en) * 2015-04-17 2015-09-09 国家电网公司 Fault monitoring system of power line
CN105245850A (en) * 2015-10-27 2016-01-13 太原市公安局 Method, device and system for tracking target across surveillance cameras
CN105788166A (en) * 2016-01-29 2016-07-20 维沃移动通信有限公司 Monitoring and alarm method and apparatus
CN108540754A (en) * 2017-03-01 2018-09-14 中国电信股份有限公司 Methods, devices and systems for more video-splicings in video monitoring
CN108234961A (en) * 2018-02-13 2018-06-29 欧阳昌君 A kind of multichannel video camera coding and video flowing drainage method and system
CN109729287A (en) * 2018-12-06 2019-05-07 浙江大华技术股份有限公司 A kind of method, apparatus and calculating equipment, storage medium of perimeter region monitoring

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112788229A (en) * 2019-11-11 2021-05-11 株式会社数据价值 Indoor self-shooting support camera system based on Internet of things
CN111405203A (en) * 2020-03-30 2020-07-10 杭州海康威视数字技术股份有限公司 Method and device for determining picture switching, electronic equipment and storage medium
CN112601021A (en) * 2020-12-14 2021-04-02 中标慧安信息技术股份有限公司 Method and system for processing monitoring video of network camera
CN112601021B (en) * 2020-12-14 2021-08-31 中标慧安信息技术股份有限公司 Method and system for processing monitoring video of network camera
CN112887666A (en) * 2021-01-06 2021-06-01 浙江大华技术股份有限公司 Video processing method and device, network camera, server and storage medium
CN113114950A (en) * 2021-05-28 2021-07-13 苏州智奇胜软件科技有限公司 IoT camera control method and control system
CN113438450A (en) * 2021-06-11 2021-09-24 深圳市大工创新技术有限公司 Dynamic target tracking monitoring method, monitoring system, electronic device and storage medium
CN113438450B (en) * 2021-06-11 2022-05-17 深圳市大工创新技术有限公司 Dynamic target tracking and monitoring method
CN114697501A (en) * 2022-03-23 2022-07-01 南京云创大数据科技股份有限公司 Monitoring camera image processing method and system based on time
CN114697501B (en) * 2022-03-23 2023-08-11 南京云创大数据科技股份有限公司 Time-based monitoring camera image processing method and system
CN116582653A (en) * 2023-07-14 2023-08-11 广东天亿马信息产业股份有限公司 Intelligent video monitoring method and system based on multi-camera data fusion
CN116582653B (en) * 2023-07-14 2023-10-27 广东天亿马信息产业股份有限公司 Intelligent video monitoring method and system based on multi-camera data fusion

Similar Documents

Publication Publication Date Title
CN110278413A (en) Image processing method, device, server and storage medium
US20130123987A1 (en) Robotic system, robot control method and robot control program
CN110267008B (en) Image processing method, image processing apparatus, server, and storage medium
US10334158B2 (en) Autonomous media capturing
CN103902046B (en) Intelligent prompt method and terminal
CN107786549B (en) Adding method, device, system and the computer-readable medium of audio file
CN110267007A (en) Image processing method, device, server and storage medium
CN110278414A (en) Image processing method, device, server and storage medium
CN105126355A (en) Child companion robot and child companioning system
CN108875667B (en) Target identification method and device, terminal equipment and storage medium
CN110177258A (en) Image processing method, device, server and storage medium
US10971152B2 (en) Imaging control method and apparatus, control device, and imaging device
CN109300476A (en) Active chat device
CN109995991A (en) A kind of image pickup method, robot and mobile terminal
CN103618944A (en) Method for controlling video and user terminal
CN108197213A (en) Action performs method, apparatus, storage medium and electronic device
CN111340848A (en) Object tracking method, system, device and medium for target area
CN107168182A (en) A kind of system and method for Indoor Robot VR applications
CN110266953B (en) Image processing method, image processing apparatus, server, and storage medium
CN106325520A (en) Photographing method and device and mobile terminal
CN110267010B (en) Image processing method, image processing apparatus, server, and storage medium
CN103929460A (en) Method for obtaining state information of contact and mobile device
CN110191324A (en) Image processing method, device, server and storage medium
CN110267009A (en) Image processing method, device, server and storage medium
CN110427502A (en) Display methods, device, electronic equipment and the storage medium of virtual content

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190924

RJ01 Rejection of invention patent application after publication