KR101619953B1 - Method for image processing of surveilance camera by using auto multi-tracing - Google Patents

Method for image processing of surveilance camera by using auto multi-tracing Download PDF

Info

Publication number
KR101619953B1
KR101619953B1 KR1020150109695A KR20150109695A KR101619953B1 KR 101619953 B1 KR101619953 B1 KR 101619953B1 KR 1020150109695 A KR1020150109695 A KR 1020150109695A KR 20150109695 A KR20150109695 A KR 20150109695A KR 101619953 B1 KR101619953 B1 KR 101619953B1
Authority
KR
South Korea
Prior art keywords
image
unit
image processing
camera
photographed
Prior art date
Application number
KR1020150109695A
Other languages
Korean (ko)
Inventor
김배훈
Original Assignee
주식회사 영국전자
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 주식회사 영국전자 filed Critical 주식회사 영국전자
Priority to KR1020150109695A priority Critical patent/KR101619953B1/en
Application granted granted Critical
Publication of KR101619953B1 publication Critical patent/KR101619953B1/en

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • G08B13/19608Tracking movement of a target, e.g. by detecting an object predefined as a target, using target direction and or velocity to predict its new position
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19617Surveillance camera constructional details
    • G08B13/19626Surveillance camera constructional details optical details, e.g. lenses, mirrors or multiple lenses
    • G08B13/19628Surveillance camera constructional details optical details, e.g. lenses, mirrors or multiple lenses of wide angled cameras and camera groups, e.g. omni-directional cameras, fish eye, single units having multiple cameras achieving a wide angle view
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19617Surveillance camera constructional details
    • G08B13/1963Arrangements allowing camera rotation to change view, e.g. pivoting camera, pan-tilt and zoom [PTZ]
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19665Details related to the storage of video surveillance data

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Studio Devices (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

The present invention relates to the following. According to another aspect of the present invention, there is provided an automatic multi-tracking image processing method including: transmitting an image captured by a camera unit to a video tracking unit and a video splitting unit, respectively; Recognizing an observation object moving in the image transmitted from the image tracking unit, generating coordinate information according to movement of the observation object, and transmitting the coordinate information to the image processing unit; Dividing the image into a plurality of images in the image dividing unit, and transmitting the divided images to the image processing unit; The image processing unit arranges at least one image among the plurality of divided images transmitted from the image dividing unit, processes the enlarged or reduced image of the image by moving or replacing the image, and performs a virtual pan-tilt, zoom- Processing an image by applying a function; And outputting the image processed by the image processing unit through a display unit.

Description

TECHNICAL FIELD [0001] The present invention relates to an automatic multi-tracking image processing method for a surveillance camera,

The present invention relates to an automatic multi-tracking image processing method for a surveillance camera, and more particularly, to a method of processing an image captured by a camera in a shooting target area, Since the image can be tracked at the same time and the image can be divided into a plurality of positions and pan, tilt, and zoom can be performed for each position, the same function is performed in a camera having no other device for pan, tilt, and zoom And more particularly, to an automatic multi-tracking image processing method for a surveillance camera in which monitoring efficiency can be improved by automatically tracking a monitored object.

Generally, a surveillance camera is installed in a surveillance area requiring surveillance, and is a device for surveillance by photographing. Such a surveillance camera is installed in an area where surveillance of homes, major facilities, and various buildings is required, and is installed so as to shoot intruders in the installed area, fire, occurrence of crime, or tracking of criminals.

The surveillance camera monitors surveillance while the surveillant monitors the image connected to the video output device. If the surveillance area is increased, the number of the video output device is increased and the quantity to be monitored by the surveillant is increased so that the observer's fatigue The monitor may be ineffectively deteriorated due to an increase in the degree of negligence or concentration, and the monitor may miss fire or intruder intrusion from the image output to the video output device.

In addition, since a typical surveillance camera simply provides images photographed in a surveillance area, it is difficult to identify an intruder or recognize a fire in the image. Therefore, if the pan, tilt, and zoom functions are added to the surveillance camera, the surveillance efficiency can be improved because the camera can move according to the position of the surveillance target object or enlarge the surveillance target object.

As described above, in order to add the pan, tilt, and zoom functions to the surveillance camera, devices capable of performing each function must be added, so that a large cost can be added as compared with a conventional camera.

In addition, in order to install a camera having a pan, tilt, and zoom function, it is necessary to replace the existing camera or to use it with the existing camera, which increases the cost and is incompatible with each other.

Accordingly, it is possible to process the image photographed by the surveillance camera and to track the movement of the monitored object while being able to pan, tilt, and zoom in the image without adding any other devices related to pan, tilt, and zoom, There is a need for an image processing method that can be applied to a camera having various functions using a camera.

SUMMARY OF THE INVENTION The present invention has been made to solve the problems described above, and it is an object of the present invention to provide an image capturing apparatus, an image capturing apparatus, The present invention also provides an automatic multiple tracking image processing method for a surveillance camera capable of performing a pan and tilt function while moving for each divided position and performing a zoom function for each position, thereby improving the monitoring efficiency.

The technical problem of the present invention is not limited to those mentioned above, and another technical problem which is not mentioned can be clearly understood by those skilled in the art from the following description.

According to an aspect of the present invention, there is provided an automatic multiple tracking image processing method for a surveillance camera, comprising: transmitting an image photographed by a camera unit to a video tracking unit and a video segmenting unit; Recognizing an observation object moving in the image transmitted from the image tracking unit, generating coordinate information according to movement of the observation object, and transmitting the coordinate information to the image processing unit; Dividing the image into a plurality of images in the image dividing unit, and transmitting the divided images to the image processing unit; The image processing unit arranges at least one image among the plurality of divided images transmitted from the image dividing unit, processes the enlarged or reduced image of the image by moving or replacing the image, and performs a virtual pan-tilt, zoom- Processing an image by applying a function; And outputting the image processed by the image processing unit through a display unit.

In addition, the image processor may process the virtual pan-tilt function by changing the arrangement position of the at least one image based on the coordinate information.

The image processing unit may process the virtual pan-tilt function by replacing the arranged at least one image with another image based on the coordinate information.

In addition, the image processing unit may process a virtual pan-tilt function by changing the arrangement position of the at least one image based on the input value input through the user interface unit.

In addition, the image processor may process the virtual pan-tilt function by replacing the at least one image with another image based on the input value input through the user interface unit.

The method may further include processing the image by displaying the movement of the observation target based on the coordinate information in the image processing unit.

The display of the movement of the observation object may give the observation object a predetermined identification display.

In addition, the predetermined identification display may be configured so that a predetermined figure is displayed on the observation object.

In addition, the camera unit may include a fish eye lens, and may include a fisheye camera having a wide angle of view to transmit the captured fisheye image to the image tracking unit and the image segmentation unit.

The camera unit may include a thermal sensing camera for sensing a thermal change, and may transmit the sensed thermal sensed image to the image tracking unit and the image dividing unit.

The camera unit may include a fisheye sensor for sensing a thermal change while widening an angle of view with a fisheye lens, and may transmit the sensed heat sensed fisheye image to the image tracking unit and the image divider.

The display unit may display the image photographed in the center, and the divided images divided into a plurality of images by the image dividing unit may be positioned on the outer circumference of the captured image, and the pan, tilt, And zooming, respectively, based on the input image data.

In addition, in the step of transmitting to the display unit, the image processing unit transmits at least one of divided position information of the divided image, photographed date information, and photographed position information to the display unit so as to be output from the display .

In addition, the image checking unit connected to the camera unit checks whether there is a video signal, and if there is no video signal, it can transmit a reset signal to the camera unit.

According to another aspect of the present invention, there is provided an automatic multiple tracking image processing method for a surveillance camera, comprising the steps of: capturing an image with a camera unit having a plurality of cameras; transmitting the captured image to the image tracking unit and the image dividing unit; Recognizing a moving object to be moved in the images transmitted from the image tracking unit, generating coordinate information according to movement of the object, and transmitting the coordinate information to the image processing unit; Dividing the images into a plurality of images in the image dividing unit, and transmitting the divided images to the image processing unit; The image processing unit combines the divided images received from the image dividing unit into one image in the image combining unit while maintaining the divided positions, and arranges at least one of the plurality of images divided in the synthesized image Processing the image by moving or replacing the image, enlarging and reducing the image, and applying a virtual pan-tilt and zoom-in-zoom-out function to the image; And outputting the image processed by the image processing unit through a display unit.

In addition, the image processor may process the virtual pan-tilt function by changing the arrangement position of the arranged at least one image based on the coordinate information.

In addition, the image processing unit may process the virtual pan-tilt function by replacing the arranged at least one image with another image based on the coordinate information.

In addition, the image processing unit may process a virtual pan-tilt function by changing the arrangement position of the at least one image based on the input value input through the user interface unit.

The image processing unit may process the virtual pan-tilt function by replacing the at least one image with another image based on the input value input through the user interface unit.

The method may further include processing the image by displaying the movement of the observation object based on the coordinate information in the image processing unit.

In addition, the display of the movement of the observation object can give the observation object a predetermined identification mark.

In addition, the predetermined identification display may be configured so that a predetermined figure is displayed on the observation object.

The camera unit may include a photographing camera that photographs an area to be photographed, and a thermal sensing camera that senses a thermal change of a photographing target area. The camera senses an image photographed by the photographing camera and a thermal sensed image photographed by the thermal sensing camera, To the image tracking unit and the image dividing unit.

When the first fisheye camera and the first fisheye camera are combined with the first fisheye camera and the first fisheye camera in a direction opposite to the first fisheye camera, And a second fisheye camera for photographing a direction opposite to the first fisheye camera so as to be an image of a full-field angle and a front elevation angle, wherein the first fisheye image photographed by the first fisheye camera and the second fisheye image photographed by the second fisheye camera And may transmit the image to the image tracking unit and the image dividing unit.

In addition, when the first panoramic camera photographed with an azimuth angle of 180 degrees in the direction in which the photographing target region is photographed, and the image of the first panoramic camera in a direction opposite to the first panoramic camera, The first panoramic image captured by the first panoramic camera and the second panoramic image captured by the second panoramic camera are captured by the image tracking unit and the second panoramic image captured by the second panoramic camera, And to the image divider.

The details of other embodiments are included in the detailed description and drawings.

According to the automatic multi-tracking image processing method for a surveillance camera according to an embodiment of the present invention, the image is divided into a plurality of regions while automatically tracking the position and movement path of the monitored object in the image captured by the camera It is possible to perform a pan and tilt function while moving for each divided position, and to enlarge each position to perform a zoom function, thereby improving the monitoring efficiency.

The effects of the present invention are not limited to the effects mentioned above, and other effects not mentioned can be clearly understood by those skilled in the art from the description of the claims.

1 is a flowchart of an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.
2 is a first configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention. .
3 is a second configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.
FIG. 4 is a first example of a fish-eye image screen output from the display unit of the system of FIG.
5 is a second example of a screen displayed on the display unit of the system of FIG.
6 is a third example of a screen displayed on the display unit of the system of FIG.
FIG. 7 is a fourth example of a screen output from the display unit of the system of FIG.
Fig. 8 is a screen when the enlargement ratio is increased in the screen of Fig.
9 is a third block diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.
10 is a fourth block diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.
11 is a fifth configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.
12 is a sixth configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.
13 is an example of a screen of a fish-eye image output from the camera unit of the system of Fig.
14 is a seventh configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.
FIG. 15 is a first example of a screen of panorama images output from the camera unit of the system of FIG.
16 is a second example of a screen of panorama images output from the camera unit of the system of Fig.

DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the accompanying drawings, so that those skilled in the art can easily carry out the present invention.

In the following description of the embodiments of the present invention, descriptions of techniques which are well known in the technical field of the present invention and are not directly related to the present invention will be omitted. This is for the sake of clarity of the present invention without omitting the unnecessary explanation.

For the same reason, some of the components in the drawings are exaggerated, omitted, or schematically illustrated. Also, the size of each component does not entirely reflect the actual size. In the drawings, the same or corresponding components are denoted by the same reference numerals.

Hereinafter, the present invention will be described with reference to the accompanying drawings for explaining an automatic multiple tracking image processing method for a surveillance camera according to embodiments of the present invention.

FIG. 1 is a flowchart of an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention. FIG. 2 is a flowchart illustrating a method for processing an automatic multiple tracking image for a surveillance camera according to an exemplary embodiment of the present invention. FIG.

Referring to FIGS. 1 and 2, an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention includes a photographing transmission step S10 for photographing a surveillance target area, a image tracking transmission step S21, A step S22, an image processing transmission step S30, and a video output step S40.

The photographing transmission step S10 is a step in which the camera section 10 having the photographing camera 11 photographs an area to be photographed and the image photographed by the photographing camera 11 is transmitted to the image tracking section 30 and the image dividing section 40 Respectively.

The photographing transmission step S10 includes a video image tracking unit 30 for photographing the photographed area to be photographed with the camera unit 10 to track the photographed image and a pan, And a zoom function in an image and transmits the image to the image dividing unit 40 for dividing the image. Here, the camera section 10 uses a photographing camera 11 for photographing a normal image, and can use a camera in which a region to be monitored is already connected.

The image checking unit 20 is electrically connected to the camera unit 10 and receives a photographed image to check whether a video signal is present. If there is no video signal, the video checking unit 20 transmits a reset signal to the camera unit 10, The part 10 can be operated to be restarted.

The image tracking transmission step S21 is a step of recognizing a moving object to be observed in the image transmitted from the image tracking unit 30 and generating coordinate information according to the movement of the object to be observed and transmitting the coordinate information to the image processing unit 50. [

That is, the image tracking unit 30 extracts coordinates by designating an object to be moved as an observation object, except a designated background, when the image taken from the camera unit 10 is received. When the extracted observation object moves, it continuously extracts the movement coordinates along the trajectory to be moved and transmits the coordinate information automatically extracted to the image processing unit 50.

The image tracking unit 30 extracts the coordinates of the observation object moving in the background specified by the frame in the photographed image, tracks the coordinates of the moving object when the observation object moves, and transmits the coordinate information to the image processing unit 50.

Accordingly, the image tracking unit 30 continuously extracts the coordinates of the trajectory to be moved after specifying the coordinates of the moving object to be observed, and then extracts the coordinates so as to display the object in a predetermined shape so as to be identified by the image processing The observer can automatically follow the trajectory of the observed object specified by the movement of the rectangle.

Here, the quadrangle is only one example for displaying the observation object, and all types of figures surrounding the outer periphery of the observation object can be used.

When a plurality of objects to be observed are displayed at the time of extracting the coordinates of the figure shape of the object to be observed by the image tracking unit 30, a plurality of coordinates are simultaneously specified for each of the plurality of objects in a single graphic form, Can be extracted to multiple coordinates so as to automatically track the individually moved locus and transmitted to the image processing unit 50.

That is, when a plurality of objects to be observed are displayed, the image tracking unit 30 specifies the positions of the objects to be observed as rectangles in the form of a rectangle, and then, in accordance with the movement of the plurality of objects to be observed, And transmits each of the moving coordinates to the image processing unit 50 so as to display the trajectory moving in a rectangular shape while the plurality of objects to be observed are automatically traced at multiple positions simultaneously in accordance with the movement. Therefore, the image tracking unit 30 can automatically track the object to be traced in a rectangular shape without regard to the quantity, thereby improving the monitoring efficiency.

The image segmentation transmission step S22 is a step of dividing the image into a plurality of segments in the image segmentation unit 40 having received the image shot by the camera unit 10 and transmitting the segmented image to the image processing unit 50 .

The image processing transmission step S30 arranges at least one image among the plurality of divided images transmitted from the image division unit 40 in the image processing unit 50 and processes the image movement or substitution, A virtual pan-tilt, and a zoom-in-zoom-out function.

In this way, the image processing unit 50 can process the virtual pan-tilt function by changing the arrangement position of at least one image arranged based on the coordinate information.

Also, the image processing unit 50 may process the virtual pan-tilt function by replacing at least one image arranged based on the coordinate information with another image.

The image processing unit 50 may process the virtual pan-tilt function by changing the arrangement position of the at least one image based on the input value input through the user interface unit 80. [ In this way, the interface unit 80 can perform a pan, tilt, and zoom function in the image by the user's operation, and can automatically track the movement along the observation object according to the coordinates extracted by the image tracking unit 30 The input value can be set.

In addition, the image processing unit 50 may process the virtual pan-tilt function by replacing the at least one image with another image based on the input value input through the user interface unit 80. [

In addition, the image processing unit 50 can display the movement of the observation object based on the coordinate information, and process the image.

The display of the movement of the observation object can give a predetermined identification mark to the observation object.

The predetermined identification display described above can be configured so that a predetermined figure is displayed on the observation object.

That is, the identification mark can be processed so as to indicate the position where the observation object is moved in the image according to the coordinates of the quadrangle tracked by the image tracking unit 30. [

In addition, in the image processing transmission step S30, at least one of the divided position information, the photographed date information, and the photographed position information of the image divided by the image processing unit 50 may be transmitted to the display unit 70. [ Such various types of information may be generated in the caption processing unit 60 in the form of a caption and transmitted to the display unit 70.

That is, the caption processing unit 60 is connected to one side of the image processing unit 50 and transmits information such as a position at which the video image is divided, a date taken at the camera unit 10, And is displayed in the form of a caption on the image according to the font size and font. In this way, the caption processing unit 60 can recognize the position and the date of the monitored image by the caption displayed on the surveillance image by the surveillant, so that the surveillance efficiency can be improved.

The image output step S40 is a step of outputting the image transmitted from the image processing unit 50 in the display unit 70. [ The video output step S40 is a step of outputting an image processed by the image processing unit 50 according to a control signal of the user interface unit 80 so that the surveillant can observe the image on the display unit 70. [ Here, the display unit 70 may include any means capable of outputting a normal image, and a monitor for outputting an image may be used.

The display unit 70 displays the images photographed in the center. The divided images divided into a plurality of images by the image dividing unit 40 are positioned on the outer circumference of the photographed image, ), It is possible to output images processed to perform the functions of pan, tilt, and zoom respectively. For example, if the photographed image is divided into eight images by the image dividing unit 40, the entire image is output to the center of the display unit 70, and the divided images are divided into the outer portions of the divided positions along the outer circumference Respectively.

That is, the display unit 70 can output the entire image and the divided image, and can output the entire image and the divided image together on one display unit 70, And output the entire image and the divided image, respectively.

When the image processing unit 50 processes the divided images so that the divided images are output, the image processing unit 50 processes the images to be divided, Can be output. In addition, the subtitles generated by the subtitle processor 60 can be displayed on the divided video and on the entire video, respectively. The image processing unit 50 may perform a zoom function to enlarge or reduce the respective divided images.

Accordingly, since the camera unit 10 can perform a panning, tilting, and zooming function virtually through the image processing unit 50 in the photographed image without moving or other devices having a panning, tilting, and zooming function, The camera can also perform a pan, tilt, and zoom function.

Particularly, the coordinates of the observation object moved in the image are extracted by the image tracking unit 30 and transmitted to the image processing unit 50, and the image processing unit 50 displays a square shaped marking on the outer circumference of the observation object And outputs an image to the display unit 70 to continuously monitor the object to be monitored. In addition, by providing images to be automatically tracked to the display unit 70 that outputs the divided images and the entire images using the pan, tilt, and zoom functions at the time of tracking in a rectangular shape according to the coordinates on which the observed object is moved, Tracking function can be performed.

FIG. 3 is a second configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention. FIG. Fig. 5 shows a second example of the screen displayed on the display unit of the system of Fig. 3, Fig. 6 shows a third example of the screen displayed on the display unit of the system of Fig. 3, FIG. 8 shows a screen in the case where the enlargement ratio is increased in the screen of FIG. 7. FIG.

3 to 8, an automatic multiple tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention includes a photographing transmission step S10 for photographing a surveillance target area with a fisheye camera 12, S21, an image division transmission step S22, an image processing transmission step S30, and a video output step S40. Here, the automatic multi-tracking image processing method for a surveillance camera according to an embodiment of the present invention is similar to the automatic multi-tracking image processing method for surveillance cameras shown in FIGS. 1 and 2 except that the fisheye camera 12 is used. Explain only the differences in how they are the same.

The photographing transmission step S10 is a step of photographing a region to be photographed by the camera section 10 having a fisheye lens 12 equipped with a fisheye lens and sending a fisheye image photographed by the camera section 10 to the image tracking section 30 To the image division unit 40, respectively. Here, the fisheye camera 12 may be photographed through a fisheye fish-shaped fisheye lens so as to be enlarged to an azimuth angle and an altitude angle of 180 degrees of the photographed area, thereby minimizing a rectangular area that is not photographed, thereby improving monitoring efficiency.

The fisheye image photographed using the fisheye camera 12 is transmitted to the image tracking unit 30 and the image segmentation unit 40 in the image transmission step S10.

Extracts the coordinates of the observation object moved in the fish-eye image transmitted to the image-tracking unit 30 in the image-tracking transmission step S21, and transmits the extracted coordinates to the image-processing unit 50. [ At this time, coordinates for displaying the outer circumference of the observation object in a rectangular shape are generated, and coordinates of a rectangular shape are transmitted to the image processing unit 50 so that the monitor can easily identify the displayed rectangular image while being traced with the observation object. In addition, when a plurality of observation objects appear in the fish-eye image, the multi-moving coordinates are extracted and transmitted to the image processing unit 50.

In the image segmentation and transmission step S22, the fisheye image transmitted to the image segmentation unit 40 is divided into a plurality of segments, and the segmented fisheye image is transmitted to the image processing unit 50. [

The image processing transmission step S30 is a step of arranging at least one fisheye image among the plurality of divided fisheye images transmitted from the image division unit 40 in the image processing unit 50 and moving or replacing the fisheye image, And processing a fisheye image by applying a virtual pan-tilt and zoom-in-zoom-out function.

In this way, the image processing unit 50 can process the virtual pan-tilt function by changing the arrangement position of at least one fisheye image disposed based on the coordinate information.

Also, the image processing unit 50 can process the virtual pan-tilt function by replacing at least one fisheye image disposed on the basis of the coordinate information with another image.

The image processing unit 50 may process the virtual pan-tilt function by changing the arrangement position of the at least one fisheye image based on the input value input through the user interface unit 80. [ The interface unit 80 can perform pan, tilt, and zoom functions in a fisheye image by operation of the user, and automatically tracks the movement along the object to be observed based on the coordinates extracted by the image tracking unit 30. [ The input value can be set.

In addition, the image processing unit 50 may process a virtual pan-tilt function by replacing at least one fisheye image disposed on the basis of the input value input through the user interface unit 80 with another fisheye image.

Further, the image processing unit 50 can display the movement of the observation target based on the coordinate information, and process the fish-eye image.

The display of the movement of the observation object can give a predetermined identification mark to the observation object.

The predetermined identification display described above can be configured so that a predetermined figure is displayed on the observation object.

That is, the identification mark can be processed so as to indicate the position where the observation object is moved in the image according to the coordinates of the quadrangle tracked by the image tracking unit 30. [

Here, the image processing unit 50 may generate information on the image during the image processing by the caption processing unit 60 in a caption format, and transmit the caption to the display unit 70 so that the caption is displayed on the image upon output.

In the video output step S40, the display unit 70 outputs the image transmitted from the image processing unit 50 that processed the fisheye image.

9 is a third block diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.

9, an automatic multi-tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention includes an image capturing transmission step S10 capturing a surveillance target area with a thermal sensing camera 13, a image capturing transmission step S21, An image division transmission step S22, an image processing transmission step S30, and a video output step S40.

Here, the automatic multi-tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention is similar to the automatic multi-tracking image processing method for a surveillance camera shown in FIGS. 1 and 2 except that a thermal sensing camera 13 is used. Only the difference method will be described as the same.

In the shooting transmission step S10, the camera unit 10 having the thermal sensing camera 13 for sensing a thermal change photographs an area to be photographed, and the thermal sensed image photographed by the camera unit 10 is transmitted to a video tracking unit 30 and the image divider 40, respectively. Here, the thermal sensing camera 13 displays the thermal distribution in the image photographed at the time of photographing of the photographed area in terms of lightness or saturation difference, and a situation in which heat is suddenly changed, for example, As a result, the surveillance efficiency can be improved because the surveillance person can clearly determine the intruder approach or fire occurrence.

The thermal sensed image photographed using the thermal sensing camera 13 is transmitted to the image tracking unit 30 and the image dividing unit 40 in the image capturing transmission step S10.

In the image tracking transmission step S21, coordinates of the moving object to be moved in the heat sensing image transmitted to the image tracking unit 30 are tracked while being extracted and transmitted to the image processing unit 50. At this time, coordinates for displaying the outer circumference of the observation object in a rectangular shape are generated, and coordinates of a rectangular shape are transmitted to the image processing unit 50 so that the monitor can easily identify the displayed rectangular image while being traced with the observation object. When a plurality of objects to be observed are displayed in the thermal sensed image, they are extracted while tracking the coordinates that are moved in multiple, and are transmitted to the image processor 50.

In the image segmentation and transmission step S22, the thermal sensed image transmitted to the image segmentation unit 40 is divided into a plurality of segments and the divided thermal sensed images are transmitted to the image processing unit 50. [

In the image processing transmission step S30, at least one thermal sensing image among the plurality of divided thermal sensing images transmitted from the image dividing unit 40 is arranged in the image processing unit 50, and movement or replacement of the thermal sensing image, Zooming in and out of the image, and processing the image by applying a virtual pan-tilt and zoom-in-zoom-out function.

In this way, the image processing unit 50 can process the virtual pan-tilt function by changing the arrangement position of at least one thermal sensed image arranged based on the coordinate information.

In addition, the image processing unit 50 may process the virtual pan-tilt function by replacing at least one thermal sensed image disposed on the basis of the coordinate information with another thermal sensed image.

The image processing unit 50 may process the virtual pan-tilt function by changing the arrangement position of the at least one thermal sensed image based on the input value input through the user interface unit 80. [ The interface unit 80 can perform a pan, tilt, and zoom function in the heat sensed image by the user's operation, and automatically moves according to the coordinates extracted by the image tracking unit 30 You can set the input value to trace.

In addition, the image processing unit 50 may process a virtual pan-tilt function by replacing at least one thermal sensed image disposed on the basis of an input value input through the user interface unit 80 with another thermal sensed image .

In addition, the image processing unit 50 can display the movement of the observation object based on the coordinate information to process the thermal sensed image.

The display of the movement of the observation object can give a predetermined identification mark to the observation object.

The predetermined identification display described above can be configured so that a predetermined figure is displayed on the observation object.

That is, the identification mark can be processed so as to indicate the position where the observation object is moved in the image according to the coordinates of the quadrangle tracked by the image tracking unit 30. [

Here, the image processing unit 50 may generate information on the image in the caption processing by the caption processing unit 60, and transmit the caption to the display unit 70 so that the caption is displayed on the image.

In the video output step S40, the display unit 70 outputs the image received from the image processing unit 50 that processes the thermal sensed image.

10 is a fourth block diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.

10, an automatic multi-tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention includes a photographing transmission step S10 for photographing a surveillance target area with a fisheye thermal sensing camera 14, a image tracking transmission step S21 ), An image segmentation and transmission step S22, an image processing transmission step S30, and a video output step S40.

Here, the automatic multi-tracking image processing method for a surveillance camera according to the embodiment of the present invention is similar to the automatic multi-tracking image processing method for a surveillance camera shown in Figs. 1 and 2 except for the method using the automatic multi- The method should be described only as to how it differs according to the same.

In the photographing transmission step S10, a camera section 10 having a fisheye lens 14 and a fisheye sensor 14 for sensing heat with a wide angle of view is photographed by the camera section 10, To the image tracking unit 30 and the image dividing unit 40, respectively. Here, the fisheye thermal sensing camera 14 photographs an image that senses heat through a fish-eye fisheye lens so as to be enlarged to an azimuth angle and an altitude angle of 180 degrees of a photographed area, It is possible to detect the thermal change in the shooting target area, so that the monitoring efficiency can be improved by shooting the heat change such as the movement of the intruder or the fire without a square.

The fisheye sensed image photographed using the fisheye thermal sensing camera 14 is transmitted to the image tracking unit 30 and the image segmentation unit 40 in the image transmission step S10.

The coordinates of the moving object to be moved in the fuzzy heat sensed image transmitted to the image tracking unit 30 in the image tracking transmission step S21 are tracked while being extracted and transmitted to the image processing unit 50. [ At this time, coordinates for displaying the outer circumference of the observation object in a rectangular shape are generated, and coordinates of a rectangular shape are transmitted to the image processing unit 50 so that the monitor can easily identify the displayed rectangular image while being traced with the observation object. When a plurality of objects to be observed are displayed in the fisheye thermal image, the coordinates are moved while being tracked, and the coordinates are extracted and transmitted to the image processor 50.

In the image segmentation and transmission step S22, the fisheye sensed image transmitted to the image segmentation unit 40 is divided into a plurality of segments and the segmented fisheye sensor image is transmitted to the image processing unit 50. [

The image processing transmission step S30 arranges at least one fuzzy heat sensed image among the plurality of divided fuzzy heat sensed images transmitted from the image division unit 40 in the image processing unit 50, Zooming in and out on the image, and applying the virtual pan-tilt and zoom-in / zoom-out functions to process the image.

In this way, the image processing unit 50 can process the virtual pan-tilt function by changing the arrangement position of at least one fisheye heat sensed image arranged based on the coordinate information.

Also, the image processing unit 50 may process the virtual pan-tilt function by replacing at least one fisheye sensed image disposed on the basis of the coordinate information with another fisheye sensed image.

The image processing unit 50 may process the virtual pan-tilt function by changing the arrangement position of the at least one fisheye thermal sensed image based on the input value input through the user interface unit 80 . The interface unit 80 can perform a pan, tilt, and zoom function within a fisheye heat sensed image by the user's manipulation, and can automatically move along the object to be observed according to the coordinates extracted by the image tracking unit 30. [ The input value can be set to track the input.

In addition, the image processing unit 50 processes a virtual pan-tilt function by replacing at least one fisheye sensed image disposed on the basis of the input value input through the user interface unit 80 with another fisheye sensed image .

In addition, the image processing unit 50 can display the movement of the observation object based on the coordinate information to process the fisheye sensed image.

The display of the movement of the observation object can give a predetermined identification mark to the observation object.

The predetermined identification display described above can be configured so that a predetermined figure is displayed on the observation object.

That is, the identification mark may be processed so as to indicate the position where the observation object is moved in the fisheye sensor image according to the coordinates of the quadrangle tracked by the image tracking unit 30.

Here, the image processing unit 50 may generate information on the image in a caption format by the caption processing unit 60, and transmit the caption to the display unit 70 so that the caption is displayed on the image.

In the video output step S40, the display unit 70 outputs the image transmitted from the image processing unit that processed the fisheye sensed image.

11 is a fifth configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention.

11, an automatic multiple tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention includes a photographing transmission step S10 for photographing a surveillance target area with a photographing camera 11 and a heat sensing camera 13, A trace transmission step S21, an image segmentation transmission step S22, an image processing transmission step S30, and a video output step S40.

Here, a fifth configuration of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention includes a method of processing an automatic multiple tracking image for a surveillance camera shown in FIGS. 1 and 2, (S30) for composing a plurality of images are different from each other, and only a difference method is explained as the principle of processing images is the same.

The photographing transmission step S10 is performed by the camera unit 10 including the photographing camera 11 for photographing the photographing target area and the heat sensing camera 13 for photographing the heat sensed image for sensing the thermal change , And transmitting the photographed image and the heat sensed image photographed by the camera section 10 to the image tracking section 30 and the image dividing section 40, respectively. Here, the camera unit 10 includes a photographing camera 11 for photographing an area to be photographed, and a heat sensing camera 13 for sensing a thermal change. The camera unit 10 includes a column for detecting a thermal change such as intrusion of an intruder or fire And transmits the sensed image and the photographed image of the surrounding area to the image tracking unit 30 and the image segmenting unit 40, respectively. That is, the camera unit 10 not only monitors the occurrence of an intruder and the occurrence of a fire through a thermal sensed image sensing a thermal change in the region to be photographed, It is possible to continuously monitor the cause and progress, thereby providing a variety of image information according to prevention and follow-up measures as well as enhancement of monitoring efficiency.

The photographed image photographed by the photographed camera 11 and the photographed photographed image photographed by the thermal sensitive camera 13 are transmitted to the image tracing unit 30 and the image dividing unit 40 in the image transmitting step S10 .

In the image tracking transmission step S21, coordinates of the moving object and the moving object moving in the thermal image are extracted and transmitted to the image processing unit 50, respectively. At this time, coordinates for displaying the outer circumference of the observation object in a rectangular shape are generated, and coordinates of a rectangular shape are transmitted to the image processing unit 50 so that the monitor can easily identify the displayed rectangular image while being traced with the observation object. In addition, when a plurality of objects to be observed appear in the photographed image and the heat-traced image, it is possible to extract and extract the multi-moved coordinates to the image processing unit 50.

In the image segmentation and transmission step S22, the photographed image and the heat sensed image transmitted to the image segmentation unit 40 are divided into a plurality of segments, respectively, and the segmented image and the heat sensed image are transmitted to the image processing unit 50. [

The image processing transmission step S30 is a step in which the image processing unit 50 generates a single image from the image combining unit 90 while maintaining the divided positions of the divided shot image and the thermal sense image transmitted from the image dividing unit 40, And arranges at least one image among the plurality of divided images in the synthesized image, and processes the image by moving or replacing the image, enlarging and reducing the image, and applying the virtual pan-tilt and zoom-in and zoom-out functions .

In this way, the image processing unit 50 can process the virtual pan-tilt function by changing the arrangement position of at least one image arranged based on the coordinate information.

Also, the image processing unit 50 may process the virtual pan-tilt function by replacing at least one image arranged based on the coordinate information with another image.

The image processing unit 50 may process the virtual pan-tilt function by changing the arrangement position of at least one image arranged on the basis of the input value input through the user interface unit 80. [ The interface unit 80 can perform the pan, tilt, and zoom functions in the image synthesized by the user, and automatically moves according to the coordinates extracted by the image tracking unit 30 along the observation object You can set the input value to trace.

In addition, the image processing unit 50 may process a virtual pan-tilt function by replacing at least one image arranged on the basis of the input value inputted through the user interface unit 80 with another image.

Further, the image processing unit 50 can display the movement of the observation object based on the coordinate information, and process the synthesized image.

The display of the movement of the observation object can give a predetermined identification mark to the observation object.

The predetermined identification display described above can be configured so that a predetermined figure is displayed on the observation object.

That is, the identification mark can be processed so as to indicate the position where the observation object is moved in the image obtained by combining the photographed image and the heat sensed image according to the coordinates of the quadrangle tracked by the image tracking unit 30. [

In the image output step S40, the display unit 70 outputs the image received from the image processing unit 50, which is a composite image of the photographed image and the heat sensed image.

FIG. 12 is a sixth configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention, FIG. 13 is a view illustrating a screen of a fish- Yes.

12 and 13, an automatic multiple tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention includes a photographing transmission step of photographing a surveillance target area with a pair of fisheye cameras 12 installed in different directions, A video image transmission step S30, a video image transmission step S30, and a video output step S40.

Hereinafter, an automatic multiple tracking image processing method for a surveillance camera according to an exemplary embodiment of the present invention will be described with reference to FIGS. 1 and 2, and a method of processing an automatic multiple tracking image for a surveillance camera and a plurality of images (S30) for composing a plurality of images, and only the difference method is explained as the principle of image processing is the same.

When the images are synthesized, the camera unit 10 including the fish-eye cameras 12, in which the 360-degree omnidirectional angle and the total altitude angle are photographed, And transmits the two fisheye images taken by the unit 10 to the image tracking unit 30 and the image dividing unit 40, respectively. Here, the camera unit 10 includes a pair of fisheye cameras 12 for taking a fisheye image having an azimuth angle and an altitude angle of 180 degrees in mutually opposite directions, respectively. When the fisheye images opposite to each other are synthesized, Since the image having the angle of view of the omnidirectional angle and the total altitude angle can be obtained, the image of the entire direction can be taken without a rectangular area, and the surveillance efficiency can be improved by enlarging the surveillance area.

The fisheye images photographed by the pair of fisheye cameras 12 installed in mutually opposite directions are transmitted to the image tracking unit 30 and the image dividing unit 40 in the photograph transmission step S10.

In the image tracking transmission step (S21), the image tracking unit (30) extracts the coordinates of the observation object moved in the two fisheye images, and transmits the extracted coordinates to the image processing unit (50). At this time, coordinates for displaying the outer circumference of the observation object in a rectangular shape are generated, and coordinates of a rectangular shape are transmitted to the image processing unit 50 so that the monitor can easily identify the displayed rectangular image while being traced with the observation object. In addition, when a plurality of observation objects are displayed in the fisheye images, the coordinates can be extracted while being tracked in multiple, and transmitted to the image processing unit 50.

In the image segmentation and transmission step S22, the two fisheye images transmitted to the image segmentation unit 40 are divided into a plurality of segments, and the divided fisheye images are transmitted to the image processing unit 50. [

The image processing transmitting step S30 is a step in which the image processing unit 50 maintains the divided positions of the divided two eicetic images transmitted from the image dividing unit 40, Arranges at least one fisheye image among the plurality of fisheye images divided in the synthesized fisheye image, processes the movement or substitution of the fisheye image, enlargement and reduction of the image, and applies the virtual pan-tilt and zoom-in and zoom-out functions And processing the fisheye image.

In this way, the image processing unit 50 can process the virtual pan-tilt function by changing the arrangement position of at least one fisheye image disposed based on the coordinate information.

In addition, the image processing unit 50 can process the virtual pan-tilt function by replacing at least one fisheye image disposed based on the coordinate information with another fisheye image.

The image processing unit 50 may process the virtual pan-tilt function by changing the arrangement position of at least one fisheye image disposed on the basis of the input value input through the user interface unit 80. [ The interface unit 80 can perform a pan, tilt, and zoom function in a fisheye image synthesized by a user, and automatically moves along the object to be observed according to the coordinates extracted by the image tracking unit 30. [ The input value can be set to track the input.

In addition, the image processing unit 50 may process a virtual pan-tilt function by replacing at least one fisheye image disposed on the basis of the input value input through the user interface unit 80 with another fisheye image.

In addition, the image processing unit 50 can display the movement of the observation object based on the coordinate information and process the synthesized fisheye image.

The display of the movement of the observation object can give a predetermined identification mark to the observation object.

The predetermined identification display described above can be configured so that a predetermined figure is displayed on the observation object.

That is, the identification mark can be processed so as to indicate the position where the observation object is moved in the fisheye image synthesized with the two fisheye images according to the coordinates of the quadrangle tracked by the image tracking unit 30. [

In the video output step S40, the display unit 70 outputs the received image from the image processing unit 50 that synthesizes the two fisheye images.

FIG. 14 is a seventh configuration diagram of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention, FIG. 15 is a diagram illustrating a screen of a panorama image output from a camera unit of the system of FIG. FIG. 16 is a second example of a screen of panorama images output from the camera unit of the system of FIG. 14; FIG.

14 to 16, a seventh configuration of a system for performing an automatic multiple tracking image processing method for a surveillance camera according to an embodiment of the present invention includes a pair of panoramic cameras A video image transmission step S21, an image division transmission step S22, an image processing transmission step S30, and a video output step S40.

Hereinafter, an automatic multi-tracking image processing method for a surveillance camera according to an embodiment of the present invention will be described with reference to FIGS. 1 and 2, and an automatic multiple tracking image processing method for a surveillance camera and a plurality of images (S30) in which a plurality of images are combined, and only a difference method will be described as the principle of image processing is the same.

The photographing transmission step S10 is performed by the camera unit 10 including the panoramic cameras 15, which are installed in pairs in mutually different directions and the images are synthesized 360 degrees in all directions. The camera unit 10 ) To the image tracking unit 30 and the image dividing unit 40, respectively. Here, the camera unit 10 includes a pair of panoramic cameras 15 for photographing a panoramic image having an azimuth angle of 180 degrees in mutually opposite directions. When the panoramic images opposite to each other are synthesized, Since the image having the angle of view can be secured, the image showing the entire azimuth angle can be photographed without the square area at the azimuth angle position, and the monitoring efficiency can be improved by enlarging the surveillance area.

The panorama images photographed by the pair of panoramic cameras 15 installed in mutually opposite directions are transmitted to the image tracing unit 30 and the image dividing unit 40 in the image capturing transmission step S10.

In the image tracking transmission step S21, the coordinates of the moving object to be moved within the two panorama images are respectively tracked and transmitted to the image processing unit 50 by the image tracking unit. At this time, coordinates for displaying the outer circumference of the observation object in a rectangular shape are generated, and coordinates of a rectangular shape are transmitted to the image processing unit 50 so that the monitor can easily identify the displayed rectangular image while being traced with the observation object. In addition, when a plurality of objects to be observed appear in the panoramic images, the coordinates can be extracted while being tracked in multiple, and transmitted to the image processing unit 50.

In the image segmentation transmission step S22, the two panorama images transmitted to the image segmentation unit are divided into a plurality of segments, and the divided panorama images are transmitted to the image processing unit 50. [

The image processing transmission step S30 is a step of transmitting the image processing unit 50 to the image synthesizing unit 90 while maintaining the divided positions of the two divided panorama images transmitted from the image dividing unit 40, And arranges at least one panorama image among the plurality of panorama images divided in the synthesized panorama image, processes the movement or replacement of the panorama image, enlargement and reduction of the image, and applies the virtual pan-tilt and zoom in-zoom-out functions And processing the panoramic image.

In this way, the image processing unit 50 can process the virtual pan-tilt function by changing the arrangement position of at least one panorama image arranged based on the coordinate information.

Also, the image processing unit 50 may process the virtual pan-tilt function by replacing at least one panoramic image arranged based on the coordinate information with another panoramic image.

The image processing unit 50 may process the virtual pan-tilt function by changing the arrangement position of at least one panorama image arranged on the basis of the input values inputted through the user interface unit 80. [ The interface unit 80 can perform the pan, tilt, and zoom functions in the panorama image synthesized by the user and automatically move along the observation object according to the coordinates extracted by the image tracking unit 30. [ The input value can be set to track the input.

In addition, the image processing unit 50 may process a virtual pan-tilt function by replacing at least one panoramic image arranged on the basis of the input value inputted through the user interface unit 80 with another panoramic image.

In addition, the image processing unit 50 can display the movement of the observation object based on the coordinate information and process the combined panorama image.

The display of the movement of the observation object can give a predetermined identification mark to the observation object.

The predetermined identification display described above can be configured so that a predetermined figure is displayed on the observation object.

That is, the identification mark can be processed so as to indicate the position where the observation object is moved in the panorama image obtained by combining the two panorama images according to the coordinates of the quadrangle tracked by the image tracking unit 30.

In the video output step S40, the display unit 70 outputs the received image from the image processing unit 50 that synthesizes the two panorama images.

While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is to be understood that the invention is not limited to the disclosed embodiments, but, on the contrary, And is not intended to limit the scope of the invention. It is to be understood by those skilled in the art that other modifications based on the technical idea of the present invention are possible in addition to the embodiments disclosed herein.

Description of the Related Art
10: camera part 11: photographing camera
12: Fish eye camera 13: Thermal camera
14: fish eye heat sensing camera 15: panoramic camera
20: image checking unit 30: image tracking unit
40: image divider 50: image processor
60: a caption processing unit 70:
80: interface unit 90:

Claims (25)

Transmitting an image of a region to be photographed, which is photographed by a camera unit having a lens having a 180 degree angle of view, to a video tracker and a video segmenter, respectively;
Recognizing an observation object moving in the image transmitted from the image tracking unit, generating coordinate information according to movement of the observation object, and transmitting the coordinate information to the image processing unit;
Dividing the image into a plurality of regions in the image division unit, and transmitting the images of the divided regions to the image processing unit;
Moving, replacing, enlarging and reducing at least one image of the plurality of divided regions received from the image dividing unit in the image processing unit; And
And outputting an image of a region to be photographed photographed by the camera section and an image of a plurality of divided regions processed by the image processing section together through a display section,
Wherein the image tracking unit simultaneously recognizes the at least two objects to be observed when two or more objects to be observed are displayed on the image and generates coordinate information of the at least two objects to be observed and transmits the coordinate information to the image processing unit,
Wherein the image of the region to be photographed is arranged at the center of the display section and the image of the plurality of divided regions is arranged at a divided position around the center of the display section, And simultaneously displaying a plurality of images of the divided regions in the entire image in one display unit,
Wherein the image processing unit realizes a virtual pan, tilt, or zoom function by enlarging or reducing an image of each of the divided regions,
Wherein the image enlarged or reduced by the image processing unit is displayed in a corresponding one of the plurality of divided areas.
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
The method according to claim 1,
Wherein the image processing unit processes the virtual pan-tilt function by changing an arrangement position of the arranged at least one image based on the coordinate information,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
The method according to claim 1,
Wherein the image processing unit processes the virtual pan-tilt function by replacing the arranged at least one image with another image based on the coordinate information,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
The method according to claim 1,
Wherein the image processing unit processes a virtual pan-tilt function by changing an arrangement position of the arranged at least one image based on an input value input through a user interface unit,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
The method according to claim 1,
Wherein the image processing unit processes the virtual pan-tilt function by replacing the at least one image with another image based on the input value input through the user interface unit,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
The method according to claim 1,
The method
And displaying the movement of the observation object based on the coordinate information in the image processing section to process the image.
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
The method according to claim 6,
The display of the movement of the observation object may include displaying the movement of the observation object
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
8. The method of claim 7,
Wherein the predetermined identification display is configured so that a predetermined figure is displayed on the observation object
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
The method according to claim 1,
Wherein the camera unit is a fisheye camera having a fisheye lens and transmits a fisheye image photographed by the fisheye camera to the image tracking unit and the image dividing unit.
delete The method according to claim 1,
Wherein the camera unit is a fisheye thermal sensing camera having a fisheye lens for sensing a thermal change and an automatic multiple tracking for a surveillance camera for transmitting a thermal sensation image photographed by the fisheye thermal sensing camera to the image tracking unit and the image dividing unit, Image processing method.
delete The method according to claim 1,
And transmitting to the display unit at least one of the divided position information of the divided image, the photographed date information, and the photographed position information in the step of transmitting to the display unit
Automatic multiple tracking image processing method.
The method according to claim 1,
Wherein the image checking unit connected to the camera unit checks whether the video signal is present, and transmits a reset signal to the camera unit when the video signal does not exist,
Automatic multiple tracking image processing method.
A step of photographing an image of a region to be photographed with a camera unit including a plurality of cameras including a camera having a lens having an angle of view of 180 degrees, and transmitting the photographed images to the image tracking unit and the image dividing unit;
Recognizing a moving object to be moved in the images transmitted from the image tracking unit, generating coordinate information according to movement of the object, and transmitting the coordinate information to the image processing unit;
Dividing the images into a plurality of regions in the image division unit, and transmitting the images of the divided regions to the image processing unit;
Wherein the image processing unit combines the images of the plurality of divided regions received from the image dividing unit into one image in the image synthesizing unit while maintaining the divided positions and outputs at least one of the synthesized images of the divided regions Moving or replacing, enlarging and reducing the image; And
And outputting an image of a region to be photographed photographed by the camera section and an image of a plurality of divided regions processed by the image processing section together through a display section,
Wherein the image tracking unit simultaneously recognizes the at least two objects to be observed when two or more objects to be observed are displayed on the image and generates coordinate information of the at least two objects to be observed and transmits the coordinate information to the image processing unit,
Wherein the image of the region to be photographed is arranged at the center of the display section and the image of the plurality of divided regions is arranged at a divided position around the center of the display section, And simultaneously displaying a plurality of images of the divided regions in the entire image in one display unit,
Wherein the processing step realizes a virtual pan, tilt or zoom function by enlarging or reducing the image of each divided area in the image processing section,
Wherein the image enlarged or reduced by the image processing unit is displayed in a corresponding one of the plurality of divided areas.
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
Wherein the image processing unit processes the virtual pan-tilt function by changing an arrangement position of the arranged at least one image based on the coordinate information,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
Wherein the image processing unit processes the virtual pan-tilt function by replacing the arranged at least one image with another image based on the coordinate information,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
Wherein the image processing unit processes a virtual pan-tilt function by changing an arrangement position of the arranged at least one image based on an input value input through a user interface unit,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
Wherein the image processing unit processes the virtual pan-tilt function by replacing the at least one image with another image based on the input value input through the user interface unit,
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
The method
And displaying the movement of the observation object based on the coordinate information in the image processing section to process the image.
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
21. The method of claim 20,
The display of the movement of the observation object may include displaying the movement of the observation object
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
22. The method of claim 21,
Wherein the predetermined identification display is configured so that a predetermined figure is displayed on the observation object
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
Wherein the camera unit comprises a fish-eye camera having a lens having an angle of view of 180 degrees to capture an image of a region to be photographed, and a thermal-sensitive camera for sensing thermal changes in the region to be photographed, Sensing image captured by the thermal sensing camera to the image-tracking unit and the image divider
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
Wherein the camera unit comprises: a first fisheye camera which is photographed with an angle of view of 180 degrees and an elevation angle of 180 degrees in a direction in which the photographing target area is photographed; and a second fisheye camera which, when combined with the image of the first fisheye camera, And a second fisheye camera for photographing the opposite direction to the first fisheye camera, wherein the first fisheye image photographed by the first fisheye camera and the second fisheye image photographed by the second fisheye camera, To the image division unit
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
16. The method of claim 15,
The camera unit includes a first panoramic camera that is photographed with an angle of view of 180 degrees in a direction in which the image of the region to be photographed is photographed, and a second panoramic camera that, when combined with the image of the first panoramic camera, The first panorama image captured by the first panorama camera and the second panorama image captured by the second panorama camera are transmitted to the image tracking unit and the image dividing unit
Automatic Multiple Tracking Image Processing Method for Surveillance Cameras.
KR1020150109695A 2015-08-03 2015-08-03 Method for image processing of surveilance camera by using auto multi-tracing KR101619953B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
KR1020150109695A KR101619953B1 (en) 2015-08-03 2015-08-03 Method for image processing of surveilance camera by using auto multi-tracing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
KR1020150109695A KR101619953B1 (en) 2015-08-03 2015-08-03 Method for image processing of surveilance camera by using auto multi-tracing

Publications (1)

Publication Number Publication Date
KR101619953B1 true KR101619953B1 (en) 2016-05-23

Family

ID=56104311

Family Applications (1)

Application Number Title Priority Date Filing Date
KR1020150109695A KR101619953B1 (en) 2015-08-03 2015-08-03 Method for image processing of surveilance camera by using auto multi-tracing

Country Status (1)

Country Link
KR (1) KR101619953B1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017213288A1 (en) * 2016-06-08 2017-12-14 한화테크윈 주식회사 Monitoring apparatus and monitoring system
WO2018038300A1 (en) * 2016-08-24 2018-03-01 한화테크윈 주식회사 Image providing device, method and computer program
KR20240069379A (en) * 2022-11-11 2024-05-20 주식회사 넥스트칩 Analog video signal transmission/reception method and apparatus for performing the method

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010244207A (en) * 2009-04-02 2010-10-28 Univ Of Tokyo Moving object tracking device, moving object tracking method, and moving object tracking program
KR101502448B1 (en) * 2014-09-25 2015-03-13 주식회사 영국전자 Video Surveillance System and Method Having Field of Views of 360 Degrees Horizontally and Vertically

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010244207A (en) * 2009-04-02 2010-10-28 Univ Of Tokyo Moving object tracking device, moving object tracking method, and moving object tracking program
KR101502448B1 (en) * 2014-09-25 2015-03-13 주식회사 영국전자 Video Surveillance System and Method Having Field of Views of 360 Degrees Horizontally and Vertically

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017213288A1 (en) * 2016-06-08 2017-12-14 한화테크윈 주식회사 Monitoring apparatus and monitoring system
KR20170138721A (en) * 2016-06-08 2017-12-18 한화테크윈 주식회사 The Apparatus And The System For Monitoring
CN108476303A (en) * 2016-06-08 2018-08-31 韩华泰科株式会社 Supervision equipment and monitoring system
CN108476303B (en) * 2016-06-08 2021-04-13 韩华泰科株式会社 Monitoring apparatus and monitoring system
US10999486B2 (en) 2016-06-08 2021-05-04 Hanwha Techwin Co., Ltd. Monitoring apparatus and monitoring system
KR102518855B1 (en) 2016-06-08 2023-04-06 한화비전 주식회사 The Apparatus And The System For Monitoring
WO2018038300A1 (en) * 2016-08-24 2018-03-01 한화테크윈 주식회사 Image providing device, method and computer program
US11042970B2 (en) 2016-08-24 2021-06-22 Hanwha Techwin Co., Ltd. Image providing device, method, and computer program
KR20240069379A (en) * 2022-11-11 2024-05-20 주식회사 넥스트칩 Analog video signal transmission/reception method and apparatus for performing the method
KR102706005B1 (en) * 2022-11-11 2024-09-12 주식회사 넥스트칩 Analog video signal transmission/reception method and apparatus for performing the method

Similar Documents

Publication Publication Date Title
US10339386B2 (en) Unusual event detection in wide-angle video (based on moving object trajectories)
US7366359B1 (en) Image processing of regions in a wide angle video camera
KR101530255B1 (en) Cctv system having auto tracking function of moving target
KR101778744B1 (en) Monitoring system through synthesis of multiple camera inputs
JP2007158860A (en) Photographing system, photographing device, image switching device, and data storage device
US20050052533A1 (en) Object tracking method and object tracking apparatus
WO2011082185A1 (en) Confined motion detection for pan-tilt cameras employing motion detection and autonomous motion tracking
KR101982751B1 (en) Video surveillance device with motion path tracking technology using multi camera
WO2009066988A2 (en) Device and method for a surveillance system
JP2016127571A (en) Camera system, display control device, display control method, and program
KR20130130544A (en) Method and system for presenting security image
KR101832274B1 (en) System for crime prevention of intelligent type by video photographing and method for acting thereof
KR20180086048A (en) Camera and imgae processing method thereof
JP2013118712A (en) System for tracking and photographing a plurality of moving objects and photographing control method
JP4309900B2 (en) Video surveillance system
KR101619953B1 (en) Method for image processing of surveilance camera by using auto multi-tracing
KR20110044101A (en) Watching apparatus using dual camera
CN110944101A (en) Image pickup apparatus and image recording method
KR20160094655A (en) The System and Method for Panoramic Video Surveillance with Multiple High-Resolution Video Cameras
KR101619956B1 (en) Apparatus for image processing of surveilance camera by using auto multi-tracing
KR20190026625A (en) Image displaying method, Computer program and Recording medium storing computer program for the same
Lin et al. Large-area, multilayered, and high-resolution visual monitoring using a dual-camera system
KR101410985B1 (en) monitoring system and monitoring apparatus using security camera and monitoring method thereof
JP4444746B2 (en) Object tracking method and object tracking device
KR102634188B1 (en) System for monitoring image

Legal Events

Date Code Title Description
AMND Amendment
AMND Amendment
X701 Decision to grant (after re-examination)
GRNT Written decision to grant
FPAY Annual fee payment

Payment date: 20190503

Year of fee payment: 4