WO2022190470A1 - Image processing device, method therefor, and image processing program - Google Patents

Image processing device, method therefor, and image processing program Download PDF

Info

Publication number
WO2022190470A1
WO2022190470A1 PCT/JP2021/043070 JP2021043070W WO2022190470A1 WO 2022190470 A1 WO2022190470 A1 WO 2022190470A1 JP 2021043070 W JP2021043070 W JP 2021043070W WO 2022190470 A1 WO2022190470 A1 WO 2022190470A1
Authority
WO
WIPO (PCT)
Prior art keywords
region
area
candidate
moving
image processing
Prior art date
Application number
PCT/JP2021/043070
Other languages
French (fr)
Japanese (ja)
Inventor
拓実 會下
卓馬 寺田
光太 星野
朋晟 平岡
Original Assignee
株式会社日立製作所
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社日立製作所 filed Critical 株式会社日立製作所
Publication of WO2022190470A1 publication Critical patent/WO2022190470A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/215Motion-based segmentation

Definitions

  • the present invention relates to an image processing apparatus and method, and an image processing program, and more particularly to a monitoring technique using moving images.
  • Surveillance technology using moving images such as video is used in various fields. For example, in manufacturing sites, for the purpose of quality control, ensuring safety of workers, improving productivity, etc., the introduction of a mechanism for grasping the work situation using images of the work environment is being promoted. In a system that grasps the work situation using images, determining the movement area of the target object as the monitoring area prevents erroneous recognition of objects other than the target object, reduces the amount of data processing, and furthermore, allows multiple target objects to be monitored. It is essential to grasp the work status of each individual.
  • the monitoring area differs depending on the target object in the work environment and the installation status of the imaging device, it is necessary to set the monitoring region of the target object for the image of each imaging device. Therefore, when installing a plurality of imaging devices, it is necessary for an operator to manually determine a monitoring region for all target objects in the images of all the imaging devices.
  • a moving vector of a monitored object extracted from a moving image and boundary line information detected from a background image are used to determine a boundary line.
  • a monitoring device is disclosed that automatically determines a monitoring area by estimating a boundary line of a moving area of a monitoring object moving in a lane having a .
  • Patent Literature 1 The technique described in Patent Literature 1 is based on the premise that a monitoring area for an object such as a vehicle moving in a lane with a boundary line is determined. do.
  • the boundary line does not necessarily exist in the area where the person moves.
  • the belt conveyor may be a boundary line, but the worker does not always work around the belt conveyor, and while moving to another place Work may proceed. In that case, the technique of Patent Document 1 cannot cope with determination of the monitoring area.
  • One possible solution is to cover all areas where people may move (flow line areas) and extract and determine monitoring areas.
  • a monitoring area is extracted from a wide flow line area, the number of candidate monitoring areas may increase.
  • the processing time and storage capacity of the computer increase, and the operating cost of the system increases. Therefore, it is desired to determine a useful monitoring area from the flow line area of moving bodies such as people, that is, to efficiently determine the monitoring area, while keeping costs as low as possible.
  • an object of the present invention is to provide an image processing technique that can efficiently determine a monitoring area based on a moving object's flow line area.
  • a preferable example of the image processing device is a moving area detection unit that detects a plurality of moving areas from moving image data; a movement area storage unit that stores the plurality of movement areas; a basis generator that generates basis vectors based on the plurality of moving regions; a region extraction unit that extracts a candidate region from the base vectors generated by the base generation unit; It is an image processing apparatus having The present invention can also be grasped as an image processing method and an image processing program executed by the image processing apparatus.
  • the present invention it is possible to efficiently determine the monitoring area based on the moving body flow line area.
  • FIG. 1 is a diagram showing a schematic configuration example of an image processing system according to an embodiment
  • FIG. It is a figure which shows the hardware structural example of an image processing apparatus.
  • 2 is a diagram illustrating an example of functional configuration of an image processing apparatus
  • FIG. 3 is a block diagram showing a processing example of an image processing device
  • FIG. 7 is a flow chart showing an example of monitoring area determination processing by an image processing apparatus
  • FIG. 4 is a diagram showing an example of a moving area
  • FIG. 4 is a diagram showing an example of basis vectors and candidate regions
  • 4 is a diagram showing an example of an area correspondence table 145
  • FIG. 14 is a diagram showing an example of an area selection table 146
  • FIG. 10 is a diagram showing an example of a display screen displaying selection results of candidate regions;
  • FIG. 1 shows a schematic configuration example of an image processing system 1 .
  • the image processing system 1 is configured by connecting an imaging device 3 and an image processing device 100 as an information processing device so as to be able to communicate with each other via a wired or wireless communication means 2 .
  • one imaging device 3 is illustrated in FIG. 1, it is preferable that a plurality of imaging devices 3 be installed.
  • a plurality of imaging devices 3 are installed so as to cover a flow line area in which moving objects such as people and machines may move, and output moving image data obtained by photographing working environments of the moving objects.
  • the plurality of imaging devices 3 for example, a camera (digital camera (RGB camera), infrared camera, thermography camera, time of flight (TOF) camera, stereo cameras, etc.).
  • the communication means 2 is, for example, a communication means conforming to various communication standards such as USB (Universal Serial Bus), RS-232C, LAN (Local Area Network), WAN (Wide Area Network), the Internet, a dedicated line, and the like. Other devices such as a mobile terminal may be connected to the communication means 2 .
  • the image processing device 100 performs processing for determining a monitoring area based on moving image data acquired by the imaging device 3 .
  • FIG. 2 shows the hardware configuration of the image processing apparatus 100.
  • the image processing apparatus 100 is an information processing apparatus (computer) and includes a processor 11 , a main memory device 12 , an auxiliary memory device 13 , an input device 14 , an output device 15 and a communication device 16 .
  • the processor 11 is, for example, a device that performs arithmetic processing, such as a CPU (Central Processing Unit), MPU (Micro Processing Unit), GPU (Graphics Processing Unit), AI (Artificial Intelligence) chip, and the like.
  • the main storage device 12 is a device that stores programs and data, and includes, for example, ROM (Read Only Memory) (SRAM (Static Random Access Memory), NVRAM (Non Volatile RAM), mask ROM (Mask Read Only Memory), PROM (Programmable ROM), etc.), RAM (Random Access Memory) (DRAM (Dynamic Random Access Memory), etc.).
  • the auxiliary storage device 13 is a hard disk drive, flash memory, SSD (Solid State Drive), optical storage device (CD (Compact Disc), DVD (Digital Versatile Disc), etc.), etc. . Programs and data stored in the auxiliary storage device 13 are read into the main storage device 12 as needed.
  • the input device 14 is a user interface that receives information from the user, such as a keyboard, mouse, card reader, touch panel, and the like.
  • the output device 15 is a user interface that outputs various information (display output, audio output, print output, etc.). These include an output device (speaker), a printing device, and the like.
  • the communication device 16 is a communication interface that communicates with other devices via the communication means 2, such as a NIC (Network Interface Card), a wireless communication module, a USB (Universal Serial Interface) module, a serial communication module, and the like.
  • the communication device 16 can also function as a device that receives information from other devices such as mobile terminals that are communicably connected.
  • the communication device 16 can also function as a device that transmits information to other devices that are communicatively connected.
  • the image processing device 100 communicates with the imaging device 3 via the communication means 2 by the communication device 16 .
  • the above functions of the image processing apparatus 100 are implemented by the processor 11 reading out and executing programs stored in the main storage device 12 . Note that it may be realized by hardware (FPGA, ASIC, AI chip, etc.) that constitutes the image processing apparatus 100 .
  • FIG. 3 shows an example of the functional configuration of the image processing apparatus 100.
  • the image processing apparatus 100 includes an acquisition unit 110 , an extraction unit 120 , a selection unit 130 and a storage unit 140 .
  • the image processing apparatus 100 may further include functions such as an operating system, a device driver, a file system, and a DBMS (DataBase Management System).
  • DBMS DataBase Management System
  • the storage unit 140 stores a movement area storage unit 141, moving image data 142, base vectors 143, candidate areas 144, area correspondence table 145, and area selection table 146.
  • the storage unit 140 stores such information as, for example, a database table provided by a DBMS or a file provided by a file system.
  • the acquisition unit 110 receives and acquires data transmitted from other devices.
  • Acquisition unit 110 includes data acquisition unit 111 and moving region detection unit 112 .
  • the data acquisition unit 111 acquires moving image data 142 transmitted from the imaging device 3 .
  • the movement area detection unit 112 detects a movement area from the moving image data 142 and stores the obtained movement area in the movement area storage unit 141 .
  • the extraction unit 120 extracts candidate areas based on a plurality of moving areas stored in the moving area storage unit 141 .
  • the extractor includes a base generator 121 and a region extractor 122 .
  • the base generation unit 121 generates a plurality of base vectors 143 based on the plurality of motion regions stored in the motion region storage unit 141 .
  • a region extraction unit 122 extracts a candidate region 144 from the basis vectors 143 .
  • the selection unit 130 selects the candidate area 144 and outputs it as a monitoring area.
  • the selection unit includes an area integration unit 131 and an area selection unit 132 .
  • the area integration unit 131 integrates the overlapping candidate areas 144 .
  • the region selection unit 132 selects candidate regions 144 based on the region selection table 146 .
  • FIG. 4 is a block diagram showing an example of processing performed by the image processing apparatus 100 when determining a monitoring area of a moving image of a working environment.
  • the moving image data 142' of the working environment captured by each of the plurality of imaging devices 3 is acquired.
  • the moving image data 142' is a moving image of a plurality of imaging devices installed in the work environment capturing the working of target objects to be monitored, such as workers, machine tools, and other manufacturing equipment. .
  • the data acquisition unit 111 of the acquisition unit 110 acquires the moving image data 142 ′ transmitted from the imaging device 3 and stores it in the moving image data 142 of the storage unit 140 . Further, the movement area detection unit 112 detects a movement area from the moving image data 142 ′ and stores the detected movement area in the movement area storage unit 141 . For example, the movement area detection unit 112 detects an area in which movement occurs between a plurality of frames as a movement area by optical flow or background subtraction.
  • the base generation unit 121 of the extraction unit 120 generates one or more base vectors 143 based on the plurality of moving regions stored in the moving region storage unit 141 . Then, the area extracting unit 122 extracts the candidate area 144 from the base vector 143 . For example, the base generation unit 121 generates one or more eigenvectors obtained by principal component analysis of a plurality of movement regions as the base vectors 143, and the region extraction unit 122 extracts predetermined A set of elements exceeding the value is extracted as a candidate region 144 .
  • the area integration unit 131 of the selection unit 130 integrates the overlapping candidate areas 144 . Then, the area selection unit 132 selects candidate areas 144 based on the area selection table 146 .
  • the image processing apparatus 100 determines the statistically extracted area as the monitoring area based on one or more basis vectors generated from the plurality of moving areas included in the moving image data 142. be able to. Therefore, even for an unknown target object, it becomes possible to determine the monitoring region from the moving region of the target object, and efficiency in determining the monitoring region can be achieved.
  • FIG. 5 is a flowchart showing monitoring area determination processing by the image processing apparatus 100 .
  • the data acquisition unit 111 acquires a plurality of moving image data 142' transmitted from the plurality of imaging devices 3 and stores them in the moving image data 142 (S510).
  • the movement area detection unit 112 detects a movement area from the moving image data 142 (S520), and stores the detected movement area in the movement area storage unit 141 (S530).
  • the movement area detection unit 112 detects an area that has moved between a plurality of frames as a movement area by optical flow, background subtraction, or the like.
  • FIG. 6 shows an example of an image included in the moving image data 142 and an example of a moving area detected from the moving image including the image.
  • FIG. 6(1) shows a worker G10-1 moving rightward on the image G1 and a robot arm G10-2 moving leftward on the image G1.
  • FIG. 6(2) shows the movement area G2 detected from the moving image including the image G1. 2 is detected.
  • the number of moving regions G20 to be detected is not limited, and a plurality of moving regions G20 do not need to be distinguished because they overlap.
  • the moving region detection unit 112 sets “1” to each pixel of the moving regions G20-1 and G20-2, and sets “0” to each pixel of the other regions.
  • the pixel value set for each pixel in the moving region G20-1 and the moving region G20-2 may be other than "1".
  • a value indicating the amount of movement of each pixel calculated by optical flow from one or more pixels corresponding to each pixel may be set as the pixel value.
  • the base generation unit 121 generates a plurality of basis vectors 143 (M1, M2, . . . , Mk (k is a predetermined integer of 1 or more )) is generated (S540). Specifically, for example, the basis generation unit 121 generates one or more eigenvectors obtained by principal component analysis of a plurality of moving regions as basis vectors (M1, M2, . . . , Mk).
  • FIG. 7 shows an eigenvector (base vector M1) of the first principal component obtained by principal component analysis of a plurality of moving regions, as an example of the base vector.
  • the value of k is generated by principal component analysis in the order of the eigenvector of the first principal component (base vector M1) and the eigenvector of the second principal component (base vector M2). may be up to the eigenvector of the k-th principal component (basis vector Mk).
  • the basis generation unit 121 may use a plurality of basis obtained by independent component analysis of a plurality of moving regions as a basis vector. Alternatively, weights of an autoencoder that match the input movement area and the output movement area may be used as basis vectors.
  • the region extracting unit 122 extracts candidate regions 144 (R1, R2, ..., Rn (n is an integer equal to or greater than 1)) from the basis vectors 143 (M1, M2, ..., Mk) (S550). . Specifically, for example, the region extracting unit 122 extracts a set of elements having a positive value exceeding a predetermined value and a set of elements having a negative value exceeding a predetermined value among the elements of the base vector M1. are extracted as a candidate region R1 for the base vector M1 and a candidate region R2 for the base vector M1, respectively. Similar operations are performed on basis vectors M1, M2, ..., Mk to obtain candidate regions R1, R2, ..., Rn.
  • FIG. 7 shows a candidate region R1 and a candidate region R2 extracted from the base vector M1 as an example of candidate regions.
  • a region O1 of a set of elements having positive values exceeding a predetermined value among the elements of the base vector M1 is extracted in the candidate region R1 of the base vector M1.
  • a region O2 of a set of elements having a negative value exceeding a predetermined value among the elements of the base vector M1 is extracted in the candidate region R1 and the candidate region R2, it is assumed that each pixel of the region O1 and the region O2 is set to "1" and each pixel of the other region is set to "0".
  • a pixel value other than "1" may be set to the pixel.
  • the region extracting unit 122 may divide the region O into a plurality of regions using a region division method for the candidate region R of the base vector M, and use each of the divided regions as the candidate region of the base vector M.
  • the candidate region R1 of the base vector M1 is divided into multiple regions (O11, O12, ..., O1p) using a region division method such as the Watershed method, and the divided regions (O11, O12 , ..., O1p) may be used as candidate regions for basis vectors M1.
  • FIG. 8 shows an example of the area correspondence table 145.
  • the region correspondence table 145 manages candidate region names and corresponding base vectors.
  • the candidate area with "1" set in the column of the basis vector Mi indicates that the candidate area is extracted from the basis vector Mi.
  • candidate regions R1 and R2 have "1" set in the column of base vector M1, so candidate regions R1 and R2 are candidate regions extracted from base vector M1.
  • the region correspondence table 145 is updated each time the region extraction unit 122 extracts a candidate region or each time the region integration unit 131 integrates the candidate regions.
  • the area integrating section 131 integrates the overlapping candidate areas 144 based on the area correspondence table 145 (S560).
  • the region integration unit 131 first refers to the region correspondence table 145 and selects one candidate region Ri from the candidate regions (R1, R2, . . . , Rn). Next, referring to the region correspondence table 145, a candidate region Rj that overlaps with the candidate region Ri is selected from candidate regions extracted from base vectors different from the candidate region Ri. Specifically, for example, using an index representing the similarity of sets such as the Jaccard coefficient or the Simpson coefficient, the pixel set included in the region Oi of the candidate region Ri and the pixel set included in the region Oj of the candidate region Rj are calculated. A degree of similarity is calculated, and if the calculated degree of similarity exceeds a predetermined value, it is determined that the candidate regions Ri and Rj overlap.
  • an index representing the similarity of sets such as the Jaccard coefficient or the Simpson coefficient
  • the region integration unit 131 compares the areas of the candidate region Ri and the candidate region Rj, and selects a candidate region with a small area or a candidate region with a large area.
  • the candidate regions Ri and Rj are integrated by deleting them from the correspondence table 145 .
  • the integration method is not limited to the method of deleting candidate regions with small or large areas.
  • the vectors may be compared, and candidate regions extracted from eigenvectors with low contribution rates may be deleted from the region correspondence table 145 .
  • the basis vectors are principal component analysis, it is possible to narrow down to a small number of candidate regions by selecting the eigenvectors (basis vectors) of the upper principal components. Furthermore, by executing the area integration process, it is possible to narrow down to a smaller number of candidate areas. As a result, it is possible to reduce the amount of calculation and the processing time required for analyzing the image data.
  • the area selection unit 132 selects the candidate area 144 based on the area selection table 146 (S570), outputs the selected candidate area 144 as the monitoring area (S580), and performs the process of determining the monitoring area. finish.
  • FIG. 9 shows an example of the area selection table 146.
  • the area selection table 146 manages candidate area names and corresponding selection results.
  • a candidate area for which "1" is set in the selection result column of the area selection table 146 indicates that it is a selected candidate area. For example, since "1" is set in the selection result column for candidate regions R1 and R2, candidate regions R1 and R2 are selected candidate regions.
  • FIG. 10 is a diagram showing an example of a display screen displaying selection results of candidate regions.
  • the candidate area selected by the area selection unit 132 is displayed on the output device 15 of the image processing apparatus 100 as a monitoring area.
  • the selection result may be displayed on a display of a terminal (not shown) connected via the communication device 16.
  • a display screen 1000 includes a monitoring area result table 1001 and a monitoring area display 1002 .
  • the monitoring area result table 1001 is data displaying the contents of the area correspondence table 145 and the area selection table 146 .
  • the candidate areas for which "1" is set in the selection result column of the area selection table 146 are displayed together with the rectangle surrounding the candidate area and the name of the candidate area superimposed.
  • the candidate region R1 and the candidate region R2 are the rectangles surrounding each candidate region and each candidate region. displayed with the first name. The user can recognize which candidate area has been selected and how close or overlapping the selected areas R1 and R2 are by looking at the displays 1001 and 1002 on the display screen.
  • statistically extracted regions are determined as monitoring regions based on one or more base vectors generated from a plurality of moving regions in the moving image data 142. can be done. Therefore, even for an unknown target object, it becomes possible to determine the monitoring region from the moving region of the target object, and efficiency in determining the monitoring region can be achieved.
  • the present invention is not limited to the above-described embodiments, and can be modified in various ways without departing from the gist of the present invention.
  • the above embodiments have been described in detail in order to explain the present invention in an easy-to-understand manner, and are not necessarily limited to those having all the described configurations.
  • moving image data acquired by the imaging device 3 is processed.
  • the moving image data may be processed in real time, or the moving image data acquired and stored in the past may be processed afterward.
  • moving image data generated for simulation by the processor 11 or another processing device may be processed as well as the moving image data acquired from the imaging device 3 .
  • each of the above configurations, functional units, processing units, processing means, etc. may be implemented in hardware, for example, by designing a part or all of them using an integrated circuit.
  • each configuration, function, etc. described above may be implemented by a program that causes the processor to implement each function.
  • Information such as programs, tables, and files that implement each function can be stored in recording devices such as memories, hard disks, SSDs (Solid State Drives), and recording media such as IC cards, SD cards, and DVDs.
  • the configuration of the database that stores the various data described above can be changed as appropriate from the viewpoint of efficient use of resources, improvement of processing efficiency, improvement of access efficiency, improvement of search efficiency, etc.
  • image processing system 2 communication means 3: imaging device 11: processor 12: main storage device 13: auxiliary storage device 14: input device 15: output device 16: communication device 100: image processing device 110: acquisition unit 111: data Acquisition unit 112: moving region detection unit 120: extraction unit 121: base generation unit 122: region extraction unit 130: selection unit 131: region integration unit 132: region selection unit 140: storage unit 141: movement region storage unit 142: moving image Data 143: Base vector 144: Candidate area 145: Area correspondence table 146: Area selection table

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The present invention efficiently determines a monitoring region based on the movement line region of a moving body. This image processing device has: a movement region detection unit 112 that detects a plurality of movement regions from moving image data; a movement region storage unit 141 that stores the plurality of movement regions; a basis generation unit 121 that generates a basis vector based on the plurality of movement regions; and a region extraction unit 122 that extracts a candidate region from the basis vector generated by the basis generation unit.

Description

画像処理装置及びその方法、並びに画像処理プログラムIMAGE PROCESSING DEVICE AND METHOD, AND IMAGE PROCESSING PROGRAM
 本発明は、画像処理装置及びその方法、並びに画像処理プログラムに係り、特に動画像を用いた監視技術に関する。 The present invention relates to an image processing apparatus and method, and an image processing program, and more particularly to a monitoring technique using moving images.
 映像等の動画像を用いた監視技術が種々の分野に利用されている。例えば、製造現場においては、品質管理や作業者の安全確保、生産性向上等を目的として、作業環境を撮影した映像を用いて作業状況を把握しようとする仕組みの導入が進められている。映像を用いて作業状況を把握する仕組みにおいて、対象物体の移動領域を監視領域に決定することは、対象物体以外の物体の誤認識を予防し、データ処理量を削減し、さらに複数の対象物体の作業状況を個別に把握するために必要不可欠である。 Surveillance technology using moving images such as video is used in various fields. For example, in manufacturing sites, for the purpose of quality control, ensuring safety of workers, improving productivity, etc., the introduction of a mechanism for grasping the work situation using images of the work environment is being promoted. In a system that grasps the work situation using images, determining the movement area of the target object as the monitoring area prevents erroneous recognition of objects other than the target object, reduces the amount of data processing, and furthermore, allows multiple target objects to be monitored. It is essential to grasp the work status of each individual.
 監視領域は、作業環境の対象物体や撮像装置の設置状況によって異なるため、個々の撮像装置の映像に対して、対象物体の監視領域を設定する必要がある。このため、複数の撮像装置を設置する場合、全ての撮像装置の映像の全ての対象物体に対して作業員が手動で監視領域を決定する必要がある。 Since the monitoring area differs depending on the target object in the work environment and the installation status of the imaging device, it is necessary to set the monitoring region of the target object for the image of each imaging device. Therefore, when installing a plurality of imaging devices, it is necessary for an operator to manually determine a monitoring region for all target objects in the images of all the imaging devices.
 動画像を用いた監視領域の決定に関して、例えば、特許文献1では、動画像から抽出された監視対象物の移動ベクトルと、背景画像から検出された境界線の情報とを用いることで、境界線を持つレーン内を移動する監視対象物の移動領域の境界線を推測し、監視領域を自動的に決定する監視装置が開示されている。 Regarding determination of a monitoring area using a moving image, for example, in Patent Document 1, a moving vector of a monitored object extracted from a moving image and boundary line information detected from a background image are used to determine a boundary line. A monitoring device is disclosed that automatically determines a monitoring area by estimating a boundary line of a moving area of a monitoring object moving in a lane having a .
特開2020-49911号公報JP 2020-49911 A
 特許文献1に記載の技術は、境界線を持つレーン内を移動する車など物体の監視領域を決定することを前提としているので、背景画像には境界線のような基準となるものを必要とする。然しながら、人が移動する領域から監視領域を決定する場合、人の移動領域には必ずしも境界線が存在するとは限らない。例えば、工場のベルトコンベアを流れる製品を扱う作業者の場合、ベルトコンベアは境界線となるかも知れないが、作業者はベルトコンベアの周辺で常に作業するとは限らず、他の場所へ移動しながら作業を進めることがある。その場合、特許文献1の技術は監視領域の決定に対応できない。 The technique described in Patent Literature 1 is based on the premise that a monitoring area for an object such as a vehicle moving in a lane with a boundary line is determined. do. However, when the monitoring area is determined from the area where the person moves, the boundary line does not necessarily exist in the area where the person moves. For example, in the case of a worker handling products flowing on a conveyor belt in a factory, the belt conveyor may be a boundary line, but the worker does not always work around the belt conveyor, and while moving to another place Work may proceed. In that case, the technique of Patent Document 1 cannot cope with determination of the monitoring area.
 その一策として、人が移動する可能性のある全ての領域(動線領域)をカバーして、監視領域を抽出、決定することが考えられる。しかし、広域な動線領域から監視領域を抽出するとなると、候補となる監視領域が増える可能性がある。その結果、多くの監視領域を抽出、解析するために、コンピュータの処理時間や記憶容量が増加して、システム上の運用コストが増える。そこで、コストを出来るだけ抑えて、人などの移動体の動線領域から有用な監視領域を決定する、すなわち監視領域を効率的に決定することが望まれる。 One possible solution is to cover all areas where people may move (flow line areas) and extract and determine monitoring areas. However, if a monitoring area is extracted from a wide flow line area, the number of candidate monitoring areas may increase. As a result, in order to extract and analyze many monitored areas, the processing time and storage capacity of the computer increase, and the operating cost of the system increases. Therefore, it is desired to determine a useful monitoring area from the flow line area of moving bodies such as people, that is, to efficiently determine the monitoring area, while keeping costs as low as possible.
 そこで、本発明の目的は、移動体の動線領域を基に監視領域を効率的に決定することが可能な画像処理技術を提供することにある。 Therefore, an object of the present invention is to provide an image processing technique that can efficiently determine a monitoring area based on a moving object's flow line area.
 本発明に係る画像処理装置の好ましい例は、
動画像データから複数の移動領域を検出する移動領域検出部と、
前記複数の移動領域を記憶する移動領域記憶部と、
前記複数の移動領域に基づいて基底ベクトルを生成する基底生成部と、
前記基底生成部により生成された前記基底ベクトルから候補領域を抽出する領域抽出部と、
を有する画像処理装置である。
  本発明はまた、上記画像処理装置により実行される画像処理方法、および画像処理プログラムとしても把握される。
A preferable example of the image processing device according to the present invention is
a moving area detection unit that detects a plurality of moving areas from moving image data;
a movement area storage unit that stores the plurality of movement areas;
a basis generator that generates basis vectors based on the plurality of moving regions;
a region extraction unit that extracts a candidate region from the base vectors generated by the base generation unit;
It is an image processing apparatus having
The present invention can also be grasped as an image processing method and an image processing program executed by the image processing apparatus.
 本発明によれば、移動体の動線領域を基に監視領域を効率的に決定することが可能となる。 According to the present invention, it is possible to efficiently determine the monitoring area based on the moving body flow line area.
一実施例に係る画像処理システムの概略構成例を示す図である。1 is a diagram showing a schematic configuration example of an image processing system according to an embodiment; FIG. 画像処理装置のハードウェア構成例を示す図である。It is a figure which shows the hardware structural example of an image processing apparatus. 画像処理装置の機能構成例を示す図である。2 is a diagram illustrating an example of functional configuration of an image processing apparatus; FIG. 画像処理装置の処理例を示すブロック図である。3 is a block diagram showing a processing example of an image processing device; FIG. 画像処理装置による監視領域の決定処理例を示すフローチャートである。7 is a flow chart showing an example of monitoring area determination processing by an image processing apparatus; 移動領域の一例を示す図である。FIG. 4 is a diagram showing an example of a moving area; 基底ベクトルと候補領域の一例を示す図である。FIG. 4 is a diagram showing an example of basis vectors and candidate regions; 領域対応表145の一例を示す図である。4 is a diagram showing an example of an area correspondence table 145; FIG. 領域選定表146の一例を示す図である。FIG. 14 is a diagram showing an example of an area selection table 146; FIG. 候補領域の選定結果を表示する表示画面の一例を示す図である。FIG. 10 is a diagram showing an example of a display screen displaying selection results of candidate regions;
 以下、図面を参照して好ましい実施形態を説明する。以下の説明において、同一のまたは類似する構成に同一の符号を付して重複した説明を省略することがある。また以下の説明において、例示や注釈、または同種の構成を区別する場合、括弧書きで、例示や注釈、識別子(数字、アルファベット等)を表記することがある。 A preferred embodiment will be described below with reference to the drawings. In the following description, the same reference numerals may be given to the same or similar configurations, and redundant description may be omitted. Further, in the following description, examples, annotations, or identifiers (numbers, alphabets, etc.) may be indicated in parentheses when distinguishing between examples, annotations, or similar configurations.
 <画像処理システム1の構成例>
 図1は、画像処理システム1の概略構成例を示す。画像処理システム1は、撮像装置3と、情報処理装置である画像処理装置100が、有線又は無線の通信手段2を介して通信可能に接続して構成される。なお、図1には、1つの撮像装置3が図示されているが、複数の撮像装置3が設置されるのが好ましい。複数の撮像装置3は、人や機械等の移動体が移動する可能性のある動線領域をカバーするように設置されて、移動体の作業環境を撮影した動画像データを出力する。複数の撮像装置3としては、例えば、動画や静止画の画像データを取得(撮影)するカメラ(デジタルカメラ(RGBカメラ)、赤外線カメラ、サーモグラフィカメラ、タイムオブフライト(TOF: Time Of Flight)カメラ、ステレオカメラ等)がある。
<Configuration example of image processing system 1>
FIG. 1 shows a schematic configuration example of an image processing system 1 . The image processing system 1 is configured by connecting an imaging device 3 and an image processing device 100 as an information processing device so as to be able to communicate with each other via a wired or wireless communication means 2 . Although one imaging device 3 is illustrated in FIG. 1, it is preferable that a plurality of imaging devices 3 be installed. A plurality of imaging devices 3 are installed so as to cover a flow line area in which moving objects such as people and machines may move, and output moving image data obtained by photographing working environments of the moving objects. As the plurality of imaging devices 3, for example, a camera (digital camera (RGB camera), infrared camera, thermography camera, time of flight (TOF) camera, stereo cameras, etc.).
 通信手段2は、例えば、USB(Universal Serial Bus)、RS-232C等の各種通信規格に準拠した通信手段、LAN(Local Area Network)、WAN(Wide Area Network)、インターネット、専用線等である。通信手段2には携帯端末等の他装置が接続されてよい。画像処理装置100は、撮像装置3によって取得される動画像データに基づき、監視領域を決定する処理を行う。 The communication means 2 is, for example, a communication means conforming to various communication standards such as USB (Universal Serial Bus), RS-232C, LAN (Local Area Network), WAN (Wide Area Network), the Internet, a dedicated line, and the like. Other devices such as a mobile terminal may be connected to the communication means 2 . The image processing device 100 performs processing for determining a monitoring area based on moving image data acquired by the imaging device 3 .
 <画像処理装置100のハードウェア構成例>
 図2は画像処理装置100のハードウェア構成を示す。
  画像処理装置100は、情報処理装置(コンピュータ)であり、プロセッサ11、主記憶装置12、補助記憶装置13、入力装置14、出力装置15、及び通信装置16を備える。
<Hardware Configuration Example of Image Processing Apparatus 100>
FIG. 2 shows the hardware configuration of the image processing apparatus 100. As shown in FIG.
The image processing apparatus 100 is an information processing apparatus (computer) and includes a processor 11 , a main memory device 12 , an auxiliary memory device 13 , an input device 14 , an output device 15 and a communication device 16 .
 プロセッサ11は、例えば演算処理を行う装置であり、CPU(Central Processing Unit)、MPU(Micro Processing Unit)、GPU(Graphics Processing Unit)、AI(Artificial Intelligence)チップ等である。主記憶装置12は、プログラムやデータを記憶する装置であり、例えば、ROM(Read Only Memory)(SRAM(Static Random Access Memory)、NVRAM(Non Volatile RAM)、マスクROM(Mask Read Only Memory)、PROM(Programmable ROM)等)、RAM(Random Access Memory)(DRAM(Dynamic Random Access Memory)等)等である。補助記憶装置13は、ハードディスクドライブ(Hard Disk Drive)、フラッシュメモリ(Flash Memory)、SSD(Solid State Drive)、光学式記憶装置(CD(Compact Disc)、DVD(Digital Versatile Disc)等)等である。補助記憶装置13に格納されているプログラムやデータは、主記憶装置12に随時読み込まれる。 The processor 11 is, for example, a device that performs arithmetic processing, such as a CPU (Central Processing Unit), MPU (Micro Processing Unit), GPU (Graphics Processing Unit), AI (Artificial Intelligence) chip, and the like. The main storage device 12 is a device that stores programs and data, and includes, for example, ROM (Read Only Memory) (SRAM (Static Random Access Memory), NVRAM (Non Volatile RAM), mask ROM (Mask Read Only Memory), PROM (Programmable ROM), etc.), RAM (Random Access Memory) (DRAM (Dynamic Random Access Memory), etc.). The auxiliary storage device 13 is a hard disk drive, flash memory, SSD (Solid State Drive), optical storage device (CD (Compact Disc), DVD (Digital Versatile Disc), etc.), etc. . Programs and data stored in the auxiliary storage device 13 are read into the main storage device 12 as needed.
 入力装置14は、ユーザから情報を受付けるユーザインタフェースであり、例えば、キーボード、マウス、カードリーダ、タッチパネル等である。出力装置15は、各種の情報を出力(表示出力、音声出力、印字出力等)するユーザインタフェースであり、例えば、各種情報を可視化する表示装置(LCD(Liquid Crystal Display)、グラフィックカード等)や音声出力装置(スピーカ)、印字装置等である。 The input device 14 is a user interface that receives information from the user, such as a keyboard, mouse, card reader, touch panel, and the like. The output device 15 is a user interface that outputs various information (display output, audio output, print output, etc.). These include an output device (speaker), a printing device, and the like.
 通信装置16は、通信手段2を介して他の装置と通信する通信インタフェースであり、例えば、NIC(Network Interface Card)、無線通信モジュール、USB(Universal Serial Interface)モジュール、シリアル通信モジュール等である。通信装置16は、通信可能に接続される携帯端末等の他装置から情報を受信する装置として機能することもできる。また通信装置16は、通信可能に接続する他の装置に情報を送信する装置として機能することもできる。画像処理装置100は、通信装置16により通信手段2を介して撮像装置3と通信する。 The communication device 16 is a communication interface that communicates with other devices via the communication means 2, such as a NIC (Network Interface Card), a wireless communication module, a USB (Universal Serial Interface) module, a serial communication module, and the like. The communication device 16 can also function as a device that receives information from other devices such as mobile terminals that are communicably connected. The communication device 16 can also function as a device that transmits information to other devices that are communicatively connected. The image processing device 100 communicates with the imaging device 3 via the communication means 2 by the communication device 16 .
 画像処理装置100が有する上記の各機能は、プロセッサ11が、主記憶装置12に格納されているプログラムを読み出して実行することで実現される。なお、画像処理装置100を構成しているハードウェア(FPGA、ASIC、AIチップ等)により実現されてもよい。 The above functions of the image processing apparatus 100 are implemented by the processor 11 reading out and executing programs stored in the main storage device 12 . Note that it may be realized by hardware (FPGA, ASIC, AI chip, etc.) that constitutes the image processing apparatus 100 .
 <画像処理装置100の機能構成例>
 図3は画像処理装置100の機能構成例を示す。
  画像処理装置100は、取得部110、抽出部120、選定部130、及び記憶部140を備える。尚、画像処理装置100は、上記の機能に加えて、例えば、オペレーティングシステム、デバイスドライバ、ファイルシステム、DBMS(DataBase Management System)等の機能を更に備えていてもよい。
<Functional Configuration Example of Image Processing Apparatus 100>
FIG. 3 shows an example of the functional configuration of the image processing apparatus 100. As shown in FIG.
The image processing apparatus 100 includes an acquisition unit 110 , an extraction unit 120 , a selection unit 130 and a storage unit 140 . In addition to the functions described above, the image processing apparatus 100 may further include functions such as an operating system, a device driver, a file system, and a DBMS (DataBase Management System).
 ここで、記憶部140は、移動領域記憶部141、動画像データ142、基底ベクトル143、候補領域144、領域対応表145、領域選定表146を記憶する。尚、記憶部140は、例えば、DBMSが提供するデータベースのテーブルや、ファイルシステムが提供するファイルとしてこれらの情報を記憶する。 Here, the storage unit 140 stores a movement area storage unit 141, moving image data 142, base vectors 143, candidate areas 144, area correspondence table 145, and area selection table 146. The storage unit 140 stores such information as, for example, a database table provided by a DBMS or a file provided by a file system.
 取得部110は、他の装置から送信されてくるデータを受信して取得する。取得部110は、データ取得部111と、移動領域検出部112を含む。データ取得部111は、撮像装置3から送信される動画像データ142を取得する。移動領域検出部112は動画像データ142から移動領域を検出し、得られた移動領域を移動領域記憶部141に記憶する。 The acquisition unit 110 receives and acquires data transmitted from other devices. Acquisition unit 110 includes data acquisition unit 111 and moving region detection unit 112 . The data acquisition unit 111 acquires moving image data 142 transmitted from the imaging device 3 . The movement area detection unit 112 detects a movement area from the moving image data 142 and stores the obtained movement area in the movement area storage unit 141 .
 抽出部120は、移動領域記憶部141に記憶された複数の移動領域に基づき、候補領域を抽出する。抽出部は、基底生成部121と、領域抽出部122を含む。基底生成部121は、移動領域記憶部141に記憶された複数の移動領域に基づき、複数の基底ベクトル143を生成する。領域抽出部122は基底ベクトル143から候補領域144を抽出する。 The extraction unit 120 extracts candidate areas based on a plurality of moving areas stored in the moving area storage unit 141 . The extractor includes a base generator 121 and a region extractor 122 . The base generation unit 121 generates a plurality of base vectors 143 based on the plurality of motion regions stored in the motion region storage unit 141 . A region extraction unit 122 extracts a candidate region 144 from the basis vectors 143 .
 選定部130は、候補領域144を選定し、監視領域として出力する。選定部は、領域統合部131と、領域選定部132を含む。領域統合部131は互いに重複する候補領域144を統合する。領域選定部132は、領域選定表146に基づき候補領域144を選定する。 The selection unit 130 selects the candidate area 144 and outputs it as a monitoring area. The selection unit includes an area integration unit 131 and an area selection unit 132 . The area integration unit 131 integrates the overlapping candidate areas 144 . The region selection unit 132 selects candidate regions 144 based on the region selection table 146 .
 <画像処理装置の処理>
 図4は、作業環境を撮影した動画像の監視領域を決定する場合の、画像処理装置100が行う処理例を示すブロック図である。
<Processing of image processing device>
FIG. 4 is a block diagram showing an example of processing performed by the image processing apparatus 100 when determining a monitoring area of a moving image of a working environment.
 (1)まず、複数の各撮像装置3により撮影された作業環境の動画像データ142´を取得する。例えば、動画像データ142´は、作業環境に設置された複数の撮像装置が、作業者や工作機械、その他製造設備などの監視の対象となる対象物体が作業する様子を撮影した動画像である。 (1) First, the moving image data 142' of the working environment captured by each of the plurality of imaging devices 3 is acquired. For example, the moving image data 142' is a moving image of a plurality of imaging devices installed in the work environment capturing the working of target objects to be monitored, such as workers, machine tools, and other manufacturing equipment. .
 (2)取得部110のデータ取得部111が、撮像装置3から送信された動画像データ142´を取得し、記憶部140の動画像データ142に記憶する。また、移動領域検出部112が動画像データ142´から移動領域を検出し、検出された移動領域を移動領域記憶部141に記憶する。
  例えば、移動領域検出部112は、オプティカルフローまたは背景差分法によって、複数のフレーム間で移動が生じた領域を移動領域として検出する。
(2) The data acquisition unit 111 of the acquisition unit 110 acquires the moving image data 142 ′ transmitted from the imaging device 3 and stores it in the moving image data 142 of the storage unit 140 . Further, the movement area detection unit 112 detects a movement area from the moving image data 142 ′ and stores the detected movement area in the movement area storage unit 141 .
For example, the movement area detection unit 112 detects an area in which movement occurs between a plurality of frames as a movement area by optical flow or background subtraction.
 (3)抽出部120の基底生成部121が、移動領域記憶部141に記憶された、複数の移動領域に基づき、1つ以上の基底ベクトル143を生成する。そして、領域抽出部122が、基底ベクトル143から候補領域144を抽出する。
  例えば、基底生成部121は、複数の移動領域の主成分分析により得られる、1つ以上の固有ベクトルを基底ベクトル143として生成し、領域抽出部122は、基底ベクトル143の要素の中で、所定の値を超える要素の集合を候補領域144として抽出する。
(3) The base generation unit 121 of the extraction unit 120 generates one or more base vectors 143 based on the plurality of moving regions stored in the moving region storage unit 141 . Then, the area extracting unit 122 extracts the candidate area 144 from the base vector 143 .
For example, the base generation unit 121 generates one or more eigenvectors obtained by principal component analysis of a plurality of movement regions as the base vectors 143, and the region extraction unit 122 extracts predetermined A set of elements exceeding the value is extracted as a candidate region 144 .
 (4)選定部130の領域統合部131が、互いに重複する候補領域144を統合する。そして、領域選定部132が、領域選定表146に基づき候補領域144を選定する。 (4) The area integration unit 131 of the selection unit 130 integrates the overlapping candidate areas 144 . Then, the area selection unit 132 selects candidate areas 144 based on the area selection table 146 .
 (5)これにより、画像処理装置100は、動画像データ142に含まれる複数の移動領域から生成される1つ以上の基底ベクトルに基づき、統計的に抽出された領域を、監視領域として決定することができる。したがって、未知の対象物体に対しても、その対象物体の移動領域から監視領域を決定することが可能になり、監視領域の決定の効率化を図ることができる。 (5) Thereby, the image processing apparatus 100 determines the statistically extracted area as the monitoring area based on one or more basis vectors generated from the plurality of moving areas included in the moving image data 142. be able to. Therefore, even for an unknown target object, it becomes possible to determine the monitoring region from the moving region of the target object, and efficiency in determining the monitoring region can be achieved.
 <監視領域の決定>
 図5は、画像処理装置100による監視領域の決定処理を示すフローチャートである。
  画像処理装置100において、まず、データ取得部111が、複数の撮像装置3から送信される複数の動画像データ142´を取得して、動画像データ142に格納する(S510)。そして、移動領域検出部112が、動画像データ142から移動領域を検出して(S520)、検出された移動領域を移動領域記憶部141に格納する(S530)。
<Determination of monitoring area>
FIG. 5 is a flowchart showing monitoring area determination processing by the image processing apparatus 100 .
In the image processing apparatus 100, first, the data acquisition unit 111 acquires a plurality of moving image data 142' transmitted from the plurality of imaging devices 3 and stores them in the moving image data 142 (S510). Then, the movement area detection unit 112 detects a movement area from the moving image data 142 (S520), and stores the detected movement area in the movement area storage unit 141 (S530).
 具体的には、例えば、移動領域検出部112は、オプティカルフローや背景差分法などによって、複数のフレーム間で移動が生じた領域を移動領域として検出する。 Specifically, for example, the movement area detection unit 112 detects an area that has moved between a plurality of frames as a movement area by optical flow, background subtraction, or the like.
 図6は、動画像データ142に含まれる画像の例およびその画像を含む動画像から検出された移動領域の例を示す。図6(1)には、画像G1上を右向きに移動する作業者G10-1と、画像G1上を左向きに移動するロボットアームG10-2が写っている。また、図6(2)には、画像G1を含む動画像から検出された移動領域G2が示され、作業者G10-1の移動領域G20-1と、ロボットアームG10-2の移動領域G20-2が検出される。なお、検出される移動領域G20の数は限定されず、複数の移動領域G20が重なることで区別することができなくてもよい。移動領域検出部112は、移動領域G2において、移動領域G20-1および移動領域G20-2の各画素に「1」を設定し、それ以外の領域の各画素に「0」を設定する。なお、移動領域G20-1および移動領域G20-2の各画素に設定される画素値は「1」以外でもよい。例えば、各画素に対応する1つ以上の画素から、オプティカルフローにより算出される、各画素の移動量を示す値を、画素値として設定してもよい。 FIG. 6 shows an example of an image included in the moving image data 142 and an example of a moving area detected from the moving image including the image. FIG. 6(1) shows a worker G10-1 moving rightward on the image G1 and a robot arm G10-2 moving leftward on the image G1. Further, FIG. 6(2) shows the movement area G2 detected from the moving image including the image G1. 2 is detected. Note that the number of moving regions G20 to be detected is not limited, and a plurality of moving regions G20 do not need to be distinguished because they overlap. In the moving region G2, the moving region detection unit 112 sets “1” to each pixel of the moving regions G20-1 and G20-2, and sets “0” to each pixel of the other regions. Note that the pixel value set for each pixel in the moving region G20-1 and the moving region G20-2 may be other than "1". For example, a value indicating the amount of movement of each pixel calculated by optical flow from one or more pixels corresponding to each pixel may be set as the pixel value.
 次に、基底生成部121が、移動領域記憶部141に格納されている複数の移動領域を用いて、複数の基底ベクトル143(M1, M2, …, Mk(kは、所定の1以上の整数))を生成する(S540)。具体的には、例えば、基底生成部121は、複数の移動領域の主成分分析により得られる、1つ以上の固有ベクトルを基底ベクトル(M1, M2, …, Mk)として生成する。 Next, the base generation unit 121 generates a plurality of basis vectors 143 (M1, M2, . . . , Mk (k is a predetermined integer of 1 or more )) is generated (S540). Specifically, for example, the basis generation unit 121 generates one or more eigenvectors obtained by principal component analysis of a plurality of moving regions as basis vectors (M1, M2, . . . , Mk).
 図7は、基底ベクトルの一例として、複数の移動領域の主成分分析により得られる第1主成分の固有ベクトル(基底ベクトルM1)を示す。kの値は、主成分分析により、第1主成分の固有ベクトル(基底ベクトルM1)、第2主成分の固有ベクトル(基底ベクトルM2)と順に生成し、寄与率または累積寄与率が所定の値を超えた第k主成分の固有ベクトル(基底ベクトルMk)までとしてもよい。また、基底生成部121は、複数の移動領域の独立成分分析により得られる複数の基底を、基底ベクトルとしてもよい。また、入力される移動領域と出力される移動領域とが一致するようなオートエンコーダの重みを基底ベクトルとしてもよい。 FIG. 7 shows an eigenvector (base vector M1) of the first principal component obtained by principal component analysis of a plurality of moving regions, as an example of the base vector. The value of k is generated by principal component analysis in the order of the eigenvector of the first principal component (base vector M1) and the eigenvector of the second principal component (base vector M2). may be up to the eigenvector of the k-th principal component (basis vector Mk). Further, the basis generation unit 121 may use a plurality of basis obtained by independent component analysis of a plurality of moving regions as a basis vector. Alternatively, weights of an autoencoder that match the input movement area and the output movement area may be used as basis vectors.
 次に、領域抽出部122が、基底ベクトル143(M1, M2, …, Mk)から、候補領域144(R1, R2, …, Rn(nは、1以上の整数))を抽出する(S550)。具体的には、例えば、領域抽出部122は、基底ベクトルM1の要素の中で、所定の値を超える正の値を持つ要素の集合と、所定の値を超える負の値を持つ要素の集合とをそれぞれ、基底ベクトルM1の候補領域R1と、基底ベクトルM1の候補領域R2として抽出する。同様の操作を基底ベクトルM1, M2, …, Mkに対して行い、候補領域R1, R2, …, Rnを得る。 Next, the region extracting unit 122 extracts candidate regions 144 (R1, R2, …, Rn (n is an integer equal to or greater than 1)) from the basis vectors 143 (M1, M2, …, Mk) (S550). . Specifically, for example, the region extracting unit 122 extracts a set of elements having a positive value exceeding a predetermined value and a set of elements having a negative value exceeding a predetermined value among the elements of the base vector M1. are extracted as a candidate region R1 for the base vector M1 and a candidate region R2 for the base vector M1, respectively. Similar operations are performed on basis vectors M1, M2, …, Mk to obtain candidate regions R1, R2, …, Rn.
 図7は、候補領域の一例として、基底ベクトルM1から抽出された、候補領域R1と候補領域R2を示す。図7では、基底ベクトルM1の候補領域R1において、基底ベクトルM1の要素の中で、所定の値を超える正の値を持つ要素の集合の領域O1が抽出される。また、基底ベクトルM1の候補領域R2において、基底ベクトルM1の要素の中で、所定の値を超える負の値を持つ要素の集合の領域O2が抽出さる。候補領域R1および候補領域R2において、領域O1および領域O2の各画素に「1」、それ以外の領域の各画素に「0」が設定される場合を想定するが、領域O1および領域O2の各画素に設定される画素値は「1」以外でもよい。また、領域抽出部122は、基底ベクトルMの候補領域Rに対して領域分割手法を用いて、領域Oを複数の領域に分割し、分割された領域それぞれを基底ベクトルMの候補領域としてもよい。例えば、基底ベクトルM1の候補領域R1に対してWatershed法などの領域分割手法を用いて、領域O1を複数の領域(O11, O12, …, O1p)に分割し、分割された領域(O11, O12, …, O1p)それぞれを基底ベクトルM1の候補領域としてもよい。 FIG. 7 shows a candidate region R1 and a candidate region R2 extracted from the base vector M1 as an example of candidate regions. In FIG. 7, in the candidate region R1 of the base vector M1, a region O1 of a set of elements having positive values exceeding a predetermined value among the elements of the base vector M1 is extracted. Also, in the candidate region R2 of the base vector M1, a region O2 of a set of elements having a negative value exceeding a predetermined value among the elements of the base vector M1 is extracted. In the candidate region R1 and the candidate region R2, it is assumed that each pixel of the region O1 and the region O2 is set to "1" and each pixel of the other region is set to "0". A pixel value other than "1" may be set to the pixel. Alternatively, the region extracting unit 122 may divide the region O into a plurality of regions using a region division method for the candidate region R of the base vector M, and use each of the divided regions as the candidate region of the base vector M. . For example, the candidate region R1 of the base vector M1 is divided into multiple regions (O11, O12, …, O1p) using a region division method such as the Watershed method, and the divided regions (O11, O12 , …, O1p) may be used as candidate regions for basis vectors M1.
 図8は、領域対応表145の一例を示す。領域対応表145は、候補領域名とそれに対応する基底ベクトルを管理する。領域対応表145において、基底ベクトルMiの列に「1」が設定されている候補領域が、その基底ベクトルMiから抽出された候補領域であることを示す。例えば、候補領域R1と候補領域R2は、基底ベクトルM1の列に「1」が設定されているため、候補領域R1と候補領域R2は、基底ベクトルM1から抽出された候補領域である。領域対応表145は、領域抽出部122が候補領域を抽出する都度、あるいは領域統合部131が候補領域を統合する都度、更新される。上記処理の後、領域統合部131は、領域対応表145に基づき、互いに重複する候補領域144を統合する(S560)。 FIG. 8 shows an example of the area correspondence table 145. FIG. The region correspondence table 145 manages candidate region names and corresponding base vectors. In the area correspondence table 145, the candidate area with "1" set in the column of the basis vector Mi indicates that the candidate area is extracted from the basis vector Mi. For example, candidate regions R1 and R2 have "1" set in the column of base vector M1, so candidate regions R1 and R2 are candidate regions extracted from base vector M1. The region correspondence table 145 is updated each time the region extraction unit 122 extracts a candidate region or each time the region integration unit 131 integrates the candidate regions. After the above process, the area integrating section 131 integrates the overlapping candidate areas 144 based on the area correspondence table 145 (S560).
 ここで、領域統合部131が、重複する候補領域144を統合する処理について説明する。領域統合部131は、まず、領域対応表145を参照して、候補領域(R1, R2, …, Rn)から1つの候補領域Riを選択する。次に、領域対応表145を参照し、候補領域Riとは異なる基底ベクトルから抽出され候補領域のうち、候補領域Riと重複する候補領域Rjを選択する。具体的には、例えば、Jaccard係数あるいはSimpson係数などの集合の類似度を表す指標を用いて、候補領域Riの領域Oiに含まれる画素集合と、候補領域Rjの領域Ojに含まれる画素集合の類似度を算出し、算出された類似度が所定の値を超える場合、候補領域Riと候補領域Rjが重複すると判定する。 Here, the process of integrating the overlapping candidate areas 144 by the area integration unit 131 will be described. The region integration unit 131 first refers to the region correspondence table 145 and selects one candidate region Ri from the candidate regions (R1, R2, . . . , Rn). Next, referring to the region correspondence table 145, a candidate region Rj that overlaps with the candidate region Ri is selected from candidate regions extracted from base vectors different from the candidate region Ri. Specifically, for example, using an index representing the similarity of sets such as the Jaccard coefficient or the Simpson coefficient, the pixel set included in the region Oi of the candidate region Ri and the pixel set included in the region Oj of the candidate region Rj are calculated. A degree of similarity is calculated, and if the calculated degree of similarity exceeds a predetermined value, it is determined that the candidate regions Ri and Rj overlap.
 次に、候補領域Riと候補領域Rjが重複すると判定された場合、領域統合部131は、候補領域Riと候補領域Rjの面積を比較し、面積の小さい候補領域あるいは面積の大きい候補領域を領域対応表145から削除することによって、候補領域Riと候補領域Rjの統合を行う。なお、統合の方法は、面積の小さいあるいは大きい候補領域を削除する方法に限定されず、例えば、基底ベクトル143が主成分分析により得られた固有ベクトルである場合、2つの候補領域が抽出された基底ベクトルを比較し、寄与率の低い固有ベクトルから抽出された候補領域を領域対応表145から削除してもよい。 Next, when it is determined that the candidate region Ri and the candidate region Rj overlap, the region integration unit 131 compares the areas of the candidate region Ri and the candidate region Rj, and selects a candidate region with a small area or a candidate region with a large area. The candidate regions Ri and Rj are integrated by deleting them from the correspondence table 145 . Note that the integration method is not limited to the method of deleting candidate regions with small or large areas. The vectors may be compared, and candidate regions extracted from eigenvectors with low contribution rates may be deleted from the region correspondence table 145 .
 上記のように、基底ベクトルが主成分分析の場合、上位の主成分の固有ベクトル(基底ベクトル)を選択することで少数の候補領域に絞ることができる。さらに、領域の統合処理を実行することでより少ない候補領域に絞ることができる。その結果、画像データの解析に要する計算量や処理時間を削減することが可能となる。 As described above, when the basis vectors are principal component analysis, it is possible to narrow down to a small number of candidate regions by selecting the eigenvectors (basis vectors) of the upper principal components. Furthermore, by executing the area integration process, it is possible to narrow down to a smaller number of candidate areas. As a result, it is possible to reduce the amount of calculation and the processing time required for analyzing the image data.
 次に、領域選定部132が、領域選定表146に基づき、候補領域144を選定し(S570)、選定された候補領域144を監視領域として出力して(S580)、監視領域を決定する処理を終了する。 Next, the area selection unit 132 selects the candidate area 144 based on the area selection table 146 (S570), outputs the selected candidate area 144 as the monitoring area (S580), and performs the process of determining the monitoring area. finish.
 図9は、領域選定表146の一例を示す。領域選定表146は、候補領域名とそれに対応する選定結果を管理する。領域選定表146の選定結果の列に「1」が設定されている候補領域が、選定された候補領域であることを示す。例えば、候補領域R1と候補領域R2は、選定結果の列に「1」が設定されているため、候補領域R1と候補領域R2は、選定された候補領域である。 FIG. 9 shows an example of the area selection table 146. FIG. The area selection table 146 manages candidate area names and corresponding selection results. A candidate area for which "1" is set in the selection result column of the area selection table 146 indicates that it is a selected candidate area. For example, since "1" is set in the selection result column for candidate regions R1 and R2, candidate regions R1 and R2 are selected candidate regions.
 図10は、候補領域の選定結果を表示する表示画面の一例を示す図である。
  領域選定部132により選定された候補領域は監視領域として、画像処理装置100の出力装置15に表示される。なお、選定結果は、通信装置16を介して接続される端末(不図示)が持つ表示器に表示されてもよい。
FIG. 10 is a diagram showing an example of a display screen displaying selection results of candidate regions.
The candidate area selected by the area selection unit 132 is displayed on the output device 15 of the image processing apparatus 100 as a monitoring area. The selection result may be displayed on a display of a terminal (not shown) connected via the communication device 16. FIG.
 表示画面1000は、監視領域結果表1001と、監視領域表示1002を含む。監視領域結果表1001は、領域対応表145と、領域選定表146の内容を表示するデータである。監視領域表示1002には、領域選定表146の選定結果の列に「1」が設定されている候補領域が、その候補領域を囲む矩形とその候補領域名と共に重畳されて表示される。例えば、候補領域R1と候補領域R2は、領域選定表146の選定結果の列に「1」が設定されているため、候補領域R1と候補領域R2が、各候補領域を囲む矩形と各候補領域名と共に表示される。ユーザは、表示画面の表示1001,1002を見て、何れの候補領域が選定されたか、および選定された領域R1,R2がどの程度近いか或いは重複しているか、等を認識することができる。 A display screen 1000 includes a monitoring area result table 1001 and a monitoring area display 1002 . The monitoring area result table 1001 is data displaying the contents of the area correspondence table 145 and the area selection table 146 . In the monitoring area display 1002, the candidate areas for which "1" is set in the selection result column of the area selection table 146 are displayed together with the rectangle surrounding the candidate area and the name of the candidate area superimposed. For example, since "1" is set in the selection result column of the region selection table 146 for the candidate region R1 and the candidate region R2, the candidate region R1 and the candidate region R2 are the rectangles surrounding each candidate region and each candidate region. displayed with the first name. The user can recognize which candidate area has been selected and how close or overlapping the selected areas R1 and R2 are by looking at the displays 1001 and 1002 on the display screen.
 以上説明したように、本実施形態によれば、動画像データ142における複数の移動領域から生成される1つ以上の基底ベクトルに基づき、統計的に抽出された領域を、監視領域として決定することができる。したがって、未知の対象物体に対しても、その対象物体の移動領域から監視領域を決定することが可能になり、監視領域の決定の効率化を図ることができる。 As described above, according to the present embodiment, statistically extracted regions are determined as monitoring regions based on one or more base vectors generated from a plurality of moving regions in the moving image data 142. can be done. Therefore, even for an unknown target object, it becomes possible to determine the monitoring region from the moving region of the target object, and efficiency in determining the monitoring region can be achieved.
 なお、本発明は上記実施形態に限定されるものではなく、その要旨を逸脱しない範囲で種々変更可能であることはいうまでもない。例えば、上記の実施形態は本発明を分かりやすく説明するために詳細に説明したものであり、説明した全ての構成を備えるものに必ずしも限定されるものではない。また上記実施形態の構成の一部について、他の構成の追加や削除、置換をすることが可能である。 It goes without saying that the present invention is not limited to the above-described embodiments, and can be modified in various ways without departing from the gist of the present invention. For example, the above embodiments have been described in detail in order to explain the present invention in an easy-to-understand manner, and are not necessarily limited to those having all the described configurations. Moreover, it is possible to add, delete, or replace a part of the configuration of the above embodiment with another configuration.
 例えば、上記実施形態では、撮像装置3が取得した動画像データを処理する。動画像データの処理はリアルタイムに実行してもよいし、過去に取得して記憶しておいた動画像データを事後的に処理してもよい。更には、撮像装置3から取得した動画像データだけでなく、プロセッサ11や他の処理装置がシミュレーション用に生成した動画像データ(仮想動画データ)を処理対象としてもよい。 For example, in the above embodiment, moving image data acquired by the imaging device 3 is processed. The moving image data may be processed in real time, or the moving image data acquired and stored in the past may be processed afterward. Furthermore, moving image data (virtual moving image data) generated for simulation by the processor 11 or another processing device may be processed as well as the moving image data acquired from the imaging device 3 .
 また、上記の各構成、機能部、処理部、処理手段等は、それらの一部または全部を、例えば、集積回路で設計する等によりハードウェアで実現してもよい。また上記の各構成、機能等は、プロセッサがそれぞれの機能を実現するプログラムにより実現してもよい。各機能を実現するプログラム、テーブル、ファイル等の情報は、メモリやハードディスク、SSD(Solid State Drive)等の記録装置、ICカード、SDカード、DVD等の記録媒体に置くことができる。 In addition, each of the above configurations, functional units, processing units, processing means, etc. may be implemented in hardware, for example, by designing a part or all of them using an integrated circuit. Further, each configuration, function, etc. described above may be implemented by a program that causes the processor to implement each function. Information such as programs, tables, and files that implement each function can be stored in recording devices such as memories, hard disks, SSDs (Solid State Drives), and recording media such as IC cards, SD cards, and DVDs.
 また、以上に説明した各情報処理装置の各種機能部、各種処理部、各種データベースの配置形態は一例に過ぎない。各種機能部、各種処理部、各種データベースの配置形態は、これらの装置が備えるハードウェアやソフトウェアの性能、処理効率、通信効率等の観点から最適な配置形態に変更し得る。 In addition, the arrangement form of various functional units, various processing units, and various databases of each information processing apparatus described above is merely an example. The arrangement form of various functional units, various processing units, and various databases can be changed to an optimum arrangement form from the viewpoint of the performance, processing efficiency, communication efficiency, etc. of hardware and software provided in these devices.
 また、上述した各種のデータを格納するデータベースの構成(スキーマ(Schema)等)は、リソースの効率的な利用、処理効率向上、アクセス効率向上、検索効率向上等の観点から適宜変更し得る。 In addition, the configuration of the database that stores the various data described above (schema, etc.) can be changed as appropriate from the viewpoint of efficient use of resources, improvement of processing efficiency, improvement of access efficiency, improvement of search efficiency, etc.
1:画像処理システム
2:通信手段
3:撮像装置
11:プロセッサ
12:主記憶装置
13:補助記憶装置
14:入力装置
15:出力装置
16:通信装置
100:画像処理装置
110:取得部
111:データ取得部
112:移動領域検出部
120:抽出部
121:基底生成部
122:領域抽出部
130:選定部
131:領域統合部
132:領域選定部
140:記憶部
141:移動領域記憶部
142:動画像データ
143:基底ベクトル
144:候補領域
145:領域対応表
146:領域選定表
1: image processing system 2: communication means 3: imaging device 11: processor 12: main storage device 13: auxiliary storage device 14: input device 15: output device 16: communication device 100: image processing device 110: acquisition unit 111: data Acquisition unit 112: moving region detection unit 120: extraction unit 121: base generation unit 122: region extraction unit 130: selection unit 131: region integration unit 132: region selection unit 140: storage unit 141: movement region storage unit 142: moving image Data 143: Base vector 144: Candidate area 145: Area correspondence table 146: Area selection table

Claims (12)

  1. 動画像データから複数の移動領域を検出する移動領域検出部と、
    前記複数の移動領域を記憶する移動領域記憶部と、
    前記複数の移動領域に基づいて基底ベクトルを生成する基底生成部と、
    前記基底生成部により生成された前記基底ベクトルから候補領域を抽出する領域抽出部と、
    を有する、画像処理装置。
    a moving area detection unit that detects a plurality of moving areas from moving image data;
    a movement area storage unit that stores the plurality of movement areas;
    a basis generator that generates basis vectors based on the plurality of moving regions;
    a region extraction unit that extracts a candidate region from the base vectors generated by the base generation unit;
    An image processing device having
  2. 前記基底生成部は、前記複数の移動領域の主成分分析により得られる固有ベクトルを、基底ベクトルとして生成する、
    請求項1の画像処理装置。
    The basis generation unit generates eigenvectors obtained by principal component analysis of the plurality of movement regions as basis vectors.
    The image processing apparatus according to claim 1.
  3. 前記領域抽出部により抽出される、互いに重複する前記候補領域を統合する領域統合部と、
    前記領域統合部による統合の結果から前記候補領域を選定する領域選定部と、
    請求項1の画像処理装置。
    a region integration unit that integrates the overlapping candidate regions extracted by the region extraction unit;
    an area selection unit that selects the candidate area from the result of integration by the area integration unit;
    The image processing apparatus according to claim 1.
  4. 移動体が移動する可能性のある領域に設置された複数の撮像装置により取得された複数の動画像データを取得するデータ取得部と、
    前記データ取得部により取得された前記複数の動画像データを格納する動画像データ記憶部と、を有し、
    前記移動領域検出部は、前記データ取得部により取得された前記複数の動画像データを基に、前記複数の移動領域を検出する
    請求項1の画像処理装置。
    a data acquisition unit that acquires a plurality of moving image data acquired by a plurality of imaging devices installed in an area where a moving object may move;
    a moving image data storage unit that stores the plurality of moving image data acquired by the data acquisition unit;
    2. The image processing apparatus according to claim 1, wherein said movement area detection section detects said plurality of movement areas based on said plurality of moving image data acquired by said data acquisition section.
  5. 前記基底生成部は、前記主成分分析により、第1主成分の固有ベクトルから、寄与率または累積寄与率が所定の値を超えた第k主成分の固有ベクトルまで、順次生成する(kは複数の整数)、
    請求項2の画像処理装置。
    The basis generator sequentially generates eigenvectors of the first principal component to eigenvectors of the k-th principal component whose contribution rate or cumulative contribution rate exceeds a predetermined value (k is a plurality of integers) by the principal component analysis. ),
    3. The image processing apparatus according to claim 2.
  6. 前記領域抽出部は、前記基底ベクトルM1の候補領域R1において、基底ベクトルM1の要素の中で、所定の値を超える正の値を持つ要素の集合の領域O1を抽出し、基底ベクトルM1の候補領域R2において、基底ベクトルM1の要素の中で、所定の値を超える負の値を持つ要素の集合の領域O2を抽出し、
    抽出した候補領域の候補領域名とそれに対応する基底ベクトルを管理する領域対応表を作成する、
    請求項1の画像処理装置。
    The region extraction unit extracts a region O1 of a set of elements having a positive value exceeding a predetermined value among the elements of the base vector M1 in the candidate region R1 of the base vector M1, and extracts a region O1 of the base vector M1 candidate. extracting a region O2 of a set of elements having a negative value exceeding a predetermined value among the elements of the base vector M1 in the region R2;
    create a region correspondence table that manages the candidate region names of the extracted candidate regions and their corresponding basis vectors;
    The image processing apparatus according to claim 1.
  7. 前記領域統合部は、複数の候補領域が重複すると判定した場合、複数の候補領域の面積を比較し、面積の小さい候補領域あるいは面積の大きい候補領域を削除することによって、候補領域の統合を行う、
    請求項3の画像処理装置。
    When determining that a plurality of candidate regions overlap, the region integration unit integrates the candidate regions by comparing the areas of the plurality of candidate regions and deleting a candidate region with a small area or a candidate region with a large area. ,
    4. The image processing apparatus according to claim 3.
  8. 前記領域選定部により選定された、候補領域の名称(候補領域名)と、選定結果と、該候補領域名に対応する基底ベクトルを表す監視領域結果表と、選定された候補領域の表示を含む表示画面を表示する出力装置を有する、
    請求項1の画像処理装置。
    Names of candidate areas selected by the area selection unit (candidate area names), selection results, a monitoring area result table showing base vectors corresponding to the candidate area names, and display of the selected candidate areas. Having an output device for displaying a display screen,
    The image processing apparatus according to claim 1.
  9. 動画像データから複数の移動領域を検出する移動領域検出ステップと、
    前記複数の移動領域を記憶する移動領域記憶ステップと、
    前記複数の移動領域に基づいて基底ベクトルを生成する基底生成ステップと、
    前記基底生成ステップにより生成された前記基底ベクトルから候補領域を抽出する領域抽出ステップと、
    を有する、画像処理方法。
    a moving region detection step of detecting a plurality of moving regions from moving image data;
    a movement area storage step of storing the plurality of movement areas;
    a basis generating step of generating basis vectors based on the plurality of moving regions;
    a region extraction step of extracting a candidate region from the base vectors generated by the base generation step;
    An image processing method comprising:
  10. 前記基底生成ステップは、前記複数の移動領域の主成分分析により得られる固有ベクトルを、基底ベクトルとして生成する、
    請求項9の画像処理方法。
    In the base generation step, eigenvectors obtained by principal component analysis of the plurality of moving regions are generated as basis vectors.
    10. The image processing method according to claim 9.
  11. 前記領域抽出ステップにより抽出される、互いに重複する前記候補領域を統合する領域統合ステップと、
    前記領域統合ステップによる統合の結果から前記候補領域を選定する領域選定ステップと、
    請求項9の画像処理方法。
    a region integration step of integrating the overlapping candidate regions extracted by the region extraction step;
    an area selection step of selecting the candidate area from the result of integration by the area integration step;
    10. The image processing method according to claim 9.
  12. コンピュータが実行する画像処理用のプログラムは、
    動画像データから複数の移動領域を検出する移動領域検出部と、
    前記複数の移動領域を記憶する移動領域記憶部と、
    前記複数の移動領域に基づいて基底ベクトルを生成する基底生成部と、
    前記基底生成部により生成された前記基底ベクトルから候補領域を抽出する領域抽出部と、
    を有する、画像処理処理用のプログラム。
    An image processing program executed by a computer is
    a moving area detection unit that detects a plurality of moving areas from moving image data;
    a movement area storage unit that stores the plurality of movement areas;
    a basis generator that generates basis vectors based on the plurality of moving regions;
    a region extraction unit that extracts a candidate region from the base vectors generated by the base generation unit;
    A program for image processing.
PCT/JP2021/043070 2021-03-08 2021-11-24 Image processing device, method therefor, and image processing program WO2022190470A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2021036724A JP2022136898A (en) 2021-03-08 2021-03-08 Image processing device, method thereof, and image processing program
JP2021-036724 2021-03-08

Publications (1)

Publication Number Publication Date
WO2022190470A1 true WO2022190470A1 (en) 2022-09-15

Family

ID=83227819

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/043070 WO2022190470A1 (en) 2021-03-08 2021-11-24 Image processing device, method therefor, and image processing program

Country Status (2)

Country Link
JP (1) JP2022136898A (en)
WO (1) WO2022190470A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000137790A (en) * 1998-10-29 2000-05-16 Hitachi Ltd Method and device for monitoring image of man conveyor
JP2002049911A (en) * 2000-08-01 2002-02-15 Fujitsu Ltd Monitor

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000137790A (en) * 1998-10-29 2000-05-16 Hitachi Ltd Method and device for monitoring image of man conveyor
JP2002049911A (en) * 2000-08-01 2002-02-15 Fujitsu Ltd Monitor

Also Published As

Publication number Publication date
JP2022136898A (en) 2022-09-21

Similar Documents

Publication Publication Date Title
US11222239B2 (en) Information processing apparatus, information processing method, and non-transitory computer-readable storage medium
Agarwal et al. Anubhav: recognizing emotions through facial expression
JP2018173940A (en) Automatic object tracking in video feed using machine learning
JP2020515983A (en) Target person search method and device, device, program product and medium
WO2006013913A1 (en) Object image detection device, face image detection program, and face image detection method
US11323577B2 (en) Image processing device for creating an album
JP2015187759A (en) Image searching device and image searching method
JP6278108B2 (en) Image processing apparatus, image sensor, and image processing method
US20200250401A1 (en) Computer system and computer-readable storage medium
WO2022190470A1 (en) Image processing device, method therefor, and image processing program
US10248853B2 (en) Image output device, image output method, and computer program product
JP7192888B2 (en) Processing device, processing method and program
TW202022796A (en) Neural network model fusion method and electronic device using the same
JP2005235041A (en) Retrieval image display method and retrieval image display program
US20220374653A1 (en) System and method for learning human activities from video demonstrations using video augmentation
US11749021B2 (en) Retrieval device, control method, and non-transitory storage medium
JP6419560B2 (en) Search device, method and program
JP2015187770A (en) Image recognition device, image recognition method, and program
KR20210041856A (en) Method and apparatus for generating learning data required to learn animation characters based on deep learning
CN112115941B (en) Fire detection method, device, equipment and storage medium
CN112966272B (en) Internet of things Android malicious software detection method based on countermeasure network
JP6897095B2 (en) Image processing program, image processing device and image processing method
US20240169709A1 (en) Training device, object detection device, training method, object detection method, and non-transitory computer-readable storage medium
WO2022079841A1 (en) Group specifying device, group specifying method, and computer-readable recording medium
JP2006059284A (en) Object identification device, object identification method and program for this method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21930316

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21930316

Country of ref document: EP

Kind code of ref document: A1