US12148176B2 - Comparison apparatus, control method, and program - Google Patents

Comparison apparatus, control method, and program Download PDF

Info

Publication number
US12148176B2
US12148176B2 US17/439,928 US201917439928A US12148176B2 US 12148176 B2 US12148176 B2 US 12148176B2 US 201917439928 A US201917439928 A US 201917439928A US 12148176 B2 US12148176 B2 US 12148176B2
Authority
US
United States
Prior art keywords
registered object
registered
comparison
camera
area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US17/439,928
Other versions
US20220189046A1 (en
Inventor
Shihono MOCHIZUKI
Satoshi TERASAWA
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp filed Critical NEC Corp
Publication of US20220189046A1 publication Critical patent/US20220189046A1/en
Assigned to NEC CORPORATION reassignment NEC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MOCHIZUKI, SHIHONO, TERASAWA, Satoshi
Application granted granted Critical
Publication of US12148176B2 publication Critical patent/US12148176B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/38Registration of image sequences
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/36User authentication by graphic or iconic representation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/207Analysis of motion for motion estimation over a hierarchy of resolutions
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/62Extraction of image or video features relating to a temporal dimension, e.g. time-based feature extraction; Pattern tracking
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/761Proximity, similarity or dissimilarity measures
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/167Detection; Localisation; Normalisation using comparisons between temporally consecutive images
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30164Workpiece; Machine component
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Definitions

  • the present invention relates to a technique of determining an object included in a captured image.
  • Patent Document 1 discloses a technique of extracting a biometric registration image associated with a mobile terminal within a predetermined range from a biometric comparison apparatus and comparing the extracted biometric registration image with input biometric information of a user. Position information of each mobile terminal is periodically transmitted to the biometric comparison apparatus.
  • the present invention has been made in view of the above-described problem, and an object of the present invention is to provide a technique of shortening a time required for determining an object included in a captured image.
  • a comparison apparatus includes: 1) an extraction unit that acquires a captured image generated by a camera and extracts an image feature of a detected object detected from the acquired captured image; and 2) a determination unit that determines the detected object using the extracted image feature and object information of each of a plurality of registered objects.
  • the object information of each of the registered objects includes an image feature of the registered object.
  • the determination unit determines a comparison order of the object information based on position information of each of the registered objects, and determines the detected object by comparing the image feature included in the object information with the extracted image feature in order from the object information whose the determined comparison order is earlier.
  • a control method is executed by a computer.
  • the control method includes: 1) an extraction step of acquiring a captured image generated by a camera and extracting an image feature of a detected object detected from the acquired captured image; and 2) a determination step of determining the detected object using the extracted image feature and object information of each of a plurality of registered objects.
  • the object information of each of the registered objects includes an image feature of the registered object.
  • a comparison order of the object information is determined based on position information of each of the registered objects, and the detected object is determined by comparing the image feature included in the object information with the extracted image feature in order from the object information whose the determined comparison order is earlier.
  • a program according to the present invention causes a computer to execute each step of the control method according to the present invention.
  • FIG. 1 is a diagram illustrating an outline of an operation of a comparison apparatus according to the present example embodiment.
  • FIG. 2 is a diagram illustrating a configuration of the comparison apparatus according to an example embodiment 1.
  • FIG. 3 is a diagram illustrating a computer for realizing the comparison apparatus.
  • FIG. 4 is a flowchart illustrating a flow of processing executed by the comparison apparatus according to the example embodiment 1.
  • FIG. 5 is a diagram illustrating a case where a place where an object moves is divided into a plurality of areas.
  • each block diagram represents a functional unit configuration, not a hardware unit configuration.
  • FIG. 1 is a diagram illustrating an outline of an operation of a comparison apparatus 2000 according to the present example embodiment.
  • FIG. 1 is a diagram illustrating a conceptual description for facilitating understanding of the operation of the comparison apparatus 2000 , and does not specifically limit the operation of the comparison apparatus 2000 .
  • the comparison apparatus 2000 is an apparatus that determines an object captured by a camera 10 using a captured image 12 obtained from the camera 10 .
  • the camera 10 is a surveillance camera provided in a facility such as an airport or a stadium.
  • a place where the camera 10 is provided is not limited to indoors, and may be outdoors.
  • determining an object means determining a registered object that matches with an object detected from the captured image 12 (detected object 20 ), from a plurality of known objects (hereinafter, referred to as registered objects).
  • registered object here means an object for which there exists information (object information 30 , described later) that associates an identifier of that object with an image feature being a feature value of that object in an image.
  • the matching apparatus 2000 extracts a feature value (an image feature 22 ) of the detected object 20 in the image, the detected object 20 being detected from the captured image 12 .
  • image feature various feature values that can be extracted from the image may be used.
  • the detected object 20 is, for example, a person.
  • processing performed by the comparison apparatus 2000 is processing of “determining a person among known persons that matches with the person detected from the captured image 12 ”.
  • the detected object 20 is not limited to a person.
  • the detected object 20 may be an animal, a machine (a car or a robot), or the like.
  • the comparison apparatus 2000 determines the detected object 20 using a plurality of pieces of object information 30 .
  • Each of the pieces of object information 30 is information in which an identifier of a registered object (an object identifier 32 ) is associated with an image feature of the registered object (image feature 34 ).
  • the object information 30 is stored in an object information storage device 60 .
  • the comparison apparatus 2000 determines an image feature 34 that matches with the image feature 22 (for example, an image feature 34 whose similarity with the image feature 22 is equal to or greater than a threshold value), by comparing the image feature 22 of the detected object 20 obtained from the captured image 12 with the image feature 34 indicated by each of the pieces of object information 30 .
  • the object identifier 32 associated with the determined image feature 34 is determined as an identifier of the detected object 20 . Thereby, the detected object 20 is determined.
  • the comparison apparatus 2000 determines an order (comparison order) of the object information 30 used for comparison of the image feature 22 and the image feature 34 , based on position information 40 of each registered object.
  • position information 40 an identifier of the registered object (object identifier 42 ), a position of the registered object (position 44 ), and a time when the position is detected (detection time 46 ) are associated with each other.
  • the detection time 46 does not necessarily have to be included in the position information 40 .
  • the comparison apparatus 2000 uses, for comparison, the object information 30 of the registered object at an earlier timing.
  • the position information 40 is stored in a position information storage device 70 .
  • the comparison apparatus 2000 determines the detected object 20 , by comparing the image feature 34 indicated in the object information 30 and the image feature 22 , in comparison order based on the position information 40 of each registered object. In this way, by performing comparison in order in which the position of the registered object is considered, the object information 30 of the registered object, which is estimated to be likely to match with the detected object 20 , is used for comparison at an earlier timing. Therefore, a time required for determining the detected object 20 can be shortened.
  • each of the pieces of object information 30 is used for comparison in order, and thus the object information 30 of each registered object is used for comparison at any timing Therefore, even when a frequency of updating the position information 40 is reduced and thus the position indicated by the position information 40 of the registered object matching with the detected object 20 is greatly different from the actual current position of the registered object, this does not result in that the object information 30 of the registered object is not used for comparison and thus the detected object 20 cannot be determined. Accordingly, the frequency of updating the position information 40 can be flexibly determined in consideration of a processing load of the comparison apparatus 2000 and the like.
  • comparison apparatus 2000 according to the present example embodiment will be described in more detail.
  • FIG. 2 is a diagram illustrating a configuration of the comparison apparatus 2000 according to an example embodiment 1.
  • the comparison apparatus 2000 includes an extraction unit 2020 and a determination unit 2040 .
  • the extraction unit 2020 extracts an image feature 22 (an image feature of the detected object 20 ) from the captured image 12 generated by the camera 10 .
  • the determination unit 2040 determines the detected object 20 using the image feature 22 and the plurality of pieces of object information 30 .
  • the determination unit 2040 determines a comparison order of the object information 30 based on the position information 40 of each registered object.
  • the determination unit 2040 determines the detected object 20 by comparing the image feature 34 indicated in the object information 30 with the image feature 22 in order from the object information 30 whose determined comparison order is earlier.
  • Each functional unit of the comparison apparatus 2000 may be realized by hardware (for example, a hard-wired electronic circuit) that realizes each functional unit, or a combination of hardware and software (for example, a combination of an electronic circuit and a program for controlling the electronic circuit).
  • hardware for example, a hard-wired electronic circuit
  • software for example, a combination of an electronic circuit and a program for controlling the electronic circuit.
  • FIG. 3 is a diagram illustrating a computer 1000 for realizing the comparison apparatus 2000 .
  • the computer 1000 is a certain computer.
  • the computer 1000 is a personal computer (PC), a server machine, a tablet terminal, or a smartphone.
  • the computer 1000 may be a dedicated computer designed for realizing the comparison apparatus 2000 , or a general-purpose computer.
  • the computer 1000 includes a bus 1020 , a processor 1040 , a memory 1060 , a storage device 1080 , an input and output interface 1100 , and a network interface 1120 .
  • the bus 1020 is a data transmission path through which the processor 1040 , the memory 1060 , the storage device 1080 , the input and output interface 1100 , and the network interface 1120 transmit and receive data to and from each other.
  • a method of connecting the processor 1040 and the like to each other is not limited to bus connection.
  • the processor 1040 is a processor such as a central processing unit (CPU), a graphics processing unit (GPU), or a field-programmable gate array (FPGA).
  • the memory 1060 is a main storage device that is realized using a random access memory (RAM) or the like.
  • the storage device 1080 is an auxiliary storage device that is realized using a hard disk drive, a solid state drive (SSD), a memory card, a read only memory (ROM), or the like.
  • the storage device 1080 may be configured by hardware similar to the hardware of the main storage device such as RAM.
  • the input and output interface 1100 is an interface for connecting the computer 1000 and an input and output device.
  • the network interface 1120 is an interface for connecting the computer 1000 to a communication network.
  • the communication network is, for example, a local area network (LAN) or a wide area network (WAN).
  • a method of connecting the network interface 1120 to the communication network may be wireless connection or a wired connection.
  • the computer 1000 is connected to the camera 10 through the network interface 1120 .
  • the storage device 1080 stores a program module that realizes the functional unit of the comparison apparatus 2000 .
  • the processor 1040 realizes a function corresponding to each program module by reading each program module into the memory 1060 and executing the program module.
  • the camera 10 is a device having a function of capturing and then generating image data (captured image 12 ) representing a result of the capturing.
  • the camera 10 may be a video camera that generates a video or a still camera that generates a still image.
  • a captured image 12 is a video frame of the video.
  • the comparison apparatus 2000 may handle a plurality of cameras 10 . In this case, the comparison apparatus 2000 performs processing of determining a registered object included in each captured image 12 obtained from each of the plurality of cameras 10 .
  • the object information storage device 60 and the position information storage device 70 are any storage devices which allow an access from the comparison apparatus 2000 .
  • the object information 30 and the position information 40 may be stored in one storage device. That is, one storage device may have both a function as the object information storage device 60 and a function as the position information storage device 70 .
  • the object information 30 and the position information 40 may be realized by one piece of information (for example, one table) obtained by merging the object information 30 and the position information 40 , or may be realized by individual information (for example, tables different from each other).
  • FIG. 4 is a flowchart illustrating a flow of processing executed by the comparison apparatus 2000 according to the example embodiment 1.
  • the extraction unit 2020 acquires a captured image 12 (S 102 ).
  • the extraction unit 2020 extracts an image feature 22 from the captured image 12 (S 104 ).
  • Steps of S 106 to S 112 correspond to a loop processing A executed for each of the pieces of object information 30 .
  • the determination unit 2040 determines whether or not there is object information 30 that has not yet been used for comparison. In a case where comparison has already been performed for all the object information 30 , the loop processing A is ended. Thereby, the processing of FIG. 4 ends without determining the detected object 20 .
  • the processing of FIG. 4 proceeds to S 108 .
  • the determination unit 2040 determines object information 30 to be used for the next comparison from the object information 30 that has not yet been used for comparison, based on the position information 40 of each registered object (S 108 ).
  • the determination unit 2040 determines whether or not the image feature 34 indicated by the determined object information 30 matches with the image feature 22 extracted in S 104 (comparison of the image feature 34 and the image feature 22 ) (S 110 ).
  • the determination unit 2040 determines the detected object 20 as a registered object having an object identifier 32 corresponding to the image feature 34 (S 114 ). The processing of FIG. 4 is ended.
  • the flowchart illustrated in FIG. 4 is an example of a flow of processing performed by the comparison apparatus 2000 and the flow of the processing performed by the comparison apparatus 2000 is not limited to the flow illustrated in FIG. 4 .
  • the object information 30 to be used for the next comparison is sequentially determined.
  • the comparison order of each of the pieces of object information 30 may be determined before starting the loop processing A, and the object information 30 may be used for comparison in order in which the determined comparison order is earlier.
  • the extraction unit 2020 acquires the captured image 12 from the camera 10 (S 102 ).
  • a technique of acquiring the captured image generated by the camera an existing technique may be used.
  • the camera 10 is configured to transmit the captured image 12 to the comparison apparatus 2000 every time a new captured image 12 is generated.
  • the comparison apparatus 2000 acquires the captured image 12 by receiving the captured image 12 transmitted from the camera 10 .
  • the camera 10 may put the generated captured image 12 into a storage device which allows an access from the comparison apparatus 2000 .
  • the comparison apparatus 2000 acquires an unacquired captured image 12 by periodically accessing the storage device.
  • the extraction unit 2020 may acquire a part of the captured images 12 generated by the camera 10 instead of all the captured images 12 .
  • the camera 10 transmits one of a plurality of captured images 12 generated during a predetermined period, to the comparison apparatus 2000 .
  • the comparison apparatus 2000 acquires only the latest captured image 12 at the time of the periodical access to a storage device in which the captured images 12 are stored.
  • the extraction unit 2020 extracts the image feature 22 from the captured image 12 (S 104 ). Before performing the extraction, the extraction unit 2020 performs processing of detecting an object from the captured image 12 .
  • the object detected by the processing is referred to as a detected object 20 .
  • an existing technique may be used as a specific technique of detecting an object from an image.
  • the extraction unit 2020 detects an image region which is not included in the background region (a so-called foreground region) from the captured image 12 , and detects the detected foreground region as a detected object 20 .
  • the extraction unit 2020 performs processing of extracting the image feature 22 from the detected object 20 detected from the captured image 12 .
  • a technique of extracting an image feature from an image region representing an object an existing technique may be used. Note that, in a case where the detected object 20 is a person, it is particularly preferable to extract an image feature from an image region representing a face.
  • the object information 30 is information in which an identifier of a registered object is associated with an image feature of the registered object.
  • There are various methods of generating the object information 30 for the registered object For example, in a case where the camera 10 is provided in a facility, a user of the facility performs use registration of the facility in advance.
  • an apparatus used for use registration is referred to as a registration apparatus.
  • the registration apparatus generates object information 30 based on a result of use registration described later, and puts the generated object information 30 into the object information storage device 60 .
  • the registration apparatus is provided in a facility, for example, the registration apparatus is provided near an entrance or a reception of the facility (for example, a check-in counter at an airport). In this case, a user performs use registration by operating the registration apparatus.
  • the registration apparatus may be realized as a server machine that receives a use registration request from a user terminal through a network.
  • the registration apparatus is realized as a web server that provides a web site for use registration.
  • a user can perform use registration of the facility by accessing the website using a user's own terminal (a PC, a sample terminal, or the like).
  • the registration apparatus receives registration of personal information such as a name and an address from a user of the facility. At this time, the user also performs registration of an image including the user own (for example, a face picture).
  • the registration apparatus extracts an image feature of the user from the registered image, and generates object information 30 in which the image feature (corresponding to the image feature 34 ) is associated with an identifier of the user (corresponding to the object identifier 32 ).
  • a value generated by the registration apparatus may be set, or a value input by the user (for example, a passport number) may be set.
  • the image of the user may be registered by registering image data prepared by the user, or may be registered by using a camera attached to the registration apparatus. In the latter case, this camera captures a user and generates image data, and the image data is registered as an image of the user.
  • an administrator or the like of the object performs registration for using the object in the facility in the same manner as the use registration.
  • an administrator of the machine performs registration for using the machine in the facility.
  • an image of the machine is registered in addition to personal information of the user.
  • the detected object 20 is an animal or the like.
  • the position information 40 indicates a position of the registered object in association with the identifier of the registered object. Further, the position information 40 may further indicate a time when the registered object is detected at the position (detection time 46 ).
  • a position of a registered object is represented by coordinates in a system used to determine a position of an object, such as a global positioning system (GPS) or a beacon system.
  • a sensor such as a GPS sensor or a beacon receiver
  • coordinates of the person are determined using a GPS sensor provided in a mobile terminal owned by the person or a beacon receiver owned by the person.
  • a sensor is provided in the registered object. Note that, as a technique of representing a position of a registered object by coordinates using a GPS or a beacon, an existing technique may be used.
  • a mobile terminal in which the sensor is provided generates the position information 40 so that the object identifier 42 indicates an identifier of the mobile terminal (an address or the like), the position 44 indicates the coordinate, and the detection time 46 indicates the detection time of the coordinate.
  • each mobile terminal periodically transmits the position information 40 to an apparatus that collects the position information 40 (hereinafter, a position information collection apparatus).
  • the position information apparatus puts the received position information 40 into the position information storage device 70 .
  • the position of the registered object may be determined using a radio wave intensity of a radio wave received by the mobile terminal associated with the registered object. More specifically, it is assumed that the registered object such as a person owns a mobile terminal. Further, it is assumed that a transmitter which transmits a predetermined radio wave, such as a wireless LAN access point, is provided in a facility in which the comparison apparatus 2000 is used. In this case, the position of the mobile terminal can be estimated based on an intensity of the radio wave received from each transmitter by the mobile terminal. Therefore, for example, the position of the registered object can be determined by a combination of “an identifier of the transmitter and an intensity of the radio wave received from the transmitter” obtained from the mobile terminal owned by the registered object.
  • the object identifier 42 an identifier of the mobile terminal may be used. Further, as the detection time 46 , a time when the mobile terminal receives the radio wave may be used. Even in this case, the position information 40 is generated by each mobile terminal.
  • the position of the mobile terminal may be calculated as a coordinate representing a position in the facility based on the position of each transmitter and the radio wave intensity of the radio wave received from each transmitter by the mobile terminal.
  • a technique of calculating the position of the receiver based on the position of each of the plurality of transmitters and the intensity of the radio wave received from each of the plurality of transmitters by the receiver an existing technique may be used.
  • each mobile terminal transmits the generated position information 40 to the position information collection apparatus.
  • the position information apparatus stores the received position information 40 in the position information storage device 70 .
  • the position of the registered object may be determined by an access point that a mobile terminal associated with the registered object accesses.
  • the access point can record a combination of an address of the mobile terminal that accesses the access point (for example, a MAC address) and an access date and time. Therefore, it is able to generate the position information 40 so that the object identifier 42 indicates the address, the position 44 indicates the identifier of the access point (such as an address or an SSID), and the detection time 46 indicates the access date and time.
  • the access point generates the position information 40 for each mobile terminal that accesses the access point, and transmits the generated position information 40 to the position information collection apparatus.
  • the position information apparatus puts the received position information 40 into the position information storage device 70 .
  • the position of the registered object may be determined by an identifier of the camera 10 that generates the captured image 12 including the registered object.
  • the detected object 20 is detected from the captured image 12 generated by the camera c 1 at a certain time t 1 .
  • the comparison apparatus 2000 determines the detected object 20 and thus the identifier of the detected object 20 is determined as obj 1 .
  • the position of the registered object determined as obj 1 at the time t 1 is within the imaging range of the camera c 1 . Therefore, for example, it is able to generate the position information 40 so that the object identifier 42 indicates obj 1 , the position 44 indicates c 1 , and the detection time 46 indicates t 1 .
  • the comparison apparatus 2000 generates the position information 40 for the registered object determined as the detected object 20 .
  • the comparison apparatus 2000 stores the generated position information 40 in the position information storage device 70 .
  • the same type of information may be used, or different types of information may be used.
  • a passport number or the like input at the time of use registration is used as the object identifier 32
  • an identifier or the like of the mobile terminal is used as the object identifier 42 .
  • the same type of identifier as the object identifier 42 is included in the object information 30 in addition to the object identifier 32 .
  • object information 30 for the user is generated.
  • the user also performs registration of information used as the object identifier 42 at the time of use registration.
  • the object information 30 includes information corresponding to the object identifier 42 .
  • the object information 30 and the position information 40 can be associated with each other.
  • the determination unit 2040 searches for the position information 40 using the information which corresponds to the object identifier 42 and is indicated in the object information 30 . In this way, the position information 40 for the registered object determined by the object information 30 can be acquired.
  • the determination unit 2040 determines the comparison order of the object information 30 based on a relationship between the position of the registered object indicated by the position information 40 of each registered object and the position of the camera 10 that generates the acquired captured image 12 . For example, as the registered object is estimated to be present at a position closer to the camera 10 , the determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier. Here, as will be described later, by considering a direction in which the registered object is to move, the comparison order of the object information 30 of the registered object is not necessarily set to be earlier as the registered object is estimated to be present at a position closer to the camera 10 .
  • the determination unit 2040 acquires the position information 40 of the registered object, and calculates a distance between the position of the registered object indicated by the position information 40 and the position of the camera 10 .
  • the determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier as the calculated distance is shorter.
  • the distance between the position of the registered object indicated by the position information 40 and the position of the camera 10 can be calculated as a distance between the coordinates.
  • the position of the registered object is represented by an identifier of a radio wave transmitter such as a wireless LAN access point.
  • the distance between the position of the registered object and the position of the camera 10 may also be represented by how close the transmitter indicated as the position of the registered object is from the camera 10 .
  • a distance between the transmitter n-th closest from the camera 10 and the camera 10 is n (n is a natural number).
  • the determination unit 2040 may handle the registered object corresponding to the position information 40 whose detection time 46 is later (a difference between the current time and the detection time 46 is smaller) as being closer to the camera 10 . This is because the position information 40 whose detection time 46 is later is considered to be more accurate (have a smaller error from the actual current position).
  • the determination unit 2040 may calculate an estimated current position of the registered object based on the position 44 and the difference between the detection time 46 and the current time. In this case, the determination unit 2040 determines the comparison order of the object information 30 based on the calculated estimated position.
  • the determination unit 2040 calculates an estimated movement distance of the registered object by multiplying an elapsed time from the generation of the position information 40 for the registered object (a difference between the current time and the detection time 46 ) by a speed of the registered object.
  • the determination unit 2040 calculates, as an estimated position of the registered object, a position obtained by moving the position of the registered object from the position 44 by the estimated movement distance along a movement direction of the registered object.
  • the speed and the movement direction of the registered object will be described.
  • the speed of the registered object a fixed value may be used regardless of the registered object, or a different value may be used for each registered object.
  • the speed of the registered object is calculated based on an attribute of the registered object.
  • the speed of the registered object is determined in association with a value of each attribute such as a gender or a age group.
  • the attribute value of each registered object is included in the object information 30 .
  • the attribute value of the registered object may be obtained by image analysis of an image obtained by capturing the registered object in the past, or may be input by the user at the time of use registration.
  • the determination unit 2040 acquires the attribute value of the registered object by referring to the object information 30 for the registered object, and uses the speed of the registered object that is determined in association with the acquired attribute value.
  • the speed of the registered object may be calculated based on a change in the past position of the registered object.
  • a temporal change in the position of the registered object can be recognized by using a plurality of combinations of the position 44 of the registered object and the detection time 46 . Therefore, the determination unit 2040 calculates the speed of the registered object using a plurality of pieces of position information 40 (that is, a history of the pieces of position information 40 ) generated so far for one registered object.
  • a direction in which the registered object is to move (hereinafter, a predetermined direction) in a place where the camera 10 is provided is determined in advance, and the predetermined direction is handled as a movement direction of the registered object.
  • the predetermined direction is determined in advance based on a structure of the facility at which the comparison apparatus 2000 is operated. For example, at an airport, the movement direction after an entrance of a security inspection counter is considered to be a direction toward a boarding gate.
  • the determination unit 2040 may determine the comparison order of the registered object based on the position of the camera 10 and the predetermined direction. Specifically, the determination unit 2040 sets an earlier comparison order to each registered object positioned at a place that is forward in the predetermined direction from the position of the camera 10 than to each registered object positioned at a place that is forward in the direction opposite to the predetermined direction from the position of the camera 10 .
  • FIG. 5 is a diagram illustrating a case where a place where the registered object moves is divided into a plurality of areas.
  • FIG. 5 illustrates a passage in a facility at which the comparison apparatus 2000 is used. The passage is divided into five areas 50 ( 50 - 1 to 50 - 5 ).
  • processing of determining the detected object 20 by the comparison apparatus 2000 is performed in each area. For this reason, a camera 10 is provided in each area.
  • an area where the camera 10 in which the captured image 12 acquired in S 102 is generated is provided is referred to as a reference area.
  • the comparison apparatus 2000 performs processing of determining the detected object 20 on the captured image 12 obtained from the camera 10 - 3 provided in the area 50 - 3 .
  • the area 50 - 3 is the reference area.
  • the determination unit 2040 may determine the order of the area and determine the comparison order of the object information 30 based on the order. For example, the determination unit 2040 sets the order of the area 50 to be earlier as the area is closer to the reference area. The determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier as the registered object is positioned in the area with the earlier order.
  • the determination unit 2040 uses the pieces of object information 30 in the order of the object information 30 of the registered object positioned in the reference area 50 - 3 , the pieces of object information 30 of the registered objects positioned in the areas 50 - 2 and 50 - 4 , and the pieces of object information 30 of the registered objects positioned in the areas 50 - 1 and 50 - 5 .
  • the determination unit 2040 determines the order of the plurality of areas 50 .
  • the predetermined direction which is a direction in which the registered object is to move, is defined in advance.
  • the determination unit 2040 sets an earlier order to the area 50 that is forward in the predetermined direction from the reference area than to the area 50 that is forward in a direction opposite to the predetermined direction from the reference area.
  • the reference area is the area 50 - 3 .
  • the order of the area 50 - 4 is earlier than the order of the area 50 - 2 .
  • the order of the area 50 - 5 is earlier than the order of the area 50 - 1 . Therefore, the order of the areas is the order of the reference area, the area 50 - 4 , the area 50 - 2 , the area 50 - 5 , and the area 50 - 1 .
  • the determination unit 2040 may set an earlier order to all the areas 50 that are forward in the predetermined direction from the reference area than to each area 50 that is forward in the direction opposite to the predetermined direction from the reference area.
  • the order of the areas is the order of the reference area, the area 50 - 4 , the area 50 - 5 , the area 50 - 2 , and the area 50 - 1 .
  • a plurality of registered objects may be positioned in the same area.
  • the determination unit 2040 may determine the order of each registered object based on the detection time 46 of the position of each registered object. For example, the determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier as the detection time 46 is later (the detection time 46 is closer to the current time).
  • the determination unit 2040 may estimate an area where the registered object is currently positioned using the position information 40 . Specifically, for each registered object, the determination unit 2040 estimates the current position of the registered object using the position 44 , the speed of the registered object, and the movement direction of the registered object. The determination unit 2040 handles the area including the current position as the area where the registered object is currently positioned. In this case, the determination unit 2040 uses the area obtained by estimation as an area where each registered object is positioned. Note that a method of estimating the current position of the registered object using the detection time 46 or the like is as described above.
  • the determination unit 2040 determines whether or not the image feature 34 indicated by the object information 30 matches with the image feature 22 .
  • an existing technique may be used. For example, the determination unit 2040 calculates a similarity between the image feature 34 and the image feature 22 . When the similarity is equal to or higher than a threshold value, the determination unit 2040 determines that the image features match with each other. On the other hand, when the similarity is lower than the threshold value, the determination unit 2040 determines that the image features do not match with each other.
  • the determination unit 2040 determines the detected object 20 as the registered object having the object identifier 32 indicated by the object information 30 (S 114 ). At this time, the determination unit 2040 may output information representing the determination result (hereinafter, output information).
  • the output information is output in various forms. For example, the determination unit 2040 causes the output information to be displayed on a display device connected to the comparison apparatus 2000 , to be transmitted to another apparatus, or to be put into a storage device.
  • the contents of the output information are various.
  • the determination unit 2040 outputs the object information 30 of the registered object determined as the detected object 20 , as the output information.
  • the determination unit 2040 may further output the captured image 12 .
  • the captured image 12 including the detected object 20 and the object information 30 can be recognized in association with each other by a user of the comparison apparatus 2000 (for example, an administrator or a security guard of a facility).
  • a user of the comparison apparatus 2000 for example, an administrator or a security guard of a facility.
  • Which information of the object information 30 is to be included in the output information is set in advance.
  • the object information 30 includes an image of the registered object in addition to the image feature of the registered object.
  • the determination unit 2040 outputs the captured image 12 together with the image of the registered object included in the object information 30 . In this way, the user of the comparison apparatus 2000 can visually compare the image of the detected object 20 with the image of the registered object determined as matching with the detected object 20 .
  • the comparison apparatus 2000 highlights (for example, encloses with a frame) the object handled as the detected object 20 in the captured image 12 included in the output information. In this way, the user of the comparison apparatus 2000 can easily recognize the registered object determined as matching with the detected object 20 .
  • the comparison apparatus 2000 outputs output information representing this fact.
  • the output information includes the fact indicating that the detected object 20 cannot be determined and the captured image 12 including the detected object 20 . In this way, the user of the comparison apparatus 2000 can visually recognize the object that cannot be determined by the comparison apparatus 2000 .
  • the object that cannot be determined by the comparison apparatus 2000 is likely to be a person who has not been registered in advance. Therefore, it can be considered that the object requires attention in security of the facility. Accordingly, by providing the output information, a person or the like who requires attention can be easily recognized.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Software Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

A comparison apparatus (2000) extracts an image feature (22) from a detected object (20) detected from a captured image (12) generated by a camera (10). The comparison apparatus (2000) determines the detected object (20) using the extracted image feature (22) and object information (30) of each of a plurality of registered objects. Here, in the object information (30) of each registered object, an object identifier (32) of the registered object is associated with an image feature (34) of the registered object. The comparison apparatus (2000) determines a comparison order of the object information (30) based on position information (40) of each registered object, and determines the detected object (20) by comparing the image feature (34) indicated in the object information (30) with the image feature (22) in order from the object information (30) of which the determined comparison order is earlier.

Description

This application is a National Stage Entry of PCT/JP2019/013243 filed on Mar. 27, 2019, the contents of all of which are incorporated herein by reference, in their entirety.
TECHNICAL FIELD
The present invention relates to a technique of determining an object included in a captured image.
BACKGROUND ART
A technique of determining an object included in a captured image obtained from a camera has been disclosed. For example, Patent Document 1 discloses a technique of extracting a biometric registration image associated with a mobile terminal within a predetermined range from a biometric comparison apparatus and comparing the extracted biometric registration image with input biometric information of a user. Position information of each mobile terminal is periodically transmitted to the biometric comparison apparatus.
RELATED DOCUMENT Patent Documents
    • [Patent Document 1] Japanese Patent Application Publication No. 2017-59060
SUMMARY OF THE INVENTION Technical Problem
In the invention disclosed in Patent Document 1, when there are many persons who own a mobile terminal within the predetermined range from the biometric comparison apparatus, many biometric registration images are extracted as comparison targets. Thus, a time required for determining a user is increased.
The present invention has been made in view of the above-described problem, and an object of the present invention is to provide a technique of shortening a time required for determining an object included in a captured image.
Solution to Problem
A comparison apparatus according to the present invention includes: 1) an extraction unit that acquires a captured image generated by a camera and extracts an image feature of a detected object detected from the acquired captured image; and 2) a determination unit that determines the detected object using the extracted image feature and object information of each of a plurality of registered objects.
The object information of each of the registered objects includes an image feature of the registered object.
The determination unit determines a comparison order of the object information based on position information of each of the registered objects, and determines the detected object by comparing the image feature included in the object information with the extracted image feature in order from the object information whose the determined comparison order is earlier.
A control method according to the present invention is executed by a computer. The control method includes: 1) an extraction step of acquiring a captured image generated by a camera and extracting an image feature of a detected object detected from the acquired captured image; and 2) a determination step of determining the detected object using the extracted image feature and object information of each of a plurality of registered objects.
The object information of each of the registered objects includes an image feature of the registered object.
In the determination step, a comparison order of the object information is determined based on position information of each of the registered objects, and the detected object is determined by comparing the image feature included in the object information with the extracted image feature in order from the object information whose the determined comparison order is earlier.
A program according to the present invention causes a computer to execute each step of the control method according to the present invention.
Advantageous Effects of Invention
According to the present invention, there is provided a technique of shortening a time required for determining an object included in a captured image.
BRIEF DESCRIPTION OF THE DRAWINGS
The above-described object and other objects, features, and advantages will be further clarified by the preferred embodiment to be described below and the accompanying drawings attached herein.
FIG. 1 is a diagram illustrating an outline of an operation of a comparison apparatus according to the present example embodiment.
FIG. 2 is a diagram illustrating a configuration of the comparison apparatus according to an example embodiment 1.
FIG. 3 is a diagram illustrating a computer for realizing the comparison apparatus.
FIG. 4 is a flowchart illustrating a flow of processing executed by the comparison apparatus according to the example embodiment 1.
FIG. 5 is a diagram illustrating a case where a place where an object moves is divided into a plurality of areas.
DESCRIPTION OF EMBODIMENTS
Hereinafter, an example embodiment of the present invention will be described with reference to the drawings. Note that, in all the drawings, the same reference numerals are given to the same components and the description will not be repeated as appropriate. Further, unless otherwise specified, in each block diagram, each block represents a functional unit configuration, not a hardware unit configuration.
<Outline>
FIG. 1 is a diagram illustrating an outline of an operation of a comparison apparatus 2000 according to the present example embodiment. FIG. 1 is a diagram illustrating a conceptual description for facilitating understanding of the operation of the comparison apparatus 2000, and does not specifically limit the operation of the comparison apparatus 2000.
The comparison apparatus 2000 is an apparatus that determines an object captured by a camera 10 using a captured image 12 obtained from the camera 10. For example, the camera 10 is a surveillance camera provided in a facility such as an airport or a stadium. Here, a place where the camera 10 is provided is not limited to indoors, and may be outdoors.
Here, “determining an object” means determining a registered object that matches with an object detected from the captured image 12 (detected object 20), from a plurality of known objects (hereinafter, referred to as registered objects). Further, “registered object” here means an object for which there exists information (object information 30, described later) that associates an identifier of that object with an image feature being a feature value of that object in an image.
The matching apparatus 2000 extracts a feature value (an image feature 22) of the detected object 20 in the image, the detected object 20 being detected from the captured image 12. As the image feature, various feature values that can be extracted from the image may be used. Here, the detected object 20 is, for example, a person. In a case where the detected object 20 is a person, it is assumed that processing performed by the comparison apparatus 2000 is processing of “determining a person among known persons that matches with the person detected from the captured image 12”. Here, the detected object 20 is not limited to a person. For example, the detected object 20 may be an animal, a machine (a car or a robot), or the like.
The comparison apparatus 2000 determines the detected object 20 using a plurality of pieces of object information 30. Each of the pieces of object information 30 is information in which an identifier of a registered object (an object identifier 32) is associated with an image feature of the registered object (image feature 34). The object information 30 is stored in an object information storage device 60.
The comparison apparatus 2000 determines an image feature 34 that matches with the image feature 22 (for example, an image feature 34 whose similarity with the image feature 22 is equal to or greater than a threshold value), by comparing the image feature 22 of the detected object 20 obtained from the captured image 12 with the image feature 34 indicated by each of the pieces of object information 30. The object identifier 32 associated with the determined image feature 34 is determined as an identifier of the detected object 20. Thereby, the detected object 20 is determined.
Here, the comparison apparatus 2000 determines an order (comparison order) of the object information 30 used for comparison of the image feature 22 and the image feature 34, based on position information 40 of each registered object. In the position information 40, an identifier of the registered object (object identifier 42), a position of the registered object (position 44), and a time when the position is detected (detection time 46) are associated with each other. Here, the detection time 46 does not necessarily have to be included in the position information 40. For example, as the position 44 indicated in the position information 40 is closer to the position of the camera 10, the comparison apparatus 2000 uses, for comparison, the object information 30 of the registered object at an earlier timing. The position information 40 is stored in a position information storage device 70.
<Representative Advantageous Effect>
The comparison apparatus 2000 according to the present example embodiment determines the detected object 20, by comparing the image feature 34 indicated in the object information 30 and the image feature 22, in comparison order based on the position information 40 of each registered object. In this way, by performing comparison in order in which the position of the registered object is considered, the object information 30 of the registered object, which is estimated to be likely to match with the detected object 20, is used for comparison at an earlier timing. Therefore, a time required for determining the detected object 20 can be shortened.
In addition, according to the comparison apparatus 2000 of the present example embodiment, each of the pieces of object information 30 is used for comparison in order, and thus the object information 30 of each registered object is used for comparison at any timing Therefore, even when a frequency of updating the position information 40 is reduced and thus the position indicated by the position information 40 of the registered object matching with the detected object 20 is greatly different from the actual current position of the registered object, this does not result in that the object information 30 of the registered object is not used for comparison and thus the detected object 20 cannot be determined. Accordingly, the frequency of updating the position information 40 can be flexibly determined in consideration of a processing load of the comparison apparatus 2000 and the like.
Hereinafter, the comparison apparatus 2000 according to the present example embodiment will be described in more detail.
<Example of Functional Configuration of Comparison Apparatus 2000>
FIG. 2 is a diagram illustrating a configuration of the comparison apparatus 2000 according to an example embodiment 1. The comparison apparatus 2000 includes an extraction unit 2020 and a determination unit 2040. The extraction unit 2020 extracts an image feature 22 (an image feature of the detected object 20) from the captured image 12 generated by the camera 10. The determination unit 2040 determines the detected object 20 using the image feature 22 and the plurality of pieces of object information 30. Here, the determination unit 2040 determines a comparison order of the object information 30 based on the position information 40 of each registered object. The determination unit 2040 determines the detected object 20 by comparing the image feature 34 indicated in the object information 30 with the image feature 22 in order from the object information 30 whose determined comparison order is earlier.
<Hardware Configuration of Comparison Apparatus 2000>
Each functional unit of the comparison apparatus 2000 may be realized by hardware (for example, a hard-wired electronic circuit) that realizes each functional unit, or a combination of hardware and software (for example, a combination of an electronic circuit and a program for controlling the electronic circuit). Hereinafter, a case where each functional unit of the comparison apparatus 2000 is realized by a combination of hardware and software will be described.
FIG. 3 is a diagram illustrating a computer 1000 for realizing the comparison apparatus 2000. The computer 1000 is a certain computer. For example, the computer 1000 is a personal computer (PC), a server machine, a tablet terminal, or a smartphone. The computer 1000 may be a dedicated computer designed for realizing the comparison apparatus 2000, or a general-purpose computer.
The computer 1000 includes a bus 1020, a processor 1040, a memory 1060, a storage device 1080, an input and output interface 1100, and a network interface 1120. The bus 1020 is a data transmission path through which the processor 1040, the memory 1060, the storage device 1080, the input and output interface 1100, and the network interface 1120 transmit and receive data to and from each other. Here, a method of connecting the processor 1040 and the like to each other is not limited to bus connection. The processor 1040 is a processor such as a central processing unit (CPU), a graphics processing unit (GPU), or a field-programmable gate array (FPGA). The memory 1060 is a main storage device that is realized using a random access memory (RAM) or the like. The storage device 1080 is an auxiliary storage device that is realized using a hard disk drive, a solid state drive (SSD), a memory card, a read only memory (ROM), or the like. Here, the storage device 1080 may be configured by hardware similar to the hardware of the main storage device such as RAM.
The input and output interface 1100 is an interface for connecting the computer 1000 and an input and output device. The network interface 1120 is an interface for connecting the computer 1000 to a communication network. The communication network is, for example, a local area network (LAN) or a wide area network (WAN). A method of connecting the network interface 1120 to the communication network may be wireless connection or a wired connection. In FIG. 3 , the computer 1000 is connected to the camera 10 through the network interface 1120.
The storage device 1080 stores a program module that realizes the functional unit of the comparison apparatus 2000. The processor 1040 realizes a function corresponding to each program module by reading each program module into the memory 1060 and executing the program module.
<Camera 10>
The camera 10 is a device having a function of capturing and then generating image data (captured image 12) representing a result of the capturing. The camera 10 may be a video camera that generates a video or a still camera that generates a still image. In a case where the camera 10 generates a video, a captured image 12 is a video frame of the video.
The comparison apparatus 2000 may handle a plurality of cameras 10. In this case, the comparison apparatus 2000 performs processing of determining a registered object included in each captured image 12 obtained from each of the plurality of cameras 10.
<Object Information Storage Device 60 and Position Information Storage Device 70>
The object information storage device 60 and the position information storage device 70 are any storage devices which allow an access from the comparison apparatus 2000. Note that the object information 30 and the position information 40 may be stored in one storage device. That is, one storage device may have both a function as the object information storage device 60 and a function as the position information storage device 70. Here, in a case where both the object information 30 and the position information 40 are stored in one storage device, the object information 30 and the position information 40 may be realized by one piece of information (for example, one table) obtained by merging the object information 30 and the position information 40, or may be realized by individual information (for example, tables different from each other).
<Flow of Processing>
FIG. 4 is a flowchart illustrating a flow of processing executed by the comparison apparatus 2000 according to the example embodiment 1. The extraction unit 2020 acquires a captured image 12 (S102). The extraction unit 2020 extracts an image feature 22 from the captured image 12 (S104).
Steps of S106 to S112 correspond to a loop processing A executed for each of the pieces of object information 30. In S106, the determination unit 2040 determines whether or not there is object information 30 that has not yet been used for comparison. In a case where comparison has already been performed for all the object information 30, the loop processing A is ended. Thereby, the processing of FIG. 4 ends without determining the detected object 20.
On the other hand, in a case where there is object information 30 that has not yet been used for comparison, the processing of FIG. 4 proceeds to S108. The determination unit 2040 determines object information 30 to be used for the next comparison from the object information 30 that has not yet been used for comparison, based on the position information 40 of each registered object (S108). The determination unit 2040 determines whether or not the image feature 34 indicated by the determined object information 30 matches with the image feature 22 extracted in S104 (comparison of the image feature 34 and the image feature 22) (S110).
In a case where it is determined that the image feature 34 does not match with the image feature 22 (NO in S110), the processing of FIG. 4 proceeds to S112 (S112). Since S112 is the end of the loop processing A, the processing of FIG. 4 returns to S106.
In a case where it is determined that the image feature 34 matches with the image feature 22 (YES in S110), the determination unit 2040 determines the detected object 20 as a registered object having an object identifier 32 corresponding to the image feature 34 (S114). The processing of FIG. 4 is ended.
Note that the flowchart illustrated in FIG. 4 is an example of a flow of processing performed by the comparison apparatus 2000 and the flow of the processing performed by the comparison apparatus 2000 is not limited to the flow illustrated in FIG. 4 . For example, in FIG. 4, in the loop processing A, the object information 30 to be used for the next comparison is sequentially determined. On the other hand, for example, the comparison order of each of the pieces of object information 30 may be determined before starting the loop processing A, and the object information 30 may be used for comparison in order in which the determined comparison order is earlier.
<Acquisition of Captured Image 12: S102>
The extraction unit 2020 acquires the captured image 12 from the camera 10 (S102). Here, as a technique of acquiring the captured image generated by the camera, an existing technique may be used. For example, the camera 10 is configured to transmit the captured image 12 to the comparison apparatus 2000 every time a new captured image 12 is generated. The comparison apparatus 2000 acquires the captured image 12 by receiving the captured image 12 transmitted from the camera 10.
In addition, for example, the camera 10 may put the generated captured image 12 into a storage device which allows an access from the comparison apparatus 2000. In this case, for example, the comparison apparatus 2000 acquires an unacquired captured image 12 by periodically accessing the storage device.
Note that the extraction unit 2020 may acquire a part of the captured images 12 generated by the camera 10 instead of all the captured images 12. For example, the camera 10 transmits one of a plurality of captured images 12 generated during a predetermined period, to the comparison apparatus 2000. In addition, for example, the comparison apparatus 2000 acquires only the latest captured image 12 at the time of the periodical access to a storage device in which the captured images 12 are stored.
<Extraction of Image Feature 22: S104>
The extraction unit 2020 extracts the image feature 22 from the captured image 12 (S104). Before performing the extraction, the extraction unit 2020 performs processing of detecting an object from the captured image 12. The object detected by the processing is referred to as a detected object 20. Note that, as a specific technique of detecting an object from an image, an existing technique may be used. For example, the extraction unit 2020 detects an image region which is not included in the background region (a so-called foreground region) from the captured image 12, and detects the detected foreground region as a detected object 20.
The extraction unit 2020 performs processing of extracting the image feature 22 from the detected object 20 detected from the captured image 12. Here, as a technique of extracting an image feature from an image region representing an object, an existing technique may be used. Note that, in a case where the detected object 20 is a person, it is particularly preferable to extract an image feature from an image region representing a face.
<Object Information 30>
The object information 30 is information in which an identifier of a registered object is associated with an image feature of the registered object. There are various methods of generating the object information 30 for the registered object. For example, in a case where the camera 10 is provided in a facility, a user of the facility performs use registration of the facility in advance. Hereinafter, an apparatus used for use registration is referred to as a registration apparatus. The registration apparatus generates object information 30 based on a result of use registration described later, and puts the generated object information 30 into the object information storage device 60.
In a case where the registration apparatus is provided in a facility, for example, the registration apparatus is provided near an entrance or a reception of the facility (for example, a check-in counter at an airport). In this case, a user performs use registration by operating the registration apparatus.
In addition, for example, the registration apparatus may be realized as a server machine that receives a use registration request from a user terminal through a network. In this case, for example, the registration apparatus is realized as a web server that provides a web site for use registration. A user can perform use registration of the facility by accessing the website using a user's own terminal (a PC, a sample terminal, or the like).
The registration apparatus receives registration of personal information such as a name and an address from a user of the facility. At this time, the user also performs registration of an image including the user own (for example, a face picture). The registration apparatus extracts an image feature of the user from the registered image, and generates object information 30 in which the image feature (corresponding to the image feature 34) is associated with an identifier of the user (corresponding to the object identifier 32). As the identifier of the user, a value generated by the registration apparatus may be set, or a value input by the user (for example, a passport number) may be set.
The image of the user may be registered by registering image data prepared by the user, or may be registered by using a camera attached to the registration apparatus. In the latter case, this camera captures a user and generates image data, and the image data is registered as an image of the user.
Here, it is assumed that an object other than a person is handled as the detected object 20. In this case, an administrator or the like of the object performs registration for using the object in the facility in the same manner as the use registration. For example, in a case where a machine such as a robot is handled as the detected object 20, an administrator of the machine performs registration for using the machine in the facility. At this time, an image of the machine is registered in addition to personal information of the user. The same applies to a case where the detected object 20 is an animal or the like.
<Position Information 40>
The position information 40 indicates a position of the registered object in association with the identifier of the registered object. Further, the position information 40 may further indicate a time when the registered object is detected at the position (detection time 46).
Here, there are various ways to represent a position of an object. For example, a position of a registered object is represented by coordinates in a system used to determine a position of an object, such as a global positioning system (GPS) or a beacon system. In this case, a sensor (such as a GPS sensor or a beacon receiver) for determining coordinates is used. For example, in a case where a person is handled as the detected object 20, coordinates of the person are determined using a GPS sensor provided in a mobile terminal owned by the person or a beacon receiver owned by the person. In a case where an object other than a person is handled as the detected object 20, for example, a sensor is provided in the registered object. Note that, as a technique of representing a position of a registered object by coordinates using a GPS or a beacon, an existing technique may be used.
In a case where a coordinate determined by a GPS sensor or a beacon receiver is used as the position 44 indicated by the position information 40, for example, a mobile terminal in which the sensor is provided generates the position information 40 so that the object identifier 42 indicates an identifier of the mobile terminal (an address or the like), the position 44 indicates the coordinate, and the detection time 46 indicates the detection time of the coordinate. For example, each mobile terminal periodically transmits the position information 40 to an apparatus that collects the position information 40 (hereinafter, a position information collection apparatus). The position information apparatus puts the received position information 40 into the position information storage device 70.
In addition, for example, the position of the registered object may be determined using a radio wave intensity of a radio wave received by the mobile terminal associated with the registered object. More specifically, it is assumed that the registered object such as a person owns a mobile terminal. Further, it is assumed that a transmitter which transmits a predetermined radio wave, such as a wireless LAN access point, is provided in a facility in which the comparison apparatus 2000 is used. In this case, the position of the mobile terminal can be estimated based on an intensity of the radio wave received from each transmitter by the mobile terminal. Therefore, for example, the position of the registered object can be determined by a combination of “an identifier of the transmitter and an intensity of the radio wave received from the transmitter” obtained from the mobile terminal owned by the registered object. In this case, as the object identifier 42, an identifier of the mobile terminal may be used. Further, as the detection time 46, a time when the mobile terminal receives the radio wave may be used. Even in this case, the position information 40 is generated by each mobile terminal.
Further, the position of the mobile terminal may be calculated as a coordinate representing a position in the facility based on the position of each transmitter and the radio wave intensity of the radio wave received from each transmitter by the mobile terminal. Note that, as a technique of calculating the position of the receiver based on the position of each of the plurality of transmitters and the intensity of the radio wave received from each of the plurality of transmitters by the receiver, an existing technique may be used.
As described above, each mobile terminal transmits the generated position information 40 to the position information collection apparatus. The position information apparatus stores the received position information 40 in the position information storage device 70.
In addition, for example, the position of the registered object may be determined by an access point that a mobile terminal associated with the registered object accesses. The access point can record a combination of an address of the mobile terminal that accesses the access point (for example, a MAC address) and an access date and time. Therefore, it is able to generate the position information 40 so that the object identifier 42 indicates the address, the position 44 indicates the identifier of the access point (such as an address or an SSID), and the detection time 46 indicates the access date and time.
In this case, the access point generates the position information 40 for each mobile terminal that accesses the access point, and transmits the generated position information 40 to the position information collection apparatus. The position information apparatus puts the received position information 40 into the position information storage device 70.
Further, the position of the registered object may be determined by an identifier of the camera 10 that generates the captured image 12 including the registered object. For example, it is assumed that the detected object 20 is detected from the captured image 12 generated by the camera c1 at a certain time t1. It is assumed that the comparison apparatus 2000 determines the detected object 20 and thus the identifier of the detected object 20 is determined as obj1. In this case, it can be considered that the position of the registered object determined as obj1 at the time t1 is within the imaging range of the camera c1. Therefore, for example, it is able to generate the position information 40 so that the object identifier 42 indicates obj1, the position 44 indicates c1, and the detection time 46 indicates t1.
In this case, the comparison apparatus 2000 generates the position information 40 for the registered object determined as the detected object 20. The comparison apparatus 2000 stores the generated position information 40 in the position information storage device 70.
<Association Between Object Information 30 and Position Information 40>
For the object identifier 32 and the object identifier 42, the same type of information may be used, or different types of information may be used. For example, in the latter case, a passport number or the like input at the time of use registration is used as the object identifier 32, and an identifier or the like of the mobile terminal is used as the object identifier 42. In this case, the same type of identifier as the object identifier 42 is included in the object information 30 in addition to the object identifier 32.
For example, it is assumed that a user performs use registration using a registration apparatus and thus object information 30 for the user is generated. In this case, the user also performs registration of information used as the object identifier 42 at the time of use registration. For example, in a case where the identifier of the mobile terminal is used as the object identifier 42, the user registers the identifier of the mobile terminal owned by the user. In this way, the object information 30 includes information corresponding to the object identifier 42. By using the information, the object information 30 and the position information 40 can be associated with each other.
For example, in a case of determining the comparison order for certain object information 30, the determination unit 2040 searches for the position information 40 using the information which corresponds to the object identifier 42 and is indicated in the object information 30. In this way, the position information 40 for the registered object determined by the object information 30 can be acquired.
<Determination of Comparison Order>
The determination unit 2040 determines the comparison order of the object information 30 based on a relationship between the position of the registered object indicated by the position information 40 of each registered object and the position of the camera 10 that generates the acquired captured image 12. For example, as the registered object is estimated to be present at a position closer to the camera 10, the determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier. Here, as will be described later, by considering a direction in which the registered object is to move, the comparison order of the object information 30 of the registered object is not necessarily set to be earlier as the registered object is estimated to be present at a position closer to the camera 10.
For example, for the registered object represented by each of the plurality of pieces of object information 30, the determination unit 2040 acquires the position information 40 of the registered object, and calculates a distance between the position of the registered object indicated by the position information 40 and the position of the camera 10. The determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier as the calculated distance is shorter.
There are various methods of calculating the distance between the position of the registered object indicated by the position information 40 and the position of the camera 10. For example, in a case where both the position of the registered object and the position of the camera 10 are represented by coordinates, the distance between the position of the registered object and the position of the camera 10 can be calculated as a distance between the coordinates. In addition, for example, it is assumed that the position of the registered object is represented by an identifier of a radio wave transmitter such as a wireless LAN access point. In this case, the distance between the position of the registered object and the position of the camera 10 may also be represented by how close the transmitter indicated as the position of the registered object is from the camera 10. Specifically, it is assumed that a distance between the transmitter n-th closest from the camera 10 and the camera 10 is n (n is a natural number).
Here, there may be a plurality of registered objects having the same distance from the camera 10. For example, there is a case where both pieces of the position information 40 of two registered objects indicate the identifier of the transmitter second closest from the camera 10. In this case, for example, the determination unit 2040 may handle the registered object corresponding to the position information 40 whose detection time 46 is later (a difference between the current time and the detection time 46 is smaller) as being closer to the camera 10. This is because the position information 40 whose detection time 46 is later is considered to be more accurate (have a smaller error from the actual current position).
Further, in a case where the information indicated in the position information 40 is old (in a case where a difference between the detection time 46 and the current time is large), the actual current position of the registered object is likely to be distant from the position represented by the position 44. Therefore, the determination unit 2040 may calculate an estimated current position of the registered object based on the position 44 and the difference between the detection time 46 and the current time. In this case, the determination unit 2040 determines the comparison order of the object information 30 based on the calculated estimated position.
For example, for each registered object, the determination unit 2040 calculates an estimated movement distance of the registered object by multiplying an elapsed time from the generation of the position information 40 for the registered object (a difference between the current time and the detection time 46) by a speed of the registered object. The determination unit 2040 calculates, as an estimated position of the registered object, a position obtained by moving the position of the registered object from the position 44 by the estimated movement distance along a movement direction of the registered object.
Here, the speed and the movement direction of the registered object will be described. As the speed of the registered object, a fixed value may be used regardless of the registered object, or a different value may be used for each registered object. In the latter case, for example, the speed of the registered object is calculated based on an attribute of the registered object. For example, the speed of the registered object is determined in association with a value of each attribute such as a gender or a age group. Further, the attribute value of each registered object is included in the object information 30. The attribute value of the registered object may be obtained by image analysis of an image obtained by capturing the registered object in the past, or may be input by the user at the time of use registration. When calculating the estimated position of a certain registered object, the determination unit 2040 acquires the attribute value of the registered object by referring to the object information 30 for the registered object, and uses the speed of the registered object that is determined in association with the acquired attribute value.
In addition, for example, the speed of the registered object may be calculated based on a change in the past position of the registered object. For example, in a case where the position information 40 for each registered object is updated, a temporal change in the position of the registered object can be recognized by using a plurality of combinations of the position 44 of the registered object and the detection time 46. Therefore, the determination unit 2040 calculates the speed of the registered object using a plurality of pieces of position information 40 (that is, a history of the pieces of position information 40) generated so far for one registered object.
In the case of calculating the estimated position of the registered object, a direction in which the registered object is to move (hereinafter, a predetermined direction) in a place where the camera 10 is provided is determined in advance, and the predetermined direction is handled as a movement direction of the registered object. For example, the predetermined direction is determined in advance based on a structure of the facility at which the comparison apparatus 2000 is operated. For example, at an airport, the movement direction after an entrance of a security inspection counter is considered to be a direction toward a boarding gate.
Further, the determination unit 2040 may determine the comparison order of the registered object based on the position of the camera 10 and the predetermined direction. Specifically, the determination unit 2040 sets an earlier comparison order to each registered object positioned at a place that is forward in the predetermined direction from the position of the camera 10 than to each registered object positioned at a place that is forward in the direction opposite to the predetermined direction from the position of the camera 10.
<<Case Where Place is Divided into Areas>>
It is considered to divide a place where the registered object moves into a plurality of areas. FIG. 5 is a diagram illustrating a case where a place where the registered object moves is divided into a plurality of areas. FIG. 5 illustrates a passage in a facility at which the comparison apparatus 2000 is used. The passage is divided into five areas 50 (50-1 to 50-5). Here, it is assumed that processing of determining the detected object 20 by the comparison apparatus 2000 is performed in each area. For this reason, a camera 10 is provided in each area.
Hereinafter, for simplicity of explanation, an area where the camera 10 in which the captured image 12 acquired in S102 is generated is provided is referred to as a reference area. For example, it is assumed that the comparison apparatus 2000 performs processing of determining the detected object 20 on the captured image 12 obtained from the camera 10-3 provided in the area 50-3. In this case, it is assumed that the area 50-3 is the reference area.
In a case where a plurality of areas are provided in this way, the determination unit 2040 may determine the order of the area and determine the comparison order of the object information 30 based on the order. For example, the determination unit 2040 sets the order of the area 50 to be earlier as the area is closer to the reference area. The determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier as the registered object is positioned in the area with the earlier order.
For example, in the example of FIG. 5 , it is assumed that the area 50-3 is the reference area. In this case, for example, the determination unit 2040 uses the pieces of object information 30 in the order of the object information 30 of the registered object positioned in the reference area 50-3, the pieces of object information 30 of the registered objects positioned in the areas 50-2 and 50-4, and the pieces of object information 30 of the registered objects positioned in the areas 50-1 and 50-5.
Here, in a case where there are a plurality of areas 50 having the same distance from the reference area, the determination unit 2040 determines the order of the plurality of areas 50. For example, as described above, the predetermined direction, which is a direction in which the registered object is to move, is defined in advance. For the plurality of areas 50 having the same distance from the reference area, the determination unit 2040 sets an earlier order to the area 50 that is forward in the predetermined direction from the reference area than to the area 50 that is forward in a direction opposite to the predetermined direction from the reference area. For example, in the example of FIG. 5 , it is assumed that the reference area is the area 50-3. The order of the area 50-4 is earlier than the order of the area 50-2. Further, the order of the area 50-5 is earlier than the order of the area 50-1. Therefore, the order of the areas is the order of the reference area, the area 50-4, the area 50-2, the area 50-5, and the area 50-1.
In addition, for example, the determination unit 2040 may set an earlier order to all the areas 50 that are forward in the predetermined direction from the reference area than to each area 50 that is forward in the direction opposite to the predetermined direction from the reference area. For example, in the example described above, the order of the areas is the order of the reference area, the area 50-4, the area 50-5, the area 50-2, and the area 50-1.
<<<Plurality of Registered Objects in Same Area>>>
Here, a plurality of registered objects may be positioned in the same area. In this case, for example, the determination unit 2040 may determine the order of each registered object based on the detection time 46 of the position of each registered object. For example, the determination unit 2040 sets the comparison order of the object information 30 of the registered object to be earlier as the detection time 46 is later (the detection time 46 is closer to the current time).
<<<Estimation of Current Area>>>
The determination unit 2040 may estimate an area where the registered object is currently positioned using the position information 40. Specifically, for each registered object, the determination unit 2040 estimates the current position of the registered object using the position 44, the speed of the registered object, and the movement direction of the registered object. The determination unit 2040 handles the area including the current position as the area where the registered object is currently positioned. In this case, the determination unit 2040 uses the area obtained by estimation as an area where each registered object is positioned. Note that a method of estimating the current position of the registered object using the detection time 46 or the like is as described above.
<Comparison: S110>
The determination unit 2040 determines whether or not the image feature 34 indicated by the object information 30 matches with the image feature 22. Here, as a specific method of determining whether or not the image features match with each other, an existing technique may be used. For example, the determination unit 2040 calculates a similarity between the image feature 34 and the image feature 22. When the similarity is equal to or higher than a threshold value, the determination unit 2040 determines that the image features match with each other. On the other hand, when the similarity is lower than the threshold value, the determination unit 2040 determines that the image features do not match with each other.
<Result of Determination>
When the image feature 34 indicated by the object information 30 and the image feature 22 match with each other, the determination unit 2040 determines the detected object 20 as the registered object having the object identifier 32 indicated by the object information 30 (S114). At this time, the determination unit 2040 may output information representing the determination result (hereinafter, output information). Here, the output information is output in various forms. For example, the determination unit 2040 causes the output information to be displayed on a display device connected to the comparison apparatus 2000, to be transmitted to another apparatus, or to be put into a storage device.
The contents of the output information are various. For example, the determination unit 2040 outputs the object information 30 of the registered object determined as the detected object 20, as the output information. At this time, the determination unit 2040 may further output the captured image 12. In this way, the captured image 12 including the detected object 20 and the object information 30 can be recognized in association with each other by a user of the comparison apparatus 2000 (for example, an administrator or a security guard of a facility). Note that it is not necessary to output all pieces of the object information 30 and only a part of the pieces of the object information 30 may be output. Which information of the object information 30 is to be included in the output information is set in advance.
Here, preferably, the object information 30 includes an image of the registered object in addition to the image feature of the registered object. In this case, for example, the determination unit 2040 outputs the captured image 12 together with the image of the registered object included in the object information 30. In this way, the user of the comparison apparatus 2000 can visually compare the image of the detected object 20 with the image of the registered object determined as matching with the detected object 20.
In a case where a plurality of objects are detected from the captured image 12, preferably, the comparison apparatus 2000 highlights (for example, encloses with a frame) the object handled as the detected object 20 in the captured image 12 included in the output information. In this way, the user of the comparison apparatus 2000 can easily recognize the registered object determined as matching with the detected object 20.
Note that, in a case where the detected object 20 cannot be determined (in a case where there is no image feature 34 matching with the image feature 22), preferably, the comparison apparatus 2000 outputs output information representing this fact. The output information includes the fact indicating that the detected object 20 cannot be determined and the captured image 12 including the detected object 20. In this way, the user of the comparison apparatus 2000 can visually recognize the object that cannot be determined by the comparison apparatus 2000.
For example, in a case where use registration of a facility is performed in advance, the object that cannot be determined by the comparison apparatus 2000 is likely to be a person who has not been registered in advance. Therefore, it can be considered that the object requires attention in security of the facility. Accordingly, by providing the output information, a person or the like who requires attention can be easily recognized.
As described above, although the example embodiment of the present invention is described with reference to the drawings, the embodiment has been presented by way of example only, and various configurations other than the embodiment may also be adopted.
A part or all of the example embodiment may be described as in the following appendix, but are not limited thereto.
    • 1. A comparison apparatus comprising:
      • an extraction unit that acquires a captured image generated by a camera, and extracts an image feature of a detected object detected from the acquired captured image; and
      • a determination unit that determines the detected object using the extracted image feature and object information of each of a plurality of registered objects,
      • wherein the object information of each of the registered objects includes an image feature of the registered object, and
      • wherein the determination unit determines a comparison order of the object information based on position information of each of the registered objects, and determines the detected object by comparing the image feature included in the object information with the extracted image feature in order from the object information whose determined comparison order is earlier.
    • 2. The comparison apparatus according to 1,
      • wherein a place where the registered object is allowed to move includes a plurality of areas,
      • wherein the position information of the registered object indicates an area where the registered object is detected, and
      • wherein the determination unit determines a comparison order of the object information based on proximity between: an area at which the registered object is positioned and which is determined based on the position information of the registered object; and an area where the camera is provided.
    • 3. The comparison apparatus according to 2,
      • wherein, in a place where the camera is provided, a predetermined direction in which the registered object is to move is defined, and
      • wherein the determination unit sets the comparison order of the object information of each of the registered objects positioned in an area that is forward in the predetermined direction from the area where the camera is provided to be earlier than the comparison order of the object information of each of the registered objects positioned in an area that is forward in a direction opposite to the predetermined direction from the area where the camera is provided.
    • 4. The comparison apparatus according to 2,
      • wherein the position information of the registered object indicates a position where the registered object is detected and a detection time of the registered object, and
      • wherein the determination unit sets the comparison order of the object information of each of the plurality of registered objects for which the same area is indicated in the position information to be earlier as the detection time indicated in the position information is later.
    • 5. The comparison apparatus according to any one of 2 to 4,
      • wherein the determination unit estimates an area where the registered object is currently positioned based on the position and the detection time indicated by the position information of the registered object and a speed of the registered object, and determines the comparison order of the object information based on proximity between the estimated area and the area where the camera is provided.
    • 6. The comparison apparatus according to 1,
      • wherein, in a place where the camera is provided, a predetermined direction in which the registered object is to move is defined, and
      • wherein the comparison order of the object information of each of the registered objects positioned in a place that is forward in the predetermined direction from the place where the camera is provided is set to be earlier than the comparison order of the object information of each of the registered objects positioned in a place that is forward in a direction opposite to the predetermined direction from the place where the camera is provided.
    • 7. The comparison apparatus according to 1,
      • wherein the determination unit calculates a distance between a position of the registered object and a position of the camera using the position information of each of the registered objects, and sets the comparison order of the object information of the registered object to be earlier as the calculated distance is shorter.
    • 8. The comparison apparatus according to 7,
      • wherein the position information of the registered object indicates a position where the registered object is detected and a detection time of the registered object, and
      • wherein the determination unit estimates a current position of the registered object based on the position and the detection time indicated by the position information of the registered object and a speed of the registered object, and calculates a distance between the estimated current position and the position of the camera.
    • 9. The comparison apparatus according to 5 or 8,
      • wherein the determination unit calculates the speed of the registered object using a plurality of pieces of the position information generated so far for the registered object.
    • 10. A control method executed by a computer, the control method comprising:
      • an extraction step of acquiring a captured image generated by a camera and extracting an image feature of a detected object detected from the acquired captured image; and
      • a determination step of determining the detected object using the extracted image feature and object information of each of a plurality of registered objects,
      • wherein the object information of each of the registered objects includes an image feature of the registered object, and
      • wherein, in the determination step, a comparison order of the object information is determined based on position information of each of the registered objects, and the detected object is determined by comparing the image feature included in the object information with the extracted image feature in order from the object information whose determined comparison order is earlier.
    • 11. The control method according to 10,
      • wherein a place where the registered object is allowed to move includes a plurality of areas,
      • wherein the position information of the registered object indicates an area where the registered object is detected, and
      • wherein, in the determination step, a comparison order of the object information is determined based on proximity between: an area at which the registered object is positioned and which is determined based on the position information of the registered object; and an area where the camera is provided.
    • 12. The control method according to 11,
      • wherein, in a place where the camera is provided, a predetermined direction in which the registered object is to move is defined, and
      • wherein, in the determination step, the comparison order of the object information of each of the registered objects positioned in an area that is forward in the predetermined direction from the area where the camera is provided is set to be earlier than the comparison order of the object information of each of the registered objects positioned in an area that is forward in a direction opposite to the predetermined direction from the area where the camera is provided.
    • 13. The control method according to 11,
      • wherein the position information of the registered object indicates a position where the registered object is detected and a detection time of the registered object, and
      • wherein, in the determination step, the comparison order of the object information of each of the plurality of registered objects for which the same area is indicated in the position information is set to be earlier as the detection time indicated in the position information is later.
    • 14. The control method according to any one of 11 to 13,
      • wherein, in the determination step, an area where the registered object is currently positioned is estimated based on the position and the detection time indicated by the position information of the registered object and a speed of the registered object, and the comparison order of the object information is determined based on proximity between the estimated area and the area where the camera is provided.
    • 15. The control method according to 10,
      • wherein, in a place where the camera is provided, a predetermined direction in which the registered object is to move is determined, and
      • wherein the comparison order of the object information of each of the registered objects positioned in a place that is forward in the predetermined direction from the place where the camera is provided is set to be earlier than the comparison order of the object information of each of the registered objects positioned in a place that is forward in a direction opposite to the predetermined direction from the place where the camera is provided.
    • 16. The control method according to 10,
      • wherein, in the determination step, a distance between a position of the registered object and a position of the camera is calculated using the position information of each of the registered objects, and the comparison order of the object information of the registered object is set to be earlier as the calculated distance is shorter.
    • 17. The control method according to 16,
      • wherein the position information of the registered object indicates a position where the registered object is detected and a detection time of the registered object, and
      • wherein, in the determination step, a current position of the registered object is estimated based on the position and the detection time indicated by the position information of the registered object and a speed of the registered object, and a distance between the estimated current position and the position of the camera is calculated.
    • 18. The control method according to 14 or 17,
      • wherein, in the determination step, the speed of the registered object is calculated using a plurality of pieces of the position information generated so far for the registered object.
    • 19. A program causing a computer to execute each step of the control method according to any one of 10 to 18.

Claims (20)

What is claimed is:
1. A comparison apparatus comprising:
at least one memory storing instructions; and
at least one processor configured to execute the instructions to:
acquire a captured image generated by a camera;
extract an image feature of a detected object detected from the acquired captured image;
acquire position information of each of a plurality of registered objects;
determine an order of comparison of the plurality of registered objects, based on the position information of each registered object; and
identify the detected object by individually comparing, in the determined order of comparison, a respective plurality of pieces of object information of the plurality of registered objects stored in an object information storage with the extracted image feature until a matching registered object of the plurality of registered objects is found, such that the detected object identified as the matching registered object,
wherein the piece of object information of each registered object includes an image feature of the each registered object.
2. The comparison apparatus according to claim 1,
wherein a place where each registered object is allowed to move includes a plurality of areas,
wherein the position information of each registered object indicates an area where the each registered object is detected, and
wherein the at least one processor is further configured to execute the instructions to:
determine the order of comparison based on, for each registered object, proximity between the area where the each registered object is detected and an area where the camera is provided.
3. The comparison apparatus according to claim 2,
wherein, in a place where the camera is provided, a predetermined direction for each of the plurality of areas in which each registered object is to move is defined, and
wherein the at least one processor is further configured to execute the instructions to:
determine the order of comparison such that each registered object in an area forward in the predetermined direction from the area where the camera is provided is set in the order of comparison ahead of each registered object in an area forward in a direction opposite to the predetermined direction from the area where the camera is provided.
4. The comparison apparatus according to claim 2,
wherein the position information of each registered object indicates a position where the each registered object is detected and a detection time of the each registered object, and
wherein the at least one processor is further configured to execute the instructions to:
determine the order of comparison such that the registered objects detected in a same area are ordered in the order of comparison based on the detection time.
5. The comparison apparatus according to claim 2,
wherein the at least one processor is further configured to execute the instructions to:
estimate the area where each registered object is currently positioned based on a position and detection time indicated by the position information of the each registered object and a speed of the each registered object; and
determine the order of comparison based on, for each registered object, the proximity between the estimated area of the each registered object and the area where the camera is provided.
6. The comparison apparatus according to claim 5,
wherein the at least one processor is further configured to execute the instructions to calculate the speed of each registered object using a plurality of pieces of the position information generated so far for the each registered object.
7. The comparison apparatus according to claim 1,
wherein, in a place where the camera is provided, a predetermined direction for each of a plurality of areas in which each registered object is to move is defined, and
wherein the at least one processor is further configured to execute the instructions to:
determine the order of comparison such that each registered object in an area forward in the predetermined direction from the area where the camera is provided is set in the order of comparison ahead of each registered object in an area forward in a direction opposite to the predetermined direction from the area where the camera is provided.
8. The comparison apparatus according to claim 1,
wherein the at least one processor is further configured to execute the instructions to:
calculate a distance between a position of each registered object and a position of the camera using the position information of the each registered object; and
determine the order of comparison in accordance with the distance between the position of each registered object and the position of the camera, from smallest distanced to greatest distance.
9. The comparison apparatus according to claim 8,
wherein the position information of each registered object indicates a position where the each registered object is detected and a detection time of the each registered object, and
wherein the at least one processor is further configured to execute the instructions to:
estimate a current position of each registered object based on the position and the detection time indicated by the position information of the each registered object and a speed of the each registered object; and
calculate the distance between the estimated current position of each registered object and the position of the camera.
10. The comparison apparatus according to claim 9,
wherein the at least one processor is further configured to execute the instructions to calculate the speed of each registered object using a plurality of pieces of the position information generated so far for the each registered object.
11. A control method performed by a computer and comprising:
acquiring a captured image generated by a camera;
extracting an image feature of a detected object detected from the acquired captured image;
acquiring position information of each of a plurality of registered objects;
determining an order of comparison of the plurality of registered objects, based on the position information of each registered object; and
determining the detected object by individually comparing, in the determined order of comparison, a respective plurality of pieces of object information of the plurality of registered objects stored in an object information storage with the extracted image feature until a matching registered object of the plurality of registered objects is found, such that the detected object identified as the matching registered object,
wherein the piece of object information of each registered object includes an image feature of the each registered object.
12. The control method according to claim 11,
wherein a place where each registered object is allowed to move includes a plurality of areas,
wherein the position information of each registered object indicates an area where the each registered object is detected, and
wherein the method comprises determining the order of comparison based on, for each registered object, proximity between the area where the each registered object is detected and an area where the camera is provided.
13. The control method according to claim 12,
wherein, in a place where the camera is provided, a predetermined direction for each of the plurality of areas in which each registered object is to move is defined, and
wherein the method comprises determining the order of comparison such that each registered object in an area forward in the predetermined direction from the area where the camera is provided is set in the order of comparison ahead of each registered object in an area forward in a direction opposite to the predetermined direction from the area where the camera is provided.
14. The control method according to claim 12,
wherein the position information of each registered object indicates a position where the each registered object is detected and a detection time of the each registered object, and
wherein the method comprises determining the order of comparison such that the registered objects detected in a same area are ordered in the order of comparison based on the detection time.
15. The control method according to claim 12,
wherein the method comprises:
estimating the area where each registered object is currently positioned based on a position and detection time indicated by the position information of the each registered object and a speed of the each registered object; and
determining the order of comparison based on, for each registered object, the proximity between the estimated area of the each registered object and the area where the camera is provided.
16. The control method according to claim 15,
wherein the method comprises calculating the speed of each registered object is calculated using a plurality of pieces of the position information generated so far for the each registered object.
17. The control method according to claim 11,
wherein, in a place where the camera is provided, a predetermined direction for each of a plurality of areas in which each registered object is to move is defined, and
wherein the method comprises determining the order of comparison such that each registered object in an area forward in the predetermined direction from the area where the camera is provided is set in the order of comparison ahead of each registered object in an area forward in a direction opposite to the predetermined direction from the area where the camera is provided.
18. The control method according to claim 11,
wherein the method comprises:
calculating a distance between a position of each registered object and a position of the camera using the position information of the each registered object; and
determining the order of comparison in accordance with the distance between the position of each registered object and the position of the camera, from smallest distanced to greatest distance.
19. The control method according to claim 18,
wherein the position information of each registered object indicates a position where the each registered object is detected and a detection time of the each registered object, and
wherein the method comprises:
estimating a current position of each registered object based on the position and the detection time indicated by the position information of the each registered object and a speed of the each registered object; and
calculating the distance between the estimated current position of each registered object and the position of the camera.
20. A non-transitory computer-readable storage medium storing a program causing a computer to execute the control method according to claim 11.
US17/439,928 2019-03-27 2019-03-27 Comparison apparatus, control method, and program Active 2040-08-21 US12148176B2 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2019/013243 WO2020194585A1 (en) 2019-03-27 2019-03-27 Verification device, control method, and program

Publications (2)

Publication Number Publication Date
US20220189046A1 US20220189046A1 (en) 2022-06-16
US12148176B2 true US12148176B2 (en) 2024-11-19

Family

ID=72609665

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/439,928 Active 2040-08-21 US12148176B2 (en) 2019-03-27 2019-03-27 Comparison apparatus, control method, and program

Country Status (5)

Country Link
US (1) US12148176B2 (en)
JP (1) JP7218795B2 (en)
AR (1) AR118429A1 (en)
TW (1) TW202040422A (en)
WO (1) WO2020194585A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7582236B2 (en) * 2022-03-15 2024-11-13 トヨタ自動車株式会社 Person identification method, person identification system, and person identification program

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008158678A (en) 2006-12-21 2008-07-10 Toshiba Corp Person authentication device, person authentication method, and entrance / exit management system
JP2010079609A (en) 2008-09-26 2010-04-08 Brother Ind Ltd Personal authentication device, personal authentication program, and intercom system with personal authentication device
US20140015978A1 (en) * 2012-07-16 2014-01-16 Cubic Corporation Barrierless gate
US20140341427A1 (en) 2013-05-17 2014-11-20 Canon Kabushiki Kaisha Surveillance camera system and surveillance camera control apparatus
US20160275354A1 (en) 2015-03-17 2016-09-22 Samsung Eletrônica da Amazônia Ltda. System and method for video context-based composition and compression from normalized spatial resolution objects
JP2017059060A (en) 2015-09-17 2017-03-23 ソフトバンク株式会社 Biological collation system, biological collation method, biological collation device and control program
WO2017146160A1 (en) 2016-02-26 2017-08-31 日本電気株式会社 Facial verification system, facial verification method, and recording medium
US20180314893A1 (en) * 2017-04-26 2018-11-01 Canon Kabushiki Kaisha Information processing device, video image monitoring system, information processing method, and recording medium
CN109186584A (en) 2018-07-18 2019-01-11 浙江臻万科技有限公司 A kind of indoor orientation method and positioning system based on recognition of face

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008158678A (en) 2006-12-21 2008-07-10 Toshiba Corp Person authentication device, person authentication method, and entrance / exit management system
JP2010079609A (en) 2008-09-26 2010-04-08 Brother Ind Ltd Personal authentication device, personal authentication program, and intercom system with personal authentication device
US20140015978A1 (en) * 2012-07-16 2014-01-16 Cubic Corporation Barrierless gate
US20140341427A1 (en) 2013-05-17 2014-11-20 Canon Kabushiki Kaisha Surveillance camera system and surveillance camera control apparatus
JP2014241578A (en) 2013-05-17 2014-12-25 キヤノン株式会社 Camera system and camera control device
US20160275354A1 (en) 2015-03-17 2016-09-22 Samsung Eletrônica da Amazônia Ltda. System and method for video context-based composition and compression from normalized spatial resolution objects
JP2017059060A (en) 2015-09-17 2017-03-23 ソフトバンク株式会社 Biological collation system, biological collation method, biological collation device and control program
WO2017146160A1 (en) 2016-02-26 2017-08-31 日本電気株式会社 Facial verification system, facial verification method, and recording medium
CN108701225A (en) 2016-02-26 2018-10-23 日本电气株式会社 Facial recognition system, facial recognition method and storage medium
US20190050631A1 (en) * 2016-02-26 2019-02-14 Nec Corporation Face recognition system, face recognition method, and storage medium
US20180314893A1 (en) * 2017-04-26 2018-11-01 Canon Kabushiki Kaisha Information processing device, video image monitoring system, information processing method, and recording medium
CN109186584A (en) 2018-07-18 2019-01-11 浙江臻万科技有限公司 A kind of indoor orientation method and positioning system based on recognition of face

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
International Search Report for PCT Application No. PCT/JP2019/013243, mailed on Jun. 11, 2019.
TW Office Action for TW Application No. 109106875, mailed on Sep. 21, 2022 with English Translation.

Also Published As

Publication number Publication date
AR118429A1 (en) 2021-10-06
TW202040422A (en) 2020-11-01
JPWO2020194585A1 (en) 2020-10-01
WO2020194585A1 (en) 2020-10-01
JP7218795B2 (en) 2023-02-07
US20220189046A1 (en) 2022-06-16

Similar Documents

Publication Publication Date Title
CN109753928B (en) Method and device for identifying illegal buildings
US8971577B2 (en) Monitoring device, reliability calculation program, and reliability calculation method
US20160171300A1 (en) Authentication apparatus and method
JP2017004252A (en) Image information processing system
US11315430B2 (en) System, program, and method for detecting information on a person from a video of an on-vehicle camera
CN113807342B (en) Method and related device for acquiring device information based on image
US20220067343A1 (en) Information processing apparatus, information processing method, and storage medium
CN111065044A (en) Data association analysis method, device and computer storage medium based on big data
Radaelli et al. Using cameras to improve wi-fi based indoor positioning
US12148176B2 (en) Comparison apparatus, control method, and program
CN113744236B (en) Loop detection method, device, storage medium and computer program product
US20190122228A1 (en) Examination device
US11896408B2 (en) Automated patient modeling and positioning
JP2018186397A (en) Information processing device, image monitoring system, information processing method, and program
CN111126159A (en) Method, apparatus, electronic device, and medium for tracking pedestrian in real time
US9817471B2 (en) Method for adding information and server apparatus
CN104424635A (en) Information processing method, system and equipment
US20220222961A1 (en) Attribute determination device, attribute determination program, and attribute determination method
CN109752001A (en) Navigation system, method and apparatus
US12348852B2 (en) Person identification method, person identification system, and computer readable medium
US20250232615A1 (en) Management apparatus, management method, and non-transitory computer-readable medium
JP6434307B2 (en) POSITION INFORMATION ESTIMATION DEVICE, POSITION INFORMATION ESTIMATION METHOD, AND PROGRAM
CN113052034B (en) Living body detection method based on binocular camera and related device
EP4264569A1 (en) Systems and methods for nose-based pet identification
JP6443144B2 (en) Information output device, information output program, information output method, and information output system

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: NEC CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MOCHIZUKI, SHIHONO;TERASAWA, SATOSHI;SIGNING DATES FROM 20230131 TO 20230201;REEL/FRAME:065233/0766

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE