WO2023286152A1 - Detection device, detection method, and non-transitory computer-readable medium - Google Patents

Detection device, detection method, and non-transitory computer-readable medium Download PDF

Info

Publication number
WO2023286152A1
WO2023286152A1 PCT/JP2021/026293 JP2021026293W WO2023286152A1 WO 2023286152 A1 WO2023286152 A1 WO 2023286152A1 JP 2021026293 W JP2021026293 W JP 2021026293W WO 2023286152 A1 WO2023286152 A1 WO 2023286152A1
Authority
WO
WIPO (PCT)
Prior art keywords
person
predicted
information
camera
video data
Prior art date
Application number
PCT/JP2021/026293
Other languages
French (fr)
Japanese (ja)
Inventor
健吾 大羽賀
Original Assignee
日本電気株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電気株式会社 filed Critical 日本電気株式会社
Priority to PCT/JP2021/026293 priority Critical patent/WO2023286152A1/en
Publication of WO2023286152A1 publication Critical patent/WO2023286152A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/40Business processes related to the transportation industry

Definitions

  • This disclosure relates to technology that supports the provision of safe transportation.
  • a means of transportation using vehicles such as trains and buses is provided. Such transportation is preferably provided in a safe manner. Therefore, information processing technology has been developed to support the provision of safe means of transportation.
  • Patent Document 1 discloses a technology for opening and closing the door at a safe timing by detecting a person near the door of a train with a sensor and predicting the person's behavior based on the detection result of the sensor.
  • Patent Literature 2 discloses a technique of detecting a person near the door of a train with a sensor and detecting rushing boarding based on the moving speed and moving direction of the person.
  • Patent Documents 1 and 2 can only monitor the behavior of people near the train door.
  • the present invention has been made in view of the above problems, and one of its purposes is to provide a new technique capable of improving the safety of transportation.
  • the detection device of the present disclosure predicts that a person who is about to board the target vehicle just before the departure of the target vehicle passes through an imaging range of a camera capable of imaging a person heading to the departure place of the target vehicle.
  • a first acquisition unit that acquires predicted passage time information indicating a passage time
  • a second acquisition unit that obtains video data generated by the camera
  • a detection unit that detects a person to be monitored who is predicted to try to board the target vehicle just before departure of the vehicle.
  • the detection method of the present disclosure is executed by a computer.
  • the detection method is a predicted passage time at which a person who is about to board the vehicle of interest is predicted to pass through an imaging range of a camera capable of imaging a person heading for the departure point of the vehicle of interest just before departure of the vehicle of interest.
  • the non-transitory computer-readable medium of the present disclosure stores a program that causes a computer to execute the detection method of the present disclosure.
  • new technology is provided that can improve the safety of transportation.
  • FIG. 4 is a diagram illustrating an overview of the operation of the detection device of Embodiment 1;
  • FIG. 2 is a block diagram illustrating the functional configuration of the detection device of Embodiment 1;
  • FIG. It is a block diagram which illustrates the hardware constitutions of the computer which implement
  • 4 is a flow chart illustrating the flow of processing executed by the detection device of Embodiment 1.
  • FIG. It is a figure which illustrates predicted passage time information.
  • 3 is a block diagram illustrating the functional configuration of a detection device having an output section;
  • FIG. FIG. 4 is a diagram illustrating output information transmitted to a mobile terminal possessed by a person requiring monitoring;
  • predetermined values such as predetermined values and reference values are stored in advance in an arbitrary storage unit in a manner accessible from a device that uses the values.
  • the storage unit is composed of one or more arbitrary number of storage devices.
  • FIG. 1 is a diagram illustrating an overview of the operation of the detection device 2000 of Embodiment 1.
  • FIG. 1 is a diagram for facilitating understanding of the outline of the detection device 2000, and the operation of the detection device 2000 is not limited to that shown in FIG.
  • the detection device 2000 detects people who are predicted to try to board the vehicle 10 just before departure (so-called rush boarding) from among the people heading to the departure place 20 of the vehicle 10 .
  • a person predicted to try to board the vehicle 10 just before departure is referred to as a person to be monitored.
  • a vehicle 10 is a vehicle used to provide a means of transportation.
  • vehicle 10 is a vehicle that carries customers.
  • the vehicle 10 has a predetermined departure time.
  • such vehicles 10 include trains, buses, and the like.
  • "just before departure of the vehicle 10" can be defined as, for example, "after a predetermined time before the departure time of the vehicle 10".
  • the predetermined time can be set as 10 seconds.
  • the departure location 20 is the location from which the vehicle 10 departs. If the vehicle 10 is a train, the departure point 20 is the platform on which the train stops. If the vehicle 10 is a bus, the departure point 20 is the stop at which the bus stops.
  • the detection device 2000 uses the video data 50 and the predicted passage time information 60 to detect the person to be monitored.
  • the video data 50 is video data generated by the camera 40 capable of capturing images of people heading to the starting point 20 .
  • a camera 40 is provided at a place where a person heading to the departure place 20 passes.
  • the predicted passage time information 60 indicates the time at which the person to be monitored is predicted to pass through the imaging range of the camera 40 (hereinafter referred to as predicted passage time).
  • the predicted passage time can also be said to be "the time at which a person who is within the imaging range of the camera 40 at the predicted passage time will get on the vehicle 10 just before departure if they hurry to the departure place 20".
  • the detection device 2000 uses the predicted passage time information 60 to detect persons to be monitored from among persons included in the video data 50 (persons captured by the camera 40).
  • the person to be monitored is, among the persons detected from the video data 50, a person whose image is captured at a time close to the predicted passage time and who is assumed to be moving in a hurry. A specific method for detecting such a person will be described later.
  • a person who is predicted to board the vehicle 10 just before departure (a person who is predicted to board the vehicle 10 at the last minute) is detected from among the people heading to the departure point 20 of the vehicle 10. . Therefore, it is possible to grasp the possibility of rushing to board a train at an earlier timing than in the case of detecting a rushing boarding person from among the people near the train door. Therefore, transportation can be provided more safely.
  • the station attendant it is possible for the station attendant to make an announcement in advance to call attention, so that the trainees refrain from rushing to get on the train.
  • the detection device 2000 of this embodiment will be described in more detail below.
  • FIG. 2 is a block diagram illustrating the functional configuration of the detection device 2000 of Embodiment 1.
  • a first acquisition unit 2020 acquires the predicted passage time information 60 .
  • a second acquisition unit 2040 acquires the video data 50 .
  • the detection unit 2060 detects a person to be monitored from persons included in the video data 50 using the predicted passage time information 60 .
  • Each functional configuration unit of the detection device 2000 may be implemented by hardware that implements each functional configuration unit (eg, hardwired electronic circuit, etc.), or a combination of hardware and software (eg, electronic A combination of a circuit and a program that controls it, etc.).
  • each functional component of the detection device 2000 is implemented by a combination of hardware and software will be further described below.
  • FIG. 3 is a block diagram illustrating the hardware configuration of the computer 500 that implements the detection device 2000.
  • Computer 500 is any computer.
  • the computer 500 is a stationary computer such as a PC (Personal Computer) or a server machine.
  • the computer 500 is a portable computer such as a smart phone or a tablet terminal.
  • Computer 500 may be a dedicated computer designed to implement detection apparatus 2000 or a general-purpose computer.
  • the functions of the detection device 2000 are implemented on the computer 500.
  • the application is composed of a program for realizing each functional component of the detection device 2000 .
  • the acquisition method of the above program is arbitrary.
  • the program can be acquired from a storage medium (DVD disc, USB memory, etc.) in which the program is stored.
  • the program can be obtained by downloading the program from a server device that manages the storage unit storing the program.
  • Computer 500 has bus 502 , processor 504 , memory 506 , storage device 508 , input/output interface 510 and network interface 512 .
  • the bus 502 is a data transmission path through which the processor 504, memory 506, storage device 508, input/output interface 510, and network interface 512 exchange data with each other.
  • the method of connecting the processors 504 and the like to each other is not limited to bus connection.
  • the processor 504 is various processors such as a CPU (Central Processing Unit), GPU (Graphics Processing Unit), or FPGA (Field-Programmable Gate Array).
  • the memory 506 is a main memory implemented using a RAM (Random Access Memory) or the like.
  • the storage device 508 is an auxiliary storage device implemented using a hard disk, SSD (Solid State Drive), memory card, ROM (Read Only Memory), or the like.
  • the input/output interface 510 is an interface for connecting the computer 500 and input/output devices.
  • the input/output interface 510 is connected to an input device such as a keyboard and an output device such as a display device.
  • a network interface 512 is an interface for connecting the computer 500 to a network.
  • This network may be a LAN (Local Area Network) or a WAN (Wide Area Network).
  • the storage device 508 stores a program that implements each functional component of the detection device 2000 (a program that implements the application described above).
  • the processor 504 implements each functional component of the detection device 2000 by reading this program into the memory 506 and executing it.
  • the detection device 2000 may be realized by one computer 500 or may be realized by a plurality of computers 500. In the latter case, the configuration of each computer 500 need not be the same, and can be different.
  • Camera 40 is any camera that takes images and produces video data representing the results.
  • some or all of the functions of the detection device 2000 may be realized by the camera 40.
  • FIG. As such a camera, a camera called an intelligent camera, a network camera, an IP (Internet Protocol) camera, or the like can be used.
  • IP Internet Protocol
  • FIG. 4 is a flowchart illustrating the flow of processing executed by the detection device 2000 of the first embodiment.
  • the first acquisition unit 2020 acquires the predicted passage time information 60 (S102).
  • the second acquisition unit 2040 acquires the video data 50 (S104).
  • the detection unit 2060 detects the person to be monitored using the video data 50 and the predicted passage time information 60 (S106).
  • a plurality of vehicles 10 can depart from the departure place 20 at different times. For example, in the case of trains, trains may depart from the same platform every few minutes to several tens of minutes. Therefore, the detection device 2000 detects a person to be monitored for each vehicle 10 .
  • the number of cameras 40 is not limited to one, and cameras 40 can be installed at each of a plurality of locations. Therefore, the detection device 2000 detects a person to be monitored for each camera 40 . That is, the detection device 2000 detects a person to be monitored for each vehicle 10 and camera 40 pair.
  • the detection device 2000 uses the predicted passage time information 60 to grasp the predicted passage time for each vehicle 10 and camera 40 pair. Furthermore, the detection device 2000 performs the following processing for each pair of the vehicle 10 and camera 40 . First, the detection device 2000 acquires the video data 50 at the timing based on the predicted passage time grasped for the target pair. This timing is, for example, when a predetermined time (5 seconds, 10 seconds, etc.) has passed from the predicted passage time. Then, the detection device 2000 uses the predicted passage time for the target pair and the video data 50 acquired for the pair to detect the person to be monitored for the pair.
  • the first acquisition unit 2020 acquires the predicted passage time information 60 (S102).
  • the predicted passage time information 60 indicates the predicted passage time.
  • a plurality of cameras 40 may exist. Therefore, the predicted passage time information 60 indicates the predicted passage time for each camera 40 . That is, the predicted passage time information 60 indicates the predicted passage time in association with the pair of the vehicle 10 and camera 40 .
  • FIG. 5 is a diagram illustrating predicted passage time information 60.
  • the predicted passage time information 60 indicates a predicted passage time 66 in association with a pair of vehicle identification information 62 and camera identification information 64 .
  • Vehicle identification information 62 indicates the identification information of the vehicle 10 .
  • the camera identification information 64 indicates identification information of the camera 40 .
  • Predicted passage time 66 indicates the predicted passage time for the corresponding vehicle 10 and camera 40 pair.
  • the first row record in FIG. 5 indicates that the predicted passage time is Tp1 for the vehicle v001 and camera c001 pair. This indicates that if a person who passes through the imaging range of camera c001 at time Tp1 rushes to departure point 20, there is a high probability that he will board vehicle v001 just before departure.
  • the first acquisition unit 2020 acquires the predicted passage time information 60.
  • the predicted passage time information 60 is pre-stored in any storage unit in a form that can be obtained from the detection device 2000 .
  • the first acquisition unit 2020 acquires the predicted passage time information 60 stored in this storage unit.
  • the first acquisition unit 2020 may acquire the predicted passage time information 60 by receiving the predicted passage time information 60 transmitted from another device (for example, a generation device, which will be described later).
  • the predicted passage time information 60 can be generated based on the departure time of the vehicle 10 and the required time required for movement from the imaging range of the camera 40 to the departure place 20 .
  • the predicted passage time indicated by the predicted passage time information 60 can be calculated by subtracting the time required for movement from the imaging range of the camera 40 to the departure place 20 from the departure time of the vehicle 10 . For example, suppose the departure time of vehicle 10 is 10:10 and the required travel time is 5 minutes. In this case, the predicted passage time can be predicted to be 10:05.
  • a device having a function of generating predicted passage time information 60 specifies the departure time of vehicle 10 and the required time required for movement from the imaging range of camera 40 to departure place 20 .
  • the departure times of the vehicles 10 can be specified by, for example, obtaining operation schedule information (diagram information) indicating the departure times of the vehicles 10 departing from the departure place 20 .
  • the time required to move from the imaging range of the camera 40 to the departure location 20 can be calculated based on the positional relationship between the imaging range of the camera 40 and the departure location 20. For example, the generation device identifies the length of the movement route by identifying the movement route from the imaging range of the camera 40 to the departure place 20 . Then, the generating device calculates the time required for moving the route of the specified length as the required time required for moving from the imaging range of the camera 40 to the departure place 20 .
  • any route search algorithm can be used as a method of specifying the travel route between two specific points and its length.
  • the generation device acquires map data of the premises of the station where the departure place 20 is provided and the surroundings of the station. It is assumed that this map data indicates the position of the departure place 20 and the position of the camera 40 .
  • the generating device specifies the movement route from the camera 40 to the starting place 20 and its length by executing the shortest route search from the camera 40 to the starting place 20 for the map data.
  • the moving route from the camera 40 to the departure place 20 is not limited to the shortest route.
  • a travel route frequently used by the user of the vehicle 10 may be specified. may be used to generate the predicted passage time information 60.
  • the generation device calculates the required time by calculating “length of travel route/assumed travel speed”.
  • the speed of a person who is moving in a hurry for example, the average speed of a person who is running
  • a function that defines the relationship between the length of the travel route and the required time may be prepared in advance.
  • the generation device can obtain the required time by inputting the length of the identified travel route to the function.
  • the required time obtained from this function also represents the required time when the person moves in a hurry.
  • the generation device may calculate the required time considering the influence of these facilities. For example, the time required for movement within these facilities is fixed in advance. In this case, for example, the generation device adds the time required for movement in each piece of equipment present in the movement route to the required time calculated by the method described above.
  • the predicted passage time information 60 is generated using the timetable information at that time. After that, when the timetable information is updated (that is, when the timetable is revised), the updated timetable information is used before the operation of the vehicle 10 starts according to the revised timetable, and the predicted passage time information 60 is generated.
  • the vehicle 10 may be operated on a timetable that differs from the normal timetable.
  • the generation device may generate temporary predicted passage time information 60 using information indicating a temporary timetable.
  • the detection device 2000 uses this predicted passage time information 60 only until the timetable returns to normal.
  • the generation device may be provided integrally with the detection device 2000, or may be provided separately.
  • the former means that the detecting device 2000 has the function of operating as a generating device.
  • the detection device 2000 has a generation unit (not shown) that functions as a generation device.
  • the second acquisition unit 2040 acquires the video data 50 (S104).
  • the camera 40 stores the video data 50 in any storage unit in a manner that can be obtained from the detection device 2000 .
  • the second acquisition unit 2040 acquires the video data 50 from the storage unit.
  • the video data 50 may be transmitted from the camera 40 to the detection device 2000 .
  • the second acquisition unit 2040 may acquire part of the video data generated by the camera 40 as the video data 50.
  • the second acquisition unit 2040 acquires, as the video data 50, only the portion of the video data generated by the camera 40 that is a predetermined length of time before and after the predicted passage time.
  • the detection unit 2060 detects a person to be monitored from the video data 50 (S106). For example, the detection unit 2060 detects persons from the video data 50, and determines whether or not each detected person is a person to be monitored based on the predicted passage time. As described above, the detection unit 2060 detects a person to be monitored for each vehicle 10 and camera 40 pair. At this time, among the predicted passage times indicated in the predicted passage time information 60, the predicted passage time corresponding to the pair of the vehicle 10 and the camera 40 to be processed is used. Also, the video data 50 used is generated by the camera 40 to be processed.
  • the detection unit 2060 calculates a score focusing on each of the following three elements for each person to be determined.
  • element 1 The image is captured at a time close to the predicted passage time.
  • element 2 Heading to departure location 20 .
  • Condition 3 I am in a hurry.
  • scores focused on element 1, element 2, and element 3 are referred to as first score, second score, and third score, respectively. A specific calculation method for each score will be described later.
  • the detection unit 2060 uses the first to third scores to calculate a total score, and uses the total score to determine whether or not the person to be determined is a person requiring surveillance. For example, the detection unit 2060 determines that the person to be determined is a person requiring monitoring when the total score is equal to or greater than a threshold. On the other hand, when the total score is less than the threshold, the detection unit 2060 determines that the person to be determined is not a person requiring monitoring.
  • the total score is calculated as, for example, the sum, weighted sum, or product of the first to third scores. The weight given to each score is determined in advance.
  • the detection unit 2060 determines whether each of the first score to the third score is equal to or greater than the threshold, and if all of the scores are equal to or greater than the threshold, the person to be determined is the person to be monitored. Determine that there is. In other words, in this case, the person to be determined satisfies all of the conditions that "the image is captured at a time close to the predicted passing time", the condition that "the person is heading to the starting point 20", and the condition that the person is "in a hurry". In this case, it is determined that the person is a person requiring surveillance. On the other hand, if any of the scores is less than the threshold, the detection unit 2060 determines that the person to be determined is not a person requiring monitoring.
  • the threshold values for each score may be the same or different.
  • first to third scores need to be used.
  • the second score may not be used.
  • the detection unit 2060 calculates, as the first score, a value representing the degree to which the period in which the person to be determined is detected from the video data 50 is close to the predicted passage time. Specifically, the detection unit 2060 calculates a higher first score as the period during which the person to be determined is detected from the video data 50 is closer to the predicted passage time. For example, if the predicted passing time is included in the period in which the person to be determined is detected from the video data 50, the detection unit 2060 calculates the maximum first score for the person to be determined. On the other hand, if the predicted passage time is not included in the period in which the person to be determined is detected from the video data 50, the detecting unit 2060 assigns a smaller first score as the difference between the period and the predicted passage time increases. calculate.
  • the detection unit 2060 calculates the maximum first score for the person to be determined. For example, when the value range of the first score is 0 or more and 1 or less, the first score is 1. On the other hand, if Tp1 ⁇ Td1, the detection unit 2060 calculates a smaller first score as the value of Td1-Tp1 increases. Similarly, if Td2 ⁇ Tp1, the detection unit 2060 calculates a smaller first score as Tp1-Td2 is larger.
  • the detection unit 2060 calculates a value representing the degree of heading toward the starting point 20 as the second score.
  • the degree of heading toward the starting point 20 can be represented by the degree of matching between the movement direction of the person to be determined and the direction toward the starting point 20 .
  • the degree of matching is represented, for example, by the size of an angle formed by a vector representing the moving direction of the person to be determined and a vector representing the direction toward the departure place 20 .
  • the second score increases as the size of the formed angle is closer to 0°, and decreases as the angle is further away from 0°.
  • the correspondence relationship between the formed angle and the second score is determined in advance as a function, for example.
  • the detection unit 2060 calculates, as the third score, a value representing the degree to which the person to be determined is in a hurry.
  • the third score is predetermined according to the type of movement action of the person to be determined.
  • types such as “walking slowly”, “walking fast”, “running slowly”, and “running fast” can be defined as types of moving motions.
  • the third score is set to a larger value for a type of movement action that allows faster movement.
  • the 3rd score for "walking slowly” is the lowest
  • the 3rd score for "walking fast” is the second lowest
  • the 3rd score for "running slowly” is the third.
  • the third score of "running fast” is the maximum.
  • the types of movement motions are not limited to the four types described above. For example, two types of "walking" and “running” may be used.
  • the detection unit 2060 identifies the type of moving motion by analyzing the motion of the person to be determined detected from the video data 50 . Then, the detection unit 2060 uses the third score corresponding to the specified type of moving action as the third score of the person to be determined.
  • the detection unit 2060 detects feature points such as tips of hands and feet and joints of the person in each video frame in which the person to be determined is detected. Then, the detection unit 2060 identifies the type of movement action of the person to be determined by pattern recognition using the movement trajectory of these feature points as a feature amount.
  • the third score may be determined based on the speed of movement of the person to be determined. Specifically, the third score in this case is determined as a value that increases as the speed of movement of the person to be determined increases. A correspondence relationship between the speed of movement and the third score is determined in advance as a function, for example. In this case, the detection unit 2060 specifies the speed of movement of the person to be determined, and inputs the speed into the above function to calculate the third score. An existing technique can be used as a technique for calculating the movement speed of a person detected from video data.
  • the third score may be determined based on the characteristics of the facial expression of the person to be determined. For example, the contours of the eyes and mouth, or the position, size, or movement of the pupils can be used as facial expression features.
  • the detection unit 2060 calculates these feature amounts from one or more video frames in which the person to be determined is detected, and performs pattern recognition using these feature amounts to detect whether the person to be determined is in a hurry. Calculate the degree.
  • FIG. 6 is a block diagram illustrating the functional configuration of the detection device 2000 having the output section 2080. As shown in FIG.
  • the output information is information indicating that there is a person to be monitored.
  • This output information is transmitted to terminals used by, for example, station staff and security guards. By providing such output information, station staff, security guards, etc. can make advance warning announcements in preparation for last-minute boarding and prepare a security system.
  • the output destination terminal may be a portable terminal such as a smartphone, or a stationary terminal such as a PC.
  • the output information may include various information related to the person under surveillance.
  • the information related to the person to be monitored includes the position of the person to be monitored, the score calculated for the person to be monitored, information about the vehicle 10 that the person to be monitored is going to use (departure place, departure time, remaining time until departure time, etc.) , or destination, etc.), or the distance from the monitored person to the starting point 20, and the like.
  • the position of the person to be monitored is indicated using a map (such as a station map).
  • the output information includes a map superimposed with a mark representing the location of the person to be monitored.
  • the output unit 2080 may set the degree of emphasis of the output information based on the information regarding the person to be monitored. For example, ⁇ the greater the score calculated for the person to be monitored, the stronger the degree of emphasis'', ⁇ the closer the distance from the person to be monitored to the departure place 20, the stronger the degree of emphasis'', or ⁇ to the departure time of the vehicle 10 The shorter the remaining time is, the stronger the degree of emphasis is.”
  • a method of increasing the degree of emphasis for example, a method of using a more conspicuous color, a method of increasing the size, and the like are conceivable.
  • a method of increasing the volume may be considered as a method of increasing the degree of emphasis.
  • the output information is not limited to information about the person to be monitored. For example, using a speaker installed in a station premises, etc., output information that expresses a voice message to the effect that a person trying to rush to get on the train has been detected, or a voice message that urges people to refrain from rushing to get on the train, etc. may be output.
  • this voice message is preferably heard by the person under surveillance. Therefore, for example, the output unit 2080 causes a speaker provided near the camera 40 that generated the video data 50 in which the person to be monitored is detected to output a voice message. It is also preferable to use each speaker provided on the route from the camera 40 to the starting point 20 .
  • the output unit 2080 may transmit output information to a mobile terminal possessed by a person requiring monitoring.
  • the output information preferably includes guidance regarding alternative means of transportation.
  • information that activates an application that can search for means of transportation, or information that guides users to a website that can search for means of transportation may be sent as output information. According to these methods, it becomes easier for the person to be monitored to grasp alternative means of transportation, so that the probability that the person to be monitored will stop rushing to board the vehicle can be increased.
  • FIG. 7 is a diagram exemplifying output information transmitted to a mobile terminal possessed by a person requiring monitoring.
  • This output information includes a map showing the current position of the person to be monitored and the position of the taxi stand, and a message prompting the use of a taxi.
  • the output information is to be transmitted to the mobile terminal possessed by the person to be monitored, it is assumed that the information on the person is registered in advance. Specifically, a pair of a person's face information (facial feature amount) and a destination address is stored in advance in the storage unit in a manner that can be obtained from the detecting device 2000 .
  • the output unit 2080 searches for a registered person who matches the person to be monitored by matching the facial feature amount of the person to be monitored detected from the video data 50 with the facial feature amount of each registered person. do. When a registered person matching the person to be monitored is detected, the output unit 2080 transmits output information to the destination address associated with the registered person.
  • the detection device 2000 uses the video data 50 obtained from each camera 40 provided at the destination of the person (on the route toward the starting point 20) to detect the person to be monitored. may continue to be detected as a person to be monitored. In other words, the detection apparatus 2000 may determine whether each person to be monitored detected using each camera 40 is a person who has already been detected as a person to be monitored. When a person who has already been detected as a person requiring surveillance is detected again as a person requiring surveillance, output information to be output to the person newly detected as a person requiring surveillance is output information to be output to that person. A different one may be used.
  • Whether or not the persons to be monitored detected from different video data 50 match each other can be determined, for example, by comparing the facial features of the persons to be monitored detected from each video data 50 .
  • the program includes instructions (or software code) that, when read into a computer, cause the computer to perform one or more functions described in the embodiments.
  • the program may be stored in a non-transitory computer-readable medium or tangible storage medium.
  • computer readable media or tangible storage media may include random-access memory (RAM), read-only memory (ROM), flash memory, solid-state drives (SSD) or other memory technology, CDs - ROM, digital versatile disc (DVD), Blu-ray disc or other optical disc storage, magnetic cassette, magnetic tape, magnetic disc storage or other magnetic storage device.
  • the program may be transmitted on a transitory computer-readable medium or communication medium.
  • transitory computer readable media or communication media include electrical, optical, acoustic, or other forms of propagated signals.
  • Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle.
  • a first acquisition unit that acquires information
  • a second acquisition unit for acquiring video data generated by the camera
  • a detection unit that detects a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle from the video data using the predicted passage time information.
  • the detection unit For each person detected from the video data, the detection unit obtains a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, and Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. 2.
  • the detection unit calculates the third score for each person detected from the video data based on the type of movement of the person, the speed of movement of the person, or the facial expression of the person. , Supplementary Note 2.
  • a first obtaining step of obtaining information a second acquisition step of acquiring video data generated by the camera; and a detecting step of detecting from the video data a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle, using the predicted passage time information.
  • a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, the person is heading to the departure location; Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored.
  • appendix 7 The detection method according to appendix 7, wherein: (Appendix 9) In the detection step, for each person detected from the video data, the third score is calculated based on the type of movement of the person, the speed of movement of the person, or the characteristics of facial expression of the person. , appendix 8. (Appendix 10) Based on the positional relationship between the camera and the departure place, a predicted required time, which is the time required for movement from the imaging range of the camera to the departure place, is calculated, and the departure time and the predicted required time of the target vehicle. 10.
  • the detection method comprising a generating step of generating the predicted transit time information based on: (Appendix 11) An output that outputs output information indicating any one or more of information related to the person to be monitored detected by the detecting step, information indicating alternative means of transportation, and information facilitating search for alternative means of transportation. 11.
  • a detection method according to any one of clauses 7 to 10, comprising steps.
  • the detection method wherein the output information is output to the associated destination address. (Appendix 13) to the computer, Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle. a first obtaining step of obtaining information; a second acquisition step of acquiring video data generated by the camera; a detection step of detecting a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle from the video data using the predicted passage time information; and non-transitory computer-readable medium.
  • a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, the person is heading to the departure location; Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. 14.
  • the third score is calculated based on the type of movement of the person, the speed of movement of the person, or the characteristics of facial expression of the person. 15. The computer-readable medium of claim 14.
  • the computer-readable medium causing an output step of outputting output information indicative of: (Appendix 18) Information that associates a person's facial features with a destination address of the person is stored in a storage unit, In the output step, among the facial feature amounts stored in the storage unit, those that match the facial feature amount of the person to be monitored obtained from the video data are specified, and the specified feature amount is specified. 18.

Landscapes

  • Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • Economics (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • Tourism & Hospitality (AREA)
  • Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Traffic Control Systems (AREA)

Abstract

A detection device (2000) acquires predicted passing time information (60). The predicted passing time information (60) indicates a predicted passing time at which a person who will attempt to board a vehicle (10) immediately before the departure of the vehicle (10) is predicted to pass through the imaging range of a camera (40). The camera (40) is capable of capturing images of persons heading to the departure location (20). The detection device (2000) acquires video data (50) generated by the camera (40). The detection device (2000) uses the predicted passing time information (60) to detect, from the video data (50), a person to be monitored who is predicted to attempt to board the vehicle (10) immediately before the departure of the vehicle (10).

Description

検出装置、検出方法、及び非一時的なコンピュータ可読媒体DETECTION APPARATUS, DETECTION METHOD, AND NON-TRANSITARY COMPUTER-READABLE MEDIUM
 本開示は、安全な交通手段の提供をサポートする技術に関する。 This disclosure relates to technology that supports the provision of safe transportation.
 列車やバスなどといった乗物を利用した交通手段が提供されている。このような交通手段は、安全に提供されることが好ましい。そこで、安全な交通手段の提供をサポートする情報処理技術が開発されている。 A means of transportation using vehicles such as trains and buses is provided. Such transportation is preferably provided in a safe manner. Therefore, information processing technology has been developed to support the provision of safe means of transportation.
 例えば特許文献1は、列車のドア付近にいる人をセンサで検知し、当該センサの検知結果に基づいてその人の行動を予測することで、安全なタイミングでドアの開閉を行う技術を開示している。その他にも例えば、特許文献2は、列車のドア付近の人をセンサで検出し、その人の移動速度及び移動方向に基づいて、駆け込み乗車を検出する技術を開示している。 For example, Patent Document 1 discloses a technology for opening and closing the door at a safe timing by detecting a person near the door of a train with a sensor and predicting the person's behavior based on the detection result of the sensor. ing. In addition, for example, Patent Literature 2 discloses a technique of detecting a person near the door of a train with a sensor and detecting rushing boarding based on the moving speed and moving direction of the person.
特開2020-064570号公報JP 2020-064570 A 特開2013-052738号公報JP 2013-052738 A
 特許文献1と2の技術ではいずれも、列車のドア付近にいる人の行動しか監視することができない。本発明は上記の課題に鑑みてなされたものであり、その目的の一つは、交通機関の安全性を向上させることができる新たな技術を提供することである。 Both the technologies of Patent Documents 1 and 2 can only monitor the behavior of people near the train door. The present invention has been made in view of the above problems, and one of its purposes is to provide a new technique capable of improving the safety of transportation.
 本開示の検出装置は、対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得部と、前記カメラによって生成されるビデオデータを取得する第2取得部と、前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出部と、を有する。 The detection device of the present disclosure predicts that a person who is about to board the target vehicle just before the departure of the target vehicle passes through an imaging range of a camera capable of imaging a person heading to the departure place of the target vehicle. a first acquisition unit that acquires predicted passage time information indicating a passage time; a second acquisition unit that obtains video data generated by the camera; and a detection unit that detects a person to be monitored who is predicted to try to board the target vehicle just before departure of the vehicle.
 本開示の検出方法は、コンピュータによって実行される。当該検出方法は、対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得部と、前記カメラによって生成されるビデオデータを取得する第2取得部と、前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出部と、を有する。 The detection method of the present disclosure is executed by a computer. The detection method is a predicted passage time at which a person who is about to board the vehicle of interest is predicted to pass through an imaging range of a camera capable of imaging a person heading for the departure point of the vehicle of interest just before departure of the vehicle of interest. a first acquisition unit for acquiring predicted passage time information indicating the target vehicle from the video data using the predicted passage time information; and a detection unit that detects a person to be monitored who is predicted to try to board the vehicle of interest just before departure of the vehicle.
 本開示の非一時的なコンピュータ可読媒体は、コンピュータに本開示の検出方法を実行させるプログラムを格納している。 The non-transitory computer-readable medium of the present disclosure stores a program that causes a computer to execute the detection method of the present disclosure.
 本開示によれば、交通機関の安全性を向上させることができる新たな技術が提供される。 According to the present disclosure, new technology is provided that can improve the safety of transportation.
実施形態1の検出装置の動作の概要を例示する図である。4 is a diagram illustrating an overview of the operation of the detection device of Embodiment 1; FIG. 実施形態1の検出装置の機能構成を例示するブロック図である。2 is a block diagram illustrating the functional configuration of the detection device of Embodiment 1; FIG. 検出装置を実現するコンピュータのハードウエア構成を例示するブロック図である。It is a block diagram which illustrates the hardware constitutions of the computer which implement|achieves a detection apparatus. 実施形態1の検出装置によって実行される処理の流れを例示するフローチャートである。4 is a flow chart illustrating the flow of processing executed by the detection device of Embodiment 1. FIG. 予測通過時刻情報を例示する図である。It is a figure which illustrates predicted passage time information. 出力部を有する検出装置の機能構成を例示するブロック図である。3 is a block diagram illustrating the functional configuration of a detection device having an output section; FIG. 要監視人物が所持する携帯端末に対して送信される出力情報を例示する図である。FIG. 4 is a diagram illustrating output information transmitted to a mobile terminal possessed by a person requiring monitoring;
 以下では、本開示の実施形態について、図面を参照しながら詳細に説明する。各図面において、同一又は対応する要素には同一の符号が付されており、説明の明確化のため、必要に応じて重複説明は省略される。また、特に説明しない限り、所定値や基準値などといった予め定められている値は、その値を利用する装置からアクセス可能な態様で、任意の記憶部に予め格納されている。さらに、特に説明しない限り、記憶部は、1つ以上の任意の数の記憶装置によって構成される。 Below, embodiments of the present disclosure will be described in detail with reference to the drawings. In each drawing, the same reference numerals are given to the same or corresponding elements, and redundant description will be omitted as necessary for clarity of description. Further, unless otherwise described, predetermined values such as predetermined values and reference values are stored in advance in an arbitrary storage unit in a manner accessible from a device that uses the values. Further, unless otherwise specified, the storage unit is composed of one or more arbitrary number of storage devices.
[実施形態1]
<概要>
 図1は、実施形態1の検出装置2000の動作の概要を例示する図である。ここで、図1は、検出装置2000の概要の理解を容易にするための図であり、検出装置2000の動作は、図1に示したものに限定されない。
[Embodiment 1]
<Overview>
FIG. 1 is a diagram illustrating an overview of the operation of the detection device 2000 of Embodiment 1. FIG. Here, FIG. 1 is a diagram for facilitating understanding of the outline of the detection device 2000, and the operation of the detection device 2000 is not limited to that shown in FIG.
 検出装置2000は、乗物10の出発場所20へ向かう人の中から、出発間際に乗物10に乗ろうとする(いわゆる駆け込み乗車をする)と予測される人を検出する。以下、出発間際に乗物10に乗ろうとすると予測される人のことを、要監視人物と呼ぶ。 The detection device 2000 detects people who are predicted to try to board the vehicle 10 just before departure (so-called rush boarding) from among the people heading to the departure place 20 of the vehicle 10 . Hereinafter, a person predicted to try to board the vehicle 10 just before departure is referred to as a person to be monitored.
 乗物10は、交通手段の提供に用いられる乗物である。言い換えれば、乗物10は、顧客を乗せる乗物である。また、乗物10には、出発時刻が予め定められている。例えばこのような乗物10としては、電車やバスなどが挙げられる。ここで、「乗物10の出発間際」は、例えば、「乗物10の出発時刻から所定時間前以降」と定めることができる。例えば所定時間は、10秒などと定めることができる。 A vehicle 10 is a vehicle used to provide a means of transportation. In other words, vehicle 10 is a vehicle that carries customers. In addition, the vehicle 10 has a predetermined departure time. For example, such vehicles 10 include trains, buses, and the like. Here, "just before departure of the vehicle 10" can be defined as, for example, "after a predetermined time before the departure time of the vehicle 10". For example, the predetermined time can be set as 10 seconds.
 出発場所20は、乗物10が出発する場所である。乗物10が電車である場合、出発場所20は、その電車が停車しているプラットホームである。乗物10がバスである場合、出発場所20は、そのバスが停まっている停留所である。 The departure location 20 is the location from which the vehicle 10 departs. If the vehicle 10 is a train, the departure point 20 is the platform on which the train stops. If the vehicle 10 is a bus, the departure point 20 is the stop at which the bus stops.
 検出装置2000は、要監視人物の検出に、ビデオデータ50及び予測通過時刻情報60を利用する。ビデオデータ50は、出発場所20へ向かう人を撮像可能なカメラ40によって生成されるビデオデータである。カメラ40は、出発場所20へ向かう人が通る場所に設けられている。予測通過時刻情報60は、要監視人物がカメラ40の撮像範囲を通ると予測される時刻(以下、予測通過時刻)を示す。予測通過時刻は、「予測通過時刻にカメラ40の撮像範囲にいる人が、急いで出発場所20へ向かうと、出発間際に乗物10に乗ることになる時刻」とも言うことができる。 The detection device 2000 uses the video data 50 and the predicted passage time information 60 to detect the person to be monitored. The video data 50 is video data generated by the camera 40 capable of capturing images of people heading to the starting point 20 . A camera 40 is provided at a place where a person heading to the departure place 20 passes. The predicted passage time information 60 indicates the time at which the person to be monitored is predicted to pass through the imaging range of the camera 40 (hereinafter referred to as predicted passage time). The predicted passage time can also be said to be "the time at which a person who is within the imaging range of the camera 40 at the predicted passage time will get on the vehicle 10 just before departure if they hurry to the departure place 20".
 検出装置2000は、予測通過時刻情報60を利用して、ビデオデータ50に含まれる人(カメラ40によって撮像された人)の中から、要監視人物を検出する。要監視人物は、ビデオデータ50から検出される人のうち、予測通過時刻に近い時刻に撮像されており、なおかつ、急いで移動していると推測される人である。このような人を検出する具体的な方法については後述する。 The detection device 2000 uses the predicted passage time information 60 to detect persons to be monitored from among persons included in the video data 50 (persons captured by the camera 40). The person to be monitored is, among the persons detected from the video data 50, a person whose image is captured at a time close to the predicted passage time and who is assumed to be moving in a hurry. A specific method for detecting such a person will be described later.
<作用効果の一例>
 本実施形態の検出装置2000によれば、乗物10の出発場所20へ向かう人の中から、出発間際に乗物10に乗ると予測される人(駆け込み乗車をすると予測される人)が検出される。そのため、列車のドア付近にいる人の中から駆け込み乗車をする人を検出するケースよりも早いタイミングで、駆け込み乗車の可能性を把握することができる。そのため、交通機関の提供をより安全に行うことができるようになる。詳しくは後述するが、例えば、駅員が事前に注意喚起のアナウンスを行うことで、駆け込み乗車を控えてもらうといったことが可能となる。
<Example of action and effect>
According to the detection device 2000 of the present embodiment, a person who is predicted to board the vehicle 10 just before departure (a person who is predicted to board the vehicle 10 at the last minute) is detected from among the people heading to the departure point 20 of the vehicle 10. . Therefore, it is possible to grasp the possibility of rushing to board a train at an earlier timing than in the case of detecting a rushing boarding person from among the people near the train door. Therefore, transportation can be provided more safely. Although the details will be described later, for example, it is possible for the station attendant to make an announcement in advance to call attention, so that the trainees refrain from rushing to get on the train.
 以下、本実施形態の検出装置2000について、より詳細に説明する。 The detection device 2000 of this embodiment will be described in more detail below.
<機能構成の例>
 図2は、実施形態1の検出装置2000の機能構成を例示するブロック図である。第1取得部2020は予測通過時刻情報60を取得する。第2取得部2040はビデオデータ50を取得する。検出部2060は、予測通過時刻情報60を用いて、ビデオデータ50に含まれる人の中から、要監視人物を検出する。
<Example of functional configuration>
FIG. 2 is a block diagram illustrating the functional configuration of the detection device 2000 of Embodiment 1. As shown in FIG. A first acquisition unit 2020 acquires the predicted passage time information 60 . A second acquisition unit 2040 acquires the video data 50 . The detection unit 2060 detects a person to be monitored from persons included in the video data 50 using the predicted passage time information 60 .
<ハードウエア構成の例>
 検出装置2000の各機能構成部は、各機能構成部を実現するハードウエア(例:ハードワイヤードされた電子回路など)で実現されてもよいし、ハードウエアとソフトウエアとの組み合わせ(例:電子回路とそれを制御するプログラムの組み合わせなど)で実現されてもよい。以下、検出装置2000の各機能構成部がハードウエアとソフトウエアとの組み合わせで実現される場合について、さらに説明する。
<Example of hardware configuration>
Each functional configuration unit of the detection device 2000 may be implemented by hardware that implements each functional configuration unit (eg, hardwired electronic circuit, etc.), or a combination of hardware and software (eg, electronic A combination of a circuit and a program that controls it, etc.). A case in which each functional component of the detection device 2000 is implemented by a combination of hardware and software will be further described below.
 図3は、検出装置2000を実現するコンピュータ500のハードウエア構成を例示するブロック図である。コンピュータ500は、任意のコンピュータである。例えばコンピュータ500は、PC(Personal Computer)やサーバマシンなどといった、据え置き型のコンピュータである。その他にも例えば、コンピュータ500は、スマートフォンやタブレット端末などといった可搬型のコンピュータである。コンピュータ500は、検出装置2000を実現するために設計された専用のコンピュータであってもよいし、汎用のコンピュータであってもよい。 FIG. 3 is a block diagram illustrating the hardware configuration of the computer 500 that implements the detection device 2000. As shown in FIG. Computer 500 is any computer. For example, the computer 500 is a stationary computer such as a PC (Personal Computer) or a server machine. In addition, for example, the computer 500 is a portable computer such as a smart phone or a tablet terminal. Computer 500 may be a dedicated computer designed to implement detection apparatus 2000 or a general-purpose computer.
 例えば、コンピュータ500に対して所定のアプリケーションをインストールすることにより、コンピュータ500で、検出装置2000の各機能が実現される。上記アプリケーションは、検出装置2000の各機能構成部を実現するためのプログラムで構成される。なお、上記プログラムの取得方法は任意である。例えば、当該プログラムが格納されている記憶媒体(DVD ディスクや USB メモリなど)から、当該プログラムを取得することができる。その他にも例えば、当該プログラムが格納されている記憶部を管理しているサーバ装置から、当該プログラムをダウンロードすることにより、当該プログラムを取得することができる。 For example, by installing a predetermined application on the computer 500, the functions of the detection device 2000 are implemented on the computer 500. The application is composed of a program for realizing each functional component of the detection device 2000 . It should be noted that the acquisition method of the above program is arbitrary. For example, the program can be acquired from a storage medium (DVD disc, USB memory, etc.) in which the program is stored. In addition, for example, the program can be obtained by downloading the program from a server device that manages the storage unit storing the program.
 コンピュータ500は、バス502、プロセッサ504、メモリ506、ストレージデバイス508、入出力インタフェース510、及びネットワークインタフェース512を有する。バス502は、プロセッサ504、メモリ506、ストレージデバイス508、入出力インタフェース510、及びネットワークインタフェース512が、相互にデータを送受信するためのデータ伝送路である。ただし、プロセッサ504などを互いに接続する方法は、バス接続に限定されない。 Computer 500 has bus 502 , processor 504 , memory 506 , storage device 508 , input/output interface 510 and network interface 512 . The bus 502 is a data transmission path through which the processor 504, memory 506, storage device 508, input/output interface 510, and network interface 512 exchange data with each other. However, the method of connecting the processors 504 and the like to each other is not limited to bus connection.
 プロセッサ504は、CPU(Central Processing Unit)、GPU(Graphics Processing Unit)、又は FPGA(Field-Programmable Gate Array)などの種々のプロセッサである。メモリ506は、RAM(Random Access Memory)などを用いて実現される主記憶装置である。ストレージデバイス508は、ハードディスク、SSD(Solid State Drive)、メモリカード、又は ROM(Read Only Memory)などを用いて実現される補助記憶装置である。 The processor 504 is various processors such as a CPU (Central Processing Unit), GPU (Graphics Processing Unit), or FPGA (Field-Programmable Gate Array). The memory 506 is a main memory implemented using a RAM (Random Access Memory) or the like. The storage device 508 is an auxiliary storage device implemented using a hard disk, SSD (Solid State Drive), memory card, ROM (Read Only Memory), or the like.
 入出力インタフェース510は、コンピュータ500と入出力デバイスとを接続するためのインタフェースである。例えば入出力インタフェース510には、キーボードなどの入力装置や、ディスプレイ装置などの出力装置が接続される。 The input/output interface 510 is an interface for connecting the computer 500 and input/output devices. For example, the input/output interface 510 is connected to an input device such as a keyboard and an output device such as a display device.
 ネットワークインタフェース512は、コンピュータ500をネットワークに接続するためのインタフェースである。このネットワークは、LAN(Local Area Network)であってもよいし、WAN(Wide Area Network)であってもよい。 A network interface 512 is an interface for connecting the computer 500 to a network. This network may be a LAN (Local Area Network) or a WAN (Wide Area Network).
 ストレージデバイス508は、検出装置2000の各機能構成部を実現するプログラム(前述したアプリケーションを実現するプログラム)を記憶している。プロセッサ504は、このプログラムをメモリ506に読み出して実行することで、検出装置2000の各機能構成部を実現する。 The storage device 508 stores a program that implements each functional component of the detection device 2000 (a program that implements the application described above). The processor 504 implements each functional component of the detection device 2000 by reading this program into the memory 506 and executing it.
 検出装置2000は、1つのコンピュータ500で実現されてもよいし、複数のコンピュータ500で実現されてもよい。後者の場合において、各コンピュータ500の構成は同一である必要はなく、それぞれ異なるものとすることができる。 The detection device 2000 may be realized by one computer 500 or may be realized by a plurality of computers 500. In the latter case, the configuration of each computer 500 need not be the same, and can be different.
<カメラ40について>
 カメラ40は、撮像を行い、その結果を表すビデオデータを生成する任意のカメラである。ここで、検出装置2000の機能の一部又は全部は、カメラ40によって実現されてもよい。このようなカメラとしては、インテリジェントカメラ、ネットワークカメラ、又は IP(Internet Protocol)カメラなどと呼ばれるカメラを利用することができる。
<About camera 40>
Camera 40 is any camera that takes images and produces video data representing the results. Here, some or all of the functions of the detection device 2000 may be realized by the camera 40. FIG. As such a camera, a camera called an intelligent camera, a network camera, an IP (Internet Protocol) camera, or the like can be used.
<処理の流れ>
 図4は、実施形態1の検出装置2000によって実行される処理の流れを例示するフローチャートである。第1取得部2020は、予測通過時刻情報60を取得する(S102)。第2取得部2040は、ビデオデータ50を取得する(S104)。検出部2060は、ビデオデータ50及び予測通過時刻情報60を用いて、要監視人物を検出する(S106)。
<Process flow>
FIG. 4 is a flowchart illustrating the flow of processing executed by the detection device 2000 of the first embodiment. The first acquisition unit 2020 acquires the predicted passage time information 60 (S102). The second acquisition unit 2040 acquires the video data 50 (S104). The detection unit 2060 detects the person to be monitored using the video data 50 and the predicted passage time information 60 (S106).
 ここで、出発場所20からは、それぞれ異なる時刻に複数の乗物10が出発しうる。例えば電車の場合、同じプラットホームから、数分から数十分おきに電車が出発しうる。そこで検出装置2000は、乗物10ごとに要監視人物の検出を行う。また、カメラ40の数は1つに限らず、複数の場所それぞれにカメラ40が設置されうる。そこで検出装置2000は、カメラ40ごとに要監視人物の検出を行う。すなわち、検出装置2000は、乗物10とカメラ40のペアごとに、要監視人物の検出を行う。 Here, a plurality of vehicles 10 can depart from the departure place 20 at different times. For example, in the case of trains, trains may depart from the same platform every few minutes to several tens of minutes. Therefore, the detection device 2000 detects a person to be monitored for each vehicle 10 . In addition, the number of cameras 40 is not limited to one, and cameras 40 can be installed at each of a plurality of locations. Therefore, the detection device 2000 detects a person to be monitored for each camera 40 . That is, the detection device 2000 detects a person to be monitored for each vehicle 10 and camera 40 pair.
 より具体的には、検出装置2000は、予測通過時刻情報60を利用して乗物10とカメラ40のペアごとに予測通過時刻を把握する。さらに検出装置2000は、乗物10とカメラ40のペアごとに、以下の処理を行う。まず検出装置2000は、対象のペアについて把握した予測通過時刻に基づくタイミングで、ビデオデータ50を取得する。このタイミングは、例えば、予測通過時刻から所定時間(5秒や10秒など)が経過した時点である。そして検出装置2000は、対象のペアについての予測通過時刻、及び当該ペアについて取得したビデオデータ50を利用して、当該ペアについて要監視人物の検出を行う。 More specifically, the detection device 2000 uses the predicted passage time information 60 to grasp the predicted passage time for each vehicle 10 and camera 40 pair. Furthermore, the detection device 2000 performs the following processing for each pair of the vehicle 10 and camera 40 . First, the detection device 2000 acquires the video data 50 at the timing based on the predicted passage time grasped for the target pair. This timing is, for example, when a predetermined time (5 seconds, 10 seconds, etc.) has passed from the predicted passage time. Then, the detection device 2000 uses the predicted passage time for the target pair and the video data 50 acquired for the pair to detect the person to be monitored for the pair.
<予測通過時刻情報60の取得:S102>
 第1取得部2020は予測通過時刻情報60を取得する(S102)。予測通過時刻情報60は、予測通過時刻を示す。ここで、出発場所20から出発する乗物10は複数存在しうる。そこで予測通過時刻情報60は、乗物10ごとに予測通過時刻を示す。また、カメラ40が複数存在しうる。そこで予測通過時刻情報60は、カメラ40ごとに予測通過時刻を示す。すなわち、予測通過時刻情報60は、乗物10とカメラ40のペアに対応づけて、予測通過時刻を示す。
<Obtaining Predicted Passing Time Information 60: S102>
The first acquisition unit 2020 acquires the predicted passage time information 60 (S102). The predicted passage time information 60 indicates the predicted passage time. Here, there may be multiple vehicles 10 departing from the departure location 20 . Therefore, the predicted passage time information 60 indicates the predicted passage time for each vehicle 10 . Also, a plurality of cameras 40 may exist. Therefore, the predicted passage time information 60 indicates the predicted passage time for each camera 40 . That is, the predicted passage time information 60 indicates the predicted passage time in association with the pair of the vehicle 10 and camera 40 .
 図5は、予測通過時刻情報60を例示する図である。予測通過時刻情報60は、乗物識別情報62とカメラ識別情報64のペアに対応づけて、予測通過時刻66を示す。乗物識別情報62は、乗物10の識別情報を示す。カメラ識別情報64は、カメラ40の識別情報を示す。予測通過時刻66は、対応する乗物10とカメラ40のペアについて、予測通過時刻を示す。例えば図5の1行目のレコードは、乗物 v001 とカメラ c001 のペアについて、予測通過時刻が Tp1 であることを示している。これは、カメラ c001 の撮像範囲を時刻 Tp1 に通過する人が急いで出発場所20に向かうと、出発間際の乗物 v001 に乗り込む蓋然性が高いことを表している。 FIG. 5 is a diagram illustrating predicted passage time information 60. FIG. The predicted passage time information 60 indicates a predicted passage time 66 in association with a pair of vehicle identification information 62 and camera identification information 64 . Vehicle identification information 62 indicates the identification information of the vehicle 10 . The camera identification information 64 indicates identification information of the camera 40 . Predicted passage time 66 indicates the predicted passage time for the corresponding vehicle 10 and camera 40 pair. For example, the first row record in FIG. 5 indicates that the predicted passage time is Tp1 for the vehicle v001 and camera c001 pair. This indicates that if a person who passes through the imaging range of camera c001 at time Tp1 rushes to departure point 20, there is a high probability that he will board vehicle v001 just before departure.
 第1取得部2020が予測通過時刻情報60を取得する方法は様々である。例えば予測通過時刻情報60は、検出装置2000から取得可能な態様で、任意の記憶部に予め格納されている。この場合、第1取得部2020は、この記憶部に格納されている予測通過時刻情報60を取得する。その他にも例えば、第1取得部2020は、他の装置(例えば後述する、生成装置)から送信された予測通過時刻情報60を受信することで、予測通過時刻情報60を取得してもよい。 There are various methods for the first acquisition unit 2020 to acquire the predicted passage time information 60. For example, the predicted passage time information 60 is pre-stored in any storage unit in a form that can be obtained from the detection device 2000 . In this case, the first acquisition unit 2020 acquires the predicted passage time information 60 stored in this storage unit. Alternatively, for example, the first acquisition unit 2020 may acquire the predicted passage time information 60 by receiving the predicted passage time information 60 transmitted from another device (for example, a generation device, which will be described later).
<予測通過時刻情報60の生成>
 ここで、予測通過時刻情報60の生成方法について例示する。予測通過時刻情報60は、乗物10の出発時刻と、カメラ40の撮像範囲から出発場所20までの移動に要する所要時間とに基づいて生成することができる。具体的には、予測通過時刻情報60が示す予測通過時刻は、乗物10の出発時刻から、カメラ40の撮像範囲から出発場所20までの移動に要する時間を引くことで算出することができる。例えば、乗物10の出発時刻が 10:10 であり、移動の所要時間が5分であるとする。この場合、予測通過時刻は 10:05 であると予測できる。
<Generation of predicted passage time information 60>
Here, a method for generating the predicted passage time information 60 will be illustrated. The predicted passage time information 60 can be generated based on the departure time of the vehicle 10 and the required time required for movement from the imaging range of the camera 40 to the departure place 20 . Specifically, the predicted passage time indicated by the predicted passage time information 60 can be calculated by subtracting the time required for movement from the imaging range of the camera 40 to the departure place 20 from the departure time of the vehicle 10 . For example, suppose the departure time of vehicle 10 is 10:10 and the required travel time is 5 minutes. In this case, the predicted passage time can be predicted to be 10:05.
 そこで、予測通過時刻情報60を生成する機能を有する装置(以下、生成装置)は、乗物10の出発時刻と、カメラ40の撮像範囲から出発場所20までの移動に要する所要時間とを特定する。乗物10の出発時刻は、例えば、出発場所20から出発する各乗物10の出発時刻を示す運行スケジュール情報(ダイヤ情報)を取得することによって特定することができる。 Therefore, a device having a function of generating predicted passage time information 60 (hereinafter referred to as generation device) specifies the departure time of vehicle 10 and the required time required for movement from the imaging range of camera 40 to departure place 20 . The departure times of the vehicles 10 can be specified by, for example, obtaining operation schedule information (diagram information) indicating the departure times of the vehicles 10 departing from the departure place 20 .
 カメラ40の撮像範囲から出発場所20までの移動に要する所要時間は、カメラ40の撮像範囲と出発場所20との位置関係に基づいて算出することができる。例えば生成装置は、カメラ40の撮像範囲から出発場所20までの移動経路を特定することで、その移動経路の長さを特定する。そして、生成装置は、特定した長さの経路の移動に要する時間を、カメラ40の撮像範囲から出発場所20までの移動に要する所要時間として算出する。 The time required to move from the imaging range of the camera 40 to the departure location 20 can be calculated based on the positional relationship between the imaging range of the camera 40 and the departure location 20. For example, the generation device identifies the length of the movement route by identifying the movement route from the imaging range of the camera 40 to the departure place 20 . Then, the generating device calculates the time required for moving the route of the specified length as the required time required for moving from the imaging range of the camera 40 to the departure place 20 .
 ここで、特定の2つの地点の間の移動経路及びその長さを特定する方法には、例えば、任意の経路探索アルゴリズムを利用することができる。例えば生成装置は、出発場所20が設けられている駅の構内や駅の周辺の地図データを取得する。この地図データには、出発場所20の位置及びカメラ40の位置が示されているとする。例えば生成装置は、当該地図データについて、カメラ40から出発場所20までの最短経路探索を実行することにより、カメラ40から出発場所20までの移動経路及びその長さを特定する。 Here, for example, any route search algorithm can be used as a method of specifying the travel route between two specific points and its length. For example, the generation device acquires map data of the premises of the station where the departure place 20 is provided and the surroundings of the station. It is assumed that this map data indicates the position of the departure place 20 and the position of the camera 40 . For example, the generating device specifies the movement route from the camera 40 to the starting place 20 and its length by executing the shortest route search from the camera 40 to the starting place 20 for the map data.
 なお、カメラ40から出発場所20までの移動経路は、最短経路に限定されない。例えば、これまでの駅などの利用実績から、乗物10の利用者によって頻繁に利用される移動経路(特に、急いでいる人によって頻繁に利用される移動経路)を特定しておき、この移動経路を予測通過時刻情報60の生成に利用してもよい。 Note that the moving route from the camera 40 to the departure place 20 is not limited to the shortest route. For example, based on past usage records of stations, etc., a travel route frequently used by the user of the vehicle 10 (particularly, a travel route frequently used by a person in a hurry) may be specified. may be used to generate the predicted passage time information 60.
 移動経路の長さから所要時間を算出する方法は様々である。例えば、要監視人物の想定移動速度を予め定めておく。この場合、生成装置は、「移動経路の長さ/想定移動速度」という計算により、所要時間を算出する。ここで、要監視人物は、急いで(例えば、走って)移動している蓋然性が高い。そこで、想定移動速度には、急いで移動している人の速さ(例えば、走って移動する人の平均的な速さ)を設定することが好適である。 There are various methods for calculating the required time from the length of the travel route. For example, an assumed moving speed of the person to be monitored is determined in advance. In this case, the generation device calculates the required time by calculating “length of travel route/assumed travel speed”. Here, there is a high probability that the person to be monitored is moving quickly (for example, running). Therefore, it is preferable to set the speed of a person who is moving in a hurry (for example, the average speed of a person who is running) as the assumed moving speed.
 その他にも例えば、移動経路の長さと所要時間との関係を定めた関数が予め用意されていてもよい。この場合、生成装置は、当該関数に対して、特定した移動経路の長さを入力することにより、所要時間を得ることができる。なお、前述したように、要監視人物は急いで移動している蓋然性が高いため、この関数から得られる所要時間も、人が急いで移動した場合についての所要時間を表すことが好適である。 In addition, for example, a function that defines the relationship between the length of the travel route and the required time may be prepared in advance. In this case, the generation device can obtain the required time by inputting the length of the identified travel route to the function. As described above, it is highly probable that the person to be monitored is moving in a hurry, so it is preferable that the required time obtained from this function also represents the required time when the person moves in a hurry.
 なお、移動経路上には、エスカレーター、エレベーター、又は階段などといった設備が設けられていることがある。生成装置は、これらの設備の影響を考慮して所要時間を算出してもよい。例えば、これらの設備の中での移動に要する時間は、予め固定で定めておく。この場合、例えば生成装置は、前述した方法で算出した所要時間に対し、移動経路の中に存在する設備それぞれの中での移動に要する時間を加算する。 In addition, there may be facilities such as escalators, elevators, or stairs on the movement route. The generation device may calculate the required time considering the influence of these facilities. For example, the time required for movement within these facilities is fixed in advance. In this case, for example, the generation device adds the time required for movement in each piece of equipment present in the movement route to the required time calculated by the method described above.
 予測通過時刻情報60を生成するタイミングは様々である。例えばまず、検出装置2000の運用を開始する前に、その時点におけるダイヤ情報を利用して、予測通過時刻情報60を生成しておく。その後、ダイヤ情報が更新された場合(すなわち、ダイヤが改正された場合)、改正後のダイヤによる乗物10の運行が開始される前に、更新されたダイヤ情報を利用して、予測通過時刻情報60を生成しておく。 There are various timings for generating the predicted passage time information 60. For example, first, before starting the operation of the detecting device 2000, the predicted passage time information 60 is generated using the timetable information at that time. After that, when the timetable information is updated (that is, when the timetable is revised), the updated timetable information is used before the operation of the vehicle 10 starts according to the revised timetable, and the predicted passage time information 60 is generated.
 また、特別なイベントの開催や事故などのアクシデントにより、通常のダイヤとは異なるダイヤで乗物10が運行されることがある。このような場合、生成装置は、臨時のダイヤを示す情報を利用して、臨時の予測通過時刻情報60を生成してもよい。検出装置2000は、この予測通過時刻情報60を、通常のダイヤに戻るまでの間だけ利用する。 In addition, due to special events or accidents such as accidents, the vehicle 10 may be operated on a timetable that differs from the normal timetable. In such a case, the generation device may generate temporary predicted passage time information 60 using information indicating a temporary timetable. The detection device 2000 uses this predicted passage time information 60 only until the timetable returns to normal.
 生成装置は、検出装置2000と一体として設けられてもよいし、別体として設けられてもよい。前者は、生成装置として動作する機能を検出装置2000に持たせることを意味する。この場合、検出装置2000は、生成装置としての機能を実現する生成部を有する(図示せず)。 The generation device may be provided integrally with the detection device 2000, or may be provided separately. The former means that the detecting device 2000 has the function of operating as a generating device. In this case, the detection device 2000 has a generation unit (not shown) that functions as a generation device.
<ビデオデータ50の取得:S104>
 第2取得部2040は、ビデオデータ50を取得する(S104)。第2取得部2040がビデオデータ50を取得する方法は様々である。例えばカメラ40が、検出装置2000から取得可能な態様で、ビデオデータ50を任意の記憶部に格納する。この場合、第2取得部2040は、当該記憶部からビデオデータ50を取得する。その他にも例えば、ビデオデータ50は、カメラ40から検出装置2000へ送信されてもよい。
<Acquisition of video data 50: S104>
The second acquisition unit 2040 acquires the video data 50 (S104). There are various methods for the second acquisition unit 2040 to acquire the video data 50 . For example, the camera 40 stores the video data 50 in any storage unit in a manner that can be obtained from the detection device 2000 . In this case, the second acquisition unit 2040 acquires the video data 50 from the storage unit. Alternatively, for example, the video data 50 may be transmitted from the camera 40 to the detection device 2000 .
 なお、第2取得部2040は、カメラ40によって生成されたビデオデータの一部を、ビデオデータ50として取得してもよい。例えば第2取得部2040は、カメラ40によって生成されたビデオデータのうち、予測通過時刻の前後所定時間の長さの部分のみを、ビデオデータ50として取得する。 Note that the second acquisition unit 2040 may acquire part of the video data generated by the camera 40 as the video data 50. For example, the second acquisition unit 2040 acquires, as the video data 50, only the portion of the video data generated by the camera 40 that is a predetermined length of time before and after the predicted passage time.
<要監視人物の検出:S106>
 検出部2060は、ビデオデータ50から要監視人物を検出する(S106)。例えば検出部2060は、ビデオデータ50から人物を検出し、予測通過時刻に基づいて、検出された各人物が要監視人物であるか否かの判定を行う。ここで前述したように、検出部2060は、乗物10とカメラ40のペアごとに、要監視人物の検出を行う。この際、予測通過時刻は、予測通過時刻情報60に示されている予測通過時刻のうち、処理対象としている乗物10とカメラ40のペアに対応するものが利用される。また、ビデオデータ50は、処理対象としているカメラ40によって生成されたものが利用される。
<Detection of person requiring surveillance: S106>
The detection unit 2060 detects a person to be monitored from the video data 50 (S106). For example, the detection unit 2060 detects persons from the video data 50, and determines whether or not each detected person is a person to be monitored based on the predicted passage time. As described above, the detection unit 2060 detects a person to be monitored for each vehicle 10 and camera 40 pair. At this time, among the predicted passage times indicated in the predicted passage time information 60, the predicted passage time corresponding to the pair of the vehicle 10 and the camera 40 to be processed is used. Also, the video data 50 used is generated by the camera 40 to be processed.
 例えば検出部2060は、判定対象の各人物について、以下の3つの要素それぞれに着目したスコアを算出する。
(要素1)予測通過時刻に近い時刻に撮像されている。
(要素2)出発場所20へ向かっている。
(条件3)急いでいる。
 以下、要素1、要素2、及び要素3に着目したスコアをそれぞれ、第1スコア、第2スコア、及び第3スコアと呼ぶ。各スコアの具体的な算出方法については後述する。
For example, the detection unit 2060 calculates a score focusing on each of the following three elements for each person to be determined.
(Element 1) The image is captured at a time close to the predicted passage time.
(Element 2) Heading to departure location 20 .
(Condition 3) I am in a hurry.
Hereinafter, scores focused on element 1, element 2, and element 3 are referred to as first score, second score, and third score, respectively. A specific calculation method for each score will be described later.
 例えば検出部2060は、第1スコアから第3スコアを用いて総合スコアを算出し、総合スコアを用いて、判定対象の人物が要監視人物であるか否かを判定する。例えば検出部2060は、総合スコアが閾値以上である場合に、判定対象の人物が要監視人物であると判定する。一方、総合スコアが閾値未満である場合、検出部2060は、判定対象の人物が要監視人物でないと判定する。なお、総合スコアは、例えば第1スコアから第3スコアの和、重み付き和、又は積などとして算出される。各スコアに与える重みは、予め定めておく。 For example, the detection unit 2060 uses the first to third scores to calculate a total score, and uses the total score to determine whether or not the person to be determined is a person requiring surveillance. For example, the detection unit 2060 determines that the person to be determined is a person requiring monitoring when the total score is equal to or greater than a threshold. On the other hand, when the total score is less than the threshold, the detection unit 2060 determines that the person to be determined is not a person requiring monitoring. Note that the total score is calculated as, for example, the sum, weighted sum, or product of the first to third scores. The weight given to each score is determined in advance.
 その他にも例えば、検出部2060は、第1スコアから第3スコアのそれぞれが閾値以上であるか否かを判定し、いずれもが閾値以上である場合に、判定対象の人物が要監視人物であると判定する。すなわちこの場合、判定対象の人物は、「予測通過時刻に近い時刻に撮像されている」という条件、「出発場所20へ向かっている」という条件、及び「急いでいる」という条件の全てを満たす場合に、要監視人物であると判定される。一方、各スコアのいずれかが閾値未満である場合、検出部2060は、判定対象の人物が要監視人物でないと判定する。ここで、各スコアの閾値は、全て同一であってもよいし、それぞれ異なっていてもよい。 In addition, for example, the detection unit 2060 determines whether each of the first score to the third score is equal to or greater than the threshold, and if all of the scores are equal to or greater than the threshold, the person to be determined is the person to be monitored. Determine that there is. In other words, in this case, the person to be determined satisfies all of the conditions that "the image is captured at a time close to the predicted passing time", the condition that "the person is heading to the starting point 20", and the condition that the person is "in a hurry". In this case, it is determined that the person is a person requiring surveillance. On the other hand, if any of the scores is less than the threshold, the detection unit 2060 determines that the person to be determined is not a person requiring monitoring. Here, the threshold values for each score may be the same or different.
 なお、必ずしも第1スコアから第3スコアの全てが利用される必要はない。例えば、カメラ40によって撮像されている通路が一方通行である場合などのように、各人物の移動方向が明らかな場合には、第2スコアを利用しなくてもよい。 It should be noted that not all of the first to third scores need to be used. For example, when the direction of movement of each person is clear, such as when the passage captured by the camera 40 is one-way, the second score may not be used.
 以下、第1スコアから第3スコアのそれぞれについて、その算出方法を例示する。 Below, the calculation method for each of the first to third scores will be exemplified.
<<第1スコアについて>>
 検出部2060は、第1スコアとして、判定対象の人物がビデオデータ50から検出された期間が予測通過時刻に近い度合いを表す値を算出する。具体的には、検出部2060は、判定対象の人物がビデオデータ50から検出された期間が予測通過時刻により近いほど、より高い第1スコアを算出する。例えば検出部2060は、ビデオデータ50から判定対象の人物が検出された期間の中に予測通過時刻が含まていれば、当該判定対象の人物について、最大の第1スコアを算出する。一方、ビデオデータ50から判定対象の人物が検出された期間の中に予測通過時刻が含まれていない場合、検出部2060は、当該期間と予測通過時刻との乖離が大きいほど小さい第1スコアを算出する。
<<About the first score>>
The detection unit 2060 calculates, as the first score, a value representing the degree to which the period in which the person to be determined is detected from the video data 50 is close to the predicted passage time. Specifically, the detection unit 2060 calculates a higher first score as the period during which the person to be determined is detected from the video data 50 is closer to the predicted passage time. For example, if the predicted passing time is included in the period in which the person to be determined is detected from the video data 50, the detection unit 2060 calculates the maximum first score for the person to be determined. On the other hand, if the predicted passage time is not included in the period in which the person to be determined is detected from the video data 50, the detecting unit 2060 assigns a smaller first score as the difference between the period and the predicted passage time increases. calculate.
 例えば、予測通過時刻が Tp1 であるとする。また、ビデオデータ50から判定対象の人物が検出された期間が時刻 Td1 から時刻 Td2 であるとする。この場合、Td1<=Tp1<=Td2 であれば、検出部2060は、判定対象の人物について、最大の第1スコアを算出する。例えば第1スコアの値域が0以上1以下の値である場合、第1スコアは1となる。これに対し、Tp1<Td1 であれば、検出部2060は、Td1-Tp1 の値が大きいほど小さい第1スコアを算出する。同様に、Td2<Tp1 であれば、検出部2060は、Tp1-Td2 が大きいほど小さい第1スコアを算出する。 For example, suppose the predicted passage time is Tp1. It is also assumed that the period during which the person to be determined is detected from the video data 50 is from time Td1 to time Td2. In this case, if Td1<=Tp1<=Td2, the detection unit 2060 calculates the maximum first score for the person to be determined. For example, when the value range of the first score is 0 or more and 1 or less, the first score is 1. On the other hand, if Tp1<Td1, the detection unit 2060 calculates a smaller first score as the value of Td1-Tp1 increases. Similarly, if Td2<Tp1, the detection unit 2060 calculates a smaller first score as Tp1-Td2 is larger.
 なお、ビデオデータ50から人物を検出する処理を行う際、予測通過時刻の前後所定期間のビデオフレームのみが処理の対象とされてもよい。こうすることで、要監視人物ではないことが明らかな人物を、スコアの算出対象から除外することができる。 It should be noted that when performing the process of detecting a person from the video data 50, only video frames in a predetermined period before and after the predicted passage time may be processed. By doing so, a person who is clearly not a person requiring monitoring can be excluded from score calculation targets.
<<第2スコアについて>>
 検出部2060は、第2スコアとして、出発場所20へ向かっている度合いを表す値を算出する。出発場所20へ向かっている度合いは、判定対象の人物の移動方向と、出発場所20へ向かう方向との一致度合いで表すことができる。この一致度合いは、例えば、判定対象の人物の移動方向を表すベクトルと、出発場所20へ向かう方向を表すベクトルとの成す角の大きさで表される。第2スコアは、上記成す角の大きさが0°に近いほど大きく、0°から離れるほど小さくなる。成す角と第2スコアとの対応関係は、例えば、関数として予め定めておく。
<<About the second score>>
The detection unit 2060 calculates a value representing the degree of heading toward the starting point 20 as the second score. The degree of heading toward the starting point 20 can be represented by the degree of matching between the movement direction of the person to be determined and the direction toward the starting point 20 . The degree of matching is represented, for example, by the size of an angle formed by a vector representing the moving direction of the person to be determined and a vector representing the direction toward the departure place 20 . The second score increases as the size of the formed angle is closer to 0°, and decreases as the angle is further away from 0°. The correspondence relationship between the formed angle and the second score is determined in advance as a function, for example.
<<第3スコアについて>>
 検出部2060は、第3スコアとして、判定対象の人物が急いでいる度合いを表す値を算出する。例えば第3スコアは、判定対象の人物の移動動作の種類に応じて、予め定められている。移動動作の種類としては、例えば、「ゆっくり歩いている」、「速めに歩いている」、「ゆっくり走っている」、及び「速めに走っている」といった種類を定めることができる。第3スコアは、より速く移動できる種類の移動動作ほど大きい値に設定される。上記の例では、「ゆっくり歩いている」の第3スコアが最小であり、「速めに歩いている」の第3スコアが2番目に小さく、「ゆっくり走っている」の第3スコアが3番目に小さく、「速めに走っている」の第3スコアが最大となる。なお、移動動作の種類は上記4種類に限定されない。例えば、「歩いている」と「走っている」の2種類にしてもよい。
<<About the third score>>
The detection unit 2060 calculates, as the third score, a value representing the degree to which the person to be determined is in a hurry. For example, the third score is predetermined according to the type of movement action of the person to be determined. For example, types such as "walking slowly", "walking fast", "running slowly", and "running fast" can be defined as types of moving motions. The third score is set to a larger value for a type of movement action that allows faster movement. In the example above, the 3rd score for "walking slowly" is the lowest, the 3rd score for "walking fast" is the second lowest, and the 3rd score for "running slowly" is the third. The third score of "running fast" is the maximum. Note that the types of movement motions are not limited to the four types described above. For example, two types of "walking" and "running" may be used.
 検出部2060は、ビデオデータ50から検出される判定対象の人物の動作を解析することにより、その移動動作の種類を特定する。そして、検出部2060は、特定された移動動作の種類に対応する第3スコアを、当該判定対象の人物の第3スコアとして利用する。 The detection unit 2060 identifies the type of moving motion by analyzing the motion of the person to be determined detected from the video data 50 . Then, the detection unit 2060 uses the third score corresponding to the specified type of moving action as the third score of the person to be determined.
 ここで、移動動作の種類を特定する方法は様々である。例えば検出部2060は、判定対象の人物が検出された各ビデオフレームについて、当該人物の手足の先端や関節などの特徴点を検出する。そして、検出部2060は、これらの特徴点の移動軌跡を特徴量としたパターン認識により、判定対象の人物の移動動作の種類を特定する。 Here, there are various methods for identifying the type of movement motion. For example, the detection unit 2060 detects feature points such as tips of hands and feet and joints of the person in each video frame in which the person to be determined is detected. Then, the detection unit 2060 identifies the type of movement action of the person to be determined by pattern recognition using the movement trajectory of these feature points as a feature amount.
 第3スコアは、判定対象の人物の移動の速さに基づいて定められてもよい。具体的には、この場合の第3スコアは、判定対象の人物の移動の速さが速いほど大きい値として定められる。移動の速さと第3スコアとの対応関係は、例えば、予め関数として定められている。この場合、検出部2060は、判定対象の人物の移動の速さを特定し、その速さを上記関数に入力することで、第3スコアを算出する。なお、ビデオデータから検出された人物についてその移動の速さを算出する技術には、既存の技術を利用することができる。 The third score may be determined based on the speed of movement of the person to be determined. Specifically, the third score in this case is determined as a value that increases as the speed of movement of the person to be determined increases. A correspondence relationship between the speed of movement and the third score is determined in advance as a function, for example. In this case, the detection unit 2060 specifies the speed of movement of the person to be determined, and inputs the speed into the above function to calculate the third score. An existing technique can be used as a technique for calculating the movement speed of a person detected from video data.
 その他にも例えば、第3スコアは、判定対象の人物の表情の特徴に基づいて定められてもよい。表情の特徴量としては、例えば、目や口の輪郭、又は瞳孔の位置、大きさ、若しくは動きなどを利用することができる。検出部2060は、判定対象の人物が検出された1つ以上のビデオフレームからこれらの特徴量を算出し、これらの特徴量を利用したパターン認識を行うことで、判定対象の人物が急いでいる度合いを算出する。 In addition, for example, the third score may be determined based on the characteristics of the facial expression of the person to be determined. For example, the contours of the eyes and mouth, or the position, size, or movement of the pupils can be used as facial expression features. The detection unit 2060 calculates these feature amounts from one or more video frames in which the person to be determined is detected, and performs pattern recognition using these feature amounts to detect whether the person to be determined is in a hurry. Calculate the degree.
<検出結果の利用方法>
 要監視人物が検出された場合、検出装置2000は、要監視人物に関する情報を出力することが好適である。以下、この情報を出力情報と呼ぶ。また、出力情報の出力を行う機能構成部を、出力部と呼ぶ。図6は、出力部2080を有する検出装置2000の機能構成を例示するブロック図である。
<How to use detection results>
When a person to be monitored is detected, the detecting device 2000 preferably outputs information about the person to be monitored. This information is hereinafter referred to as output information. A functional configuration unit that outputs output information is called an output unit. FIG. 6 is a block diagram illustrating the functional configuration of the detection device 2000 having the output section 2080. As shown in FIG.
 出力情報には様々な情報を採用できる。例えば出力情報は、要監視人物が存在することを示す情報である。この出力情報は、例えば、駅員や警備員などが利用する端末へ送信される。このような出力情報を提供することにより、駅員や警備員などは、駆け込み乗車に備えて事前に注意喚起のアナウンスを行ったり、警備体制を整えたりすることができる。なお、出力先の端末は、スマートフォンなどの可搬型の端末であってもよいし、PC などの据え置き型の端末であってもよい。 Various information can be adopted as the output information. For example, the output information is information indicating that there is a person to be monitored. This output information is transmitted to terminals used by, for example, station staff and security guards. By providing such output information, station staff, security guards, etc. can make advance warning announcements in preparation for last-minute boarding and prepare a security system. The output destination terminal may be a portable terminal such as a smartphone, or a stationary terminal such as a PC.
 出力情報は、要監視人物が存在することに加え、要監視人物に関連する種々の情報を含んでもよい。例えば要監視人物に関連する情報は、要監視人物の位置、要監視人物について算出されたスコア、要監視人物が利用しようとしている乗物10に関する情報(出発場所、出発時刻、出発時刻までの残り時間、又は目的地など)、又は要監視人物から出発場所20までの距離などである。ここで、要監視人物の位置は、地図(駅の構内図など)を利用して示されることが好適である。例えば、要監視人物の位置を表すマークが重畳された地図が、出力情報に含められる。 In addition to the presence of the person under surveillance, the output information may include various information related to the person under surveillance. For example, the information related to the person to be monitored includes the position of the person to be monitored, the score calculated for the person to be monitored, information about the vehicle 10 that the person to be monitored is going to use (departure place, departure time, remaining time until departure time, etc.) , or destination, etc.), or the distance from the monitored person to the starting point 20, and the like. Here, it is preferable that the position of the person to be monitored is indicated using a map (such as a station map). For example, the output information includes a map superimposed with a mark representing the location of the person to be monitored.
 ここで、出力部2080は、要監視人物に関する情報に基づいて、出力情報の強調度合いを設定してもよい。例えば、「要監視人物について算出されたスコアが大きいほど強調度合いを強くする」、「要監視人物から出発場所20までの距離が近いほど強調度合いを強くする」、又は「乗物10の出発時刻までの残り時間が短いほど強調度合いを強くする」などといった方法が考えられる。強調度合いを強くする方法としては、例えば、より目立つ色を利用する方法や、大きさをより大きくする方法などが考えられる。また、出力情報が音声で出力される場合には、強調度合いを強くする方法として、音量を大きくする方法などが考えられる。 Here, the output unit 2080 may set the degree of emphasis of the output information based on the information regarding the person to be monitored. For example, ``the greater the score calculated for the person to be monitored, the stronger the degree of emphasis'', ``the closer the distance from the person to be monitored to the departure place 20, the stronger the degree of emphasis'', or ``to the departure time of the vehicle 10 The shorter the remaining time is, the stronger the degree of emphasis is." As a method of increasing the degree of emphasis, for example, a method of using a more conspicuous color, a method of increasing the size, and the like are conceivable. Also, when the output information is output by voice, a method of increasing the volume may be considered as a method of increasing the degree of emphasis.
 出力情報は、要監視人物に関する情報に限定されない。例えば、駅の構内などに設けられているスピーカーを利用して、駆け込み乗車を行おうとしている人物が検出された旨の音声メッセージや、駆け込み乗車を控えるように促す音声メッセージなどを表す出力情報が出力されてもよい。ここで、この音声メッセージは、要監視人物に聞かせることが好適である。そこで例えば、出力部2080は、要監視人物が検出されたビデオデータ50を生成したカメラ40の近くに設けられているスピーカーに、音声メッセージを出力させる。また、そのカメラ40から出発場所20までの経路上に設けられている各スピーカーを利用することも好適である。  The output information is not limited to information about the person to be monitored. For example, using a speaker installed in a station premises, etc., output information that expresses a voice message to the effect that a person trying to rush to get on the train has been detected, or a voice message that urges people to refrain from rushing to get on the train, etc. may be output. Here, this voice message is preferably heard by the person under surveillance. Therefore, for example, the output unit 2080 causes a speaker provided near the camera 40 that generated the video data 50 in which the person to be monitored is detected to output a voice message. It is also preferable to use each speaker provided on the route from the camera 40 to the starting point 20 .
 その他にも例えば、出力部2080は、要監視人物が所持する携帯端末に対して出力情報が送信してもよい。この場合、例えば出力情報は、代替の交通手段に関する案内を含むことが好適である。また、交通手段の検索を行えるアプリケーションを起動させる情報や、交通手段の検索を行える web サイトへ誘導する情報(当該 web サイトのリンクを示す情報など)が、出力情報として送信されてもよい。これらの方法によれば、要監視人物にとって代替の交通手段の把握が容易になるため、要監視人物が駆け込み乗車をやめる蓋然性を高めることができる。 In addition, for example, the output unit 2080 may transmit output information to a mobile terminal possessed by a person requiring monitoring. In this case, for example, the output information preferably includes guidance regarding alternative means of transportation. In addition, information that activates an application that can search for means of transportation, or information that guides users to a website that can search for means of transportation (information indicating a link to the website, etc.) may be sent as output information. According to these methods, it becomes easier for the person to be monitored to grasp alternative means of transportation, so that the probability that the person to be monitored will stop rushing to board the vehicle can be increased.
 図7は、要監視人物が所持する携帯端末に対して送信される出力情報を例示する図である。この出力情報には、要監視人物の現在位置とタクシー乗り場の位置とが示された地図、及びタクシーの利用を促すメッセージが含まれている。 FIG. 7 is a diagram exemplifying output information transmitted to a mobile terminal possessed by a person requiring monitoring. This output information includes a map showing the current position of the person to be monitored and the position of the taxi stand, and a message prompting the use of a taxi.
 ここで、要監視人物が所持する携帯端末に対して出力情報が送信されるようにする場合、予め、その人物に関する情報が登録されているものとする。具体的には、人物の顔情報(顔の特徴量)と宛先アドレスとのペアが、検出装置2000から取得可能な態様で、予め記憶部に格納されている。出力部2080は、ビデオデータ50から検出された要監視人物の顔の特徴量を、登録されている各人物の顔の特徴量と照合することにより、要監視人物とマッチする登録済み人物を検索する。要監視人物とマッチする登録済み人物が検出された場合、出力部2080は、当該登録済み人物に対応づけられている宛先アドレスに対して、出力情報を送信する。 Here, when the output information is to be transmitted to the mobile terminal possessed by the person to be monitored, it is assumed that the information on the person is registered in advance. Specifically, a pair of a person's face information (facial feature amount) and a destination address is stored in advance in the storage unit in a manner that can be obtained from the detecting device 2000 . The output unit 2080 searches for a registered person who matches the person to be monitored by matching the facial feature amount of the person to be monitored detected from the video data 50 with the facial feature amount of each registered person. do. When a registered person matching the person to be monitored is detected, the output unit 2080 transmits output information to the destination address associated with the registered person.
 なお、要監視人物が検出されたら、検出装置2000は、その人物の移動先(出発場所20へ向かう経路上)に設けられている各カメラ40から得られるビデオデータ50を利用して、当該人物が要監視人物として検出され続けるか否かを判定してもよい。言い換えれば、検出装置2000は、各カメラ40を用いて検出された各要監視人物について、既に要監視人物として検出されている人物であるか否かを判定してもよい。既に要監視人物として検出されている人物が、再度要監視人物として検出された場合、その人物に対して出力する出力情報として、新たに要監視人物として検出された人物に対して出力する出力情報とは異なるものが利用されてもよい。例えば、要監視人物として検出された回数が多いほど、よりその内容が強調された出力情報が出力されるようにする。互いに異なるビデオデータ50から検出された要監視人物同士が一致するか否かの判定は、例えば、各ビデオデータ50から検出された要監視人物の顔の特徴を比較することで行うことができる。 Note that, when a person to be monitored is detected, the detection device 2000 uses the video data 50 obtained from each camera 40 provided at the destination of the person (on the route toward the starting point 20) to detect the person to be monitored. may continue to be detected as a person to be monitored. In other words, the detection apparatus 2000 may determine whether each person to be monitored detected using each camera 40 is a person who has already been detected as a person to be monitored. When a person who has already been detected as a person requiring surveillance is detected again as a person requiring surveillance, output information to be output to the person newly detected as a person requiring surveillance is output information to be output to that person. A different one may be used. For example, the greater the number of times a person is detected as a person to be monitored, the more emphasized the content of the output information is output. Whether or not the persons to be monitored detected from different video data 50 match each other can be determined, for example, by comparing the facial features of the persons to be monitored detected from each video data 50 .
 以上、実施の形態を参照して本願発明を説明したが、本願発明は上記実施形態に限定されるものではない。本願発明の構成や詳細には、本願発明のスコープ内で当業者が理解し得る様々な変更をすることができる。 Although the present invention has been described with reference to the embodiments, the present invention is not limited to the above embodiments. Various changes that can be understood by those skilled in the art can be made to the configuration and details of the present invention within the scope of the present invention.
 なお、上述の例において、プログラムは、コンピュータに読み込まれた場合に、実施形態で説明された1又はそれ以上の機能をコンピュータに行わせるための命令群(又はソフトウェアコード)を含む。プログラムは、非一時的なコンピュータ可読媒体又は実体のある記憶媒体に格納されてもよい。限定ではなく例として、コンピュータ可読媒体又は実体のある記憶媒体は、random-access memory(RAM)、read-only memory(ROM)、フラッシュメモリ、solid-state drive(SSD)又はその他のメモリ技術、CD-ROM、digital versatile disc(DVD)、Blu-ray(登録商標)ディスク又はその他の光ディスクストレージ、磁気カセット、磁気テープ、磁気ディスクストレージ又はその他の磁気ストレージデバイスを含む。プログラムは、一時的なコンピュータ可読媒体又は通信媒体上で送信されてもよい。限定ではなく例として、一時的なコンピュータ可読媒体又は通信媒体は、電気的、光学的、音響的、またはその他の形式の伝搬信号を含む。 It should be noted that in the above examples, the program includes instructions (or software code) that, when read into a computer, cause the computer to perform one or more functions described in the embodiments. The program may be stored in a non-transitory computer-readable medium or tangible storage medium. By way of example, and not limitation, computer readable media or tangible storage media may include random-access memory (RAM), read-only memory (ROM), flash memory, solid-state drives (SSD) or other memory technology, CDs - ROM, digital versatile disc (DVD), Blu-ray disc or other optical disc storage, magnetic cassette, magnetic tape, magnetic disc storage or other magnetic storage device. The program may be transmitted on a transitory computer-readable medium or communication medium. By way of example, and not limitation, transitory computer readable media or communication media include electrical, optical, acoustic, or other forms of propagated signals.
 上記の実施形態の一部又は全部は、以下の付記のようにも記載されうるが、以下には限られない。
 (付記1)
 対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得部と、
 前記カメラによって生成されるビデオデータを取得する第2取得部と、
 前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出部と、を有する検出装置。
 (付記2)
 前記検出部は、前記ビデオデータから検出される各人物について、その人物が前記カメラによって撮像された期間が前記予測通過時刻に近い度合いを表す第1スコア、その人物が前記出発場所に向かっている度合いを表す第2スコア、及びその人物が急いでいる度合いを表す第3スコアのいずれか1つ以上を算出し、算出したスコアに基づいて、その人物が前記要監視人物であるか否かを判定する、付記1に記載の検出装置。
 (付記3)
 前記検出部は、前記ビデオデータから検出される各人物について、その人物の移動動作の種類、その人物の移動の速さ、又はその人物の表情の特徴に基づいて、前記第3スコアを算出する、付記2に記載の検出装置。
 (付記4)
 前記カメラと前記出発場所との位置関係に基づいて、前記カメラの撮像範囲から前記出発場所までの移動に要する時間である予測所要時間を算出し、前記対象の乗物の出発時刻と前記予測所要時間とに基づいて前記予測通過時刻情報を生成する生成部を有する、付記1から3いずれか一項に記載の検出装置。
 (付記5)
 前記検出部によって検出された前記要監視人物に関連する情報、代替の交通手段を示す情報、及び代替の交通手段の検索を容易にする情報のいずれか1つ以上を示す出力情報を出力する出力部を有する、付記1から4いずれか一項に記載の検出装置。
 (付記6)
 人物の顔の特徴量とその人物の宛先アドレスとを対応づけた情報が記憶部に格納されており、
 前記出力部は、前記記憶部に格納されている顔の特徴量の中から、前記ビデオデータから得られる前記要監視人物の顔の特徴量とマッチするものと特定し、前記特定した特徴量に対応づけられている宛先アドレスへ前記出力情報を出力する、付記5に記載の検出装置。
 (付記7)
 コンピュータによって実行される検出方法であって、
 対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得ステップと、
 前記カメラによって生成されるビデオデータを取得する第2取得ステップと、
 前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出ステップと、を有する検出方法。
 (付記8)
 前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物が前記カメラによって撮像された期間が前記予測通過時刻に近い度合いを表す第1スコア、その人物が前記出発場所に向かっている度合いを表す第2スコア、及びその人物が急いでいる度合いを表す第3スコアのいずれか1つ以上を算出し、算出したスコアに基づいて、その人物が前記要監視人物であるか否かを判定する、付記7に記載の検出方法。
 (付記9)
 前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物の移動動作の種類、その人物の移動の速さ、又はその人物の表情の特徴に基づいて、前記第3スコアを算出する、付記8に記載の検出方法。
 (付記10)
 前記カメラと前記出発場所との位置関係に基づいて、前記カメラの撮像範囲から前記出発場所までの移動に要する時間である予測所要時間を算出し、前記対象の乗物の出発時刻と前記予測所要時間とに基づいて前記予測通過時刻情報を生成する生成ステップを有する、付記7から9いずれか一項に記載の検出方法。
 (付記11)
 前記検出ステップによって検出された前記要監視人物に関連する情報、代替の交通手段を示す情報、及び代替の交通手段の検索を容易にする情報のいずれか1つ以上を示す出力情報を出力する出力ステップを有する、付記7から10いずれか一項に記載の検出方法。
 (付記12)
 人物の顔の特徴量とその人物の宛先アドレスとを対応づけた情報が記憶部に格納されており、
 前記出力ステップにおいて、前記記憶部に格納されている顔の特徴量の中から、前記ビデオデータから得られる前記要監視人物の顔の特徴量とマッチするものと特定し、前記特定した特徴量に対応づけられている宛先アドレスへ前記出力情報を出力する、付記11に記載の検出方法。
 (付記13)
 コンピュータに、
 対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得ステップと、
 前記カメラによって生成されるビデオデータを取得する第2取得ステップと、
 前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出ステップと、を実行させるプログラムを格納している非一時的なコンピュータ可読媒体。
 (付記14)
 前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物が前記カメラによって撮像された期間が前記予測通過時刻に近い度合いを表す第1スコア、その人物が前記出発場所に向かっている度合いを表す第2スコア、及びその人物が急いでいる度合いを表す第3スコアのいずれか1つ以上を算出し、算出したスコアに基づいて、その人物が前記要監視人物であるか否かを判定する、付記13に記載のコンピュータ可読媒体。
 (付記15)
 前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物の移動動作の種類、その人物の移動の速さ、又はその人物の表情の特徴に基づいて、前記第3スコアを算出する、付記14に記載のコンピュータ可読媒体。
 (付記16)
 前記カメラと前記出発場所との位置関係に基づいて、前記カメラの撮像範囲から前記出発場所までの移動に要する時間である予測所要時間を算出し、前記対象の乗物の出発時刻と前記予測所要時間とに基づいて前記予測通過時刻情報を生成する生成ステップを有する、付記13から15いずれか一項に記載のコンピュータ可読媒体。
 (付記17)
 前記プログラムは、前記コンピュータに、前記検出ステップによって検出された前記要監視人物に関連する情報、代替の交通手段を示す情報、及び代替の交通手段の検索を容易にする情報のいずれか1つ以上を示す出力情報を出力する出力ステップを実行させる、付記13から16いずれか一項に記載のコンピュータ可読媒体。
 (付記18)
 人物の顔の特徴量とその人物の宛先アドレスとを対応づけた情報が記憶部に格納されており、
 前記出力ステップにおいて、前記記憶部に格納されている顔の特徴量の中から、前記ビデオデータから得られる前記要監視人物の顔の特徴量とマッチするものと特定し、前記特定した特徴量に対応づけられている宛先アドレスへ前記出力情報を出力する、付記17に記載のコンピュータ可読媒体。
Some or all of the above-described embodiments can also be described in the following supplementary remarks, but are not limited to the following.
(Appendix 1)
Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle. a first acquisition unit that acquires information;
a second acquisition unit for acquiring video data generated by the camera;
a detection unit that detects a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle from the video data using the predicted passage time information.
(Appendix 2)
For each person detected from the video data, the detection unit obtains a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, and Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. 2. The detection device of claim 1, wherein the detection device determines.
(Appendix 3)
The detection unit calculates the third score for each person detected from the video data based on the type of movement of the person, the speed of movement of the person, or the facial expression of the person. , Supplementary Note 2.
(Appendix 4)
Based on the positional relationship between the camera and the departure place, a predicted required time, which is the time required for movement from the imaging range of the camera to the departure place, is calculated, and the departure time and the predicted required time of the target vehicle. 4. The detection device according to any one of appendices 1 to 3, comprising a generation unit that generates the predicted passage time information based on and.
(Appendix 5)
Output for outputting output information indicating any one or more of information related to the person to be monitored detected by the detection unit, information indicating alternative means of transportation, and information facilitating search for alternative means of transportation 5. The detection device according to any one of appendices 1 to 4, comprising a portion.
(Appendix 6)
Information that associates a person's facial features with a destination address of the person is stored in a storage unit,
The output unit identifies, from the facial feature amounts stored in the storage unit, those that match the facial feature amount of the person to be monitored obtained from the video data, and determines the identified feature amount. 6. The detection device according to appendix 5, wherein the output information is output to the associated destination address.
(Appendix 7)
A computer-implemented detection method comprising:
Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle. a first obtaining step of obtaining information;
a second acquisition step of acquiring video data generated by the camera;
and a detecting step of detecting from the video data a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle, using the predicted passage time information.
(Appendix 8)
In the detecting step, for each person detected from the video data, a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, the person is heading to the departure location; Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. The detection method according to appendix 7, wherein:
(Appendix 9)
In the detection step, for each person detected from the video data, the third score is calculated based on the type of movement of the person, the speed of movement of the person, or the characteristics of facial expression of the person. , appendix 8.
(Appendix 10)
Based on the positional relationship between the camera and the departure place, a predicted required time, which is the time required for movement from the imaging range of the camera to the departure place, is calculated, and the departure time and the predicted required time of the target vehicle. 10. The detection method according to any one of appendices 7 to 9, comprising a generating step of generating the predicted transit time information based on:
(Appendix 11)
An output that outputs output information indicating any one or more of information related to the person to be monitored detected by the detecting step, information indicating alternative means of transportation, and information facilitating search for alternative means of transportation. 11. A detection method according to any one of clauses 7 to 10, comprising steps.
(Appendix 12)
Information that associates a person's facial features with a destination address of the person is stored in a storage unit,
In the output step, among the facial feature amounts stored in the storage unit, those that match the facial feature amount of the person to be monitored obtained from the video data are specified, and the specified feature amount is specified. 12. The detection method according to appendix 11, wherein the output information is output to the associated destination address.
(Appendix 13)
to the computer,
Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle. a first obtaining step of obtaining information;
a second acquisition step of acquiring video data generated by the camera;
a detection step of detecting a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle from the video data using the predicted passage time information; and non-transitory computer-readable medium.
(Appendix 14)
In the detecting step, for each person detected from the video data, a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, the person is heading to the departure location; Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. 14. The computer-readable medium of Clause 13, wherein determining.
(Appendix 15)
In the detection step, for each person detected from the video data, the third score is calculated based on the type of movement of the person, the speed of movement of the person, or the characteristics of facial expression of the person. 15. The computer-readable medium of claim 14.
(Appendix 16)
Based on the positional relationship between the camera and the departure place, a predicted required time, which is the time required for movement from the imaging range of the camera to the departure place, is calculated, and the departure time and the predicted required time of the target vehicle. 16. The computer-readable medium of any one of Clauses 13-15, comprising a generating step of generating the predicted transit time information based on:
(Appendix 17)
The program provides to the computer any one or more of information related to the person to be monitored detected by the detecting step, information indicating alternative means of transportation, and information facilitating search for alternative means of transportation. 17. The computer-readable medium according to any one of clauses 13 to 16, causing an output step of outputting output information indicative of:
(Appendix 18)
Information that associates a person's facial features with a destination address of the person is stored in a storage unit,
In the output step, among the facial feature amounts stored in the storage unit, those that match the facial feature amount of the person to be monitored obtained from the video data are specified, and the specified feature amount is specified. 18. The computer-readable medium of clause 17, outputting the output information to an associated destination address.
10      乗物
20      出発場所
40      カメラ
50      ビデオデータ
60      予測通過時刻情報
62      乗物識別情報
64      カメラ識別情報
66      予測通過時刻
500      コンピュータ
502      バス
504      プロセッサ
506      メモリ
508      ストレージデバイス
510      入出力インタフェース
512      ネットワークインタフェース
2000     検出装置
2020     第1取得部
2040     第2取得部
2060     検出部
2080     出力部
10 vehicle 20 starting point 40 camera 50 video data 60 predicted passage time information 62 vehicle identification information 64 camera identification information 66 predicted passage time 500 computer 502 bus 504 processor 506 memory 508 storage device 510 input/output interface 512 network interface 2000 detector 2020 1 acquisition unit 2040 2nd acquisition unit 2060 detection unit 2080 output unit

Claims (18)

  1.  対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得部と、
     前記カメラによって生成されるビデオデータを取得する第2取得部と、
     前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出部と、を有する検出装置。
    Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle. a first acquisition unit that acquires information;
    a second acquisition unit for acquiring video data generated by the camera;
    a detection unit that detects a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle from the video data using the predicted passage time information.
  2.  前記検出部は、前記ビデオデータから検出される各人物について、その人物が前記カメラによって撮像された期間が前記予測通過時刻に近い度合いを表す第1スコア、その人物が前記出発場所に向かっている度合いを表す第2スコア、及びその人物が急いでいる度合いを表す第3スコアのいずれか1つ以上を算出し、算出したスコアに基づいて、その人物が前記要監視人物であるか否かを判定する、請求項1に記載の検出装置。 For each person detected from the video data, the detection unit obtains a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, and Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. 2. The detection device of claim 1, for determining.
  3.  前記検出部は、前記ビデオデータから検出される各人物について、その人物の移動動作の種類、その人物の移動の速さ、又はその人物の表情の特徴に基づいて、前記第3スコアを算出する、請求項2に記載の検出装置。 The detection unit calculates the third score for each person detected from the video data based on the type of movement of the person, the speed of movement of the person, or the facial expression of the person. 3. A detection device according to claim 2.
  4.  前記カメラと前記出発場所との位置関係に基づいて、前記カメラの撮像範囲から前記出発場所までの移動に要する時間である予測所要時間を算出し、前記対象の乗物の出発時刻と前記予測所要時間とに基づいて前記予測通過時刻情報を生成する生成部を有する、請求項1から3いずれか一項に記載の検出装置。 Based on the positional relationship between the camera and the departure place, a predicted required time, which is the time required for movement from the imaging range of the camera to the departure place, is calculated, and the departure time and the predicted required time of the target vehicle. 4. The detection device according to any one of claims 1 to 3, further comprising a generation unit that generates the predicted passage time information based on and.
  5.  前記検出部によって検出された前記要監視人物に関連する情報、代替の交通手段を示す情報、及び代替の交通手段の検索を容易にする情報のいずれか1つ以上を示す出力情報を出力する出力部を有する、請求項1から4いずれか一項に記載の検出装置。 Output for outputting output information indicating any one or more of information related to the person to be monitored detected by the detection unit, information indicating alternative means of transportation, and information facilitating search for alternative means of transportation 5. A detection device according to any one of claims 1 to 4, comprising a portion.
  6.  人物の顔の特徴量とその人物の宛先アドレスとを対応づけた情報が記憶部に格納されており、
     前記出力部は、前記記憶部に格納されている顔の特徴量の中から、前記ビデオデータから得られる前記要監視人物の顔の特徴量とマッチするものと特定し、前記特定した特徴量に対応づけられている宛先アドレスへ前記出力情報を出力する、請求項5に記載の検出装置。
    Information that associates a person's facial features with a destination address of the person is stored in a storage unit,
    The output unit identifies, from the facial feature amounts stored in the storage unit, those that match the facial feature amount of the person to be monitored obtained from the video data, and determines the identified feature amount. 6. The detecting device according to claim 5, wherein said output information is output to the associated destination address.
  7.  コンピュータによって実行される検出方法であって、
     対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得ステップと、
     前記カメラによって生成されるビデオデータを取得する第2取得ステップと、
     前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出ステップと、を有する検出方法。
    A computer-implemented detection method comprising:
    Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle. a first obtaining step of obtaining information;
    a second acquisition step of acquiring video data generated by the camera;
    and a detecting step of detecting from the video data a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle, using the predicted passage time information.
  8.  前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物が前記カメラによって撮像された期間が前記予測通過時刻に近い度合いを表す第1スコア、その人物が前記出発場所に向かっている度合いを表す第2スコア、及びその人物が急いでいる度合いを表す第3スコアのいずれか1つ以上を算出し、算出したスコアに基づいて、その人物が前記要監視人物であるか否かを判定する、請求項7に記載の検出方法。 In the detecting step, for each person detected from the video data, a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, the person is heading to the departure location; Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. 8. A detection method according to claim 7, comprising determining.
  9.  前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物の移動動作の種類、その人物の移動の速さ、又はその人物の表情の特徴に基づいて、前記第3スコアを算出する、請求項8に記載の検出方法。 In the detection step, for each person detected from the video data, the third score is calculated based on the type of movement of the person, the speed of movement of the person, or the characteristics of facial expression of the person. 9. The detection method according to claim 8.
  10.  前記カメラと前記出発場所との位置関係に基づいて、前記カメラの撮像範囲から前記出発場所までの移動に要する時間である予測所要時間を算出し、前記対象の乗物の出発時刻と前記予測所要時間とに基づいて前記予測通過時刻情報を生成する生成ステップを有する、請求項7から9いずれか一項に記載の検出方法。 Based on the positional relationship between the camera and the departure place, a predicted required time, which is the time required for movement from the imaging range of the camera to the departure place, is calculated, and the departure time and the predicted required time of the target vehicle. 10. A detection method according to any one of claims 7 to 9, comprising a generating step of generating said predicted transit time information based on:
  11.  前記検出ステップによって検出された前記要監視人物に関連する情報、代替の交通手段を示す情報、及び代替の交通手段の検索を容易にする情報のいずれか1つ以上を示す出力情報を出力する出力ステップを有する、請求項7から10いずれか一項に記載の検出方法。 An output that outputs output information indicating any one or more of information related to the person to be monitored detected by the detecting step, information indicating alternative means of transportation, and information facilitating search for alternative means of transportation. 11. A detection method according to any one of claims 7 to 10, comprising steps.
  12.  人物の顔の特徴量とその人物の宛先アドレスとを対応づけた情報が記憶部に格納されており、
     前記出力ステップにおいて、前記記憶部に格納されている顔の特徴量の中から、前記ビデオデータから得られる前記要監視人物の顔の特徴量とマッチするものと特定し、前記特定した特徴量に対応づけられている宛先アドレスへ前記出力情報を出力する、請求項11に記載の検出方法。
    Information that associates a person's facial features with a destination address of the person is stored in a storage unit,
    In the output step, among the facial feature amounts stored in the storage unit, those that match the facial feature amount of the person to be monitored obtained from the video data are specified, and the specified feature amount is specified. 12. The detection method according to claim 11, wherein the output information is output to the associated destination address.
  13.  コンピュータに、
     対象の乗物の出発間際に前記対象の乗物に乗ろうとする人物が、前記対象の乗物の出発場所へ向かう人物を撮像可能なカメラの撮像範囲を通過すると予測される予測通過時刻を示す予測通過時刻情報を取得する第1取得ステップと、
     前記カメラによって生成されるビデオデータを取得する第2取得ステップと、
     前記予測通過時刻情報を用いて、前記ビデオデータから、前記対象の乗物の出発間際に前記対象の乗物に乗ろうとすると予測される要監視人物を検出する検出ステップと、を実行させるプログラムを格納している非一時的なコンピュータ可読媒体。
    to the computer,
    Predicted passage time indicating a predicted passage time at which a person who is about to board the target vehicle is predicted to pass through an imaging range of a camera capable of capturing an image of a person heading to the departure place of the target vehicle just before the departure of the target vehicle. a first obtaining step of obtaining information;
    a second acquisition step of acquiring video data generated by the camera;
    a detection step of detecting a person to be monitored who is predicted to board the target vehicle just before departure of the target vehicle from the video data using the predicted passage time information; and non-transitory computer-readable medium.
  14.  前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物が前記カメラによって撮像された期間が前記予測通過時刻に近い度合いを表す第1スコア、その人物が前記出発場所に向かっている度合いを表す第2スコア、及びその人物が急いでいる度合いを表す第3スコアのいずれか1つ以上を算出し、算出したスコアに基づいて、その人物が前記要監視人物であるか否かを判定する、請求項13に記載のコンピュータ可読媒体。 In the detecting step, for each person detected from the video data, a first score representing the degree to which the period during which the person was captured by the camera is close to the predicted passage time, the person is heading to the departure location; Any one or more of a second score representing the degree of urgency and a third score representing the degree of urgency of the person is calculated, and based on the calculated score, it is determined whether or not the person is the person to be monitored. 14. The computer readable medium of claim 13, wherein determining.
  15.  前記検出ステップにおいて、前記ビデオデータから検出される各人物について、その人物の移動動作の種類、その人物の移動の速さ、又はその人物の表情の特徴に基づいて、前記第3スコアを算出する、請求項14に記載のコンピュータ可読媒体。 In the detection step, for each person detected from the video data, the third score is calculated based on the type of movement of the person, the speed of movement of the person, or the characteristics of facial expression of the person. 15. The computer readable medium of claim 14.
  16.  前記カメラと前記出発場所との位置関係に基づいて、前記カメラの撮像範囲から前記出発場所までの移動に要する時間である予測所要時間を算出し、前記対象の乗物の出発時刻と前記予測所要時間とに基づいて前記予測通過時刻情報を生成する生成ステップを有する、請求項13から15いずれか一項に記載のコンピュータ可読媒体。 Based on the positional relationship between the camera and the departure place, a predicted required time, which is the time required for movement from the imaging range of the camera to the departure place, is calculated, and the departure time and the predicted required time of the target vehicle. 16. The computer-readable medium of any one of claims 13-15, comprising a generating step of generating the predicted transit time information based on:
  17.  前記プログラムは、前記コンピュータに、前記検出ステップによって検出された前記要監視人物に関連する情報、代替の交通手段を示す情報、及び代替の交通手段の検索を容易にする情報のいずれか1つ以上を示す出力情報を出力する出力ステップを実行させる、請求項13から16いずれか一項に記載のコンピュータ可読媒体。 The program provides to the computer any one or more of information related to the person to be monitored detected by the detecting step, information indicating alternative means of transportation, and information facilitating search for alternative means of transportation. 17. A computer readable medium according to any one of claims 13 to 16, causing an output step to be performed to output output information indicative of a.
  18.  人物の顔の特徴量とその人物の宛先アドレスとを対応づけた情報が記憶部に格納されており、
     前記出力ステップにおいて、前記記憶部に格納されている顔の特徴量の中から、前記ビデオデータから得られる前記要監視人物の顔の特徴量とマッチするものと特定し、前記特定した特徴量に対応づけられている宛先アドレスへ前記出力情報を出力する、請求項17に記載のコンピュータ可読媒体。
    Information that associates a person's facial features with a destination address of the person is stored in a storage unit,
    In the output step, among the facial feature amounts stored in the storage unit, those that match the facial feature amount of the person to be monitored obtained from the video data are specified, and the specified feature amount is specified. 18. The computer-readable medium of claim 17, outputting the output information to an associated destination address.
PCT/JP2021/026293 2021-07-13 2021-07-13 Detection device, detection method, and non-transitory computer-readable medium WO2023286152A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/026293 WO2023286152A1 (en) 2021-07-13 2021-07-13 Detection device, detection method, and non-transitory computer-readable medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/026293 WO2023286152A1 (en) 2021-07-13 2021-07-13 Detection device, detection method, and non-transitory computer-readable medium

Publications (1)

Publication Number Publication Date
WO2023286152A1 true WO2023286152A1 (en) 2023-01-19

Family

ID=84919727

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/026293 WO2023286152A1 (en) 2021-07-13 2021-07-13 Detection device, detection method, and non-transitory computer-readable medium

Country Status (1)

Country Link
WO (1) WO2023286152A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08244612A (en) * 1995-03-10 1996-09-24 Hitachi Ltd Dash-on-car-passenger monitor
JP2003179909A (en) * 2001-12-11 2003-06-27 Sharp Corp Monitoring system
JP2009302598A (en) * 2008-06-10 2009-12-24 Omron Corp Notification controller
JP2013052738A (en) * 2011-09-02 2013-03-21 Japan Transport Engineering Co Detector for rushing-into-train
JP2013143088A (en) * 2012-01-12 2013-07-22 Denso Corp Drowsy driving prevention device
JP2019041207A (en) * 2017-08-24 2019-03-14 株式会社日立国際電気 Monitoring system and monitoring method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08244612A (en) * 1995-03-10 1996-09-24 Hitachi Ltd Dash-on-car-passenger monitor
JP2003179909A (en) * 2001-12-11 2003-06-27 Sharp Corp Monitoring system
JP2009302598A (en) * 2008-06-10 2009-12-24 Omron Corp Notification controller
JP2013052738A (en) * 2011-09-02 2013-03-21 Japan Transport Engineering Co Detector for rushing-into-train
JP2013143088A (en) * 2012-01-12 2013-07-22 Denso Corp Drowsy driving prevention device
JP2019041207A (en) * 2017-08-24 2019-03-14 株式会社日立国際電気 Monitoring system and monitoring method

Similar Documents

Publication Publication Date Title
US10776627B2 (en) Human flow analysis method, human flow analysis apparatus, and human flow analysis system
US11205275B2 (en) Information processing apparatus, control method, and program
CN112364696B (en) Method and system for improving family safety by utilizing family monitoring video
US11776274B2 (en) Information processing apparatus, control method, and program
US8971577B2 (en) Monitoring device, reliability calculation program, and reliability calculation method
CN113177469B (en) Training method and device of human attribute detection model, electronic equipment and medium
JP2020077000A (en) Interaction device, interaction method, program, and vehicle control method
JP2010128594A (en) Motion detection device and motion detection method
CN111611894A (en) Personnel trajectory prediction method and device, computer equipment and storage medium
JP2015002477A (en) Information processing apparatus, information processing system, and information processing method
KR101979375B1 (en) Method of predicting object behavior of surveillance video
CN113269111B (en) Video monitoring-based elevator abnormal behavior detection method and system
WO2023286152A1 (en) Detection device, detection method, and non-transitory computer-readable medium
US11527091B2 (en) Analyzing apparatus, control method, and program
US11195404B2 (en) Interpreting reactions of other people for physically impaired during an emergency situation
KR102648004B1 (en) Apparatus and Method for Detecting Violence, Smart Violence Monitoring System having the same
US20200219412A1 (en) System and method for sensor fusion from a plurality of sensors and determination of a responsive action
JP2021190051A (en) Behavior body identification system
WO2020075283A1 (en) Abnormal person prediction system, abnormal person prediction method, and program
US20230334867A1 (en) Image processing apparatus, image processing method, and non-transitory computer-readable medium
JP6954416B2 (en) Information processing equipment, information processing methods, and programs
JP7476965B2 (en) Notification control device, control method, and program
WO2022259902A1 (en) Guidance information generation device, control method, and non-transitory computer-readable medium
WO2022009339A1 (en) Conversation monitoring device, control method, and computer-readable medium
JP2022052538A (en) Image processing device, image processing method, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21950095

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE