WO2021131024A1 - 処理装置、処理方法及びプログラム - Google Patents
処理装置、処理方法及びプログラム Download PDFInfo
- Publication number
- WO2021131024A1 WO2021131024A1 PCT/JP2019/051452 JP2019051452W WO2021131024A1 WO 2021131024 A1 WO2021131024 A1 WO 2021131024A1 JP 2019051452 W JP2019051452 W JP 2019051452W WO 2021131024 A1 WO2021131024 A1 WO 2021131024A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- reporting
- operated
- notification
- person
- condition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/188—Capturing isolated or intermittent images triggered by the occurrence of a predetermined event, e.g. an object reaching a predetermined position
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B25/00—Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/183—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/08—Mouthpieces; Microphones; Attachments therefor
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/06—Decision making techniques; Pattern matching strategies
Definitions
- the present invention relates to a processing device, a processing method and a program.
- a reporting device that notifies the police to that effect may be installed on the street or on the street.
- Patent Document 1 discloses a self-supporting security light that allows a whistleblower to make a call with the police via an intercom when an emergency such as an incident or accident occurs. Further, Patent Document 1 discloses that a voice message is transmitted to a person in the vicinity of a security light in order to prevent the reporting device from being pressed due to mischief or erroneous operation.
- Patent Document 1 only gives a warning regarding mischief and erroneous operation by voice. If the reporting device is pressed due to mischief or erroneous operation, the call processing with the police will be performed. Responding to reports due to mischief or erroneous operation puts a heavy burden on the responder (police).
- An object of the present invention is to reduce the frequency of reports due to mischief or erroneous operation.
- a detection means for detecting that the reporting device has been operated, and When it is detected that the reporting device has been operated, an acquisition means for acquiring at least one of an image generated by the camera and a voice collected by the microphone, and A means for determining whether at least one of the image and the sound satisfies the reporting condition, and When it is determined that the notification condition is satisfied, an output means for outputting notification information indicating that the notification device has been operated and an output means.
- a processing device having the above is provided.
- the computer Detects that the reporting device has been operated and When it is detected that the reporting device has been operated, at least one of the image generated by the camera and the voice collected by the microphone is acquired. Judging whether at least one of the image and the sound satisfies the reporting condition, When it is determined that the notification condition is satisfied, a processing method for outputting notification information indicating that the notification device has been operated is provided.
- Detection means for detecting that the reporting device has been operated, When it is detected that the reporting device has been operated, an acquisition means for acquiring at least one of an image generated by the camera and a voice collected by the microphone. A means for determining whether at least one of the image and the sound satisfies the reporting condition, An output means that outputs notification information indicating that the notification device has been operated when it is determined that the notification condition is satisfied.
- a program is provided to function as.
- the frequency of reports due to mischief or erroneous operation is reduced.
- the reporting device 1 is installed at an arbitrary position on the street, on the street, in a public facility, or the like.
- a data collection device 2 including at least one of a camera and a microphone is installed around the notification device 1.
- the camera is installed so as to photograph the surroundings of the reporting device 1.
- the microphone is installed so as to collect the sound around the reporting device 1.
- the reporting device 1 and the processing device 10 are connected to each other so as to be able to communicate with each other by wire and / or wirelessly. Further, the data collection device 2 and the processing device 10 are connected to each other so as to be able to communicate with each other by wire and / or wirelessly.
- the processing device 10 When the processing device 10 detects that the reporting device 1 has been operated based on the signal from the reporting device 1, the data (at least one of the image data and the voice data) generated by the data collecting device 2 satisfies a predetermined reporting condition. To judge. Then, when the processing device 10 determines that the reporting condition is satisfied, the processing device 10 transmits the reporting information indicating that the reporting device 1 has been operated to the reporting destination device 3.
- the reporting device 3 and the processing device 10 are connected to each other so as to be able to communicate with each other by wire and / or wirelessly.
- the reporting destination device 3 may be, for example, a police device, a security company device, or a device managed by another person.
- the processing device 10 determines the necessity of reporting based on the data generated by the data collecting device 2, that is, whether it is a mischief or an erroneous operation, and determines that the reporting is necessary. You can make a report in case. According to such a processing device 10, the frequency of reporting due to mischief or erroneous operation is reduced.
- Each functional unit of the processing device 10 is a storage unit (stored from the stage of shipping the device in advance) such as a CPU (Central Processing Unit) of an arbitrary computer, a memory, a program loaded in the memory, and a hard disk for storing the program.
- a storage unit such as a CPU (Central Processing Unit) of an arbitrary computer, a memory, a program loaded in the memory, and a hard disk for storing the program.
- it can also store programs downloaded from storage media such as CDs (Compact Discs) and servers on the Internet), and it is realized by any combination of hardware and software centered on the network connection interface. Program.
- CDs Compact Discs
- FIG. 2 is a block diagram illustrating a hardware configuration of the processing device 10.
- the processing device 10 includes a processor 1A, a memory 2A, an input / output interface 3A, a peripheral circuit 4A, and a bus 5A.
- the peripheral circuit 4A includes various modules.
- the processing device 10 does not have to have the peripheral circuit 4A.
- the processing device 10 may be composed of a plurality of physically and / or logically separated devices, or may be composed of one physically and / or logically integrated device. When the processing device 10 is composed of a plurality of physically and / or logically separated devices, each of the plurality of devices can be provided with the above hardware configuration.
- the bus 5A is a data transmission path for the processor 1A, the memory 2A, the peripheral circuit 4A, and the input / output interface 3A to send and receive data to and from each other.
- the processor 1A is, for example, an arithmetic processing unit such as a CPU or a GPU (Graphics Processing Unit).
- the memory 2A is, for example, a memory such as a RAM (RandomAccessMemory) or a ROM (ReadOnlyMemory).
- the input / output interface 3A includes an interface for acquiring information from an input device, an external device, an external server, an external sensor, a camera, etc., an interface for outputting information to an output device, an external device, an external server, etc. ..
- the input device is, for example, a keyboard, a mouse, a microphone, a physical button, a touch panel, or the like.
- the output device is, for example, a display, a speaker, a printer, a mailer, or the like.
- the processor 1A can issue commands to each module and perform calculations based on the calculation results thereof.
- FIG. 3 shows an example of a functional block diagram of the processing device 10.
- the processing device 10 includes a detection unit 11, an acquisition unit 12, a determination unit 13, and an output unit 14.
- the detection unit 11 detects that the reporting device 1 has been operated.
- the reporting device 1 is installed at an arbitrary position on the street, on the street, in a public facility, or the like.
- the police and the like are notified to that effect.
- the reporting device 1 may include a button.
- the operation to be performed in an emergency may be the pressing of the button.
- the reporting device 1 may include other input devices such as a touch panel, a mouse, and a keyboard. Then, the operation performed in an emergency may be a predetermined input operation via these input devices.
- the operation performed in an emergency is preferably simple and easy, such as pressing a button.
- the reporting device 1 When the reporting device 1 receives a predetermined operation (eg, pressing a button, etc.), it transmits to that effect to the processing device 10.
- the detection unit 11 detects that the reporting device 1 has been operated based on the signal received from the reporting device 1.
- reporting device 1 and the processing device 10 may have a many-to-one relationship or a one-to-one relationship.
- the acquisition unit 12 acquires the data generated by the data collecting device 2.
- the data acquisition device 2 includes at least one of a camera and a microphone.
- the data generated by such a data acquisition device 2 includes at least one of an image and an audio.
- the camera is installed so as to take a picture of the surroundings of the reporting device 1.
- the camera is installed at a position and orientation capable of photographing a person who operates the reporting device 1, a person who is in the vicinity of the person who operates the reporting device 1, and the like.
- the microphone is installed so as to collect the sound around the reporting device 1.
- the microphone is in a position where it can collect the remarks of the person who operates the reporting device 1, the remarks of people around the person who operates the reporting device 1, the voices generated by the actions of these people, and the like. And installed in the orientation.
- the data generated by the data collecting device 2 is included in the storage device in the data collecting device 2, the storage device in the processing device 10, and the storage device in another device configured to be communicable with the data collecting device 2. It accumulates in at least one. Then, when it is detected that the reporting device 1 has been operated, the acquisition unit 12 acquires a predetermined part of the data accumulated as described above.
- the acquisition unit 12 acquires the data generated by the data collection device 2 from the start timing before the operation timing when the notification device 1 is operated to the end timing after the operation timing. That is, the acquisition unit 12 acquires data before and after the operation timing.
- the start timing is t1 second before the operation timing.
- the end timing is t2 seconds after the operation timing.
- t1 and t2 may have the same value or may have different values.
- the acquisition unit 12 may acquire the data generated by the data collection device 2 from the operation timing in which the notification device 1 is operated to the end timing after the operation timing. That is, the acquisition unit 12 may acquire data after the operation timing.
- the end timing is t2 seconds after the operation timing.
- the acquisition unit 12 may acquire the data generated by the data collection device 2 from the start timing before the operation timing when the notification device 1 is operated to the operation timing. That is, the acquisition unit 12 may acquire data before the operation timing.
- the start timing is t1 second before the operation timing.
- acquisition means “the own device goes to fetch the data stored in another device or storage medium” based on the user input or the instruction of the program (active). Acquisition) ”, for example, requesting or inquiring about another device to receive the data, accessing another device or a storage medium to read the data, and the like. Further, “acquisition” means “inputting data output from another device to the own device (passive acquisition)” based on user input or program instruction, for example, distribution (or distribution (or). , Transmission, push notification, etc.) may be included. In addition, “acquisition” means to select and acquire from received data or information, and “edit data (text conversion, data sorting, partial data extraction, file format change, etc.)". It may include “to generate new data and acquire the new data”.
- the determination unit 13 determines whether the data (at least one of the image and the sound) acquired by the acquisition unit 12 satisfies the reporting condition.
- the determination unit 13 extracts at least one of the person who operated the reporting device 1 and the person who is in the vicinity of the person who operated the reporting device 1 from the images acquired by the acquiring unit 12 as analysis targets. For example, the determination unit 13 may extract a person who comes into contact with the reporting device 1 (a person who touches it by hand, etc.) at the timing when the reporting device 1 is operated as a person who operates the reporting device 1. Then, the determination unit 13 is "a person who is within a predetermined distance from the person who operated the reporting device 1", "another person who is within a predetermined distance from the reporting device 1", or "another person in the image".
- the "other person” is a person other than the person who operated the reporting device 1. Since the means for extracting a person from the image and the means for estimating the actual distance between two points in the image are widely known, the description thereof is omitted here.
- the determination unit 13 determines whether the reporting condition is satisfied based on at least one of the behavior and facial expression of the person extracted as the analysis target. In the present embodiment, the determination unit 13 determines that the notification condition is satisfied when at least one of the behavior and facial expression of the person extracted as the analysis target corresponds to the predefined behavior and facial expression required for notification.
- the behaviors required for reporting by the person who operated the reporting device 1 include "care about the surroundings" and "restlessness".
- the detection of the behavior that cares about the surrounding state may be detected based on, for example, the movement of the face.
- the detection of restless behavior may be detected based on, for example, the movement of the body.
- a part of the body is constantly moving without resting for a predetermined time or longer and the speed of change (calculated based on the movement distance of the moving part within a predetermined time, etc.) is equal to or higher than the reference level, calm down. You may judge that there is no such thing.
- the facial expression required for reporting by the person who operated the reporting device 1 is fear or the like.
- the behavior and facial expression required for reporting by a person in the vicinity of the person who operated the reporting device 1 may be the same as the behavior and facial expression required for reporting by the person who operated the reporting device 1. This assumes that the person around the person who operated the reporting device 1 is a friend of the person who operated the reporting device 1 and is involved in some trouble like the person who operated the reporting device 1. It was done.
- the behavior and facial expression required for reporting by a person in the vicinity of the person who operates the reporting device 1 may be the behavior and facial expression that such a person tends to take.
- the behaviors required for reporting by a person in the vicinity of the person who operated the reporting device 1 are "to stare at the person who operated the reporting device 1 without distracting the line of sight” and “to hit or kick a nearby object”. , “Open your mouth wide and say something to intimidate” and so on.
- the detection of the behavior of staring at the person who operates the reporting device 1 without diverting the line of sight may be detected based on, for example, the movement of the face. As an example, when the face is always facing the person who operated the reporting device 1 for a predetermined time or longer, it is determined that the person who operated the reporting device 1 is staring at the person who operated the reporting device 1 without turning his / her line of sight. May be good.
- the behavior of hitting or kicking a surrounding object may be detected based on, for example, the movement of the body.
- a part of the body (arms and legs) moves and comes into contact with surrounding objects, and the speed of the part of the body (calculated based on the movement distance of the moving part within a predetermined time, etc.) is above the standard level. If this is the case, it may be determined that the player has hit or kicked a nearby object.
- the detection of the behavior of threatening to say something with the mouth wide open may be detected based on, for example, the orientation of the face or the movement of the mouth. As an example, if the mouth moves with the face facing the person who operated the reporting device 1, and there is a moment when the mouth opens wider than the reference level during the movement, open the mouth wide and say something. You may judge that it is intimidating.
- the facial expressions required for reporting by a person in the vicinity of the person who operated the reporting device 1 are anger and the like.
- the determination unit 13 may determine that the reporting condition is satisfied when the person around the person who operates the reporting device 1 is on the blacklist of dangerous persons prepared in advance.
- the appearance feature amount (facial feature amount, etc.) of the dangerous person may be registered. Then, the determination unit 13 may determine whether or not the person around the person who operates the reporting device 1 is a person on the blacklist based on the feature amount.
- the determination unit 13 determines that the reporting condition is satisfied when the voice includes a predefined voice that requires reporting.
- the voice required for reporting may be a scream, a yelling voice, or the like, or may be a predetermined remark content.
- the prescribed content of the statement may be something that the victim may say, such as "stop” or "forgive", or the perpetrator may say “wait”, "do not allow", etc. It may be one with. Since the means for detecting screams and yells from the voice and the means for identifying the content of the remark contained in the voice are widely known, the description thereof is omitted here.
- the voice required for reporting may be the voice of a person on the blacklist of dangerous persons prepared in advance. That is, the determination unit 13 may determine that the reporting condition is satisfied when the voice includes the voice of a person on the blacklist of the dangerous person.
- the voiceprint of the voice of a dangerous person may be registered in the blacklist. Then, the determination unit 13 may determine whether or not the voice of the person on the blacklist is included in the voice based on the voiceprint.
- the output unit 14 When it is determined that the notification condition is satisfied, the output unit 14 outputs notification information indicating that the notification device 1 has been operated. Specifically, the output unit 14 transmits the report information to the predetermined report destination device 3.
- the reporting destination device 3 may be, for example, a police device, a security company device, or a device managed by another person.
- the output unit 14 may include the data (at least one of the image and the voice) acquired by the acquisition unit 12 and the information indicating the installed position of the operated notification device 1 in the notification information.
- the flowchart of FIG. 4 shows the entire flow from the operation detection of the reporting device 1 to the output of the reporting information.
- the acquisition unit 12 receives the data generated by the data collection device 2 (at least one of the image and the voice). (S11). For example, the acquisition unit 12 receives data generated by the data collection device 2 between the start timing before the operation timing when the reporting device 1 is operated and the end timing after the operation timing, that is, the data before and after the operation timing. To get.
- the determination unit 13 determines whether the data acquired by the acquisition unit 12 in S11 satisfies a predetermined reporting condition (S12). Then, when it is determined that the reporting condition is satisfied (Yes in S13), the output unit 14 transmits the reporting information indicating that the reporting device 1 has been operated to the reporting destination device 3 (S14). On the other hand, if it is not determined that the reporting condition is satisfied (No in S13), the output unit 14 does not transmit the reporting information to the reporting destination device 3.
- the output unit 14 may include the data (at least one of the image and the voice) acquired by the acquisition unit 12 and the information indicating the installed position of the operated notification device 1 in the notification information.
- the report destination device 3 that has received the report information is provided with information indicating the operated report device 1 on the terminal of a person (police officer, security guard, etc.) near the operated report device 1. You may send a request to rush there. It should be noted that the request may include at least one of an image taken around the operated reporting device 1 and a voice collected around the operated reporting device 1. Then, the reporting destination device 3 may receive an answer from the terminal as to whether or not it is possible to rush to the operated reporting device 1. In this case, the reporting destination device 3 manages the position information of each terminal of a plurality of persons (police officers, guards, etc.), and based on the management information, the person who is near the operated reporting device 1 is selected. Can be identified.
- FIG. 5 shows an example of the processing flow of S12 of FIG.
- the acquisition unit 12 acquires both the image and the sound in S11, and the determination unit 13 determines whether the reporting condition is satisfied based on both the image and the sound.
- the determination unit 13 determines that the reporting condition is satisfied (S22).
- the determination unit 13 determines that the report condition is satisfied (S22) when the image does not satisfy the report condition (No in S20) but the voice satisfies the report condition (Yes in S21).
- the determination unit 13 determines that the reporting condition is not satisfied (S23).
- FIG. 6 shows an example of the processing flow for determining whether the reporting condition is satisfied based on the image.
- the determination unit 13 determines whether the reporting condition is satisfied based on both the behavior and facial expression of the person to be analyzed.
- the determination unit 13 extracts at least one of the person who operated the reporting device 1 and the person around the person who operated the reporting device 1 from the image as an analysis target (S30).
- the determination unit 13 determines that the notification condition is satisfied (S32).
- the determination unit 13 determines that the notification conditions are not satisfied (S33).
- FIG. 7 shows an example of the processing flow for determining whether the reporting condition is satisfied based on voice.
- the judgment unit 13 determines that the notification condition is satisfied (S41).
- the voice does not include the voice required for reporting (No in S40)
- the processing device 10 of the present embodiment described above determines the necessity of reporting based on the data generated by the data collecting device 2, that is, whether it is mischief or erroneous operation, and determines that it is necessary. If you do, you can make a report. According to such a processing device 10, the frequency of reporting due to mischief or erroneous operation is reduced.
- the processing device 10 can determine whether or not it is a mischief or an erroneous operation based on the behavior and facial expression of at least one of the person who operated the reporting device 1 and the person around it. Specifically, the processing device 10 can determine whether or not the behavior or facial expression of those persons is mischief or erroneous operation based on whether or not the person needs to be notified. According to such a processing device 10, mischief and erroneous operation can be detected with high accuracy.
- the processing device 10 determines whether or not it is a mischief or an erroneous operation based on the content of remarks made by at least one of the person who operated the reporting device 1 and the people around it, and the voice generated by the actions of those people. Can be done. Specifically, the processing device 10 can determine whether or not there is mischief or erroneous operation based on whether or not the voice that requires notification is included in the voice. According to such a processing device 10, mischief and erroneous operation can be detected with high accuracy.
- the processing device 10 can determine whether or not there is mischief or erroneous operation based on not only the image and voice after the operation timing when the reporting device 1 is operated but also the image and voice before the operation timing. According to such a processing device 10, mischief and erroneous operation can be detected with high accuracy.
- the determination unit 13 of the present embodiment determines whether or not the reporting condition is satisfied based on whether or not at least one of the behavior and facial expression of the person extracted as the analysis target corresponds to the predefined behavior and facial expression that does not require notification. .. The determination unit 13 determines that the notification condition is not satisfied when the behavior and facial expression do not require notification.
- the behavior of the person who operated the reporting device 1 that does not require reporting is “laughing”, “smiling”, “yawning”, “does not care about the surroundings”, and “calmness”. There is an example.
- facial expressions of the person who operated the reporting device 1 that do not require reporting are smile, joy, peace, calmness, and the like.
- the behavior and facial expression of the person who operates the reporting device 1 that does not require reporting is the same as the behavior and facial expression of the person who operates the reporting device 1 that does not require reporting. This assumes that the person around the person who operated the reporting device 1 is a friend of the person who operated the reporting device 1 and is mischievous with the person who operated the reporting device 1. Is.
- the determination unit 13 sets the reporting conditions. You may decide that it does not meet.
- the appearance feature amount (facial feature amount, etc.) of the mischievous addict may be registered in the blacklist. Then, the determination unit 13 may determine whether or not the person who operated the reporting device 1 and the person in the vicinity thereof are on the blacklist based on the feature amount.
- the determination unit 13 can determine that the notification condition is not satisfied when the voice includes a predefined voice that does not require notification.
- the voice that does not require a report may be a laughing voice or the like, or may be a predetermined remark content.
- the prescribed content may be something that a mischievous person may say, such as “Stop it", “I don't know”, “Do it quickly”, “I made a mistake”, “I made a mistake”, " It may be something that a person who has made an erroneous operation may make a statement, such as "I've done it" or "What should I do?"
- the voice that does not need to be reported may be the voice of a person on the blacklist of mischievous addicts prepared in advance. That is, the determination unit 13 may determine that the reporting condition is not satisfied when the voice includes the voice of a person on the blacklist of the mischievous addict.
- the voiceprint of the voice of a mischievous addict may be registered in the blacklist. Then, the determination unit 13 may determine whether or not the voice of the person on the blacklist is included in the voice based on the voiceprint.
- FIG. 8 shows an example of the flow of processing for determining whether the reporting condition is satisfied based on the image.
- the determination unit 13 determines whether the reporting condition is satisfied based on both the behavior and facial expression of the person to be analyzed.
- the determination unit 13 extracts at least one of the person who operated the reporting device 1 and the person around the person who operated the reporting device 1 from the image as an analysis target (S50).
- the determination unit 13 determines that the notification condition is not satisfied (S53).
- the determination unit 13 determines that the notification conditions are satisfied (S52). ..
- FIG. 9 shows an example of the flow of processing for determining whether the reporting condition is satisfied based on voice.
- the judgment unit 13 determines that the notification condition is not satisfied (S62).
- the voice does not include a predefined voice that does not require notification (No in S60)
- the processing device 10 of the present embodiment determines whether or not the behavior or facial expression of at least one of the person who operated the reporting device 1 and the surrounding persons is mischief or erroneous operation based on whether or not the reporting is unnecessary. Can be done. According to such a processing device 10, mischief and erroneous operation can be detected with high accuracy.
- the processing device 10 whether or not the voice generated by the person who operated the reporting device 1 and at least one of the surrounding people and the voice generated by the actions of those people include voices that do not require reporting. It is possible to judge whether it is a mischief or an erroneous operation based on the fact. According to such a processing device 10, mischief and erroneous operation can be detected with high accuracy.
- the processing devices 10 of the first and second embodiments transmit the report information to the report destination device 3 in response to the operation of the report device 1.
- the processing device 10 may have a function of transmitting report information to the report destination device 3 without operating the report device 1.
- the processing device 10 may perform a process of constantly monitoring the data generated by the data collecting device 2 and detecting an abnormality. Then, when the processing device 10 detects an abnormality in the data generated by the data collecting device 2, the processing device 10 may transmit the report information to the report destination device 3.
- the content detected by the processing device 10 as an abnormality may be, for example, that a person runs toward the reporting device 1. This can be detected based on the image.
- the processing device 10 may detect a scream, a yelling voice, or the like as an abnormality from the voice.
- the processing device 10 may detect violent acts such as hitting and kicking based on an image.
- the processing device 10 that can make a report without operating the report device 1 based on the situation around the report device 1, relief for a person or the like who has fallen into such a situation is realized.
- the acquisition means is at least the image generated by the camera from the start timing before the operation timing when the notification device is operated to the end timing after the operation timing, and the sound collected by the microphone.
- the processing apparatus according to 1, wherein one is acquired. 3.
- the processing device according to 1 or 2, wherein the camera photographs the surroundings of the reporting device, and the microphone collects sounds around the reporting device.
- the judgment means is The person who operated the reporting device was extracted from the image as an analysis target.
- the processing apparatus according to any one of 1 to 3, which determines whether or not the reporting condition is satisfied based on the behavior of the person extracted as the analysis target.
- the judgment means is From the image, a person in the vicinity of the person who operated the reporting device is extracted as an analysis target.
- the processing apparatus according to any one of 1 to 4, which determines whether or not the reporting condition is satisfied based on at least one of the behavior and facial expression of the person extracted as the analysis target. 6. 4.
- the determination means determines that the notification condition is satisfied when at least one of the behavior and facial expression of the person extracted as the analysis target corresponds to the predefined behavior and facial expression required for notification.
- the determination means is described in 4 or 5 for determining that the notification condition is not satisfied when at least one of the behavior and facial expression of the person extracted as the analysis target corresponds to the predefined behavior and facial expression that does not require notification.
- the judgment means is The processing device according to any one of 1 to 7, which determines that the notification condition is satisfied when the voice includes a voice that requires a notification that is defined in advance.
- the judgment means is The processing device according to any one of 1 to 7, which determines that the notification condition is not satisfied when the voice includes a predefined voice that does not require notification. 10.
- the computer Detects that the reporting device has been operated and When it is detected that the reporting device has been operated, at least one of the image generated by the camera and the voice collected by the microphone is acquired. Judging whether at least one of the image and the sound satisfies the reporting condition, A processing method for outputting notification information indicating that the notification device has been operated when it is determined that the notification condition is satisfied.
- 11. Computer, Detection means for detecting that the reporting device has been operated When it is detected that the reporting device has been operated, an acquisition means for acquiring at least one of an image generated by the camera and a voice collected by the microphone.
- a means for determining whether at least one of the image and the sound satisfies the reporting condition An output means that outputs notification information indicating that the notification device has been operated when it is determined that the notification condition is satisfied.
- a program that functions as.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Acoustics & Sound (AREA)
- Business, Economics & Management (AREA)
- Emergency Management (AREA)
- Alarm Systems (AREA)
- Closed-Circuit Television Systems (AREA)
Priority Applications (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/JP2019/051452 WO2021131024A1 (ja) | 2019-12-27 | 2019-12-27 | 処理装置、処理方法及びプログラム |
| US17/788,785 US11869244B2 (en) | 2019-12-27 | 2019-12-27 | Notification device, notification method, and notification storage medium |
| JP2021566728A JP7355120B2 (ja) | 2019-12-27 | 2019-12-27 | 処理装置、処理方法及びプログラム |
| US18/512,309 US20240087326A1 (en) | 2019-12-27 | 2023-11-17 | Notification device, notification method, and notification storage medium |
| US19/071,299 US20250265843A1 (en) | 2019-12-27 | 2025-03-05 | Notification device, notification method, and notification storage medium |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/JP2019/051452 WO2021131024A1 (ja) | 2019-12-27 | 2019-12-27 | 処理装置、処理方法及びプログラム |
Related Child Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/788,785 A-371-Of-International US11869244B2 (en) | 2019-12-27 | 2019-12-27 | Notification device, notification method, and notification storage medium |
| US18/512,309 Continuation US20240087326A1 (en) | 2019-12-27 | 2023-11-17 | Notification device, notification method, and notification storage medium |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2021131024A1 true WO2021131024A1 (ja) | 2021-07-01 |
Family
ID=76573847
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2019/051452 Ceased WO2021131024A1 (ja) | 2019-12-27 | 2019-12-27 | 処理装置、処理方法及びプログラム |
Country Status (3)
| Country | Link |
|---|---|
| US (3) | US11869244B2 (enExample) |
| JP (1) | JP7355120B2 (enExample) |
| WO (1) | WO2021131024A1 (enExample) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP7355050B2 (ja) * | 2021-03-04 | 2023-10-03 | トヨタ自動車株式会社 | 車両制御装置、車両制御方法およびプログラム |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH0935133A (ja) * | 1995-07-14 | 1997-02-07 | Safety- Net:Kk | 非常通報機能を有した自動販売機及びこれを用いた通信警備システム |
| JP2011073819A (ja) * | 2009-09-30 | 2011-04-14 | Mitsubishi Electric Building Techno Service Co Ltd | エレベータのインターホン装置 |
Family Cites Families (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2007072541A (ja) | 2005-09-05 | 2007-03-22 | Hitachi Ltd | 誤報いたずら対策用無線タグ付名札および誤報いたずら対策用通報装置 |
| JP2007251697A (ja) * | 2006-03-16 | 2007-09-27 | Chugoku Electric Power Co Inc:The | 機器監視制御システム、方法およびプログラム |
| KR101659027B1 (ko) * | 2014-05-15 | 2016-09-23 | 엘지전자 주식회사 | 이동 단말기 및 차량 제어 장치 |
| JP7296838B2 (ja) | 2019-09-26 | 2023-06-23 | 株式会社デンソーテン | 緊急通報装置及び方法 |
-
2019
- 2019-12-27 JP JP2021566728A patent/JP7355120B2/ja active Active
- 2019-12-27 WO PCT/JP2019/051452 patent/WO2021131024A1/ja not_active Ceased
- 2019-12-27 US US17/788,785 patent/US11869244B2/en active Active
-
2023
- 2023-11-17 US US18/512,309 patent/US20240087326A1/en not_active Abandoned
-
2025
- 2025-03-05 US US19/071,299 patent/US20250265843A1/en active Pending
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH0935133A (ja) * | 1995-07-14 | 1997-02-07 | Safety- Net:Kk | 非常通報機能を有した自動販売機及びこれを用いた通信警備システム |
| JP2011073819A (ja) * | 2009-09-30 | 2011-04-14 | Mitsubishi Electric Building Techno Service Co Ltd | エレベータのインターホン装置 |
Also Published As
| Publication number | Publication date |
|---|---|
| JP7355120B2 (ja) | 2023-10-03 |
| US20220358766A1 (en) | 2022-11-10 |
| US20250265843A1 (en) | 2025-08-21 |
| US20240087326A1 (en) | 2024-03-14 |
| JPWO2021131024A1 (enExample) | 2021-07-01 |
| US11869244B2 (en) | 2024-01-09 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US7944353B2 (en) | System and method for detecting and broadcasting a critical event | |
| JP4912606B2 (ja) | 監視装置、監視センタ、監視システムおよび監視方法 | |
| JP5902966B2 (ja) | 救助活動支援システム | |
| JP6801459B2 (ja) | 情報処理装置、いじめ発見方法、情報処理システム、および、コンピュータプログラム | |
| CN103338362B (zh) | 一种基于视频监控系统的事件处理方法及系统 | |
| JP6084257B2 (ja) | 救助活動支援システム | |
| CN116170566A (zh) | 一种智慧楼宇监控管理方法、装置、电子设备及存储介质 | |
| CN105227920A (zh) | 一种通过网络摄像机进行监控的方法和装置 | |
| KR101321447B1 (ko) | 네트워크를 통한 현장 모니터링 방법, 및 이에 사용되는 관리 서버 | |
| JP7666603B2 (ja) | 監視装置、監視方法、及びプログラム | |
| JP6026115B2 (ja) | 救助活動支援システム | |
| US20250265843A1 (en) | Notification device, notification method, and notification storage medium | |
| CN110659603A (zh) | 一种数据处理方法及装置 | |
| US11909914B2 (en) | Forwarding emergency messages from IoT devices to PSAPs | |
| JP2013225248A (ja) | 音識別システム、音識別装置、音識別方法およびプログラム | |
| JP6360919B2 (ja) | 救助活動支援システム | |
| JP2019040495A (ja) | 情報処理装置、情報処理方法、及び、プログラム | |
| KR102240089B1 (ko) | 개인 위험 상태 감지 기반 상황 정보 획득 장치 | |
| CN112435423A (zh) | 一种监控方法及装置 | |
| JP2004192459A (ja) | 異常事象検出装置 | |
| CN118506808A (zh) | 霸凌识别与处理方法、装置、电子设备及存储介质 | |
| CN110675607A (zh) | 一种基于印章的信息提示方法及装置 | |
| JP6351125B2 (ja) | 異常通報システム | |
| JP7778201B2 (ja) | システム | |
| CN112580390A (zh) | 基于智能音箱的安防监控方法、装置、音箱和介质 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19957167 Country of ref document: EP Kind code of ref document: A1 |
|
| ENP | Entry into the national phase |
Ref document number: 2021566728 Country of ref document: JP Kind code of ref document: A |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 19957167 Country of ref document: EP Kind code of ref document: A1 |