WO2022134504A9 - Image detection method and apparatus, electronic device, and storage medium - Google Patents
Image detection method and apparatus, electronic device, and storage medium Download PDFInfo
- Publication number
- WO2022134504A9 WO2022134504A9 PCT/CN2021/101619 CN2021101619W WO2022134504A9 WO 2022134504 A9 WO2022134504 A9 WO 2022134504A9 CN 2021101619 W CN2021101619 W CN 2021101619W WO 2022134504 A9 WO2022134504 A9 WO 2022134504A9
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- area
- escalator
- state
- image
- elevator
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 153
- 238000000034 method Methods 0.000 claims abstract description 76
- 238000012545 processing Methods 0.000 claims description 41
- 238000004590 computer program Methods 0.000 claims description 20
- 230000011218 segmentation Effects 0.000 claims description 16
- 238000003672 processing method Methods 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 26
- 238000004891 communication Methods 0.000 description 11
- 230000006870 function Effects 0.000 description 11
- 230000007246 mechanism Effects 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 8
- 238000012549 training Methods 0.000 description 7
- 230000005540 biological transmission Effects 0.000 description 6
- 238000013527 convolutional neural network Methods 0.000 description 6
- 230000008569 process Effects 0.000 description 5
- 230000004913 activation Effects 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 230000018109 developmental process Effects 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 3
- 230000002829 reductive effect Effects 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000012850 discrimination method Methods 0.000 description 2
- 238000005265 energy consumption Methods 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000006698 induction Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000001788 irregular Effects 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000007774 longterm Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 238000013145 classification model Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000009440 infrastructure construction Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000000670 limiting effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B66—HOISTING; LIFTING; HAULING
- B66B—ELEVATORS; ESCALATORS OR MOVING WALKWAYS
- B66B25/00—Control of escalators or moving walkways
- B66B25/003—Methods or algorithms therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2132—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/136—Segmentation; Edge detection involving thresholding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/751—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/809—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of classification results, e.g. where the classifiers operate on the same input data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- the present disclosure relates to the technical field of image detection and the technical field of computer vision, and in particular, to an image detection method and apparatus, an electronic device and a storage medium.
- the embodiment of the present disclosure proposes an image detection technical solution.
- an image detection method including: acquiring a first image of an escalator; performing area detection on the first image, and determining an image of the escalator in the first image.
- the first area perform elevator state recognition on the image of the first area corresponding to the first area, and determine at least one state recognition result of the escalator, where the state recognition result includes that the escalator is in an empty state or in non-empty elevator state; according to the at least one state identification result, determine the state of the escalator.
- the determining the state of the escalator according to the at least one state identification result includes: in the case of multiple state identification results, determining the state of the escalator according to the multiple state identification results and The weights of the plurality of state identification results determine the state discriminating value of the escalator; when the state discriminating value is greater than or equal to the first threshold, it is determined that the escalator is in an empty state.
- the state identification result includes a first state identification result
- the elevator state identification is performed on a first area image corresponding to the first area to determine at least one state of the escalator
- the identification result includes: classifying the first area image to obtain the first state identification result of the escalator.
- the state identification result includes a second state identification result
- the elevator state identification is performed on the first area image corresponding to the first area to determine at least one state of the escalator
- the identification result includes: segmenting the first area image, dividing the first area image into a background area and a foreground area where the escalator is located; adjusting the pixel values of the background area to obtain an adjusted
- the second area image of the escalator is obtained; the second area image is classified and processed to obtain the second state recognition result of the escalator.
- the state identification result includes a third state identification result
- the elevator state identification is performed on the first area image corresponding to the first area to determine at least one state of the escalator
- the identification result includes: performing pixel matching on the first area image and a preset reference image, and determining the matching area ratio between the first area image and the reference image, the reference image including The area image corresponding to the escalator in the elevator state; when the proportion of the matching area is greater than or equal to the second threshold, it is determined that the third state recognition result is that the escalator is in an empty state.
- the state identification result includes a fourth state identification result
- the elevator state identification is performed on the first area image corresponding to the first area to determine at least one state of the escalator
- the identification result includes: performing a first target detection on the first area image, and determining whether the first target exists in the first area image; and in the case where the first target does not exist in the first area image, determining The fourth state recognition result is that the escalator is in an empty state.
- the method further includes: when the escalator is in an empty state, sending an elevator stop signal, where the elevator stop signal is used to instruct the escalator to stop running.
- the method further includes: when the escalator is in a non-empty state and the escalator has stopped running, sending an elevator start signal, where the elevator start signal is used for Instruct the escalator to run.
- the method further includes: performing second target detection on the first image, and determining a third area of the second target in the first image; The positional relationship between the third areas determines a detection result of the second target, where the detection result includes that the second target is on the escalator or not on the escalator.
- the determining the detection result of the second target according to the positional relationship between the first area and the third area includes: between the fourth area and the third area When the area ratio between the areas is greater than or equal to a third threshold, it is determined that the detection result is that the second target is on the escalator, and the fourth area includes the first area and the third Intersection area between regions.
- an image detection apparatus including: an image acquisition module configured to acquire a first image of an escalator; an area detection module configured to perform area detection on the first image, and determine the first area of the escalator in the first image; the state recognition module is configured to perform elevator state recognition on the first area image corresponding to the first area, and determine at least one state recognition of the escalator As a result, the state recognition result includes that the escalator is in an empty state or a non-empty state; the state determination module is configured to determine the state of the escalator according to the at least one state recognition result.
- an electronic device comprising: a processor; a memory configured to store instructions executable by the processor; wherein the processor is configured to invoke the instructions stored in the memory to Perform the above method.
- a computer-readable storage medium having computer program instructions stored thereon, the computer program instructions implementing the above method when executed by a processor.
- a computer program product including computer-readable code, and when the computer-readable code is executed in an electronic device, a processor in the electronic device executes the above method .
- the escalator it is possible to detect the area where the escalator is located in the image, identify at least one state recognition result according to the area image of the escalator, and determine the empty state or non-empty state of the escalator according to the at least one state recognition result
- the state can improve the positioning accuracy of the escalator area and improve the recognition accuracy of the running state of the escalator.
- FIG. 1 shows a flowchart of an image detection method according to an embodiment of the present disclosure.
- Fig. 2a shows a schematic diagram of region detection of an image detection method according to an embodiment of the present disclosure.
- FIG. 2b shows a schematic diagram of area detection of an image detection method according to an embodiment of the present disclosure.
- FIG. 3 shows a schematic diagram of a first region according to an embodiment of the present disclosure.
- Figure 4a shows a schematic diagram of a second object according to an embodiment of the present disclosure.
- Figure 4b shows a schematic diagram of a second object according to an embodiment of the present disclosure.
- FIG. 5 shows a schematic diagram of a processing procedure of an image detection method according to an embodiment of the present disclosure.
- FIG. 6 shows a block diagram of an image detection apparatus according to an embodiment of the present disclosure.
- FIG. 7 shows a block diagram of an electronic device according to an embodiment of the present disclosure.
- FIG. 8 shows a block diagram of an electronic device according to an embodiment of the present disclosure.
- the artificial intelligence wave based on deep learning, industry + artificial intelligence has become one of the goals of industrial development in the new era.
- New fields such as smart factories, smart stores, and smart agriculture are emerging one after another.
- AI Artificial Intelligence
- smart security the industrial application of AI for security, called smart security
- the purpose of the smart elevator is to improve the work efficiency of the management department and strengthen the supervision of the daily operation of the elevator.
- only part of the detection function for escalators can usually be realized, and there is no complete smart elevator solution.
- the image detection method according to the embodiment of the present disclosure can be applied to scenes such as shopping malls, office buildings, and public transportation.
- the positioning of the elevator area, the discrimination of the empty and non-empty status of the elevator, and the detection and identification of key targets (such as baby strollers, wheelchairs, luggage, etc.) Effectively, reduce the cost of elevator operation and reduce the risk of safety accidents.
- FIG. 1 shows a flowchart of an image detection method according to an embodiment of the present disclosure. As shown in FIG. 1 , the image detection method includes:
- step S11 a first image of the escalator is acquired
- step S12 performing region detection on the first image to determine a first region of the escalator in the first image
- step S13 perform elevator state recognition on the first area image corresponding to the first area, and determine at least one state recognition result of the escalator, where the state recognition result includes that the escalator is in an empty state or in a non-empty state;
- step S14 the state of the escalator is determined according to the at least one state identification result.
- the image detection method may be executed by an electronic device such as a terminal device or a server, and the terminal device may be a user equipment (User Equipment, UE), a mobile device, a user terminal, a terminal, a handheld device, a computing equipment, vehicle-mounted equipment, wearable equipment, etc.
- the method can be implemented by the processor calling the computer-readable instructions stored in the memory.
- the method may be performed by a server.
- At least one image acquisition device can be installed at the location where the escalator to be detected is located, for example, at least one camera facing the escalator can be installed, so as to collect the video stream of the escalator, and to monitor the escalator and the escalator in the video stream.
- the target of the escalator (such as pedestrians, items carried by pedestrians, etc.) is detected.
- the embodiments of the present disclosure do not limit the installation position of the image acquisition device, the acquisition method of the video stream, and the area corresponding to the video stream.
- a first image of the escalator may be acquired.
- Each image frame of the video stream can be used as the first image; the video stream can also be sampled at a certain time interval, and the sampled image frame can be used as the first image; key frames in the video stream can also be obtained as the first image .
- the embodiment of the present disclosure does not limit the acquisition manner of the first image.
- region detection may be performed on the first image through the trained region detection network to determine the first region of the escalator in the first image.
- the region detection network may be, for example, a convolutional neural network, and the embodiments of the present disclosure do not limit the network structure and training method of the region detection network.
- FIG. 2a and 2b illustrate schematic diagrams of region detection in an image detection method according to an embodiment of the present disclosure.
- Fig. 2a shows the first image of the escalator;
- Fig. 2b shows the region detection result in the first image.
- the region detection network may adopt an object detection model or a region segmentation model.
- the detected first area can be the rectangular detection frame of the area where the escalator is located, such as the detection frame 21 in Fig. 2b; when the area segmentation model is used, the detected first area can be the escalator The irregular region in which it is located, such as region 22 in Figure 2b.
- the embodiment of the present disclosure does not limit the network model adopted by the area detection network.
- the background area can be reduced as much as possible while the detection frame includes the complete elevator area, thereby reducing the influence of background noise on the subsequent processing process.
- the lower left corner of the dark elevator area is used as the reference, as the coordinates of the lower left vertex of the detection frame 21; the upper right corner of the dark elevator area is used as the reference, as the detection The coordinates of the upper right vertex of box 21.
- the region segmentation model is used for processing, the dark elevator region in the detection frame can be segmented, and a polygon can be drawn with the elevator handrail as the boundary, such as region 22 in FIG. 2b.
- a first area image corresponding to the first area may be intercepted from the first image, and the escalator may be identified according to the first area image state, that is, to identify whether the escalator is in an empty state or a non-empty state.
- the first area image state that is, to identify whether the escalator is in an empty state or a non-empty state.
- At least one identification manner may be used to identify the images of the first area respectively to obtain a corresponding state identification result.
- the identification method may include, for example, at least one of the following: directly classifying the image of the first area; adjusting the pixel value of the background area in the image of the first area, and classifying the adjusted image; comparing the image of the first area with the reference image of the empty elevator Comparison; detect whether there are objects such as pedestrians on the elevator.
- the embodiments of the present disclosure do not limit the identification methods used.
- the state of the escalator may be determined according to at least one state identification result.
- the state of the escalator is comprehensively determined according to the value and weight of each state identification result.
- an elevator stop signal can be sent to instruct the escalator to stop running; when the escalator is in a non-empty state and has stopped running, an elevator stop signal can be sent. Send an elevator start signal to instruct the escalator to start running.
- the area where the escalator is located in the image can be detected, at least one state recognition result is recognized according to the area image of the escalator, and the empty state of the escalator or the escalator is determined according to the at least one state recognition result.
- the positioning accuracy of the escalator area can be improved, and the recognition accuracy of the running state of the escalator can be improved.
- the video stream of the area where the escalator is located can be captured by the camera, and the captured video stream can be transmitted to electronic devices such as a local front-end server or a cloud server.
- the electronic device can decode the video stream to obtain the decoded video stream.
- a first image of the escalator may be acquired.
- the first image may be an image frame of the decoded video stream.
- the region detection is performed on the first image through the trained region detection network, and the first region of the escalator in the first image is determined, for example, the detection frame of the region where the escalator is located.
- step S13 the first area image corresponding to the first area may be intercepted from the first image, and at least one identification method is used to respectively identify the first area image to obtain at least one state identification result.
- FIG. 3 shows a schematic diagram of a first region according to an embodiment of the present disclosure.
- the first image includes the first areas 31 and 32 of two elevators.
- the first area images corresponding to the first areas 31 and 32 can be intercepted respectively, and the two first area images can be recognized in parallel. Thereby improving the processing efficiency.
- the state identification result includes the first state identification result.
- Step S13 may include:
- the first area image is classified and processed to obtain a first state recognition result of the escalator.
- the first classification network may be, for example, a convolutional neural network, including a convolutional layer, a fully connected layer, an activation layer, etc.
- the embodiment of the present disclosure does not limit the network structure and training method of the first classification network.
- the first region image can be input into the first classification network, and the first state recognition result can be output, indicating that the escalator is in an empty elevator state or in a non-empty elevator state, for example, when the escalator is in an empty elevator state It outputs 1 when it is in the state, and outputs 0 when it is in the non-empty state. In this way, the status of the escalator can be identified simply and efficiently.
- the state identification result includes the second state identification result.
- Step S13 may include:
- the second area image is classified and processed to obtain the second state recognition result of the escalator.
- the first area image corresponding to the first area can be segmented by the trained segmentation network, and the first area image is segmented into the background area and the motorized area.
- the segmentation network may be, for example, a convolutional neural network, including a convolutional layer, a fully connected layer, an activation layer, etc.
- the embodiments of the present disclosure do not limit the network structure and training method of the segmentation network.
- the area image corresponding to the enclosing rectangular frame of the first area may be used as the first area image.
- the first area can be directly used as the foreground area; the area other than the first area in the first area image is used as the background area to realize the segmentation of the first area image.
- the pixel values of the pixels in the background area may be adjusted, for example, the pixel values of the background area are all adjusted to zero (black) to obtain an adjusted image of the second area.
- the pixel value of the background area may also be adjusted to other values, which is not limited in this embodiment of the present disclosure.
- the images of the second region may be classified by a trained classification network (herein referred to as a second classification network).
- the second classification network may be, for example, a convolutional neural network, including a convolutional layer, a fully connected layer, an activation layer, etc.
- the second classification network may have the same network structure as the first classification network, but with different network parameters.
- the embodiments of the present disclosure do not limit the network structure and training method of the second classification network.
- the second region image is input into the second classification network, and the second state recognition result is output, indicating that the escalator is in an empty state or in a non-empty state, for example, when the escalator is in an empty state It outputs 1 when it is in the empty state, and outputs 0 when it is in an empty state.
- the second region image after pixel adjustment is classified and processed, which can improve the accuracy of elevator state identification.
- the state identification result includes a third state identification result.
- Step S13 may include:
- Pixel matching is performed on the first area image and a preset reference image, and the matching area ratio between the first area image and the reference image is determined, and the reference image includes an escalator in an empty state. the corresponding area image;
- the third state recognition result is that the escalator is in an empty state.
- a single or multiple area images corresponding to the escalator can be acquired, for example, an image of the escalator is acquired by a camera, and an area image is obtained by performing area detection on the image.
- the area image can be used as a reference image; if multiple area images are acquired, the multiple area images can be fused to obtain a reference image (or called the empty elevator template).
- a reference image or called the empty elevator template.
- a segmentation model is used to obtain the labels of all pixels in the elevator area, and then an empty elevator template is obtained through a Gaussian mixture model.
- the embodiments of the present disclosure do not limit the generation manner of the reference image.
- the reference images may be stored in a database.
- pixel matching can be performed on the first area image and the reference image to determine the number of matching pixels; The proportion of matching areas.
- the proportion of the matching area is greater than or equal to the second threshold, it may be determined that the third state recognition result is that the escalator is in an empty state; otherwise, if the proportion of the matching area is less than the second threshold , it can be determined that the third state recognition result is that the escalator is in a non-empty state.
- the second threshold can set the second threshold according to the actual situation, for example, 0.8, and the embodiment of the present disclosure does not limit the actual value of the second threshold.
- the state identification result includes a fourth state identification result.
- Step S13 may include:
- the fourth state recognition result is that the escalator is in an empty state.
- a trained target detection network (which may be referred to as a first detection network) may be used to perform a first target detection on an image of the first area to determine whether there is a first target in the image of the first area.
- the first target may include, for example, a pedestrian, an item, etc., which is not limited in this embodiment of the present disclosure.
- the first detection network may be, for example, a convolutional neural network, and the embodiment of the present disclosure does not limit the network structure and training method of the first detection network.
- the fourth state recognition result is that the escalator is in a non-empty state; otherwise, if the first area image does not exist the first target If a target is reached, it can be determined that the fourth state recognition result is that the escalator is in an empty state.
- step S14 may include: in the case of multiple state identification results, determining the state identification value of the escalator according to the multiple state identification results and the weights of the multiple state identification results;
- the state of the escalator can be directly determined according to the state recognition result; if there are multiple state recognition results, the state of the escalator can be comprehensively determined according to the multiple state recognition results.
- the weight of each state recognition result may be preset, the weight of the state recognition result with a higher accuracy rate is set to be higher, and the weight of the state recognition result with a lower accuracy rate is set to be higher low, the sum of the weights of each state recognition result is 1.
- the state discrimination value of the escalator may be determined according to multiple state identification results and the weights of the multiple state identification results with reference to a voting mechanism.
- the state identification result may include at least two of the above-mentioned first, second, third and fourth state identification results.
- the state recognition result includes the first state recognition result a 1 , the second state recognition result a 2 , the third state recognition result a 3 , and the fourth state recognition result a 4
- the weights of the third and fourth state identification results are set as w 1 , w 2 , w 3 and w 4 respectively
- the state discrimination value w 1 *a 1 +w 2 *a 2 +w 3 *a 3 +w 4 *a 4
- w 1 +w 2 +w 3 +w 4 1, a 1 , a 2 , a 3 , and a 4 output 1 when the escalator is in an empty state, and output 0 when it is in a non-empty state.
- the state discriminating value is greater than or equal to the preset first threshold, it can be considered that the escalator is in an empty state; on the contrary, if the state discriminating value is smaller than the preset first threshold, it can be considered that the escalator is in an empty state; The escalator is not empty. As shown in FIG. 3 , the escalator in area 31 is in an empty state, and the escalator in area 32 is in a non-empty state.
- the first threshold value for example, 0.5, according to the actual situation, and the embodiment of the present disclosure does not limit the value of the first threshold value.
- the running state of the escalator can be accurately identified, and the error rate of the elevator state judgment can be significantly reduced.
- the image detection method according to the embodiment of the present disclosure may further include:
- an elevator stop signal is sent, and the elevator stop signal is used to instruct the escalator to stop running.
- an elevator stop signal can be generated and sent to instruct the escalator to stop running.
- the elevator stop signal is sent to the elevator control device, so that the elevator control device controls the escalator to stop running; the elevator stop signal can also be sent to the staff, so that the staff can control the escalator to stop running.
- the embodiment of the present disclosure does not limit the type and transmission manner of the elevator outage signal.
- the elevator can be stopped when it is empty, thereby reducing the cost of elevator operation.
- the image detection method according to the embodiment of the present disclosure may further include:
- an elevator start signal is sent, and the elevator start signal is used to instruct the escalator to run.
- an elevator start signal can be generated and sent to instruct the escalator to run .
- the elevator start signal can be sent to the elevator control device, so that the elevator control device controls the operation of the escalator; the elevator start signal can also be sent to the staff, so that the staff can control the operation of the escalator.
- the embodiment of the present disclosure does not limit the type and transmission mode of the elevator start signal.
- the elevator can be started when someone takes the elevator and the elevator is out of service, thereby ensuring the normal use of the elevator.
- the image detection method according to the embodiment of the present disclosure may further include:
- a detection result of the second target is determined, and the detection result includes that the second target is on the escalator or not on the electric escalator on the escalator.
- a trained target detection network (which may be referred to as a second detection network) can also be used to perform second target detection on the first image to determine the first image. Whether there is a second target in .
- the second target may include items prohibited from entering the escalator, such as a baby stroller, a wheelchair, a large luggage, etc.
- the embodiment of the present disclosure does not limit the type of the second target.
- the second detection network may be, for example, a convolutional neural network, and the embodiment of the present disclosure does not limit the network structure and training method of the second detection network.
- the second detection network may determine a region (which may be referred to as a third region) of the second object in the first image.
- 4a and 4b show schematic diagrams of a second object according to an embodiment of the present disclosure.
- the second object in FIG. 4 a is the stroller 41
- the second object in FIG. 4 b is the luggage 42 .
- the detection frame where the second target is located that is, the third area can be determined.
- the detection result of the second target may be determined through the positional relationship between the first area and the third area. That is, it is determined whether the second target is on the escalator according to the positional relationship.
- the step of determining the detection result of the second target according to the positional relationship between the first area and the third area may include:
- the area ratio between the fourth area and the third area is greater than or equal to a third threshold, it is determined that the detection result is that the second target is on the escalator, and the fourth area includes all The intersection area between the first area and the third area.
- IOU Intersection-over-Union
- the area ratio between the fourth area and the third area can be obtained, that is, the area of the intersection area of the area where the second target is located and the elevator area is equal to the The ratio of the area of the area where the second target is located.
- the area ratio is greater than or equal to a preset third threshold, it may be determined that the detection result is that the second target is on the escalator; otherwise, if the area ratio is less than the preset third threshold , it can be determined that the detection result is that the second target is not on the escalator.
- a preset third threshold it may be determined that the detection result is that the second target is on the escalator; otherwise, if the area ratio is less than the preset third threshold , it can be determined that the detection result is that the second target is not on the escalator.
- the second detection network may also directly perform second target detection on the first area image of the elevator area to determine whether there is a second target in the first area image. If the second object exists in the first area image, it may be determined that the second object is on the escalator.
- the embodiments of the present disclosure do not limit the adopted processing manner.
- the second target includes items prohibited from entering the escalator
- the image detection method according to an embodiment of the present disclosure may further include: when the second target is on the escalator Next, send alarm information.
- alarm information can be sent to remind or directly control the escalator to stop running.
- the alarm information can be sent to the elevator monitoring equipment and/or staff in the monitoring room, so that the staff can control the escalator to stop running and/or go to the elevator for processing; the alarm information can also be sent to the elevator control equipment to make the elevator
- the control device controls the escalator to stop running.
- the embodiments of the present disclosure do not limit the type and sending manner of the alarm information.
- an alarm may be issued in response to the second object appearing on the elevator for the first time, and the second object that appears in succession will repeat the alarm within a specified interval to avoid sending alarm information too frequently.
- FIG. 5 shows a schematic diagram of a processing procedure of an image detection method according to an embodiment of the present disclosure.
- the image of the escalator can be input in step 51 of image input; step 52 of elevator positioning and step 53 of key target detection are respectively performed on the image to improve the processing efficiency of the event; Step 52 of elevator positioning determines the elevator area 54 in the image; then, through step 55 of judging empty elevators, it is determined that the elevator is empty/non-empty 57, and then corresponding processing is performed.
- the key target position 56 in the image is determined through the key target detection step 53, and the key target is the item that is prohibited from entering the elevator (violation item); according to the elevator area 54 and the key target position 56, it is determined whether the key target is on the elevator ; If the key target is on the elevator, alarm 58 of illegal items. In this way, the entire processing process of the image detection method according to the embodiment of the present disclosure can be realized.
- step 52 of elevator positioning input the first image of the escalator as shown in Figure 2a, that is, the elevator, build a target detection or area segmentation model based on the deep learning algorithm, and output the elevator detection frame 21 as shown in Figure 2b, Or output the segmentation result of the region 22 as shown in Fig. 2b.
- the target detection method needs to ensure that the detection frame includes the elevator area while reducing the background area as much as possible, so as to prevent background noise from interfering with subsequent elevator services. Therefore, the bounding box of the elevator as shown in Fig. 2b is defined as the detection box 21. At the same time, pay attention to the position of the elevator.
- Figure 2b is offset to the left, with the lower left corner of the elevator area as the benchmark (the lower left coordinate of the bounding box is determined from this), and the right boundary is based on the upper right corner of the elevator area as the benchmark. If the segmentation model is used to build a more accurate elevator area positioning module, the elevator area is defined as the elevator area within the detection frame 21, and a polygon is drawn with the elevator handrail as the boundary, such as area 22 in Figure 2b.
- step 55 of judging empty elevators build a multi-channel parallel elevator state classification model based on deep learning, receive the processing results of the elevator positioning module (taking the detection results as an example), perform data processing simultaneously through various technical solutions, and finally make the results Fusion, determine the idle state of the elevator: an empty elevator or a non-empty elevator, as shown in FIG. 3 , the left elevator 31 is in an idle state, and the right elevator 32 is in a non-idle state.
- This process can be accomplished in several ways:
- the first method is to extract the detection frame area, and output the elevator status through the classification network model
- Method 2 Extract the detection frame area, refer to the segmentation area, set the pixel value of the background area in the detection frame to 0 (black), and output the elevator status through the classification network model;
- the segmentation model is used to obtain the labels of all pixels in all elevator areas, and finally the template of the empty elevator (equivalent to the reference image) is obtained through the Gaussian mixture model, and the proportion of the matching area is obtained through the same The ratio of the number of pixels/total number of pixels is obtained;
- Method 4 Based on the pedestrian and key target detection model, the elevator area is used as the input. If there is no target of interest in the elevator area, the output is an idle elevator, otherwise it is a non-empty elevator.
- step S56 of the key target location detection of key targets such as baby carriages, wheelchairs, and luggage is supported, and various event alarms are supported.
- the output results of the model are a detection target of a stroller 41 (equivalent to the second target) as shown in 4a, and a detection target of a suitcase 42 shown in 4b (equivalent to the second target).
- step 52 of elevator positioning it is judged whether the detection target exists in the elevator area, so as to give an alarm.
- the entrance of the smart elevator solution can be constructed based on the positioning of the elevator area; in the elevator area, the multi-channel parallel empty elevator discrimination method and the result voting mechanism are used to determine whether the elevator is an empty elevator, Improve the accuracy of empty elevator prediction; through the empty elevator alarm, guide the management department to reduce the elevator running speed or even stop running, reduce energy consumption while ensuring the safety and convenience of travel; based on the target detection algorithm, locate the key targets in the elevator area, Give an alarm to the illegal objects, and improve the supervision of the management department.
- elevator area positioning can be performed based on a deep learning target detection/segmentation model, and multi-target detection is supported to locate multiple elevator areas at the same time.
- the two area definition methods proposed by this method can reduce the interference of background noise while ensuring complete coverage of the elevator area.
- the image detection method of the embodiment of the present disclosure it is possible to judge whether the elevator is empty based on the multi-channel parallel empty elevator state discrimination method combined with the result voting mechanism, which can greatly reduce the error rate of output results, and also supports multi-area parallel classification.
- the empty state of the elevator is usually judged by a single method, and the accuracy rate is low.
- the image detection method of the embodiment of the present disclosure by training a deep neural network to detect specific illegal items and identify all key targets within the field of view, the problem of inability to detect illegal items such as baby strollers, wheelchairs, and large suitcases in the related art is solved , and based on the elevator area and key target area, a specific IOU calculation method is used to judge whether the illegal items are on the elevator, which improves the accuracy of the judgment.
- an intelligent elevator system based on automatic positioning and key target detection is proposed, which can form a complete and stable intelligent elevator solution and can be applied to all current public elevator scenarios.
- the method can be applied to smart cameras to locate elevator areas and identify empty elevators in elevator scenes, reduce energy consumption when the elevator is in an empty state, ensure that the elevators are empty before the elevator stops, and improve the work efficiency of the management department; Elevator area positioning and key target detection, alarm the illegal items on the elevator, and strengthen the supervision of the management department.
- the embodiments of the present disclosure also provide image detection apparatuses, electronic devices, computer-readable storage media, and programs, all of which can be used to implement any image detection method provided by the embodiments of the present disclosure, and the corresponding technical solutions and descriptions and refer to the methods part of the corresponding records.
- FIG. 6 shows a block diagram of an image detection apparatus according to an embodiment of the present disclosure. As shown in FIG. 6 , the apparatus includes:
- an image acquisition module 61 configured to acquire a first image of the escalator
- an area detection module 62 configured to perform area detection on the first image, and determine a first area of the escalator in the first image
- the state recognition module 63 is configured to perform elevator state recognition on the first area image corresponding to the first area, and determine at least one state recognition result of the escalator, where the state recognition result includes that the escalator is empty state or in a non-empty state;
- the state determination module 64 is configured to determine the state of the escalator according to the at least one state identification result.
- the state determination module includes: a discriminant value determination sub-module, configured to identify multiple states according to the multiple state identification results and the multiple state identification results when there are multiple state identification results The weight of the result determines the state discriminating value of the escalator; the state determination submodule is configured to determine that the escalator is in an empty state when the state discriminating value is greater than or equal to a first threshold.
- the state identification result includes a first state identification result
- the state identification module includes: a first result determination sub-module, configured to perform classification processing on the first area image, and obtain the The first state recognition result of the escalator is described.
- the state identification result includes a second state identification result
- the state identification module includes: a segmentation sub-module configured to segment the first region image, and divide the first region image into The image is divided into a background area and a foreground area where the escalator is located; the pixel adjustment sub-module is configured to adjust the pixel values of the background area to obtain an adjusted second area image; the second result determination sub-module is configured to In order to classify the image of the second area, a second state recognition result of the escalator is obtained.
- the state identification result includes a third state identification result
- the state identification module includes: a pixel matching sub-module configured to perform pixel matching between the first region image and a preset reference image. matching, to determine the proportion of the matching area between the first area image and the reference image, where the reference image includes the area image corresponding to the escalator in the empty elevator state; the third result determination sub-module is configured to When the proportion of the matching area is greater than or equal to the second threshold, it is determined that the third state recognition result is that the escalator is in an empty state.
- the state identification result includes a fourth state identification result
- the state identification module includes: a detection sub-module configured to perform a first target detection on the first area image, and determine the Whether the first target exists in the first area image; the fourth result determination submodule is configured to determine that the fourth state recognition result is the escalator when the first target does not exist in the first area image in an empty state.
- the device further includes: a shutdown signal sending module configured to send an elevator shutdown signal when the escalator is in an empty state, and the elevator shutdown signal is configured as The escalator is instructed to stop running.
- the device further includes: a start signal sending module configured to send an elevator start signal when the escalator is in a non-empty state and the escalator has stopped running,
- the elevator activation signal is configured to instruct the escalator to operate.
- the apparatus further includes: a target detection module, configured to perform second target detection on the first image, and determine a third area of the second target in the first image; detecting A result determination module, configured to determine a detection result of the second target according to the positional relationship between the first area and the third area, where the detection result includes that the second target is on the escalator or not on the escalator.
- a target detection module configured to perform second target detection on the first image, and determine a third area of the second target in the first image
- a result determination module configured to determine a detection result of the second target according to the positional relationship between the first area and the third area, where the detection result includes that the second target is on the escalator or not on the escalator.
- the detection result determination module includes: a determination sub-module configured to determine, when the area ratio between the fourth region and the third region is greater than or equal to a third threshold, The detection result is that the second target is on the escalator, and the fourth area includes an intersection area between the first area and the third area.
- the second target includes items that are prohibited from entering the escalator
- the apparatus further includes: an alarm information sending module configured to, when the second target is on the escalator, , send alarm information.
- the functions or modules included in the apparatus provided in the embodiments of the present disclosure may be configured to execute the methods described in the above method embodiments, and the specific implementation may refer to the descriptions in the above method embodiments.
- Embodiments of the present disclosure further provide a computer-readable storage medium, on which computer program instructions are stored, and when the computer program instructions are executed by a processor, the foregoing method is implemented.
- the computer-readable storage medium may be a non-volatile computer-readable storage medium.
- An embodiment of the present disclosure further provides an electronic device, comprising: a processor; a memory configured to store instructions executable by the processor; wherein the processor is configured to invoke the instructions stored in the memory to execute the above method.
- Embodiments of the present disclosure also provide a computer program product, including computer-readable code, when the computer-readable code is run on a device, a processor in the device executes a method configured to implement the image detection method provided by any of the above embodiments. instruction.
- Embodiments of the present disclosure further provide another computer program product configured to store computer-readable instructions, which, when executed, cause the computer to perform the operations of the image detection method provided by any of the foregoing embodiments.
- the electronic device may be provided as a terminal, server or other form of device.
- FIG. 7 shows a block diagram of an electronic device 800 according to an embodiment of the present disclosure.
- electronic device 800 may be a mobile phone, computer, digital broadcast terminal, messaging device, game console, tablet device, medical device, fitness device, personal digital assistant, etc. terminal.
- an electronic device 800 may include one or more of the following components: a processing component 802, a memory 804, a power supply component 806, a multimedia component 808, an audio component 810, an input/output (I/O) interface 812 , sensor component 814 and communication component 816 .
- the processing component 802 generally controls the overall operations of the electronic device 800, such as operations associated with display, phone calls, data communications, camera operations, and recording operations.
- the processing component 802 can include one or more processors 820 to execute instructions to perform all or some of the steps of the methods described above.
- processing component 802 may include one or more modules that facilitate interaction between processing component 802 and other components.
- processing component 802 may include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
- Memory 804 is configured to store various types of data to support operation at electronic device 800 . Examples of such data include instructions for any application or method configured to operate on electronic device 800, contact data, phonebook data, messages, pictures, videos, and the like.
- Memory 804 may be implemented by any type of volatile or non-volatile storage device or combination thereof, such as Static Random-Access Memory (SRAM), Electrically Erasable Programmable Read-Only Memory (EEPROM) , Static Random-Access Memory), Erasable Programmable Read-Only Memory (EPROM, Electrically Erasable Programmable Read-Only Memory), Programmable Read-Only Memory (PROM, Programmable Read-Only Memory), Read-Only Memory (ROM, Read Only Memory), magnetic memory, flash memory, magnetic disk or optical disk.
- SRAM Static Random-Access Memory
- EEPROM Electrically Erasable Programmable Read-Only Memory
- EPROM Electrically Erasable Programmable Read-Only Memory
- PROM Programmable Read-Only Memory
- Read-Only Memory
- Power supply assembly 806 provides power to various components of electronic device 800 .
- Power supply components 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power to electronic device 800 .
- Multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user.
- the screen may include a liquid crystal display (LCD, Liquid Crystal Display) and a touch panel (TP, Touch Panel). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from a user.
- the touch panel includes one or more touch sensors to sense touch, swipe, and gestures on the touch panel. The touch sensor may not only sense the boundaries of a touch or swipe action, but also detect the duration and pressure associated with the touch or swipe action.
- the multimedia component 808 includes a front-facing camera and/or a rear-facing camera. When the electronic device 800 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera may receive external multimedia data. Each of the front and rear cameras can be a fixed optical lens system or have focal length and optical zoom capability.
- Audio component 810 is configured to output and/or input audio signals.
- the audio component 810 includes a microphone (MIC, Microphone) that is configured to receive external audio signals when the electronic device 800 is in an operating mode, such as a calling mode, a recording mode, and a voice recognition mode.
- the received audio signal may be further stored in memory 804 or transmitted via communication component 816 .
- audio component 810 also includes a speaker configured to output audio signals.
- the I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module, which may be a keyboard, a click wheel, a button, or the like. These buttons may include, but are not limited to: home button, volume buttons, start button, and lock button.
- Sensor assembly 814 includes one or more sensors configured to provide status assessment of various aspects of electronic device 800 .
- the sensor component 814 can detect the open/closed state of the electronic device 800, the relative positioning of components, such as the display and the keypad of the electronic device 800, and the sensor component 814 can also detect the electronic device 800 or one of the electronic devices 800. Changes in the position of components, presence or absence of user contact with the electronic device 800 , orientation or acceleration/deceleration of the electronic device 800 and changes in the temperature of the electronic device 800 .
- Sensor assembly 814 may include a proximity sensor configured to detect the presence of nearby objects in the absence of any physical contact.
- Sensor assembly 814 may also include a light sensor, such as a Complementary Metal-Oxide-Semiconductor (CMOS) or Charge Coupled Device (CCD) image sensor, for use in imaging applications.
- CMOS Complementary Metal-Oxide-Semiconductor
- CCD Charge Coupled Device
- the sensor assembly 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
- Communication component 816 is configured to facilitate wired or wireless communication between electronic device 800 and other devices.
- the electronic device 800 can access a wireless network based on a communication standard, such as a wireless network (Wi-Fi, Wireless Fidelity), a second-generation mobile communication technology (2G, The 2nd Generation) or a third-generation mobile communication technology (3G, The 3nd Generation) Generation), or their combination.
- the communication component 816 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel.
- the communication component 816 further includes a Near Field Communication (NFC, Near Field Communication) module to facilitate short-range communication.
- NFC Near Field Communication
- the NFC module can be based on Radio Frequency Identification (RFID, Radio Frequency Identification) technology, Infrared Data Association (IrDA, Infrared Data Association) technology, Ultra Wide Band (UWB, Ultra Wide Band) technology, Bluetooth (BT, Blue Tooth) technology and other technologies to achieve.
- RFID Radio Frequency Identification
- IrDA Infrared Data Association
- UWB Ultra Wide Band
- Bluetooth Bluetooth
- electronic device 800 may be implemented by one or more Application Specific Integrated Circuit (ASIC), Digital Signal Processor (DSP), Digital Signal Processing Device (DSPD) Processing Device), Programmable Logic Device (PLD, Programmable Logic Device), Field Programmable Gate Array (FPGA, Field Programmable Gate Array), controller, microcontroller, microprocessor or other electronic component implementation, configured to perform the above method.
- ASIC Application Specific Integrated Circuit
- DSP Digital Signal Processor
- DSPD Digital Signal Processing Device
- PLD Programmable Logic Device
- FPGA Field Programmable Gate Array
- controller microcontroller, microprocessor or other electronic component implementation, configured to perform the above method.
- a non-volatile computer-readable storage medium such as a memory 804 comprising computer program instructions executable by the processor 820 of the electronic device 800 to perform the above method is also provided.
- FIG. 8 shows a block diagram of an electronic device 1900 according to an embodiment of the present disclosure.
- the electronic device 1900 may be provided as a server.
- electronic device 1900 includes processing component 1922, which may include one or more processors, and a memory resource represented by memory 1932 configured to store instructions executable by processing component 1922, such as an application program.
- An application program stored in memory 1932 may include one or more modules, each corresponding to a set of instructions.
- the processing component 1922 is configured to execute instructions to perform the above-described methods.
- the electronic device 1900 may also include a power supply assembly 1926 configured to perform power management of the electronic device 1900, a wired or wireless network interface 1950 configured to connect the electronic device 1900 to a network, and an input output (I/O) interface 1958 .
- the electronic device 1900 can operate based on an operating system stored in the memory 1932, such as a Microsoft server operating system (Windows ServerTM), a graphical user interface based operating system (Mac OS XTM) introduced by Apple, a multi-user multi-process computer operating system (UnixTM). ), a free and open source Unix-like operating system (LinuxTM), an open source Unix-like operating system (FreeBSDTM) or similar systems.
- a non-volatile computer-readable storage medium such as memory 1932 comprising computer program instructions executable by the processing component 1922 of the electronic device 1900 to accomplish the above-described method is also provided.
- Embodiments of the present disclosure may be systems, methods and/or computer program products.
- the computer program product may include a computer-readable storage medium having computer-readable program instructions loaded thereon for causing a processor to implement various aspects of the embodiments of the present disclosure.
- a computer-readable storage medium may be a tangible device that can hold and store instructions for use by the instruction execution device.
- the computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- Computer-readable storage media may include: portable computer disks, hard disks, random access memory (RAM, Random Access Memory), read-only memory, erasable programmable read-only memory (EPROM or flash memory), static random access memory, Portable Compact Disc Read-Only Memory (CD-ROM, Compact Disc Read-Only Memory), Digital Versatile Disc (DVD, Digital Video Disc), memory stick, floppy disk, mechanical coding device, such as a punch card on which instructions are stored Or the protruding structure in the groove, and any suitable combination of the above.
- RAM Random Access Memory
- EPROM or flash memory erasable programmable read-only memory
- static random access memory Portable Compact Disc Read-Only Memory
- CD-ROM Compact Disc Read-Only Memory
- DVD Digital Versatile Disc
- memory stick floppy disk
- mechanical coding device such as a punch card on which instructions are stored Or the protruding structure in the groove, and any suitable combination of the above.
- Computer-readable storage media are not to be construed as transient signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through waveguides or other transmission media (eg, light pulses through fiber optic cables), or through electrical wires transmitted electrical signals.
- the computer readable program instructions described herein may be downloaded to various computing/processing devices from a computer readable storage medium, or to an external computer or external storage device over a network such as the Internet, a local area network, a wide area network, and/or a wireless network.
- the network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer-readable program instructions from a network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in each computing/processing device .
- the computer program instructions for carrying out the operations of the present disclosure may be assembly instructions, Industry Standard Architecture (ISA) instructions, machine instructions, machine-dependent instructions, pseudocode, firmware instructions, state setting data, or in one or more Source or object code written in any combination of programming languages, including object-oriented programming languages such as Smalltalk, C++, etc., and conventional procedural programming languages such as the "C" language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server implement.
- the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or Wide Area Network (WAN), or it may be connected to an external computer (eg, using Internet service provider to connect via the Internet).
- LAN Local Area Network
- WAN Wide Area Network
- electronic circuits such as programmable logic circuits, field programmable gate arrays, or programmable logic arrays, that can execute computer readable program instructions are personalized by utilizing state information of computer readable program instructions , thereby implementing various aspects of the present disclosure.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer or other programmable data processing apparatus to produce a machine that causes the instructions when executed by the processor of the computer or other programmable data processing apparatus , resulting in means for implementing the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- These computer readable program instructions can also be stored in a computer readable storage medium, these instructions cause a computer, programmable data processing apparatus and/or other equipment to operate in a specific manner, so that the computer readable medium on which the instructions are stored includes An article of manufacture comprising instructions for implementing various aspects of the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- Computer readable program instructions can also be loaded onto a computer, other programmable data processing apparatus, or other equipment to cause a series of operational steps to be performed on the computer, other programmable data processing apparatus, or other equipment to produce a computer-implemented process , thereby causing instructions executing on a computer, other programmable data processing apparatus, or other device to implement the functions/acts specified in one or more blocks of the flowcharts and/or block diagrams.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more functions for implementing the specified logical function(s) executable instructions.
- the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or actions , or can be implemented in a combination of dedicated hardware and computer instructions.
- the computer program product can be implemented in hardware, software or a combination thereof.
- the computer program product is embodied as a computer storage medium, and in another optional embodiment, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK) and the like.
- the first area of the escalator in the first image is determined;
- the first area image of the escalator is used for elevator status recognition, and at least one status recognition result of the escalator is determined, and the status recognition result includes that the escalator is in an empty elevator state or in a non-empty elevator state;
- the state of the escalator is determined; in this way, the recognition accuracy of the running state of the escalator can be improved, and a complete and stable intelligent elevator solution can be formed, which can be applied to all current public elevator scenarios.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Geometry (AREA)
- Computational Linguistics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Mathematical Physics (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Business, Economics & Management (AREA)
- Emergency Management (AREA)
- Image Analysis (AREA)
- Escalators And Moving Walkways (AREA)
Abstract
Description
Claims (25)
- 一种拍摄处理方法,包括:A shooting processing method, comprising:获取电动扶梯的第一图像;Get the first image of the escalator;对所述第一图像进行区域检测,确定所述电动扶梯在所述第一图像中的第一区域;Perform area detection on the first image to determine the first area of the escalator in the first image;对与所述第一区域对应的第一区域图像进行电梯状态识别,确定所述电动扶梯的至少一个状态识别结果,所述状态识别结果包括所述电动扶梯处于空梯状态或处于非空梯状态;Perform elevator state recognition on the first area image corresponding to the first area, and determine at least one state recognition result of the escalator, where the state recognition result includes that the escalator is in an empty elevator state or in a non-empty elevator state ;根据所述至少一个状态识别结果,确定所述电动扶梯的状态。According to the at least one state identification result, the state of the escalator is determined.
- 根据权利要求1所述的方法,其中,所述根据所述至少一个状态识别结果,确定所述电动扶梯的状态,包括:The method according to claim 1, wherein the determining the state of the escalator according to the at least one state identification result comprises:在所述状态识别结果为多个的情况下,根据多个状态识别结果及所述多个状态识别结果的权重,确定所述电动扶梯的状态判别值;In the case of multiple state identification results, determining the state identification value of the escalator according to the multiple state identification results and the weights of the multiple state identification results;在所述状态判别值大于或等于第一阈值的情况下,确定所述电动扶梯处于空梯状态。When the state discrimination value is greater than or equal to a first threshold value, it is determined that the escalator is in an empty state.
- 根据权利要求1或2所述的方法,其中,所述状态识别结果包括第一状态识别结果,The method according to claim 1 or 2, wherein the state identification result comprises a first state identification result,所述对与所述第一区域对应的第一区域图像进行电梯状态识别,确定所述电动扶梯的至少一个状态识别结果,包括:The performing elevator state identification on the first area image corresponding to the first area to determine at least one state identification result of the escalator includes:对所述第一区域图像进行分类处理,得到所述电动扶梯的第一状态识别结果。The first area image is classified and processed to obtain a first state recognition result of the escalator.
- 根据权利要求1至3任一项所述的方法,其中,所述状态识别结果包括第二状态识别结果,The method according to any one of claims 1 to 3, wherein the state identification result comprises a second state identification result,所述对与所述第一区域对应的第一区域图像进行电梯状态识别,确定所述电动扶梯的至少一个状态识别结果,包括:The performing elevator state identification on the first area image corresponding to the first area to determine at least one state identification result of the escalator includes:对所述第一区域图像进行分割,将所述第一区域图像分割为背景区域及所述电动扶梯所在的前景区域;Segmenting the first area image, and dividing the first area image into a background area and a foreground area where the escalator is located;对所述背景区域的像素值进行调整,得到调整后的第二区域图像;Adjust the pixel value of the background area to obtain the adjusted second area image;对所述第二区域图像进行分类处理,得到所述电动扶梯的第二状态识别结果。The second area image is classified and processed to obtain the second state recognition result of the escalator.
- 根据权利要求1至4任一项所述的方法,其中,所述状态识别结果包括第三状态识别结果,The method according to any one of claims 1 to 4, wherein the state identification result comprises a third state identification result,所述对与所述第一区域对应的第一区域图像进行电梯状态识别,确定所述电动扶梯的至少一个状态识别结果,包括:The performing elevator state identification on the first area image corresponding to the first area to determine at least one state identification result of the escalator includes:对所述第一区域图像与预设的参考图像进行像素匹配,确定所述第一区域图像与所述参考图像之间的匹配区域占比,所述参考图像包括与处于空梯状态的电动扶梯对应的区域图像;Pixel matching is performed on the first area image and a preset reference image, and the matching area ratio between the first area image and the reference image is determined, and the reference image includes an escalator in an empty state. the corresponding area image;在所述匹配区域占比大于或等于第二阈值的情况下,确定所述第三状态识别结果为所述电动扶梯处于空梯状态。When the proportion of the matching area is greater than or equal to the second threshold, it is determined that the third state recognition result is that the escalator is in an empty state.
- 根据权利要求1至5任一项所述的方法,其中,所述状态识别结果包括第四状态识别结果,The method according to any one of claims 1 to 5, wherein the state identification result comprises a fourth state identification result,所述对与所述第一区域对应的第一区域图像进行电梯状态识别,确定所述电动扶梯的至少一个状态识别结果,包括:The performing elevator state identification on the first area image corresponding to the first area to determine at least one state identification result of the escalator includes:对所述第一区域图像进行第一目标检测,确定所述第一区域图像中是否存在第一目标;performing a first target detection on the first area image to determine whether there is a first target in the first area image;在所述第一区域图像中不存在第一目标的情况下,确定所述第四状态识别结果为所 述电动扶梯处于空梯状态。In the case that the first target does not exist in the first area image, it is determined that the fourth state recognition result is that the escalator is in an empty state.
- 根据权利要求1至6任一项所述的方法,其中,所述方法还包括:The method according to any one of claims 1 to 6, wherein the method further comprises:在所述电动扶梯处于空梯状态的情况下,发送电梯停运信号,所述电梯停运信号用于指示所述电动扶梯停止运行。When the escalator is in an empty state, an elevator stop signal is sent, and the elevator stop signal is used to instruct the escalator to stop running.
- 根据权利要求1至7任一项所述的方法,其中,所述方法还包括:The method according to any one of claims 1 to 7, wherein the method further comprises:在所述电动扶梯处于非空梯状态,且所述电动扶梯已停止运行的情况下,发送电梯启动信号,所述电梯启动信号用于指示所述电动扶梯运行。When the escalator is in a non-empty state and the escalator has stopped running, an elevator start signal is sent, and the elevator start signal is used to instruct the escalator to run.
- 根据权利要求1至8任一项所述的方法,其中,所述方法还包括:The method according to any one of claims 1 to 8, wherein the method further comprises:对所述第一图像进行第二目标检测,确定第二目标在所述第一图像中的第三区域;performing second target detection on the first image, and determining a third area of the second target in the first image;根据所述第一区域与所述第三区域之间的位置关系,确定所述第二目标的检测结果,所述检测结果包括所述第二目标处于所述电动扶梯上或未处于所述电动扶梯上。According to the positional relationship between the first area and the third area, a detection result of the second target is determined, and the detection result includes that the second target is on the escalator or not on the electric escalator on the escalator.
- 根据权利要求9所述的方法,其中,所述根据所述第一区域与所述第三区域之间的位置关系,确定所述第二目标的检测结果,包括:The method according to claim 9, wherein the determining the detection result of the second target according to the positional relationship between the first area and the third area comprises:在第四区域与所述第三区域之间的面积比值大于或等于第三阈值的情况下,确定所述检测结果为所述第二目标处于所述电动扶梯上,所述第四区域包括所述第一区域与所述第三区域之间的交集区域。In the case that the area ratio between the fourth area and the third area is greater than or equal to a third threshold, it is determined that the detection result is that the second target is on the escalator, and the fourth area includes all The intersection area between the first area and the third area.
- 根据权利要求9或10所述的方法,其中,所述第二目标包括禁止进入电动扶梯的物品,所述方法还包括:The method of claim 9 or 10, wherein the second target includes items prohibited from entering the escalator, the method further comprising:在所述第二目标处于所述电动扶梯上的情况下,发送告警信息。When the second target is on the escalator, alarm information is sent.
- 一种图像检测装置,包括:An image detection device, comprising:图像获取模块,配置为获取电动扶梯的第一图像;an image acquisition module, configured to acquire a first image of the escalator;区域检测模块,配置为对所述第一图像进行区域检测,确定所述电动扶梯在所述第一图像中的第一区域;an area detection module, configured to perform area detection on the first image, and determine a first area of the escalator in the first image;状态识别模块,配置为对与所述第一区域对应的第一区域图像进行电梯状态识别,确定所述电动扶梯的至少一个状态识别结果,所述状态识别结果包括所述电动扶梯处于空梯状态或处于非空梯状态;A state recognition module configured to perform elevator state recognition on a first area image corresponding to the first area, and determine at least one state recognition result of the escalator, where the state recognition result includes that the escalator is in an empty state or in a non-empty state;状态确定模块,配置为根据所述至少一个状态识别结果,确定所述电动扶梯的状态。The state determination module is configured to determine the state of the escalator according to the at least one state identification result.
- 根据权利要求12所述的装置,其中,所述状态确定模块包括:The apparatus of claim 12, wherein the state determination module comprises:判别值确定子模块,配置为在所述状态识别结果为多个的情况下,根据多个状态识别结果及所述多个状态识别结果的权重,确定所述电动扶梯的状态判别值;a discriminant value determination submodule, configured to determine the state discriminant value of the escalator according to the multiple state identification results and the weights of the multiple state identification results when the state identification results are multiple;状态确定子模块,配置为在所述状态判别值大于或等于第一阈值的情况下,确定所述电动扶梯处于空梯状态。The state determination submodule is configured to determine that the escalator is in an empty state when the state discrimination value is greater than or equal to a first threshold.
- 根据权利要求12或13所述的装置,其中,所述状态识别结果包括第一状态识别结果,所述状态识别模块包括:The apparatus according to claim 12 or 13, wherein the state identification result includes a first state identification result, and the state identification module includes:第一结果确定子模块,配置为对所述第一区域图像进行分类处理,得到所述电动扶梯的第一状态识别结果。The first result determination sub-module is configured to perform classification processing on the first area image to obtain a first state recognition result of the escalator.
- 根据权利要求12至14任一项所述的装置,其中,所述状态识别结果包括第二状态识别结果,所述状态识别模块包括:The device according to any one of claims 12 to 14, wherein the state identification result includes a second state identification result, and the state identification module includes:分割子模块,配置为对所述第一区域图像进行分割,将所述第一区域图像分割为背景区域及所述电动扶梯所在的前景区域;a segmentation submodule, configured to segment the first area image, and divide the first area image into a background area and a foreground area where the escalator is located;像素调整子模块,配置为对所述背景区域的像素值进行调整,得到调整后的第二区域图像;a pixel adjustment sub-module, configured to adjust the pixel value of the background area to obtain an adjusted second area image;第二结果确定子模块,配置为对所述第二区域图像进行分类处理,得到所述电动扶梯的第二状态识别结果。The second result determination sub-module is configured to perform classification processing on the second area image to obtain a second state recognition result of the escalator.
- 根据权利要求12至15任一项所述的装置,其中,所述状态识别结果包括第三状态识别结果,所述状态识别模块包括:The device according to any one of claims 12 to 15, wherein the state identification result includes a third state identification result, and the state identification module includes:像素匹配子模块,配置为对所述第一区域图像与预设的参考图像进行像素匹配,确定所述第一区域图像与所述参考图像之间的匹配区域占比,所述参考图像包括与处于空梯状态的电动扶梯对应的区域图像;A pixel matching sub-module, configured to perform pixel matching on the first area image and a preset reference image, and determine the matching area ratio between the first area image and the reference image, the reference image including and The image of the area corresponding to the escalator in the empty state;第三结果确定子模块,配置为在所述匹配区域占比大于或等于第二阈值的情况下,确定所述第三状态识别结果为所述电动扶梯处于空梯状态。The third result determination sub-module is configured to determine that the third state identification result is that the escalator is in an empty state when the proportion of the matching area is greater than or equal to the second threshold.
- 根据权利要求12至16任一项所述的装置,其中,所述状态识别结果包括第四状态识别结果,所述状态识别模块包括:The device according to any one of claims 12 to 16, wherein the state identification result includes a fourth state identification result, and the state identification module includes:检测子模块,配置为对所述第一区域图像进行第一目标检测,确定所述第一区域图像中是否存在第一目标;a detection submodule, configured to perform a first target detection on the first area image to determine whether there is a first target in the first area image;第四结果确定子模块,配置为在所述第一区域图像中不存在第一目标的情况下,确定所述第四状态识别结果为所述电动扶梯处于空梯状态。The fourth result determination sub-module is configured to determine that the fourth state recognition result is that the escalator is in an empty state when the first target does not exist in the first area image.
- 根据权利要求12至17任一项所述的装置,其中,所述装置还包括:The apparatus of any one of claims 12 to 17, wherein the apparatus further comprises:停运信号发送模块,配置为在所述电动扶梯处于空梯状态的情况下,发送电梯停运信号,所述电梯停运信号配置为指示所述电动扶梯停止运行。A shutdown signal sending module is configured to send an elevator shutdown signal when the escalator is in an empty state, and the elevator shutdown signal is configured to instruct the escalator to stop running.
- 根据权利要求12至18任一项所述的装置,其中,所述装置还包括:The apparatus of any one of claims 12 to 18, wherein the apparatus further comprises:启动信号发送模块,配置为在所述电动扶梯处于非空梯状态,且所述电动扶梯已停止运行的情况下,发送电梯启动信号,所述电梯启动信号配置为指示所述电动扶梯运行。A start signal sending module is configured to send an elevator start signal when the escalator is in a non-empty state and the escalator has stopped running, and the elevator start signal is configured to instruct the escalator to run.
- 根据权利要求12至19任一项所述的装置,其中,所述装置还包括:The apparatus of any one of claims 12 to 19, wherein the apparatus further comprises:目标检测模块,配置为对所述第一图像进行第二目标检测,确定第二目标在所述第一图像中的第三区域;a target detection module, configured to perform second target detection on the first image, and determine a third area of the second target in the first image;检测结果确定模块,配置为根据所述第一区域与所述第三区域之间的位置关系,确定所述第二目标的检测结果,所述检测结果包括所述第二目标处于所述电动扶梯上或未处于所述电动扶梯上。a detection result determination module, configured to determine a detection result of the second target according to the positional relationship between the first area and the third area, the detection result includes that the second target is in the escalator on or not on the escalator.
- 根据权利要求20所述的装置,其中,所述检测结果确定模块包括:The apparatus according to claim 20, wherein the detection result determination module comprises:确定子模块,配置为在第四区域与所述第三区域之间的面积比值大于或等于第三阈值的情况下,确定所述检测结果为所述第二目标处于所述电动扶梯上,所述第四区域包括所述第一区域与所述第三区域之间的交集区域。A determination sub-module configured to determine that the detection result is that the second target is on the escalator when the area ratio between the fourth area and the third area is greater than or equal to a third threshold, so The fourth area includes an intersection area between the first area and the third area.
- 根据权利要求20或21所述的装置,其中,所述第二目标包括禁止进入电动扶梯的物品,所述装置还包括:The apparatus of claim 20 or 21, wherein the second target includes items prohibited from entering the escalator, the apparatus further comprising:告警信息发送模块,配置为在所述第二目标处于所述电动扶梯上的情况下,发送告警信息。The alarm information sending module is configured to send alarm information when the second target is on the escalator.
- 一种电子设备,包括:An electronic device comprising:处理器;processor;配置为存储处理器可执行指令的存储器;a memory configured to store processor executable instructions;其中,所述处理器被配置为调用所述存储器存储的指令,以执行权利要求1至11中任意一项所述的方法。wherein the processor is configured to invoke instructions stored in the memory to perform the method of any one of claims 1-11.
- 一种计算机可读存储介质,其上存储有计算机程序指令,所述计算机程序指令被处理器执行时实现权利要求1至11中任意一项所述的方法。A computer-readable storage medium having computer program instructions stored thereon, the computer program instructions implementing the method of any one of claims 1 to 11 when executed by a processor.
- 一种计算机程序,包括计算机可读代码,当所述计算机可读代码在电子设备中运行时,所述电子设备中的处理器执行配置为实现权利要求1至11中任意一项所述的方法。A computer program comprising computer readable code, when the computer readable code is run in an electronic device, the execution of the processor in the electronic device is configured to implement the method of any one of claims 1 to 11 .
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2022532078A JP2023510477A (en) | 2020-12-25 | 2021-06-22 | Image detection method and device, electronic device, and storage medium |
KR1020227018450A KR20220095218A (en) | 2020-12-25 | 2021-06-22 | Image detection method and apparatus, electronic device and storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011559951.4 | 2020-12-25 | ||
CN202011559951.4A CN112560986B (en) | 2020-12-25 | 2020-12-25 | Image detection method and device, electronic equipment and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2022134504A1 WO2022134504A1 (en) | 2022-06-30 |
WO2022134504A9 true WO2022134504A9 (en) | 2022-07-21 |
Family
ID=75034245
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/101619 WO2022134504A1 (en) | 2020-12-25 | 2021-06-22 | Image detection method and apparatus, electronic device, and storage medium |
Country Status (4)
Country | Link |
---|---|
JP (1) | JP2023510477A (en) |
KR (1) | KR20220095218A (en) |
CN (2) | CN112560986B (en) |
WO (1) | WO2022134504A1 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112560986B (en) * | 2020-12-25 | 2022-01-04 | 上海商汤智能科技有限公司 | Image detection method and device, electronic equipment and storage medium |
CN114155483A (en) * | 2021-11-11 | 2022-03-08 | 鸿富锦精密电子(郑州)有限公司 | Monitoring alarm method, device, storage medium and computer equipment |
CN114036987B (en) * | 2021-11-12 | 2024-05-31 | 上海擎朗智能科技有限公司 | Staircase detection method and device, mobile equipment and storage medium |
CN116363575B (en) * | 2023-02-15 | 2023-11-03 | 南京诚勤教育科技有限公司 | Classroom monitoring management system based on wisdom campus |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103699878B (en) * | 2013-12-09 | 2017-05-03 | 安维思电子科技(广州)有限公司 | Method and system for recognizing abnormal operation state of escalator |
JP6134641B2 (en) * | 2013-12-24 | 2017-05-24 | 株式会社日立製作所 | Elevator with image recognition function |
EP4134626A1 (en) * | 2017-06-02 | 2023-02-15 | Apple Inc. | Venues map application and system |
CN107273852A (en) * | 2017-06-16 | 2017-10-20 | 华南理工大学 | Escalator floor plates object and passenger behavior detection algorithm based on machine vision |
CN107416630B (en) * | 2017-09-05 | 2020-04-28 | 日立楼宇技术(广州)有限公司 | Method and system for detecting abnormal closing of elevator |
CN109353907B (en) * | 2017-09-05 | 2020-09-15 | 日立楼宇技术(广州)有限公司 | Safety prompting method and system for elevator operation |
CN107832730A (en) * | 2017-11-23 | 2018-03-23 | 高域(北京)智能科技研究院有限公司 | Improve the method and face identification system of face recognition accuracy rate |
US11029810B2 (en) * | 2018-05-07 | 2021-06-08 | Otis Elevator Company | Equipment service graphical interface |
CN108639921A (en) * | 2018-07-05 | 2018-10-12 | 江苏瑞奇海力科技有限公司 | A kind of staircase passenger safety prior-warning device and method |
CN110342357A (en) * | 2019-05-24 | 2019-10-18 | 深圳壹账通智能科技有限公司 | A kind of elevator scheduling method, device, computer equipment and storage medium |
CN110427741B (en) * | 2019-07-31 | 2021-07-27 | Oppo广东移动通信有限公司 | Fingerprint identification method and related product |
CN110852253A (en) * | 2019-11-08 | 2020-02-28 | 杭州宇泛智能科技有限公司 | Ladder control scene detection method and device and electronic equipment |
CN111339846B (en) * | 2020-02-12 | 2022-08-12 | 深圳市商汤科技有限公司 | Image recognition method and device, electronic equipment and storage medium |
CN111325188A (en) * | 2020-03-24 | 2020-06-23 | 通力电梯有限公司 | Method for monitoring escalator and device for monitoring escalator |
CN111913857A (en) * | 2020-07-08 | 2020-11-10 | 浙江大华技术股份有限公司 | Method and device for detecting operation behavior of intelligent equipment |
CN111924695A (en) * | 2020-07-09 | 2020-11-13 | 上海市隧道工程轨道交通设计研究院 | Intelligent safety protection system for subway escalator and working method of intelligent safety protection system |
CN111807204A (en) * | 2020-07-27 | 2020-10-23 | 苏州雷格特智能设备股份有限公司 | Intelligent elevator safety monitoring system |
CN112102407A (en) * | 2020-09-09 | 2020-12-18 | 北京市商汤科技开发有限公司 | Display equipment positioning method and device, display equipment and computer storage medium |
CN111931701B (en) * | 2020-09-11 | 2021-01-15 | 平安国际智慧城市科技股份有限公司 | Gesture recognition method and device based on artificial intelligence, terminal and storage medium |
CN112560986B (en) * | 2020-12-25 | 2022-01-04 | 上海商汤智能科技有限公司 | Image detection method and device, electronic equipment and storage medium |
-
2020
- 2020-12-25 CN CN202011559951.4A patent/CN112560986B/en active Active
- 2020-12-25 CN CN202111579927.1A patent/CN114283305A/en not_active Withdrawn
-
2021
- 2021-06-22 JP JP2022532078A patent/JP2023510477A/en not_active Withdrawn
- 2021-06-22 WO PCT/CN2021/101619 patent/WO2022134504A1/en active Application Filing
- 2021-06-22 KR KR1020227018450A patent/KR20220095218A/en active Search and Examination
Also Published As
Publication number | Publication date |
---|---|
CN114283305A (en) | 2022-04-05 |
CN112560986A (en) | 2021-03-26 |
KR20220095218A (en) | 2022-07-06 |
JP2023510477A (en) | 2023-03-14 |
CN112560986B (en) | 2022-01-04 |
WO2022134504A1 (en) | 2022-06-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022134504A9 (en) | Image detection method and apparatus, electronic device, and storage medium | |
WO2022183661A1 (en) | Event detection method and apparatus, electronic device, storage medium, and program product | |
TWI766286B (en) | Image processing method and image processing device, electronic device and computer-readable storage medium | |
EP3163498B1 (en) | Alarming method and device | |
US11410001B2 (en) | Method and apparatus for object authentication using images, electronic device, and storage medium | |
TWI728621B (en) | Image processing method and device, electronic equipment, computer readable storage medium and computer program | |
US11321575B2 (en) | Method, apparatus and system for liveness detection, electronic device, and storage medium | |
US20210133468A1 (en) | Action Recognition Method, Electronic Device, and Storage Medium | |
US11288531B2 (en) | Image processing method and apparatus, electronic device, and storage medium | |
US20210166040A1 (en) | Method and system for detecting companions, electronic device and storage medium | |
WO2022183663A1 (en) | Event detection method and apparatus, and electronic device, storage medium and program product | |
US20200250495A1 (en) | Anchor determination method and apparatus, electronic device, and storage medium | |
US20190035104A1 (en) | Object detection and tracking | |
TWI757756B (en) | Pedestrian event detection method and detection device, electronic device and computer-readable storage medium | |
WO2022160569A1 (en) | Method and apparatus for detecting security check anomaly event, electronic device, and storage medium | |
US11222231B2 (en) | Target matching method and apparatus, electronic device, and storage medium | |
WO2022134388A1 (en) | Method and device for rider fare evasion detection, electronic device, storage medium, and computer program product | |
CN107122743B (en) | Security monitoring method and device and electronic equipment | |
WO2023071174A1 (en) | Occupancy detection method and apparatus, electronic device, and storage medium | |
CN111435422B (en) | Action recognition method, control method and device, electronic equipment and storage medium | |
US20210326649A1 (en) | Configuration method and apparatus for detector, storage medium | |
Liu et al. | Infrasee: An unobtrusive alertness system for pedestrian mobile phone users | |
CN112464898A (en) | Event detection method and device, electronic equipment and storage medium | |
CN116778415A (en) | Crowd counting network model for unmanned aerial vehicle and counting method | |
CN113569749A (en) | Elevator trapping detection method and device, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2022532078 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20227018450 Country of ref document: KR Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21908512 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21908512 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 08.12.2023) |