CN112379777A - Digital exhibition room gesture recognition system based on target tracking - Google Patents
Digital exhibition room gesture recognition system based on target tracking Download PDFInfo
- Publication number
- CN112379777A CN112379777A CN202011320355.0A CN202011320355A CN112379777A CN 112379777 A CN112379777 A CN 112379777A CN 202011320355 A CN202011320355 A CN 202011320355A CN 112379777 A CN112379777 A CN 112379777A
- Authority
- CN
- China
- Prior art keywords
- module
- image
- target tracking
- output
- gesture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000002452 interceptive effect Effects 0.000 claims abstract description 18
- 238000001514 detection method Methods 0.000 claims abstract description 16
- 230000002457 bidirectional effect Effects 0.000 claims abstract description 15
- 230000010365 information processing Effects 0.000 claims abstract description 10
- 230000009471 action Effects 0.000 claims description 18
- 238000012545 processing Methods 0.000 claims description 16
- 238000000605 extraction Methods 0.000 claims description 14
- 230000005611 electricity Effects 0.000 claims description 12
- 230000000007 visual effect Effects 0.000 claims description 11
- 230000005540 biological transmission Effects 0.000 claims description 7
- 239000004576 sand Substances 0.000 claims description 4
- 239000004509 smoke generator Substances 0.000 claims description 2
- 238000013459 approach Methods 0.000 abstract description 5
- 238000000034 method Methods 0.000 description 10
- 238000005516 engineering process Methods 0.000 description 8
- 230000033001 locomotion Effects 0.000 description 7
- 230000008569 process Effects 0.000 description 7
- 230000003993 interaction Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 238000005259 measurement Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000026058 directional locomotion Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000001454 recorded image Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3231—Monitoring the presence, absence or movement of users
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
Abstract
The invention relates to the technical field of digital exhibition, in particular to a target tracking-based digital exhibition room gesture recognition system which comprises a processor, a main server and a detection unit, wherein the output end of the processor is in bidirectional electric connection with the input end of the main server, the detection unit comprises a distance measuring sensor, an infrared sensor and a camera, and the output ends of the distance measuring sensor and the infrared sensor are in unidirectional electric connection with an information processing module; the invention can detect whether a visitor approaches or leaves and can control the interactive equipment to be turned on or off, thereby saving electric power.
Description
Technical Field
The invention relates to the technical field of digital exhibition, in particular to a target tracking-based digital exhibition room gesture recognition system.
Background
The digital exhibition hall is also called as a digital exhibition hall, a multimedia digital exhibition hall and the like, and means an exhibition hall form which takes multimedia and digital technologies as exhibition technologies, uses the latest movie and animation technology, combines unique graphic digital and multimedia technologies, attracts visitors by various novel technologies and realizes a man-machine interaction mode. The digital exhibition hall integrates various multimedia exhibition display systems into a whole, and comprises a digital sand table, a circular screen/arc screen/dome screen movie hall, a welcome ground screen system, an interactive bar counter, an interactive mirror surface, a touch screen and the like. Meanwhile, various high and new technology is integrated, so that the exhibition hall has great connotation and attraction, the background and significance contained in the exhibition object are deeply excavated through the combined application of media such as videos, sounds, animations and the like, the high-tech visual impact is brought to audiences, and the brand value is greatly improved.
The gesture recognition system is needed to be used in the interaction process of the digital exhibition hall, the gesture of a visitor is recognized through the gesture recognition system, and related equipment is driven to interact, however, some existing gesture recognition systems cannot detect whether the visitor is close to or far away from the equipment, when the visitor is far away from the equipment, the system cannot control the equipment to be closed, the interactive equipment still runs, a large amount of electric power is wasted, some recognition systems are low in recognition degree, some gestures are not easy to recognize, the recognition effect is not good enough, so that the gesture recognition system of the digital exhibition hall, which is good in recognition effect, can detect whether the visitor is close to or far away from the equipment, and can control the interactive equipment to be opened and closed, and the problem is solved through the gesture recognition system of the digital exhibition hall, which is good.
Disclosure of Invention
The present invention is directed to a digital exhibition hall gesture recognition system based on target tracking, so as to solve the problems mentioned in the background art.
In order to achieve the purpose, the invention provides the following technical scheme: a digital exhibition room gesture recognition system based on target tracking comprises a processor, a main server and a detection unit, wherein the output end of the processor is in bidirectional electric connection with the input end of the main server, the detection unit comprises a distance measurement sensor, an infrared sensor and a camera, the output ends of the distance measurement sensor and the infrared sensor are in unidirectional electric connection with an information processing module, the output end of the camera is in unidirectional electric connection with an image processing module, the output end of the information processing module is in unidirectional electric connection with the input end of the main server, the output end of the image processing module is in unidirectional electric connection with an image transmission module, the output end of the image transmission module is in unidirectional electric connection with the input end of the main server, the output end of the main server is in bidirectional electric connection with a driving module, and the output end of the driving module is in bidirectional electric connection, the output end of the processor is respectively and bidirectionally electrically connected with a target tracking system and a recognition processing system, the recognition processing system comprises a gesture storage module, an information extraction module and an image feature matching module, the output end of the processor is bidirectionally and electrically connected with the input end of the image feature matching module, and the gesture storage module, the information extraction module and the image feature matching module are electrically connected in a bidirectional way.
Preferably, the target tracking system comprises an image detection module, an image recognition module, an image tracking module and an action extraction module, wherein the output end of the image detection module is bidirectionally and electrically connected with the input end of the image recognition module, the output end of the image recognition module is bidirectionally and electrically connected with the input end of the image tracking module, and the electrical connection relationship between the image tracking module and the action extraction module is bidirectionally and electrically connected.
Preferably, the digital interaction unit comprises a visual display device, an auditory display device and a physical display device, and the output end of the driving module is respectively and bidirectionally electrically connected with the input ends of the visual display device, the auditory display device and the physical display device.
Preferably, the visual display equipment comprises a large display screen, a projector and a smoke generator, and the physical display equipment comprises an interactive robot, a movable sand table and a mechanical arm.
Preferably, the input end of the processor is electrically connected with a setting module, and the output end of the setting module is unidirectionally and electrically connected with the input end of the processor.
Preferably, the input end of the gesture storage module is electrically connected with an input module, and the output end of the input module is bidirectionally and electrically connected with the input end of the gesture storage module.
Preferably, the data stored in the gesture storage module includes an image, a track and a corresponding meaning of the gesture.
Preferably, the model of the distance measuring sensor is KLH-01T-20hz, and the model of the infrared sensor is P228.
Compared with the prior art, the invention has the following beneficial effects:
1. the invention can detect whether a visitor approaches or leaves and can control the interactive equipment to be turned on or off, thereby saving electric power;
2. through the arrangement of the image detection module, the image recognition module, the image tracking module and the action extraction module, the images fed back to the processor by the camera are detected, recognized and tracked in a matched mode, the gesture action of a visitor in the images is recognized, and through the arrangement of the visual display equipment, the auditory display equipment and the physical display equipment, the driving equipment can control the visual display equipment, the auditory display equipment and the physical display equipment to bring visual and auditory interactive display for the visitor, so that the interestingness of the visitor in the visiting process is improved;
3. through the setting of the setting module, a worker can conveniently change and set various parameters in the processor, and through the setting of the input module, the worker can conveniently input different gesture actions and related meanings into the gesture storage module.
Drawings
FIG. 1 is a schematic diagram of the system of the present invention;
FIG. 2 is a functional block diagram of the recognition processing system of the present invention;
FIG. 3 is a functional block diagram of the target tracking system of the present invention;
FIG. 4 is a schematic block diagram of a digital interactive unit according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1-4, a digital exhibition hall gesture recognition system based on target tracking comprises a processor, a main server and a detection unit, wherein the output end of the processor is bidirectionally electrically connected with the input end of the main server, the detection unit comprises a distance measurement sensor, an infrared sensor and a camera, the output ends of the distance measurement sensor and the infrared sensor are unidirectionally electrically connected with an information processing module, the output end of the camera is unidirectionally electrically connected with an image processing module, the output end of the information processing module is unidirectionally electrically connected with the input end of the main server, the output end of the image processing module is unidirectionally electrically connected with an image transmission module, the output end of the image transmission module is unidirectionally electrically connected with the input end of the main server, the output end of the main server is bidirectionally electrically connected with a driving module, the output end of the driving module is bidirectionally electrically connected with a digital interaction unit, and the output end, the recognition processing system comprises a gesture storage module, an information extraction module and an image feature matching module, the output end of the processor is in bidirectional electric connection with the input end of the image feature matching module, the gesture storage module and the information extraction module and the image feature matching module are in bidirectional electric connection with each other, the system can detect whether a visitor approaches or leaves, and can control the interactive equipment to be turned on or turned off, so that electric power is saved.
In this embodiment, the target tracking system includes an image detection module, an image recognition module, an image tracking module and an action extraction module, an output end of the image detection module is bidirectionally electrically connected with an input end of the image recognition module, an output end of the image recognition module is bidirectionally electrically connected with an input end of the image tracking module, an electrical connection relationship between the image tracking module and the action extraction module is bidirectionally electrically connected, and the image detection module, the image recognition module, the image tracking module and the action extraction module are cooperatively used to detect, recognize and track an image fed back to the processor by the camera and recognize a gesture action of a visitor in the image.
In this embodiment, the digital interactive unit includes vision display device, sense of hearing display device and physics display device, drive module's output respectively with vision display device, the two-way electricity of input of sense of hearing display device and physics display device is connected, through vision display device, the setting of sense of hearing display device and physics display device, drive device can control vision display device, sense of hearing display device and physics display device bring the interactive show of vision and sense of hearing for the visitor, the interest of visitor in the visiting process has been improved.
In this embodiment, the visual display device includes large-scale display screen, projecting apparatus and fog generator, and the physics display device includes interactive robot, removal sand table and arm.
In this embodiment, the input electricity of treater is connected and is provided with the setting module, and the output of setting module is connected with the one-way electricity of the input of treater, and through the setting of setting module, it is convenient for the workman to change and set for each parameter in the treater.
In this embodiment, the input end of the gesture storage module is electrically connected with the input module, the output end of the input module is electrically connected with the input end of the gesture storage module in a bidirectional manner, and through the arrangement of the input module, a worker can conveniently input different gesture actions and related meanings into the gesture storage module.
In this embodiment, the data stored in the gesture storage module includes an image, a trajectory, and a corresponding meaning of the gesture.
In this embodiment, the distance measuring sensor is of the type KLH-01T-20hz, and the infrared sensor is of the type P228.
The working principle is as follows: when the intelligent digital interactive device works, the infrared sensor can detect whether a visitor approaches, the distance measuring sensor can detect the position of the visitor, when the visitor approaches the display device, the infrared sensor and the distance measuring sensor transmit electric signals to the information processing module, the information processing module then transmits the signals to the main server, the main server transmits the signals to the processor, the processor transmits the signals back to the main server after finishing processing the information, the main server transmits a starting signal to the driving module, the driving module can drive the digital interactive device to operate and interact with the visitor, in the interaction process, the camera records the action of the visitor and transmits the action to the image processing module, the image processing module converts the recorded image into data and transmits the data to the main server through the image transmission module, and the main server transmits the data to the processor, at the moment, the processor transmits the image data to the target tracking system, the target tracking system tracks, captures and identifies the gesture motion in the image data, the gesture motion information is transmitted back to the processor, the gesture motion of the visitor can be easily recognized, then the processor uploads the gesture motion information to the recognition processing system, the gesture motion information enters the image feature matching module for matching, meanwhile, the information extraction module can extract and transmit the gesture information in the gesture storage module to the image feature matching module, after the image feature matching module matches corresponding gesture information, the meanings of the related gestures are extracted and uploaded to the processor, then the processor transmits a control instruction to the driving module through the main server, and the driving module drives equipment to interact according to the meanings of the related gestures, so that interactivity and interestingness are increased; after the activity is finished, the visitor keeps away from the equipment, and distance measuring sensor and infrared sensor transmit the signal to the processor through information processing module and main server this moment, and the processor will close the instruction and transmit to drive module through main server afterwards, and then drive module control relevant equipment is closed, reaches the purpose of practicing thrift electric power.
The target tracking technology is one of the hot spots in the field of computer vision research, and has wide application prospects in various aspects such as military reconnaissance, accurate guidance, fire fighting, battlefield evaluation, security monitoring and the like. The non-directional motion of the target changes the appearance pattern of the target and the scene, the non-rigid target structure, the shielding between the targets and between the target and the scene, the motion of the camera and the like, so that the target tracking task becomes more difficult. Tracking is often applied in those application environments where knowledge of the position and shape of each frame of the target is required, and assumptions are often used to constrain the tracking problem in a particular application environment. Tracking can be defined simply as estimating the trajectory in the image plane of an object as it moves around a scene, i.e. a tracking system assigns consistent labels to tracked objects in different frames of the same video. The target tracking system adopts advanced image detection, identification and tracking technology and is matched with a precise motion control system to realize continuous and rapid tracking and capturing of the actions of a plurality of targets in a scene, integrates the tracking detection and video analysis functions of a plurality of targets in a large scene into an independent system, and automatically collects, classifies and links data of the behaviors and events of the targets by intelligently analyzing video information collected by a front-end camera.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.
Claims (8)
1. The utility model provides a digital exhibition room gesture recognition system based on target tracking, includes treater, main server and detecting element, its characterized in that: the output of treater is connected with the two-way electricity of main server's input, detecting element includes range sensor, infrared sensor and camera, the one-way electricity of output of range sensor and infrared sensor is connected with information processing module, the one-way electricity of output of camera is connected with image processing module, information processing module's the output is connected with the one-way electricity of main server's input, the one-way electricity of output of image processing module is connected with image transmission module, image transmission module's the one-way electricity of output and main server's the input is connected, the two-way electricity of output of main server is connected with drive module, the two-way electricity of output of drive module is connected with digital interactive unit, the output of treater is two-way electricity respectively connected with target tracking system and discernment processing system, discernment processing system includes gesture storage module, the output end of the processor is in bidirectional electric connection with the input end of the image feature matching module, and the electric connection relationship among the gesture storage module, the information extraction module and the image feature matching module is bidirectional electric connection.
2. The digital exhibition hall gesture recognition system based on target tracking according to claim 1, characterized in that: the target tracking system comprises an image detection module, an image recognition module, an image tracking module and an action extraction module, wherein the output end of the image detection module is in bidirectional electric connection with the input end of the image recognition module, the output end of the image recognition module is in bidirectional electric connection with the input end of the image tracking module, and the image tracking module is in bidirectional electric connection with the action extraction module in an electric connection relationship.
3. The digital exhibition hall gesture recognition system based on target tracking according to claim 1, characterized in that: the digital interactive unit comprises visual display equipment, auditory display equipment and physical display equipment, and the output end of the driving module is respectively in bidirectional electric connection with the input ends of the visual display equipment, the auditory display equipment and the physical display equipment.
4. The digital exhibition hall gesture recognition system based on target tracking according to claim 3, characterized in that: the visual display equipment comprises a large display screen, a projector and a smoke generator, and the physical display equipment comprises an interactive robot, a movable sand table and a mechanical arm.
5. The digital exhibition hall gesture recognition system based on target tracking according to claim 1, characterized in that: the input end of the processor is electrically connected with a setting module, and the output end of the setting module is unidirectionally and electrically connected with the input end of the processor.
6. The digital exhibition hall gesture recognition system based on target tracking according to claim 1, characterized in that: the input end of the gesture storage module is electrically connected with the input module, and the output end of the input module is electrically connected with the input end of the gesture storage module in a bidirectional mode.
7. The digital exhibition hall gesture recognition system based on target tracking according to claim 1, characterized in that: the data stored by the gesture storage module comprises images, tracks and corresponding meanings of gestures.
8. The digital exhibition hall gesture recognition system based on target tracking according to claim 1, characterized in that: the model of the distance measuring sensor is KLH-01T-20hz, and the model of the infrared sensor is P228.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011320355.0A CN112379777A (en) | 2020-11-23 | 2020-11-23 | Digital exhibition room gesture recognition system based on target tracking |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011320355.0A CN112379777A (en) | 2020-11-23 | 2020-11-23 | Digital exhibition room gesture recognition system based on target tracking |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112379777A true CN112379777A (en) | 2021-02-19 |
Family
ID=74588460
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011320355.0A Pending CN112379777A (en) | 2020-11-23 | 2020-11-23 | Digital exhibition room gesture recognition system based on target tracking |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112379777A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113190108A (en) * | 2021-03-26 | 2021-07-30 | 特斯联科技集团有限公司 | Museum exhibition non-inductive touch and sound linkage method and system |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104050859A (en) * | 2014-05-08 | 2014-09-17 | 南京大学 | Interactive digital stereoscopic sand table system |
CN106710490A (en) * | 2016-12-26 | 2017-05-24 | 上海斐讯数据通信技术有限公司 | Show window system and practice method thereof |
CN206584716U (en) * | 2017-03-07 | 2017-10-24 | 杭州杭景模型有限公司 | Electronic digit sand table system |
CN107688390A (en) * | 2017-08-28 | 2018-02-13 | 武汉大学 | A kind of gesture recognition controller based on body feeling interaction equipment |
CN207882853U (en) * | 2018-03-19 | 2018-09-18 | 深圳市安星装饰设计工程有限公司 | A kind of intelligent information release system |
CN109116987A (en) * | 2018-08-13 | 2019-01-01 | 连云港易圣游网络科技有限公司 | A kind of holographic display system based on Kinect gesture control |
CN110209285A (en) * | 2019-06-19 | 2019-09-06 | 哈尔滨拓博科技有限公司 | A kind of sand table display systems based on gesture control |
CN111078036A (en) * | 2020-01-14 | 2020-04-28 | 宁波新文三维股份有限公司 | Interactive multimedia display system and network-based control method thereof |
-
2020
- 2020-11-23 CN CN202011320355.0A patent/CN112379777A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104050859A (en) * | 2014-05-08 | 2014-09-17 | 南京大学 | Interactive digital stereoscopic sand table system |
CN106710490A (en) * | 2016-12-26 | 2017-05-24 | 上海斐讯数据通信技术有限公司 | Show window system and practice method thereof |
CN206584716U (en) * | 2017-03-07 | 2017-10-24 | 杭州杭景模型有限公司 | Electronic digit sand table system |
CN107688390A (en) * | 2017-08-28 | 2018-02-13 | 武汉大学 | A kind of gesture recognition controller based on body feeling interaction equipment |
CN207882853U (en) * | 2018-03-19 | 2018-09-18 | 深圳市安星装饰设计工程有限公司 | A kind of intelligent information release system |
CN109116987A (en) * | 2018-08-13 | 2019-01-01 | 连云港易圣游网络科技有限公司 | A kind of holographic display system based on Kinect gesture control |
CN110209285A (en) * | 2019-06-19 | 2019-09-06 | 哈尔滨拓博科技有限公司 | A kind of sand table display systems based on gesture control |
CN111078036A (en) * | 2020-01-14 | 2020-04-28 | 宁波新文三维股份有限公司 | Interactive multimedia display system and network-based control method thereof |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113190108A (en) * | 2021-03-26 | 2021-07-30 | 特斯联科技集团有限公司 | Museum exhibition non-inductive touch and sound linkage method and system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Elharrouss et al. | A review of video surveillance systems | |
Kong et al. | Mmact: A large-scale dataset for cross modal human action understanding | |
CN101699862B (en) | Acquisition method of high-resolution region-of-interest image of PTZ camera | |
Dai et al. | Pedestrian detection and tracking in infrared imagery using shape and appearance | |
Wang et al. | Recent developments in human motion analysis | |
Bodor et al. | Vision-based human tracking and activity recognition | |
Hampapur et al. | Smart video surveillance: exploring the concept of multiscale spatiotemporal tracking | |
CN105844659B (en) | The tracking and device of moving component | |
CN201845345U (en) | Facial expression identifying data collecting system based on active vision | |
CN101399971A (en) | Video surveillance system employing video primitives | |
CN104049749A (en) | Method and apparatus to generate haptic feedback from video content analysis | |
CN102164270A (en) | Intelligent video monitoring method and system capable of exploring abnormal events | |
KR20120119144A (en) | Apparatus and method of camera-based intelligent management | |
CN102103409A (en) | Man-machine interaction method and device based on motion trail identification | |
CN103514429A (en) | Method for detecting specific part of object and image processing equipment | |
CN204305213U (en) | The interactive sighting device of multi-cam human body tracking | |
CN106710490A (en) | Show window system and practice method thereof | |
CN202231803U (en) | Intelligent monitoring system for multi-aspect video object tracking detection | |
CN112379777A (en) | Digital exhibition room gesture recognition system based on target tracking | |
CN112785564B (en) | Pedestrian detection tracking system and method based on mechanical arm | |
CN111711796A (en) | Intelligent human shape detection system | |
Boghossian et al. | Image processing system for pedestrian monitoring using neural classification of normal motion patterns | |
CN116189041A (en) | Violence action recognition method based on visual optical flow space-time analysis | |
Bhaltilak et al. | Human motion analysis with the help of video surveillance: a review | |
de Diego et al. | Scalable and flexible wireless distributed architecture for intelligent video surveillance systems |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |