GB2496742A - Hazardous device detection training system - Google Patents
Hazardous device detection training system Download PDFInfo
- Publication number
- GB2496742A GB2496742A GB1220213.1A GB201220213A GB2496742A GB 2496742 A GB2496742 A GB 2496742A GB 201220213 A GB201220213 A GB 201220213A GB 2496742 A GB2496742 A GB 2496742A
- Authority
- GB
- United Kingdom
- Prior art keywords
- text
- virtual
- sensor
- environment
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012549 training Methods 0.000 title abstract description 27
- 238000001514 detection method Methods 0.000 title description 9
- 231100001261 hazardous Toxicity 0.000 title description 5
- 239000002360 explosive Substances 0.000 claims abstract description 4
- 230000001419 dependent effect Effects 0.000 claims description 2
- 230000003993 interaction Effects 0.000 claims description 2
- 238000004088 simulation Methods 0.000 claims description 2
- 239000002689 soil Substances 0.000 claims description 2
- 239000000126 substance Substances 0.000 claims description 2
- 230000000694 effects Effects 0.000 claims 1
- 239000003673 groundwater Substances 0.000 claims 1
- 239000000463 material Substances 0.000 claims 1
- 239000004081 narcotic agent Substances 0.000 claims 1
- 239000011824 nuclear material Substances 0.000 claims 1
- 238000000034 method Methods 0.000 abstract description 12
- 238000011161 development Methods 0.000 abstract description 2
- 239000000383 hazardous chemical Substances 0.000 abstract 1
- 230000003190 augmentative effect Effects 0.000 description 3
- 230000004044 response Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000011065 in-situ storage Methods 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000149 penetrating effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B9/00—Simulators for teaching or training purposes
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B9/00—Simulators for teaching or training purposes
- G09B9/003—Simulators for teaching or training purposes for military purposes and tactics
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B9/00—Simulators for teaching or training purposes
- G09B9/006—Simulators for teaching or training purposes for locating or ranging of objects
Abstract
The present invention involves the extension of the well known technique of virtual reality to encompass the use of virtual or real sensors as a means of detecting virtual threats including mines, improvised explosive devices and hazardous materials in complex threat environments and to allow fully immersive computer based basic training, pre-deployment training or mission specific training. It also allows the development of techniques, tactics and procedures in a representative environment to improve the effectiveness of operations and to explore a range of scenarios which might be experienced in real life by an individual or a group of personnel using sensors including both handheld sensors and sensors mounted on machines including robotic platforms to detect the presence of threats.
Description
Hazardous Device Detection Training System The present invention relates to a system for simulating a user environment in which a sensor or detector is used operationally for the location detection and marking of mines, improvised explosive devices, ammunition or arms or other similar threats.
Virtual reality (VR) training systems are themselves reasonably well-known. For example, such systems have been disclosed for training purposes in sea mine disposal based on physical search. Other known systems are used for remote hazardous incident training for emergency services personnel including the remote operation of a vehicle in military situations or landmine clearance training using wireless physical simulated landmines and the remote operation of a robotic arm vehicle for clearing landmines. Such systems are considered the closest art to the present invention in terms of disclosing VR or remote operation systems for use in hazardous and-based or littoral environments.
However, none of these prior art systems employ simulated virtual threats whose characteristics are detected using virtual sensors.
A number of systems with differing types of sensors are used for improvised explosive device (lED) detection. Whilst a number of sensors can prove effective in counter lED detection and in the clearance of landmines as well as in counter terrorism and operations to search locations in civil as well as conflict or post conflict scenarios, they rely on the effective training and operation by the user of the sensor for maximum effectiveness. This in turn requires significant levels of training and demanding levels of concentration to ensure that swept ground is covered sufficiently, objects are not missed and false alarms are minimised. The creation of such environments with representative threat scenarios and the provision of representative devices is expensive and time consuming. The location of these threats can also be learned by students, thus making objective training and assessment problematic. The present invention allows a systematic training environment which enables objective assessment of the effectiveness of the operators, reduces the training burden and maximises their overall future operational effectiveness.
According to the present invention there is provided a system for simulating the operation of a sensor for detecting a device, the system comprising: means for generating a virtual environment and for providing data relating to the environment to a user; a physical interface device representing the sensor and receiving input from a user so that the user can interact with the virtual environment; and means for receiving input from the physical interface device and for interacting with the virtual environment generating means to produce a virtual representation of the sensor in the virtual environment, the virtual representation including the manner of operation of the sensor provided in the virtual representation.
An example of the present invention will now be described with reference to the figures, in which: Figure 1 is virtual sensor in a virtual environment in which the operator experiences the virtual threat as if they were using a real detector; Figure 2 is a representation of an area being searched by the detector, showing the track of the virtual sensor and the location of virtual threats (left hand side) and a transparent ground showing a representation of the virtual sensor and virtual threats (right hand side) which indicate to the user a means of ensuring that the detector is level, moved at the right speed and separation from the ground for optimal use; Figure 3 shows the same representation of an area as figure 2 (left hand side) and a virtual vehicle and robot with radar sensors (right hand side); Figure 4a shows an illustration of a co-operative training version of the present invention in which the users do not need headsets since the users are themselves in a real world environment, but the sensors respond to the virtual environment. Figure 4b shows a person wearing a headset to simulate a realistic visual virtual environment. In this virtual environment, the surroundings, the sensor interface and the threats, including ground signs such as disturbed ground marks that could visually alert the user to the potential of a threat, are all virtual. The example of the present invention depicted in figure 4b could be used in a scenario with a single user as shown or in a co-operative training environment as demonstrated by the example in figure 4a.
Figure 5 shows a virtual representation of a typical headset used for fully immersive virtual sensor reality training.
In one example, the invention provides a virtual environment that is fully visually immersive through the use of a headset or equivalent means of displaying the virtual environment to the user in which the user uses a dummy or real detector either directly or through interface to a remotely operated sensor, for example, mounted on a robotically controlled platform. An example of such a headset is shown in figure 5. In another example, the invention provides an augmented environment that is partially visually immersive through the use of a visor or equivalent means of displaying the augmented environment to the user in which the user uses a dummy or real detector. For example, the augmented reality could involve the superposition of virtual objects onto the user's vision of the real environment. In another example, the invention provides a real environment in which the user may be alerted to virtual threats through the use of a dummy or real detector. In addition to the potential physical headset or visor, the system also comprises a physical interface device that the user operates to interact with the virtual environment. In the example of figure 1, this is a device configured to represent a handheld mine detector. This device normally has a location and/or motion detection system associated with it so that the training system can detect the device's position and/or movement and represent this within the virtual environment that it generates. The position of the sensor is monitored by a location sensor such that it is possible to make the sensor respond to the environment and threats emplaced within it as if it were a real sensor responding to a real environment, thus allowing the user real-time interaction with the virtual surroundings as shown in figure 1. In this example, a virtual landrnine detector is generated by the system of the invention and its operation is simulated to operate in the same way as a real detector might, with variations in sensitivity and detection output dependent on the terrain which is being simulated and the nature of the device to be detected, all of which is portrayed to the user.
The system operates in a manner which closely represents the behaviour of a real detector responding to a real target. It could also generate, in the headset or visor, an image which represents the terrain being simulated but including the ability to make the ground or other real opaque features, such as a wall, to be represented in three dimensions as an opaque or transparent feature in a representative environment. The system can embed measured, simulated or a combination of measured and simulated sensor responses to virtual stimuli to replicate the behaviour of real sensors and real targets within the virtual sensor reality environment. The system of the invention allows initial and mission training and has access to reference environments and devices for the generation of a number of scenarios. It is also capable of receiving and processing updates and can enable the definition of complex scenarios. It also allows operation of the sensor in such an environment to allow playback analysis and guidance to the students and users as to how to move the detector in order to optimise the use of the detector. The virtual environment allows the simulation of different terrains, including soil types, false targets and weather conditions including snow and water. It also allows techniques, tactics and procedures to be developed as new threats are identified and characterised.
A key aspect of the virtual reality approach is the ability to let the user either mask the ground and sweep for buried threats in the way they would in usual operation or to work with a transparent or semi-transparent representation of a feature that would otherwise conceal the virtual threat. Within the virtual environment, it is possible to switch the virtual representation of the ground or other features such as a wall to be semi-transparent so that a mine or lED or other threat can be viewed in-situ by a person using the system to aid correlation of the sensor response with the true location of the threat as shown in figure 2.
In figure 2, there is a representation of the area being searched showing the track of the virtual sensor and the location of virtual threats (left hand side) and transparent ground showing a representation of a virtual sensor and virtual threats (right hand side) showing indication to the user a means of ensuring that the detector is level, moved at the right speed and separation from the ground for optimal use.
The system can be configured to provide a virtual representation indicating the path and area over which the virtual sensor has traversed. This can provide feedback either in real time or as feedback to the operator of the invention, such as a student or instructor, and also allow quantitative assessment of the probability of detection of a threat or false alarm rate or concealed target or false target based on the area covered. It can also provide feedback on the system settings used and the speed and orientation of the sensor or other attributes which dictate its effectiveness in operation so as to improve operator effectiveness or to compare the capabilities of different sensors or settings or search procedures or risk factors or as part of developing tactics, techniques and procedures.
Figure 3 shows how the system of the invention can be employed for simulating vehicle mounted or robot mounted detection systems in a similar fashion to the handheld detectors as shown in figure 1.
Additionally it is possible to link multiple users together within a virtual space as shown in figure 4; these users can share information between them or send information to a central node for performance evaluation purposes. Equally, a single user or a number of users can interact with a number of virtual users within the virtual space. Of course, this could also be applied to the vehicle and robot mounted examples of figureS.
Efficiency of training in the use of such systems which are safety critical and used in the clearance of land mines and other threats is absolutely vital. The invention offers the ability to significantly increase the effectiveness of training using virtual sensors in a VR environment. The invention allows training against a range of existing and new threats and includes the development of training techniques and procedures.
The invention also allows training against new targets and scenarios to be undertaken; once data on a single threat has been characterised and recorded in the form of a datafile or entry of characteristics in a database for use with the system this can be transmitted and uploaded to the system and used for training elsewhere or in multiple locations as new threats emerge.
The concept is applicable too as an aid to training personnel in the operation of such sensors either individually and directly by personnel or remotely through use of wireless or wired operation and monitoring, whether static or mobile.
Application of the invention is envisaged primarily for and and littoral operations either alone or in collaboration with other personnel, either with single or multiple sensors either individually or in the form of a collaborative application. The concept is applicable to a multiplicity of types of sensors and deployment scenarios including, for example, hazardous or remote environments to aid real operational use particularly where the personnel are remote from the locality of the sensor and whereby the application of the virtual sensor will aid the efficacy of the use of the sensor and for which setting up realistic training scenarios is complex and costly.
In addition to handheld, ground penetrating or other sensors including through wall sensors, the technique is applicable to other sensors including metal detector or current flow or magnetometers or non linear junction or chemical or biological or radioactivity or nuclear or acoustic or sonar or radar or optical or hyperspectral or infrared sensors. The technique is also applicable to multi-sensor or multi-spectral sensors and to both close-in and stand-off modes of operation.
The invention allows users to play back their performance including showing the user's search track and desired sweep pattern overlaid in their vision or assist the trainer to assess the user by displaying the results on another display. It will also allow quantitative assessment of search effectiveness in a scenario or between scenarios of equivalent complexity but different threat emplacements to avoid students learning scenarios.
Claims (1)
- <claim-text>CLAIMS1. A system for simulating the operation of a sensor for detecting a device, the system comprising: means for generating a virtual environment and for providing data relating to the environment to a user; a physical interface device representing the sensor and receiving input from a user so that the user can interact with the virtual environment; and means for receiving input from the physical interface device and for interacting with the virtual environment generating means to produce a virtual representation of the sensor in the virtual environment, the virtual representation including the manner of operation of the sensor provided in the virtual representation.</claim-text> <claim-text>2. The system of claim 1, wherein the sensor is of the type arranged to detect at least one of mines or improvised explosive devices.</claim-text> <claim-text>3. A system of claim 1, wherein the sensor is of the type arranged to detect at least one of ammunition, arms, narcotics, nuclear material, radiological material, biological or chemical weaponry.</claim-text> <claim-text>4. The system of claim 2 or 3, wherein the sensor is a handheld device.</claim-text> <claim-text>5. The system of claim 4, wherein the physical interface device is a real sensor of the type being simulated.</claim-text> <claim-text>6. The system of claim 2 or 3, wherein the sensor is mounted on a vehicle or a robot.</claim-text> <claim-text>7. The system of any preceding claim wherein the sensor is an array of sensors.B</claim-text> <claim-text>8. The system of any preceding claim wherein the data relating to the virtual environment includes the data representing the location of at least one virtual device representing a device of the type to be detected by the sensor being simulated.</claim-text> <claim-text>9. The system of any preceding claim, wherein the means for generating a virtual environment generates the environment with reference to a database comprising data for the simulation of different terrains including soil types and false targets and weather conditions including snow and ground water conditions.</claim-text> <claim-text>10. The system according claim 9, wherein the database can be updated by a trainer and in which one or more devices to be detected can be placed or relocated by a trainer.</claim-text> <claim-text>11. The system of any preceding claim comprising a plurality of physical interface devices, such that the virtual environment can be shared between and interacted with by multiple users or operators.</claim-text> <claim-text>12. The system of any preceding claim, wherein the manner of operation of the virtual representation of the sensor is varied dependent upon the virtual environment to reflect effects comprising at least one of reflection, refraction, multipath, attenuation or other scattering from a target or environment.</claim-text> <claim-text>13. The system of any preceding claim, wherein the data relating to the virtual environment includes image data representing the virtual environment and wherein the system further comprises means for providing that image data to a user.</claim-text> <claim-text>14. The system of any preceding claim wherein the means for generating data for the environment further generates at least one of an audible, tactile, odour or taste-based data for provision to the user.</claim-text> <claim-text>15. The system of claim 13 or 14, wherein the data is used to augment the interaction of a user with the virtual environment.</claim-text> <claim-text>16. The system of claim 13, 14 or 15, wherein the means for generating an image generates an image showing the user a virtual representation of features of the device being detected to aid them in understanding how the sensor responds.</claim-text> <claim-text>17. The system of claims 13 to 16, wherein the means for generating an image generates an image showing the user a virtual representation indicating the path and area which the virtual sensor has traversed.</claim-text>
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GBGB1119456.0A GB201119456D0 (en) | 2011-11-11 | 2011-11-11 | Hazardous device detection training system |
Publications (3)
Publication Number | Publication Date |
---|---|
GB201220213D0 GB201220213D0 (en) | 2012-12-26 |
GB2496742A true GB2496742A (en) | 2013-05-22 |
GB2496742B GB2496742B (en) | 2013-11-27 |
Family
ID=45421602
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GBGB1119456.0A Ceased GB201119456D0 (en) | 2011-11-11 | 2011-11-11 | Hazardous device detection training system |
GB1220213.1A Active GB2496742B (en) | 2011-11-11 | 2012-11-09 | Hazardous device detection training system |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GBGB1119456.0A Ceased GB201119456D0 (en) | 2011-11-11 | 2011-11-11 | Hazardous device detection training system |
Country Status (3)
Country | Link |
---|---|
US (1) | US20130125028A1 (en) |
EP (1) | EP2592611A1 (en) |
GB (2) | GB201119456D0 (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2016535379A (en) * | 2013-10-16 | 2016-11-10 | パスポート システムズ, インク.Passport Systems, Inc. | Injection of simulated sources into a networked sensor system |
US10650700B2 (en) * | 2015-01-08 | 2020-05-12 | Lawrence Livermore National Security, Llc | Incident exercise in a virtual environment |
US11025498B2 (en) * | 2017-08-23 | 2021-06-01 | Sap Se | Device model to thing model mapping |
US10990683B2 (en) | 2018-05-25 | 2021-04-27 | At&T Intellectual Property I, L.P. | Virtual reality for security augmentation in home and office environments |
US11087049B2 (en) * | 2018-11-27 | 2021-08-10 | Hitachi, Ltd. | Online self-driving car virtual test and development system |
CN111680736A (en) * | 2020-06-03 | 2020-09-18 | 长春博立电子科技有限公司 | Artificial intelligence behavior analysis model training system and method based on virtual reality |
TWI783790B (en) * | 2021-11-23 | 2022-11-11 | 遠東科技大學 | Method, computer program product, and computer readable medium of using tactile robot to assist computational thinking course |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2326266A (en) * | 1997-06-05 | 1998-12-16 | Aea Technology Plc | Field instrument simulator device |
US20020052724A1 (en) * | 2000-10-23 | 2002-05-02 | Sheridan Thomas B. | Hybrid vehicle operations simulator |
US20030210228A1 (en) * | 2000-02-25 | 2003-11-13 | Ebersole John Franklin | Augmented reality situational awareness system and method |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2115589B (en) * | 1982-02-08 | 1985-08-07 | Marconi Instruments Ltd | Training apparatus |
US5240416A (en) * | 1988-11-23 | 1993-08-31 | Bennington Thomas E | Simulator apparatus employing actual craft and simulators |
US5219290A (en) * | 1991-10-16 | 1993-06-15 | Lapp Philip W | Tool for control of a hazard fighting operation |
US5304065A (en) * | 1992-11-13 | 1994-04-19 | Consultec Scientific, Inc. | Instrument simulator system |
US5292254A (en) * | 1993-01-04 | 1994-03-08 | Motorola, Inc. | Method for determining minefield effects in a simulated battlefield |
GB9519098D0 (en) * | 1995-09-19 | 1995-11-22 | Pike Steven D | Contamination training simulator |
US6430997B1 (en) * | 1995-11-06 | 2002-08-13 | Trazer Technologies, Inc. | System and method for tracking and assessing movement skills in multidimensional space |
US6950054B1 (en) * | 2001-12-03 | 2005-09-27 | Cyterra Corporation | Handheld radar frequency scanner for concealed object detection |
US7507089B2 (en) * | 2005-07-15 | 2009-03-24 | Raytheon Company | Methods and apparatus to provide training against improvised explosive devices |
US7922491B2 (en) * | 2005-09-28 | 2011-04-12 | Raytheon Company | Methods and apparatus to provide training against improvised explosive devices |
GB2449694B (en) * | 2007-05-31 | 2010-05-26 | Sony Comp Entertainment Europe | Entertainment system and method |
US8088004B2 (en) * | 2007-10-16 | 2012-01-03 | International Business Machines Corporation | System and method for implementing environmentally-sensitive simulations on a data processing system |
US8615383B2 (en) * | 2008-01-18 | 2013-12-24 | Lockheed Martin Corporation | Immersive collaborative environment using motion capture, head mounted display, and cave |
US9165475B2 (en) * | 2008-02-20 | 2015-10-20 | Hazsim, Llc | Hazardous material detector simulator and training system |
US20090263770A1 (en) * | 2008-02-20 | 2009-10-22 | Ambrose Philip L | System and method for simulating hazardous environments for portable detection meters used by first responders |
US9600067B2 (en) * | 2008-10-27 | 2017-03-21 | Sri International | System and method for generating a mixed reality environment |
US8732592B2 (en) * | 2009-06-08 | 2014-05-20 | Battelle Energy Alliance, Llc | Methods and systems relating to an augmented virtuality environment |
US9002535B2 (en) * | 2010-05-11 | 2015-04-07 | Irobot Corporation | Navigation portals for a remote vehicle control user interface |
US9244133B2 (en) * | 2011-01-13 | 2016-01-26 | Icove And Associates, Llc | Handheld devices and structures to detect sticky devices having magnets |
US8763442B2 (en) * | 2011-08-27 | 2014-07-01 | The Boeing Company | Combined acoustic excitation and standoff chemical sensing for the remote detection of buried explosive charges |
-
2011
- 2011-11-11 GB GBGB1119456.0A patent/GB201119456D0/en not_active Ceased
-
2012
- 2012-11-09 EP EP12192054.0A patent/EP2592611A1/en not_active Ceased
- 2012-11-09 GB GB1220213.1A patent/GB2496742B/en active Active
- 2012-11-09 US US13/673,292 patent/US20130125028A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2326266A (en) * | 1997-06-05 | 1998-12-16 | Aea Technology Plc | Field instrument simulator device |
US20030210228A1 (en) * | 2000-02-25 | 2003-11-13 | Ebersole John Franklin | Augmented reality situational awareness system and method |
US20020052724A1 (en) * | 2000-10-23 | 2002-05-02 | Sheridan Thomas B. | Hybrid vehicle operations simulator |
Also Published As
Publication number | Publication date |
---|---|
EP2592611A1 (en) | 2013-05-15 |
GB201119456D0 (en) | 2011-12-21 |
GB201220213D0 (en) | 2012-12-26 |
US20130125028A1 (en) | 2013-05-16 |
GB2496742B (en) | 2013-11-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2592611A1 (en) | Hazardous device detection training system | |
US20160217578A1 (en) | Systems and methods for mapping sensor feedback onto virtual representations of detection surfaces | |
US20170148214A1 (en) | Virtual reality training | |
US20140167953A1 (en) | Emergency response scenario simulators and simulation techniques | |
US20140323157A1 (en) | Systems and methods for hazardous material simulations and games using internet-connected mobile devices | |
US20200117840A1 (en) | Injection of simulated sources in a system of networked sensors | |
Fernández et al. | Design of a training tool for improving the use of hand‐held detectors in humanitarian demining | |
Bajic et al. | Impact of Flooding on mine action in Bosnia and Herzegovina, Croatia, and Serbia | |
Finn et al. | Evaluating autonomous ground‐robots | |
Raybourn et al. | Applying Model-Based Situational Awareness and Augmented Reality to Next-Generation Physical Security Systems | |
KR102511206B1 (en) | Radiation safety education system based on mixed reality | |
Zhu et al. | Integration of underwater sonar simulation with a geografical information system | |
Schneider et al. | Unmanned systems for radiological and nuclear measuring and mapping | |
GUNES et al. | A Serious Game Application For The Detection Of Explosives | |
Tunick et al. | Developing Neural Scene Understanding for Autonomous Robotic Missions in Realistic Environments | |
Bartel et al. | Metrics for Assessing Underwater Demonstrations for Detection and Classification of Unexploded Ordnance (Presentation) | |
Juarez III | Discrete event simulation model of the Polaris 2.1 Gamma Ray Imaging Radiation Detection Device | |
Hanshaw | Multisensor application for mines and minelike target detection in the operational environment | |
GUNES et al. | A Serious Game Model Proposal for Detecting Explosives | |
Cannarsa et al. | Exploring the Advantages of a Simulation-Based Mission Planning for Underwater Robotics | |
Shimizu et al. | Proposal of simulation platform for robot operations with sound | |
Kaye et al. | MMRP Technology Update | |
Sandoval | International Maritime and Border Security Technologies & Capabilities at Sandia National Laboratories. | |
Sheng et al. | Challenges in standardizing ram testing for small unmanned robotic systems | |
Fernández Saavedra et al. | Design of a training tool for improving the use of hand-held detectors in humanitarian demining |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
732E | Amendments to the register in respect of changes of name or changes affecting rights (sect. 32/1977) |
Free format text: REGISTERED BETWEEN 20150723 AND 20150729 |