EP1860395A1 - Spotter scope - Google Patents
Spotter scope Download PDFInfo
- Publication number
- EP1860395A1 EP1860395A1 EP07107673A EP07107673A EP1860395A1 EP 1860395 A1 EP1860395 A1 EP 1860395A1 EP 07107673 A EP07107673 A EP 07107673A EP 07107673 A EP07107673 A EP 07107673A EP 1860395 A1 EP1860395 A1 EP 1860395A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- bullet
- video
- target
- point
- intended
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Images
Classifications
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F41—WEAPONS
- F41G—WEAPON SIGHTS; AIMING
- F41G1/00—Sighting devices
- F41G1/46—Sighting devices for particular applications
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F41—WEAPONS
- F41G—WEAPON SIGHTS; AIMING
- F41G1/00—Sighting devices
- F41G1/38—Telescopic sights specially adapted for smallarms or ordnance; Supports or mountings therefor
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F41—WEAPONS
- F41G—WEAPON SIGHTS; AIMING
- F41G3/00—Aiming or laying means
- F41G3/02—Aiming or laying means using an independent line of sight
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F41—WEAPONS
- F41G—WEAPON SIGHTS; AIMING
- F41G3/00—Aiming or laying means
- F41G3/14—Indirect aiming means
- F41G3/142—Indirect aiming means based on observation of a first shoot; using a simulated shoot
Definitions
- U.S. military sniper teams generally consist of a shooter and an observer (or spotter).
- the observer uses a non-electronic glass optics-based spotting scope to observe a target, determine distance, and estimate wind speed and direction before a shot is fired.
- the spotter conveys this information to the shooter for point of aim adjustments prior to shooting. Distance is estimated manually.
- the spotter After the shooter fires, the spotter tries to observe the actual path of the bullet (trace) to the intended target (point of impact) through the spotting scope. The spotter then attempts to determine if the target was hit based on the observed trace trajectory. If the target was not hit, the spotter determines where the bullet crossed the plane of the target and suggests an aiming correction to the shooter. Observing target can only be performed during daylight and the trace is extremely difficult to observe even under ideal daylight conditions. Trace observations are also subject to very large errors. Also, if no spotter is present, then observation of the trace trajectory is not possible.
- the present invention provides systems and methods for automatically generating an aim point correction for sniper operations.
- the present invention reduces spotter/sniper workload and improves trace spotting analysis.
- An example system includes a scope, a video capture component, an output device, and a processor in signal communication with the video capture component and the output device.
- the video capture component captures video of a bullet from when the bullet left a weapon to at least when the bullet crossed a previously determined target range.
- the processor determines from the captured video where the bullet was located relative to an intended target when the bullet was at the target range, generates a new aim point if the bullet was determined to have missed an intended hit point, and outputs the generated new aim point to the output device.
- the intended hit point is the intended target.
- the video capture component includes a digital video camera and/or an infrared video camera.
- FIGURE 1 illustrates a perspective view of an example spotter scope formed in accordance with an embodiment of the present invention
- FIGURE 2 illustrates a block diagram of components of the scope shown in FIGURE 1;
- FIGURE 3 is a flow diagram of an example process performed by the scope of FIGURES 1 and 2;
- FIGURE 4 is an example image viewable by a user of the scope.
- FIGURE 5 is a perspective view of a sniper's gun-mounted scope.
- FIGURE 1 shows an example spotter scope 20 formed in accordance with an embodiment of the present invention.
- the scope 20 may be hand-held or mounted to a support device, such as a tripod 40.
- the scope 20 includes a housing 24 with a scope lens 34, a video lens 36, and an infrared lens 38 located at a first end of the housing 24.
- eye pieces 28 At a second end of the housing 24 are eye pieces 28 that correspond to the lenses 34-38, user interface controls 30, and a display device 32.
- the scope 20 includes a processor 60 that is in data communication with user interface controls 30, the display device 32, and an output device 42.
- An example of the output device 42 is a digital micro mirror device (DMD) that is controlled by a Digital Signal Processing (DSP) chip for presenting images in the field of view through the scope lens 34 and via an associated eye piece.
- DMD digital micro mirror device
- DSP Digital Signal Processing
- the processor 60 includes video capture components 80, video processing components 82, and a targeting component 88.
- the video capture components 80 includes a digital video camera associated with the video lens 36 and an infrared video capture component associated with the infrared lens 38.
- the video capture components 80 capture video images of a trajectory of a bullet expelled by a nearby weapon.
- the captured video is sent to the video processing components 82 for analysis.
- the video captured by the digital video camera is processed to determine trajectory of the bullet and at night the video captured by the infrared camera is used to determine bullet trajectory.
- Daytime video capture with the digital video camera can be augmented by the infrared camera where conditions warrant.
- the processing component 82 determines where the bullet was most likely to have crossed the plane of the intended target. If the processing component 82 determines that the trajectory of the bullet shows that the bullet did not hit the intended target, then the targeting component 88 determines an aiming correction location.
- the processing component 82 and the targeting component 88 includes a display component for generating an image of the location of where the bullet crossed the target plane (processing component 82) and an image for a new aiming point (targeting component 88).
- the images are sent to the display device 32 and/or the output device 44 for presentation within the field of view of the scope, other video capture devices may be used.
- the processor 60 may output the captured video to the display device 32.
- the display device 32 may present scope status information, activateable user controls (e.g., touch screen control buttons), previously stored information, or information received (wirelessly or via wire) from another system.
- FIGURE 3 is a flow diagram of an example process 120 performed by the components of the scope 20.
- one of the video capture components 80 records video at some point prior to firing of the weapon that is in close proximity to the scope 20.
- the video capture components 80 may be activated manually by the user interacting with the user interface controls 30 or the display device 32, by activation of a remote control that is in wired or wireless signal communication with the processor 60.
- the remote control device may be a voice capturing device and the processor 60 includes a voice processing component (not shown) that interprets voice signals sent to it via the remote control.
- Activation or deactivation of the capturing of video images can be performed automatically, for example, by sensing activation of the weapon and by deactivating after a predefined period of time from when the weapon was activated.
- image analysis of the captured video is automatically performed in order to determine trajectory of the bullet.
- the processor 60 automatically determines the point where the bullet crossed the intended target based on the determined trajectory, the frame rate of the captured video, a predicted range of the intended target, and a determination of when the bullet left the weapon or when the trigger was pulled.
- the determination of when the bullet left the weapon or trigger activation may be based on a sensed event, such as sound or shock as sensed by a sensing device (not shown).
- processor 60 outputs a dot, such as a red dot, to represent the determined point where the bullet crossed the intended target.
- the outputted dot is presented on the output device 42. If, at the decision block 136, it was determined that the bullet did hit the target, then the process is done, See block 138. However, if the bullet did not hit the target as determined at the decision block 136, the processor 60, at a block 140, determines an aiming correction point based on the point determined at the block 32 and the previous aiming point. At a block 42, a corrected pipper location or aim point location is generated and displayed and outputted by the output device 142 or the display device 32. The determination by the processor 60 of whether the bullet hit the target is based on comparing the point determined at the block 132 to a stored image that is sized according to the determined predicted range of the target.
- FIGURE 4 illustrates an image 160 that a viewer sees through the scope 20.
- a center pipper 166 in this example is located at the center of the intended target.
- the point 168 is displayed to one viewing the image 160 in order to show where the point is that was determined at the block 132.
- a new pipper 170 is generated and outputted according to the block 142.
- the point 168 and pipper 170 are presented within the scope by a DMD and DSP chip.
- the corrected pipper location such as the pipper 170 of FIGURE 4, is conveyed to the sniper.
- the sniper viewing the target through gun-mounted scope 180 adjusts their targeting in order to match the new aim location, .See aim point 188. If it is determined that the new aim location is outside of the MILDOT settings of a typical scope, then the sniper will activate a dial 190 in order to adjust the targeting aim point according to the new aim point.
- the range of the target is predicted manually by the spotter or shooter or automatically by the processor 60.
- the spotter or shooter determines range by known techniques and enters the determined range into the processor 60 using the user interface controls 30 or the display device 32.
- the processor 60 automatically determines range by using image analysis of a center portion of an image recorded by one of the video capture components 80 after the user has placed the crosshair on the intended target and instructed the processor 60 to calculate range.
- the processor 60 performs image matching that matches a prestored target object (upper body human form) to a similar object in the captured image. After a match has been determined, range is determined by determining a width and/or a height dimensions of the matched object in the captured image and comparing that to predefined width and height dimensions for a typical or predefined target.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Aiming, Guidance, Guns With A Light Source, Armor, Camouflage, And Targets (AREA)
- Closed-Circuit Television Systems (AREA)
Abstract
Description
- U.S. military sniper teams generally consist of a shooter and an observer (or spotter). The observer uses a non-electronic glass optics-based spotting scope to observe a target, determine distance, and estimate wind speed and direction before a shot is fired. The spotter conveys this information to the shooter for point of aim adjustments prior to shooting. Distance is estimated manually.
- After the shooter fires, the spotter tries to observe the actual path of the bullet (trace) to the intended target (point of impact) through the spotting scope. The spotter then attempts to determine if the target was hit based on the observed trace trajectory. If the target was not hit, the spotter determines where the bullet crossed the plane of the target and suggests an aiming correction to the shooter. Observing target can only be performed during daylight and the trace is extremely difficult to observe even under ideal daylight conditions. Trace observations are also subject to very large errors. Also, if no spotter is present, then observation of the trace trajectory is not possible.
- Therefore, there exists a need for an improved spotter scope.
- The present invention provides systems and methods for automatically generating an aim point correction for sniper operations. The present invention reduces spotter/sniper workload and improves trace spotting analysis.
- An example system includes a scope, a video capture component, an output device, and a processor in signal communication with the video capture component and the output device. The video capture component captures video of a bullet from when the bullet left a weapon to at least when the bullet crossed a previously determined target range. The processor determines from the captured video where the bullet was located relative to an intended target when the bullet was at the target range, generates a new aim point if the bullet was determined to have missed an intended hit point, and outputs the generated new aim point to the output device.
- In one aspect of the invention, the intended hit point is the intended target.
- In another aspect of the invention, the video capture component includes a digital video camera and/or an infrared video camera.
- Preferred and alternative embodiments of the present invention are described in detail below with reference to the following drawings:
- FIGURE 1 illustrates a perspective view of an example spotter scope formed in accordance with an embodiment of the present invention;
- FIGURE 2 illustrates a block diagram of components of the scope shown in FIGURE 1;
- FIGURE 3 is a flow diagram of an example process performed by the scope of FIGURES 1 and 2;
- FIGURE 4 is an example image viewable by a user of the scope; and
- FIGURE 5 is a perspective view of a sniper's gun-mounted scope.
- FIGURE 1 shows an
example spotter scope 20 formed in accordance with an embodiment of the present invention. Thescope 20 may be hand-held or mounted to a support device, such as atripod 40. Thescope 20 includes ahousing 24 with ascope lens 34, avideo lens 36, and aninfrared lens 38 located at a first end of thehousing 24. At a second end of thehousing 24 areeye pieces 28 that correspond to the lenses 34-38, user interface controls 30, and adisplay device 32. - As shown in FIGURE 2, the
scope 20 includes aprocessor 60 that is in data communication withuser interface controls 30, thedisplay device 32, and anoutput device 42. An example of theoutput device 42 is a digital micro mirror device (DMD) that is controlled by a Digital Signal Processing (DSP) chip for presenting images in the field of view through thescope lens 34 and via an associated eye piece. - In one embodiment, the
processor 60 includesvideo capture components 80,video processing components 82, and atargeting component 88. Thevideo capture components 80 includes a digital video camera associated with thevideo lens 36 and an infrared video capture component associated with theinfrared lens 38. Thevideo capture components 80 capture video images of a trajectory of a bullet expelled by a nearby weapon. The captured video is sent to thevideo processing components 82 for analysis. In a daytime situation, the video captured by the digital video camera is processed to determine trajectory of the bullet and at night the video captured by the infrared camera is used to determine bullet trajectory. Daytime video capture with the digital video camera can be augmented by the infrared camera where conditions warrant. Once the trajectory has been determined from one or both of the generated video images, theprocessing component 82 determines where the bullet was most likely to have crossed the plane of the intended target. If theprocessing component 82 determines that the trajectory of the bullet shows that the bullet did not hit the intended target, then thetargeting component 88 determines an aiming correction location. Theprocessing component 82 and thetargeting component 88 includes a display component for generating an image of the location of where the bullet crossed the target plane (processing component 82) and an image for a new aiming point (targeting component 88). The images are sent to thedisplay device 32 and/or the output device 44 for presentation within the field of view of the scope, other video capture devices may be used. - The
processor 60 may output the captured video to thedisplay device 32. Also, thedisplay device 32 may present scope status information, activateable user controls (e.g., touch screen control buttons), previously stored information, or information received (wirelessly or via wire) from another system. - FIGURE 3 is a flow diagram of an
example process 120 performed by the components of thescope 20. First, at ablock 126, one of thevideo capture components 80 records video at some point prior to firing of the weapon that is in close proximity to thescope 20. Thevideo capture components 80 may be activated manually by the user interacting with theuser interface controls 30 or thedisplay device 32, by activation of a remote control that is in wired or wireless signal communication with theprocessor 60. In one embodiment, the remote control device may be a voice capturing device and theprocessor 60 includes a voice processing component (not shown) that interprets voice signals sent to it via the remote control. Activation or deactivation of the capturing of video images can be performed automatically, for example, by sensing activation of the weapon and by deactivating after a predefined period of time from when the weapon was activated. Next, at ablock 128 image analysis of the captured video is automatically performed in order to determine trajectory of the bullet. At ablock 132, theprocessor 60 automatically determines the point where the bullet crossed the intended target based on the determined trajectory, the frame rate of the captured video, a predicted range of the intended target, and a determination of when the bullet left the weapon or when the trigger was pulled. The determination of when the bullet left the weapon or trigger activation may be based on a sensed event, such as sound or shock as sensed by a sensing device (not shown). - At
block 134,processor 60 outputs a dot, such as a red dot, to represent the determined point where the bullet crossed the intended target. The outputted dot is presented on theoutput device 42. If, at thedecision block 136, it was determined that the bullet did hit the target, then the process is done, Seeblock 138. However, if the bullet did not hit the target as determined at thedecision block 136, theprocessor 60, at a block 140, determines an aiming correction point based on the point determined at theblock 32 and the previous aiming point. At ablock 42, a corrected pipper location or aim point location is generated and displayed and outputted by theoutput device 142 or thedisplay device 32. The determination by theprocessor 60 of whether the bullet hit the target is based on comparing the point determined at theblock 132 to a stored image that is sized according to the determined predicted range of the target. - FIGURE 4 illustrates an
image 160 that a viewer sees through thescope 20. Acenter pipper 166 in this example is located at the center of the intended target. After the weapon has been fired and the analysis has been performed atblocks point 168 is displayed to one viewing theimage 160 in order to show where the point is that was determined at theblock 132. After the correction determination is made at the block 140, anew pipper 170 is generated and outputted according to theblock 142. Thepoint 168 andpipper 170 are presented within the scope by a DMD and DSP chip. - The corrected pipper location, such as the
pipper 170 of FIGURE 4, is conveyed to the sniper. The sniper viewing the target through gun-mountedscope 180 adjusts their targeting in order to match the new aim location, .See aim point 188. If it is determined that the new aim location is outside of the MILDOT settings of a typical scope, then the sniper will activate adial 190 in order to adjust the targeting aim point according to the new aim point. - In one embodiment, the range of the target is predicted manually by the spotter or shooter or automatically by the
processor 60. The spotter or shooter determines range by known techniques and enters the determined range into theprocessor 60 using the user interface controls 30 or thedisplay device 32. Theprocessor 60 automatically determines range by using image analysis of a center portion of an image recorded by one of thevideo capture components 80 after the user has placed the crosshair on the intended target and instructed theprocessor 60 to calculate range. Theprocessor 60 performs image matching that matches a prestored target object (upper body human form) to a similar object in the captured image. After a match has been determined, range is determined by determining a width and/or a height dimensions of the matched object in the captured image and comparing that to predefined width and height dimensions for a typical or predefined target.
Claims (10)
- A method for automatically generating an aim point correction, the method comprising:capturing video of a bullet from when the bullet left a weapon to at least when the bullet crossed a previously determined target range;automatically determining from the captured video where the bullet was located relative to an intended target when the bullet was at the target range;automatically generating a new aim point if the bullet was determined to have missed an intended hit point; andoutputting the generated new aim point.
- The method of Claim 1, wherein the intended hit point is the intended target and wherein capturing includes capturing daytime video images.
- The method of Claim 1, wherein capturing includes capturing infrared video images.
- The method of Claim 1, further comprising automatically determining range of the target.
- The method of Claim 1, wherein outputting includes displaying the generated new aim point in a field of view of a scope.
- A system for automatically generating an aim point correction, the system comprising:a scope;a video capture component configured to capture video of a bullet fromwhen the bullet left a weapon to at least when the bullet crossed a previously determined target range;an output device;a processor in signal communication with the video capture component and the output device, the processor comprising:a first component configured to determine from the captured video where the bullet was located relative to an intended target when the bullet was at the target range;a second component configured to generate a new aim point if the bullet was determined to have missed an intended hit point; anda third component configured to output the generated new aim point to the output device.
- The system of Claim 6, wherein the intended hit point is the intended target and wherein the video capture component includes a digital video camera.
- The system of Claim 6, wherein the video capture component includes an infrared video camera.
- The system of Claim 6, wherein the processor comprises a fourth component configured to determine range of the target.
- The system of Claim 6, wherein the output device includes a component for outputting the generated new aim point in a field of view of the scope.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US74673606P | 2006-05-08 | 2006-05-08 | |
US11/696,050 US20080022575A1 (en) | 2006-05-08 | 2007-04-03 | Spotter scope |
Publications (1)
Publication Number | Publication Date |
---|---|
EP1860395A1 true EP1860395A1 (en) | 2007-11-28 |
Family
ID=38537683
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP07107673A Withdrawn EP1860395A1 (en) | 2006-05-08 | 2007-05-08 | Spotter scope |
Country Status (2)
Country | Link |
---|---|
US (1) | US20080022575A1 (en) |
EP (1) | EP1860395A1 (en) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8353454B2 (en) * | 2009-05-15 | 2013-01-15 | Horus Vision, Llc | Apparatus and method for calculating aiming point information |
US20110315767A1 (en) * | 2010-06-28 | 2011-12-29 | Lowrance John L | Automatically adjustable gun sight |
US8457179B2 (en) | 2010-09-13 | 2013-06-04 | Honeywell International Inc. | Devices, methods, and systems for building monitoring |
US8285305B2 (en) | 2010-09-13 | 2012-10-09 | Honeywell International Inc. | Notifying a user of an event |
CN101975530B (en) * | 2010-10-19 | 2013-06-12 | 李丹韵 | Electronic sighting device and method for regulating and determining graduation thereof |
US20150287224A1 (en) * | 2013-10-01 | 2015-10-08 | Technology Service Corporation | Virtual tracer methods and systems |
US20160069643A1 (en) * | 2014-09-06 | 2016-03-10 | Philip Lyren | Weapon Targeting System |
US10184758B2 (en) * | 2014-09-19 | 2019-01-22 | Philip Lyren | Weapon targeting system |
DE102015120205A1 (en) * | 2015-09-18 | 2017-03-23 | Rheinmetall Defence Electronics Gmbh | Remote weapon station and method of operating a remote weapon station |
DE102015120036A1 (en) * | 2015-11-19 | 2017-05-24 | Rheinmetall Defence Electronics Gmbh | Remote weapon station and method of operating a remote weapon station |
DE102016007624A1 (en) * | 2016-06-23 | 2018-01-11 | Diehl Defence Gmbh & Co. Kg | 1Procedure for file correction of a weapon system |
US11287638B2 (en) | 2019-08-20 | 2022-03-29 | Francesco E. DeAngelis | Reflex sight with superluminescent micro-display, dynamic reticle, and metadata overlay |
IL283793B1 (en) * | 2021-06-07 | 2024-04-01 | Smart Shooter Ltd | System and method for zeroing of smart aiming device |
US11644277B2 (en) * | 2021-07-01 | 2023-05-09 | Raytheon Canada Limited | Digital booster for sights |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4494198A (en) * | 1981-03-12 | 1985-01-15 | Barr & Stroud Limited | Gun fire control systems |
WO2000025082A1 (en) * | 1998-10-23 | 2000-05-04 | Precision Remotes, Inc. | Rapid aiming telepresent system |
US6247259B1 (en) * | 1997-10-09 | 2001-06-19 | The State Of Israel, Atomic Energy Commission, Soreq Nuclear Research Center | Method and apparatus for fire control |
US6252706B1 (en) * | 1997-03-12 | 2001-06-26 | Gabriel Guary | Telescopic sight for individual weapon with automatic aiming and adjustment |
WO2005047805A2 (en) * | 2003-11-12 | 2005-05-26 | Horus Vision, Llc | Apparatus and method for calculating aiming point information |
US20050268521A1 (en) * | 2004-06-07 | 2005-12-08 | Raytheon Company | Electronic sight for firearm, and method of operating same |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5026158A (en) * | 1988-07-15 | 1991-06-25 | Golubic Victor G | Apparatus and method for displaying and storing impact points of firearm projectiles on a sight field of view |
SE506468C2 (en) * | 1996-01-08 | 1997-12-22 | Tommy Andersson | Hit position marker for shotgun shooting |
US5924868A (en) * | 1996-09-18 | 1999-07-20 | Rod; Samuel R. | Method and apparatus for training a shooter of a firearm |
US7158167B1 (en) * | 1997-08-05 | 2007-01-02 | Mitsubishi Electric Research Laboratories, Inc. | Video recording device for a targetable weapon |
US7832137B2 (en) * | 1997-12-08 | 2010-11-16 | Horus Vision, Llc | Apparatus and method for calculating aiming point information |
US7937878B2 (en) * | 1997-12-08 | 2011-05-10 | Horus Vision Llc | Apparatus and method for calculating aiming point information |
US20060005447A1 (en) * | 2003-09-12 | 2006-01-12 | Vitronics Inc. | Processor aided firing of small arms |
US7603804B2 (en) * | 2003-11-04 | 2009-10-20 | Leupold & Stevens, Inc. | Ballistic reticle for projectile weapon aiming systems and method of aiming |
US7269920B2 (en) * | 2004-03-10 | 2007-09-18 | Raytheon Company | Weapon sight with ballistics information persistence |
US7404268B1 (en) * | 2004-12-09 | 2008-07-29 | Bae Systems Information And Electronic Systems Integration Inc. | Precision targeting system for firearms |
US7210262B2 (en) * | 2004-12-23 | 2007-05-01 | Raytheon Company | Method and apparatus for safe operation of an electronic firearm sight depending upon detected ambient illumination |
DE102005013117A1 (en) * | 2005-03-18 | 2006-10-05 | Rudolf Koch | Rifle with a aiming device |
US20070097351A1 (en) * | 2005-11-01 | 2007-05-03 | Leupold & Stevens, Inc. | Rotary menu display and targeting reticles for laser rangefinders and the like |
US8464451B2 (en) * | 2006-05-23 | 2013-06-18 | Michael William McRae | Firearm system for data acquisition and control |
-
2007
- 2007-04-03 US US11/696,050 patent/US20080022575A1/en not_active Abandoned
- 2007-05-08 EP EP07107673A patent/EP1860395A1/en not_active Withdrawn
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4494198A (en) * | 1981-03-12 | 1985-01-15 | Barr & Stroud Limited | Gun fire control systems |
US6252706B1 (en) * | 1997-03-12 | 2001-06-26 | Gabriel Guary | Telescopic sight for individual weapon with automatic aiming and adjustment |
US6247259B1 (en) * | 1997-10-09 | 2001-06-19 | The State Of Israel, Atomic Energy Commission, Soreq Nuclear Research Center | Method and apparatus for fire control |
WO2000025082A1 (en) * | 1998-10-23 | 2000-05-04 | Precision Remotes, Inc. | Rapid aiming telepresent system |
WO2005047805A2 (en) * | 2003-11-12 | 2005-05-26 | Horus Vision, Llc | Apparatus and method for calculating aiming point information |
US20050268521A1 (en) * | 2004-06-07 | 2005-12-08 | Raytheon Company | Electronic sight for firearm, and method of operating same |
Also Published As
Publication number | Publication date |
---|---|
US20080022575A1 (en) | 2008-01-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1860395A1 (en) | Spotter scope | |
TWI278596B (en) | Electronic firearm sight, and method of operating same | |
JP4874248B2 (en) | Electronic sight for small firearms and operation method thereof | |
KR101252929B1 (en) | Method and apparatus for safe operation of an electronic firearm sight depending upon the detection of a selected color | |
KR101252871B1 (en) | Method and apparatus for safe operation of an electronic firearm sight | |
US8908045B2 (en) | Camera device to capture and generate target lead and shooting technique data and images | |
US7926219B2 (en) | Digital scope with horizontally compressed sidefields | |
EP3034987A1 (en) | System for identifying a position of impact of a weapon shot on a target | |
US20110315767A1 (en) | Automatically adjustable gun sight | |
EA031066B1 (en) | Firearm aiming system (embodiments) and method of operating the firearm | |
WO2012068423A2 (en) | Firearm sight having uhd video camera | |
EP2111612A1 (en) | Image orientation correction method and system | |
US10480903B2 (en) | Rifle scope and method of providing embedded training | |
US20110030545A1 (en) | Weapons control systems | |
US20180202775A1 (en) | Shooting Game for Multiple Players with Dynamic Shot Position Recognition and Remote Sensors | |
JP2016166731A (en) | Shooting system, gun, and data processing device | |
CN214950858U (en) | Image gun and image shooting system with same | |
EP2746716A1 (en) | Optical device including a mode for grouping shots for use with precision guided firearms | |
EP1350132A1 (en) | A device for viewing objects at a distance from a user of the device | |
KR101912754B1 (en) | Shooting and display system for shooting target | |
KR101402758B1 (en) | System and method for shooting game | |
KR101977234B1 (en) | Assembled shooting simulation system using of fish-eye lens camera | |
KR20120055382A (en) | Apparatus for controlling remote firing and method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC MT NL PL PT RO SE SI SK TR |
|
AX | Request for extension of the european patent |
Extension state: AL BA HR MK YU |
|
17P | Request for examination filed |
Effective date: 20080527 |
|
AKX | Designation fees paid |
Designated state(s): DE FR GB |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
GRAC | Information related to communication of intention to grant a patent modified |
Free format text: ORIGINAL CODE: EPIDOSCIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20090523 |