US20170195569A1 - Wide-area aerial camera systems - Google Patents
Wide-area aerial camera systems Download PDFInfo
- Publication number
- US20170195569A1 US20170195569A1 US15/462,212 US201715462212A US2017195569A1 US 20170195569 A1 US20170195569 A1 US 20170195569A1 US 201715462212 A US201715462212 A US 201715462212A US 2017195569 A1 US2017195569 A1 US 2017195569A1
- Authority
- US
- United States
- Prior art keywords
- camera
- detail
- view
- field
- shows
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000007246 mechanism Effects 0.000 claims abstract description 26
- 230000033001 locomotion Effects 0.000 claims description 56
- 238000000034 method Methods 0.000 claims description 22
- 238000003384 imaging method Methods 0.000 claims description 17
- 230000000694 effects Effects 0.000 claims description 11
- 230000003287 optical effect Effects 0.000 abstract description 25
- 238000013461 design Methods 0.000 description 19
- 230000006870 function Effects 0.000 description 19
- 238000010586 diagram Methods 0.000 description 13
- 238000009987 spinning Methods 0.000 description 13
- 239000013598 vector Substances 0.000 description 9
- 230000008569 process Effects 0.000 description 8
- 238000010304 firing Methods 0.000 description 7
- 238000005070 sampling Methods 0.000 description 7
- 238000003860 storage Methods 0.000 description 7
- 238000013459 approach Methods 0.000 description 6
- 238000004364 calculation method Methods 0.000 description 6
- 230000009897 systematic effect Effects 0.000 description 5
- 230000006872 improvement Effects 0.000 description 4
- 230000010355 oscillation Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 230000006835 compression Effects 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 238000012937 correction Methods 0.000 description 3
- 238000009826 distribution Methods 0.000 description 3
- 238000009434 installation Methods 0.000 description 3
- 239000004065 semiconductor Substances 0.000 description 3
- 230000006641 stabilisation Effects 0.000 description 3
- 241001466538 Gymnogyps Species 0.000 description 2
- WYTGDNHDOZPMIW-RCBQFDQVSA-N alstonine Natural products C1=CC2=C3C=CC=CC3=NC2=C2N1C[C@H]1[C@H](C)OC=C(C(=O)OC)[C@H]1C2 WYTGDNHDOZPMIW-RCBQFDQVSA-N 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 2
- 238000001444 catalytic combustion detection Methods 0.000 description 2
- 230000003111 delayed effect Effects 0.000 description 2
- 208000007345 glycogen storage disease Diseases 0.000 description 2
- 230000010354 integration Effects 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000005096 rolling process Methods 0.000 description 2
- 230000003595 spectral effect Effects 0.000 description 2
- 229910000831 Steel Inorganic materials 0.000 description 1
- XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 description 1
- 229910052782 aluminium Inorganic materials 0.000 description 1
- 239000004411 aluminium Substances 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000005452 bending Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000000701 chemical imaging Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000011143 downstream manufacturing Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000013011 mating Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 239000005304 optical glass Substances 0.000 description 1
- 238000000275 quality assurance Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 239000010959 steel Substances 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/58—Means for changing the camera field of view without moving the camera body, e.g. nutating or panning of optics or image sensors
-
- H04N5/23238—
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B64—AIRCRAFT; AVIATION; COSMONAUTICS
- B64D—EQUIPMENT FOR FITTING IN OR TO AIRCRAFT; FLIGHT SUITS; PARACHUTES; ARRANGEMENT OR MOUNTING OF POWER PLANTS OR PROPULSION TRANSMISSIONS IN AIRCRAFT
- B64D47/00—Equipment not otherwise provided for
- B64D47/08—Arrangements of cameras
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C11/00—Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
- G01C11/02—Picture taking arrangements specially adapted for photogrammetry or photographic surveying, e.g. controlling overlapping of pictures
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B26/00—Optical devices or arrangements for the control of light using movable or deformable optical elements
- G02B26/08—Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B26/00—Optical devices or arrangements for the control of light using movable or deformable optical elements
- G02B26/08—Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
- G02B26/0816—Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light by means of one or more reflecting elements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/64—Imaging systems using optical elements for stabilisation of the lateral and angular position of the image
- G02B27/644—Imaging systems using optical elements for stabilisation of the lateral and angular position of the image compensating for large deviations, e.g. maintaining a fixed line of sight while a vehicle on which the system is mounted changes course
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
- G03B15/006—Apparatus mounted on flying objects
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B17/00—Details of cameras or camera bodies; Accessories therefor
- G03B17/56—Accessories
- G03B17/561—Support related camera accessories
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B35/00—Stereoscopic photography
- G03B35/02—Stereoscopic photography by sequential recording
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B37/00—Panoramic or wide-screen photography; Photographing extended surfaces, e.g. for surveying; Photographing internal surfaces, e.g. of pipe
- G03B37/02—Panoramic or wide-screen photography; Photographing extended surfaces, e.g. for surveying; Photographing internal surfaces, e.g. of pipe with scanning movement of lens or cameras
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B37/00—Panoramic or wide-screen photography; Photographing extended surfaces, e.g. for surveying; Photographing internal surfaces, e.g. of pipe
- G03B37/04—Panoramic or wide-screen photography; Photographing extended surfaces, e.g. for surveying; Photographing internal surfaces, e.g. of pipe with cameras or projectors providing touching or overlapping fields of view
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/243—Image signal generators using stereoscopic image cameras using three or more 2D image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/681—Motion detection
- H04N23/6812—Motion detection based on additional sensors, e.g. acceleration sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/682—Vibration or motion blur correction
- H04N23/685—Vibration or motion blur correction performed by mechanical compensation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/682—Vibration or motion blur correction
- H04N23/685—Vibration or motion blur correction performed by mechanical compensation
- H04N23/687—Vibration or motion blur correction performed by mechanical compensation by shifting the lens or sensor position
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/69—Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/40—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
- H04N25/41—Extracting pixel data from a plurality of image sensors simultaneously picking up an image, e.g. for increasing the field of view by combining the outputs of a plurality of sensors
-
- H04N5/2328—
-
- H04N5/23296—
-
- H04N5/247—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/57—Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/71—Charge-coupled device [CCD] sensors; Charge-transfer registers specially adapted for CCD sensors
- H04N25/711—Time delay and integration [TDI] registers; TDI shift registers
Definitions
- 14/687,544 is a continuation-in-part of application Ser. No. 14/478,380, filed Sep. 5, 2014, the entire contents of which are incorporated herein by reference.
- application Ser. No. 14/478,380 is a continuation-in-part of application Ser. No. 14/310,523, filed Jun. 20, 2014, the entire contents of which are incorporated herein by reference.
- the present invention relates to efficient aerial camera systems and efficient methods for creating orthomosaics from aerial photos.
- orthomosaics Accurately georeferenced mosaics of orthophotos, referred to as orthomosaics, are becoming popular alternatives to traditional pictorial maps because they can be created automatically from aerial photos, and because they show actual useful detail on the ground.
- orthomosaic requires the systematic capture of overlapping aerial photos of the area of interest, both to ensure complete coverage of the area of interest, and to ensure that there is sufficient redundancy in the imagery to allow accurate bundle adjustment, orthorectification and alignment of the photos.
- Bundle adjustment is the process by which redundant estimates of ground points and camera poses are refined. Modern bundle adjustment is described in detail in “Bundle Adjustment—A Modern Synthesis” (Triggs et al.) in Vision Algorithms: Theory and Practice (Lecture Notes in Computer Science, Volume 1883, Springer 2000).
- Bundle adjustment may operate on the positions of manually-identified ground points, or, increasingly, on the positions of automatically-identified ground features which are automatically matched between overlapping photos.
- Overlapping aerial photos are typically captured by navigating a survey aircraft in a serpentine pattern over the area of interest.
- the survey aircraft carries an aerial camera system, and the serpentine flight pattern ensures that the photos captured by the camera system overlap both along flight lines within the flight pattern and between adjacent flight lines.
- the aerial camera system carried by a survey aircraft sees a larger view of the ground when the aircraft is operated at a higher altitude. This allows the spacing of flight lines to be increased and hence survey efficiency to be improved. However, to maintain a particular ground sampling distance, i.e. the spacing of camera pixels on the ground, larger camera image sensors or more cameras may need to used.
- the present invention provides a system for capturing aerial images, the system comprising four steerable detail camera modules, each steerable detail camera module comprising a detail camera and a beam-steering mechanism in an optical path of the detail camera module whereby the pointing direction of the detail camera is time-multiplexed to provide a wider effective field of view, the beam-steering mechanism comprising a first steerable mirror.
- the first steerable mirror may be adapted to tilt with respect to an optical axis of the detail camera module to effect beam steering.
- the first steerable mirror may be tilted with respect to an optical axis of the detail camera module, and the first steerable mirror may be adapted to spin about the optical axis to effect beam steering.
- the steering axes of the steerable camera modules may be angularly distributed approximately 90 degrees apart, with at least one steering axis angled at approximately 45 degrees to the direction of flight.
- the steering axes of at least two of the steerable detail camera modules may be substantially parallel.
- each steerable detail camera module may be asymmetric about a vertical pointing direction.
- the wider effective fields of view provided by at least two of the steerable detail camera modules may together form a still wider effective field of view that is symmetric about a vertical pointing direction.
- the system may be attachable, above a camera hole, to at least one of a floor of an aircraft and a floor of a pod carried by an aircraft, thereby providing the at least one detail camera with a view of the ground below the aircraft through the camera hole.
- the system may comprise at least one overview camera, the focal length of the overview camera shorter than the focal length of the detail camera.
- the detail camera may have a lens selected from the group comprising: a dioptric lens, a catoptric lens, and a catadioptric lens.
- the first steerable mirror may be stationary during an exposure period of the detail camera.
- the first steerable mirror may have a non-zero angular velocity during an exposure period of the detail camera
- the steerable detail camera module may comprise a second steerable mirror in the optical path configured to correct the effect, on the detail camera, of the angular velocity of the first steerable mirror during the exposure period, thereby to ensure the detail camera points in a consistent direction during the exposure period.
- the angular velocity of the second steerable mirror may be sinusoidal, and the average angular velocity of the second steerable mirror may be equal to the average angular velocity of the first steerable mirror during the exposure period.
- the angular velocity of the second steerable mirror may be constant during the exposure period.
- the system may further comprise at least one angular motion compensation (AMC) unit, the at least one AMC unit configured to correct the effect, on the detail camera, of angular motion of the aircraft, thereby to ensure the detail camera points in a consistent direction over time.
- AMC angular motion compensation
- AMC may be provided via a steerable mirror in the optical path.
- the system may comprise at least one forward motion compensation (FMC) mechanism, the at least one FMC mechanism configured to correct the effect, on the detail camera, of forward motion of the aircraft, thereby to reduce motion blur in photos captured by the detail camera.
- FMC forward motion compensation
- FMC may be provided via a steerable mirror in the optical path.
- FMC may be provided via time delayed integration of adjacent lines of pixels in an image sensor of the detail camera.
- FMC may be provided by moving an image sensor of the detail camera.
- FIG. 1 shows a front view of a V5-300 HyperCamera unit, i.e. facing forwards towards the front of the aircraft.
- FIG. 2 shows a back view of the camera unit.
- FIG. 3 shows a top view of the camera unit.
- FIG. 4 shows a bottom view of the camera unit.
- FIG. 5 shows an exploded view of the camera unit.
- FIG. 6 shows an exploded view of the cameras and the central support of the camera unit, with the field of view of each camera.
- FIG. 7 shows the camera unit from below, with its combined fields of view passing through the aperture of an aircraft camera hole.
- FIG. 8 shows the overview field of view and the five overlapping detail fields of view of the camera unit.
- FIG. 9 shows the adapter plate of the HyperCamera mounted on seat tracks on the floor of an aircraft.
- FIG. 10 shows an exploded view of the camera unit, the adapter plate, and seat tracks on the floor of an aircraft.
- FIG. 11 shows an exploded view of the cameras and the central support of a wider-angle version of the camera unit.
- FIG. 12 shows the overview field of view and the five overlapping detail fields of view of the V5-180 HyperCamera unit.
- FIG. 13 shows a plan view of the HyperCamera installed in a Cessna 208 aircraft.
- FIG. 14 shows a detailed plan view of the V5-300 HyperCamera installed in a Cessna 208 aircraft.
- FIG. 15 shows a front elevation of a Cessna 208 aircraft carrying a V5-300 HyperCamera, and the resultant overview and aggregate detail fields of view.
- FIG. 16 shows a side elevation of a Cessna 208 aircraft carrying a V5-300 HyperCamera, and the resultant overview and aggregate detail fields of view.
- FIG. 17 shows the overlapping fields of view of three successive shots.
- FIG. 18 shows the overlapping fields of view of shots in adjacent flight lines.
- FIG. 19 shows the overlapping aggregate detail fields of view of a succession of shots along three adjacent flight lines.
- FIG. 20 shows a block diagram of a power and control system for the HyperCamera.
- FIG. 21 shows a photogrammetric process flow for efficiently creating an orthomosaic from multi-resolution HyperCamera photos.
- FIG. 22 shows a tabulation of ground sampling distance (GSD) as a function of altitude and camera focal length between 100 mm and 600 mm.
- GSD ground sampling distance
- FIG. 23 shows a tabulation of ground sampling distance (GSD) as a function of altitude and camera focal length between 700 mm and 1200 mm.
- GSD ground sampling distance
- FIGS. 24A, 24B and 24C show the overview field of view and the overlapping detail fields of view of the V5-300, V10-600 and V15-900 HyperCameras respectively, at an altitude of 12,000 feet.
- FIGS. 25A, 25B and 25C show the overlapping fields of view of three successive shots of the V5-300, V10-600 and V15-900 HyperCameras respectively, at an altitude of 12,000 feet.
- FIGS. 26A, 26B and 26C show the overlapping fields of view of shots in adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at an altitude of 12,000 feet.
- FIG. 27 shows a front elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera, consisting of two V5-600 HyperCamera units, and the resultant overview and aggregate detail fields of view.
- FIG. 28 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIG. 29 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIG. 30 shows a front elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units, and the resultant overview and aggregate detail fields of view.
- the dashed line shows the division between the fields of view of the camera units.
- FIG. 31 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIG. 32 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera, comprising three V5-900 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIGS. 33A, 33B and 33C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 24,000 feet.
- the dashed grid in the figures has a 1 km spacing.
- FIGS. 34A, 34B and 34C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 36,000 feet.
- the dashed grid in the figures has a 1 km spacing.
- FIG. 35A and FIG. 35B show front and side elevations respectively of a Cessna 208 aircraft carrying five HyperCameras installed in the aircraft cabin: a vertical V10-600, a right oblique R10-600, a left oblique L10-600, a front oblique F10-600, and a back oblique B10-600, and the resultant overview and aggregate detail fields of view of each.
- FIG. 36 shows a side elevations of a Cessna 208 aircraft carrying five HyperCameras installed in a standard Cessna 208 belly-mounted cargo pod: a vertical V10-600, a right oblique R10-600, a left oblique L10-600, a front oblique F10-600, and a back oblique B10-600, and the resultant overview and aggregate detail fields of view of each.
- FIG. 37 shows a bottom view of the camera unit with an alternative arrangement of the cameras.
- FIG. 38 shows a camera unit implemented as an array of cameras and their respective optical axes.
- FIG. 39 shows a camera unit implemented as a single camera time-multiplexed via a steerable mirror to produce multiple optical paths.
- FIG. 40 shows an alternative configuration of a camera unit implemented as a single camera time-multiplexed via a steerable mirror to produce multiple optical paths.
- FIG. 41 shows a camera unit implemented as a single camera time-multiplexed via a steerable mirror to produce multiple optical paths, with a steerable mirror for angular motion compensation (AMC).
- AMC angular motion compensation
- FIG. 42 shows a diagram and equation relating to the calculation of the field of view and resolution of a camera.
- FIG. 43 shows a diagram and equations relating to the calculation of the field of view of an array of cameras.
- FIG. 44 shows a diagram and equations relating to the calculation of the resolution of tilted cameras.
- FIG. 45 shows a steerable camera module.
- FIG. 46A shows the steerable camera module with its mirror positioned for imaging at a vertical viewing angle.
- FIG. 46B shows the steerable camera module with its mirror positioned for imaging at an oblique viewing angle.
- FIG. 47 shows a WS-300 HyperCamera unit comprising two half-field steerable detail camera modules and two fixed overview cameras.
- FIG. 48A shows the elevation and plan of the wide-angle field of view of a WS-300 camera unit operating at 16,000 feet.
- FIG. 48B shows the elevation and plan of the wide-angle field of view of a WS-600 camera unit operating at 28,000 feet.
- FIG. 49A shows a schematic plan view of a full-field steerable detail camera module.
- FIG. 49B shows a schematic plan view of a full-field steerable detail camera module comprising a pair of half-field steerable detail camera modules.
- FIG. 49C shows a schematic plan view of a full-field steerable overview camera module.
- FIG. 49D shows a schematic plan view of a camera unit comprising a pair of half-field steerable detail camera modules and two overview cameras.
- FIG. 49E shows a schematic plan view of a camera unit comprising a full-field steerable detail camera module and two overview cameras.
- FIG. 50A shows a schematic plan view of a camera unit comprising a full-field steerable detail camera module and a full-field steerable overview camera module.
- FIG. 50B shows a plan view of a set of vertical and four-way oblique overview and detail fields of view.
- FIG. 51A shows an elevation of the field of view of a wide-angle camera unit, comprising vertical, oblique and interstitial fields of view.
- FIG. 51B shows the elevation and plan of the continuous wide-angle field of view of a WS-600 camera unit operating at 28,000 feet.
- FIG. 52A shows the overlapping lateral fields of view of a wide-angle camera unit from three adjacent flight lines.
- FIG. 52B shows the overlapping lateral fields of view of a wide-angle camera unit from three adjacent flight lines, showing individual camera fields.
- FIG. 53A shows a plan view of the fields of view of a WS-600 camera unit with both overview and detail cameras operating at 28,000 feet.
- FIG. 53B shows a plan view of the field of view of a WS-600 camera unit with just detail cameras operating at 28,000 feet.
- FIG. 53C shows a camera unit comprising two full-field steerable detail camera modules.
- FIG. 53D shows a plan view of the fields of view of a double-field WS-600 camera unit operating at 28,000 feet.
- FIG. 54 shows a WS-300 HyperCamera unit comprising four half-field steerable detail camera modules arranged in an X configuration and two fixed overview cameras.
- FIG. 55A shows a plan view of the fields of view of an X-configuration WS-300 camera unit operating at 16,000 feet.
- FIG. 55B shows a plan view of the fields of view of an X-configuration WS-600 camera unit operating at 28,000 feet.
- FIG. 56A shows a camera unit comprising two pairs of half-field steerable detail camera modules arranged in an X configuration.
- FIG. 56B shows a camera unit comprising two full-field steerable detail camera modules arranged orthogonally.
- FIG. 57A shows a plan view of the fields of view of an X+1-configuration WS-600 camera unit operating at 28,000 feet.
- FIG. 57B shows a camera unit comprising two pairs of half-field steerable detail camera modules arranged in an X configuration, as well as a full-field steerable detail camera module.
- FIG. 58A shows a plan view of three successive shots of an X-configuration WS-600 camera unit operating at 28,000 feet.
- FIG. 58B shows a plan view of shots from two adjacent flight lines of an X-configuration WS-600 camera unit operating at 28,000 feet.
- FIG. 59 shows a diagram and equations relating to the calculation of the longitudinal field of view and shot rate of an aerial camera.
- FIG. 60 shows a diagram and equations relating to the calculation of the lateral field of view of an aerial camera system and area capture rate.
- FIG. 61 shows a diagram and equations relating to the effect of field of view rotation on aerial camera system shot, frame and area capture rates.
- FIG. 62 shows an electronic and mechanical block diagram of a camera unit with steerable camera modules.
- FIG. 63 shows an electronic and mechanical block diagram of a steerable camera module.
- FIG. 64 shows a tabulation, for a 36Mpixel X-configuration steerable camera unit and a 7 cm GSD, of shot rate, frame rate and area capture rate as a function of altitude, aircraft speed and usable field of view angle.
- FIG. 65 shows a tabulation, for a 25Mpixel X-configuration steerable camera unit and a 5 cm GSD, of shot rate, frame rate and area capture rate as a function of altitude, aircraft speed and usable field of view angle.
- FIG. 66 shows a tabulation, for a 12Mpixel X-configuration steerable camera unit and a 5 cm GSD, of shot rate, frame rate and area capture rate as a function of altitude, aircraft speed and usable field of view angle.
- FIG. 67A shows a plan view of the fields of view of a three-field WS-600 camera unit operating at 28,000 feet.
- FIG. 67B shows a plan view of the fields of view of a four-field WS-600 camera unit operating at 28,000 feet.
- FIG. 68A shows a plan view of the fields of view of a five-field WS-600 camera unit operating at 28,000 feet.
- FIG. 68B shows a plan view of the staggered field of view of a single-field WS-600 camera unit operating at 28,000 feet.
- FIG. 69 shows a plan view of systematic motion vectors to be corrected.
- FIG. 70 shows a graph of angular position as a function of time for a multiplexing mirror and a motion-compensation mirror when the latter is used to correct the motion of the former.
- FIG. 71 shows a graph of the linear motion of a multiplexing mirror modulated by a sinusoid, periodically slowing the mirror.
- FIG. 72A shows an alternative steerable camera module utilising a spinning primary mirriro, with its mirror positioned for imaging at a vertical viewing angle.
- FIG. 72B shows the steerable camera module with its mirror positioned for imaging at an oblique viewing angle.
- FIG. 73A shows a plan view of the steerable camera module with its mirror positioned for imaging at an oblique viewing angle.
- FIG. 73B shows a back elevation of the steerable camera module with its mirror positioned for imaging at an oblique viewing angle.
- FIG. 74 shows a plan view of a WS-300 HyperCamera unit comprising two full-field steerable detail camera modules and one fixed overview camera.
- FIG. 75A shows a plan view of the fields of view of an X-configuration WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet.
- FIG. 75B shows a plan view of three successive shots of the X-configuration WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet.
- FIG. 76A shows a double-field steerable camera module.
- FIG. 76B shows detail of the mirror arrangement from FIG. 76A .
- FIG. 77 shows a plan view of the WS-300 HyperCamera unit of FIG. 76A comprising a double-field steerable detail camera module and one fixed overview camera.
- FIG. 78A shows a plan view of the fields of view of a four-field WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet.
- FIG. 78B shows a plan view of the fields of view of a five-field WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet.
- FIG. 79 shows a plan view of a WS-300 HyperCamera unit comprising four half-field spinning-mirror steerable detail camera modules, arranged in an X configuration, and one fixed overview camera.
- FIG. 80 shows a plan view of the WS-300 HyperCamera unit of FIG. 79 , with the detail fields of view shown at an extreme angle.
- FIG. 81 shows a perspective view of the WS-300 HyperCamera unit of FIG. 79 , with the detail fields of view shown at both a vertical angle and at an extreme angle.
- HyperCameraTM is a range of aerial camera systems suitable for deployment in a wide range of aircraft, large and small.
- the camera systems are modular, and designed to be installed above one or more standard camera holes where appropriate, as are typically provided through the floor of a survey aircraft or airborne pod.
- Each HyperCamera model is defined by its pointing direction (P), the number of cameras (N) in its array, and the focal length (f) of the cameras, and is identified by the designator PN-f.
- P pointing direction
- N number of cameras
- f focal length
- V5-300 model a 300 mm five-camera vertical HyperCamera
- Wide-angle steerable HyperCamera models are identified by the designator WS-f.
- a 300 mm wide-angle steerable HyperCamera is referred to as a WS-300 model.
- the level of detail captured by an aerial camera is typically characterized by the ground sampling distance (GSD), i.e. the distance between adjacent pixel centers when projected onto the ground within the camera's field of view.
- GSD ground sampling distance
- the GSD is determined by the focal length (f, 252 ) of the camera lens, the altitude above ground level (a, 254 ), and the pixel pitch (p, 260 ) of the image sensor, per EQ 1 in FIG. 42 .
- the efficiency of aerial imaging is typically characterized by the area captured per unit time (e.g. square km per hour). This is proportional to the speed of the aircraft and the width of the field of view (FOV) of the aerial camera system, referred to as the swath width.
- FOV field of view
- the swath width (w, 258 ) of a single camera is determined by the focal length (f, 252 ) of the camera lens, the altitude above ground level (a, 254 ), and the lateral size (s, 256 ) of the image sensor, per EQ 2 in FIG. 42 . Doubling the altitude doubles the swath width.
- the lateral field of view (beta, 250 ) of a single camera is determined by the focal length (f, 252 ) of the camera lens and the lateral size (s, 256 ) of the image sensor, per EQ 4 in FIG. 42 . Doubling the focal length approximately halves the field of view.
- the lateral field of view (gamma, 270 ) of an array of cameras is determined by the field of view of each camera (beta, 250 ), the number of cameras (N), and their angular overlap (omega, 272 ), per EQ 7 in FIG. 43 . Doubling the number of cameras approximately doubles the field of view.
- the swath width (w(N), 278 ) of the array of cameras is given by EQ 8 in FIG. 43 .
- the practical field of view of an aerial camera system is limited by the acceptable level of obliqueness in the aerial imagery—e.g. how much buildings are allowed to lean when captured at the edge of the field of view.
- the practical field of view is usually limited to 50 degrees or less.
- the HyperCamera range includes models suited to a variety of operating altitudes, and supporting a range of GSDs and capture efficiencies.
- a HyperCamera system may be installed in any aircraft suitable for aerial imaging.
- this includes (but is not limited to) piston aircraft such as a Cessna 210, turboprop aircraft such as a Cessna 208, and turbofan (jet) aircraft such as a Cessna Citation, allowing aerial imaging from low altitudes to altitudes in excess of 40,000 feet, at speeds ranging from less than 100 knots to 400 knots.
- each camera hole may be open or contain an optical glass window as appropriate.
- Each camera hole may be optionally protected by a door which can be closed when the HyperCamera is not in operation.
- a V5-300 HyperCamera comprises a camera unit 100 incorporating five detail cameras 110 and a relatively wider-angle overview camera 112 .
- Each detail camera 110 has a detail lens 114
- the overview camera 112 has a overview lens 116 .
- the overview lens 116 is characterised by having a significantly wider angle than the detail lens 114 . While it may be a true wide-angle lens, it may also be a normal lens or even a telephoto lens so long as it is significantly wider than the detail lens 114 . Likewise, while the detail lens 114 may be a true telephoto lens, it may also be a normal lens or even a wide-angle lens so long as it is significantly narrower than the overview lens 116 .
- the cameras 110 and 112 are preferably commercial off-the-shelf (COTS) digital SLR (DSLR) cameras.
- COTS commercial off-the-shelf
- DSLR digital SLR
- the use of COTS cameras allows the system to be readily adapted to the latest and best available cameras.
- dedicated camera designs may also be utilised.
- High-resolution COTS cameras are available with typical pixel counts ranging from 24Mpixels to 36Mpixels, from vendors such as Nikon and Canon.
- the 36Mpixel Nikon D800 (and D810) DSLR camera is a particularly good choice for the present system.
- DSLR cameras offer a wide range of high-quality lenses, allowing the system to be readily configured to operate at different altitudes and resolutions.
- the system is readily adapted to a mixture of cameras.
- a relatively more expensive camera with a higher pixel count may be employed as the overview camera.
- 70Mpixel DSLR cameras are expected to be available in the near future, and a 70Mpixel camera would be a good choice for the overview camera.
- the detail lenses 114 of the detail cameras 110 all have the same focal length, and the detail cameras 110 all have the same pixel size.
- the camera unit 100 embodies two distinct camera resolutions—overview and detail. This is readily extended to multiple resolutions greater than two through the use detail lenses 114 with different focal lengths, and/or the use of detail cameras 110 with different pixel sizes.
- the camera unit 100 may also incorporate multiple overview cameras with different resolutions.
- Each detail lens 114 and overview lens 116 may be a fixed-focus lens focused at infinity or a variable-focus lens. In the latter case the corresponding camera 110 and/or 112 incorporates an auto-focus mechanism.
- Each detail camera 110 is bolted to a camera mount 140 , which in turn is bolted to a center support 122 .
- Each detail camera lens 114 is further secured by a clamp 144 which is bolted to the detail camera mount 140 .
- the overview camera is bolted to a camera mount 142 , which in turn is bolted to the center support 122 .
- the overview camera lens 116 is further secured by a clamp 146 which is bolted to the overview camera mount 142 .
- the camera mounts 140 and 142 isolate much of the structure of camera unit 100 from the specifics of individual camera models and lens sizes.
- the center support 122 is attached to a pair of side supports 124 a and 124 b , and each side support 124 is in turn attached to a rear support 126 and a front support 128 to form a rigid frame 120 .
- Each side support 124 is attached to mount point block 130 via a set of four bolts, and the mount point block 130 is in turn attached to the rear support 126 or front support 128 , as appropriate, via a further set of four bolts.
- the mount point blocks 130 thereby provide the attachment mechanism between the side supports 124 and the rear and front supports 126 and 128 .
- Each of the side supports 124 and the rear and front supports 126 and 128 has a C-shaped cross-sectional profile to minimise weight while maximising rigidity, while the center support 122 is pocketed to minimise weight while maximising rigidity.
- Each mount point block 130 is solid, and serves the additional purpose of providing a point of attachment between the camera unit 100 and a survey aircraft, as described below.
- the rear support 126 and the front support 128 hold three power and control distribution boxes 150 .
- Each box 150 distributes power and control signals to a pair of cameras.
- the power and control cabling between the boxes 150 and the cameras 110 and 112 is omitted in the figures.
- each detail camera 110 has a lens 114 with a focal length of 300 mm suitable for high-resolution imaging at relatively high altitudes.
- a 300 mm lens allows a ground sampling distance (GSD) of 30 cm at 60,000 feet, 20 cm at 40,000 feet, 10 cm at 20,000 feet, 8 cm at 16,000 feet, 6 cm at 12,000 feet, 4 cm at 8,000 feet, 2 cm at 4,000 feet, 1 cm at 2,000 feet, etc.
- the overview camera 112 ideally has a lens 116 with a focal length that is between 4 and 8 times shorter than the focal length of the detail lens 114 , as discussed further below. I.e. for a 300 mm detail lens 114 , suitable focal lengths for the overview lens 116 range from about 40 mm to 75 mm. For illustrative purposes the present system utilises a 50 mm overview lens 116 .
- FIG. 6 shows the 6.90-degree lateral field of view 162 of each of the five detail cameras 110 with 300 mm lenses 114 , and the 39.60-degree lateral field of the overview camera 112 with a 50 mm lens 116 .
- the lateral direction is the direction perpendicular to the direction of flight 220
- the longitudinal direction is the direction parallel to the direction of flight 220 .
- the detail cameras are angled 6 degrees apart laterally, i.e. slightly less than their 6.90-degree fields of view 162 , so that the fields of view 162 overlap slightly.
- the five detail cameras 110 have an aggregate field of view with a pixel count of approximately 160Mpixels, i.e. excluding overlap.
- Stock telephoto lenses suitable for use as detail lenses 114 are available in a variety of focal lengths, typically including 85 mm, 105 mm, 135 mm, 180 mm, 200 mm, 300 mm, 400 mm, 500 mm, 600 mm, 700 mm, and 800 mm.
- a 400 mm lens on a Nikon D800 camera allows a GSD of 7.4 cm, a 600 mm lens a GSD of 5.0 cm, and an 800 mm lens a GSD of 3.7 cm.
- Stock normal and wide-angle lenses suitable for use as an overview lens 116 are available in a variety of focal lengths, typically including 10.5 mm, 14 mm, 16 mm, 18 mm, 20 mm, 21 mm, 24 mm, 28 mm, 35 mm, 40 mm, 45 mm, 50 mm, 55 mm, 60 mm and 70 mm.
- the camera unit 100 is readily adapted for different models and sizes of cameras 110 (and 112 ) and lenses 114 (and 116 ) via different camera mounts 140 (and 142 ) and clamps 144 (and 146 ). For extremely long lenses a taller center support 122 can be used.
- the detail cameras are angled inwards so that their fields of view 162 cross over directly below the camera unit 100 , creating a waist of minimum diameter where the fields of view pass through the camera hole 212 .
- FIG. 8 shows the projection of the three-dimensional fields of view 160 and 170 of the detail cameras 110 and overview camera 112 onto a ground plane. It shows how the detail field of views 160 overlap in a direction perpendicular to the direction of flight 220 .
- FIG. 9 shows an adapter plate 200 that attaches to the seat tracks 214 of an aircraft, a Cessna 208 in this case, via standard seat track fasteners 202 .
- the adapter plate has an aperture 216 which exposes a camera hole 212 through the floor 210 of the aircraft.
- FIG. 10 shows an exploded view of the camera unit 100 , adapter plate 200 , and the aircraft floor 210 .
- the adapter plate 200 is designed to attach the camera unit 100 to a particular aircraft, and isolates the design of the camera unit 100 from aircraft specifics.
- a different adapter plate is designed for each aircraft attachment variation, e.g. due to different seat track spacings, or because the aircraft's camera hole installation includes its own mounting points.
- mount points 132 are bolted to the adapter plate, with each mount point 132 mating with a recess in the base of its corresponding mount point block 134 .
- a mount bolt 143 securely attaches each mount point block 134 to its corresponding mount point 132 , thus attaching the camera unit 100 to the adapter plate 200 .
- the adapter plate 200 allows the camera unit 100 to be easily installed in and subsequently removed from an aircraft via installation and removal of the four mount bolts 143 .
- the adapter plate 200 is itself generally easily installed in and removed from an aircraft, requiring no modification to the aircraft (assuming a suitable camera hole is already installed).
- the installation of external camera pod is generally a much more complicated operation.
- FIG. 11 shows a V5-180 HyperCamera unit 100 that utilises shorter 180 mm lenses 114 for the detail cameras 110 , and a matching 28 mm lens 116 for the overview camera 112 .
- a 180 mm lens allows a ground sampling distance (GSD) of 9.9 cm at 12,000 feet, 8.3 cm at 10,000 feet, 6.6 cm at 8,000 feet, 5 cm at 6,000 feet, 3.3 cm at 4,000 feet, and so on.
- GSD ground sampling distance
- FIG. 11 shows the 11.40-degree lateral field of view 162 of each of the five detail cameras 110 with 180 mm lenses 114 , and the 65.50-degree lateral field of the overview camera 112 with a 28 mm lens 116 .
- the detail cameras are angled 10.50 degrees apart laterally, i.e. slightly less than their 11.40-degree fields of view 162 , so that the fields of view 162 overlap slightly.
- FIG. 12 shows the projection of the three-dimensional fields of view 160 and 170 of the detail cameras 110 and overview camera 112 of FIG. 10 onto a ground plane. It shows how the detail field of views 160 overlap in a direction perpendicular to the direction of flight 220 , and how the wider fields of view associated with the shorter lenses leads to a lower flying altitude for the same footprint on the ground, i.e. in comparison to FIG. 8 .
- FIG. 13 and FIG. 14 show plan views of a Cessna 208 survey aircraft 230 carrying a camera unit 100 installed centrally over a camera hole.
- the figures also show a camera control unit 310 (CCU) and battery unit 320 used to control and power the camera unit 100 . These are described in more detail below. For clarity, the cabling connecting the CCU 310 , battery unit 320 and camera unit 100 is omitted.
- FIG. 15 shows a front elevation of the Cessna 208 survey aircraft 230 carrying a HyperCamera, and shows the lateral overview field of view 172 of the camera unit 100 , and the aggregate lateral detail field of view 182 of the camera unit 100 .
- the aggregate lateral detail field of view 182 is the aggregate of the five individual overlapping lateral detail fields of view 162 .
- FIG. 16 shows a side elevation of the Cessna 208 survey aircraft 230 carrying a HyperCamera, and shows the longitudinal overview field of view 174 of the camera unit 100 , and the longitudinal detail field of view 164 of the camera unit 100 .
- FIG. 17 shows the overlapping overview fields of view 170 and aggregate detail fields of view 180 of three successive shots in the direction of flight 220 .
- the aggregate detail field of view 180 is the aggregate of the five individual overlapping detail fields of view 160 .
- the aggregate detail fields of view 180 overlap by about 20% longitudinally, while the overview fields of view 170 overlap by about 85% longitudinally.
- FIG. 18 shows the overlapping overview fields of view 170 and aggregate detail fields of view 180 of two shots from adjacent flight lines, i.e. flown in opposite directions 220 .
- the aggregate detail fields of view 180 overlap by between 20% and 25% laterally, while the overview fields of view 170 overlap by about 40% laterally.
- the size of the lateral overview field of view 172 and the size of the lateral aggregate detail field of view 182 are similar when the ratio of the focal length of the detail camera lens 114 to the focal length of the overview camera lens 116 is about 6, and useful lens combinations can be chosen with focal length ratios between about 4 and 8.
- FIG. 19 shows the overlapping aggregate detail fields of view 180 of a succession of shots along three adjacent flight lines that are part of a typical serpentine flight path 222 , i.e. a subset of the flight lines that would make up a typical large-area survey. For clarity the corresponding overview fields of view 170 are omitted.
- the figure also shows the shot position 224 corresponding to each aggregate detail field of view 180 , i.e. the position of the survey aircraft 230 .
- the HyperCamera Compared with a traditional single-resolution aerial camera system and a comparable aggregate detail pixel count (e.g. 160Mpixel), the HyperCamera is between 2 and 3 times more efficient, as detailed below, with respect to both reduced survey flying time and fewer photos to process.
- the HyperCamera also has a higher efficiency than many aerial camera systems due to its high (detail) pixel count alone.
- the HyperCamera can be used to capture detail photos only, with higher overlap (e.g. 60/40 rather than 20/20), to allow the creation of an orthomosaic with higher spatial accuracy, but at greater capture and processing cost.
- the overview camera 112 can be omitted or used for another purpose.
- FIG. 20 shows a block diagram of a power and control system for the camera unit 100 .
- the detail cameras 110 and overview camera 112 are controlled by a computer 300 via a camera interface 308 , which may consist of a set of digital switches that trigger the remote control interfaces of the cameras.
- the computer 300 uses one or more Global Navigation Satellite System (GNSS) receiver 304 to monitor the position and speed of the survey aircraft 230 in real time.
- GNSS Global Navigation Satellite System
- the GNSS receiver(s) may be compatible with a variety of space-based satellite navigation systems, including the Global Positioning System (GPS), GLONASS, Galileo and BeiDou.
- GPS Global Positioning System
- GLONASS Global Positioning System
- Galileo Galileo
- BeiDou BeiDou.
- the computer 300 provides precisely-timed firing signals to the cameras 110 and 112 via the camera interface 308 , to trigger camera exposure, according to a stored flight plan and the real-time position and speed of the aircraft. If a camera 110 and/or 112 incorporates an auto-focus mechanism then the computer 300 also provides a focus signal to each such camera to trigger auto-focus prior to exposure.
- the computer 300 fires the overview camera 112 and the detail cameras 110 at the same rate.
- the computer 300 may fire the overview camera 112 at a different rate to the detail cameras 110 , i.e. either a higher rate or lower rate, to achieve a different overlap between successive overview photos, i.e. either a higher overlap or a lower overlap, independent of the overlap between successive detail photos.
- the computer 300 may fire the cameras simultaneously, or it may stagger the timing of the firing, e.g. to achieve a different alignment of photos longitudinally, or to reduce peak power consumption.
- the flight plan describes each flight line making up the survey, and the nominal camera firing rate along each flight line required to ensure that the necessary overlap is maintained between successive shots.
- the firing rate is sensitive to the elevation of the terrain below the aircraft, i.e. the higher the terrain the higher the firing rate needs to be. It is adjusted by the computer 300 according to the actual ground speed of the aircraft, which may vary from its nominal speed due to wind and the pilot's operation of the aircraft.
- the computer 300 also uses the flight plan and real-time GNSS position to guide the pilot along each flight line via a pilot display 302 .
- the position data from the GNSS receiver is optionally augmented with orientation information (roll, pitch and yaw) from an inertial measurement unit 306 (IMU). It allows the computer 300 to provide enhanced feedback to the pilot on how closely the pilot is following the flight plan, and allows more accurate processing of the photos. In the absence of the IMU 306 the GNSS receiver connects directly to the computer 300 .
- orientation information roll, pitch and yaw
- IMU inertial measurement unit
- Each camera 110 and 112 stores its shots locally, e.g. in removable flash memory. This eliminates the need for centralised storage in the HyperCamera system, and the need for a high-bandwidth data communication channel between the cameras and the centralised storage.
- the GNSS position of each shot may be delivered to each camera 110 and 112 , to allow the camera to tag each photo with its GNSS position.
- One or more optional angular motion compensation (AMC) units 330 responsive to the orientation reported by the IMU 306 , correct the orientation of the cameras so that they maintain a consistent pointing direction over time, despite the aircraft rolling, pitching or yawing during flight. This ensures that the captured photos can be used to create an orthomosaic without gaps, while allowing the overlap between successive shots and between adjacent flight lines to be minimised.
- AMC angular motion compensation
- the AMC unit 330 may consist of a platform with two or three axes of rotation (i.e. roll and pitch, or roll, pitch and yaw) upon which the HyperCamera unit 100 is mounted.
- Commercially-available AMC or stabilisation platforms 330 include the PAV series from Leica Geosystems, including the PAV100, and the PV-15G from PV Labs, e.g. as described in U.S. Pat. No. 6,263,160 (Lewis), the contents of which are herein incorporated by reference.
- the AMC units 330 may comprise one or more beam-steering mechanisms in the optical path of each camera (or group of cameras), whereby the pointing direction of the cameras is corrected by beam-steering.
- pitch and roll compensation may be provided by beam-steering mechanism(s), while yaw compensation is provided by a separate AMC platform 330 .
- Angular motion compensation becomes increasingly important as the flying altitude is increased and/or the GSD is decreased.
- Motion blur due to the forward motion of the aircraft is equal to the speed of the aircraft multiplied by the exposure time of the camera.
- FMC forward motion compensation
- FMC can be provided in a number of ways, including translating or rotating the optical axis of the camera (by moving the image sensor, or an intermediate mirror, or the camera itself), or by time delayed integration (TDI) of adjacent lines of pixels in the image sensor.
- the computer 300 stores the GNSS position of each shot. This is used during subsequent processing of the photos to produce an accurate orthomosaic.
- the computer 300 also stores the orientation of each shot if the IMU 306 is present, and stores the corrected orientation of each shot if an AMC unit 330 is present, based on correction information from the AMC.
- the cameras 110 and 112 are powered by a battery unit 320 .
- the battery unit 320 provides a voltage higher than the voltage required by all connected components, e.g. between 24V and 28V, and the voltage requirement of each connected component is provided via a DC-DC converter 326 .
- a Nikon D800 camera requires less than 10V.
- Additional DC-DC converters 326 also provide appropriate voltages to power the computer 300 , the pilot display 302 , the GNSS receiver 304 , and the IMU 306 . For clarity these power connections are omitted in FIG. 20 .
- the battery unit 320 contains two 12V (nominal) batteries or a single 24V (nominal) battery. It contains a charging circuit that allows it to be trickle-charged from an aircraft with a suitable auxiliary power source 322 , allowing it to remain charged at all times. It may also be charged on the ground from a ground power unit 324 (GPU).
- GPU ground power unit
- the camera interface 308 and DC-DC converters 326 may be housed in a camera control unit 310 (CCU). This may additionally include a USB interface to allow the computer 300 to control the camera interface.
- CCU camera control unit
- the DC-DC converters 326 that provide power to the cameras 110 and 112 may be located in the CCU 310 or closer to the cameras in the distribution boxes 150 .
- FIG. 21 shows a photogrammetric process flow for efficiently creating an orthomosaic from multi-resolution HyperCamera photos.
- the process operates on detail photos 400 captured by the detail cameras 110 , and overview photos 402 captured by the overview cameras 112 .
- the process consists of four main steps: (1) features are automatically detected in each of the photos 400 and 402 and matched between photos (step 410 ); bundle adjustment is used to iteratively refine initial estimates of the real-world three-dimensional position of each feature, as well as the camera pose (three-dimensional position and orientation) and camera calibration (focal length, radial and tangential distortion) associated with each photo (at step 412 ); each detail photo 400 is orthorectified according to its camera pose and terrain elevation data (at step 414 ); and the orthorectified photos (orthophotos) are blended to form the final orthomosaic 404 (at step 416 ).
- the elevation data (used at step 414 ) may be derived from the three-dimensional feature positions (refined during step 412 ); and/or may be derived from dense multi-view stereo matching between photos (see, for example, U.S. Pat. No. 8,331,615 (Furukawa), and U.S. Pat. No. 8,634,637 (Hirschmueller et al), the contents of both of which are herein incorporated by reference); and/or may be obtained from another source such as a LIDAR survey of the area of interest.
- the elevation data may be in the form of a raster or vector digital elevation model (DEM), or in the form a three-dimensional point cloud, or in the form of a three-dimensional geometric model. It may include surface features such as vegetation and buildings, or may describe the bare earth.
- the accuracy of the orthomosaic 404 derives from the high overlap between lower-resolution overview photos 402 , while detail in the orthomosaic 404 derives from the higher-resolution detail photos 400 .
- a survey may be flown with higher overlap between the detail photos 400 , and the orthomosaic may be created from the detail photos 400 only.
- the orthomosaic is typically stored as an image pyramid, i.e. within which different (binary) zoom levels are pre-computed for fast access at any zoom level.
- Lower zoom levels in the pyramid are generated from higher zoom levels by low-pass filtering and subsampling, thus the entire pyramid may be generated from the detail-resolution orthomosaic.
- lower zoom levels may be generated from an orthomosaic created from the overview photos 402 , in which case the overview photos 402 are also orthorectified and blended as described above for the detail photos 400 .
- An initial estimate of the camera pose of each photo, subsequently refined by the bundle adjustment process (at step 412 ), is derived from the GNSS position of each photo, as well as its IMU-derived orientation, if available.
- Automatically detected ground features may be augmented with manually-identified ground points, each of which may have an accurate surveyed real-world position (and is then referred to as a ground control point).
- V5-300 and V5-180 HyperCameras are two models from the HyperCamera range. Even greater efficiency is achieved by operating a higher altitudes with longer focal lengths, and additional HyperCamera configurations are described below.
- FIG. 22 tabulates the GSD as a function of the flying altitude above ground (in feet) and the lens focal length from 100 mm to 600 mm (in 100 mm increments). It also tabulates the swath width (in km) of 30-degree and 45-degree fields of view as a function of altitude. In addition, it tabulates the number of detail cameras required to cover these fields of view as a function of focal length.
- FIG. 23 further tabulates the GSD as a function of focal length from 700 mm to 1200 mm.
- a 30-degree V10-600 HyperCamera can be realised using two 15-degree V5-600 HyperCamera units in tandem, each unit mounted on a wedge-shaped adapter plate 200 that induces the required 7.5-degree lateral tilt.
- a 45-degree V15-600 HyperCamera can be realised using three 15-degree V5-600 HyperCamera units in tandem, the central unit mounted flat, and each of the two outside units mounted on a wedge-shaped adapter plate 200 that induces the required 15-degree lateral tilt.
- a 30-degree V15-900 HyperCamera can be realised using three 10-degree V5-900 HyperCamera units in tandem, the central unit mounted flat, and each of the two outside units mounted on a wedge-shaped adapter plate 200 that induces the required 10-degree lateral tilt.
- the adapter plate 200 can comprise one or more gross and/or fine tilt adjustment mechanisms to allow the tilt of the camera unit 100 to be adjusted both nominally as well as to compensate for variations in the tilt of the aircraft floor.
- the tilt mechanisms may comprise one or more shims, one or more biasing screws, or similar.
- Tilt adjustment can also be provided via the roll compensation mechanism of the angular motion compensation (AMC) unit 330 , if present, by adjusting the pre-set nominal roll of the camera unit 100 .
- AMC angular motion compensation
- FIGS. 24A, 24B and 24C show the footprints of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 12,000 feet.
- the dashed grid in the figures has a 1 km spacing.
- the figures show the same swath width but with increasing GSD.
- FIGS. 25A, 25B and 25C show three successive overlapping footprints of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 12,000 feet. In order to ensure adequate longitudinal overlap, the shot rate increases with increasing resolution.
- FIGS. 26A, 26B and 26C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 12,000 feet.
- FIG. 27 shows a front elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units, and the resultant overview and aggregate detail fields of view.
- the dashed line shows the division between the fields of view of the camera units.
- FIG. 28 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIG. 29 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIG. 30 shows a front elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units, and the resultant overview and aggregate detail fields of view.
- the dashed line shows the division between the fields of view of the camera units.
- FIG. 31 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIG. 32 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera, comprising three V5-900 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit.
- FIGS. 33A, 33B and 33C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 24,000 feet.
- the dashed grid in the figures has a 1 km spacing.
- FIGS. 34A, 34B and 34C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 36,000 feet.
- the dashed grid in the figures has a 1 km spacing.
- a particular HyperCamera model can be realised using a number of identical smaller camera units 100 , each mounted at the required lateral angle. As shown in the FIGS. 27 to 32 , the individual camera units 100 can be mounted along the longitudinal axis of the aircraft. The firing of each camera unit 100 can be staggered to account for its longitudinal position and the aircraft speed, or the longitudinal offsets between photos from different camera units can be accounted for during downstream processing.
- the HyperCamera design is thus modular, and suited to deployment in conjunction with one or more standard camera holes.
- An oblique HyperCamera i.e. with a pointing direction other than vertical (and typically with pointing direction of 45 degrees), can be realised using one or more vertical camera units 100 mounted at the required oblique angle, e.g. via a tilted adapter plate 200 , or using one or more oblique camera units 100 , each comprising an array of cameras mounted at the required oblique angle within an otherwise vertically-oriented frame.
- Oblique aerial photos have several uses. They can be used to produce an oblique georeferenced orthomosaic to provide a useful complement to a vertical orthomosaic. They can be used along with vertical photos during photogrammetric bundle adjustment (as described above in relation to FIG. 21 ) to achieve higher accuracy, including higher elevation data accuracy. And they can be used to texture an elevation model or a full three-dimensional model of the area of interest, to support three-dimensional visualisation and interaction.
- FIGS. 35A and 35B show front and side elevations respectively of a Cessna 208 aircraft carrying five HyperCameras installed in the aircraft cabin: a vertical V10-600, a right oblique R10-600, a left oblique L10-600, a front oblique F10-600, and a back oblique B10-600, and the resultant overview and aggregate detail fields of view of each.
- the oblique GSD ( 282 ) of a camera tilted at an angle ( 280 ) away from the vertical is given by the EQ 9 in FIG. 44 .
- the vertical GSD ( 262 ) is increased by a factor of the square of the secant of the tilt angle.
- the first secant term is due to the greater tilt-induced viewing distance, while the second secant term is due to the tilt-induced foreshortening of the ground.
- a tilted camera therefore needs to have a correspondingly longer focal length to match the GSD of a vertical camera.
- the oblique GSD ( 286 ) is then given by EQ 10 in FIG. 44 .
- the vertical GSD ( 262 ) is then increased only by a factor of the square root of two, implying that a 45-degree tilted-sensor oblique camera needs to have 1.4 times the focal length of a vertical camera to match the GSD of the vertical camera.
- the smaller GSD on horizontal surfaces is accompanied by a larger GSD on vertical surfaces such as the sides of buildings.
- This in turn may motivate a strategy of capturing oblique photos in multiple passes, with each pass optimised for a particular surface orientation via image sensor tilt.
- tilting the image sensor of any non-nadir camera to be (more) parallel to the ground can be used to reduce perspective foreshortening in photos captured by the camera and thus improve the GSD.
- Each 900 mm oblique HyperCamera has a 33% larger GSD than the 600 mm vertical HyperCamera.
- a 1200 mm oblique HyperCamera would have the same GSD as the 600 mm vertical HyperCamera.
- An 850 mm tilted-sensor oblique HyperCamera would also have the same GSD as the 600 mm vertical HyperCamera.
- FIG. 37 shows a bottom view of the camera unit 100 where the detail cameras are arranged in an alternative “X” pattern to provide more room for longer detail lenses 114 , i.e. to accommodate the larger diameters of longer lenses.
- This arrangement also provides room for a second lower-resolution camera (i.e. with a shorter lens) which can be utilised either for a second overview camera or for a dedicated spectral camera, such as a near-infrared (NIR) camera.
- NIR near-infrared
- Including a second overview camera allows both cameras to have a longer focal length and therefore provide greater resolution and hence photogrammetric accuracy.
- Including a near-infrared camera allows a near-infrared band to be included in the orthomosaic, in turn supporting applications such as identifying vegetation.
- a lighter and physically shorter lens with the same long focal length may be realised using a reflecting design, i.e. using curved mirrors for focusing, either using mirrors alone (catoptric) or in conjunction with corrective refractive elements (catadioptric).
- Astronomical telescopes which are characterised by long focal lengths and large apertures, typically utilise reflecting designs.
- Classic modern designs include the Schmidt-Cassegrain, which combines spherical primary and secondary mirrors with a Schmidt correction plate, and the Ritchey-Chretien, which utilises hyperbolic primary and secondary mirrors.
- Reflecting detail lenses 114 can thus be used to significantly reduce the size and weight of the camera unit 100 , and this is particularly beneficial for longer focal lengths such as 900 mm and longer.
- FIG. 38 shows a camera unit 100 implemented as an array of cameras 600 , arranged in a fan, and the optical axis 602 of each camera.
- Each camera 600 consists of a camera body (e.g. 110 or 112 ) and a camera lens (e.g. 114 or 116 ).
- FIG. 39 shows a camera unit 100 implemented using a steerable camera module 610 comprising a single camera 600 time-multiplexed via a steerable mirror 604 to produce multiple optical paths that sweep out the required angular field.
- a steerable camera module 610 comprising a single camera 600 time-multiplexed via a steerable mirror 604 to produce multiple optical paths that sweep out the required angular field.
- This allows a single camera to be used to implement an array of cameras using time-multiplexing, and thus reduces the number of physical cameras required to implement a HyperCamera, in turn reducing the size, weight and cost of each camera unit 100 .
- any suitable beam-steering mechanism may be used as an alternative to a steerable mirror 604 .
- FIG. 40 shows an alternative configuration of a camera unit 100 implemented using a steerable camera module 610 comprising a single camera 600 time-multiplexed via a steerable mirror 604 to produce multiple optical paths.
- the addition of a fixed mirror 606 allows the camera 600 to be mounted vertically, allowing the camera unit 100 to have a smaller footprint.
- FIG. 41 shows a camera unit 100 implemented using a steerable camera module 610 comprising a single camera 600 time-multiplexed via a steerable mirror 604 to produce multiple optical paths, with an additional steerable mirror 608 providing angular motion compensation (AMC).
- AMC angular motion compensation
- This implements the IMU-driven AMC unit 330 .
- any suitable beam-steering mechanism may be used.
- the steerable mirror 608 may additionally or alternatively be used to implement a forward motion compensation (FMC) mechanism.
- FMC forward motion compensation
- the mirror In FMC mode the mirror is smoothly pitched backwards during exposure to match the forward motion of the aircraft, and is pitched forwards again between exposures.
- the pitch of the mirror 608 can be driven directly, or can be oscillated at an appropriate rate to provide FMC, with camera exposure synchronised with the backwards phase of the oscillation.
- AMC and FMC may also be provided by separate cascaded mirrors.
- the multiplexing steerable mirror 604 may be stopped at each of a set of discrete positions within the required angular field, or may be rotated continuously to sweep out the required angular field. In the latter case the motion of the mirror provides an additional source of systematic motion during camera exposure, like the forward motion of the aircraft, and this may be corrected using any of the FMC mechanisms described above, including via the steerable motion-compensation mirror 608 , or via a separate stabilisation platform 330 .
- Vibration of the aircraft may also be detected via the IMU, or via an additional vibration-sensing device, and vibration compensation (VC) may be provided using any of the AMC or FMC mechanisms described above, including via the steerable motion-compensation mirror 608 , or via a separate stabilisation platform 330 .
- vibration compensation may be provided using any of the AMC or FMC mechanisms described above, including via the steerable motion-compensation mirror 608 , or via a separate stabilisation platform 330 .
- AMC, FMC and VC may also be provided via the multiplexing mirror 604 .
- FIG. 45 shows one realisation of the steerable camera module 610 of FIG. 41 .
- the steerable multiplexing mirror 604 is attached via a hinge to a rod 632 of a linear motor 630 , and the linear motor 630 is attached via a hinge to a fixed mount 634 .
- the mirror 604 is also attached via a hinge to the mount 634 , about which it rotates in response to linear motion of the rod 632 .
- the linear motor 630 may utilise any suitable drive mechanism, including a piezoelectric drive mechanism, e.g. as described in U.S. Pat. No. 6,800,984 (Marth), the contents of which are incorporated herein by reference, or as embodied in the Physik Instrumente (PI) M-272 linear piezomotor, which also incorporates a linear encoder for precise positioning.
- PI Physik Instrumente
- the steerable motion-compensation mirror 608 is attached to a tip-tilt stage 636 which provides rotation in two dimensions.
- the stage 636 can be limited to rotation in one dimension, e.g. to only provide correction for motion of the multiplexing mirror 604 .
- the tip-tilt stage 636 may utilise any suitable motors or actuators, including piezoelectric actuators, e.g. as embodied in the PI S-340, which may also incorporate a strain-gauge sensor for precise closed-loop control.
- the steerable camera module 610 in FIG. 45 allows the field of view 612 of the camera 600 to be steered across an angular range of approximately 50 degrees, depending on camera aperture and hence beam diameter.
- the steerable camera module 610 may be used to provide steering across a symmetric range or an asymmetric range.
- An asymmetric range may form part of a larger symmetric range.
- FIG. 46A and FIG. 46B show the steerable camera module 610 oriented to cover an asymmetric range from vertical to oblique, i.e. half of an equivalent symmetric range.
- the multiplexing mirror 604 is rotated to an extreme position clockwise, and the camera field of view 612 points vertically.
- FIG. 46B the multiplexing mirror 604 is rotated to an extreme position anti-clockwise, and the camera field of view 612 points obliquely to the right.
- FIG. 47 shows a camera unit 100 implemented using a pair of steerable detail camera modules 614 and a pair of fixed overview cameras 112 .
- Each steerable detail camera module 614 consists of a steerable camera module 610 comprising a detail camera body 110 and detail camera lens 114 . Together the two steerable detail camera modules 614 cover a continuous lateral angular field of view of approximately 100 degrees.
- FIG. 48A shows the elevation and plan of the wide-angle field of view of a 300 mm wide-angle steerable WS-300 camera unit 100 , with a pair of fixed overview cameras 122 (e.g. per FIG. 47 ), operating at 16,000 feet.
- the field of view comprises the vertical field of view 180 and the two oblique fields of view 510 and 530 , arranged symmetrically relative to the flight line 226 .
- FIG. 48B shows the corresponding field of view of a 600 mm WS-600 camera unit operating at 28,000 feet.
- FIG. 48A is based on using a WS-300 camera unit comprising Nikon D800 cameras (or similar), while FIG. 48B is based on using a WS-600 camera unit 100 comprising detail cameras utilising the ON Semiconductor VITA 25K 25Mpixel CMOS image sensor (or similar), as described further below.
- VITA 25K-based cameras or similar
- FIG. 49A shows a schematic plan view of a steerable detail camera module 616 .
- the detail camera steering axis 620 corresponds to the lateral axis of the aggregate detail field of view 180 .
- a steerable detail camera module 616 is referred to below, a fixed array of cameras could also be used.
- the steerable detail camera module 616 may be implemented using one or more steerable camera modules 610 , each covering part of the required field of view.
- FIG. 49B shows the camera module 616 implemented using a pair of half-field steerable detail camera modules 614 .
- the steerable detail camera module 616 may also be implemented using one or more swinging camera modules, as described in U.S. patent application Ser. No. 10/313,727 (Partynski et al), and U.S. patent application Ser. No. 11/607,511 (Pechatnikov et al), the contents of both of which are herein incorporated by reference.
- FIG. 49C shows a schematic plan view of a steerable overview camera module 618 .
- the overview camera steering axis 622 corresponds to the lateral axis of the aggregate overview field of view 190 .
- the steerable overview camera module 618 may be implemented in the same variety of ways as the steerable detail camera module 616 .
- FIG. 49D shows a schematic plan view of a camera unit 100 comprising a pair of half-field steerable detail camera modules 614 and two overview cameras 112 , i.e. corresponding to the camera unit 100 of FIG. 47 .
- FIG. 49E shows a schematic plan view of a camera unit 100 comprising a full-field steerable detail camera module 616 and two overview cameras 112 , i.e. providing equivalent functionality to the camera unit 100 of FIG. 47 and FIG. 49D .
- FIG. 50A shows a schematic plan view of a camera unit 100 comprising a full-field steerable detail camera module 616 and a full-field steerable overview camera module 618 .
- FIG. 50B shows a plan view of a set of vertical and four-way oblique overview and detail fields of view, corresponding to the fields of view illustrated in FIG. 35A and FIG. 35B .
- the five views can be captured using five fixed-array camera units 100 as described in relation to FIG. 35A and FIG. 35B , or using three of the steerable camera units 100 shown in FIG. 50A .
- FIG. 51A shows an elevation of the field of view of a wide-angle camera unit 100 , comprising the vertical field of view 180 , oblique fields of view 510 and 530 , and interstitial fields of view 654 a and 654 b .
- Capturing the full wide field, i.e. including the interstitial fields of view 654 increases the overlap between adjacent flight lines and the observability of points within both the vertical and oblique fields of view.
- FIG. 51B shows the elevation and plan of the continuous wide-angle field of view of a WS-600 camera unit 100 operating at 28,000 feet.
- FIG. 52A shows the overlapping lateral fields of view of a wide-angle camera unit 100 from three adjacent flight lines, and illustrates how the interstitial fields of view overlap with the vertical and oblique fields of view.
- FIG. 52B shows the overlapping lateral fields of view of a wide-angle camera unit 100 from three adjacent flight lines, showing individual camera fields.
- FIG. 53A shows a plan view of the continuous wide-area fields of view of a WS-600 camera unit 100 , with both overview cameras 112 and detail cameras 110 , operating at 28,000 feet, while FIG. 53B shows the field of view of the same camera unit with just detail cameras 110 .
- the camera unit 100 can be used to capture detail photos only, with higher overlap, to allow the creation of an orthomosaic with higher spatial accuracy, and elevation data with higher resolution and accuracy.
- Longitudinal overlap i.e. in the flight direction 220 , is typically used to capture redundant photos of ground points with stereo disparity.
- the longitudinal field of view angle of a detail camera 110 is sufficiently small that little stereo disparity is present in longitudinally overlapping photos.
- FIG. 53C shows a camera unit 100 comprising two full-field steerable detail camera modules 616 for this purpose.
- the first camera module 616 a captures imagery vertically with respect to the lateral axis
- the second camera module 616 b captures imagery at e.g. a 20-degree tilt about the lateral axis.
- the tilt is chosen to maximise the observability of ground points without compromising feature-matching accuracy.
- FIG. 53D shows a plan view of the resulting un-tilted field of view 650 a and tilted field of view 650 b of such a double-field WS-600 camera unit 100 operating at 28,000 feet.
- An alternative approach is to perform two orthogonal passes over the area of interest, and capture two high-quality lateral obliques during each pass.
- Another approach is to capture all obliques with the same narrow horizontal field of view during a single pass, and reduce the flight-line spacing as necessary.
- the optimum approach i.e. the approach that maximises the flight-line spacing and hence the survey efficiency
- the four oblique directions are to arrange the four oblique directions at 45 degrees to the flight line.
- FIG. 54 shows a compact WS-300 HyperCamera unit 100 comprising four half-field steerable detail camera modules 614 arranged in an “X” configuration for the purposes of optimally capturing four high-quality oblique views in a single pass.
- the WS-300 camera unit 100 also optionally comprises a pair of fixed overview cameras that may be used for dual-resolution imaging or for capturing additional spectral bands such as NIR.
- the flight lines of the flight plan can be oriented at 45 degrees to the cardinal directions.
- FIG. 55A shows a plan view of the complete field of view 660 of an X-configuration WS-300 camera unit operating at 16,000 feet, consisting of two orthogonal fields of view 650 a and 650 b
- FIG. 55B shows the complete field of view 660 of an X-configuration WS-600 camera unit operating at 28,000 feet.
- the two overlapping fields 650 provide sufficient overlap and observability that there is no need for additional longitudinal overlap between shots, i.e. beyond the minimum required to avoid longitudinal gaps.
- FIG. 56A shows a schematic plan view of the compact X-configuration camera unit 100 of FIG. 54
- FIG. 56B shows less compact X-configuration camera unit 100 comprising two full-field steerable detail camera modules 616 arranged orthogonally. These can be deployed over two camera holes if necessary.
- an additional detail camera module 616 can be used to capture a field perpendicular to the flight direction 220 . This additional field may then be used to create the vertical orthomosaic, while the two orthogonal X-configuration fields are used to create oblique orthomosaics (as well as to increase the observability of all ground points).
- FIG. 57A shows a plan view of the fields of view 650 a , 650 b and 650 c of such an X+1-configuration WS-600 camera unit 100 operating at 28,000 feet.
- FIG. 57B shows a schematic plan view of a corresponding camera unit 100 comprising two pairs of half-field steerable detail camera modules 614 arranged in an X configuration, as well as a full-field steerable detail camera module 616 for the perpendicular vertical field.
- FIG. 58A shows a plan view of three successive shots 660 a , 660 b and 660 c of an X-configuration WS-600 camera unit 100 operating at 28,000 feet.
- FIG. 58B shows a plan view of shots 660 a and 660 b from two adjacent flight lines 226 a and 226 b of an X-configuration WS-600 camera unit 100 operating at 28,000 feet.
- the longitudinal footprint (b, 672 ) of an aerial camera is determined by its longitudinal field of view angle (delta, 670 ) and its altitude (a, 254 ), per EQ 11 in FIG. 59 .
- the longitudinal shot spacing (d, 676 ) is determined by the longitudinal footprint (b, 672 ) and the desired longitudinal overlap (c, 674 ), per EQ 12 in FIG. 59 .
- the shot rate (r) is determined by the shot spacing (d, 676 ) and the velocity (v, 678 ) of the aircraft, per EQ 13 in FIG. 59 .
- the lateral field of view angle (gamma, 270 ) of an array of cameras is a function of the number of cameras (N), the lateral field of view angle (beta, 250 ) of each camera, and the lateral angular overlap (omega, 272 ) between adjacent cameras.
- the number of cameras (N) required to cover a particular field of view angle is determined by the target lateral field of view angle (gamma, 270 ), the lateral field of view (beta, 250 ) of each camera, and the angular overlap (omega, 272 ) between adjacent cameras, per EQ 14 in FIG. 60 .
- the frame rate (q) of each camera is determined by the required number of cameras (N), the number of physical cameras (C), the shot rate (r), and the duty cycle (u) of the time-multiplexing beam-steering mechanism, per EQ 15 in FIG. 60 .
- the usable lateral field of view angle (phi, 680 ) is the field of view angle within which imagery is sufficiently consistent to be used to create a vertical orthomosaic. This may be as low as 20 degrees, or as high as 45 degrees, depending on the required quality of the resultant orthomosaic. This in turn dictates the flight-line spacing (e, 682 ), per EQ 16 in FIG. 60 .
- the flight-line spacing (e, 682 ) should be minimised to maximise the observability of ground points and avoid unobserved ground points (e.g. in urban canyons). This is particularly important for generating high-quality digital surface models, true orthomosaics, and textured 3D surfaces.
- the rate at which ground area is captured by an aerial camera system is determined by the flight-line spacing (e, 682 ) and the velocity (v, 678 ) of the aircraft, per EQ 17 in FIG. 60 .
- the resultant shot spacing (d(lambda), 690 ) is larger, per EQ 18 in FIG. 61 .
- the resultant shot rate (r(lambda)) is smaller, per EQ 19 in FIG. 61 .
- the resultant flight-line spacing (e(lambda), 692 ) is smaller, per EQ 20 in FIG. 61 .
- the resultant frame rate (q(lambda)) is smaller, per EQ 21 in FIG. 61 .
- the resultant area capture rate (R(lambda)) is smaller, per EQ 22 in FIG. 61 .
- FIG. 62 shows an electronic and mechanical block diagram of the camera unit 100 with one or more steerable camera modules 610 , and represents a variation of the block diagram in FIG. 20 .
- Blocks have the same functions as in FIG. 20 , except as described below. Power is omitted for clarity.
- the computer 300 controls one or more steerable camera modules 610 , described further below.
- Real-time imagery from the camera modules is compressed by one or more image input/output modules 700 and stored in one or more image storage devices 702 .
- the image input/output module(s) 700 may also provide the computer with access to the imagery, both for real-time monitoring and for later retrieval.
- the image storage device(s) 702 comprise removable media to allow dispatch from the field to a processing facility.
- the input/output module(s) 700 and image storage device(s) 702 may be incorporated in each individual camera 600 (e.g. when the camera 600 is a DSLR such as a Nikon D800), or may be separate.
- Image compression can utilise any suitable scheme, including JPEG or a similar DCT-based compression scheme, or JPEG2000 or a similar wavelet-based compression scheme, implemented in hardware or software.
- the computer 300 may monitor the quality of photos captured by the cameras and direct the pilot (or auto-pilot) to re-fly affected parts of the flight plan.
- the computer 300 may monitor the photos to detect poor focus, poor exposure, the presence of cloud or other unwanted objects such as other aircraft, etc.
- the pilot input device 704 allow the pilot to provide input to the computer, including to select flight plans, acknowledge alerts, navigate imagery, perform quality assurance, etc. It may comprise any suitable device such as a touch screen (in conjunction with the pilot display 302 ), tablet, keyboard, etc.
- the computer 300 may control the aircraft auto-pilot 706 to direct the aircraft along, and between, each flight line in the flight plan.
- FIG. 63 shows an electronic and mechanical block diagram of the steerable camera module 610 . Mechanical couplings are shown dashed.
- Shot-level timing is maintained by the computer 300 , and is provided by the computer 300 to a microcontroller 710 via signals to the microcontroller's control interface 712 .
- the control interface 712 may comprise a serial, USB or Ethernet interface, or any other suitable interface.
- Frame-level timing is maintained by the microcontroller 710 , and is provided by the microcontroller 710 to the camera 600 via the camera's control interface 714 .
- the microcontroller 710 provides frame-level control of the multiplexing mirror 604 via signals to its controller 640 , and the motion-compensation (MC) mirror 608 via signals to its controller 642 .
- the microcontroller 710 reports the lateral steering angle of each photo to the computer 300 , allowing the computer 300 to include the lateral steering angle in the initial estimate of the camera pose of each photo. This is subsequently refined by the bundle adjustment process (at step 412 in FIG. 21 ), as described above.
- Photo metadata including exposure, timing and pose, may be stored by the computer 300 and/or stored alongside each photo on the image storage device(s) 702 .
- the camera 600 produces a stream of images on its image stream interface 716 .
- the image stream interface 716 may comprise a serial, USB or Ethernet interface, a high-rate interface such as CoaXPress (CXP) or CameraLink (CL), utilising multiple channels if necessary, or any other suitable interface. If the image stream interface 716 comprises a CXP interface then it may also provide the camera control interface 714 .
- the multiplexing mirror 604 is attached to a goniometer stage 638 which is controlled by the microcontroller 710 via a controller 640 .
- An optional position sensor 644 supports closed-loop control.
- the goniometer stage 638 may utilise a linear motor as described in relation to FIG. 45 above, or any other suitable actuator or motor including a linear actuator or motor (such as a voice coil actuator) and a rotating motor (such as a stepper motor, DC motor or brushless motor) coupled via a gearbox or worm drive.
- the controller 640 may be any suitable controller, e.g. a PI C-867 motion controller when used in conjunction with a PI M-272 piezomotor.
- the position sensor 644 may be of any suitable type, including an optical sensor, capacitive sensor, strain gauge sensor, and eddy-current sensor.
- the goniometer stage 638 may provide two axes of rotation if the multiplexing mirror is also used to provide motion compensation.
- the motion-compensation (MC) mirror 608 is attached to a tip-tilt stage 636 which is controlled by the microcontroller 710 via a controller 642 .
- An optional position sensor 646 supports closed-loop control.
- the tip-tilt stage 636 may utilise piezoelectric actuators as described in relation to FIG. 45 above, or any other suitable actuators or motors including linear actuators or motors (such as voice coil actuators) and rotating motors (such as stepper motors, DC motors or brushless motors) coupled via a gearbox or worm drive.
- the controller 642 may be any suitable controller, e.g. a PI E-616 controller when used in conjunction with a PI S-340 tip-tilt stage.
- the position sensor 646 may be of any suitable type, including an optical sensor, capacitive sensor, strain gauge sensor, and eddy-current sensor.
- FIG. 64 shows a tabulation, for an X-configuration steerable camera unit 100 comprising 36Mpixel cameras, of shot rate (r), frame rate (q) and area capture rate (R) as a function of altitude (a, 254 ), aircraft speed (v, 678 ) and usable field of view angle (phi, 680 ).
- the tabulation is based on a constant GSD of 7 cm (and hence a focal length that increases with altitude).
- the daily capture rate is based on six survey hours per day.
- the required frame rate of the camera 600 is below the 4 Hz frame rate of the 36Mpixel Nikon D800 DSLR, and up 22,000 feet the frame rate is below the 5 Hz frame rate of the Nikon D810 DSLR.
- a camera with a higher frame rate can be used.
- CMOS image sensors with a rolling electronic shutter. This leads to geometric distortion in the presence of motion, and makes them incompatible with most motion compensation mechanisms.
- a smaller number of cameras utilise CMOS image sensors with a global electronic shutter, making them suitable for use in the steerable camera module 610 .
- Cameras which utilise CCD image sensors are also suitable, since CCDs intrinsically have a global electronic shutter.
- Cameras with external (e.g. mechanical shutters) can also be suitable.
- CMOS image sensors are widely used, including the ON Semiconductor 12Mpixel KAC-12040, 12Mpixel VITA 12K, 16Mpixel VITA 16K, and 25Mpixel VITA 25K; and the CMOSIS 12Mpixel CMV12000 and 20Mpixel CMV20000.
- the VITA 25K has a pixel pitch of 4.5 microns and supports frame rates up to 53 Hz.
- the CMV20000 has a pixel pitch of 6.4 microns and support frame rates up to 30 Hz (at 12 bits/pixel).
- the CMV12000 has a pixel pitch of 5.5 microns and support frame rates up to 180 Hz (at 12 bits/pixel) and 300 Hz (at 10 bits/pixel).
- Machine vision cameras utilising the 25Mpixel VITA 25K include the 32 fps Adimec S-25A30, the 53 fps CMOSVision Condor, and the 53 fps ISVI IC-X25CXP.
- Machine vision cameras utilising the 20Mpixel CMV20000 include the 30 fps Jai SP-20000C-PMCL, and the 30 fps Ximea CB200CG-CM.
- Machine vision cameras utilising the 12Mpixel CMV12000 include the 66 fps Adimec Q-12A65, the 62 fps Basler Beat beA4000-62 kc, and the 149 fps IO Industries Flare 12M180-CX.
- FIG. 65 shows a tabulation, for a constant 5 cm GSD and a 25Mpixel X-configuration steerable camera unit 100 (e.g. assuming a camera 600 utilising the ON Semiconductor VITA 25K image sensor), of shot rate, frame rate and area capture rate.
- the CMOSVision Condor and the ISVI IC-X25CXP support 5 cm GSD capture up to 54,000 feet.
- FIG. 66 shows a tabulation, for a constant 5 cm GSD and a 12Mpixel X-configuration steerable camera unit 100 (e.g. assuming a camera 600 utilising the CMOSIS CMV12000 image sensor), of shot rate, frame rate and area capture rate.
- the Flare 12M180-CX supports 5 cm GSD capture beyond 60,000 feet.
- the observability of ground points and the quality of 3D textures can be further increased by increasing the number of viewing directions beyond the four of the X-configuration.
- FIG. 67A shows a plan view of the fields of view of a three-field WS-600 camera unit 100 operating at 28,000 feet.
- FIG. 67B shows a plan view of the fields of view of a four-field WS-600 camera unit 100 operating at 28,000 feet.
- FIG. 68A shows a plan view of the fields of view of a five-field WS-600 camera unit 100 operating at 28,000 feet.
- These fields of view can also be captured using multiple camera units 100 , with each camera unit 100 deployed over a separate camera hole.
- four-field capture can be performed using two X-configuration camera units 100 rotated 45 degrees relative to each other.
- the fields should be evenly spaced angularly, i.e. 180/M degrees apart, and should be arranged symmetrically with respect to the flight direction, i.e. one field should be rotated 90/M degrees from the flight direction.
- the minimum adjusted flight-line spacing given by EQ 20 in FIG. 61
- the maximum adjusted frame rate given by EQ 21 in FIG. 61
- Successive frames captured by a time-multiplexed camera 600 in a steerable camera unit 610 are necessarily displaced in the direction of flight due to the forward motion of the aircraft, an effect ignored for clarity in previous figures.
- Adjacent frames are therefore slightly staggered, as illustrated in FIG. 68B for a single-field dual-camera WS-600 camera unit 100 operating at 28,000 feet.
- the two cameras 600 of the camera unit 100 are simultaneously steered from the centre outwards.
- the induced stagger has no effect on the degree of overlap within and between flight lines since the effect is systematic.
- FIG. 69 shows a plan view of the systematic motion vectors to be corrected for each of the four cameras 600 of an X-configuration camera unit 100 . If the steerable multiplexing mirror 604 is moved continuously then its motion vector 720 is dominant. The motion of the aircraft adds a smaller constant forward motion vector 722 , producing the resultant motion vector 724 (shown dashed).
- FIG. 70 shows angular position (y, 730 ) as a function of time (t, 732 ) for the multiplexing mirror 604 and the motion-compensation mirror 608 when the latter is used to correct the motion of the former.
- Graph 734 shows the linear motion of the multiplexing mirror 604
- the graphs 736 a and 736 b show two possible sinusoidal motions of the motion-compensation mirror 608 , each with its maximum velocity aligned with the centre of the exposure interval 738 .
- the angular position (y, 730 ) of the motion-compensation mirror 608 is a function of the amplitude (A) and frequency (f) of the oscillation, and time (t, 732 ), per EQ 23 in FIG. 70 .
- the angular velocity (dy/dt) of the motion-compensation mirror 608 is the derivative of EQ 23, per EQ 24, and it is at a maximum when the cosine is one, per EQ 25.
- the average angular velocity of the mirror 608 over the exposure interval 738 is the integral of EQ 24, divided by the exposure interval, per EQ 26.
- the angular velocity of the motion-compensation mirror 608 can be matched to the angular velocity of the multiplexing mirror 604 by varying both its amplitude and its frequency.
- the frequency must be an integer multiple of the camera frame rate.
- the linear motion of the multiplexing mirror 604 may be modulated by a sinusoid, periodically slowing and even reversing it. This allows the angular velocity of the motion-compensation mirror 608 to be matched to the angular velocity of the multiplexing mirror 604 via a low-amplitude sinusoid 736 , with a slight phase delay to account for the asymmetry of the motion of the multiplexing mirror 604 . Once again the average velocities are matched over the exposure interval.
- FIG. 72A shows an alternative realisation of the steerable camera module 610 , wherein the multiplexing mirror 604 is mounted at 45 degrees to the optical axis and configured to spin about the optical axis, thereby allowing the field of view 612 of the camera 600 to be steered from side to side.
- the mirror 604 is coupled to and spun by a rotating motor 648 .
- the motor 648 may be any suitable rotating motor (such as a stepper motor, DC motor or brushless motor).
- the mirror may be coupled to a linear actuator or linear motor via a gear.
- motion-compensation mirror 608 (or fixed mirror 606 in its place) provides the function of bending the optical axis to allow more convenient vertical mounting of the camera 600 . As shown, it bends the optical axis approximately 90 degrees, from vertical to horizontal. Motion-compensation mirror 608 may also provide the motion-compensation functions described earlier (in conjunction with motor or actuator 636 , not shown).
- FIG. 72A shows steerable camera module 610 with the multiplexing mirror 604 positioned for vertical imaging
- FIG. 72B shows the steerable camera module 610 with the multiplexing mirror 604 positioned for imaging at an oblique angle.
- FIG. 73A and FIG. 73B show a plan view and back elevation, respectively, of the steerable camera module 610 of FIG. 72A , both with the multiplexing mirror 604 positioned for imaging at an oblique angle.
- the steerable camera module 610 of FIG. 72A supports a wider steering range than the steerable camera module 610 of FIG. 45 , thus allowing the design of more compact and/or less expensive camera units 100 .
- FIG. 74 shows a plan view of a WS-300 HyperCamera unit 100 comprising two full-field steerable detail camera modules 616 utilising the spinning multiplexing mirror design of FIG. 72A , and one fixed overview camera 112 .
- This camera unit 100 supports the same complete detail field of view 660 as the camera unit 100 of FIG. 54 , but using two detail cameras instead of four.
- Each of the two detail cameras operate at a frame rate twice that of each of the four detail cameras.
- FIG. 75A shows a plan view of the complete field of view 660 of an alternative X-configuration WS-600 camera unit 100 utilising the spinning design and operating at 28,000 feet.
- the complete field of view 660 is equivalent to the field of view 660 of FIG. 55B .
- the spinning multiplexing mirror 604 induces rotation in individual camera fields, as shown in FIG. 75A
- successive shots provide complete coverage, as shown in FIG. 75B .
- FIG. 76A shows a double-field steerable camera module 624 comprising two full-field steerable camera modules 610 utilising the spinning design, but sharing a single camera 600 via a switching mirror 656 .
- the mirror 656 also provides the functions of the motion-compensation mirror 608 or fixed mirror 606 .
- the switching mirror 656 when the switching mirror 656 is facing in a first direction it directs the field of view of the camera 600 to a first multiplexing mirror 604 a , and when it is facing in a second direction it directs the field of view of the camera 600 to a second multiplexing mirror 604 b mounted at right angles to the first.
- the switching mirror 656 is coupled to and rotated by a rotating motor 658 .
- the motor 658 may be any suitable rotating motor (such as a stepper motor, DC motor or brushless motor).
- the mirror 656 may be coupled to a linear actuator or motor via a gear.
- FIG. 77 shows a plan view of a WS-300 HyperCamera unit 100 comprising a double-field steerable detail camera module 626 utilising the spinning design, and one fixed overview camera 112 .
- This camera unit 100 supports the same complete detail field of view 660 as the camera unit 100 of FIG. 54 , but using one detail camera instead of four.
- the single detail camera operates at a frame rate quadruple that of each of the four detail cameras.
- FIG. 78A shows a plan view of the fields of view of a four-field WS-600 camera unit 100 using the spinning design
- FIG. 78B shows a plan view of the fields of view of a five-field WS-600 camera unit 100 using the spinning design, both operating at 28,000 feet.
- These fields can be captured using a camera unit 100 comprising a suitable number of full-field detail camera modules 616 utilising the spinning design, or double-field detail camera modules 626 utilising the spinning design. As before, these fields of view can also be captured using multiple camera units 100 utilising the spinning design, with each camera unit 100 deployed over a separate camera hole. For example, four-field capture can be performed using two X-configuration camera units 100 rotated 45 degrees relative to each other.
- These fields can also be captured utilising a single steerable detail camera module 616 utilising the spinning design, by rotating the mirrors 604 and 608 (or 606 ) (and their motors) in the horizontal plane about the vertical axis of the detail camera 110 .
- the camera 110 is ideally (but not crucially) kept fixed.
- the mirrors can be rotated successively to each of a set of discrete angular positions, followed by the capture of a full field at each such position.
- Photos can be captured with arbitrary density purely as a function of the number of discrete angular positions, subject to the maximum frame rate of the camera and the maximum spin rate of the multiplexing mirror 604 .
- the mirrors can be rotated about the camera axis continuously to capture photos in a spiral pattern, with suitable motion compensation.
- FIG. 79 shows a plan view of a WS-300 HyperCamera unit 100 comprising four half-field steerable detail camera modules 614 utilising the spinning multiplexing mirror design of FIG. 72A , and one fixed overview camera 112 .
- This camera unit 100 supports the same complete detail field of view 660 as the camera unit 100 of FIG. 54 .
- FIG. 79 shows the detail fields of view 160 pointing vertically
- FIG. 80 shows the detail fields of view 160 at their oblique extremes.
- FIG. 81 shows a perspective view of the WS-300 HyperCamera unit 100 comprising four half-field steerable detail camera modules 614 , and illustrates the detail fields of views 160 pointing vertically as well as at their oblique extremes.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Optics & Photonics (AREA)
- Aviation & Aerospace Engineering (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Stereoscopic And Panoramic Photography (AREA)
- Studio Devices (AREA)
Abstract
A system for capturing aerial images, the system comprising four steerable detail camera modules, each steerable detail camera module comprising a detail camera and a beam-steering mechanism in an optical path of the detail camera module whereby the pointing direction of the detail camera is time-multiplexed to provide a wider effective field of view, the beam-steering mechanism comprising a first steerable mirror.
Description
- This is a continuation of application Ser. No. 15/009,181, filed Jan. 28, 2016, the entire contents of which are incorporated herein by reference. application Ser. No. 15/009,181 is a continuation-in-part of application Ser. No. 14/928,542, filed Oct. 30, 2015, the entire contents of which are incorporated herein by reference. application Ser. No. 14/928,542 is a continuation of application Ser. No. 14/735,504, filed Jun. 10, 2015, the entire contents of which are incorporated herein by reference. application Ser. No. 14/735,504 is a continuation-in-part of application Ser. No. 14/687,544, filed Apr. 15, 2015, the entire contents of which are incorporated herein by reference. application Ser. No. 14/687,544 is a continuation-in-part of application Ser. No. 14/478,380, filed Sep. 5, 2014, the entire contents of which are incorporated herein by reference. application Ser. No. 14/478,380 is a continuation-in-part of application Ser. No. 14/310,523, filed Jun. 20, 2014, the entire contents of which are incorporated herein by reference.
- The present invention relates to efficient aerial camera systems and efficient methods for creating orthomosaics from aerial photos.
- Accurately georeferenced mosaics of orthophotos, referred to as orthomosaics, are becoming popular alternatives to traditional pictorial maps because they can be created automatically from aerial photos, and because they show actual useful detail on the ground.
- The creation of accurate orthomosaics from aerial photos is well described in the literature. See, for example, Elements of Photogrammetry with Application in GIS, Fourth Edition (Wolf et al.) (McGraw-Hill 2014), and the Manual of Photogrammetry, Sixth Edition (American Society for Photogrammetry and Remote Sensing (ASPRS) 2013).
- The creation of an orthomosaic requires the systematic capture of overlapping aerial photos of the area of interest, both to ensure complete coverage of the area of interest, and to ensure that there is sufficient redundancy in the imagery to allow accurate bundle adjustment, orthorectification and alignment of the photos.
- Bundle adjustment is the process by which redundant estimates of ground points and camera poses are refined. Modern bundle adjustment is described in detail in “Bundle Adjustment—A Modern Synthesis” (Triggs et al.) in Vision Algorithms: Theory and Practice (Lecture Notes in Computer Science, Volume 1883, Springer 2000).
- Bundle adjustment may operate on the positions of manually-identified ground points, or, increasingly, on the positions of automatically-identified ground features which are automatically matched between overlapping photos.
- Overlapping aerial photos are typically captured by navigating a survey aircraft in a serpentine pattern over the area of interest. The survey aircraft carries an aerial camera system, and the serpentine flight pattern ensures that the photos captured by the camera system overlap both along flight lines within the flight pattern and between adjacent flight lines.
- The aerial camera system carried by a survey aircraft sees a larger view of the ground when the aircraft is operated at a higher altitude. This allows the spacing of flight lines to be increased and hence survey efficiency to be improved. However, to maintain a particular ground sampling distance, i.e. the spacing of camera pixels on the ground, larger camera image sensors or more cameras may need to used.
- In a first aspect, the present invention provides a system for capturing aerial images, the system comprising four steerable detail camera modules, each steerable detail camera module comprising a detail camera and a beam-steering mechanism in an optical path of the detail camera module whereby the pointing direction of the detail camera is time-multiplexed to provide a wider effective field of view, the beam-steering mechanism comprising a first steerable mirror.
- The first steerable mirror may be adapted to tilt with respect to an optical axis of the detail camera module to effect beam steering.
- The first steerable mirror may be tilted with respect to an optical axis of the detail camera module, and the first steerable mirror may be adapted to spin about the optical axis to effect beam steering.
- The steering axes of the steerable camera modules may be angularly distributed approximately 90 degrees apart, with at least one steering axis angled at approximately 45 degrees to the direction of flight.
- The steering axes of at least two of the steerable detail camera modules may be substantially parallel.
- The wider effective field of view provided by each steerable detail camera module may be asymmetric about a vertical pointing direction.
- The wider effective fields of view provided by at least two of the steerable detail camera modules may together form a still wider effective field of view that is symmetric about a vertical pointing direction.
- The system may be attachable, above a camera hole, to at least one of a floor of an aircraft and a floor of a pod carried by an aircraft, thereby providing the at least one detail camera with a view of the ground below the aircraft through the camera hole.
- The system may comprise at least one overview camera, the focal length of the overview camera shorter than the focal length of the detail camera.
- The detail camera may have a lens selected from the group comprising: a dioptric lens, a catoptric lens, and a catadioptric lens.
- The first steerable mirror may be stationary during an exposure period of the detail camera.
- The first steerable mirror may have a non-zero angular velocity during an exposure period of the detail camera, and the steerable detail camera module may comprise a second steerable mirror in the optical path configured to correct the effect, on the detail camera, of the angular velocity of the first steerable mirror during the exposure period, thereby to ensure the detail camera points in a consistent direction during the exposure period.
- The angular velocity of the second steerable mirror may be sinusoidal, and the average angular velocity of the second steerable mirror may be equal to the average angular velocity of the first steerable mirror during the exposure period.
- The angular velocity of the second steerable mirror may be constant during the exposure period.
- The system may further comprise at least one angular motion compensation (AMC) unit, the at least one AMC unit configured to correct the effect, on the detail camera, of angular motion of the aircraft, thereby to ensure the detail camera points in a consistent direction over time.
- AMC may be provided via a steerable mirror in the optical path.
- The system may comprise at least one forward motion compensation (FMC) mechanism, the at least one FMC mechanism configured to correct the effect, on the detail camera, of forward motion of the aircraft, thereby to reduce motion blur in photos captured by the detail camera.
- FMC may be provided via a steerable mirror in the optical path.
- FMC may be provided via time delayed integration of adjacent lines of pixels in an image sensor of the detail camera.
- FMC may be provided by moving an image sensor of the detail camera.
-
FIG. 1 shows a front view of a V5-300 HyperCamera unit, i.e. facing forwards towards the front of the aircraft. -
FIG. 2 shows a back view of the camera unit. -
FIG. 3 shows a top view of the camera unit. -
FIG. 4 shows a bottom view of the camera unit. -
FIG. 5 shows an exploded view of the camera unit. -
FIG. 6 shows an exploded view of the cameras and the central support of the camera unit, with the field of view of each camera. -
FIG. 7 shows the camera unit from below, with its combined fields of view passing through the aperture of an aircraft camera hole. -
FIG. 8 shows the overview field of view and the five overlapping detail fields of view of the camera unit. -
FIG. 9 shows the adapter plate of the HyperCamera mounted on seat tracks on the floor of an aircraft. -
FIG. 10 shows an exploded view of the camera unit, the adapter plate, and seat tracks on the floor of an aircraft. -
FIG. 11 shows an exploded view of the cameras and the central support of a wider-angle version of the camera unit. -
FIG. 12 shows the overview field of view and the five overlapping detail fields of view of the V5-180 HyperCamera unit. -
FIG. 13 shows a plan view of the HyperCamera installed in a Cessna 208 aircraft. -
FIG. 14 shows a detailed plan view of the V5-300 HyperCamera installed in a Cessna 208 aircraft. -
FIG. 15 shows a front elevation of a Cessna 208 aircraft carrying a V5-300 HyperCamera, and the resultant overview and aggregate detail fields of view. -
FIG. 16 shows a side elevation of a Cessna 208 aircraft carrying a V5-300 HyperCamera, and the resultant overview and aggregate detail fields of view. -
FIG. 17 shows the overlapping fields of view of three successive shots. -
FIG. 18 shows the overlapping fields of view of shots in adjacent flight lines. -
FIG. 19 shows the overlapping aggregate detail fields of view of a succession of shots along three adjacent flight lines. -
FIG. 20 shows a block diagram of a power and control system for the HyperCamera. -
FIG. 21 shows a photogrammetric process flow for efficiently creating an orthomosaic from multi-resolution HyperCamera photos. -
FIG. 22 shows a tabulation of ground sampling distance (GSD) as a function of altitude and camera focal length between 100 mm and 600 mm. -
FIG. 23 shows a tabulation of ground sampling distance (GSD) as a function of altitude and camera focal length between 700 mm and 1200 mm. -
FIGS. 24A, 24B and 24C show the overview field of view and the overlapping detail fields of view of the V5-300, V10-600 and V15-900 HyperCameras respectively, at an altitude of 12,000 feet. -
FIGS. 25A, 25B and 25C show the overlapping fields of view of three successive shots of the V5-300, V10-600 and V15-900 HyperCameras respectively, at an altitude of 12,000 feet. -
FIGS. 26A, 26B and 26C show the overlapping fields of view of shots in adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at an altitude of 12,000 feet. -
FIG. 27 shows a front elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera, consisting of two V5-600 HyperCamera units, and the resultant overview and aggregate detail fields of view. -
FIG. 28 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIG. 29 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIG. 30 shows a front elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units, and the resultant overview and aggregate detail fields of view. The dashed line shows the division between the fields of view of the camera units. -
FIG. 31 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIG. 32 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera, comprising three V5-900 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIGS. 33A, 33B and 33C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 24,000 feet. The dashed grid in the figures has a 1 km spacing. -
FIGS. 34A, 34B and 34C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 36,000 feet. The dashed grid in the figures has a 1 km spacing. -
FIG. 35A andFIG. 35B show front and side elevations respectively of a Cessna 208 aircraft carrying five HyperCameras installed in the aircraft cabin: a vertical V10-600, a right oblique R10-600, a left oblique L10-600, a front oblique F10-600, and a back oblique B10-600, and the resultant overview and aggregate detail fields of view of each. -
FIG. 36 shows a side elevations of a Cessna 208 aircraft carrying five HyperCameras installed in a standard Cessna 208 belly-mounted cargo pod: a vertical V10-600, a right oblique R10-600, a left oblique L10-600, a front oblique F10-600, and a back oblique B10-600, and the resultant overview and aggregate detail fields of view of each. -
FIG. 37 shows a bottom view of the camera unit with an alternative arrangement of the cameras. -
FIG. 38 shows a camera unit implemented as an array of cameras and their respective optical axes. -
FIG. 39 shows a camera unit implemented as a single camera time-multiplexed via a steerable mirror to produce multiple optical paths. -
FIG. 40 shows an alternative configuration of a camera unit implemented as a single camera time-multiplexed via a steerable mirror to produce multiple optical paths. -
FIG. 41 shows a camera unit implemented as a single camera time-multiplexed via a steerable mirror to produce multiple optical paths, with a steerable mirror for angular motion compensation (AMC). -
FIG. 42 shows a diagram and equation relating to the calculation of the field of view and resolution of a camera. -
FIG. 43 shows a diagram and equations relating to the calculation of the field of view of an array of cameras. -
FIG. 44 shows a diagram and equations relating to the calculation of the resolution of tilted cameras. -
FIG. 45 shows a steerable camera module. -
FIG. 46A shows the steerable camera module with its mirror positioned for imaging at a vertical viewing angle. -
FIG. 46B shows the steerable camera module with its mirror positioned for imaging at an oblique viewing angle. -
FIG. 47 shows a WS-300 HyperCamera unit comprising two half-field steerable detail camera modules and two fixed overview cameras. -
FIG. 48A shows the elevation and plan of the wide-angle field of view of a WS-300 camera unit operating at 16,000 feet. -
FIG. 48B shows the elevation and plan of the wide-angle field of view of a WS-600 camera unit operating at 28,000 feet. -
FIG. 49A shows a schematic plan view of a full-field steerable detail camera module. -
FIG. 49B shows a schematic plan view of a full-field steerable detail camera module comprising a pair of half-field steerable detail camera modules. -
FIG. 49C shows a schematic plan view of a full-field steerable overview camera module. -
FIG. 49D shows a schematic plan view of a camera unit comprising a pair of half-field steerable detail camera modules and two overview cameras. -
FIG. 49E shows a schematic plan view of a camera unit comprising a full-field steerable detail camera module and two overview cameras. -
FIG. 50A shows a schematic plan view of a camera unit comprising a full-field steerable detail camera module and a full-field steerable overview camera module. -
FIG. 50B shows a plan view of a set of vertical and four-way oblique overview and detail fields of view. -
FIG. 51A shows an elevation of the field of view of a wide-angle camera unit, comprising vertical, oblique and interstitial fields of view. -
FIG. 51B shows the elevation and plan of the continuous wide-angle field of view of a WS-600 camera unit operating at 28,000 feet. -
FIG. 52A shows the overlapping lateral fields of view of a wide-angle camera unit from three adjacent flight lines. -
FIG. 52B shows the overlapping lateral fields of view of a wide-angle camera unit from three adjacent flight lines, showing individual camera fields. -
FIG. 53A shows a plan view of the fields of view of a WS-600 camera unit with both overview and detail cameras operating at 28,000 feet. -
FIG. 53B shows a plan view of the field of view of a WS-600 camera unit with just detail cameras operating at 28,000 feet. -
FIG. 53C shows a camera unit comprising two full-field steerable detail camera modules. -
FIG. 53D shows a plan view of the fields of view of a double-field WS-600 camera unit operating at 28,000 feet. -
FIG. 54 shows a WS-300 HyperCamera unit comprising four half-field steerable detail camera modules arranged in an X configuration and two fixed overview cameras. -
FIG. 55A shows a plan view of the fields of view of an X-configuration WS-300 camera unit operating at 16,000 feet. -
FIG. 55B shows a plan view of the fields of view of an X-configuration WS-600 camera unit operating at 28,000 feet. -
FIG. 56A shows a camera unit comprising two pairs of half-field steerable detail camera modules arranged in an X configuration. -
FIG. 56B shows a camera unit comprising two full-field steerable detail camera modules arranged orthogonally. -
FIG. 57A shows a plan view of the fields of view of an X+1-configuration WS-600 camera unit operating at 28,000 feet. -
FIG. 57B shows a camera unit comprising two pairs of half-field steerable detail camera modules arranged in an X configuration, as well as a full-field steerable detail camera module. -
FIG. 58A shows a plan view of three successive shots of an X-configuration WS-600 camera unit operating at 28,000 feet. -
FIG. 58B shows a plan view of shots from two adjacent flight lines of an X-configuration WS-600 camera unit operating at 28,000 feet. -
FIG. 59 shows a diagram and equations relating to the calculation of the longitudinal field of view and shot rate of an aerial camera. -
FIG. 60 shows a diagram and equations relating to the calculation of the lateral field of view of an aerial camera system and area capture rate. -
FIG. 61 shows a diagram and equations relating to the effect of field of view rotation on aerial camera system shot, frame and area capture rates. -
FIG. 62 shows an electronic and mechanical block diagram of a camera unit with steerable camera modules. -
FIG. 63 shows an electronic and mechanical block diagram of a steerable camera module. -
FIG. 64 shows a tabulation, for a 36Mpixel X-configuration steerable camera unit and a 7 cm GSD, of shot rate, frame rate and area capture rate as a function of altitude, aircraft speed and usable field of view angle. -
FIG. 65 shows a tabulation, for a 25Mpixel X-configuration steerable camera unit and a 5 cm GSD, of shot rate, frame rate and area capture rate as a function of altitude, aircraft speed and usable field of view angle. -
FIG. 66 shows a tabulation, for a 12Mpixel X-configuration steerable camera unit and a 5 cm GSD, of shot rate, frame rate and area capture rate as a function of altitude, aircraft speed and usable field of view angle. -
FIG. 67A shows a plan view of the fields of view of a three-field WS-600 camera unit operating at 28,000 feet. -
FIG. 67B shows a plan view of the fields of view of a four-field WS-600 camera unit operating at 28,000 feet. -
FIG. 68A shows a plan view of the fields of view of a five-field WS-600 camera unit operating at 28,000 feet. -
FIG. 68B shows a plan view of the staggered field of view of a single-field WS-600 camera unit operating at 28,000 feet. -
FIG. 69 shows a plan view of systematic motion vectors to be corrected. -
FIG. 70 shows a graph of angular position as a function of time for a multiplexing mirror and a motion-compensation mirror when the latter is used to correct the motion of the former. -
FIG. 71 shows a graph of the linear motion of a multiplexing mirror modulated by a sinusoid, periodically slowing the mirror. -
FIG. 72A shows an alternative steerable camera module utilising a spinning primary mirriro, with its mirror positioned for imaging at a vertical viewing angle. -
FIG. 72B shows the steerable camera module with its mirror positioned for imaging at an oblique viewing angle. -
FIG. 73A shows a plan view of the steerable camera module with its mirror positioned for imaging at an oblique viewing angle. -
FIG. 73B shows a back elevation of the steerable camera module with its mirror positioned for imaging at an oblique viewing angle. -
FIG. 74 shows a plan view of a WS-300 HyperCamera unit comprising two full-field steerable detail camera modules and one fixed overview camera. -
FIG. 75A shows a plan view of the fields of view of an X-configuration WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet. -
FIG. 75B shows a plan view of three successive shots of the X-configuration WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet. -
FIG. 76A shows a double-field steerable camera module. -
FIG. 76B shows detail of the mirror arrangement fromFIG. 76A . -
FIG. 77 shows a plan view of the WS-300 HyperCamera unit ofFIG. 76A comprising a double-field steerable detail camera module and one fixed overview camera. -
FIG. 78A shows a plan view of the fields of view of a four-field WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet. -
FIG. 78B shows a plan view of the fields of view of a five-field WS-600 camera unit using the alternative steerable camera module and operating at 28,000 feet. -
FIG. 79 shows a plan view of a WS-300 HyperCamera unit comprising four half-field spinning-mirror steerable detail camera modules, arranged in an X configuration, and one fixed overview camera. -
FIG. 80 shows a plan view of the WS-300 HyperCamera unit ofFIG. 79 , with the detail fields of view shown at an extreme angle. -
FIG. 81 shows a perspective view of the WS-300 HyperCamera unit ofFIG. 79 , with the detail fields of view shown at both a vertical angle and at an extreme angle. -
-
- 100 Camera unit.
- 110 Detail camera body.
- 112 Overview camera body.
- 114 Detail camera lens.
- 116 Overview camera lens.
- 118 Lens of special-purpose (e.g. NIR) camera.
- 120 Frame.
- 122 Frame center support.
- 124 Frame side support.
- 126 Frame rear support.
- 128 Frame front support.
- 130 Mount point block.
- 132 Mount point.
- 134 Mount bolt.
- 140 Mount for detail camera.
- 142 Mount for overview camera.
- 144 Clamp for detail camera lens.
- 146 Clamp for overview camera lens.
- 150 Power and control distribution box.
- 160 Detail field of view.
- 162 Lateral detail field of view.
- 164 Longitudinal detail field of view.
- 170 Overview field of view.
- 172 Lateral overview field of view.
- 174 Longitudinal overview field of view.
- 180 Aggregate detail field of view.
- 182 Lateral aggregate detail field of view.
- 190 Aggregate overview field of view.
- 192 Lateral aggregate overview field of view.
- 194 Nominal 30-degree field of view.
- 196 One-km grid.
- 200 Adapter plate.
- 202 Seat track fastener.
- 210 Aircraft floor.
- 212 Camera hole.
- 214 Seat track.
- 216 Adapter plate aperture.
- 220 Direction of flight.
- 222 Flight path.
- 224 Shot position.
- 226 Flight line.
- 230 Aerial survey aircraft.
- 232 Aerial survey aircraft belly-mounted pod.
- 250 Field of view angle (beta).
- 252 Focal length (f).
- 254 Altitude above ground level (a).
- 256 Image sensor width (s).
- 258 Swath width (w).
- 260 Image sensor pixel pitch (p).
- 262 Ground sampling distance (GSD) (g).
- 270 Aggregate field of view angle (gamma).
- 272 Angular camera overlap (omega).
- 278 Aggregate swath width (w(N)).
- 280 Camera tilt angle (theta).
- 282 Tilted camera GSD (g(theta)).
- 284 Tilted camera field of view.
- 286 Tilted camera with tilted focal plane GSD (h(theta)).
- 288 Tilted camera with tilted focal plane field of view.
- 300 Computer.
- 302 Pilot display.
- 304 Inertial Measurement Unit (IMU).
- 306 Global Navigation Satellite System (GNSS) receiver.
- 308 Camera interface.
- 310 Camera control unit (CCU).
- 320 Battery unit.
- 322 Aircraft auxiliary power.
- 324 Ground power unit (GPU).
- 326 DC-DC converters.
- 330 Angular motion compensation (AMC) unit(s).
- 400 Detail photos.
- 402 Overview photos.
- 404 Orthomosaic.
- 410 Match features step.
- 412 Solve pose and positions step.
- 414 Orthorectify step.
- 416 Blend step.
- 510 Right oblique detail field of view.
- 512 Lateral right oblique detail field of view.
- 514 Longitudinal right oblique detail field of view.
- 520 Right oblique overview field of view.
- 522 Lateral right oblique overview field of view.
- 524 Longitudinal right oblique overview field of view.
- 530 Left oblique detail field of view.
- 532 Lateral left oblique detail field of view.
- 534 Longitudinal left oblique detail field of view.
- 540 Left oblique overview field of view.
- 542 Lateral left oblique overview field of view.
- 544 Longitudinal left oblique overview field of view.
- 550 Front oblique detail field of view.
- 552 Lateral front oblique detail field of view.
- 554 Longitudinal front oblique detail field of view.
- 560 Front oblique overview field of view.
- 562 Lateral front oblique overview field of view.
- 564 Longitudinal front oblique overview field of view.
- 570 Back oblique detail field of view.
- 572 Lateral back oblique detail field of view.
- 574 Longitudinal back oblique detail field of view.
- 580 Back oblique overview field of view.
- 582 Lateral back oblique overview field of view.
- 584 Longitudinal back oblique overview field of view.
- 600 Camera.
- 602 Camera optical axis.
- 604 Steerable multiplexing mirror.
- 606 Fixed mirror.
- 608 Steerable motion compensation mirror.
- 610 Steerable camera module.
- 612 Camera field of view.
- 614 Half-field steerable detail camera module.
- 616 Full-field steerable detail camera module.
- 618 Full-field steerable overview camera module.
- 620 Detail camera steering axis.
- 622 Overview camera steering axis.
- 624 Double-field steerable camera module.
- 626 Double-field detail steerable camera module.
- 630 Multiplexing mirror linear motor.
- 632 Multiplexing mirror actuator rod.
- 634 Multiplexing mirror mount.
- 636 Motion compensation mirror stage.
- 638 Multiplexing mirror stage.
- 640 Multiplexing mirror controller.
- 642 Motion compensation mirror controller.
- 644 Multiplexing mirror sensor.
- 646 Motion compensation mirror sensor.
- 648 Multiplexing mirror rotating motor.
- 650 Wide-angle aggregate detail field of view.
- 652 Wide-angle aggregate overview field of view.
- 654 Interstitial field of view.
- 656 Switching mirror.
- 658 Switching mirror rotating motor.
- 660 X-configuration combined detail field of view.
- 670 Longitudinal field of view angle (delta).
- 672 Longitudinal footprint (b).
- 674 Longitudinal overlap (c).
- 676 Longitudinal shot spacing (d).
- 678 Aircraft ground velocity (v).
- 680 Usable vertical field of view angle (phi).
- 682 Lateral flight-line spacing (e).
- 684 Lateral field of view axis.
- 686 Field of view axis rotation angle (lambda).
- 688 Rotated lateral field of view axis.
- 690 Longitudinal shot spacing with rotated axis (d(lambda)).
- 692 Flight-line spacing with rotated axis (e(lambda)).
- 700 Image input/output module.
- 702 Image storage device.
- 704 Pilot input device.
- 706 Auto-pilot.
- 710 Camera microcontroller.
- 712 Camera module control interface.
- 714 Camera control interface.
- 716 Image stream interface.
- 720 Steering motion vector.
- 722 Forward motion vector.
- 724 Resultant motion vector.
- 730 Angular position (y).
- 732 Time (t).
- 734 Multiplexing mirror angular position graph.
- 736 Motion-compensation mirror angular position graph.
- 738 Exposure interval.
- HyperCamera™ is a range of aerial camera systems suitable for deployment in a wide range of aircraft, large and small. The camera systems are modular, and designed to be installed above one or more standard camera holes where appropriate, as are typically provided through the floor of a survey aircraft or airborne pod.
- Each HyperCamera model is defined by its pointing direction (P), the number of cameras (N) in its array, and the focal length (f) of the cameras, and is identified by the designator PN-f. For example, a 300 mm five-camera vertical HyperCamera is referred to as a V5-300 model.
- Wide-angle steerable HyperCamera models are identified by the designator WS-f. For example, a 300 mm wide-angle steerable HyperCamera is referred to as a WS-300 model.
- The level of detail captured by an aerial camera is typically characterized by the ground sampling distance (GSD), i.e. the distance between adjacent pixel centers when projected onto the ground within the camera's field of view.
- The GSD is determined by the focal length (f, 252) of the camera lens, the altitude above ground level (a, 254), and the pixel pitch (p, 260) of the image sensor, per
EQ 1 inFIG. 42 . - The efficiency of aerial imaging is typically characterized by the area captured per unit time (e.g. square km per hour). This is proportional to the speed of the aircraft and the width of the field of view (FOV) of the aerial camera system, referred to as the swath width.
- The swath width (w, 258) of a single camera is determined by the focal length (f, 252) of the camera lens, the altitude above ground level (a, 254), and the lateral size (s, 256) of the image sensor, per
EQ 2 inFIG. 42 . Doubling the altitude doubles the swath width. - The lateral field of view (beta, 250) of a single camera is determined by the focal length (f, 252) of the camera lens and the lateral size (s, 256) of the image sensor, per
EQ 4 inFIG. 42 . Doubling the focal length approximately halves the field of view. - The lateral field of view (gamma, 270) of an array of cameras is determined by the field of view of each camera (beta, 250), the number of cameras (N), and their angular overlap (omega, 272), per
EQ 7 inFIG. 43 . Doubling the number of cameras approximately doubles the field of view. The swath width (w(N), 278) of the array of cameras is given byEQ 8 inFIG. 43 . - The practical field of view of an aerial camera system is limited by the acceptable level of obliqueness in the aerial imagery—e.g. how much buildings are allowed to lean when captured at the edge of the field of view. The practical field of view is usually limited to 50 degrees or less.
- Given a limit on the practical field of view, higher capture efficiency (and therefore lower capture cost) can be achieved by flying at higher speed and/or at a higher altitude. Flying at a higher altitude requires longer focal lengths to maintain the same GSD. Since each camera then has a narrower individual field of view, this in turn requires a larger number of cameras to maintain the overall field of view.
- The HyperCamera range includes models suited to a variety of operating altitudes, and supporting a range of GSDs and capture efficiencies. Depending on target speed and altitude, a HyperCamera system may be installed in any aircraft suitable for aerial imaging. By way of example, this includes (but is not limited to) piston aircraft such as a
Cessna 210, turboprop aircraft such as a Cessna 208, and turbofan (jet) aircraft such as a Cessna Citation, allowing aerial imaging from low altitudes to altitudes in excess of 40,000 feet, at speeds ranging from less than 100 knots to 400 knots. - The aircraft may be unpressurised or pressurised, and each camera hole may be open or contain an optical glass window as appropriate. Each camera hole may be optionally protected by a door which can be closed when the HyperCamera is not in operation.
- In a preferred embodiment, as illustrated in
FIGS. 1 through 5 , a V5-300 HyperCamera comprises acamera unit 100 incorporating five detail cameras 110 and a relatively wider-angle overview camera 112. Each detail camera 110 has a detail lens 114, and theoverview camera 112 has aoverview lens 116. - The
overview lens 116 is characterised by having a significantly wider angle than the detail lens 114. While it may be a true wide-angle lens, it may also be a normal lens or even a telephoto lens so long as it is significantly wider than the detail lens 114. Likewise, while the detail lens 114 may be a true telephoto lens, it may also be a normal lens or even a wide-angle lens so long as it is significantly narrower than theoverview lens 116. - The
cameras 110 and 112 are preferably commercial off-the-shelf (COTS) digital SLR (DSLR) cameras. The use of COTS cameras allows the system to be readily adapted to the latest and best available cameras. Alternatively or additionally, to provide imaging features not available in COTS cameras, such as non-mosaiced RGB imaging, multi-spectral imaging, and forward motion compensation, dedicated camera designs may also be utilised. - High-resolution COTS cameras are available with typical pixel counts ranging from 24Mpixels to 36Mpixels, from vendors such as Nikon and Canon. The 36Mpixel Nikon D800 (and D810) DSLR camera is a particularly good choice for the present system.
- DSLR cameras offer a wide range of high-quality lenses, allowing the system to be readily configured to operate at different altitudes and resolutions.
- The system is readily adapted to a mixture of cameras. For example, a relatively more expensive camera with a higher pixel count may be employed as the overview camera. 70Mpixel DSLR cameras are expected to be available in the near future, and a 70Mpixel camera would be a good choice for the overview camera.
- In the preferred embodiment the detail lenses 114 of the detail cameras 110 all have the same focal length, and the detail cameras 110 all have the same pixel size. Thus the
camera unit 100 embodies two distinct camera resolutions—overview and detail. This is readily extended to multiple resolutions greater than two through the use detail lenses 114 with different focal lengths, and/or the use of detail cameras 110 with different pixel sizes. Thecamera unit 100 may also incorporate multiple overview cameras with different resolutions. - Each detail lens 114 and
overview lens 116 may be a fixed-focus lens focused at infinity or a variable-focus lens. In the latter case the corresponding camera 110 and/or 112 incorporates an auto-focus mechanism. - Each detail camera 110 is bolted to a camera mount 140, which in turn is bolted to a
center support 122. Each detail camera lens 114 is further secured by a clamp 144 which is bolted to the detail camera mount 140. - The overview camera is bolted to a
camera mount 142, which in turn is bolted to thecenter support 122. Theoverview camera lens 116 is further secured by aclamp 146 which is bolted to theoverview camera mount 142. - The camera mounts 140 and 142 isolate much of the structure of
camera unit 100 from the specifics of individual camera models and lens sizes. - The
center support 122 is attached to a pair of side supports 124 a and 124 b, and each side support 124 is in turn attached to arear support 126 and afront support 128 to form a rigid frame 120. - Each side support 124 is attached to mount point block 130 via a set of four bolts, and the mount point block 130 is in turn attached to the
rear support 126 orfront support 128, as appropriate, via a further set of four bolts. The mount point blocks 130 thereby provide the attachment mechanism between the side supports 124 and the rear andfront supports - Each of the side supports 124 and the rear and
front supports center support 122 is pocketed to minimise weight while maximising rigidity. - Each mount point block 130 is solid, and serves the additional purpose of providing a point of attachment between the
camera unit 100 and a survey aircraft, as described below. - All parts are made from light-weight aluminium, except for fasteners which are made from steel.
- The
rear support 126 and thefront support 128 hold three power and control distribution boxes 150. Each box 150 distributes power and control signals to a pair of cameras. For clarity, the power and control cabling between the boxes 150 and thecameras 110 and 112 is omitted in the figures. - In the preferred embodiment each detail camera 110 has a lens 114 with a focal length of 300 mm suitable for high-resolution imaging at relatively high altitudes. For example, when using a 36Mpixel Nikon D800 camera (which has a 4.88 micron pixel pitch), a 300 mm lens allows a ground sampling distance (GSD) of 30 cm at 60,000 feet, 20 cm at 40,000 feet, 10 cm at 20,000 feet, 8 cm at 16,000 feet, 6 cm at 12,000 feet, 4 cm at 8,000 feet, 2 cm at 4,000 feet, 1 cm at 2,000 feet, etc.
- Assuming the detail cameras 110 and
overview camera 112 have similar pixel counts and pixel sizes, theoverview camera 112 ideally has alens 116 with a focal length that is between 4 and 8 times shorter than the focal length of the detail lens 114, as discussed further below. I.e. for a 300 mm detail lens 114, suitable focal lengths for theoverview lens 116 range from about 40 mm to 75 mm. For illustrative purposes the present system utilises a 50mm overview lens 116. -
FIG. 6 shows the 6.90-degree lateral field of view 162 of each of the five detail cameras 110 with 300 mm lenses 114, and the 39.60-degree lateral field of theoverview camera 112 with a 50mm lens 116. - In this specification, the lateral direction is the direction perpendicular to the direction of
flight 220, and the longitudinal direction is the direction parallel to the direction offlight 220. - As shown, the detail cameras are angled 6 degrees apart laterally, i.e. slightly less than their 6.90-degree fields of view 162, so that the fields of view 162 overlap slightly.
- Using 36Mpixel Nikon D800 cameras, the five detail cameras 110 have an aggregate field of view with a pixel count of approximately 160Mpixels, i.e. excluding overlap.
- Stock telephoto lenses suitable for use as detail lenses 114 are available in a variety of focal lengths, typically including 85 mm, 105 mm, 135 mm, 180 mm, 200 mm, 300 mm, 400 mm, 500 mm, 600 mm, 700 mm, and 800 mm.
- At 20,000 feet a 400 mm lens on a Nikon D800 camera allows a GSD of 7.4 cm, a 600 mm lens a GSD of 5.0 cm, and an 800 mm lens a GSD of 3.7 cm.
- Stock normal and wide-angle lenses suitable for use as an
overview lens 116 are available in a variety of focal lengths, typically including 10.5 mm, 14 mm, 16 mm, 18 mm, 20 mm, 21 mm, 24 mm, 28 mm, 35 mm, 40 mm, 45 mm, 50 mm, 55 mm, 60 mm and 70 mm. - The
camera unit 100 is readily adapted for different models and sizes of cameras 110 (and 112) and lenses 114 (and 116) via different camera mounts 140 (and 142) and clamps 144 (and 146). For extremely long lenses ataller center support 122 can be used. - As shown in
FIG. 6 andFIG. 7 , the detail cameras are angled inwards so that their fields of view 162 cross over directly below thecamera unit 100, creating a waist of minimum diameter where the fields of view pass through thecamera hole 212. This makes thecamera unit 100 compatible with standard 20-inch camera holes, as well as camera holes as small as about 17 inches. -
FIG. 8 shows the projection of the three-dimensional fields ofview 160 and 170 of the detail cameras 110 andoverview camera 112 onto a ground plane. It shows how the detail field of views 160 overlap in a direction perpendicular to the direction offlight 220. -
FIG. 9 shows anadapter plate 200 that attaches to the seat tracks 214 of an aircraft, a Cessna 208 in this case, via standard seat track fasteners 202. The adapter plate has anaperture 216 which exposes acamera hole 212 through thefloor 210 of the aircraft. -
FIG. 10 shows an exploded view of thecamera unit 100,adapter plate 200, and theaircraft floor 210. Theadapter plate 200 is designed to attach thecamera unit 100 to a particular aircraft, and isolates the design of thecamera unit 100 from aircraft specifics. A different adapter plate is designed for each aircraft attachment variation, e.g. due to different seat track spacings, or because the aircraft's camera hole installation includes its own mounting points. - Four mount points 132 are bolted to the adapter plate, with each mount point 132 mating with a recess in the base of its corresponding mount point block 134. A mount bolt 143 securely attaches each mount point block 134 to its corresponding mount point 132, thus attaching the
camera unit 100 to theadapter plate 200. - The
adapter plate 200 allows thecamera unit 100 to be easily installed in and subsequently removed from an aircraft via installation and removal of the four mount bolts 143. Theadapter plate 200 is itself generally easily installed in and removed from an aircraft, requiring no modification to the aircraft (assuming a suitable camera hole is already installed). The installation of external camera pod is generally a much more complicated operation. -
FIG. 11 shows a V5-180 HyperCamera unit 100 that utilises shorter 180 mm lenses 114 for the detail cameras 110, and a matching 28mm lens 116 for theoverview camera 112. - When using a 36Mpixel Nikon D800 camera (which has a 4.88 micron pixel pitch), a 180 mm lens allows a ground sampling distance (GSD) of 9.9 cm at 12,000 feet, 8.3 cm at 10,000 feet, 6.6 cm at 8,000 feet, 5 cm at 6,000 feet, 3.3 cm at 4,000 feet, and so on.
-
FIG. 11 shows the 11.40-degree lateral field of view 162 of each of the five detail cameras 110 with 180 mm lenses 114, and the 65.50-degree lateral field of theoverview camera 112 with a 28mm lens 116. - As shown, the detail cameras are angled 10.50 degrees apart laterally, i.e. slightly less than their 11.40-degree fields of view 162, so that the fields of view 162 overlap slightly.
-
FIG. 12 shows the projection of the three-dimensional fields ofview 160 and 170 of the detail cameras 110 andoverview camera 112 ofFIG. 10 onto a ground plane. It shows how the detail field of views 160 overlap in a direction perpendicular to the direction offlight 220, and how the wider fields of view associated with the shorter lenses leads to a lower flying altitude for the same footprint on the ground, i.e. in comparison toFIG. 8 . -
FIG. 13 andFIG. 14 show plan views of a Cessna 208survey aircraft 230 carrying acamera unit 100 installed centrally over a camera hole. The figures also show a camera control unit 310 (CCU) andbattery unit 320 used to control and power thecamera unit 100. These are described in more detail below. For clarity, the cabling connecting theCCU 310,battery unit 320 andcamera unit 100 is omitted. -
FIG. 15 shows a front elevation of the Cessna 208survey aircraft 230 carrying a HyperCamera, and shows the lateral overview field ofview 172 of thecamera unit 100, and the aggregate lateral detail field ofview 182 of thecamera unit 100. The aggregate lateral detail field ofview 182 is the aggregate of the five individual overlapping lateral detail fields of view 162. -
FIG. 16 shows a side elevation of the Cessna 208survey aircraft 230 carrying a HyperCamera, and shows the longitudinal overview field ofview 174 of thecamera unit 100, and the longitudinal detail field ofview 164 of thecamera unit 100. -
FIG. 17 shows the overlapping overview fields ofview 170 and aggregate detail fields ofview 180 of three successive shots in the direction offlight 220. The aggregate detail field ofview 180 is the aggregate of the five individual overlapping detail fields of view 160. At the camera firing rate illustrated in the figure (i.e. as implied by the longitudinal overlap), the aggregate detail fields ofview 180 overlap by about 20% longitudinally, while the overview fields ofview 170 overlap by about 85% longitudinally. -
FIG. 18 shows the overlapping overview fields ofview 170 and aggregate detail fields ofview 180 of two shots from adjacent flight lines, i.e. flown inopposite directions 220. At the flight-line spacing illustrated in the figure, the aggregate detail fields ofview 180 overlap by between 20% and 25% laterally, while the overview fields ofview 170 overlap by about 40% laterally. - Assuming the detail cameras 110 and the
overview camera 112 have similar pixel counts and pixel sizes, the size of the lateral overview field ofview 172 and the size of the lateral aggregate detail field ofview 182 are similar when the ratio of the focal length of the detail camera lens 114 to the focal length of theoverview camera lens 116 is about 6, and useful lens combinations can be chosen with focal length ratios between about 4 and 8. -
FIG. 19 shows the overlapping aggregate detail fields ofview 180 of a succession of shots along three adjacent flight lines that are part of a typicalserpentine flight path 222, i.e. a subset of the flight lines that would make up a typical large-area survey. For clarity the corresponding overview fields ofview 170 are omitted. The figure also shows the shot position 224 corresponding to each aggregate detail field ofview 180, i.e. the position of thesurvey aircraft 230. - As already noted, traditional single-resolution aerial surveys are typically operated with 60/40 overlap, i.e. 60% forward (or longitudinal) overlap, and 40% side (or lateral) overlap. With the multi-resolution HyperCamera operated as shown in
FIGS. 17 through 19 , overview photos are captured with better than 85/40 overlap, and detail photos are captured with only 20/20 overlap at best. - Compared with a traditional single-resolution aerial camera system and a comparable aggregate detail pixel count (e.g. 160Mpixel), the HyperCamera is between 2 and 3 times more efficient, as detailed below, with respect to both reduced survey flying time and fewer photos to process. The HyperCamera also has a higher efficiency than many aerial camera systems due to its high (detail) pixel count alone.
- As an alternative to capturing both overview and detail photos, the HyperCamera can be used to capture detail photos only, with higher overlap (e.g. 60/40 rather than 20/20), to allow the creation of an orthomosaic with higher spatial accuracy, but at greater capture and processing cost. In this case the
overview camera 112 can be omitted or used for another purpose. - To analyse the relative efficiency of a multi-resolution HyperCamera, assume a multi-resolution HyperCamera configuration with a lateral overlap of X %, a longitudinal overlap of Y %, N detail cameras 110, and
M overview cameras 112, and for comparison, a single-resolution HyperCamera configuration with lateral overlap of A %, longitudinal overlap of B %, N detail cameras, and no overview camera. Assuming X is smaller than A, the improvement in lateral efficiency, as reflected in a greater flight-line spacing and shorter flying time and fewer detail photos captured, is given by (1−X)/(1−A). Likewise, assuming Y is smaller than B, the improvement in longitudinal efficiency, as reflected in a greater shot spacing and shorter flying time and fewer detail photos captured, is given by (1−Y)/(1−B). The overall improvement in efficiency is given by (1−X)(1−Y)/(1−A)(1−B). This needs to be discounted by the overhead of capturing overview photos, i.e. multiplied by a factor of (N/(N+M)). For X/Y=20/20, A/B=60/40, N=5, and M=1, the net efficiency improvement is 2.2. - The greater efficiency comes at the cost of performing some photogrammetric calculations at the lower resolution of the
overview camera 112 rather than at the higher resolution of the detail cameras 110. However, this is at least partially compensated for by the greater overlap between overview photos than in traditional practice. -
FIG. 20 shows a block diagram of a power and control system for thecamera unit 100. The detail cameras 110 andoverview camera 112 are controlled by acomputer 300 via acamera interface 308, which may consist of a set of digital switches that trigger the remote control interfaces of the cameras. - The
computer 300 uses one or more Global Navigation Satellite System (GNSS)receiver 304 to monitor the position and speed of thesurvey aircraft 230 in real time. The GNSS receiver(s) may be compatible with a variety of space-based satellite navigation systems, including the Global Positioning System (GPS), GLONASS, Galileo and BeiDou. - The
computer 300 provides precisely-timed firing signals to thecameras 110 and 112 via thecamera interface 308, to trigger camera exposure, according to a stored flight plan and the real-time position and speed of the aircraft. If a camera 110 and/or 112 incorporates an auto-focus mechanism then thecomputer 300 also provides a focus signal to each such camera to trigger auto-focus prior to exposure. - The
computer 300 fires theoverview camera 112 and the detail cameras 110 at the same rate. Alternatively, thecomputer 300 may fire theoverview camera 112 at a different rate to the detail cameras 110, i.e. either a higher rate or lower rate, to achieve a different overlap between successive overview photos, i.e. either a higher overlap or a lower overlap, independent of the overlap between successive detail photos. Thecomputer 300 may fire the cameras simultaneously, or it may stagger the timing of the firing, e.g. to achieve a different alignment of photos longitudinally, or to reduce peak power consumption. - The flight plan describes each flight line making up the survey, and the nominal camera firing rate along each flight line required to ensure that the necessary overlap is maintained between successive shots. The firing rate is sensitive to the elevation of the terrain below the aircraft, i.e. the higher the terrain the higher the firing rate needs to be. It is adjusted by the
computer 300 according to the actual ground speed of the aircraft, which may vary from its nominal speed due to wind and the pilot's operation of the aircraft. - The
computer 300 also uses the flight plan and real-time GNSS position to guide the pilot along each flight line via apilot display 302. - As shown in
FIG. 20 , the position data from the GNSS receiver is optionally augmented with orientation information (roll, pitch and yaw) from an inertial measurement unit 306 (IMU). It allows thecomputer 300 to provide enhanced feedback to the pilot on how closely the pilot is following the flight plan, and allows more accurate processing of the photos. In the absence of theIMU 306 the GNSS receiver connects directly to thecomputer 300. - Each
camera 110 and 112 stores its shots locally, e.g. in removable flash memory. This eliminates the need for centralised storage in the HyperCamera system, and the need for a high-bandwidth data communication channel between the cameras and the centralised storage. - The GNSS position of each shot may be delivered to each
camera 110 and 112, to allow the camera to tag each photo with its GNSS position. - One or more optional angular motion compensation (AMC)
units 330, responsive to the orientation reported by theIMU 306, correct the orientation of the cameras so that they maintain a consistent pointing direction over time, despite the aircraft rolling, pitching or yawing during flight. This ensures that the captured photos can be used to create an orthomosaic without gaps, while allowing the overlap between successive shots and between adjacent flight lines to be minimised. - The
AMC unit 330 may consist of a platform with two or three axes of rotation (i.e. roll and pitch, or roll, pitch and yaw) upon which theHyperCamera unit 100 is mounted. Commercially-available AMC orstabilisation platforms 330 include the PAV series from Leica Geosystems, including the PAV100, and the PV-15G from PV Labs, e.g. as described in U.S. Pat. No. 6,263,160 (Lewis), the contents of which are herein incorporated by reference. - Alternatively or additionally, the
AMC units 330 may comprise one or more beam-steering mechanisms in the optical path of each camera (or group of cameras), whereby the pointing direction of the cameras is corrected by beam-steering. For example, pitch and roll compensation may be provided by beam-steering mechanism(s), while yaw compensation is provided by aseparate AMC platform 330. - Angular motion compensation becomes increasingly important as the flying altitude is increased and/or the GSD is decreased.
- Motion blur due to the forward motion of the aircraft is equal to the speed of the aircraft multiplied by the exposure time of the camera. Once motion blur becomes a significant fraction of (or exceeds) the GSD it becomes useful to provide a forward motion compensation (FMC) mechanism to reduce or eliminate motion blur. FMC can be provided in a number of ways, including translating or rotating the optical axis of the camera (by moving the image sensor, or an intermediate mirror, or the camera itself), or by time delayed integration (TDI) of adjacent lines of pixels in the image sensor.
- The
computer 300 stores the GNSS position of each shot. This is used during subsequent processing of the photos to produce an accurate orthomosaic. Thecomputer 300 also stores the orientation of each shot if theIMU 306 is present, and stores the corrected orientation of each shot if anAMC unit 330 is present, based on correction information from the AMC. - The
cameras 110 and 112 are powered by abattery unit 320. Thebattery unit 320 provides a voltage higher than the voltage required by all connected components, e.g. between 24V and 28V, and the voltage requirement of each connected component is provided via a DC-DC converter 326. For example, a Nikon D800 camera requires less than 10V. Additional DC-DC converters 326 also provide appropriate voltages to power thecomputer 300, thepilot display 302, theGNSS receiver 304, and theIMU 306. For clarity these power connections are omitted inFIG. 20 . - The
battery unit 320 contains two 12V (nominal) batteries or a single 24V (nominal) battery. It contains a charging circuit that allows it to be trickle-charged from an aircraft with a suitableauxiliary power source 322, allowing it to remain charged at all times. It may also be charged on the ground from a ground power unit 324 (GPU). - The
camera interface 308 and DC-DC converters 326 may be housed in a camera control unit 310 (CCU). This may additionally include a USB interface to allow thecomputer 300 to control the camera interface. - The DC-
DC converters 326 that provide power to thecameras 110 and 112 may be located in theCCU 310 or closer to the cameras in the distribution boxes 150. - Photos captured by the HyperCamera are intended to be seamlessly stitched into an orthomosaic, and
FIG. 21 shows a photogrammetric process flow for efficiently creating an orthomosaic from multi-resolution HyperCamera photos. The process operates ondetail photos 400 captured by the detail cameras 110, andoverview photos 402 captured by theoverview cameras 112. - The process consists of four main steps: (1) features are automatically detected in each of the
photos detail photo 400 is orthorectified according to its camera pose and terrain elevation data (at step 414); and the orthorectified photos (orthophotos) are blended to form the final orthomosaic 404 (at step 416). - The elevation data (used at step 414) may be derived from the three-dimensional feature positions (refined during step 412); and/or may be derived from dense multi-view stereo matching between photos (see, for example, U.S. Pat. No. 8,331,615 (Furukawa), and U.S. Pat. No. 8,634,637 (Hirschmueller et al), the contents of both of which are herein incorporated by reference); and/or may be obtained from another source such as a LIDAR survey of the area of interest. The elevation data may be in the form of a raster or vector digital elevation model (DEM), or in the form a three-dimensional point cloud, or in the form of a three-dimensional geometric model. It may include surface features such as vegetation and buildings, or may describe the bare earth.
- The accuracy of the
orthomosaic 404 derives from the high overlap between lower-resolution overview photos 402, while detail in theorthomosaic 404 derives from the higher-resolution detail photos 400. - As an alternative, as noted above, a survey may be flown with higher overlap between the
detail photos 400, and the orthomosaic may be created from thedetail photos 400 only. - The orthomosaic is typically stored as an image pyramid, i.e. within which different (binary) zoom levels are pre-computed for fast access at any zoom level. Lower zoom levels in the pyramid are generated from higher zoom levels by low-pass filtering and subsampling, thus the entire pyramid may be generated from the detail-resolution orthomosaic. As an alternative, lower zoom levels may be generated from an orthomosaic created from the
overview photos 402, in which case theoverview photos 402 are also orthorectified and blended as described above for thedetail photos 400. - An initial estimate of the camera pose of each photo, subsequently refined by the bundle adjustment process (at step 412), is derived from the GNSS position of each photo, as well as its IMU-derived orientation, if available.
- Automatically detected ground features may be augmented with manually-identified ground points, each of which may have an accurate surveyed real-world position (and is then referred to as a ground control point).
- The V5-300 and V5-180 HyperCameras are two models from the HyperCamera range. Even greater efficiency is achieved by operating a higher altitudes with longer focal lengths, and additional HyperCamera configurations are described below.
-
FIG. 22 tabulates the GSD as a function of the flying altitude above ground (in feet) and the lens focal length from 100 mm to 600 mm (in 100 mm increments). It also tabulates the swath width (in km) of 30-degree and 45-degree fields of view as a function of altitude. In addition, it tabulates the number of detail cameras required to cover these fields of view as a function of focal length. -
FIG. 23 further tabulates the GSD as a function of focal length from 700 mm to 1200 mm. - From the tables in
FIGS. 22 and 23 , it is instructive to compare the characteristics of the V5-300, V10-600 and V15-900 HyperCamera configurations. Each configuration has approximately the same 30-degree lateral field of view, but assuming a 5 cm GSD the corresponding swath widths are 1.6 km, 3.3 km and 4.9 km respectively (corresponding to flying altitudes of 10,000, 20,000 and 30,000 feet respectively). Crucially, the relative capture efficiency is then 1×, 2× and 3× respectively (assuming the same aircraft speed). - A 30-degree V10-600 HyperCamera can be realised using two 15-degree V5-600 HyperCamera units in tandem, each unit mounted on a wedge-shaped
adapter plate 200 that induces the required 7.5-degree lateral tilt. - A 45-degree V15-600 HyperCamera can be realised using three 15-degree V5-600 HyperCamera units in tandem, the central unit mounted flat, and each of the two outside units mounted on a wedge-shaped
adapter plate 200 that induces the required 15-degree lateral tilt. - A 30-degree V15-900 HyperCamera can be realised using three 10-degree V5-900 HyperCamera units in tandem, the central unit mounted flat, and each of the two outside units mounted on a wedge-shaped
adapter plate 200 that induces the required 10-degree lateral tilt. - Alternatively or additionally to utilising a wedge shape to induce the required lateral tilt, the
adapter plate 200 can comprise one or more gross and/or fine tilt adjustment mechanisms to allow the tilt of thecamera unit 100 to be adjusted both nominally as well as to compensate for variations in the tilt of the aircraft floor. The tilt mechanisms may comprise one or more shims, one or more biasing screws, or similar. - Tilt adjustment can also be provided via the roll compensation mechanism of the angular motion compensation (AMC)
unit 330, if present, by adjusting the pre-set nominal roll of thecamera unit 100. -
FIGS. 24A, 24B and 24C show the footprints of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 12,000 feet. The dashed grid in the figures has a 1 km spacing. The figures show the same swath width but with increasing GSD. -
FIGS. 25A, 25B and 25C show three successive overlapping footprints of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 12,000 feet. In order to ensure adequate longitudinal overlap, the shot rate increases with increasing resolution. -
FIGS. 26A, 26B and 26C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 12,000 feet. -
FIG. 27 shows a front elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units, and the resultant overview and aggregate detail fields of view. The dashed line shows the division between the fields of view of the camera units. -
FIG. 28 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIG. 29 shows a side elevation of a Cessna 208 aircraft carrying a V10-600 HyperCamera comprising two V5-600 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIG. 30 shows a front elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units, and the resultant overview and aggregate detail fields of view. The dashed line shows the division between the fields of view of the camera units. -
FIG. 31 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera comprising three V5-900 camera units installed in the aircraft cabin, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIG. 32 shows a side elevation of a Cessna 208 aircraft carrying a V15-900 HyperCamera, comprising three V5-900 camera units installed in a standard Cessna 208 belly-mounted cargo pod, and the resultant overview and aggregate detail fields of view of each camera unit. -
FIGS. 33A, 33B and 33C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 24,000 feet. The dashed grid in the figures has a 1 km spacing. -
FIGS. 34A, 34B and 34C show two overlapping footprints from two adjacent flight lines of the V5-300, V10-600 and V15-900 HyperCameras respectively, at a constant flying altitude of 36,000 feet. The dashed grid in the figures has a 1 km spacing. - In general, a particular HyperCamera model can be realised using a number of identical
smaller camera units 100, each mounted at the required lateral angle. As shown in theFIGS. 27 to 32 , theindividual camera units 100 can be mounted along the longitudinal axis of the aircraft. The firing of eachcamera unit 100 can be staggered to account for its longitudinal position and the aircraft speed, or the longitudinal offsets between photos from different camera units can be accounted for during downstream processing. - The HyperCamera design is thus modular, and suited to deployment in conjunction with one or more standard camera holes.
- An oblique HyperCamera, i.e. with a pointing direction other than vertical (and typically with pointing direction of 45 degrees), can be realised using one or more
vertical camera units 100 mounted at the required oblique angle, e.g. via a tiltedadapter plate 200, or using one or moreoblique camera units 100, each comprising an array of cameras mounted at the required oblique angle within an otherwise vertically-oriented frame. - Oblique aerial photos have several uses. They can be used to produce an oblique georeferenced orthomosaic to provide a useful complement to a vertical orthomosaic. They can be used along with vertical photos during photogrammetric bundle adjustment (as described above in relation to
FIG. 21 ) to achieve higher accuracy, including higher elevation data accuracy. And they can be used to texture an elevation model or a full three-dimensional model of the area of interest, to support three-dimensional visualisation and interaction. -
FIGS. 35A and 35B show front and side elevations respectively of a Cessna 208 aircraft carrying five HyperCameras installed in the aircraft cabin: a vertical V10-600, a right oblique R10-600, a left oblique L10-600, a front oblique F10-600, and a back oblique B10-600, and the resultant overview and aggregate detail fields of view of each. - The oblique GSD (282) of a camera tilted at an angle (280) away from the vertical is given by the
EQ 9 inFIG. 44 . The vertical GSD (262) is increased by a factor of the square of the secant of the tilt angle. The first secant term is due to the greater tilt-induced viewing distance, while the second secant term is due to the tilt-induced foreshortening of the ground. A tilted camera therefore needs to have a correspondingly longer focal length to match the GSD of a vertical camera. - For a tilt of 45 degrees the vertical GSD is increased by a factor of two, implying that a 45-degree oblique camera needs to have twice the focal length of a vertical camera to match the GSD of the vertical camera. In practice, however, there is no strict requirement to match the GSDs of vertical and oblique cameras, and any suitable focal length can be used for the oblique cameras.
- If the image sensor of the camera is tilted to be parallel to the ground then the second secant term disappears. The oblique GSD (286) is then given by
EQ 10 inFIG. 44 . For a tilt of 45 degrees the vertical GSD (262) is then increased only by a factor of the square root of two, implying that a 45-degree tilted-sensor oblique camera needs to have 1.4 times the focal length of a vertical camera to match the GSD of the vertical camera. - However, the smaller GSD on horizontal surfaces is accompanied by a larger GSD on vertical surfaces such as the sides of buildings. This in turn may motivate a strategy of capturing oblique photos in multiple passes, with each pass optimised for a particular surface orientation via image sensor tilt.
- More generally, tilting the image sensor of any non-nadir camera to be (more) parallel to the ground can be used to reduce perspective foreshortening in photos captured by the camera and thus improve the GSD.
- Each 900 mm oblique HyperCamera has a 33% larger GSD than the 600 mm vertical HyperCamera. A 1200 mm oblique HyperCamera would have the same GSD as the 600 mm vertical HyperCamera. An 850 mm tilted-sensor oblique HyperCamera would also have the same GSD as the 600 mm vertical HyperCamera.
-
FIG. 37 shows a bottom view of thecamera unit 100 where the detail cameras are arranged in an alternative “X” pattern to provide more room for longer detail lenses 114, i.e. to accommodate the larger diameters of longer lenses. This arrangement also provides room for a second lower-resolution camera (i.e. with a shorter lens) which can be utilised either for a second overview camera or for a dedicated spectral camera, such as a near-infrared (NIR) camera. Including a second overview camera allows both cameras to have a longer focal length and therefore provide greater resolution and hence photogrammetric accuracy. Including a near-infrared camera allows a near-infrared band to be included in the orthomosaic, in turn supporting applications such as identifying vegetation. - As the focal length of the detail lenses 114 is increased, the size and weight of a purely refractive (dioptric) lens design becomes significant. A lighter and physically shorter lens with the same long focal length may be realised using a reflecting design, i.e. using curved mirrors for focusing, either using mirrors alone (catoptric) or in conjunction with corrective refractive elements (catadioptric). Astronomical telescopes, which are characterised by long focal lengths and large apertures, typically utilise reflecting designs. Classic modern designs include the Schmidt-Cassegrain, which combines spherical primary and secondary mirrors with a Schmidt correction plate, and the Ritchey-Chretien, which utilises hyperbolic primary and secondary mirrors.
- Reflecting detail lenses 114 can thus be used to significantly reduce the size and weight of the
camera unit 100, and this is particularly beneficial for longer focal lengths such as 900 mm and longer. -
FIG. 38 shows acamera unit 100 implemented as an array ofcameras 600, arranged in a fan, and the optical axis 602 of each camera. This figuratively represents any of the arrays of cameras in the foregoing description, detail or overview, and provides context for the following description. Eachcamera 600 consists of a camera body (e.g. 110 or 112) and a camera lens (e.g. 114 or 116). -
FIG. 39 shows acamera unit 100 implemented using asteerable camera module 610 comprising asingle camera 600 time-multiplexed via asteerable mirror 604 to produce multiple optical paths that sweep out the required angular field. This allows a single camera to be used to implement an array of cameras using time-multiplexing, and thus reduces the number of physical cameras required to implement a HyperCamera, in turn reducing the size, weight and cost of eachcamera unit 100. As an alternative to asteerable mirror 604, any suitable beam-steering mechanism may be used. -
FIG. 40 shows an alternative configuration of acamera unit 100 implemented using asteerable camera module 610 comprising asingle camera 600 time-multiplexed via asteerable mirror 604 to produce multiple optical paths. The addition of a fixedmirror 606 allows thecamera 600 to be mounted vertically, allowing thecamera unit 100 to have a smaller footprint. -
FIG. 41 shows acamera unit 100 implemented using asteerable camera module 610 comprising asingle camera 600 time-multiplexed via asteerable mirror 604 to produce multiple optical paths, with an additionalsteerable mirror 608 providing angular motion compensation (AMC). This implements the IMU-drivenAMC unit 330. As an alternative to asteerable mirror 608, any suitable beam-steering mechanism may be used. - The
steerable mirror 608 may additionally or alternatively be used to implement a forward motion compensation (FMC) mechanism. In FMC mode the mirror is smoothly pitched backwards during exposure to match the forward motion of the aircraft, and is pitched forwards again between exposures. The pitch of themirror 608 can be driven directly, or can be oscillated at an appropriate rate to provide FMC, with camera exposure synchronised with the backwards phase of the oscillation. AMC and FMC may also be provided by separate cascaded mirrors. - The multiplexing
steerable mirror 604 may be stopped at each of a set of discrete positions within the required angular field, or may be rotated continuously to sweep out the required angular field. In the latter case the motion of the mirror provides an additional source of systematic motion during camera exposure, like the forward motion of the aircraft, and this may be corrected using any of the FMC mechanisms described above, including via the steerable motion-compensation mirror 608, or via aseparate stabilisation platform 330. - Vibration of the aircraft may also be detected via the IMU, or via an additional vibration-sensing device, and vibration compensation (VC) may be provided using any of the AMC or FMC mechanisms described above, including via the steerable motion-
compensation mirror 608, or via aseparate stabilisation platform 330. - AMC, FMC and VC may also be provided via the
multiplexing mirror 604. -
FIG. 45 shows one realisation of thesteerable camera module 610 ofFIG. 41 . Thesteerable multiplexing mirror 604 is attached via a hinge to arod 632 of alinear motor 630, and thelinear motor 630 is attached via a hinge to a fixedmount 634. Themirror 604 is also attached via a hinge to themount 634, about which it rotates in response to linear motion of therod 632. Thelinear motor 630 may utilise any suitable drive mechanism, including a piezoelectric drive mechanism, e.g. as described in U.S. Pat. No. 6,800,984 (Marth), the contents of which are incorporated herein by reference, or as embodied in the Physik Instrumente (PI) M-272 linear piezomotor, which also incorporates a linear encoder for precise positioning. - The steerable motion-
compensation mirror 608 is attached to a tip-tilt stage 636 which provides rotation in two dimensions. Alternatively thestage 636 can be limited to rotation in one dimension, e.g. to only provide correction for motion of themultiplexing mirror 604. The tip-tilt stage 636 may utilise any suitable motors or actuators, including piezoelectric actuators, e.g. as embodied in the PI S-340, which may also incorporate a strain-gauge sensor for precise closed-loop control. - The
steerable camera module 610 inFIG. 45 allows the field ofview 612 of thecamera 600 to be steered across an angular range of approximately 50 degrees, depending on camera aperture and hence beam diameter. Suitably oriented, thesteerable camera module 610 may be used to provide steering across a symmetric range or an asymmetric range. An asymmetric range may form part of a larger symmetric range. -
FIG. 46A andFIG. 46B show thesteerable camera module 610 oriented to cover an asymmetric range from vertical to oblique, i.e. half of an equivalent symmetric range. InFIG. 46A themultiplexing mirror 604 is rotated to an extreme position clockwise, and the camera field ofview 612 points vertically. InFIG. 46B themultiplexing mirror 604 is rotated to an extreme position anti-clockwise, and the camera field ofview 612 points obliquely to the right. -
FIG. 47 shows acamera unit 100 implemented using a pair of steerable detail camera modules 614 and a pair of fixedoverview cameras 112. Each steerable detail camera module 614 consists of asteerable camera module 610 comprising a detail camera body 110 and detail camera lens 114. Together the two steerable detail camera modules 614 cover a continuous lateral angular field of view of approximately 100 degrees. -
FIG. 48A shows the elevation and plan of the wide-angle field of view of a 300 mm wide-angle steerable WS-300camera unit 100, with a pair of fixed overview cameras 122 (e.g. perFIG. 47 ), operating at 16,000 feet. The field of view comprises the vertical field ofview 180 and the two oblique fields ofview flight line 226.FIG. 48B shows the corresponding field of view of a 600 mm WS-600 camera unit operating at 28,000 feet. -
FIG. 48A is based on using a WS-300 camera unit comprising Nikon D800 cameras (or similar), whileFIG. 48B is based on using a WS-600camera unit 100 comprising detail cameras utilising the ON Semiconductor VITA 25K 25Mpixel CMOS image sensor (or similar), as described further below. For illustrative purposes the following figures are all based on using VITA 25K-based cameras (or similar). -
FIG. 49A shows a schematic plan view of a steerabledetail camera module 616. The detailcamera steering axis 620 corresponds to the lateral axis of the aggregate detail field ofview 180. Wherever a steerabledetail camera module 616 is referred to below, a fixed array of cameras could also be used. - The steerable
detail camera module 616 may be implemented using one or moresteerable camera modules 610, each covering part of the required field of view. For example,FIG. 49B shows thecamera module 616 implemented using a pair of half-field steerable detail camera modules 614. - The steerable
detail camera module 616 may also be implemented using one or more swinging camera modules, as described in U.S. patent application Ser. No. 10/313,727 (Partynski et al), and U.S. patent application Ser. No. 11/607,511 (Pechatnikov et al), the contents of both of which are herein incorporated by reference. -
FIG. 49C shows a schematic plan view of a steerableoverview camera module 618. The overviewcamera steering axis 622 corresponds to the lateral axis of the aggregate overview field ofview 190. The steerableoverview camera module 618 may be implemented in the same variety of ways as the steerabledetail camera module 616. -
FIG. 49D shows a schematic plan view of acamera unit 100 comprising a pair of half-field steerable detail camera modules 614 and twooverview cameras 112, i.e. corresponding to thecamera unit 100 ofFIG. 47 . -
FIG. 49E shows a schematic plan view of acamera unit 100 comprising a full-field steerabledetail camera module 616 and twooverview cameras 112, i.e. providing equivalent functionality to thecamera unit 100 ofFIG. 47 andFIG. 49D . -
FIG. 50A shows a schematic plan view of acamera unit 100 comprising a full-field steerabledetail camera module 616 and a full-field steerableoverview camera module 618. -
FIG. 50B shows a plan view of a set of vertical and four-way oblique overview and detail fields of view, corresponding to the fields of view illustrated inFIG. 35A andFIG. 35B . The five views can be captured using five fixed-array camera units 100 as described in relation toFIG. 35A andFIG. 35B , or using three of thesteerable camera units 100 shown inFIG. 50A . -
FIG. 51A shows an elevation of the field of view of a wide-angle camera unit 100, comprising the vertical field ofview 180, oblique fields ofview view -
FIG. 51B shows the elevation and plan of the continuous wide-angle field of view of a WS-600camera unit 100 operating at 28,000 feet. -
FIG. 52A shows the overlapping lateral fields of view of a wide-angle camera unit 100 from three adjacent flight lines, and illustrates how the interstitial fields of view overlap with the vertical and oblique fields of view. -
FIG. 52B shows the overlapping lateral fields of view of a wide-angle camera unit 100 from three adjacent flight lines, showing individual camera fields. -
FIG. 53A shows a plan view of the continuous wide-area fields of view of a WS-600camera unit 100, with bothoverview cameras 112 and detail cameras 110, operating at 28,000 feet, whileFIG. 53B shows the field of view of the same camera unit with just detail cameras 110. - As discussed above, the
camera unit 100 can be used to capture detail photos only, with higher overlap, to allow the creation of an orthomosaic with higher spatial accuracy, and elevation data with higher resolution and accuracy. - Longitudinal overlap, i.e. in the
flight direction 220, is typically used to capture redundant photos of ground points with stereo disparity. However, when operating at higher altitudes, with longer focal lengths, the longitudinal field of view angle of a detail camera 110 is sufficiently small that little stereo disparity is present in longitudinally overlapping photos. - An alternative approach is to capture overlapping imagery using a second camera module that is tilted about the lateral axis.
FIG. 53C shows acamera unit 100 comprising two full-field steerabledetail camera modules 616 for this purpose. Thefirst camera module 616 a captures imagery vertically with respect to the lateral axis, while thesecond camera module 616 b captures imagery at e.g. a 20-degree tilt about the lateral axis. The tilt is chosen to maximise the observability of ground points without compromising feature-matching accuracy. -
FIG. 53D shows a plan view of the resulting un-tilted field ofview 650 a and tilted field ofview 650 b of such a double-field WS-600camera unit 100 operating at 28,000 feet. - The traditional approach to capturing four-way obliques, as exemplified in
FIG. 50B , is to capture obliques in the two lateral directions and the two longitudinal directions simultaneously. However, this results in a more consistent viewing angle in the lateral obliques than the longitudinal obliques, since the longitudinal obliques have a wider horizontal field of view to accommodate a practical flight-line spacing. - An alternative approach is to perform two orthogonal passes over the area of interest, and capture two high-quality lateral obliques during each pass.
- Another approach is to capture all obliques with the same narrow horizontal field of view during a single pass, and reduce the flight-line spacing as necessary. When capturing four obliques with the same narrow horizontal field simultaneously, the optimum approach (i.e. the approach that maximises the flight-line spacing and hence the survey efficiency) is to arrange the four oblique directions at 45 degrees to the flight line.
-
FIG. 54 shows a compact WS-300HyperCamera unit 100 comprising four half-field steerable detail camera modules 614 arranged in an “X” configuration for the purposes of optimally capturing four high-quality oblique views in a single pass. The WS-300camera unit 100 also optionally comprises a pair of fixed overview cameras that may be used for dual-resolution imaging or for capturing additional spectral bands such as NIR. - If obliques are desired for the four cardinal directions (north, south, east and west), then the flight lines of the flight plan can be oriented at 45 degrees to the cardinal directions.
-
FIG. 55A shows a plan view of the complete field ofview 660 of an X-configuration WS-300 camera unit operating at 16,000 feet, consisting of two orthogonal fields ofview FIG. 55B shows the complete field ofview 660 of an X-configuration WS-600 camera unit operating at 28,000 feet. The two overlappingfields 650 provide sufficient overlap and observability that there is no need for additional longitudinal overlap between shots, i.e. beyond the minimum required to avoid longitudinal gaps. -
FIG. 56A shows a schematic plan view of the compactX-configuration camera unit 100 ofFIG. 54 , whileFIG. 56B shows less compactX-configuration camera unit 100 comprising two full-field steerabledetail camera modules 616 arranged orthogonally. These can be deployed over two camera holes if necessary. - In order to avoid the reduction in flight-line spacing required by the X-configuration, an additional
detail camera module 616 can be used to capture a field perpendicular to theflight direction 220. This additional field may then be used to create the vertical orthomosaic, while the two orthogonal X-configuration fields are used to create oblique orthomosaics (as well as to increase the observability of all ground points). - If the flight-line spacing is not reduced to accommodate the rotated X-configuration fields, then the oblique fields of view within each wide-angle field are correspondingly larger, resulting in slightly lower-quality obliques.
-
FIG. 57A shows a plan view of the fields ofview camera unit 100 operating at 28,000 feet. -
FIG. 57B shows a schematic plan view of acorresponding camera unit 100 comprising two pairs of half-field steerable detail camera modules 614 arranged in an X configuration, as well as a full-field steerabledetail camera module 616 for the perpendicular vertical field. -
FIG. 58A shows a plan view of threesuccessive shots camera unit 100 operating at 28,000 feet. -
FIG. 58B shows a plan view ofshots adjacent flight lines camera unit 100 operating at 28,000 feet. - The longitudinal footprint (b, 672) of an aerial camera is determined by its longitudinal field of view angle (delta, 670) and its altitude (a, 254), per
EQ 11 inFIG. 59 . The longitudinal shot spacing (d, 676) is determined by the longitudinal footprint (b, 672) and the desired longitudinal overlap (c, 674), perEQ 12 inFIG. 59 . The shot rate (r) is determined by the shot spacing (d, 676) and the velocity (v, 678) of the aircraft, perEQ 13 inFIG. 59 . - As described in relation to
FIG. 43 above, the lateral field of view angle (gamma, 270) of an array of cameras is a function of the number of cameras (N), the lateral field of view angle (beta, 250) of each camera, and the lateral angular overlap (omega, 272) between adjacent cameras. Likewise, the number of cameras (N) required to cover a particular field of view angle is determined by the target lateral field of view angle (gamma, 270), the lateral field of view (beta, 250) of each camera, and the angular overlap (omega, 272) between adjacent cameras, perEQ 14 inFIG. 60 . - If the required number of cameras (N) is realised virtually by time-multiplexing a smaller number of physical cameras (C), then the frame rate (q) of each camera is determined by the required number of cameras (N), the number of physical cameras (C), the shot rate (r), and the duty cycle (u) of the time-multiplexing beam-steering mechanism, per
EQ 15 inFIG. 60 . - The usable lateral field of view angle (phi, 680) is the field of view angle within which imagery is sufficiently consistent to be used to create a vertical orthomosaic. This may be as low as 20 degrees, or as high as 45 degrees, depending on the required quality of the resultant orthomosaic. This in turn dictates the flight-line spacing (e, 682), per EQ 16 in
FIG. 60 . - In general, the flight-line spacing (e, 682) should be minimised to maximise the observability of ground points and avoid unobserved ground points (e.g. in urban canyons). This is particularly important for generating high-quality digital surface models, true orthomosaics, and textured 3D surfaces.
- The rate at which ground area is captured by an aerial camera system is determined by the flight-line spacing (e, 682) and the velocity (v, 678) of the aircraft, per
EQ 17 inFIG. 60 . - When the field of view is rotated by a non-zero axis rotation angle (lambda, 686), i.e. relative to a nominal line perpendicular to the
flight line 226, various capture parameters change. The resultant shot spacing (d(lambda), 690) is larger, perEQ 18 inFIG. 61 . The resultant shot rate (r(lambda)) is smaller, perEQ 19 inFIG. 61 . The resultant flight-line spacing (e(lambda), 692) is smaller, perEQ 20 inFIG. 61 . The resultant frame rate (q(lambda)) is smaller, perEQ 21 inFIG. 61 . And the resultant area capture rate (R(lambda)) is smaller, perEQ 22 inFIG. 61 . -
FIG. 62 shows an electronic and mechanical block diagram of thecamera unit 100 with one or moresteerable camera modules 610, and represents a variation of the block diagram inFIG. 20 . Blocks have the same functions as inFIG. 20 , except as described below. Power is omitted for clarity. - The
computer 300 controls one or moresteerable camera modules 610, described further below. Real-time imagery from the camera modules is compressed by one or more image input/output modules 700 and stored in one or moreimage storage devices 702. The image input/output module(s) 700 may also provide the computer with access to the imagery, both for real-time monitoring and for later retrieval. The image storage device(s) 702 comprise removable media to allow dispatch from the field to a processing facility. The input/output module(s) 700 and image storage device(s) 702 may be incorporated in each individual camera 600 (e.g. when thecamera 600 is a DSLR such as a Nikon D800), or may be separate. - Image compression can utilise any suitable scheme, including JPEG or a similar DCT-based compression scheme, or JPEG2000 or a similar wavelet-based compression scheme, implemented in hardware or software.
- The
computer 300 may monitor the quality of photos captured by the cameras and direct the pilot (or auto-pilot) to re-fly affected parts of the flight plan. Thecomputer 300 may monitor the photos to detect poor focus, poor exposure, the presence of cloud or other unwanted objects such as other aircraft, etc. - The
pilot input device 704 allow the pilot to provide input to the computer, including to select flight plans, acknowledge alerts, navigate imagery, perform quality assurance, etc. It may comprise any suitable device such as a touch screen (in conjunction with the pilot display 302), tablet, keyboard, etc. - The
computer 300 may control the aircraft auto-pilot 706 to direct the aircraft along, and between, each flight line in the flight plan. -
FIG. 63 shows an electronic and mechanical block diagram of thesteerable camera module 610. Mechanical couplings are shown dashed. - Shot-level timing is maintained by the
computer 300, and is provided by thecomputer 300 to amicrocontroller 710 via signals to the microcontroller'scontrol interface 712. Thecontrol interface 712 may comprise a serial, USB or Ethernet interface, or any other suitable interface. - Frame-level timing is maintained by the
microcontroller 710, and is provided by themicrocontroller 710 to thecamera 600 via the camera'scontrol interface 714. Themicrocontroller 710 provides frame-level control of themultiplexing mirror 604 via signals to itscontroller 640, and the motion-compensation (MC)mirror 608 via signals to itscontroller 642. - The
microcontroller 710 reports the lateral steering angle of each photo to thecomputer 300, allowing thecomputer 300 to include the lateral steering angle in the initial estimate of the camera pose of each photo. This is subsequently refined by the bundle adjustment process (atstep 412 inFIG. 21 ), as described above. Photo metadata, including exposure, timing and pose, may be stored by thecomputer 300 and/or stored alongside each photo on the image storage device(s) 702. - The
camera 600 produces a stream of images on itsimage stream interface 716. Theimage stream interface 716 may comprise a serial, USB or Ethernet interface, a high-rate interface such as CoaXPress (CXP) or CameraLink (CL), utilising multiple channels if necessary, or any other suitable interface. If theimage stream interface 716 comprises a CXP interface then it may also provide thecamera control interface 714. - The
multiplexing mirror 604 is attached to agoniometer stage 638 which is controlled by themicrocontroller 710 via acontroller 640. Anoptional position sensor 644 supports closed-loop control. Thegoniometer stage 638 may utilise a linear motor as described in relation toFIG. 45 above, or any other suitable actuator or motor including a linear actuator or motor (such as a voice coil actuator) and a rotating motor (such as a stepper motor, DC motor or brushless motor) coupled via a gearbox or worm drive. Thecontroller 640 may be any suitable controller, e.g. a PI C-867 motion controller when used in conjunction with a PI M-272 piezomotor. Theposition sensor 644 may be of any suitable type, including an optical sensor, capacitive sensor, strain gauge sensor, and eddy-current sensor. Thegoniometer stage 638 may provide two axes of rotation if the multiplexing mirror is also used to provide motion compensation. - The motion-compensation (MC)
mirror 608 is attached to a tip-tilt stage 636 which is controlled by themicrocontroller 710 via acontroller 642. Anoptional position sensor 646 supports closed-loop control. The tip-tilt stage 636 may utilise piezoelectric actuators as described in relation toFIG. 45 above, or any other suitable actuators or motors including linear actuators or motors (such as voice coil actuators) and rotating motors (such as stepper motors, DC motors or brushless motors) coupled via a gearbox or worm drive. Thecontroller 642 may be any suitable controller, e.g. a PI E-616 controller when used in conjunction with a PI S-340 tip-tilt stage. Theposition sensor 646 may be of any suitable type, including an optical sensor, capacitive sensor, strain gauge sensor, and eddy-current sensor. -
FIG. 64 shows a tabulation, for an X-configurationsteerable camera unit 100 comprising 36Mpixel cameras, of shot rate (r), frame rate (q) and area capture rate (R) as a function of altitude (a, 254), aircraft speed (v, 678) and usable field of view angle (phi, 680). The tabulation is based on a constant GSD of 7 cm (and hence a focal length that increases with altitude). The daily capture rate is based on six survey hours per day. - Up to 18,000 feet the required frame rate of the
camera 600 is below the 4 Hz frame rate of the 36Mpixel Nikon D800 DSLR, and up 22,000 feet the frame rate is below the 5 Hz frame rate of the Nikon D810 DSLR. For higher altitudes (or higher speeds at lower altitudes), a camera with a higher frame rate can be used. - A variety of high-speed machine vision cameras and high-resolution video cameras are available. Many such cameras utilise CMOS image sensors with a rolling electronic shutter. This leads to geometric distortion in the presence of motion, and makes them incompatible with most motion compensation mechanisms. A smaller number of cameras utilise CMOS image sensors with a global electronic shutter, making them suitable for use in the
steerable camera module 610. Cameras which utilise CCD image sensors are also suitable, since CCDs intrinsically have a global electronic shutter. Cameras with external (e.g. mechanical shutters) can also be suitable. - Several medium-resolution and high-resolution global shutter CMOS image sensors are widely used, including the ON Semiconductor 12Mpixel KAC-12040, 12Mpixel VITA 12K, 16Mpixel VITA 16K, and 25Mpixel VITA 25K; and the CMOSIS 12Mpixel CMV12000 and 20Mpixel CMV20000.
- The VITA 25K has a pixel pitch of 4.5 microns and supports frame rates up to 53 Hz. The CMV20000 has a pixel pitch of 6.4 microns and support frame rates up to 30 Hz (at 12 bits/pixel). The CMV12000 has a pixel pitch of 5.5 microns and support frame rates up to 180 Hz (at 12 bits/pixel) and 300 Hz (at 10 bits/pixel).
- Machine vision cameras utilising the 25Mpixel VITA 25K include the 32 fps Adimec S-25A30, the 53 fps CMOSVision Condor, and the 53 fps ISVI IC-X25CXP.
- Machine vision cameras utilising the 20Mpixel CMV20000 include the 30 fps Jai SP-20000C-PMCL, and the 30 fps Ximea CB200CG-CM.
- Machine vision cameras utilising the 12Mpixel CMV12000 include the 66 fps Adimec Q-12A65, the 62 fps Basler Beat beA4000-62 kc, and the 149 fps IO Industries Flare 12M180-CX.
-
FIG. 65 shows a tabulation, for a constant 5 cm GSD and a 25Mpixel X-configuration steerable camera unit 100 (e.g. assuming acamera 600 utilising the ON Semiconductor VITA 25K image sensor), of shot rate, frame rate and area capture rate. The CMOSVision Condor and the ISVI IC-X25CXP support 5 cm GSD capture up to 54,000 feet. -
FIG. 66 shows a tabulation, for a constant 5 cm GSD and a 12Mpixel X-configuration steerable camera unit 100 (e.g. assuming acamera 600 utilising the CMOSIS CMV12000 image sensor), of shot rate, frame rate and area capture rate. The Flare 12M180-CX supports 5 cm GSD capture beyond 60,000 feet. - The observability of ground points and the quality of 3D textures can be further increased by increasing the number of viewing directions beyond the four of the X-configuration.
-
FIG. 67A shows a plan view of the fields of view of a three-field WS-600camera unit 100 operating at 28,000 feet.FIG. 67B shows a plan view of the fields of view of a four-field WS-600camera unit 100 operating at 28,000 feet.FIG. 68A shows a plan view of the fields of view of a five-field WS-600camera unit 100 operating at 28,000 feet. - These fields of view can also be captured using
multiple camera units 100, with eachcamera unit 100 deployed over a separate camera hole. For example, four-field capture can be performed using twoX-configuration camera units 100 rotated 45 degrees relative to each other. - In general, for an M-
field camera unit 100, the fields should be evenly spaced angularly, i.e. 180/M degrees apart, and should be arranged symmetrically with respect to the flight direction, i.e. one field should be rotated 90/M degrees from the flight direction. - If the absolute rotation of the fields relative to the flight direction varies (as in
FIG. 67A ,FIG. 67B andFIG. 68A ), then the minimum adjusted flight-line spacing, given byEQ 20 inFIG. 61 , is determined by the field with the smallest absolute rotation difference, while the maximum adjusted frame rate, given byEQ 21 inFIG. 61 , is determined by the field with the largest absolute rotation difference. - Successive frames captured by a time-multiplexed
camera 600 in asteerable camera unit 610 are necessarily displaced in the direction of flight due to the forward motion of the aircraft, an effect ignored for clarity in previous figures. Adjacent frames are therefore slightly staggered, as illustrated inFIG. 68B for a single-field dual-camera WS-600camera unit 100 operating at 28,000 feet. In the figure the twocameras 600 of thecamera unit 100 are simultaneously steered from the centre outwards. The induced stagger has no effect on the degree of overlap within and between flight lines since the effect is systematic. -
FIG. 69 shows a plan view of the systematic motion vectors to be corrected for each of the fourcameras 600 of anX-configuration camera unit 100. If thesteerable multiplexing mirror 604 is moved continuously then itsmotion vector 720 is dominant. The motion of the aircraft adds a smaller constantforward motion vector 722, producing the resultant motion vector 724 (shown dashed). - If linear angular motion of the
multiplexing mirror 604 is corrected by oscillating the motion-compensation mirror 608, then the average angular velocity of the motion-compensation mirror 608 during the exposure period is matched to the constant angular velocity of themultiplexing mirror 604, and the phase of the oscillation is aligned with the exposure interval so that the maximum correcting angular velocity is aligned with the centre of the exposure interval. -
FIG. 70 shows angular position (y, 730) as a function of time (t, 732) for themultiplexing mirror 604 and the motion-compensation mirror 608 when the latter is used to correct the motion of the former.Graph 734 shows the linear motion of themultiplexing mirror 604, while thegraphs compensation mirror 608, each with its maximum velocity aligned with the centre of theexposure interval 738. - The angular position (y, 730) of the motion-
compensation mirror 608 is a function of the amplitude (A) and frequency (f) of the oscillation, and time (t, 732), perEQ 23 inFIG. 70 . The angular velocity (dy/dt) of the motion-compensation mirror 608 is the derivative ofEQ 23, per EQ 24, and it is at a maximum when the cosine is one, per EQ 25. The average angular velocity of themirror 608 over theexposure interval 738 is the integral of EQ 24, divided by the exposure interval, perEQ 26. - Since the angular velocity (dy/dt) is proportional to both the amplitude (A) and the frequency (f) of the oscillation, per EQ 24, the angular velocity of the motion-
compensation mirror 608 can be matched to the angular velocity of themultiplexing mirror 604 by varying both its amplitude and its frequency. To maintain phase alignment, however, the frequency must be an integer multiple of the camera frame rate.Sinusoid 736 a has twice the amplitude and half the frequency ofsinusoid 736 b, hence they have the same angular velocity where they're in phase (e.g. at t=0 inFIG. 70 ). - As shown by
graph 734 inFIG. 71 , the linear motion of themultiplexing mirror 604 may be modulated by a sinusoid, periodically slowing and even reversing it. This allows the angular velocity of the motion-compensation mirror 608 to be matched to the angular velocity of themultiplexing mirror 604 via a low-amplitude sinusoid 736, with a slight phase delay to account for the asymmetry of the motion of themultiplexing mirror 604. Once again the average velocities are matched over the exposure interval. -
FIG. 72A shows an alternative realisation of thesteerable camera module 610, wherein themultiplexing mirror 604 is mounted at 45 degrees to the optical axis and configured to spin about the optical axis, thereby allowing the field ofview 612 of thecamera 600 to be steered from side to side. Themirror 604 is coupled to and spun by arotating motor 648. Themotor 648 may be any suitable rotating motor (such as a stepper motor, DC motor or brushless motor). Alternatively the mirror may be coupled to a linear actuator or linear motor via a gear. - As usual, motion-compensation mirror 608 (or fixed
mirror 606 in its place) provides the function of bending the optical axis to allow more convenient vertical mounting of thecamera 600. As shown, it bends the optical axis approximately 90 degrees, from vertical to horizontal. Motion-compensation mirror 608 may also provide the motion-compensation functions described earlier (in conjunction with motor oractuator 636, not shown). -
FIG. 72A showssteerable camera module 610 with themultiplexing mirror 604 positioned for vertical imaging, whileFIG. 72B shows thesteerable camera module 610 with themultiplexing mirror 604 positioned for imaging at an oblique angle. -
FIG. 73A andFIG. 73B show a plan view and back elevation, respectively, of thesteerable camera module 610 ofFIG. 72A , both with themultiplexing mirror 604 positioned for imaging at an oblique angle. - The
steerable camera module 610 ofFIG. 72A supports a wider steering range than thesteerable camera module 610 ofFIG. 45 , thus allowing the design of more compact and/or lessexpensive camera units 100. -
FIG. 74 shows a plan view of a WS-300HyperCamera unit 100 comprising two full-field steerabledetail camera modules 616 utilising the spinning multiplexing mirror design ofFIG. 72A , and one fixedoverview camera 112. Thiscamera unit 100 supports the same complete detail field ofview 660 as thecamera unit 100 ofFIG. 54 , but using two detail cameras instead of four. Each of the two detail cameras operate at a frame rate twice that of each of the four detail cameras. The field of view 160 of eachcamera module 616 is shown at its two oblique extremes (R=right, L=left). -
FIG. 75A shows a plan view of the complete field ofview 660 of an alternative X-configuration WS-600camera unit 100 utilising the spinning design and operating at 28,000 feet. The complete field ofview 660 is equivalent to the field ofview 660 ofFIG. 55B . While thespinning multiplexing mirror 604 induces rotation in individual camera fields, as shown inFIG. 75A , successive shots provide complete coverage, as shown inFIG. 75B . -
FIG. 76A shows a double-fieldsteerable camera module 624 comprising two full-fieldsteerable camera modules 610 utilising the spinning design, but sharing asingle camera 600 via aswitching mirror 656. Themirror 656 also provides the functions of the motion-compensation mirror 608 or fixedmirror 606. The field ofview 612 associated with each steering axis is shown at nadir and its two oblique extremes (N=nadir, R=right, L=left). - As shown in more detail in
FIG. 76B , when theswitching mirror 656 is facing in a first direction it directs the field of view of thecamera 600 to afirst multiplexing mirror 604 a, and when it is facing in a second direction it directs the field of view of thecamera 600 to asecond multiplexing mirror 604 b mounted at right angles to the first. Theswitching mirror 656 is coupled to and rotated by arotating motor 658. Themotor 658 may be any suitable rotating motor (such as a stepper motor, DC motor or brushless motor). Alternatively themirror 656 may be coupled to a linear actuator or motor via a gear. -
FIG. 77 shows a plan view of a WS-300HyperCamera unit 100 comprising a double-field steerabledetail camera module 626 utilising the spinning design, and one fixedoverview camera 112. Thiscamera unit 100 supports the same complete detail field ofview 660 as thecamera unit 100 ofFIG. 54 , but using one detail camera instead of four. The single detail camera operates at a frame rate quadruple that of each of the four detail cameras. The field of view 160 associated with each steering axis is shown at its two oblique extremes (R=right, L=left). -
FIG. 78A shows a plan view of the fields of view of a four-field WS-600camera unit 100 using the spinning design, whileFIG. 78B shows a plan view of the fields of view of a five-field WS-600camera unit 100 using the spinning design, both operating at 28,000 feet. - These fields can be captured using a
camera unit 100 comprising a suitable number of full-fielddetail camera modules 616 utilising the spinning design, or double-fielddetail camera modules 626 utilising the spinning design. As before, these fields of view can also be captured usingmultiple camera units 100 utilising the spinning design, with eachcamera unit 100 deployed over a separate camera hole. For example, four-field capture can be performed using twoX-configuration camera units 100 rotated 45 degrees relative to each other. - These fields can also be captured utilising a single steerable
detail camera module 616 utilising the spinning design, by rotating themirrors 604 and 608 (or 606) (and their motors) in the horizontal plane about the vertical axis of the detail camera 110. The camera 110 is ideally (but not crucially) kept fixed. The mirrors can be rotated successively to each of a set of discrete angular positions, followed by the capture of a full field at each such position. Photos can be captured with arbitrary density purely as a function of the number of discrete angular positions, subject to the maximum frame rate of the camera and the maximum spin rate of themultiplexing mirror 604. Alternatively, the mirrors can be rotated about the camera axis continuously to capture photos in a spiral pattern, with suitable motion compensation. -
FIG. 79 shows a plan view of a WS-300HyperCamera unit 100 comprising four half-field steerable detail camera modules 614 utilising the spinning multiplexing mirror design ofFIG. 72A , and one fixedoverview camera 112. Thiscamera unit 100 supports the same complete detail field ofview 660 as thecamera unit 100 ofFIG. 54 . -
FIG. 79 shows the detail fields of view 160 pointing vertically, whileFIG. 80 shows the detail fields of view 160 at their oblique extremes. -
FIG. 81 shows a perspective view of the WS-300HyperCamera unit 100 comprising four half-field steerable detail camera modules 614, and illustrates the detail fields of views 160 pointing vertically as well as at their oblique extremes. - The present invention has been described with reference to a number of preferred embodiments. It will be appreciated by someone of ordinary skill in the art that a number of alternative embodiments of the present invention exist, and that the scope of the invention is only limited by the attached claims.
Claims (13)
1-20. (canceled)
21. A method of imaging a survey area, the method comprising:
(a) flying an aircraft above the survey area along a set of flight lines;
(b) capturing, during flight and via a plurality of detail cameras carried by the aircraft, a set of detail images, each detail image depicting a part of the survey area; and
(c) steering, during flight, a pointing direction of each detail camera along a corresponding steering axis, thereby to capture a subset of the detail images along the steering axis;
wherein each steering axis is angled at substantially 45 degrees to a direction of flight, and at least two of the steering axes are angled at substantially 90 degrees to each other.
22. The method of claim 21 , wherein the pointing direction during steering varies from substantially vertical to at least 45 degrees from vertical.
23. The method of claim 21 , wherein each image in the subset of images at least partially overlaps its successor image.
24. The method of claim 21 , wherein each subset of images at least partially overlaps its successor subset of images.
25. The method of claim 21 , wherein the direction of flight is angled at approximately 45 degrees to a cardinal direction.
26. The method of claim 21 further comprising utilizing an angular motion compensation mechanism to correct the effect, on each detail camera, of angular motion of the aircraft, thereby to ensure the detail cameras point in a consistent direction over time.
27. The method of claim 21 further comprising utilizing a forward motion compensation mechanism to correct the effect, on each detail camera, of forward motion of the aircraft, thereby to reduce motion blur in the images captured by the detail cameras.
28. The method of claim 21 further comprising:
identifying, in a plurality of the detail images, common features corresponding to common ground points;
estimating, via bundle adjustment and using locations of the identified ground points in the detail images, a camera pose associated with each detail image and a three-dimensional position associated with each ground point;
orthorectifying each of a plurality of the detail images at least partly in reliance on its corresponding camera pose and at least partly in reliance on some of the estimated three-dimensional ground point positions; and
merging the orthorectified detail images to create an orthomosaic of the survey area.
29. The method of claim 21 further comprising capturing, during flight and via at least one overview camera carried by the aircraft, a sequence of overview images, the focal length of the overview camera shorter than the focal length of each detail camera.
30. The method of claim 29 further comprising:
identifying, in a plurality of the overview images and detail images, common features corresponding to common ground points;
estimating, via bundle adjustment and using locations of the identified ground points in the overview images and detail images, a camera pose associated with each detail image and a three-dimensional position associated with each ground point;
orthorectifying each of a plurality of the detail images at least partly in reliance on its corresponding camera pose and at least partly in reliance on some of the estimated three-dimensional ground point positions; and
merging the orthorectified detail images to create an orthomosaic of the survey area.
31. The method of claim 21 , wherein the number of detail cameras is two.
32. The method of claim 21 , wherein the number of detail cameras is four.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/462,212 US20170195569A1 (en) | 2014-06-20 | 2017-03-17 | Wide-area aerial camera systems |
Applications Claiming Priority (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/310,523 US9046759B1 (en) | 2014-06-20 | 2014-06-20 | Compact multi-resolution aerial camera system |
US14/478,380 US9052571B1 (en) | 2014-06-20 | 2014-09-05 | Wide-area aerial camera systems |
US14/687,544 US9440750B2 (en) | 2014-06-20 | 2015-04-15 | Wide-area aerial camera systems |
US14/735,504 US9185290B1 (en) | 2014-06-20 | 2015-06-10 | Wide-area aerial camera systems |
US14/928,542 US9462185B2 (en) | 2014-06-20 | 2015-10-30 | Wide-area aerial camera systems |
US15/009,181 US9641736B2 (en) | 2014-06-20 | 2016-01-28 | Wide-area aerial camera systems |
US15/462,212 US20170195569A1 (en) | 2014-06-20 | 2017-03-17 | Wide-area aerial camera systems |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/009,181 Continuation US9641736B2 (en) | 2014-06-20 | 2016-01-28 | Wide-area aerial camera systems |
Publications (1)
Publication Number | Publication Date |
---|---|
US20170195569A1 true US20170195569A1 (en) | 2017-07-06 |
Family
ID=56011489
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/009,181 Active US9641736B2 (en) | 2014-06-20 | 2016-01-28 | Wide-area aerial camera systems |
US15/462,212 Abandoned US20170195569A1 (en) | 2014-06-20 | 2017-03-17 | Wide-area aerial camera systems |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/009,181 Active US9641736B2 (en) | 2014-06-20 | 2016-01-28 | Wide-area aerial camera systems |
Country Status (1)
Country | Link |
---|---|
US (2) | US9641736B2 (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170324932A1 (en) * | 2015-09-30 | 2017-11-09 | Cisco Technology, Inc. | Camera system for video conference endpoints |
CN107704681A (en) * | 2017-09-30 | 2018-02-16 | 江西洪都航空工业集团有限责任公司 | A kind of computational methods of aircraft floor field range |
CN110022444A (en) * | 2017-12-06 | 2019-07-16 | 凌通科技股份有限公司 | The panorama photographic method of unmanned flight's machine and the unmanned flight's machine for using it |
US10378895B2 (en) * | 2014-08-29 | 2019-08-13 | Spookfish Innovagtions PTY LTD | Aerial survey image capture system |
WO2020136632A1 (en) * | 2018-12-27 | 2020-07-02 | Simplex Mapping Solutions Sb Ltd. | A compact interval sweeping imaging system and method |
WO2021126941A1 (en) * | 2019-12-18 | 2021-06-24 | Bae Systems Information And Electronic Systems Integration Inc. | Method for co-locating dissimilar optical systems in a single aperture |
US20210325182A1 (en) * | 2019-03-27 | 2021-10-21 | Chengdu Rainpoo Technology Co., Ltd. | Aerial survey method and apparatus capable of eliminating redundant aerial photos |
WO2022086976A1 (en) * | 2020-10-19 | 2022-04-28 | Pictometry International Corp. | Variable focal length multi-camera aerial imaging system and method |
US11474363B2 (en) | 2019-12-18 | 2022-10-18 | Bae Systems Information And Electronic Systems Integration Inc. | Method for co-locating dissimilar optical systems in a single aperture |
US11722776B2 (en) | 2021-06-28 | 2023-08-08 | nearmap australia pty ltd. | Hyper camera with shared mirror |
WO2023170472A1 (en) * | 2022-03-07 | 2023-09-14 | Stuart Nixon | Airborne and spaceborne imaging survey platform |
US12015853B2 (en) | 2021-06-28 | 2024-06-18 | nearmap australia pty ltd. | Hyper camera with shared mirror |
Families Citing this family (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9688403B2 (en) | 2014-05-20 | 2017-06-27 | Infatics, Inc. | Method for adaptive mission execution on an unmanned aerial vehicle |
US9641736B2 (en) * | 2014-06-20 | 2017-05-02 | nearmap australia pty ltd. | Wide-area aerial camera systems |
US9903811B2 (en) * | 2014-08-12 | 2018-02-27 | The Boeing Company | Multi-spectral reflectometer |
US20170081026A1 (en) | 2014-09-03 | 2017-03-23 | Infatics, Inc. (DBA DroneDeploy) | System and methods for hosting missions with unmanned aerial vehicles |
CN106796276A (en) | 2014-10-08 | 2017-05-31 | 斯布克费舍创新私人有限公司 | Aerocamera system |
IL236606B (en) | 2015-01-11 | 2020-09-30 | Gornik Amihay | Systems and methods for agricultural monitoring |
CN106017862B (en) * | 2016-05-10 | 2018-10-30 | 湖北航天技术研究院总体设计所 | A kind of integrated airborne optical gondola test system |
CN106019800B (en) * | 2016-07-26 | 2018-10-09 | 广州市红鹏直升机遥感科技有限公司 | Staged double-axle rotation inclined camera |
WO2018073878A1 (en) * | 2016-10-17 | 2018-04-26 | エスゼット ディージェイアイ テクノロジー カンパニー リミテッド | Three-dimensional-shape estimation method, three-dimensional-shape estimation system, flying body, program, and recording medium |
KR102609464B1 (en) * | 2016-10-18 | 2023-12-05 | 삼성전자주식회사 | The Electronic Device Shooting Image |
EP3529978B1 (en) * | 2016-10-20 | 2022-05-18 | Spookfish Innovations Pty Ltd. | An image synthesis system |
GB2561329A (en) * | 2016-12-05 | 2018-10-17 | Gaist Solutions Ltd | Method and system for creating images |
US10621780B2 (en) | 2017-02-02 | 2020-04-14 | Infatics, Inc. | System and methods for improved aerial mapping with aerial vehicles |
CN108151711B (en) * | 2017-03-01 | 2018-10-16 | 哈尔滨工业大学 | A kind of optical satellite ring sweeps ultra-wide imaging method |
US11082566B2 (en) * | 2017-06-20 | 2021-08-03 | Omnivision Technologies, Inc. | Multiple camera calibration chart |
US10627232B2 (en) | 2017-06-27 | 2020-04-21 | Infatics, Inc. | Method and system for aerial image processing |
DE102017114962B4 (en) | 2017-07-05 | 2019-08-29 | Dallmeier Electronic Gmbh & Co. Kg | Multi-sensor camera |
WO2020051208A1 (en) * | 2018-09-04 | 2020-03-12 | Chosid Jessica | Method for obtaining photogrammetric data using a layered approach |
GB2578735A (en) * | 2018-11-05 | 2020-05-27 | Thales Holdings Uk Plc | Camera system |
US20200191568A1 (en) * | 2018-12-17 | 2020-06-18 | Paul Lapstun | Multi-View Aerial Imaging |
US10846558B2 (en) | 2018-12-17 | 2020-11-24 | Vergent Research Pty Ltd | Multi-view scanning aerial cameras |
US11190751B2 (en) | 2018-12-17 | 2021-11-30 | Nearmap Australia Pty Ltd | Multiplexed multi-view scanning aerial cameras |
US10848654B2 (en) | 2018-12-17 | 2020-11-24 | Vergent Research Pty Ltd | Oblique scanning aerial cameras |
CN109612440B (en) * | 2019-02-22 | 2021-01-05 | 安徽省川佰科技有限公司 | Calibration object profile analysis method based on oblique photography |
US10997747B2 (en) | 2019-05-09 | 2021-05-04 | Trimble Inc. | Target positioning with bundle adjustment |
US11002541B2 (en) * | 2019-07-23 | 2021-05-11 | Trimble Inc. | Target positioning with electronic distance measuring and bundle adjustment |
CN114270815A (en) * | 2019-06-06 | 2022-04-01 | 弗劳恩霍夫应用研究促进协会 | Multi-channel imaging device and device having multi-aperture imaging device |
US11036983B1 (en) * | 2019-06-20 | 2021-06-15 | Sentera, Inc. | Aerial imaging wide and narrow stitching |
CN111457897B (en) * | 2020-04-23 | 2024-02-23 | 中国科学院上海技术物理研究所 | Swing-scanning type multi-view aviation oblique photography camera and imaging method |
US11284008B2 (en) | 2020-10-30 | 2022-03-22 | Nearmap Australia Pty Ltd | Multiplexed multi-view scanning aerial cameras |
WO2022135632A1 (en) * | 2020-12-22 | 2022-06-30 | Sam - Dimension Ug (Haftungsbeschränkt) | Apparatus and method for area mapping |
WO2023149963A1 (en) | 2022-02-01 | 2023-08-10 | Landscan Llc | Systems and methods for multispectral landscape mapping |
CN116952838A (en) * | 2023-07-19 | 2023-10-27 | 西南林业大学 | Panel surface image acquisition device based on machine vision |
Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3212420A (en) * | 1962-11-23 | 1965-10-19 | Dynasciences Corp | Image motion compensator |
US3580150A (en) * | 1968-04-15 | 1971-05-25 | Itek Corp | Panoramic camera with forward image motion compensation by optical rotation of image |
US6473119B1 (en) * | 1997-04-08 | 2002-10-29 | Carl-Zeiss-Stiftung | Photogrammetic camera |
US20090256909A1 (en) * | 2008-04-11 | 2009-10-15 | Nixon Stuart | Systems and methods of capturing large area images in detail including cascaded cameras and/or calibration features |
US20100026822A1 (en) * | 2008-07-31 | 2010-02-04 | Itt Manufacturing Enterprises, Inc. | Multiplexing Imaging System for Area Coverage and Point Targets |
US7893957B2 (en) * | 2002-08-28 | 2011-02-22 | Visual Intelligence, LP | Retinal array compound camera system |
US8497905B2 (en) * | 2008-04-11 | 2013-07-30 | nearmap australia pty ltd. | Systems and methods of capturing large area images in detail including cascaded cameras and/or calibration features |
US9046759B1 (en) * | 2014-06-20 | 2015-06-02 | nearmap australia pty ltd. | Compact multi-resolution aerial camera system |
US9052571B1 (en) * | 2014-06-20 | 2015-06-09 | nearmap australia pty ltd. | Wide-area aerial camera systems |
US9063391B2 (en) * | 2010-08-12 | 2015-06-23 | Rafael Advanced Defense Systems Ltd. | Method and system for increasing the size of the area scanned by an airborne electro-optic reconnaissance system in a given time |
US9185290B1 (en) * | 2014-06-20 | 2015-11-10 | Nearmap Australia Pty Ltd | Wide-area aerial camera systems |
US20160229555A1 (en) * | 2015-02-10 | 2016-08-11 | nearmap australia pty ltd. | Corridor capture |
US9440750B2 (en) * | 2014-06-20 | 2016-09-13 | nearmap australia pty ltd. | Wide-area aerial camera systems |
US9641736B2 (en) * | 2014-06-20 | 2017-05-02 | nearmap australia pty ltd. | Wide-area aerial camera systems |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1797849A (en) | 1926-12-16 | 1931-03-24 | Firm Photogrammetrie G M B H | Multiple chamber for taking photographs from aeroplanes |
US1735109A (en) | 1927-02-08 | 1929-11-12 | Leon T Eliel | Process of and apparatus for making aerial photographs |
US2665481A (en) | 1951-03-02 | 1954-01-12 | James E Henry | Method of aerial mapping |
US3610825A (en) | 1969-06-04 | 1971-10-05 | Hughes Aircraft Co | Image control apparatus for a vision system |
US4044364A (en) * | 1976-07-02 | 1977-08-23 | The Goodyear Tire & Rubber Company | Gravity stabilized camera and operator mount |
US5550669A (en) | 1993-04-19 | 1996-08-27 | Martin Marietta Corporation | Flexure design for a fast steering scanning mirror |
US5668593A (en) * | 1995-06-07 | 1997-09-16 | Recon/Optical, Inc. | Method and camera system for step frame reconnaissance with motion compensation |
US5894323A (en) | 1996-03-22 | 1999-04-13 | Tasc, Inc, | Airborne imaging system using global positioning system (GPS) and inertial measurement unit (IMU) data |
US6737591B1 (en) | 1999-05-25 | 2004-05-18 | Silverbrook Research Pty Ltd | Orientation sensing device |
DE10034601B4 (en) | 2000-07-14 | 2013-05-23 | Leica Geosystems Ag | Camera system with at least two first and second cameras |
US7597489B2 (en) | 2005-09-12 | 2009-10-06 | Honeywell International Inc. | Apparatus and method for providing pointing capability for a fixed camera |
US20070188610A1 (en) | 2006-02-13 | 2007-08-16 | The Boeing Company | Synoptic broad-area remote-sensing via multiple telescopes |
US9262818B2 (en) | 2007-05-01 | 2016-02-16 | Pictometry International Corp. | System for detecting image abnormalities |
US8331615B2 (en) | 2007-08-01 | 2012-12-11 | The Board Of Trustees Of The University Of Illinois | Match, expand, and filter technique for multi-view stereopsis |
JP4970296B2 (en) | 2008-01-21 | 2012-07-04 | 株式会社パスコ | Orthophoto image generation method and photographing apparatus |
US7899311B1 (en) | 2008-04-29 | 2011-03-01 | Optech International, Inc. | Removable shutter for a camera |
IL201682A0 (en) | 2009-10-22 | 2010-11-30 | Bluebird Aero Systems Ltd | Imaging system for uav |
US8548313B2 (en) * | 2010-09-30 | 2013-10-01 | Trimble Germany Gmbh | Aerial digital camera and method of controlling the same |
EP2673591B1 (en) | 2011-02-10 | 2014-12-03 | BAE Systems PLC | Image capturing |
US9336568B2 (en) | 2011-06-17 | 2016-05-10 | National Cheng Kung University | Unmanned aerial vehicle image processing system and method |
-
2016
- 2016-01-28 US US15/009,181 patent/US9641736B2/en active Active
-
2017
- 2017-03-17 US US15/462,212 patent/US20170195569A1/en not_active Abandoned
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3212420A (en) * | 1962-11-23 | 1965-10-19 | Dynasciences Corp | Image motion compensator |
US3580150A (en) * | 1968-04-15 | 1971-05-25 | Itek Corp | Panoramic camera with forward image motion compensation by optical rotation of image |
US6473119B1 (en) * | 1997-04-08 | 2002-10-29 | Carl-Zeiss-Stiftung | Photogrammetic camera |
US7893957B2 (en) * | 2002-08-28 | 2011-02-22 | Visual Intelligence, LP | Retinal array compound camera system |
US8497905B2 (en) * | 2008-04-11 | 2013-07-30 | nearmap australia pty ltd. | Systems and methods of capturing large area images in detail including cascaded cameras and/or calibration features |
US20090256909A1 (en) * | 2008-04-11 | 2009-10-15 | Nixon Stuart | Systems and methods of capturing large area images in detail including cascaded cameras and/or calibration features |
US20100026822A1 (en) * | 2008-07-31 | 2010-02-04 | Itt Manufacturing Enterprises, Inc. | Multiplexing Imaging System for Area Coverage and Point Targets |
US9063391B2 (en) * | 2010-08-12 | 2015-06-23 | Rafael Advanced Defense Systems Ltd. | Method and system for increasing the size of the area scanned by an airborne electro-optic reconnaissance system in a given time |
US9046759B1 (en) * | 2014-06-20 | 2015-06-02 | nearmap australia pty ltd. | Compact multi-resolution aerial camera system |
US9052571B1 (en) * | 2014-06-20 | 2015-06-09 | nearmap australia pty ltd. | Wide-area aerial camera systems |
US9185290B1 (en) * | 2014-06-20 | 2015-11-10 | Nearmap Australia Pty Ltd | Wide-area aerial camera systems |
US9440750B2 (en) * | 2014-06-20 | 2016-09-13 | nearmap australia pty ltd. | Wide-area aerial camera systems |
US9462185B2 (en) * | 2014-06-20 | 2016-10-04 | nearmap australia pty ltd. | Wide-area aerial camera systems |
US9641736B2 (en) * | 2014-06-20 | 2017-05-02 | nearmap australia pty ltd. | Wide-area aerial camera systems |
US20160229555A1 (en) * | 2015-02-10 | 2016-08-11 | nearmap australia pty ltd. | Corridor capture |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11175135B2 (en) | 2014-08-29 | 2021-11-16 | Spookfish Innovations Pty Ltd | Aerial survey image capture systems and methods |
US10378895B2 (en) * | 2014-08-29 | 2019-08-13 | Spookfish Innovagtions PTY LTD | Aerial survey image capture system |
US10171771B2 (en) * | 2015-09-30 | 2019-01-01 | Cisco Technology, Inc. | Camera system for video conference endpoints |
US20170324932A1 (en) * | 2015-09-30 | 2017-11-09 | Cisco Technology, Inc. | Camera system for video conference endpoints |
CN107704681A (en) * | 2017-09-30 | 2018-02-16 | 江西洪都航空工业集团有限责任公司 | A kind of computational methods of aircraft floor field range |
CN110022444A (en) * | 2017-12-06 | 2019-07-16 | 凌通科技股份有限公司 | The panorama photographic method of unmanned flight's machine and the unmanned flight's machine for using it |
WO2020136632A1 (en) * | 2018-12-27 | 2020-07-02 | Simplex Mapping Solutions Sb Ltd. | A compact interval sweeping imaging system and method |
US20210325182A1 (en) * | 2019-03-27 | 2021-10-21 | Chengdu Rainpoo Technology Co., Ltd. | Aerial survey method and apparatus capable of eliminating redundant aerial photos |
US11927442B2 (en) * | 2019-03-27 | 2024-03-12 | Chengdu Rainpoo Technology Co., Ltd. | Aerial survey method and apparatus capable of eliminating redundant aerial photos |
WO2021126941A1 (en) * | 2019-12-18 | 2021-06-24 | Bae Systems Information And Electronic Systems Integration Inc. | Method for co-locating dissimilar optical systems in a single aperture |
US11474363B2 (en) | 2019-12-18 | 2022-10-18 | Bae Systems Information And Electronic Systems Integration Inc. | Method for co-locating dissimilar optical systems in a single aperture |
US11226436B2 (en) | 2019-12-18 | 2022-01-18 | Bae Systems Information And Electronic Systems Integration Inc. | Method for co-locating dissimilar optical systems in a single aperture |
WO2022086976A1 (en) * | 2020-10-19 | 2022-04-28 | Pictometry International Corp. | Variable focal length multi-camera aerial imaging system and method |
GB2615041A (en) * | 2020-10-19 | 2023-07-26 | Pictometry Int Corp | Variable focal length multi-camera aerial imaging system and method |
US12063443B2 (en) | 2021-06-28 | 2024-08-13 | nearmap australia pty ltd. | Hyper camera with shared mirror |
US11722776B2 (en) | 2021-06-28 | 2023-08-08 | nearmap australia pty ltd. | Hyper camera with shared mirror |
US12015853B2 (en) | 2021-06-28 | 2024-06-18 | nearmap australia pty ltd. | Hyper camera with shared mirror |
WO2023170472A1 (en) * | 2022-03-07 | 2023-09-14 | Stuart Nixon | Airborne and spaceborne imaging survey platform |
Also Published As
Publication number | Publication date |
---|---|
US20160150142A1 (en) | 2016-05-26 |
US9641736B2 (en) | 2017-05-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9641736B2 (en) | Wide-area aerial camera systems | |
US9462185B2 (en) | Wide-area aerial camera systems | |
US9440750B2 (en) | Wide-area aerial camera systems | |
US9706117B2 (en) | Wide-area aerial camera systems | |
US9188838B1 (en) | Compact multi-resolution aerial camera system | |
US10848654B2 (en) | Oblique scanning aerial cameras | |
US20200191568A1 (en) | Multi-View Aerial Imaging | |
US11281931B2 (en) | Multi-view scanning aerial cameras | |
US12081724B2 (en) | Multiplexed multi-view scanning aerial cameras |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |