US5508734A - Method and apparatus for hemispheric imaging which emphasizes peripheral content - Google Patents

Method and apparatus for hemispheric imaging which emphasizes peripheral content Download PDF

Info

Publication number
US5508734A
US5508734A US08/281,331 US28133194A US5508734A US 5508734 A US5508734 A US 5508734A US 28133194 A US28133194 A US 28133194A US 5508734 A US5508734 A US 5508734A
Authority
US
United States
Prior art keywords
image
optical
output
view
field
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US08/281,331
Other languages
English (en)
Inventor
Robert G. Baker
Kevin Kettler
Gustavo A. Suarez
Kenneth A. Uplinger
Candace J. F. Freedenberg
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
MediaTek Inc
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US08/281,331 priority Critical patent/US5508734A/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORP. reassignment INTERNATIONAL BUSINESS MACHINES CORP. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: UPLINGER, KENNETH ALLEN, KETTLER, KEVIN, FREEDENBERG, CANDACE JOY FLATTERY, BAKER, ROBERT GROVER, SUAREZ, GUSTAVO ARMANDO
Priority to JP07122200A priority patent/JP3103008B2/ja
Priority to CA002152314A priority patent/CA2152314C/en
Priority to BR9502919A priority patent/BR9502919A/pt
Priority to US08/496,742 priority patent/US5686957A/en
Priority to DE69515087T priority patent/DE69515087T2/de
Priority to EP95304930A priority patent/EP0695085B1/de
Publication of US5508734A publication Critical patent/US5508734A/en
Application granted granted Critical
Priority to JP2000220238A priority patent/JP3485261B2/ja
Assigned to MEDIATEK INC. reassignment MEDIATEK INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTERNATIONAL BUSINESS MACHINES CORPORATION
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • H04N7/142Constructional details of the terminal equipment, e.g. arrangements of the camera and the display
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B13/00Optical objectives specially designed for the purposes specified below
    • G02B13/06Panoramic objectives; So-called "sky lenses" including panoramic objectives having reflecting surfaces
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B3/00Simple or compound lenses
    • G02B3/0087Simple or compound lenses with index gradient
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B6/00Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
    • G02B6/04Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B6/00Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
    • G02B6/04Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres
    • G02B6/06Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres the relative position of the fibres being the same at both ends, e.g. for transporting images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2628Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • the present invention relates generally to visual imaging systems, and more particularly to visual imaging systems and techniques which provide useful electronic manipulation of wide angle hemispheric scenes.
  • a unique class of multimedia data sets is that of hemispheric visual data.
  • Known multimedia methods and apparatus attempt to combine various multimedia imaging data, such as still and motion (or video) images, with audio content using storage media such as photographic film, computer diskettes, compact discs (CDs), and interactive CDs. These are used in traditional multimedia applications in various fields, such as entertainment and education.
  • Non-multimedia applications also exist that would employ hemispheric visual data, such as in security, surveillance, unmanned exploration, and fire and police situations.
  • the known methods and apparatus have certain limitations in capturing and manipulating valuable information of hemispheric scenes in a rapid (i.e., real-time) and cost-effective manner.
  • One well-known multimedia technique is used at theme parks, wherein visual information from a scene is displayed on a screen or collection of screens that covers almost 360 degrees field of view.
  • Such a technique unfortunately results in the consumption of vast quantities of film collected from multiple cameras, requires specially-designed carriages to carry and support the cameras during filming of the scene, and necessitates synchronization of shots during capture and display.
  • the technique is also limited in that the visual image cannot be obtained with a single camera nor manipulated for display, e.g., pan, tilt, zoom, etc., after initial acquisition.
  • this technique while providing entertainment, is unable to fulfill critical technical requirements of many functional applications.
  • U.S. Pat. Nos. 4,125,862; 4,442,453; and 5,185,667 are described in U.S. Pat. Nos. 4,125,862; 4,442,453; and 5,185,667.
  • U.S. Pat. No. 4,125,862 a system is disclosed that converts signal information from a scene into digital form, stores the data of the digitized scene serially in two-dimensional format, and reads out the data by repetitive scan in a direction orthogonally related to the direction in which the data was stored.
  • U.S. Pat. No. 4,442,453 discloses a system in which a landscape is photographed and stored on film. The film is then developed, with display accomplished by scanning with electro-optical sensors at "near real-time" rates.
  • a system of the type described by the '667 patent has limitations in obtaining sufficient information of critical and useful details. This is particularly true when the camera is oriented with the central axis of the lens perpendicular to the plane bounding the hemisphere of acquisition (i.e. lens pointing straight up).
  • the majority of critical detail in a scene is contained in areas of the field along the horizon and little or no useful details are contained in central areas of the field located closer to the axis of the lens (the horizon being defined as the plane parallel to the image or camera plane and perpendicular to the optical axis of the imaging system).
  • the imaging system is aimed upward and the majority of the critical detail in the scene includes people, buildings, trees, etc.--most of which are located within only a few degrees along the horizon (i.e., this is the peripheral content). Also, in this example, although the sky makes up the larger central area of the view, it contains little or no useful information requiring higher relative resolution.
  • the technique should be able to differentiate between the relevant visual information along the horizon and the remaining visual information in the scene in order to provide greater resolution in areas of higher importance.
  • the system described by the '667 patent does not differentiate between this relevant visual information contained along the horizon and the remaining visual information in this scene. Thus, it fails to yield a sufficient quality representation of the critical detail of the scene for projected applications.
  • U.S. Pat. No. 4,170,400 describes a wide-angle optical system employing a fiber optic bundle that has differing geometric shapes at the imaging ends. Although this is useful in itself for collecting and repositioning image data, bending of light is a natural characteristic of optical fibers and not exclusive to that patent. Further, U.S. Pat. No. 4,170,400 employs a portion of a spherical mirror to gather optical information, rendering a very reduced subset of the periphery in the final imaging result. This configuration is significantly different from the multi-element lens combination described in the present invention.
  • Imperfections in the image representation of any field inherently result from the nature of creating an image with any spherical glass (or plastic) medium such as a lens.
  • the magnitude of these imperfections increases proportionally to the distance a point in the field is from the axis perpendicular to the optical imaging system.
  • aberrations of the corresponding image increase proportional to this angle cubed.
  • aberrations are more highly exaggerated in the peripheral areas with respect to more central areas of a hemispheric image.
  • the lens types above achieve a view of a large field, the valuable content from the peripheral areas lacks in potential image quality (resolution) mapping because the imaging device and system does not differentiate between these areas and the central areas of less valuable detail.
  • the difference between the imaging capabilities between the two areas is compensated for by using only the central portion of a lens to capture the scene ("stopping the lens down"). This works in effect to reduce the image quality of both areas such that the difference in error is a lesser percentage of the smallest area even the central area can resolve.
  • this compensation technique further degrades the performance of the lens by limiting the amount of light which is allowed to enter the lens, and thus reducing the overall intensity of the image.
  • the peripheral content imaged by a conventional lens is so degraded in comparison with the central area that the lens allows for only a minimal area of the periphery to be recorded by the film or electronic imager.
  • the relevant information of the horizon in the scene can be underutilized or worse yet--lost.
  • the present invention provides a new and useful visual imaging system that emphasizes the peripheral content of a hemispheric field of view using a single camera.
  • the captured visual information can be stored as a single image using conventional chemical-based (film) or electronic data storage techniques.
  • the invention allows user-selected portions of a hemispheric scene to be electronically manipulated and displayed from the stored visual database in a real-time and cost-effective manner.
  • the visual imaging system of the present invention involves either a still image or a moving picture camera, electronic or otherwise, having a lens with enhanced peripheral content imaging capabilities.
  • the lens provides an enhanced view of the valuable information in the scene's periphery by imaging a field to the image plane such that the ratio of the size of the smallest detail contained within the periphery of the scene to the size of the smallest resolving pixel of an image device is increased.
  • the peripheral content must map to a larger percentage of a given image detector area and, simultaneously, the mapped image of the central area of the scene must be minimized by the lens so that it does not interfere with the peripheral content now covering a wider annulus in the image plane.
  • Information in the image plane is then detected by an imager device (either a photographic film or electronic imager or video detector array). The detected information of the entire hemispheric scene is then stored as a single image in memory using traditional methods.
  • a transform processor subsystem electronically manipulates the scene for display as a perspective-correct image on a display device, such as a conventional monitor or TV, as if the particular portion of the scene had been photographed with a conventional camera.
  • the transform processor subsystem compensates for the distortion or difference in magnification between the central and peripheral areas of the scene caused by the lens by applying appropriate correction criteria to bring the selected portion of the scene into standard viewing format.
  • the transform processor subsystem can also more fully compensate for any aberrations of the enhanced peripheral image because of the image's improved resolution as it covers a larger portion of the image device (increased number of pixels used to detect and measure the smallest detail in the periphery image). More pixels equates to more measurement data, hence more accurate data collection.
  • the stored image can also be manipulated by the transform processor subsystem to display an operator-selected portion of the image through particular movements, such as pan, zoom, up/down, tilt, rotation, etc.
  • the visual imaging system can use a single camera to capture the relevant visual information within a panoramic field of view existing along the horizon, while being able to conventionally store and easily display the scene, or portions thereof, in real-time.
  • Using a single optical system and camera is not only cost-effective but keeps all hemispheric visual data automatically time-synchronized.
  • the unique visual imaging system lens can capture information from a hemispheric scene by emphasizing the peripheral portion of the hemispheric field of view and thus provide greater resolution with existing imaging devices for the relevant visual information in the scene.
  • an ordinary fish-eye lens focuses the lowest 15 degrees up from the horizon on 10% of the imager at the objective plane and the peripheral-enhancing lens focuses that same 15 degrees on 50% of the imager, there is a 5-fold increase in resolution using the same imaging device.
  • Another advantage of the present invention is that the captured scene information having an enhanced peripheral content can be stored as a single image using conventional storage techniques, but need not be stored in standard formats only.
  • Still another advantage of the present invention is that the image can be read out from storage and electronically manipulated at any time after acquisition to provide flexibility in display options, allowing post-acquisition selection of any particular view desired.
  • FIG. 1 is a schematic illustration of the various components and organization of the visual imaging system of the present invention
  • FIG. 2A is a cross-sectional diagram of a wide angle lens system of the prior art
  • FIG. 2B is a box diagram indicating the field input and output rays and the resulting relative field coverage a lens of the prior art typically provides in an image plane for detection by an imager device;
  • FIG. 2C is a cross-sectional diagram of a lens system of the prior art indicating angular key to visual field
  • FIG. 3A is a cross-sectional diagram of one embodiment of a peripheral image enhancing lens system of the present invention.
  • FIG. 3B is a box diagram of the annulus field input of a bundled fiber array portion of the lens system of FIG. 3A;
  • FIG. 3C is a box diagram of the imager interface output of a bundled fiber array portion of the lens system of FIG. 3A;
  • FIG. 4A is a cross-sectional diagram of another embodiment of a peripheral image enhancing lens system of the present invention.
  • FIG. 4B is a box diagram indicating the field input and output rays indicating the resulting field coverage of the lens system of FIG. 4A;
  • FIG. 5 is a schematic representation of the mapping locations of the lens systems of the present invention resulting upon an imaging device
  • FIG. 6 is a schematic block diagram of the image processing circuitry of the present invention including the transform processor subsystem thereof;
  • FIG. 7 is a block diagram schematically indicating further integration of a portion of the image processing circuitry of FIG. 6, and
  • FIG. 8 is an alternate embodiment of the image processing circuitry of the present invention incorporating the integrated circuitry elements of FIG. 7.
  • the present invention stems from the realization by the inventors that in many of the technical hemispheric field applications, where the image detector is parallel to the horizon, much of the relevant visual information in the scene (e.g., trees, mountains, people, etc.) is found only in a small angle with respect to the horizon. Although the length of the arc from the horizon containing the relevant information varies depending upon the particular application, the inventors have determined that in many situations, almost all the relevant visual information is contained within about 10 to 45 degrees with respect to the horizon.
  • the "central" area contains less detailed information, such as a solid white ceiling or a clear or lightly-clouded sky, it is allowable to maximize completely the dedication of the available image detection area to the peripheral field portion by reducing the portion of the imager device representing the "central" area to near zero.
  • the present invention provides two manners (Example I and Example II) for capturing, storing, and selectively displaying the critical visual information in a scene for many important applications.
  • the visual imaging system of the invention includes a still image or moving picture camera 10, having a lens, indicated generally at 14, designed to capture and enhance the peripheral content of a hemispheric scene.
  • the captured scene can be stored onto an assortment of media, e.g., photographic film 16, electronic storage 18, or other conventional storage means.
  • Electronic storage 18 is preferred because of the ease of electronic manipulation thereof.
  • photographic film 16 requires an image scanner 20 or other capture-and-conversion method to change the image into electronic format before electronic manipulation can be performed.
  • the stored electronic image data is then selectively accessed by a transform processor engine 22 and can be electronically manipulated according to user-defined criteria, such as pan, up/down, zoom, etc.
  • the transform processor 22 corrects the image for display on a conventional display device 28 in a normal viewer format or on head-mounted displays 30, in which integrated orientation-sensing devices having, for example, wearer eye focus or head position detection function, can be used to manipulate and define the user controls.
  • the camera 10 for the visual imaging system is an optical device that is capable of receiving a focused image from a lens and transforming that image into an electronic signal or into hard copy storage such as photographic film.
  • Various types of cameras for wide-angle viewing are known to those in the art, such as 35 mm cameras, 8 mm cameras, NTSC, RS170 and HDTV-type cameras.
  • the present invention is designed to be compatible with most commercially-available two-dimensional cameras, with scaling of the lens geometries. It also has the technological capability to be applied to three-dimensional cameras.
  • the camera can be mounted and supported in a conventional manner.
  • the fundamental principle behind the enhanced peripheral content lens is the selective magnification of the periphery and the focusing of more of that content on the objective plane.
  • This recognizes the current limitations of imaging devices and film with regard to resolution.
  • the more of the peripheral content that can be focused on the objective plane surface the more points of data that can be resolved with a given density of imaging device or material. Therefore, for this new class of selective magnification lenses, the surface area of the imaging plane reserved for peripheral content will be large relative to the central content and roughly similar for all lenses in this class, regardless of whether the lens is designed for 1-10 degree peripheral emphasis or 1-45 degree peripheral emphasis.
  • the lens with 1-10 degree emphasis will have much better resolution for the same objects than the lens with 1-45 degree emphasis.
  • the lens 14 for camera 10 provides a collection of data for enhanced digital processing of the peripheral portion of a given field of view.
  • the lens uniquely achieves this by filling the greater available area of an imager device with the peripheral areas rather than the central areas of the captured scene.
  • a periphery-enhancing optical system suitable for achieving the goals of the present invention can be configured in various ways to present an image of the field free of detrimental values of critical aberrations. Two examples or embodiments are preferred.
  • Example I as illustrated by FIGS. 3A-3C, preferably is a multi-medium system comprising a wide-angle multi-element optical lens and a fiber optic imager device.
  • Example II as illustrated by FIGS. 4A-4B, preferably is a combination system comprised of multiple refractive optical elements, one of which is of gradient index material and hemispherical in shape. The inventors define that other configurations relying on binary optics could also be utilized to accomplish similar results.
  • Example I The system of Example I is best suited for applications where the capture of the minimal information contained in the central area is irrelevant.
  • a coordinated fiber bundle array 40 either rigid or flexible, carefully aligned and assembled by methods known to those skilled in the art, is used to capture a peripheral annulus of the wide field image provided by the multi-element wide field lens.
  • the multi-element wide field lens 42 therefore, although covering the same field as standard wide field lenses, is required to provide a larger image of the entire field. This can be accomplished by those skilled in the art by scaling a standard lens to a larger focal length until the desired image size is reached.
  • the fibers arranged in a specific mechanical manner, deliver the information in a geometrically annular shape from the peripheral image area to either the rectangular or circular shape of a specific imaging device.
  • Specific configurations can be devised for desired peripheral coverage (up to 10 degrees through 45 degrees) and matched to required image detector plane geometry. (REF. Sect. 13, Fiber Optics, Handbook of Optics, McGraw-Hill Inc., 1978.)
  • a desired angle ⁇ (theta)
  • a standard wide-field lens design is scaled in "f", the focal length, to provide r y , such that the minimum pick-up area of the camera imaging device (2r') is met.
  • the equation for fiber bundle radius r' as related to the imager is:
  • a configuration in the form of Example I achieves an image in which the peripheral portion (0-45 degrees with respect to the horizon) of the field of view preferably encompasses between approximately 90% to 100% of the entire usable area of the imager device, rather than the 35% or less that is typical with commercially-available wide-angle, fish-eye, or other conventional lenses.
  • a configuration in the form of Example II achieves an image in which the peripheral portion preferably encompasses between 50% to 70% of the usable area of the imaging device surface.
  • the system of Example II is best suited for applications where all information contained in the field of view is relevant, although to differing degrees.
  • the combination system of Example II relies on the principle use of a Bravais System where a hemispherical lens 52 or a combination of lenslets magnify an object to an image, wherein the object and image lie in the same plane.
  • the hemispherical lens 52 is made of gradient index material such that points impinging the image plane incur different factors of magnification depending on what index of refraction and what portion of the hemispherical curve or lenslets they pass through.
  • This concept is used in combination with a multi-element refractive lens 54 that is designed to capture a wide field of view and also compensate for the color aberration induced by insertion of the hemispherical lens.
  • This color compensation can be designed by those skilled in the art by using a computer optimization routine.(REF. M. Horimoto and U.S. Pat. No. 4,256,373.)
  • the hemispherical gradient index unit With a wide field multi-element lens, the portion of the camera dedicated to the periphery is increased, thereby increasing the relative resolution of information detectable by the imaging device sensing elements. With the gradient index hemisphere, index values decrease from the center, such that:
  • FIGS. 2A-B, 4A and 4B The operation of the optical systems is schematically illustrated in FIGS. 2A-B, 4A and 4B.
  • arc X represents the field of view of the "center" of the lens system
  • arc Y represents the practically usable portion of the "peripheral" field.
  • the areas X' and Y' in FIG. 2A represent the resultant image focal locations on the imager at the object plane.
  • Arc Z and area Z' represent areas outside the normal imaging range of the lens. (No actual demarcation lines would exist in actual use; this is merely for explanatory purposes.)
  • FIG. 2A represents a typical wide-angle type lens 32
  • FIG. 4A represents a lens constructed according to the principles of the present invention.
  • a typical wide-angle type lens has a fairly significant portion of the image surface dedicated to the central field of the lens; while the lens constructed according to the present invention has a fairly significant portion of the objective surface dedicated to the peripheral field--and consequently less of the surface dedicated to the central field.
  • the portion of the image surface used for the peripheral portion of the scene can vary depending upon the particular prescription of lens specified, which is selected to capture items of interest for a given application.
  • the relevant visual information may be contained within 10 degrees of the horizon.
  • the lens of the present invention can thus be designed to enhance only the field of view within 10 degrees of the horizon.
  • the relevant visual information may include objects on walls and thus be contained within about 45 degrees from the horizon.
  • a peripheral enhancing lens can thus also be designed to enhance the field of view up to 45 degrees from the horizon.
  • the enhanced portion of the field of view depends upon the needs of the particular application, with the enhanced portion preferably falling somewhere between these two extremes.
  • the principles of the present invention can be applied to these types of situations with equal success, with any appropriate corrections being made with the image transform processors, as will be described herein in more detail.
  • a preferred form of the Example I form of optical system comprises a standard wide field lens 42 and a coordinated fiber array 40.
  • the focal length of the wide field lens is scaled in order to match the peripheral field of view desired.
  • the coordinated fiber array consists of an annular input face 44 that collects the image projected from the standard wide field lens.
  • the fiber bundle array then redirects the information from the peripheral view to its output end by total internal reflection. Fibers capable of 3 micron accuracy, covering the area of the annular region, are coordinated into a rectangular or circular shape at their output, depending on the geometry of the corresponding imaging device. The size of output is also matched to the camera imaging device used.
  • An electronic camera 10 used for exemplary purposes includes an imager device that records the optical image from the lens at the object plane.
  • the imaging medium is film
  • the imaging medium is an electronic device such as a charge-coupled device (CCD) or charge-injected device (CID).
  • CCD charge-coupled device
  • CID charge-injected device
  • electronic processes are typically preferred over photographic processes as they are easier to electronically manipulate. Photographic processes, however, can be preferred in certain situations. Numerous makes and models of film and electronic imaging devices are known to those skilled in the art that provide uniform resolution across the receiving surface.
  • the imager device When the imager device is used with a lens constructed according to the principles of the present invention, the imager device will collect more information along the horizon than it will from the central area of the hemispheric scene. With only a given limited resolution of either film emulsions or CCD pixel density, the lens focuses more useful information at the objective plane. The peripheral portion of the scene will therefore have a higher relative resolution in the resultant transformed image than the central portion. Thus, the details of any objects along the horizon will be highly accentuated. Further, any distortion (e.g., spherical aberration) that occurs in the peripheral region of the lens will be imaged onto a larger surface and can thus be more easily and fully compensated for.
  • any distortion e.g., spherical aberration
  • the image mapped onto the imager device can be described by a series of concentric circles, as schematically illustrated in FIG. 5.
  • each circle a, b, c, d, etc. on the imager device can be described by radii of arbitrary units, e.g., 2, 5, 9, 14, etc., respectively.
  • the radii of the circles depends on the magnification of the different regions of the hemisphere, with the sections of the outer circles having a greater area as the magnification of the peripheral region increases.
  • each concentric circle represents 18 degrees field of view from the horizontal plane, with the outer circumference of the outermost circle being level with the horizon.
  • the inventors have determined that the arc subtending the two outer circles (i.e., 36 degrees from the horizon) contains the relevant information in many hemispheric scenes for many applications (although this value can be varied depending upon the particular application).
  • Example I circular-output optical system is configured with a square or rectangular imaging device, the corner areas of the imaging device are not useful for photographic purposes. However, these areas can be used to store other information, for example, such as digitized audio content or other secondary documentation from the scene, if the imager is radially mapped 1:1 with electronic memory.
  • the image received on the imager device is passed on to the system components for storage.
  • a storage device might be film; while for electronic processes, the storage device might be electronic storage in the form of random access memories, a conventional diskette or hard file, or video recording tape.
  • the entire display of the scene (along with any secondary documentation) can be stored as a single image on the storage device.
  • the image is stored in a "warped” form.
  • the warped image is caused not only by the wide-angle nature of the lens (i.e. the “keystoning” effect), but also by the enhanced peripheral field of view of the lens (i.e., magnification along the periphery).
  • the underlying concept is that a partial slice of the scene can be reproduced with the proper aspect ratio for the human visual system (i.e., as a perspective corrected view).
  • the stored image is loaded into the source image buffer 40 if it has been stored in electronic form on a host system such as a personal computer or controller.
  • the image can be brought in for processing without going through storage.
  • an analog signal from a video camera can connect into an NTSC-to-digital converter 60. This converts the image from analog information into a digital bit map (i.e., into "pixels").
  • the source image is then loaded into the source image frame buffer 62.
  • any type of camera can be used to provide the electronic input to buffer 62.
  • the buffer preferably operates with sufficient speed so that real-time viewing is possible.
  • the stored image can be selectively accessed and transformed for display. If the storage is photographic film, an image scanner may be used to convert the stored image into an electronic format for subsequent manipulation. In order to recreate a proper display of the scene in two dimensions for perspective-correct viewing, processor logic in transform processor engine 22 is utilized.
  • the transform processors may be made of collections of small-scale, medium-scale, large-scale, or very-large-scale integrated (VLSI) circuits, examples of which are image resampling sequencers such as the TMC2301 and TMC2302, commercially available from Raytheon Semiconductors (formerly TRW LSI Products, Inc., LaJolla, Calif.).
  • resampling sequencers control the address sequencing of the pixels in the source image buffer 62 through a multiply/accumulate unit 64, and from there into the warped image buffer 66.
  • the sequencers control the filtering or remapping of 2-dimensional images from a set of Cartesian coordinates (x,y) as defined within each sector (e1, d1, etc.) onto a newly transformed set of coordinates (u,v).
  • the "fish-eye" type of transformations described in U.S. Pat. No. 5,185,667 are based on non-constant second-order derivatives. A different set of second-order derivatives would be employed for the transforms associated with the peripheral-enhancing configurations of the present invention.
  • the sequencers can also handle three-dimensional images by resampling them from a set of Cartesian coordinates (x,y,z) into a new, transformed set (u,v,w).
  • these sequencers can support nearest-neighbor, bilinear interpolation or convolution resampling, and can operate at speeds allowing real-time operation.
  • Remapped pixel locations (i.e., interpolation "kernels") of more than one pixel in the bit map require an external interpolation coefficient look-up table 68 and the multiply/accumulate unit 64.
  • a table "walk” is typically performed on each source pixel, thus providing a smoother image by summing the products of the original lens image data with the appropriate interpolation coefficients.
  • the warp engine can be programmed to perform a perspective correction, much like an inverse keystoning effect.
  • the remapping of the pixel locations is matched to the differential magnification of the particular periphery-enhancing lens system used.
  • Direct access to the interpolation coefficient look-up table 68 and to the transformation parameters is also desirable to allow dynamic modification of the interpolation algorithm.
  • a local interpolation coefficient buffer 70 to update the varying transform parameters is included to allow for real-time still and motion image transformations.
  • the row and column warping engines 72a, 72b of the transform processor 22 supply addresses to the source image buffer 62.
  • the addresses are determined by the interpolation algorithm chosen.
  • the multiply/accumulate unit 64 takes the pixels supplied by the source image buffer 62 under warping engine control and multiplies the pixels together using combinational logic with weighting factors dependent on the algorithm. Compensation for aberration (e.g., spherical aberration) can also be made at this point.
  • the composed interpolated pixels are sent to the warped image buffer 66.
  • the address location within the warped image buffer is again determined by the warping.
  • the algorithm parameters from look-up table 68 are input to the registers of the row and column warping engines 72a, 72b, as well as into the interpolation coefficient buffer 70.
  • the memory controller/clock circuitry 74 provides refresh control to the source and warped image buffers 62, 66. In addition, all clock sources are synchronized through this circuitry.
  • the bus interface and control circuitry 76 also provide an interface to the host system bus (i.e., for MCA, ISA, etc.) and the remapping circuitry. This interface logic serves to load control information into the remapping circuitry and to provide a path to transport warped images to the system display buffer (not shown; part of host system), or store images to disk via the system bus prior to warping.
  • An optional random access memory digital-to-analog converter (RAMDAC) 78 provides support for a local display connection if desired.
  • One feature of the transform processors is the valid source address flag within the transform processor engine. This allows the user to construct abutting subimages in the (x,y) plane without danger of edge interference. Thus, edge detection of the unused areas outside the circular image of FIG. 5 can alert the system to ignore these values.
  • the image capture function can be accomplished with either still or motion video devices or as pre-recorded digital data. All types of image data are input to the source image buffer 62 for processing as desired. Note that while the preferred mode for still image capture will come from previously captured images through a local host bus interface 76, the NTSC digitizer 60, for example, can provide real-time data from an external video camera. Any similar device that converts an image to the appropriate digital format for input to the source image buffer can be substituted for digitizer 60. In the same fashion, an electronic still imager, such as an electronic still camera, line scanner, or table scanner, can provide still image data for processing. Prerecorded distorted images generated through this invention's optical system can also be input through an additional conversion device to allow dynamic manipulation of previously recorded image data.
  • the image transformation performed on the captured digitized image from modified hemispheric coordinates to planar coordinates for display is one of a multitude of possible image transformations, any of which can be invoked in real-time for smooth merging of effects. These transformations include, but are not limited to pans, up/downs, zooms, tilts, rotations, scaling, cropping and image shear, which can be controlled using human or computer input. Image filtering can be performed as well as edge detection in associated processes during the course of manipulation. These services can be applied to any system image loaded into the source image buffer, thus providing a host of added features beyond the simple application of the hemispheric lens and display system.
  • a security camera application can be implemented to view an entire panorama such that the security monitor will display full-motion rate images in real-time.
  • Source image coordinates for still images can also be sequenced, allowing perceived animation or full-motion renditions by simply reloading new source image coordinates into the warping engines as frames are drawn from memory. Details from other stored images can be utilized to give the effect of full-motion panning of the horizon within the captured image by the lens.
  • An additional feature of the present invention is its ability to create motion video presentations with a reduced set of still frames, as compared with the number of full-motion frames ordinarily required. For example, when a set of building storefronts is filmed with a traditional motion picture camera from the back of a moving truck, each frame contains the time-based content of the image entering the traditional lens and only one limited field of view is available at a time. With the system of the present invention, motion can be reconstructed from a greatly reduced subset of frames because the greater field of captured data already contains picture content from the next frame and previous frame due to its exceptionally wide angle.
  • intermediate frame equivalents sufficient to complete a full-motion data set can be constructed and fed to the source image buffer in sequence for processing.
  • separate circuits operating at real-time rates can interpolate intermediate values and supply changes in transformation parameters for existing still images fast enough to synthesize motion sequences. Preference for one technique over another will be dependent on the specific application requirements and other development/market considerations.
  • the transform processor subsystem can produce multiple different outputs simultaneously from individual stored or currently-converted images.
  • multiple outputs may be generated from a single image source, either motion or still, with individual effects for each scene as desired, allowing several scenes on different display devices or several windows on a single display. This is accomplished by incorporating several image processing subsystems 80 within one overall system, as shown in FIG. 8.
  • spherical aberrations can be more fully and completely compensated for, as the aberrations are spread across a greater area on the imager device.
  • the present invention provides a visual imaging system that efficiently captures, stores, and displays visual information about an enhanced hemispheric field of view existing particularly along the horizon, and that allows electronic manipulation and selective display thereof even after acquisition and storage, while minimizing distortion.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)
  • Closed-Circuit Television Systems (AREA)
  • Optical Fibers, Optical Fiber Cores, And Optical Fiber Bundles (AREA)
  • Image Input (AREA)
  • Cameras In General (AREA)
  • Lenses (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
US08/281,331 1994-07-27 1994-07-27 Method and apparatus for hemispheric imaging which emphasizes peripheral content Expired - Lifetime US5508734A (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
US08/281,331 US5508734A (en) 1994-07-27 1994-07-27 Method and apparatus for hemispheric imaging which emphasizes peripheral content
JP07122200A JP3103008B2 (ja) 1994-07-27 1995-05-22 半球状視野の電子結像及び処理のためのシステム及び方法
CA002152314A CA2152314C (en) 1994-07-27 1995-06-21 Method and apparatus for hemisperic imaging
BR9502919A BR9502919A (pt) 1994-07-27 1995-06-23 Sistemas para formação e/ou manipulação eletrônica e/ou ótica de imagens e/ou cenas dentro de um campo de visão hemisférico bem como seus métodos de captura armazenamento manipulação e geração de sequencias de imagens
US08/496,742 US5686957A (en) 1994-07-27 1995-06-30 Teleconferencing imaging system with automatic camera steering
DE69515087T DE69515087T2 (de) 1994-07-27 1995-07-14 Verfahren und Vorrichtung zur hemisphärischen Bildaufnahme
EP95304930A EP0695085B1 (de) 1994-07-27 1995-07-14 Verfahren und Vorrichtung zur hemisphärischen Bildaufnahme
JP2000220238A JP3485261B2 (ja) 1994-07-27 2000-07-21 半球状視野の電子結像及び処理のためのシステム及び方法

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US08/281,331 US5508734A (en) 1994-07-27 1994-07-27 Method and apparatus for hemispheric imaging which emphasizes peripheral content

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US08/496,742 Continuation-In-Part US5686957A (en) 1994-07-27 1995-06-30 Teleconferencing imaging system with automatic camera steering

Publications (1)

Publication Number Publication Date
US5508734A true US5508734A (en) 1996-04-16

Family

ID=23076831

Family Applications (2)

Application Number Title Priority Date Filing Date
US08/281,331 Expired - Lifetime US5508734A (en) 1994-07-27 1994-07-27 Method and apparatus for hemispheric imaging which emphasizes peripheral content
US08/496,742 Expired - Fee Related US5686957A (en) 1994-07-27 1995-06-30 Teleconferencing imaging system with automatic camera steering

Family Applications After (1)

Application Number Title Priority Date Filing Date
US08/496,742 Expired - Fee Related US5686957A (en) 1994-07-27 1995-06-30 Teleconferencing imaging system with automatic camera steering

Country Status (6)

Country Link
US (2) US5508734A (de)
EP (1) EP0695085B1 (de)
JP (2) JP3103008B2 (de)
BR (1) BR9502919A (de)
CA (1) CA2152314C (de)
DE (1) DE69515087T2 (de)

Cited By (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5686957A (en) * 1994-07-27 1997-11-11 International Business Machines Corporation Teleconferencing imaging system with automatic camera steering
US5703605A (en) * 1994-12-05 1997-12-30 Olympus Optical Co., Ltd. Image display apparatus
WO1997050252A1 (en) * 1996-06-24 1997-12-31 Behere Corporation Panoramic camera
US5815199A (en) * 1991-01-31 1998-09-29 Matsushita Electric Works, Ltd. Interphone with television
US5870135A (en) * 1995-07-27 1999-02-09 Sensormatic Electronics Corporation Image splitting forming and processing device and method for use with no moving parts camera
US6043837A (en) * 1997-05-08 2000-03-28 Be Here Corporation Method and apparatus for electronically distributing images from a panoptic camera system
US6222683B1 (en) 1999-01-13 2001-04-24 Be Here Corporation Panoramic imaging arrangement
US6301447B1 (en) * 1991-05-13 2001-10-09 Interactive Pictures Corporation Method and system for creation and interactive viewing of totally immersive stereoscopic images
WO2001082592A1 (de) * 2000-04-25 2001-11-01 Rainer Mockler Digitalkamera oder digitale videokamera und verfahren zur bilderfassung
US6313865B1 (en) 1997-05-08 2001-11-06 Be Here Corporation Method and apparatus for implementing a panoptic camera system
US6331869B1 (en) 1998-08-07 2001-12-18 Be Here Corporation Method and apparatus for electronically distributing motion panoramic images
US6341044B1 (en) 1996-06-24 2002-01-22 Be Here Corporation Panoramic imaging arrangement
EP1178352A1 (de) * 1996-06-24 2002-02-06 Behere Corporation Verfahren und Vorrichtung zur Präsentation von Panoramabildern auf einem lokalen Empfänger, sowie zugehöriges Computerprogramm
US6369818B1 (en) 1998-11-25 2002-04-09 Be Here Corporation Method, apparatus and computer program product for generating perspective corrected data from warped information
US6373642B1 (en) 1996-06-24 2002-04-16 Be Here Corporation Panoramic imaging arrangement
US20020122113A1 (en) * 1999-08-09 2002-09-05 Foote Jonathan T. Method and system for compensating for parallax in multiple camera systems
US20020147991A1 (en) * 2001-04-10 2002-10-10 Furlan John L. W. Transmission of panoramic video via existing video infrastructure
US6466254B1 (en) 1997-05-08 2002-10-15 Be Here Corporation Method and apparatus for electronically distributing motion panoramic images
US20020180877A1 (en) * 2001-03-26 2002-12-05 Susumu Kikuchi Image acquisition apparatus
US6493031B1 (en) * 1995-03-31 2002-12-10 Canon Kabushiki Kaisha Visual information processing method and apparatus for extracting feature quantities from a two-dimensional image signal
US6493032B1 (en) 1996-06-24 2002-12-10 Be Here Corporation Imaging arrangement which allows for capturing an image of a view at different resolutions
CN1096387C (zh) * 1998-01-26 2002-12-18 塞克斯坦特飞行系统公司 具有电子式视野切换功能的飞行中摄像系统
US6621518B2 (en) 2001-04-25 2003-09-16 Denis R. Lietgeb Video surveillance system
US20040013300A1 (en) * 2002-07-17 2004-01-22 Lee Harry C. Algorithm selector
US6686955B1 (en) 1999-01-29 2004-02-03 International Business Machines Corporation Lightweight image manipulation for maximizing the capacity of digital photo storage media
US6766035B1 (en) 2000-05-03 2004-07-20 Koninklijke Philips Electronics N.V. Method and apparatus for adaptive position determination video conferencing and other applications
US20040202380A1 (en) * 2001-03-05 2004-10-14 Thorsten Kohler Method and device for correcting an image, particularly for occupant protection
US20050030406A1 (en) * 2003-07-04 2005-02-10 Koichi Yoshikawa Photographing apparatus, photographing method, and method for measuring converted data representing a fiber-array image
US6924832B1 (en) 1998-08-07 2005-08-02 Be Here Corporation Method, apparatus & computer program product for tracking objects in a warped video image
US20070263093A1 (en) * 2006-05-11 2007-11-15 Acree Elaine S Real-time capture and transformation of hemispherical video images to images in rectilinear coordinates
US7924310B1 (en) * 2006-05-31 2011-04-12 Hewlett-Packard Development Company, L.P. Image data processing methods, image data processing systems, and articles of manufacture
US20110216158A1 (en) * 2010-03-05 2011-09-08 Tessera Technologies Ireland Limited Object Detection and Rendering for Wide Field of View (WFOV) Image Acquisition Systems
US20120023281A1 (en) * 1993-09-17 2012-01-26 Shumpei Kawasaki Single-chip microcomputer
US8508652B2 (en) 2011-02-03 2013-08-13 DigitalOptics Corporation Europe Limited Autofocus method
WO2014005783A1 (en) 2012-07-03 2014-01-09 DigitalOptics Corporation Europe Limited A method and system for correcting a distorted input image
US8723959B2 (en) 2011-03-31 2014-05-13 DigitalOptics Corporation Europe Limited Face and other object tracking in off-center peripheral regions for nonlinear lens geometries
US8849064B2 (en) 2013-02-14 2014-09-30 Fotonation Limited Method and apparatus for viewing images
US8860816B2 (en) 2011-03-31 2014-10-14 Fotonation Limited Scene enhancements in off-center peripheral regions for nonlinear lens geometries
US8896703B2 (en) 2011-03-31 2014-11-25 Fotonation Limited Superresolution enhancment of peripheral regions in nonlinear lens geometries
US8982180B2 (en) 2011-03-31 2015-03-17 Fotonation Limited Face and other object detection and tracking in off-center peripheral regions for nonlinear lens geometries
US9412149B2 (en) 2011-02-10 2016-08-09 Panasonic Intellectual Property Management Co., Ltd. Display device, computer program, and computer-implemented method

Families Citing this family (125)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5764276A (en) * 1991-05-13 1998-06-09 Interactive Pictures Corporation Method and apparatus for providing perceived video viewing experiences using still images
US6731334B1 (en) * 1995-07-31 2004-05-04 Forgent Networks, Inc. Automatic voice tracking camera system and method of operation
US5757424A (en) * 1995-12-19 1998-05-26 Xerox Corporation High-resolution video conferencing system
JPH09289631A (ja) * 1996-02-20 1997-11-04 Canon Inc 撮像制御方法及び装置及び撮像システムと前記方法を実行するプログラムを記憶した記憶媒体
US5959667A (en) * 1996-05-09 1999-09-28 Vtel Corporation Voice activated camera preset selection system and method of operation
US5828838A (en) * 1996-06-20 1998-10-27 Intel Corporation Method and apparatus for conducting multi-point electronic conferences
US6009210A (en) * 1997-03-05 1999-12-28 Digital Equipment Corporation Hands-free interface to a virtual reality environment using head tracking
US6381367B1 (en) * 1997-05-29 2002-04-30 Macrovision Corp. Method and apparatus for compression compatible video fingerprinting
FR2770072B1 (fr) * 1997-08-19 2000-07-28 Serge Dulin Camera virtuelle
US6281931B1 (en) * 1997-11-04 2001-08-28 Tien Ren Tsao Method and apparatus for determining and correcting geometric distortions in electronic imaging systems
US6034716A (en) * 1997-12-18 2000-03-07 Whiting; Joshua B. Panoramic digital camera system
JP3157769B2 (ja) * 1998-03-02 2001-04-16 静岡日本電気株式会社 Tv音声制御装置
KR20010074470A (ko) 1998-04-02 2001-08-04 추후보정 카메라의 어레이를 이용한 가항성 텔레프레즌스 방법과시스템
US6522325B1 (en) * 1998-04-02 2003-02-18 Kewazinga Corp. Navigable telepresence method and system utilizing an array of cameras
US6593956B1 (en) * 1998-05-15 2003-07-15 Polycom, Inc. Locating an audio source
US6570606B1 (en) * 1998-05-29 2003-05-27 3Com Corporation Method and apparatus for controlling transmission of media signals over a data network in response to triggering events at participating stations
FR2782877B1 (fr) * 1998-08-31 2000-10-13 France Telecom Systeme automatique de prise de son et d'images
US6130704A (en) * 1998-10-22 2000-10-10 Sensormatics Electronics Corporation Controlling movement of video surveillance cameras
KR100779409B1 (ko) * 1998-11-11 2007-11-26 코닌클리케 필립스 일렉트로닉스 엔.브이. 향상된 신호 위치 측정 장치
US6128145A (en) * 1998-11-25 2000-10-03 Fit Corporation Image pick-up device, image display device and information recording medium comprising a fisheye lens
DE19854373B4 (de) * 1998-11-25 2005-02-24 Robert Bosch Gmbh Verfahren zur Steuerung der Empfindlichkeit eines Mikrofons
US7057636B1 (en) 1998-12-22 2006-06-06 Koninklijke Philips Electronics N.V. Conferencing system and method for the automatic determination of preset positions corresponding to participants in video-mediated communications
US6185152B1 (en) 1998-12-23 2001-02-06 Intel Corporation Spatial sound steering system
JP3126955B2 (ja) * 1999-02-12 2001-01-22 株式会社アドバネット 画像変換用演算装置
GB2351628B (en) * 1999-04-14 2003-10-01 Canon Kk Image and sound processing apparatus
US7117157B1 (en) 1999-03-26 2006-10-03 Canon Kabushiki Kaisha Processing apparatus for determining which person in a group is speaking
GB9908545D0 (en) * 1999-04-14 1999-06-09 Canon Kk Image processing apparatus
EP1047264B1 (de) 1999-04-22 2007-05-09 Leo Vision Vorrichtung und Verfahren zur Bildverarbeitung und -wiederherstellung mit erneuter Abtastung
US7653925B2 (en) 1999-11-17 2010-01-26 Ricoh Company, Ltd. Techniques for receiving information during multimedia presentations and communicating the information
US7299405B1 (en) 2000-03-08 2007-11-20 Ricoh Company, Ltd. Method and system for information management to facilitate the exchange of ideas during a collaborative effort
US6976032B1 (en) * 1999-11-17 2005-12-13 Ricoh Company, Ltd. Networked peripheral for visitor greeting, identification, biographical lookup and tracking
US6917693B1 (en) * 1999-12-20 2005-07-12 Ford Global Technologies, Llc Vehicle data acquisition and display assembly
US20020075295A1 (en) * 2000-02-07 2002-06-20 Stentz Anthony Joseph Telepresence using panoramic imaging and directional sound
JP2002083285A (ja) * 2000-07-07 2002-03-22 Matsushita Electric Ind Co Ltd 画像合成装置および画像合成方法
US6897892B2 (en) * 2000-10-13 2005-05-24 Alexander L. Kormos System and method for forming images for display in a vehicle
US6577333B2 (en) 2000-12-12 2003-06-10 Intel Corporation Automatic multi-camera video composition
US20020140804A1 (en) * 2001-03-30 2002-10-03 Koninklijke Philips Electronics N.V. Method and apparatus for audio/image speaker detection and locator
US20020149672A1 (en) * 2001-04-13 2002-10-17 Clapp Craig S.K. Modular video conferencing system
FR2826221B1 (fr) * 2001-05-11 2003-12-05 Immervision Internat Pte Ltd Procede d'obtention et d'affichage d'une image panoramique numerique a resolution variable
US6937266B2 (en) * 2001-06-14 2005-08-30 Microsoft Corporation Automated online broadcasting system and method using an omni-directional camera system for viewing meetings over a computer network
FR2827680B1 (fr) * 2001-07-20 2003-10-10 Immervision Internat Pte Ltd Procede de capture d'une image panoramique au moyen d'un capteur d'image de forme rectangulaire
JP2004538724A (ja) * 2001-08-07 2004-12-24 ポリコム・インコーポレイテッド 高解像度のテレビ会議のシステム及び方法
US6959095B2 (en) * 2001-08-10 2005-10-25 International Business Machines Corporation Method and apparatus for providing multiple output channels in a microphone
US7668907B1 (en) 2001-08-16 2010-02-23 Microsoft Corporation Method and system for selectively viewing participants of a multimedia network conference
WO2003019471A2 (en) * 2001-08-25 2003-03-06 Eyesee360,Inc. Method and apparatus for encoding photogrraphic images
FR2830128A1 (fr) * 2001-09-26 2003-03-28 Egg Solution Sa Capteur photoelectrique avec des elements de capture de pixels disposes en cercles concentriques et dispositif d'acquisition d'images panoramiques comprenant ce capteur
US6980485B2 (en) * 2001-10-25 2005-12-27 Polycom, Inc. Automatic camera tracking using beamforming
US7747655B2 (en) 2001-11-19 2010-06-29 Ricoh Co. Ltd. Printable representations for time-based media
US7861169B2 (en) 2001-11-19 2010-12-28 Ricoh Co. Ltd. Multimedia print driver dialog interfaces
US6812956B2 (en) 2001-12-21 2004-11-02 Applied Minds, Inc. Method and apparatus for selection of signals in a teleconference
US6789901B1 (en) 2002-01-04 2004-09-14 Raytheon Company System and method for providing images for an operator of a vehicle
US20040021764A1 (en) * 2002-01-28 2004-02-05 Be Here Corporation Visual teleconferencing apparatus
US20040008423A1 (en) * 2002-01-28 2004-01-15 Driscoll Edward C. Visual teleconferencing apparatus
US20030220971A1 (en) * 2002-05-23 2003-11-27 International Business Machines Corporation Method and apparatus for video conferencing with audio redirection within a 360 degree view
US20040001091A1 (en) * 2002-05-23 2004-01-01 International Business Machines Corporation Method and apparatus for video conferencing system with 360 degree view
US6808274B2 (en) * 2002-06-04 2004-10-26 Raytheon Company Method and system for deploying a mirror assembly from a recessed position
US6815680B2 (en) * 2002-06-05 2004-11-09 Raytheon Company Method and system for displaying an image
US7146014B2 (en) 2002-06-11 2006-12-05 Intel Corporation MEMS directional sensor system
US7058237B2 (en) * 2002-06-28 2006-06-06 Microsoft Corporation Real-time wide-angle image correction system and method for computer image viewing
FI118247B (fi) * 2003-02-26 2007-08-31 Fraunhofer Ges Forschung Menetelmä luonnollisen tai modifioidun tilavaikutelman aikaansaamiseksi monikanavakuuntelussa
US7552381B2 (en) * 2003-03-31 2009-06-23 Ricoh Co., Ltd. Check boxes for identifying and processing stored documents
US7509569B2 (en) 2003-03-31 2009-03-24 Ricoh Co., Ltd. Action stickers for nested collections
US7536638B2 (en) 2003-03-31 2009-05-19 Ricoh Co., Ltd. Action stickers for identifying and processing stored documents
US7703002B2 (en) 2003-03-31 2010-04-20 Ricoh Company, Ltd. Method and apparatus for composing multimedia documents
US20070050696A1 (en) * 2003-03-31 2007-03-01 Piersol Kurt W Physical key for accessing a securely stored digital document
US7757162B2 (en) 2003-03-31 2010-07-13 Ricoh Co. Ltd. Document collection manipulation
US7739583B2 (en) 2003-03-31 2010-06-15 Ricoh Company, Ltd. Multimedia document sharing method and apparatus
US7275159B2 (en) 2003-08-11 2007-09-25 Ricoh Company, Ltd. Multimedia output device having embedded encryption functionality
JP2005086365A (ja) * 2003-09-05 2005-03-31 Sony Corp 通話装置、会議装置および撮像条件調整方法
US7590941B2 (en) * 2003-10-09 2009-09-15 Hewlett-Packard Development Company, L.P. Communication and collaboration system using rich media environments
FR2861525B1 (fr) * 2003-10-24 2006-04-28 Winlight System Finance Procede et dispositif de capture d'une image grand champ et d'une region d'interet de celle-ci
US7689712B2 (en) 2003-11-26 2010-03-30 Ricoh Company, Ltd. Techniques for integrating note-taking and multimedia information
DE60336514D1 (de) 2003-12-23 2011-05-05 Telecom Italia Spa Vorrichtung zum betrachten von bildern für ein vid und verfahren zum gebrauch
GB0405455D0 (en) * 2004-03-11 2004-04-21 Mitel Networks Corp High precision beamsteerer based on fixed beamforming approach beampatterns
US20050221852A1 (en) * 2004-04-05 2005-10-06 D Avello Robert F Methods for controlling processing of inputs to a vehicle wireless communication interface
WO2005125186A1 (en) * 2004-06-15 2005-12-29 Quantum Semiconductor Llc Imaging devices operable with multiple aspect ratios
US7623156B2 (en) * 2004-07-16 2009-11-24 Polycom, Inc. Natural pan tilt zoom camera motion to preset camera positions
US8730322B2 (en) 2004-07-30 2014-05-20 Eyesee360, Inc. Telepresence using panoramic imaging and directional sound and motion
US7500795B2 (en) * 2004-09-09 2009-03-10 Paul Sandhu Apparatuses, systems and methods for enhancing telemedicine, video-conferencing, and video-based sales
KR100754385B1 (ko) * 2004-09-30 2007-08-31 삼성전자주식회사 오디오/비디오 센서를 이용한 위치 파악, 추적 및 분리장치와 그 방법
JP4729927B2 (ja) * 2005-01-11 2011-07-20 ソニー株式会社 音声検出装置、自動撮像装置、および音声検出方法
US7524119B2 (en) * 2005-02-03 2009-04-28 Paul Sandhu Apparatus and method for viewing radiographs
US8386909B2 (en) * 2005-04-07 2013-02-26 Hewlett-Packard Development Company, L.P. Capturing and presenting interactions with image-based media
US8457614B2 (en) * 2005-04-07 2013-06-04 Clearone Communications, Inc. Wireless multi-unit conference phone
US8805929B2 (en) 2005-06-20 2014-08-12 Ricoh Company, Ltd. Event-driven annotation techniques
US7554576B2 (en) 2005-06-20 2009-06-30 Ricoh Company, Ltd. Information capture and recording system for controlling capture devices
JP4356663B2 (ja) * 2005-08-17 2009-11-04 ソニー株式会社 カメラ制御装置および電子会議システム
US8036402B2 (en) 2005-12-15 2011-10-11 Harman International Industries, Incorporated Distortion compensation
JP2007318438A (ja) * 2006-05-25 2007-12-06 Yamaha Corp 音声状況データ生成装置、音声状況可視化装置、音声状況データ編集装置、音声データ再生装置、および音声通信システム
AU2007221976B2 (en) * 2006-10-19 2009-12-24 Polycom, Inc. Ultrasonic camera tracking system and associated methods
US20080255840A1 (en) * 2007-04-16 2008-10-16 Microsoft Corporation Video Nametags
CN100505837C (zh) * 2007-05-10 2009-06-24 华为技术有限公司 一种控制图像采集装置进行目标定位的系统及方法
US8526632B2 (en) * 2007-06-28 2013-09-03 Microsoft Corporation Microphone array for a camera speakerphone
US8330787B2 (en) * 2007-06-29 2012-12-11 Microsoft Corporation Capture device movement compensation for speaker indexing
US8165416B2 (en) * 2007-06-29 2012-04-24 Microsoft Corporation Automatic gain and exposure control using region of interest detection
JP4458128B2 (ja) * 2007-07-31 2010-04-28 ソニー株式会社 方向検出装置、方向検出方法および方向検出プログラム、ならびに、方向制御装置、方向制御方法および方向制御プログラム
US9445193B2 (en) * 2008-07-31 2016-09-13 Nokia Technologies Oy Electronic device directional audio capture
US10904658B2 (en) 2008-07-31 2021-01-26 Nokia Technologies Oy Electronic device directional audio-video capture
CN102113319A (zh) * 2008-08-04 2011-06-29 皇家飞利浦电子股份有限公司 具有周边观看装置的通信设备
JP5317630B2 (ja) * 2008-11-07 2013-10-16 キヤノン株式会社 画像配信装置、方法及びプログラム
US20100123785A1 (en) * 2008-11-17 2010-05-20 Apple Inc. Graphic Control for Directional Audio Input
US8380866B2 (en) 2009-03-20 2013-02-19 Ricoh Company, Ltd. Techniques for facilitating annotations
US8831681B1 (en) * 2010-01-04 2014-09-09 Marvell International Ltd. Image guided audio processing
US8698872B2 (en) * 2011-03-02 2014-04-15 At&T Intellectual Property I, Lp System and method for notification of events of interest during a video conference
US20130201272A1 (en) * 2012-02-07 2013-08-08 Niklas Enbom Two mode agc for single and multiple speakers
US9323912B2 (en) * 2012-02-28 2016-04-26 Verizon Patent And Licensing Inc. Method and system for multi-factor biometric authentication
US9612663B2 (en) 2012-03-26 2017-04-04 Tata Consultancy Services Limited Multimodal system and method facilitating gesture creation through scalar and vector data
US9591418B2 (en) 2012-04-13 2017-03-07 Nokia Technologies Oy Method, apparatus and computer program for generating an spatial audio output based on an spatial audio input
US9779093B2 (en) * 2012-12-19 2017-10-03 Nokia Technologies Oy Spatial seeking in media files
JP6253031B2 (ja) * 2013-02-15 2017-12-27 パナソニックIpマネジメント株式会社 キャリブレーション方法
CN105144768B (zh) 2013-04-26 2019-05-21 英特尔Ip公司 频谱共享情境中的共享频谱重新分配
DE102013010233B4 (de) 2013-06-18 2018-08-30 Volkswagen Aktiengesellschaft Verfahren zum Anzeigen von Umgebungsinformationen in einem Fahrzeug und Anzeigesystem für ein Fahrzeug
US9753119B1 (en) * 2014-01-29 2017-09-05 Amazon Technologies, Inc. Audio and depth based sound source localization
US10586114B2 (en) * 2015-01-13 2020-03-10 Vivint, Inc. Enhanced doorbell camera interactions
CN109074632B (zh) 2016-02-16 2022-03-01 6115187加拿大公司暨伊美景象公司 图像失真变换方法和设备
US10531187B2 (en) 2016-12-21 2020-01-07 Nortek Security & Control Llc Systems and methods for audio detection using audio beams
US10560636B1 (en) * 2017-03-24 2020-02-11 Cisco Technology, Inc. Crop synchronized auto focus and exposure
JP7292853B2 (ja) 2017-12-26 2023-06-19 キヤノン株式会社 撮像装置及びその制御方法及びプログラム
JP2019117375A (ja) * 2017-12-26 2019-07-18 キヤノン株式会社 撮像装置及びその制御方法及びプログラム
US10951859B2 (en) 2018-05-30 2021-03-16 Microsoft Technology Licensing, Llc Videoconferencing device and method
US11102395B2 (en) * 2019-08-16 2021-08-24 Sony Corporation Generation of media content having multi-focus video and multi-directional audio
TWI902903B (zh) * 2021-09-17 2025-11-01 信驊科技股份有限公司 視訊內容提供方法及視訊內容提供裝置
WO2023049773A1 (en) 2021-09-21 2023-03-30 Shure Acquisition Holdings, Inc. Conferencing systems and methods for room intelligence
JP2025520891A (ja) 2022-06-30 2025-07-03 シュアー アクイジッション ホールディングス インコーポレイテッド 話者追跡およびカメラ位置付けのための会議システムおよび方法
WO2024039892A1 (en) 2022-08-19 2024-02-22 Shure Acquisition Holdings, Inc. System and method for camera motion stabilization using audio localization

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4070098A (en) * 1976-04-28 1978-01-24 The University Of Arizona Foundation Fisheye projection lens system for 35mm motion pictures
US4125862A (en) * 1977-03-31 1978-11-14 The United States Of America As Represented By The Secretary Of The Navy Aspect ratio and scan converter system
US4170400A (en) * 1977-07-05 1979-10-09 Bert Bach Wide angle view optical system
US4322726A (en) * 1979-12-19 1982-03-30 The Singer Company Apparatus for providing a simulated view to hand held binoculars
US4559555A (en) * 1982-02-24 1985-12-17 Arnold Schoolman Stereoscopic remote viewing system
US4747675A (en) * 1987-06-01 1988-05-31 Albert Nagler Ultra wide-angle eyepiece
US4772942A (en) * 1986-01-11 1988-09-20 Pilkington P.E. Limited Display system having wide field of view
US4985762A (en) * 1988-12-14 1991-01-15 Horisonscan Inc. Panoramic interactive system
US5023725A (en) * 1989-10-23 1991-06-11 Mccutchen David Method and apparatus for dodecahedral imaging system
US5040055A (en) * 1988-12-14 1991-08-13 Horizonscan Inc. Panoramic interactive system
US5097325A (en) * 1990-12-17 1992-03-17 Eol3 Company, Inc. Circular scanning system for an integrated camera and panoramic catadioptric display
US5185667A (en) * 1991-05-13 1993-02-09 Telerobotics International, Inc. Omniview motionless camera orientation system

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4264928A (en) * 1979-11-05 1981-04-28 Schober Gary W Conference video system
US4442453A (en) 1982-02-09 1984-04-10 The United States Of America As Represented By The Secretary Of The Air Force Photo-reconnaissance system
US4832472A (en) 1987-07-17 1989-05-23 Lockheed Missiles & Space Company, Inc. Mapping lens
GB8722403D0 (en) * 1987-09-23 1987-10-28 Secretary Trade Ind Brit Automatic vehicle guidance systems
JPH0771279B2 (ja) * 1988-08-17 1995-07-31 富士通株式会社 テレビ会議用画像処理装置
JPH05346950A (ja) * 1991-12-19 1993-12-27 Eastman Kodak Co 三次元シーンを感知するための方法及び装置
FR2692375B1 (fr) * 1992-06-12 1994-12-23 Thomson Csf Dispositif de veille omnidirectionnel à objectif grand angulaire couvrant tout l'espace environnement.
US5508734A (en) * 1994-07-27 1996-04-16 International Business Machines Corporation Method and apparatus for hemispheric imaging which emphasizes peripheral content
JPH1163079A (ja) * 1997-08-19 1999-03-05 Showa:Kk フロントフォーク

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4070098A (en) * 1976-04-28 1978-01-24 The University Of Arizona Foundation Fisheye projection lens system for 35mm motion pictures
US4125862A (en) * 1977-03-31 1978-11-14 The United States Of America As Represented By The Secretary Of The Navy Aspect ratio and scan converter system
US4170400A (en) * 1977-07-05 1979-10-09 Bert Bach Wide angle view optical system
US4322726A (en) * 1979-12-19 1982-03-30 The Singer Company Apparatus for providing a simulated view to hand held binoculars
US4559555A (en) * 1982-02-24 1985-12-17 Arnold Schoolman Stereoscopic remote viewing system
US4772942A (en) * 1986-01-11 1988-09-20 Pilkington P.E. Limited Display system having wide field of view
US4747675A (en) * 1987-06-01 1988-05-31 Albert Nagler Ultra wide-angle eyepiece
US4985762A (en) * 1988-12-14 1991-01-15 Horisonscan Inc. Panoramic interactive system
US5040055A (en) * 1988-12-14 1991-08-13 Horizonscan Inc. Panoramic interactive system
US5023725A (en) * 1989-10-23 1991-06-11 Mccutchen David Method and apparatus for dodecahedral imaging system
US5097325A (en) * 1990-12-17 1992-03-17 Eol3 Company, Inc. Circular scanning system for an integrated camera and panoramic catadioptric display
US5185667A (en) * 1991-05-13 1993-02-09 Telerobotics International, Inc. Omniview motionless camera orientation system

Cited By (76)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5815199A (en) * 1991-01-31 1998-09-29 Matsushita Electric Works, Ltd. Interphone with television
US6301447B1 (en) * 1991-05-13 2001-10-09 Interactive Pictures Corporation Method and system for creation and interactive viewing of totally immersive stereoscopic images
US20120023281A1 (en) * 1993-09-17 2012-01-26 Shumpei Kawasaki Single-chip microcomputer
US5686957A (en) * 1994-07-27 1997-11-11 International Business Machines Corporation Teleconferencing imaging system with automatic camera steering
US5703605A (en) * 1994-12-05 1997-12-30 Olympus Optical Co., Ltd. Image display apparatus
US6493031B1 (en) * 1995-03-31 2002-12-10 Canon Kabushiki Kaisha Visual information processing method and apparatus for extracting feature quantities from a two-dimensional image signal
US5870135A (en) * 1995-07-27 1999-02-09 Sensormatic Electronics Corporation Image splitting forming and processing device and method for use with no moving parts camera
US20030193607A1 (en) * 1996-06-24 2003-10-16 Be Here Corporation Panoramic camera
US6493032B1 (en) 1996-06-24 2002-12-10 Be Here Corporation Imaging arrangement which allows for capturing an image of a view at different resolutions
US6593969B1 (en) 1996-06-24 2003-07-15 Be Here Corporation Preparing a panoramic image for presentation
US6583815B1 (en) 1996-06-24 2003-06-24 Be Here Corporation Method and apparatus for presenting images from a remote location
US7242425B2 (en) 1996-06-24 2007-07-10 Be Here, Corporation Panoramic camera
US6337708B1 (en) 1996-06-24 2002-01-08 Be Here Corporation Method and apparatus for electronically distributing motion panoramic images
US6341044B1 (en) 1996-06-24 2002-01-22 Be Here Corporation Panoramic imaging arrangement
EP1178352A1 (de) * 1996-06-24 2002-02-06 Behere Corporation Verfahren und Vorrichtung zur Präsentation von Panoramabildern auf einem lokalen Empfänger, sowie zugehöriges Computerprogramm
US7486324B2 (en) 1996-06-24 2009-02-03 B.H. Image Co. Llc Presenting panoramic images with geometric transformation
WO1997050252A1 (en) * 1996-06-24 1997-12-31 Behere Corporation Panoramic camera
US6426774B1 (en) 1996-06-24 2002-07-30 Be Here Corporation Panoramic camera
US6459451B2 (en) 1996-06-24 2002-10-01 Be Here Corporation Method and apparatus for a panoramic camera to capture a 360 degree image
US20030193606A1 (en) * 1996-06-24 2003-10-16 Be Here Corporation Panoramic camera
US6542184B1 (en) 1996-06-24 2003-04-01 Edward Driscoll, Jr. Methods, apparatus, and program products for presenting panoramic images of a remote location
USRE44087E1 (en) 1996-06-24 2013-03-19 B.H. Image Co. Llc Presenting panoramic images with geometric transformation
US6515696B1 (en) 1996-06-24 2003-02-04 Be Here Corporation Method and apparatus for presenting images from a remote location
US6480229B1 (en) 1996-06-24 2002-11-12 Be Here Corporation Panoramic camera
US6373642B1 (en) 1996-06-24 2002-04-16 Be Here Corporation Panoramic imaging arrangement
US6043837A (en) * 1997-05-08 2000-03-28 Be Here Corporation Method and apparatus for electronically distributing images from a panoptic camera system
US6466254B1 (en) 1997-05-08 2002-10-15 Be Here Corporation Method and apparatus for electronically distributing motion panoramic images
US6392687B1 (en) 1997-05-08 2002-05-21 Be Here Corporation Method and apparatus for implementing a panoptic camera system
US6313865B1 (en) 1997-05-08 2001-11-06 Be Here Corporation Method and apparatus for implementing a panoptic camera system
US6219089B1 (en) 1997-05-08 2001-04-17 Be Here Corporation Method and apparatus for electronically distributing images from a panoptic camera system
US7280134B1 (en) * 1998-01-26 2007-10-09 Thales Avionics, Inc. Landscape camera system with electronic field of view switching
CN1096387C (zh) * 1998-01-26 2002-12-18 塞克斯坦特飞行系统公司 具有电子式视野切换功能的飞行中摄像系统
AU758465B2 (en) * 1998-01-26 2003-03-20 Thales Avionics In-Flight Systems, Llc In flight camera system with electronic field of view switching
US6331869B1 (en) 1998-08-07 2001-12-18 Be Here Corporation Method and apparatus for electronically distributing motion panoramic images
US6924832B1 (en) 1998-08-07 2005-08-02 Be Here Corporation Method, apparatus & computer program product for tracking objects in a warped video image
US6369818B1 (en) 1998-11-25 2002-04-09 Be Here Corporation Method, apparatus and computer program product for generating perspective corrected data from warped information
US6222683B1 (en) 1999-01-13 2001-04-24 Be Here Corporation Panoramic imaging arrangement
US6686955B1 (en) 1999-01-29 2004-02-03 International Business Machines Corporation Lightweight image manipulation for maximizing the capacity of digital photo storage media
US7015954B1 (en) 1999-08-09 2006-03-21 Fuji Xerox Co., Ltd. Automatic video system using multiple cameras
US7277118B2 (en) 1999-08-09 2007-10-02 Fuji Xerox Co., Ltd. Method and system for compensating for parallax in multiple camera systems
US20020122113A1 (en) * 1999-08-09 2002-09-05 Foote Jonathan T. Method and system for compensating for parallax in multiple camera systems
US20060125921A1 (en) * 1999-08-09 2006-06-15 Fuji Xerox Co., Ltd. Method and system for compensating for parallax in multiple camera systems
US7710463B2 (en) 1999-08-09 2010-05-04 Fuji Xerox Co., Ltd. Method and system for compensating for parallax in multiple camera systems
US7436433B2 (en) * 2000-04-25 2008-10-14 Rainer Mockler Digital camera or digital video camera and method for acquiring images
DE10191605B4 (de) * 2000-04-25 2005-08-11 Rainer Mockler Digitalkamera oder digitale Videokamera
WO2001082592A1 (de) * 2000-04-25 2001-11-01 Rainer Mockler Digitalkamera oder digitale videokamera und verfahren zur bilderfassung
US20030164880A1 (en) * 2000-04-25 2003-09-04 Rainer Mockler Digital camera or digital video camera and method for acquiring images
US6766035B1 (en) 2000-05-03 2004-07-20 Koninklijke Philips Electronics N.V. Method and apparatus for adaptive position determination video conferencing and other applications
US20040202380A1 (en) * 2001-03-05 2004-10-14 Thorsten Kohler Method and device for correcting an image, particularly for occupant protection
US7304680B2 (en) * 2001-03-05 2007-12-04 Siemens Aktiengesellschaft Method and device for correcting an image, particularly for occupant protection
US7173658B2 (en) * 2001-03-26 2007-02-06 Olympus Corporation Arranging a plurality of imaging pixels in an imaging acquisition apparatus
US20020180877A1 (en) * 2001-03-26 2002-12-05 Susumu Kikuchi Image acquisition apparatus
US20020147991A1 (en) * 2001-04-10 2002-10-10 Furlan John L. W. Transmission of panoramic video via existing video infrastructure
US6621518B2 (en) 2001-04-25 2003-09-16 Denis R. Lietgeb Video surveillance system
US20040013300A1 (en) * 2002-07-17 2004-01-22 Lee Harry C. Algorithm selector
US7184590B2 (en) * 2002-07-17 2007-02-27 Lockheed Martin Corporation Algorithm selector
US7477312B2 (en) * 2003-07-04 2009-01-13 Sony Corporation Photographing apparatus, photographing method, and method for measuring converted data representing a fiber-array image
US20050030406A1 (en) * 2003-07-04 2005-02-10 Koichi Yoshikawa Photographing apparatus, photographing method, and method for measuring converted data representing a fiber-array image
US20070263093A1 (en) * 2006-05-11 2007-11-15 Acree Elaine S Real-time capture and transformation of hemispherical video images to images in rectilinear coordinates
US8160394B2 (en) * 2006-05-11 2012-04-17 Intergraph Software Technologies, Company Real-time capture and transformation of hemispherical video images to images in rectilinear coordinates
US7924310B1 (en) * 2006-05-31 2011-04-12 Hewlett-Packard Development Company, L.P. Image data processing methods, image data processing systems, and articles of manufacture
US8692867B2 (en) 2010-03-05 2014-04-08 DigitalOptics Corporation Europe Limited Object detection and rendering for wide field of view (WFOV) image acquisition systems
US20110216158A1 (en) * 2010-03-05 2011-09-08 Tessera Technologies Ireland Limited Object Detection and Rendering for Wide Field of View (WFOV) Image Acquisition Systems
US8872887B2 (en) 2010-03-05 2014-10-28 Fotonation Limited Object detection and rendering for wide field of view (WFOV) image acquisition systems
US8508652B2 (en) 2011-02-03 2013-08-13 DigitalOptics Corporation Europe Limited Autofocus method
US9412149B2 (en) 2011-02-10 2016-08-09 Panasonic Intellectual Property Management Co., Ltd. Display device, computer program, and computer-implemented method
US11651471B2 (en) 2011-02-10 2023-05-16 Panasonic Intellectual Property Management Co., Ltd. Display device, computer program, and computer-implemented method
US8723959B2 (en) 2011-03-31 2014-05-13 DigitalOptics Corporation Europe Limited Face and other object tracking in off-center peripheral regions for nonlinear lens geometries
US8860816B2 (en) 2011-03-31 2014-10-14 Fotonation Limited Scene enhancements in off-center peripheral regions for nonlinear lens geometries
US8896703B2 (en) 2011-03-31 2014-11-25 Fotonation Limited Superresolution enhancment of peripheral regions in nonlinear lens geometries
US8947501B2 (en) 2011-03-31 2015-02-03 Fotonation Limited Scene enhancements in off-center peripheral regions for nonlinear lens geometries
US8982180B2 (en) 2011-03-31 2015-03-17 Fotonation Limited Face and other object detection and tracking in off-center peripheral regions for nonlinear lens geometries
WO2014005783A1 (en) 2012-07-03 2014-01-09 DigitalOptics Corporation Europe Limited A method and system for correcting a distorted input image
US8928730B2 (en) 2012-07-03 2015-01-06 DigitalOptics Corporation Europe Limited Method and system for correcting a distorted input image
US9262807B2 (en) 2012-07-03 2016-02-16 Fotonation Limited Method and system for correcting a distorted input image
US8849064B2 (en) 2013-02-14 2014-09-30 Fotonation Limited Method and apparatus for viewing images

Also Published As

Publication number Publication date
CA2152314C (en) 2000-02-29
US5686957A (en) 1997-11-11
EP0695085B1 (de) 2000-02-16
JP2001091825A (ja) 2001-04-06
BR9502919A (pt) 1996-03-05
EP0695085A1 (de) 1996-01-31
DE69515087D1 (de) 2000-03-23
JPH0855215A (ja) 1996-02-27
DE69515087T2 (de) 2000-09-14
JP3485261B2 (ja) 2004-01-13
JP3103008B2 (ja) 2000-10-23
CA2152314A1 (en) 1996-01-28

Similar Documents

Publication Publication Date Title
US5508734A (en) Method and apparatus for hemispheric imaging which emphasizes peripheral content
JP3290993B2 (ja) 球形画像を作成する方法および装置
US7123777B2 (en) System and method for panoramic imaging
US6118474A (en) Omnidirectional imaging apparatus
US6459451B2 (en) Method and apparatus for a panoramic camera to capture a 360 degree image
JP4023827B2 (ja) 全方向性撮像装置
JP4247113B2 (ja) 長方形の画像センサによりパノラマ画像を撮像する方法
US5586231A (en) Method and device for processing an image in order to construct from a source image a target image with charge of perspective
KR100967826B1 (ko) 화상처리장치 및 방법, 프로그램 및 프로그램 기록매체,및 데이터구조 및 데이터 기록매체
US20090040291A1 (en) Omniview motionless camera orientation system
US20030095338A1 (en) System and method for panoramic imaging
JP3922543B2 (ja) 撮像装置、および画像表示装置
EP0448630A1 (de) Panoramisches interaktives system
CA2324802C (en) Panoramic imaging apparatus and method
KR200378727Y1 (ko) 전방위 파노라마 재생장치
CA2251879C (en) Omnidirectional imaging apparatus
JPH09261523A (ja) 画像入力装置及び画像記録再生装置

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORP., NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BAKER, ROBERT GROVER;KETTLER, KEVIN;SUAREZ, GUSTAVO ARMANDO;AND OTHERS;REEL/FRAME:007231/0878;SIGNING DATES FROM 19940727 TO 19940816

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 8

AS Assignment

Owner name: MEDIATEK INC., TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:017083/0887

Effective date: 20051228

FPAY Fee payment

Year of fee payment: 12

FEPP Fee payment procedure

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY