US10063848B2 - Perspective altering display system - Google Patents

Perspective altering display system Download PDF

Info

Publication number
US10063848B2
US10063848B2 US12/197,635 US19763508A US10063848B2 US 10063848 B2 US10063848 B2 US 10063848B2 US 19763508 A US19763508 A US 19763508A US 10063848 B2 US10063848 B2 US 10063848B2
Authority
US
United States
Prior art keywords
viewer
scene
operative
camera
display system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/197,635
Other versions
US20090051699A1 (en
Inventor
John G. Posa
Barry H. Schwab
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Videa LLC
Original Assignee
Videa LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US95784507P priority Critical
Application filed by Videa LLC filed Critical Videa LLC
Priority to US12/197,635 priority patent/US10063848B2/en
Assigned to VIDEA, LLC reassignment VIDEA, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: POSA, JOHN G., SCHWAB, BARRY H.
Publication of US20090051699A1 publication Critical patent/US20090051699A1/en
Application granted granted Critical
Publication of US10063848B2 publication Critical patent/US10063848B2/en
Application status is Active legal-status Critical
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/275Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals
    • H04N13/0468
    • H04N13/0278
    • H04N13/047
    • H04N13/0475
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/275Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals
    • H04N13/279Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals the virtual viewpoint locations being selected by the viewers or determined by tracking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking
    • H04N13/368Image reproducers using viewer tracking for two or more viewers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking
    • H04N13/373Image reproducers using viewer tracking for tracking forward-backward translational head movements, i.e. longitudinal movements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42202Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] environmental sensors, e.g. for detecting temperature, luminosity, pressure, earthquakes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/4223Cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network, synchronizing decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network, synchronizing decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44218Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/10Mixing of images, i.e. displayed pixel being the result of an operation, e.g. adding, on the corresponding input pixels
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/12Overlay of images, i.e. displayed pixel being the result of switching between the corresponding input pixels
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/12Overlay of images, i.e. displayed pixel being the result of switching between the corresponding input pixels
    • G09G2340/125Overlay of images, i.e. displayed pixel being the result of switching between the corresponding input pixels wherein one of the images is motion video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry
    • H04N5/4403User interfaces for controlling a television receiver or set top box [STB] through a remote control device, e.g. graphical user interfaces [GUI]; Remote control devices therefor
    • H04N2005/4405Hardware details of remote control devices
    • H04N2005/4428Non-standard components, e.g. timer, speaker, sensors for detecting position, direction or movement of the remote control, microphone, battery charging device
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2213/00Details of stereoscopic systems
    • H04N2213/006Pseudo-stereoscopic systems, i.e. systems wherein a stereoscopic effect is obtained without sending different images to the viewer's eyes

Abstract

The perception of a displayed image is altered for viewers moving relative to the position of the display system screen, thereby imparting a sense of three-dimensional immersion in the scene being displayed. A display generator generates a scene having foreground and background elements, and a display screen displaying the scene. A sensor detects the position of a viewer relative to the display screen, and a processor is operative to shift the relative position of the foreground and background elements in the displayed scene as a function of viewer position, such that the viewer's perspective of the scene changes as the viewer moves relative to the display screen. The foreground and background elements may be presented in the form of multiple superimposed graphics planes, and/or a camera may be used to record the scene through panning at sequential angles. The system may be used to implement virtual windows, virtual mirrors and other effects without requiring viewers or users to modify behavior or wear glasses, beacons, etc.

Description

REFERENCE TO RELATED APPLICATION

This application claims priority from U.S. Provisional Patent Application Ser. No. 60/957,845, filed Aug. 24, 2007, the entire content of which is incorporated herein by reference.

FIELD OF THE INVENTION

This invention relates generally to display systems and, more particularly, to systems that alter perspective, synthesize depth perception and provide other capabilities, thereby enhancing the viewing experience.

BACKGROUND OF THE INVENTION

Flat panel displays are growing in size and falling in price. At this time, non-projection, true HD (i.e.—1080 p) 50″ displays are available in the $1,000 to $2,000 price range, and new models are introduced on a regular basis. When edge-lit liquid-crystal display (LCD) panels are replaced with back-lit white light-emitting diodes, another leap in technology will occur. When organic LED panels become viable, flexible, affordable wall-sized displays, including wrap-around configurations, should be possible.

As large displays proliferate, users are finding uses for them beyond just “watching TV.” In some cases, for example, users are displaying pictures or rotating sequences of pictures on these displays, thereby creating, in effect, large picture frames. However, existing systems for video display are restricted either in their utility or in their realism, due to the inherent limitations in a two-dimensional presentation unit. Of these, the most important impact to the perception of a scene may be its lack of visual perspective as the viewer changes his position relative to the display unit.

The use of 3-D glasses or other paraphernalia is oppressive in those circumstances where a casual effect is desired, or where no user intervention is to be required. Thus, prior-art systems relying on still or even video images, or various photographic or video projection techniques, cannot achieve the level of perception of reality that is desired for many residential or commercial display applications.

SUMMARY OF THE INVENTION

This invention resides in apparatus and methods providing a unique experience for the viewer of a display, particularly large wall-mounted panels. In the preferred embodiments, the perception of a displayed image is altered for viewers moving relative to the position of the display system screen, thereby imparting a sense of three-dimensional immersion in the scene being displayed.

A perspective-altering display system according to the invention comprises a display generator for generating a scene having foreground and background elements, and a display screen displaying the scene. A sensor detects the position of a viewer relative to the display screen, and a processor is operative to shift the relative position of the foreground and background elements in the displayed scene as a function of viewer position, such that the viewer's perspective of the scene changes as the viewer moves relative to the display screen.

The foreground and background elements may be presented in the form of multiple superimposed graphics planes. At least one of the graphics planes may include prerecorded material or material received through a transmission medium or camera. A camera may be used to record the scene through panning at sequential angles, with a memory being used to store the images obtained at the sequential angles for later recall as a function of user movement. An interpolator may be used to “fill in” visual gaps in the scene.

The display screen may be mounted on a wall having a backside, and a camera may be mounted on the backside of the wall which pans as a viewer moves, thereby imaging a scene representative of the display being a virtual window through the wall. A camera with a zoom capability may be used for recording the scene. With the sensor being operative to detect the viewer's distance from the display screen, the processor is further operative to zoom in the camera as the viewer moves closer to the display screen, and zoom out the camera as the viewer moves away from the display screen.

The sensor may be operative to detect the viewer's distance from the display screen, with the processor being operative to increase the resolution of the scene as the viewer moves closer to the display screen and decrease the resolution of the scene as the viewer moves away from the display screen. A camera with a tilt capability may be used for recording the scene, with the processor being further operative to tilt the camera in response to the viewer's up/down movement. The camera may have a field of view which includes a viewer of the display screen, enabling the display to function a virtual mirror.

A plurality of cameras may be used for capturing the scene, with the processor being further operative to construct a three-dimensional image for display on the screen. A user control may be provided enabling a viewer to select a specific camera or cameras to see how others would view the user from different perspectives.

The sensor may include an infrared CCD (charge-coupled device) camera. Regardless, the camera may have an image sensor, such that an image of a person in front of the display may be focused onto the image sensor as a spot or group of pixels, allowing the movements of the person to be tracked with no moving parts.

The sensor may be a camera having at least a pan mount that tracks the movement of a viewer, with the processor being operative to shift the relative position of the foreground and background elements as a function of the tracking. The sensor may be a camera having at least a pan/tilt mount that tracks the movement of a viewer, with the processor being operative to shift the relative position of the foreground and background elements as a function of the tracking. The sensor may be a camera having an auto-focus capability.

The sensor may be operative to sense a plurality of individuals in front of the display screen. The processor may be operative to shift the relative position of the foreground and background elements by favoring larger clusters of individuals as opposed to smaller clusters or single individuals. Alternatively, moving individuals may be favored over stationary individuals, or individuals actually looking at the display screen may be favored over those who are not.

One or more transducers may be provided for producing sounds associated with the scene, with the processor operative to alter the reproduction of the sounds as a function of viewer movement.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1A shows a viewer having a field of view walking from the right towards the left relative to a display screen (shown as viewed from above);

FIG. 1B is a simplified representation of what the person might see on the display screen according to the invention from the position shown in FIG. 1A;

FIG. 2A shows the person moving to the left, closer to the center of the display screen;

FIG. 2B shows how closer objects have shifted laterally to the greatest degree, as opposed to more distant objects, much as a train passenger would experience while looking out the window of the train as the train moves;

FIG. 3A illustrates how the viewer has moved to the left-most portion of the display;

FIG. 3B shows how close objects have shifted to the extent that they now partially overlap with an object further away;

FIG. 4A depicts the image of a foreground object being gathered by a camera and recorded by a recorder;

FIG. 4B shows mid-range objects being recorded;

FIG. 4C shows distant or background objects are being recorded;

FIG. 5 depicts an alternative technique for implementing perspective alteration according to the invention;

FIG. 6A shows how an infrared CCD (charge-coupled device) camera, preferably with a wide-angle lens, may be used as a sensor according to the invention;

FIG. 6B shows the use of a panning camera;

FIG. 6C depicts three persons generating a composite thermal field on an image sensor;

FIG. 7A illustrates a “virtual picture window” embodiment, wherein the movement of a viewer causes an outdoor camera to pan back/forth, thereby allowing the viewer to visualize the outdoor scene as if the display were a hole in the wall;

FIG. 7B shows how the invention enables virtual windows on insides walls, which may be useful in homes and businesses such as restaurants, bars and nightclubs;

FIG. 7C shows how the display can function as a “virtual mirror”; and

FIG. 8 depicts how the virtual mirror embodiment of the invention may be used in bathrooms and dressing rooms.

DETAILED DESCRIPTION OF THE INVENTION

This invention employs a variety of techniques to provide a unique experience for the viewer of a display, particularly large wall-mounted panels. In the preferred embodiments, the perception of a displayed image is altered for viewers moving relative to the position of the display system screen, thereby imparting a sense of three-dimensional immersion in the scene being displayed.

FIG. 1A shows a viewer 102 having a field of view 104 walking from the right towards the left relative to a display screen 106 driven by display generator 107. One or more sensors 108, which may use visible-light, infrared, ultrasonic, or other modalities described in further detail below, are used to track at least the lateral position of individual 102.

FIG. 1B is a simplified representation of what the person 102 might see on the display screen 106 according to the invention from the position shown in FIG. 1A. Relatively close objects are shown at 115, 116. Less close objects are seen at 114. Somewhat distant objects are shown at 112, and distant objects are shown at 110. Although four relative distances are mentioned, the invention is not limited in this regard, and is applicable to more or fewer such relative distances.

In FIG. 2A, the person 102 has moved to the left, closer to the center of the screen 106. The sensor(s) 108 have detected this movement and, in response, the perspective of the scene has been altered. As shown in FIG. 2B, closer objects have shifted laterally to the greatest degree, followed by objects 114 and 112 in order, much as a train passenger would experience while looking out the window of the train as the train moves. Object 110, being significantly distant, would shift little, if at all.

This process continues in FIG. 3A, where the viewer has moved to the left-most portion of the display. Close object 116 has shifted to the extent that it now partially overlaps with an object further away, and object 115 has begun to move off the screen 106, as shown in FIG. 3B. The ways in which the invention makes this possible will now be described in further detail.

In the preferred embodiment, the invention employs a technique similar to that utilized in animation films: multiple superimposed graphics planes. In FIG. 4A, the image of a foreground object 412 is gathered by camera 402 and recorded by recorder 410. Depending upon the circumstances, a blue or other solid-color background 414 may be used for chroma-keying.

In FIG. 4B, mid range objects 420 are being recorded whereas, in FIG. 4C, distant or background objects are being recorded. Commercially available software packages, or customized software tuned to specific program content, can be utilized to derive the desired material from among multiple subjects representing different focal points, and, once identified, can track the subjects as they change their position and even their orientation. In an alternative approach, multiple cameras, positioned to capture three-dimensional information, may be utilized to derive a three-dimensional array for each frame of motion, thereby allowing the producer to select “slices” which can be captured as graphics plane images for manipulation by the graphics processor of the instant invention.

The camera(s) may record moving video images for the foreground graphics plane(s), the background graphics plane(s), or any combination thereof. For example, the background graphic plane may be based upon a still picture, while the foreground cameras record motion imagery. In this embodiment, position sensor(s) 108 detect the location of the viewer relative to the screen, and reposition the foreground graphics plane(s) as the viewer moves, thereby conveying to the viewer the impression, for example, that he is looking through a window at an outdoor scene, with, perhaps, a nearer image, such as a tree branch, that the viewer can see around by simply shifting his position relative to the display screen. The tree branch (or other object) may also be moving, as it would in a breeze, for example.

The video source for these graphics planes may include prerecorded material supplied by playback from any recording devices. Other sources include broadcast, satellite, cable, or other programming sources, material delivered over broadband or other telecommunication links, privately recorded material, live video from cameras (including security and monitoring cameras), computer-generated graphics and the like, or any other source of image material. Graphics planes displaying text information may be superimposed over, or under, other graphics planes.

FIG. 5 depicts an alternative technique for implementing perspective alteration according to the invention. Here a camera 502 having at least a pan mount 506 records an actual scene 512 at incremental angles suggested by arrow 508. The view at each angle is recorded by unit 510 for later replay. The number of increments depends upon the desired resolution, room dimensions, and other factors. For example, at a very high resolution, single-degree increments may be recorded through a full 180 degrees. At a lesser resolution, single-degree increments may be recorded across a smaller angle of view, or larger-degree increments may be used at angle of view with or without interpolation to fill in any ‘gaps’ during replay.

Thus, a variety of novel video applications are enabled by virtue of the invention, such as a simulated window which has a view that is not possible from that position if there were an actual window in that position (i.e., viewing a sunset from an eastern exposure), or a view which is not possible at all (i.e., a winter scene during the summer, or a scene from a different country). These represent a clear improvement over any of the typical “light box” with photographic transparency or other attempt to simulate a window view with conventional means.

FIG. 7A illustrates a “virtual picture window” embodiment, wherein the movement of viewer 102, detected by sensors 108, cause outdoor camera 704 on mount 706 to pan back/forth, thereby allowing the viewer 102 to “see” the outdoor scene 720 “through” the wall 702 on display 106 using graphics processing system 710. Depending upon the movement of the viewer toward/away from the screen 106, in this and other applicable embodiments the camera 704 may zoom as the viewer comes closer, pan when the user moves away and tilt if the viewer goes up/down, as might be the case on a staircase, for example.

Through the use of distance detection, the display system can provide a three-dimensional effect, by applying modifications to an image as the viewer changes his position. Thus, for example, an image of a painting might capture the artist's intentions when viewed at a distance, while a close-up examination of the video display by the viewer could display the details of the brush strokes as the viewer changed his position relative to the screen.

The invention is not limited to virtual windows through outside walls. As shown in FIG. 7B, the invention enables virtual windows on insides walls, which may be useful in homes and businesses such as restaurants, bars and nightclubs. Similar to the embodiment depicted in FIG. 7A, as user 102 moves relative to screen 106, camera 704 pans (or tilts or zooms), enabling the user 102 to see people 730 or other objects through the wall 702.

In still another aspect, the display system can function as a “virtual mirror.” Such an embodiment is depicted in FIG. 7C, wherein graphics processing system 710 directs the camera 704 to at least pan at an angle “A” substantially equal to angle “B” formed by the location of the user and line 722 perpendicular to the plane of display 106. This allows the viewer to see a synthesized reflection depicted by broken line 720, typically including the subject him/herself.

The mirror embodiment of the invention may be used in bathrooms and dressing rooms, as illustrated in FIG. 8. Multiple cameras 802, 804, 806, 808 disposed around the display screen 106 allow the processor unit 812 to construct a three-dimensional image for display on the screen, and the proximity sensing devices may be utilized to create an on-screen image which is representative of what a person should see as he re-positions himself, in three dimensions, about the display screen. It is a simple matter to provide a left-to-right reversed-image, in keeping with the mirror aspect of the invention. A user control 810 allows the user to select a specific camera or cameras to see how others would view them from different perspectives.

In the dressing room/bathroom embodiments of the invention, the camera(s) capturing the image of the viewer may be placed behind a semi-transparent screen allowing better visualization such as eye-level contact to be maintained. One or more video cameras may be provided on a bendable tether—or wireless hookup—enabling a user to view hard-to-reach places such as cars, nose, mouth etc. Variable degrees of magnification may be provided, based upon detected distance from a surface being viewed, for example.

In terms of position sensing, the preferred embodiment uses an infrared CCD (charge-coupled device) camera, preferably with a wide-angle lens 206 shown in FIG. 6A. The user 202 generates a thermal image 204, which is focused onto camera array 210 as a spot or group of pixels, allowing the system 220 to know where a person is with no moving parts. Processor 106 can then cause the perspective, depth perception, or other characteristics to change accordingly. If no tilt or zoom functions are provided, a linear sensor may be substituted for a 2D sensor.

As an alternative to a fixed camera with sufficiently wide-angle lens, a panning camera may be used, as shown in FIG. 6B. Here the camera 230 is trained on a subject and, as that subject moves positional information is sensed by pan mount 234 and communicated to alter screen 106 through electronics unit 240. Again, if tilt/zoom functions are provided, the camera 230 may use tilt and/or auto focus to determine other positional aspects of the viewer.

The invention may handle multiple viewers in different ways. These solutions include (1) favoring clusters of potential viewers over singular viewers; (2) favoring moving viewers over stationary viewers; and (3) favoring viewers actually looking at the screen over those looking away. One advantage of the sensor system of FIG. 6A is that clustering is naturally accommodated. In FIG. 6C, three persons 242, 244, 246 generate a composite thermal field 240, which generates a relatively large imprint 251 on sensor 210. In contrast, the narrower thermal field 250 of single person 248 results in a smaller spot 241 on array 210 such that, in this embodiment of the invention, the perspective of persons in the group would be favored.

The approach of FIG. 6A also naturally addresses the favoring of moving viewers over stationary viewers. Referring again to FIG. 6C, if the group consisting of persons 242, 244, 246 were stationary, it would be a straightforward processing task to detect that an individual is moving. In this embodiment of the invention, the moving individual 248 would be preferred over the stationary group.

In all embodiments of the invention, sensing in the visible region of the spectrum may be used instead of—or in concert with—IR sensing. This presents advantages and disadvantages. In terms of advantages, visible light sensing may allow a single camera (or cameras) to detect the image and position of the viewer (as in the mirror embodiments). Another advantage is that recognition techniques may be used to determine if a particular person is actually looking at the display in which case that person may be favored over individuals looking away. If multiple persons are looking at the display, other techniques such as clustering and motion favoring may also be used. Perhaps the only disadvantage is that image recognition and other operations require additional processing power, however, that is easily accommodated with modern processors.

In each case, specialized graphics processing provides the management of the graphics planes and any audio material, while processing rules (for example—“take image modification instructions from the position of the closest viewer only”) ensure that the system will not be misdirected by movement of viewers that are on the opposite side of the room. Where multiple display systems are in use (as for simulating an array of “structural” windows), an overall system for management of the displays is utilized, thus providing an integrated, coordinated system of imaging displays. For example—an overall image, larger than the entire display system, may be utilized, or alternative schemes, in which image planes or other data may “flow” from one display screen to another display screen next to it.

While described herein with reference to flat-panel displays (LED, LCD, Plasma, etc.) the principles disclosed may be applied with suitable results to any number of display technologies currently available or in development today (CRT-type, front or rear projection, Electroluminescence, OLEDs, etc.). Furthermore, adjustments may be applied to the image data to correct for any geometric distortions introduced due to the position of the camera(s) or display unit(s). In addition, alternative embodiments may utilize additional graphics planes to enhance the effect of the display.

In all cases, audio may be included, representing material that may or may not be related to the video images presented on the screen. Thus, the system can serve the function of an enhanced video display terminal, a television viewing screen, a security monitoring system, a video entertainment system, or any other system for which display of graphics material is of value to the viewer.

With further regard to audio, if the system is provided with stereo or surround sound, the sound reproduction may be altered as a function with user position, with or without a change in visual perspective. For example, as an individual walks past the display screen the sound of elements in the scene (i.e., birds, vehicles, etc.) may be varied whether or not the individual is looking at the screen. If the user moves toward the screen, sounds may be enhanced or attenuated. For example, if a viewer moves toward a frog or a bird in the scene, the sounds of that creature may be enhanced, or diminished as the user moves away.

Claims (23)

We claim:
1. A perspective-altering display system, comprising:
a display screen for displaying a scene having opposing side edges;
a sensor for detecting the movement of a viewer relative to the display screen;
at least one memory for storing foreground and background imagery;
wherein the memory has a capacity to store imagery that extends beyond one or both of the side edges of the display screen;
at least one processor interconnected to the memory and the sensor, the processor being operative to perform the following functions:
generate a scene based upon multiple superimposed graphics planes, including a foreground graphics plane containing the foreground imagery and a background graphics plane containing the background imagery, and
receive a signal from the sensor and shift the relative position of the foreground and background graphics planes in the displayed scene as a function of viewer movement;
wherein the multiple superimposed graphics planes include non-visible imagery beyond one of both of the side edges display screen; and
wherein the foreground graphics plane is shifted by way of a translation to a greater extent than the background graphics plane as a viewer moves from side to side relative to the display screen, causing the non-visible imagery to become visible.
2. The display system of claim 1, wherein at least one of the graphics planes includes prerecorded imagery.
3. The display system of claim 1, wherein at least one of the graphics planes includes imagery received through a transmission medium.
4. The display system of claim 1, wherein at least one of the graphics planes includes imagery received from a camera.
5. The display system of claim 1, wherein at least one of the graphics planes includes textual information.
6. The display system of claim 1, further including:
a camera that records the scene through panning at sequential angles; and
wherein the memory stores imagery obtained at the sequential angles for later recall as a function of user movement.
7. The display system of claim 6, further including:
an interpolator for filling in visual gaps in the scene.
8. The display system of claim 1, wherein:
the display screen is mounted on a wall having a backside; and
further including a camera mounted on the backside of the wall which pans as a viewer moves, thereby imaging a scene representative of the display being a virtual window through the wall.
9. The display system of claim 1, further including:
a camera with a zoom capability for recording the scene;
the sensor being operative to detect the viewer's distance from the display screen; and
wherein the processor is further operative to zoom in the camera as the viewer moves closer to the display screen and zoom out the camera as the viewer moves away from the display screen.
10. The display system of claim 1, wherein:
the sensor is operative to detect the viewer's distance from the display screen; and
the processor is operative to increase the resolution of the scene as the viewer moves closer to the display screen and decrease the resolution of the scene as the viewer moves away from the display screen.
11. The display system of claim 1, further including:
a camera with a tilt capability for recording the scene;
the sensor being operative to detect the viewer's up/down movement relative to the display screen; and
wherein the processor is further operative to tilt the camera in response to the viewer's up/down movement.
12. The display system of claim 1, further including:
a camera for recording the scene, the camera having a field of view including a viewer of the display screen; and
wherein the processor is further operative to flip the scene horizontally such that the display functions as a virtual mirror.
13. The display system of claim 1, further including:
a plurality of cameras for recording the scene; and
wherein the processor is further operative to construct a three-dimensional image for display on the screen.
14. The display system of claim 1, further including:
a plurality of cameras for recording the scene;
the processor being further operative to construct a three-dimensional image for display on the screen; and
a user control a viewer to select a specific camera or cameras to see how others would view the user from different perspectives.
15. The display system of claim 1, wherein the sensor is an infrared CCD (charge-coupled device) camera.
16. The display system of claim 1, wherein:
the sensor is a camera having an image sensor; and
an image of a person in front of the display is focused onto the image sensor as a spot or group of pixels, allowing the movements of the person to be tracked with no moving parts.
17. The display system of claim 1, wherein:
the sensor is a camera having at least a pan mount that tracks the movement of a viewer; and
the processor is operative to shift the relative position of the foreground and background elements as a function of the tracking.
18. The display system of claim 1, wherein:
the sensor is a camera having a pan/tilt mount that tracks the movement of a viewer;
the processor is operative to shift the relative position of the foreground and background elements as a function of the tracking.
19. The display system of claim 1, wherein:
the sensor is a camera having an auto-focus capability to track the movement of a viewer;
the processor is operative to shift the relative position of the foreground and background elements as a function of the tracking.
20. The display system of claim 1, wherein:
the sensor is operative to sense a plurality of individuals in front of the display screen; and
the processor is operative to shift the relative position of the foreground and background elements by favoring clusters of individuals.
21. The display system of claim 1, wherein:
the sensor is operative to sense a plurality of individuals in front of the display screen; and
the processor is operative to shift the relative position of the foreground and background elements by favoring moving individuals over stationary individuals.
22. The display system of claim 1, wherein:
the sensor is operative to sense a whether an individual is looking toward the display screen; and
the processor is operative to shift the relative position of the foreground and background elements by favoring individuals looking toward the display screen.
23. The display system of claim 1, further including one or more transducers for producing sounds associated with the scene; and
the processor is operative to alter the reproduction of the sounds as a function of viewer movement.
US12/197,635 2007-08-24 2008-08-25 Perspective altering display system Active 2034-06-23 US10063848B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US95784507P true 2007-08-24 2007-08-24
US12/197,635 US10063848B2 (en) 2007-08-24 2008-08-25 Perspective altering display system

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US12/197,635 US10063848B2 (en) 2007-08-24 2008-08-25 Perspective altering display system
US16/046,065 US10397556B2 (en) 2007-08-24 2018-07-26 Perspective altering display system
US16/452,901 US20190320163A1 (en) 2007-08-24 2019-06-26 Multi-screen perspective altering display system

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/046,065 Continuation US10397556B2 (en) 2007-08-24 2018-07-26 Perspective altering display system

Publications (2)

Publication Number Publication Date
US20090051699A1 US20090051699A1 (en) 2009-02-26
US10063848B2 true US10063848B2 (en) 2018-08-28

Family

ID=40381715

Family Applications (3)

Application Number Title Priority Date Filing Date
US12/197,635 Active 2034-06-23 US10063848B2 (en) 2007-08-24 2008-08-25 Perspective altering display system
US16/046,065 Active US10397556B2 (en) 2007-08-24 2018-07-26 Perspective altering display system
US16/452,901 Pending US20190320163A1 (en) 2007-08-24 2019-06-26 Multi-screen perspective altering display system

Family Applications After (2)

Application Number Title Priority Date Filing Date
US16/046,065 Active US10397556B2 (en) 2007-08-24 2018-07-26 Perspective altering display system
US16/452,901 Pending US20190320163A1 (en) 2007-08-24 2019-06-26 Multi-screen perspective altering display system

Country Status (1)

Country Link
US (3) US10063848B2 (en)

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8427535B2 (en) * 2008-06-07 2013-04-23 Rick Davis Personal grooming visual display system
US20110069158A1 (en) * 2009-09-21 2011-03-24 Dekel Shiloh Virtual window system and method
US8791949B1 (en) * 2010-04-06 2014-07-29 The Pnc Financial Services Group, Inc. Investment management marketing tool
US8780115B1 (en) * 2010-04-06 2014-07-15 The Pnc Financial Services Group, Inc. Investment management marketing tool
US9167289B2 (en) * 2010-09-02 2015-10-20 Verizon Patent And Licensing Inc. Perspective display systems and methods
US8896631B2 (en) * 2010-10-25 2014-11-25 Hewlett-Packard Development Company, L.P. Hyper parallax transformation matrix based on user eye positions
US20120320080A1 (en) * 2011-06-14 2012-12-20 Microsoft Corporation Motion based virtual object navigation
US9509922B2 (en) * 2011-08-17 2016-11-29 Microsoft Technology Licensing, Llc Content normalization on digital displays
US20130083252A1 (en) * 2011-10-04 2013-04-04 David John Boyes Gesture recognition capable picture video frame
US9274597B1 (en) * 2011-12-20 2016-03-01 Amazon Technologies, Inc. Tracking head position for rendering content
US9389682B2 (en) 2012-07-02 2016-07-12 Sony Interactive Entertainment Inc. Methods and systems for interaction with an expanded information space
EP2693363B1 (en) * 2012-07-31 2015-07-22 Sick Ag Camera system and method for recording a flow of objects
US10116911B2 (en) * 2012-12-18 2018-10-30 Qualcomm Incorporated Realistic point of view video method and apparatus
US9342467B1 (en) 2013-07-17 2016-05-17 Frederick Thomas McGrath Virtual window system
US20160205492A1 (en) * 2013-08-21 2016-07-14 Thomson Licensing Video display having audio controlled by viewing direction
KR20160045084A (en) * 2013-08-21 2016-04-26 톰슨 라이센싱 Video display with pan function controlled by viewing direction
US20150221064A1 (en) * 2014-02-03 2015-08-06 Nvidia Corporation User distance based modification of a resolution of a display unit interfaced with a data processing device and/or a display area size thereon
US20150262236A1 (en) 2014-03-13 2015-09-17 Ebay Inc. Follow-up messages after in-store shopping
US10134187B2 (en) 2014-08-07 2018-11-20 Somo Innvoations Ltd. Augmented reality with graphics rendering controlled by mobile device position
KR20170035608A (en) * 2015-09-23 2017-03-31 삼성전자주식회사 Videotelephony System, Image Display Apparatus, Driving Method of Image Display Apparatus, Method for Generation Realistic Image and Computer Readable Recording Medium
WO2018027110A1 (en) * 2016-08-05 2018-02-08 University Of Rochester Virtual window
US20180063444A1 (en) * 2016-08-29 2018-03-01 Razmik Karabed View friendly monitor systems
WO2018110821A1 (en) 2016-12-14 2018-06-21 Samsung Electronics Co., Ltd. Display apparatus and method for controlling the display apparatus
US10264222B2 (en) * 2017-02-08 2019-04-16 Phazr, Inc. Window-installed wireless communications device

Citations (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2815310A (en) 1952-03-01 1957-12-03 Pictorial Prod Inc Process of assembling in the art of changeable picture display devices
US3463581A (en) 1966-01-17 1969-08-26 Intermountain Res & Eng System for three-dimensional panoramic static-image motion pictures
US4135502A (en) 1976-09-07 1979-01-23 Donald Peck Stereoscopic patterns and method of making same
US4896210A (en) 1987-11-16 1990-01-23 Brokenshire Daniel A Stereoscopic graphics display terminal with image data processing
US4956705A (en) 1989-03-10 1990-09-11 Dimensional Visions Group Electronic method and apparatus for stereoscopic photography
US4987487A (en) 1988-08-12 1991-01-22 Nippon Telegraph And Telephone Corporation Method of stereoscopic images display which compensates electronically for viewer head movement
US5083199A (en) 1989-06-23 1992-01-21 Heinrich-Hertz-Institut For Nachrichtentechnik Berlin Gmbh Autostereoscopic viewing device for creating three-dimensional perception of images
US5101268A (en) 1989-12-05 1992-03-31 Sony Corporation Visual point position control apparatus
US5253051A (en) 1991-03-05 1993-10-12 Mcmanigal Paul G Video artificial window apparatus
US5543964A (en) 1993-12-28 1996-08-06 Eastman Kodak Company Depth image apparatus and method with angularly changing display information
US5574836A (en) * 1996-01-22 1996-11-12 Broemmelsiek; Raymond M. Interactive display apparatus and method with viewer position compensation
US5589980A (en) 1989-02-27 1996-12-31 Bass; Robert Three dimensional optical viewing system
US5613048A (en) 1993-08-03 1997-03-18 Apple Computer, Inc. Three-dimensional image synthesis using view interpolation
US5841439A (en) * 1994-07-22 1998-11-24 Monash University Updating graphical objects based on object validity periods
US5956180A (en) 1996-12-31 1999-09-21 Bass; Robert Optical viewing system for asynchronous overlaid images
US5963303A (en) 1997-09-04 1999-10-05 Allen; Dann M. Stereo pair and method of making stereo pairs
US6005967A (en) 1994-02-18 1999-12-21 Matushita Electric Industrial Co., Ltd. Picture synthesizing apparatus and method
US6317151B1 (en) 1997-07-10 2001-11-13 Mitsubishi Denki Kabushiki Kaisha Image reproducing method and image generating and reproducing method
US20020113791A1 (en) * 2001-01-02 2002-08-22 Jiang Li Image-based virtual reality player with integrated 3D graphics objects
US20030035001A1 (en) * 2001-08-15 2003-02-20 Van Geest Bartolomeus Wilhelmus Damianus 3D video conferencing
US20030080937A1 (en) * 2001-10-30 2003-05-01 Light John J. Displaying a virtual three-dimensional (3D) scene
US6654019B2 (en) 1998-05-13 2003-11-25 Imove, Inc. Panoramic movie which utilizes a series of captured panoramic images to display movement as observed by a viewer looking in a selected direction
US6867775B2 (en) * 2000-02-08 2005-03-15 Daimlerchrysler Ag Method and device for displaying a multidimensional image of an object
US20050059488A1 (en) * 2003-09-15 2005-03-17 Sony Computer Entertainment Inc. Method and apparatus for adjusting a view of a scene being displayed according to tracked head motion
US7079706B2 (en) 2001-06-20 2006-07-18 Paul Peterson Methods and apparatus for generating a multiple composite image
US20070060346A1 (en) 2005-06-28 2007-03-15 Samsung Electronics Co., Ltd. Tool for video gaming system and method
US20070291035A1 (en) * 2004-11-30 2007-12-20 Vesely Michael A Horizontal Perspective Representation
US7513623B2 (en) * 2006-04-17 2009-04-07 Third Dimension Ip Llc System and methods for angular slice true 3-D display
US7557824B2 (en) 2003-12-18 2009-07-07 University Of Durham Method and apparatus for generating a stereoscopic image
US7663689B2 (en) * 2004-01-16 2010-02-16 Sony Computer Entertainment Inc. Method and apparatus for optimizing capture device settings through depth information
US20100039380A1 (en) * 2004-10-25 2010-02-18 Graphics Properties Holdings, Inc. Movable Audio/Video Communication Interface System
US20100053310A1 (en) 2008-08-31 2010-03-04 Maxson Brian D Transforming 3d video content to match viewer position
US20110234584A1 (en) * 2010-03-25 2011-09-29 Fujifilm Corporation Head-mounted display device
US8704879B1 (en) * 2010-08-31 2014-04-22 Nintendo Co., Ltd. Eye tracking enabling 3D viewing on conventional 2D display
US20140118509A1 (en) * 2011-06-22 2014-05-01 Koninklijke Philips N.V. Method and apparatus for generating a signal for a display
US20150085076A1 (en) * 2013-09-24 2015-03-26 Amazon Techologies, Inc. Approaches for simulating three-dimensional views
US20150130799A1 (en) * 2013-11-12 2015-05-14 Fyusion, Inc. Analysis and manipulation of images and video for generation of surround views

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8730156B2 (en) * 2010-03-05 2014-05-20 Sony Computer Entertainment America Llc Maintaining multiple views on a shared stable virtual space

Patent Citations (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2815310A (en) 1952-03-01 1957-12-03 Pictorial Prod Inc Process of assembling in the art of changeable picture display devices
US3463581A (en) 1966-01-17 1969-08-26 Intermountain Res & Eng System for three-dimensional panoramic static-image motion pictures
US4135502A (en) 1976-09-07 1979-01-23 Donald Peck Stereoscopic patterns and method of making same
US4896210A (en) 1987-11-16 1990-01-23 Brokenshire Daniel A Stereoscopic graphics display terminal with image data processing
US4987487A (en) 1988-08-12 1991-01-22 Nippon Telegraph And Telephone Corporation Method of stereoscopic images display which compensates electronically for viewer head movement
US5589980A (en) 1989-02-27 1996-12-31 Bass; Robert Three dimensional optical viewing system
US4956705A (en) 1989-03-10 1990-09-11 Dimensional Visions Group Electronic method and apparatus for stereoscopic photography
US5083199A (en) 1989-06-23 1992-01-21 Heinrich-Hertz-Institut For Nachrichtentechnik Berlin Gmbh Autostereoscopic viewing device for creating three-dimensional perception of images
US5101268A (en) 1989-12-05 1992-03-31 Sony Corporation Visual point position control apparatus
US5253051A (en) 1991-03-05 1993-10-12 Mcmanigal Paul G Video artificial window apparatus
US5613048A (en) 1993-08-03 1997-03-18 Apple Computer, Inc. Three-dimensional image synthesis using view interpolation
US5543964A (en) 1993-12-28 1996-08-06 Eastman Kodak Company Depth image apparatus and method with angularly changing display information
US6005967A (en) 1994-02-18 1999-12-21 Matushita Electric Industrial Co., Ltd. Picture synthesizing apparatus and method
US5841439A (en) * 1994-07-22 1998-11-24 Monash University Updating graphical objects based on object validity periods
US5574836A (en) * 1996-01-22 1996-11-12 Broemmelsiek; Raymond M. Interactive display apparatus and method with viewer position compensation
US5956180A (en) 1996-12-31 1999-09-21 Bass; Robert Optical viewing system for asynchronous overlaid images
US6317151B1 (en) 1997-07-10 2001-11-13 Mitsubishi Denki Kabushiki Kaisha Image reproducing method and image generating and reproducing method
US5963303A (en) 1997-09-04 1999-10-05 Allen; Dann M. Stereo pair and method of making stereo pairs
US6654019B2 (en) 1998-05-13 2003-11-25 Imove, Inc. Panoramic movie which utilizes a series of captured panoramic images to display movement as observed by a viewer looking in a selected direction
US6867775B2 (en) * 2000-02-08 2005-03-15 Daimlerchrysler Ag Method and device for displaying a multidimensional image of an object
US20020113791A1 (en) * 2001-01-02 2002-08-22 Jiang Li Image-based virtual reality player with integrated 3D graphics objects
US7079706B2 (en) 2001-06-20 2006-07-18 Paul Peterson Methods and apparatus for generating a multiple composite image
US20030035001A1 (en) * 2001-08-15 2003-02-20 Van Geest Bartolomeus Wilhelmus Damianus 3D video conferencing
US20030080937A1 (en) * 2001-10-30 2003-05-01 Light John J. Displaying a virtual three-dimensional (3D) scene
US20050059488A1 (en) * 2003-09-15 2005-03-17 Sony Computer Entertainment Inc. Method and apparatus for adjusting a view of a scene being displayed according to tracked head motion
US7557824B2 (en) 2003-12-18 2009-07-07 University Of Durham Method and apparatus for generating a stereoscopic image
US7663689B2 (en) * 2004-01-16 2010-02-16 Sony Computer Entertainment Inc. Method and apparatus for optimizing capture device settings through depth information
US20100039380A1 (en) * 2004-10-25 2010-02-18 Graphics Properties Holdings, Inc. Movable Audio/Video Communication Interface System
US20070291035A1 (en) * 2004-11-30 2007-12-20 Vesely Michael A Horizontal Perspective Representation
US20070060346A1 (en) 2005-06-28 2007-03-15 Samsung Electronics Co., Ltd. Tool for video gaming system and method
US7513623B2 (en) * 2006-04-17 2009-04-07 Third Dimension Ip Llc System and methods for angular slice true 3-D display
US20100053310A1 (en) 2008-08-31 2010-03-04 Maxson Brian D Transforming 3d video content to match viewer position
US20110234584A1 (en) * 2010-03-25 2011-09-29 Fujifilm Corporation Head-mounted display device
US8704879B1 (en) * 2010-08-31 2014-04-22 Nintendo Co., Ltd. Eye tracking enabling 3D viewing on conventional 2D display
US20140118509A1 (en) * 2011-06-22 2014-05-01 Koninklijke Philips N.V. Method and apparatus for generating a signal for a display
US20150085076A1 (en) * 2013-09-24 2015-03-26 Amazon Techologies, Inc. Approaches for simulating three-dimensional views
US20150130799A1 (en) * 2013-11-12 2015-05-14 Fyusion, Inc. Analysis and manipulation of images and video for generation of surround views

Also Published As

Publication number Publication date
US20090051699A1 (en) 2009-02-26
US20190320163A1 (en) 2019-10-17
US10397556B2 (en) 2019-08-27
US20180343441A1 (en) 2018-11-29

Similar Documents

Publication Publication Date Title
Uyttendaele et al. Image-based interactive exploration of real-world environments
US6084979A (en) Method for creating virtual reality
US5721585A (en) Digital video panoramic image capture and display system
US8762846B2 (en) Method and system for adaptive viewport for a mobile device based on viewing angle
US7520615B2 (en) Display apparatus and image pickup apparatus
US6665003B1 (en) System and method for generating and displaying panoramic images and movies
CN102084650B (en) Telepresence system, method and video capture device
KR100730406B1 (en) Three-dimensional display apparatus using intermediate elemental images
CN101411207B (en) Method and system for acquiring and displaying 3d light fields
CN102362495B (en) A plurality of remote video cameras having a display wall conference endpoints combined view presentation apparatus and method of operation
Fehn et al. Interactive 3-DTV-concepts and key technologies
US7015954B1 (en) Automatic video system using multiple cameras
US20050053274A1 (en) System and method for 3D photography and/or analysis of 3D images and/or display of 3D images
KR101313740B1 (en) OSMU( One Source Multi Use)-type Stereoscopic Camera and Method of Making Stereoscopic Video Content thereof
CN102197653B (en) Three-dimensional display system
CA2488925C (en) Method for producing stereoscopic images from monoscopic images
JP2010505174A (en) Menu display
US5963247A (en) Visual display systems and a system for producing recordings for visualization thereon and methods therefor
US8666241B2 (en) Stereo camera with automatic control of interocular distance based on lens settings
US5448287A (en) Spatial video display system
US8199186B2 (en) Three-dimensional (3D) imaging based on motionparallax
US7079173B2 (en) Displaying a wide field of view video image
US20080030573A1 (en) Volumetric panoramic sensor systems
US7224382B2 (en) Immersive imaging system
US8928659B2 (en) Telepresence systems with viewer perspective adjustment

Legal Events

Date Code Title Description
AS Assignment

Owner name: VIDEA, LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:POSA, JOHN G.;SCHWAB, BARRY H.;REEL/FRAME:021438/0511;SIGNING DATES FROM 20080821 TO 20080825

Owner name: VIDEA, LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:POSA, JOHN G.;SCHWAB, BARRY H.;SIGNING DATES FROM 20080821 TO 20080825;REEL/FRAME:021438/0511

STCF Information on status: patent grant

Free format text: PATENTED CASE