WO2016041953A1 - A system comprising multiple digital cameras viewing a large scene - Google Patents
A system comprising multiple digital cameras viewing a large scene Download PDFInfo
- Publication number
- WO2016041953A1 WO2016041953A1 PCT/EP2015/071083 EP2015071083W WO2016041953A1 WO 2016041953 A1 WO2016041953 A1 WO 2016041953A1 EP 2015071083 W EP2015071083 W EP 2015071083W WO 2016041953 A1 WO2016041953 A1 WO 2016041953A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- cameras
- scene
- elevator
- model
- view
- Prior art date
Links
- 230000003068 static effect Effects 0.000 claims abstract description 21
- 238000000034 method Methods 0.000 claims description 63
- 238000004422 calculation algorithm Methods 0.000 claims description 28
- 230000001360 synchronised effect Effects 0.000 claims description 18
- 230000033001 locomotion Effects 0.000 claims description 9
- 230000000694 effects Effects 0.000 claims description 6
- 238000013507 mapping Methods 0.000 claims description 6
- 238000001514 detection method Methods 0.000 claims description 5
- 239000011159 matrix material Substances 0.000 claims description 5
- 240000004759 Inga spectabilis Species 0.000 claims description 2
- 238000012937 correction Methods 0.000 claims description 2
- 238000012545 processing Methods 0.000 description 18
- 238000009877 rendering Methods 0.000 description 7
- 238000005286 illumination Methods 0.000 description 5
- 230000015572 biosynthetic process Effects 0.000 description 4
- 238000005094 computer simulation Methods 0.000 description 4
- 238000000926 separation method Methods 0.000 description 4
- 238000003786 synthesis reaction Methods 0.000 description 4
- 238000013459 approach Methods 0.000 description 3
- 238000009434 installation Methods 0.000 description 3
- 238000012544 monitoring process Methods 0.000 description 3
- 239000003086 colorant Substances 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 206010052128 Glare Diseases 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 238000007519 figuring Methods 0.000 description 1
- 239000012634 fragment Substances 0.000 description 1
- 230000004313 glare Effects 0.000 description 1
- 230000008821 health effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003032 molecular docking Methods 0.000 description 1
- 238000011328 necessary treatment Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/361—Reproducing mixed stereoscopic images; Reproducing mixed monoscopic and stereoscopic images, e.g. a stereoscopic image overlay window on a monoscopic image background
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B37/00—Panoramic or wide-screen photography; Photographing extended surfaces, e.g. for surveying; Photographing internal surfaces, e.g. of pipe
- G03B37/04—Panoramic or wide-screen photography; Photographing extended surfaces, e.g. for surveying; Photographing internal surfaces, e.g. of pipe with cameras or projectors providing touching or overlapping fields of view
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B66—HOISTING; LIFTING; HAULING
- B66B—ELEVATORS; ESCALATORS OR MOVING WALKWAYS
- B66B13/00—Doors, gates, or other apparatus controlling access to, or exit from, cages or lift well landings
- B66B13/30—Constructional features of doors or gates
- B66B13/303—Details of door panels
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
- G06T15/205—Image-based rendering
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/243—Image signal generators using stereoscopic image cameras using three or more 2D image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/275—Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals
- H04N13/279—Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals the virtual viewpoint locations being selected by the viewers or determined by tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/296—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0081—Depth or disparity estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0085—Motion estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0088—Synthesising a monoscopic image signal from stereoscopic images, e.g. synthesising a panoramic or high resolution monoscopic image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0092—Image segmentation from stereoscopic image signals
Definitions
- implementations of this invention are based on conventional consumer-grade digital video cameras, which significantly reduces the cost of the solution. These implementations do not cause adverse health effects with emissions and do not interact with other active systems.
- US 20140015832 describes a system reconstructing a 3D model from a set of video files obtained from shooting a single scene.
- the video files are synchronized with a global event (e.g. a flash) then the individual position of each camera is calculated for every frame.
- a global event e.g. a flash
- the video cameras are not fixed immovably.
- a 3D model reconstruction is processed, the textures are selected and mapped on the 3D model, and weather conditions are identified and simulated.
- the system is mostly intended for car tracking. Unlike the system we describe, this system does not rely upon permanently fixed cameras, the reconstruction is undertaken for each frame individually, which makes real-time processing in high-resolution impossible.
- US 6674461 describes a method of shooting a scene with a few video cameras; the footage obtained is mapped on a synthesized virtual scene.
- the approach described is designed for shooting a sole object, not an entire scene, so its authors do not provide a solution for fixing the lighting difference across the scene.
- the problem of setting up the cameras for real time shooting is not solved either.
- the system unlike our one, has a limitation of 3 cameras only.
- Another aspect is an elevator including a display showing images generated by a method as described above.
- FIG. 8 shows the hardware used in the system.
- the system controls white balance, camera shutter opening times, shutter speed, aperture and the camera frame rate automatically, analyzing the images obtained from each camera individually.
- Settings can be applied to a specific camcorder and transmitted in parametric form to each camera; each setting is transmitted as a function of or depending on the particular camera's position.
- An automatic tuning system of the cameras allows displaying correctly the area of the scene shot simultaneously by multiple cameras, even if the shooting conditions prevent the use of automatic settings.
- the obtained video stream is processed and used for scene modeling.
- the solution proposed combines static modeling of 3D scenes and dynamic simulation of individual parts of the scene. When using only dynamic simulation, significant computing power is required, the precision of the model is sufficiently low, the dependence on weather conditions is considerably high, and the recognition applies on for example Lambertian surfaces only (which excludes glares, water surface, etc.).
- the resulting image sent to the displaying device is obtained from the render of the current 3D model angle with mapped textures, from the point of the virtual camera's current location.
- the angle can be set and modified manually by the user at runtime to obtain the desired look. Also the angle can be set automatically based on data obtained from external sensors. For example, when designing a virtual window in a lift, the angle is determined by the height sensor reading the lift position. Or, for example, to display a panorama, the data for figuring out the required angle can be obtained from the virtual binoculars direction sensor.
- the system supports rendering a scene from various angles simultaneously.
- the set is synchronized by exposure, white balance and frame per second rate.
- G-R-M master G-R-M camera of the group (R takes values from 1 to K). This type of camera might be also used for obtaining a scene image if the group consists of 2 cameras only.
- Camcorders CI - C5 are used to capture the scene. Each camera or group of cameras takes a specific part of the scene. Scopes of the cameras are partly overlapping each other.
- This solution might be used whenever 2D or 3D real-time reconstruction of complex and extended in space scenes is required.
- o 360 degrees view systems providing a panoramic view outside the vehicle o top view systems, providing a virtual view from the top
- the system contains the following blocks:
- the groups might be missing, or there might be few of them, depending on the number of objects that can't be covered by one video camera.
- Camcorders C-l,.., C-P, G-R-C-l,..., G-R-C-N, G-R-M transfer the video stream to CC computer
- CC computer interacts with camcorders G-R-C-l,..., G-R-C-N, G-R-M for calculating and making up the shooting settings for the groups of cameras
- the displays can be integrated with the elevator or landing doors, as shown in Figures 1 - 6.
- Figure 1 shows the system with large LCD displays integrated into the elevator landing doors (i.e. the doors in each floor of the building).
- FIG 2 shows the system with large LCD displays integrated into the elevator car doors (i.e. the doors in the elevator car).
- the LCD displays can be positioned anywhere in the elevator car, and can cover (all of some of) one, two or more walls (and also the roof and floor).
- Figure 3 shows the system with the displays integrated into the elevator landing doors (i.e. the doors in each floor of the building); in this case projectors mounted over the doors and in the landing door frame project the image onto the displays (which can be a conventional projector screens).
- Figure 6 shows the system with large projection displays integrated into the elevator car doors (i.e. the doors in the elevator car).
- the displays can be positioned anywhere in the elevator car, and can cover (all of some of) one, two or more walls (and also the roof and floor). Projectors are integrated into the elevator car ceiling.
- the 3D model combines both static elements that do not alter in real-time, and also dynamic elements that do alter in real-time or near real-time.
- a method of showing a scene viewed by multiple digital cameras comprising the steps of:
- a system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which:
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- Studio Devices (AREA)
- Processing Or Creating Images (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Multiple digital cameras view a large scene, such as a part of a city. Some of the cameras view different parts of that scene, and video feeds from the cameras are processed at a computer to generate a photo-realistic synthetic 3D model of the scene. This enables the scene to be viewed from any viewing angle, including angles that the original, real cameras do not occupy - i.e. as though viewed from a 'virtual camera' that can be positioned in any arbitrary position. The 3D model combines both static elements that do not alter in real-time, and also dynamic elements that do alter in real-time or near real-time.
Description
A SYSTEM COMPRISING MULTIPLE DIGITAL CAMERAS VIEWING A LARGE SCENE
BACKGROUND OF THE INVENTION
1. Field of the Invention
This invention relates to a system comprising multiple digital cameras viewing a large scene, such as a part of a city. Some of the cameras view different parts of that scene, and video feeds from the cameras are processed at a computer to generate a photo-realistic synthetic 3D model of the scene. This enables the scene to be viewed from any viewing angle, including angles that the original, real cameras do not occupy - i.e. as though viewed from a 'virtual camera' that can be positioned in any arbitrary position.
2. Description of the Prior Art
Existing 'virtual camera' solutions do not allow setting a virtual viewing angle (i.e. a viewing angle of a scene that does not coincide with the actual shooting angle of a real camera) without noticeable distortion. Nor do earlier systems keep the correct proportions of objects in the frame as the position of the virtual camera alters. Nor do earlier systems enable viewing angle amendments to be made in real time.
Also the existing solutions do not allow obtaining images from an arbitrary number of cameras that might be used in, for instance, an automatic multi-camera computer vision system. Automatic settings of each video camera in a multi-camera system - white balance, precise shutter opening speed synchronization, shutter, aperture and the frame rate - would be very different for all of these cameras, which makes working with and processing a set of images obtained from an array of these cameras very complicated.
Unlike the existing modeling systems (for example, lidar-based ones), the performance of implementations of this invention are based on conventional consumer-grade digital video
cameras, which significantly reduces the cost of the solution. These implementations do not cause adverse health effects with emissions and do not interact with other active systems.
• US 20140015832 describes a system reconstructing a 3D model from a set of video files obtained from shooting a single scene. The video files are synchronized with a global event (e.g. a flash) then the individual position of each camera is calculated for every frame. Presumably the video cameras are not fixed immovably. Then a 3D model reconstruction is processed, the textures are selected and mapped on the 3D model, and weather conditions are identified and simulated. The system is mostly intended for car tracking. Unlike the system we describe, this system does not rely upon permanently fixed cameras, the reconstruction is undertaken for each frame individually, which makes real-time processing in high-resolution impossible.
• US6940538 describes a video synthesis system based on data obtained from a real time video camera and an additional synthetic object. The system contains one fixed video camera providing textures for a pre-defined model. A synthetic object is introduced in the model. The video is synthesized from the position of a real camera and the scope coinciding with the real camera and reflecting the synthetic object as a part of the scene. This solution relies upon one video camera usage, which does not allow shooting a wide scale scene or amending the viewing position from which the video is synthesized.
• US 6674461 describes a method of shooting a scene with a few video cameras; the footage obtained is mapped on a synthesized virtual scene. The approach described is designed for shooting a sole object, not an entire scene, so its authors do not provide a solution for fixing the lighting difference across the scene. The problem of setting up the cameras for real time shooting is not solved either. Also the system, unlike our one, has a limitation of 3 cameras only.
• US 20120314089 is a method of time-based synchronizing a system of cameras. Auto calibration based on analysis of images obtained from the cameras is used here. Unlike
our system, this method relies upon time-based synchronizing only; no solution for synchronizing white balance and exposure settings is provided.
• EP1352521 describes a system of a few video cameras used for generating an optimal image of an object. Along with the cameras covering partial view only, a wide-angle video camera covering the entire view is used here. Unlike our system, this approach does not provide any solution for synchronizing of settings, and the wide-angle camera's settings are not used for controlling other cameras.
• EP2695371 describes a system designed for 3D photo shooting, with light distributed evenly between a couple of cameras using a mirror, which achieves the same level of illumination for both of those cameras. Unlike our system, the method relies upon a mechanical synchronization (a mirror) only, no digital system of setting synchronization of two cameras is described (e.g., one camera's settings are not transferred to another one).
SUMMARY OF THE INVENTION
1. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the computer is programmed such that
(c) the 3D model combines both static elements that do not alter in real-time, and also dynamic elements that do alter in real-time or near real-time.
Optional features include any one or more of the following:
• the computer can selectively turn-on dynamic modeling of specific portions of the scene.
• a motion detection algorithm can determine if a specific portion of the scene is changing and hence requires dynamic modeling.
• the computer controls the parameters or settings on multiple cameras, such as white balance, shutter speed, aperture settings, frame rate, so that the 3D model can seamlessly present the view from those multiple cameras.
• the parameters or settings are synchronized or made equal at levels that the computer determines is optimal.
• the parameters or settings are varied to levels that the computer determines is optimal to ensure that no areas are under or over-exposed.
• the system also includes one or more cameras whose parameters are independent and hence not are synchronized or made equal with any other cameras.
• the cameras include a master camera whose view overlaps the view from several of the other cameras.
• the feed from the master camera determines white balance and/or exposure for some of the other cameras.
• the feed from the master camera is not used to generate the photo-realistic 3D model of the scene.
• the image feed from several cameras is processed to generate a texture for a specific surface or object in the scene.
• effects associated with real-time weather conditions are added to the 3D model.
Another aspect is an elevator including a display showing images generated by a method as described above.
A third aspect are landing doors including a display showing images generated by a method as described above.
BRIEF DESCRIPTION OF THE DRAWINGS
Figures 1 - 6 are schematic illustrations of an implementation of the invention used in an elevator.
Figure 7 is a schematic illustration of an implementation of the invention showing cameras mounted on a building and looking out over a city scene.
Figure 8 shows the hardware used in the system.
Figure 9 is a schematic illustration of an implementation of the invention showing cameras, including a master camera, viewing a building
DETAILED DESCRIPTION
An implementation of the invention is a hardware-software system that displays realistic three-dimensional scenes based on 3D-models and video streams. It also allows displaying the scene from different angles, i.e. as a virtual camera system.
The video stream is taken from an array of video cameras, shooting the desired scene. This array of video cameras has a number of settings that are controlled in a real-time mode.
The system controls white balance, camera shutter opening times, shutter speed, aperture and the camera frame rate automatically, analyzing the images obtained from each camera individually. Settings can be applied to a specific camcorder and transmitted in parametric form to each camera; each setting is transmitted as a function of or depending on the particular camera's position. An automatic tuning system of the cameras allows displaying correctly the area of the scene shot simultaneously by multiple cameras, even if the shooting conditions prevent the use of automatic settings.
The obtained video stream is processed and used for scene modeling. The solution proposed combines static modeling of 3D scenes and dynamic simulation of individual parts of the scene. When using only dynamic simulation, significant computing power is required, the precision of the model is sufficiently low, the dependence on weather conditions is considerably high, and the recognition applies on for example Lambertian surfaces only (which excludes glares, water surface, etc.).
When using static modeling only, the presence of moving objects or new objects in the scene causes significant model distortion.
Using both static and dynamic modeling allows avoiding most of the drawbacks of these approaches. It significantly reduces the computational load required for the model creation, a considerable number of non-Lambertian surfaces can be treated statically, static objects can be modeled with a maximum possible accuracy (up to 1 pixel of the input image from a camera), motion of objects in the scene (e.g., trees) are processed in a dynamic manner (e.g. in real time), new objects in the scene are detected and modeled dynamically as well. In diverse weather conditions, when the usage of algorithms for dynamic modeling is unfeasible, the system goes on operating based on static modeling.
In addition, for the parts of the scene with no moving objects, only static simulation might be used, which brings the consumption of computing resources to a minimum.
The implementation allows switching on dynamic modeling for any zone of the scene with flexibility, which is essential for quick modification of the model in case the scene is changing. The system uses a set of algorithms for modeling and working with images:
• mapping frames obtained from different cameras to get the full texture of an object (for example, if there is an object in the scene partially overlapping another object, multiple cameras are used for obtaining the full texture of the second object)
• combining object textures obtained from multiple cameras into a single texture to design the best view of the object (for example, in the case of insufficiency of the camera matrix, or dynamic range, then multiple cameras with different exposure settings would be used to generate the texture)
• joining images obtained from multiple cameras to get a panoramic view or for modeling an image that cannot be shot with one camera
• separation of objects from their background based on the information about the colors and brightness of the objects or the background for scene modeling on different levels (for example, making a tree popping up on the foreground)
• separation of objects from their background using the effect of optical focus on the object and the camcorder segmentation algorithms (e.g., highlighting a tree or a banner on the front or middle ground)
• selection of objects using data obtained from multiple cameras with stereovision algorithms
• getting the depth map of a whole scene or its part using stereovision algorithms
• highlighting background and moving objects on the images obtained from the cameras using motion detection algorithms
• tracking images of moving objects using tracking algorithm
• depth map generation at the stage of statistical modeling to be used in dynamic simulation algorithms
Also the method provides the following options:
• Adding extra artificial elements to the generated scene model or editing existing elements of the scene (for example, adding information banners, displaying static or moving objects, hiding unwanted elements of the scene, etc.)
• Recording and playing back 3D scenes at any instance of the system (for example, a football game recorded can be played back on home TV with the option of changing the rendering angle at any time)
Depending on the type of scene, different sets of algorithms can be used. The choice of set can be made manually or automatically basing on the attributes of the scene model.
For example, if freestanding poles or trees are presented in the scene, an algorithm of separation from the background based on the color or brightness characteristics of the
background should be used. If moving objects are pictured in the scene, such as vehicles, people, etc., then an allocation algorithm and a tracking moving objects and background extraction algorithm should be used. Flags on the stage can be defined either by stereovision algorithms or by separating them from the background based on the color or brightness characteristics of the flags or of the background.
The resulting image sent to the displaying device is obtained from the render of the current 3D model angle with mapped textures, from the point of the virtual camera's current location. The angle can be set and modified manually by the user at runtime to obtain the desired look. Also the angle can be set automatically based on data obtained from external sensors. For example, when designing a virtual window in a lift, the angle is determined by the height sensor reading the lift position. Or, for example, to display a panorama, the data for figuring out the required angle can be obtained from the virtual binoculars direction sensor. The system supports rendering a scene from various angles simultaneously.
Weather conditions can be recognized and reconstructed in a scene model. Also, while rendering, a virtual glass effect relevant to the weather conditions can be added. For example, water drops on the glass may be added.
Processing algorithms can be performed either on a PC, or on dedicated computers, such as:
• GPU as parts of PC
• dedicated evaluation boards, including DSP or FPGA
• computing servers, e.g. based on GPU, DSP, FPGA
• remote computing cluster
The system allows distributing the load between the calculators to ensure optimal system performance in real-time mode and to minimize its cost.
Thus, only the final result of the processing is transmitted to the subsystem for displaying on screens. The processing subsystem of the computing cluster executes capture from cameras and all the necessary treatment.
Core features of an implementation are a software-hardware system including a set of cameras, a set of PC computing modules, a description of the static and dynamic elements for the 3D model scene, a description of camcorders' configurations, a description of rendered scene output configurations, a description of the computing modules configuration. The number of cameras, PCs, computing modules, configurations and descriptions may vary depending on the particular scene the system is used for.
The system has the following features:
• possibility of static and dynamic modeling based on data obtained from cameras located at predetermined fixed positions or computed positions
• possibility to use only dynamic or only static scene modeling
• support of retrieving static model developed in another application designed for 3D modeling (e.g., Adobe 3D Studio Max)
• support of the static model generating by dynamic algorithms of the system
• support of statically or dynamically configurable set of scene modeling algorithms:
o separation of objects from their background based on the data of colors and brightness of objects or background for modeling different levels of the scene o highlighting objects using data obtained from multiple cameras using stereovision algorithms
o obtaining the depth map of the entire scene or its part using stereovision algorithms
o highlighting background and moving objects on the images obtained from the cameras using motion detection algorithms
o tracking images of moving objects using tracking algorithms
o generating a depth map at the stage of statistical modeling for usage in dynamic simulation algorithms
• Support for statically or dynamically configurable sets of algorithms for textures:
o joining the fragments of texture obtained from multiple cameras to get the solid object texture
o mapping the textures of one object obtained from multiple cameras into a joint texture to provide the best quality image of the object
o docking or merging images obtained from multiple cameras to get a panoramic view, or for creating an image that can not be obtained from one camera
• support for connectivity with additional scene modeling or texture processing algorithms
• support for real-time modeling and rendering
• support for manual and automatic angle selection for rendering:
o in the manual mode, the user can change the angle while the system is configurable within the constraints of the model
o in the automatic mode, the angle can be set up based on the data obtained from external sensors
• support for simultaneous rendering from multiple angles
• support for various types of calculators with option of distributing the load between them:
o PC
o GPU as a part of these PCs
o specialized evaluation boards, including DSP or FPGA
o computing servers, including GPU, DSP, FPGA based
o remote computing cluster
• ability to bring the reconstructed scene to any number of screens
• adding an additional artificial element to the generated scene model, or editing the existing elements of the scene
• possibility of recording and playing 3D scenes on any instance of the system
• support for recognition and modeling of weather conditions
• the mutual positions of the camcorders and their shooting direction is known at any time
• camera shooting direction is chosen based on the need to cover a certain scene containing both still and moving objects. Coverage is determined by the multi-angle scene display.
• The camcorders sets with special controlling cameras may be selected in the system.
The view of the controlling camcorder is partly overlapping the view of each camera in the set. Such sets of cameras and camcorders are allocated and installed to synchronize the parameters of the cameras in the group while shooting certain areas. In some particular cases these sets of cameras may consist of 2 cameras with partially overlapping views and with shooting options synchronized
• The set is synchronized by exposure, white balance and frame per second rate.
• The cameras are integrated into the network so that the entire scene image can be obtained at any time
• The cameras can capture images sequentially with different settings (for example, to combine multiple images into a single high dynamic range they may use HDR image)
• Usage of different video cameras and lens parameters (focal length, diaphragm) is enabled
• Simultaneous usage of multiple video cameras with various resolutions and different frame rates is enabled
System structure
The system contains the following components:
• C-l,.., C-P - individual cameras for obtaining a scene with independent shooting parameters. There might be a few cameras of this type, or there might be none of them. The system also may contain this type of camera only.
• G-l, G-K - groups of cameras synchronized based on the shooting parameters. The groups might be missing, or there might be few of them, depending on the number of objects that can't be covered by only one camera.
• G-R-C-l,..., G-R-C-N - group G-R cameras synchronized basing on the shooting parameters with the master G-R-M camera (R takes values from 1 to K).
• G-R-M - master G-R-M camera of the group (R takes values from 1 to K). This type of camera might be also used for obtaining a scene image if the group consists of 2 cameras only.
• CC - processing and controlling computer. It is recording video data from all cameras, performing processing algorithms, managing video cameras. There might be one or a few of those to distribute the load. They can be placed remotely.
• LAN - local computer network for transmitting data from cameras to the processing and controlling computer
• MON-1,..., MON-Q. - displays for monitoring the image results
System installation
From the installation point of view the system has some specific features, such as:
• Pre-determined mutual positioning and shooting directions for all the cameras
• The entire scene is covered
• The shooting areas of different cameras may partly overlap each other
• For the groups of cameras that require synchronization of shooting parameters an additional master camera shooting the same area (wholly or partly) might be installed.
• There is an option of installing a few video cameras for shooting a certain part of the scene to extend the dynamic range of shooting.
Regarding the installation of computers for processing and controlling, they might be located inside the same building, as in a remote data center, if the existing bandwidth is high enough. There might be from 1 to an entire array of computers, depending on the number of cameras used and on the complexity of the scene processed.
The display device comprises an array of monitors, from 1 to many. The monitors might be arranged inside the same building or in remote buildings.
System Functions
From the point of view of processing the images obtained from cameras and generating previews, the system performs the following cycle:
• obtaining image from video cameras
• reconstructing 3D model's dynamic elements
• obtaining texture of the objects
• mapping the texture onto a 3D model
• exposing the model display angle in accordance with user requirements or indication of external sensors
• rendering model in accordance with the configuration of displaying devices
• outputting rendered image on displaying device
From the point of view of controlling the cameras, the system performs the following cycle:
• obtaining the shooting settings: the camcorder interface for obtaining automatically calculated shooting settings
• parameters calculation: collecting scene data and establishing the optimal parameters for tuning cameras in the group together with the master camcorders
o computing the unified optimal shooting settings for cameras in the group o calculating shooting settings to provide the wide dynamic range of shooting the particular part of the scene
• applying the parameters for all the cameras in the group
Structure of the system
The system consists of:
• CI - a camera with individual shooting settings, shoots the sky with a fast shutter speed
• Gl - a group containing video cameras C2, C3, C4 with synchronized shooting settings and a master camcorder Ml, shoot Building A
• Ml - a master camcorder computing white balance and exposure for Gl group of cameras
• C5 - a camera with individual shooting settings, shoots the foreground with long exposure
• LAN - a local computer network for transmitting video data from cameras to a host computer and transferring control commands from the control computer to the camcorders
• PC - a computer for processing and operating the system, it's recording video data obtained from all cameras, setting up the shooting time parameters, operating the processing algorithms, controlling the shooting settings for the groups of cameras. There might be one or few of those to distribute the load. They can be placed remotely.
• MON - displays for monitoring the image results
• Building A - a high-rise building sitting on the stage
• Building B - a high-rise building with the cameras, LAN and PC are installed on stationary
Camcorders CI - C5 are used to capture the scene. Each camera or group of cameras takes a specific part of the scene. Scopes of the cameras are partly overlapping each other.
Ml camcorder is shooting the same part of the scene as the Gl group of cameras is (cameras C2-C4 in total). Ml camera settings are transferred to PC via LA. MC computer is establishing shooting settings for Gl group of cameras basing on the settings transferred from Ml.
CI - C5 cameras are calculating individual shooting settings automatically basing on frames regions in use.
Only images obtained from CI - C5 cameras are used for displaying the scene.
Advantages of the system
The advantages are as follows:
• obtaining the scene image of any scale in real time
• possibility of moving the display angle or virtual viewing position around the stage or scene
• possibility of balance loading by scaling system components
• possibility of processing static views and the views containing dynamic objects
• high quality images
• use of standard camcorders that are easy to install and maintain
• the system does not interfere with other devices and doesn't cause any harm to people, as the system is passive
• quality of footage for the entire scene or its parts may be improved by extending the shooting dynamic range
• resolution of the object displayed can be increased by shooting its parts with different cameras sharing synchronized settings
• one or more angles of a scene can be changed at any time, as the scene is covered completely by all cameras with synchronized settings
The system allows solving the following scene shooting problems:
• presence of dynamic objects - due to the dynamic modeling, the objects are displayed on the stage/scene correctly
• texture distortion with angle changing - due to the texture mapping on a 3D model, as well as due to the possibility of combining textures obtained from different cameras, a high quality image taken from any arbitrary angle is available
• significant scatter of light levels at different parts of the scene - in such cases a dynamic range of one camera matrix is not sufficient to shoot the entire scene. The system allows configuring the scene shooting using multiple cameras with different exposure settings automatically, so that the scene is without underexposed or overexposed areas at any time
• diverse white balances and illumination levels across the scene - as a result of differences between illumination levels and the types of lighting sources in the particular parts of the scene where white balance and exposure are measured. It is particularly complicated if the junctions of the areas are filmed with different cameras and the joints between them don't fit the real joints between the objects in the scene. For example, if the lower part of the building is filmed with one camcorder in artificial
light and the upper part of it is filmed with another camcorder in much more natural bright light, it's essential that there's no a sharp transition between the parts of the building when displaying the entire image. The system allows synchronizing white balance for the cameras shooting these parts of scene that need to have the same white balance and illumination level
• time varying illumination of different parts of a scene - areas of a scene can be illuminated diversely at different times of day, so static camera tune-up becomes impossible. For example, in afternoon the entire scene can be evenly brightly lit, but at sunset only the top parts of the scene will be lit brightly enough, and the bottom parts will be in shade. The system can automatically adjust for light amendments using master cameras
Areas of application
This solution might be used whenever 2D or 3D real-time reconstruction of complex and extended in space scenes is required.
Thus, implementation of the following systems becomes feasible:
• real-time city panorama displaying and synthesis of its view from any point placed within the area covered by the cameras
o If the cameras are positioned vertically, a synthesis of views from any point on the vertical line between the lowest and the uppermost cameras becomes available. Movement of an elevator might be synthesized
o If the cameras are positioned horizontally, a synthesis of horizontal view is available. A cyclorama might be synthesized
o Matrix camera positioning allows moving the point of synthetic scene view in two dimensions within the coverage of the cameras system. A cycloramic elevator's motion view might be synthesized
• 3D reconstruction of moving objects (e.g. people, cars) by pre-constructing a stationary objects' model and predicting the moving objects' motion
• Identification and selection of objects in the scene by installing cameras with overlapping views in the areas where the recognition is required
• Adding artificial objects to the scene. City panorama with markers of attractions, information boards, etc. might be synthesized
• Creation of "virtual" windows in buildings with option of moving the angle automatically (for example, if the virtual window is placed in a moving elevator)
• Systems of assistance for drivers and crews:
o 360 degrees view systems providing a panoramic view outside the vehicle o top view systems, providing a virtual view from the top
• Shooting and 3D reconstruction of events. The system allows moving the virtual camera and seeing the event from the most advantageous angles, in real time or as a playback.
• Virtual guiding. The system allows to track a person, to comment on what he can see or guide him along the route
• Obtaining a nonstop image of a production line
• Obtaining a nonstop image of trains
• High effective resolution shooting by multiple cameras with narrower angle lens
Flowchart
Blocks description
The system contains the following blocks:
• C-l,.., C-P - single video cameras for obtaining the scene view with independent shooting settings
• G-l, G-K - groups of cameras synchronized by shooting settings. The groups might be missing, or there might be few of them, depending on the number of objects that can't be covered by one video camera.
• G-R-C-l,..., G-R-C-N - video cameras of G-R group synchronized by shooting settings with G-R-M master camcorder (R takes values from 1 to K).
• G-R-M - master-camcorder of G-R group (R takes values from 1 to K). This camera might be also used for obtaining scene image if the group consists of 2 cameras only.
• CC - a computer for processing and operating the system. There might be one or a few of those to distribute the load.
• LAN - local computer network for transmitting data from cameras to the processing and controlling computer
• MON-1,..., MON-Q - displays for monitoring the image results
Blocks interfaces
• Camcorders C-l,.., C-P, G-R-C-l,..., G-R-C-N, G-R-M transfer the video stream to CC computer
• Camcorders G-R-C-l,..., G-R-C-N support software interface of the system, real-time shooting settings white balance, exposure
• CC computer interacts with camcorders G-R-C-l,..., G-R-C-N, G-R-M for calculating and making up the shooting settings for the groups of cameras
• CC computer is interacting with MON-1,..., MON-Q. displays providing the output image
• All the cameras, operating computers and displays are interacting through LAN local computer network
• Sensors of angle automatic settings for image are connected to a display subsystem Figures
The displays can be integrated with the elevator or landing doors, as shown in Figures 1 - 6.
Figure 1 shows the system with large LCD displays integrated into the elevator landing doors (i.e. the doors in each floor of the building).
Figure 2 shows the system with large LCD displays integrated into the elevator car doors (i.e. the doors in the elevator car). The LCD displays can be positioned anywhere in the elevator car, and can cover (all of some of) one, two or more walls (and also the roof and floor).
Figure 3 shows the system with the displays integrated into the elevator landing doors (i.e. the doors in each floor of the building); in this case projectors mounted over the doors and in the landing door frame project the image onto the displays (which can be a conventional projector screens).
Figure 4 shows the system with large projection displays integrated into the elevator car doors (i.e. the doors in the elevator car). The displays can be positioned anywhere in the elevator car, and can cover (all of some of) one, two or more walls (and also the roof and floor). Projectors are integrated into the elevator car in the door frame.
Figure 5 shows the system with the displays integrated into the elevator landing doors (i.e. the doors in each floor of the building); in this case projectors mounted over the doors and in the ceiling project the image onto the displays (which can be a conventional projector screens).
Figure 6 shows the system with large projection displays integrated into the elevator car doors (i.e. the doors in the elevator car). The displays can be positioned anywhere in the elevator car, and can cover (all of some of) one, two or more walls (and also the roof and floor). Projectors are integrated into the elevator car ceiling.
Figure 7 shows the above system shooting a scene for modulating a virtual window in an elevator in Building B ; there is a display in one or more elevators in Building B and the display shows the view of the city scene outside of Building B, including the Building A grouping, as though the elevator were a glass elevator- i.e. the small number of cameras fixed to the outside of Building A enable the system to generate a virtual camera positioned at any height on the Building B and so the display shows a smoothly changing scene that accurately corresponds to what would be seen from the height the elevator has reached. Permanent objects in the scene, such as roads, buildings, trees, are statically processed and a photo-realistic, fully textured computer generated 3D model of these objects is created which can be viewed from any angle. Moving objects, such as cars, trains, people, are dynamically captured in real time and
modeled. The display in the elevator can hence show a combination of highly detailed permanent objects, plus real-time, dynamically generated CGI of moving objects.
Figure 8 shows the general scheme of camcorders connections in the system
Figure 9 shows shooting a building using several cameras including the master camcorder, and shows the different parts of the scene that are captured and that are synthesized into the accurate 3D model.
Features
1. Cameras - analog and digital ones, able to obtain video stream in real time mode
2. Cameras network - any, allowing the connection and video stream transmission to network, directly or via devices (e.g., Ethernet, Fiber, USB)
3. Camera models - arbitrary, if shooting synchronization (white balance, exposure and ideally the shutter opening as well) is not required. Although if synchronization is essential, only cameras with this option available might be used (e.g., Beward BD3570 (in some cases), IDS 5481VSE-C)
4. Camera lenses - any, fitting the interface of the camera in use. It's advisable to use the lenses with minimal distortion. The lenses with a fixed focal length with distortion less than 1% and a resolution of 3 megapixels or more would suit perfectly (e.g. Kowa LM5JC10M, Kowa LM3NCM)
5. Camera parameters required - any cameras obtaining digital video stream in real time mode are suitable. If shooting synchronization (white balance, exposure and ideally the shutter opening as well) is required, only cameras with this option available might be used (e.g., Beward BD3570 (in some cases), IDS 5481VSE-C).
Key Concepts
Each of the following concepts A - G can be combined with any other concept A -G; all optional features can be combined with any other optional feature.
A. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the computer is programmed such that
(c) the 3D model combines both static elements that do not alter in real-time, and also dynamic elements that do alter in real-time or near real-time.
A system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which:
the 3D model combines both static elements that do not alter in real-time, and also dynamic elements that do alter in real-time or near real-time.
Note that the term 'computer' is not limited to a single computer, but can be any number of computers, for example operating in a distributed manner, with some computing elements optimized for off-line processing and some for real-time processing.
Other optional features:
• System can selectively turn-on dynamic modeling of specific portions of the scene.
• Motion detection algorithm can determine if a specific portion of the scene is changing and hence requires dynamic modeling
• Static elements are processed off-line whereas the dynamic elements are processed in real-time
B. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the computer controls the parameters or settings on multiple cameras, such as white balance, shutter speed, aperture settings, frame rate, so that the 3D model can seamlessly present the view from those multiple cameras.
A system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which:
the computer controls the parameters or settings on multiple cameras, such as white balance, shutter speed, aperture settings, frame rate, so that the 3D model can seamlessly present the view from those multiple cameras.
Other optional features:
• The parameters or settings are synchronized or made equal at levels that the computer determines is optimal.
• The parameters or settings are varied to levels that the computer determines is optimal to ensure that no areas are under or over-exposed
• System also includes one or more cameras whose parameters are independent and hence not are synchronized or made equal with any other cameras
C. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the cameras include a master camera whose view overlaps the view from several of the other cameras.
A system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which:
the cameras include a master camera whose view overlaps the view from several of the other cameras.
Other optional features:
• feed from the Master camera determines white balance and/or exposure for some of the other cameras
• feed from the master camera is not used to generate the photo-realistic 3D model of the scene
D. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the image feed from several cameras is processed to generate a texture for a specific surface or object in the scene.
A system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which:
the image feed from several cameras is processed to generate a texture for a specific surface or object in the scene.
E. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which effects associated with real-time weather conditions are added to the 3D model.
A system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D
model of the scene, that photo-realistic model being viewable on one or more displays; and in which:
effects associated with real-time weather conditions are added to the 3D model.
F. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the 3D model shows part of the scene in real-time, and other parts are computer-generated images from an imagined historical view, such as tens, hundreds or thousands of years ago.
A system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which:
the 3D model shows part of the scene in real-time, and other parts are computer- generated images from an imagined historical view, such as tens, hundreds or thousands of years ago.
Other optional features:
• the imagined historical view alters as the location of the display alters
• the display is in a lift and as the lift moves, then the view moves forward in time, for example starting at the distant past at the ground floor and moving forward in time as the lift ascends or descends
• the view at the top of the lift's position corresponds to the contemporary realtime view
• the view at the top of the lift's position corresponds not to the contemporary real-time view, but a view from a number of years previously, to give an element of surprise when a person sees the actual view.
G. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the elevator includes one or more of the displays showing the photo-realistic model.
A system comprising multiple digital cameras viewing a large scene, such as a part of a city, some of the cameras viewing different parts of that scene, and one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which: the elevator includes one or more of the displays showing the photo-realistic model.
Other optional features:
• the display is formed as part of the elevator car door(s), viewable when inside the elevator
• the display is formed as part of the landing door(s), viewable when waiting for an elevator
• when the display is formed on the landing door panels, we may offer 2 images: looking through the building (e.g. real time at the level of landing stop) or looking along the
elevator ride, synchronized with selected elevator car's current position (this indicates how far from the landing the car is now, when it will arrive etc. )
• the elevator doors and/or landing doors include large, flat panel displays, such as OLED or LCD.
• the landing door frame or nearby structures includes projector(s) which project the photo-realistic model onto the landing doors, viewable when waiting for an elevator
• the elevator includes projector(s) which project the photo-realistic model onto the elevator doors, viewable when inside the elevator.
Optional features relevant to all concepts A - G above
• A display can show the photo-realistic 3D model of the scene from a viewing position not occupied by any specific camera
• A display can show the photo-realistic 3D model of the scene moving between viewing positions not occupied by any specific camera
• Viewing position is set automatically depending on viewer's location, including height above ground
• Display is a display in an elevator in a building and the cameras are placed to provide a view from that building
• Display is a display on or near the doors of an elevator, and shows the view of the 3D model from the height at which the elevator is currently at.
• Display is a display in an elevator in a building and the cameras are placed anywhere in the world to provide real-time changing views from a different building, anywhere in the world, at the same height above ground as the elevator, as it moves up or down.
• Viewing position is set manually by a viewer, such as by selecting a view from a viewer- selected location, and the viewer can then alter the direction (e.g. compass heading) from which the scene is viewed.
• Viewing position is set manually by a viewer, such as by selecting a view from a viewer- selected location, and the viewer can then alter the height above ground that the scene is viewed from.
• Display is a display on or near the doors of an elevator, and shows the view of the 3D model from the height at which the elevator is currently at.
• Viewing position is independent of viewer's location
• Cameras are vertically aligned
• Cameras are horizontally aligned
• Cameras are arranged in a matrix
• Video feed deliver high dynamic range images
• Cameras are fixed to a building and point out to the environment around the building
• Any arbitrary number of cameras can be used
• Cameras can be in any number of different locations, such as in different cities and countries
• Cameras are fixed inside a building and point to the environment inside the building
• Cameras are fixed to a vehicle and point out to the environment around the vehicle
• Cameras are fixed to a wearable computing device and point out to the environment around the person wearing the device
• Virtual images, such as signage, are added to the 3D model
• Elevator displays could be used for any informational purpose, not only, or even, for the photo-realistic 3D model
• A depth scanner in the elevator (e.g. lidar, IR mapping, stereo camera) detects the presence of a single person traveling within the elevator and the system then works out their point of view and makes corresponding corrections to the picture perspective/parallax for the photo-realistic 3D model.
Claims
1. A method of showing a scene viewed by multiple digital cameras, comprising the steps of:
(a) some of the cameras viewing different parts of that scene;
(b) one or more video feeds from some or all of the cameras being processed at a computer to generate a photo-realistic 3D model of the scene, that photo-realistic model being viewable on one or more displays; and in which the computer is programmed such that
(c) the 3D model combines both static elements that do not alter in real-time, and also dynamic elements that do alter in real-time or near real-time.
2. The method of Claim 1 in which the computer can selectively turn-on dynamic modeling of specific portions of the scene.
3. The method of Claim 1 or 2 in which a motion detection algorithm running on the computer can determine if a specific portion of the scene is changing and hence requires dynamic modeling.
4. The method of any preceding Claim in which the static elements are processed off-line whereas the dynamic elements are processed in real-time.
5. The method of any preceding Claim in which the computer controls the parameters or settings on multiple cameras, such as white balance, shutter speed, aperture settings, frame rate, so that the 3D model can seamlessly present the view from those multiple cameras.
6. The method of preceding Claim 5 in which the parameters or settings are synchronized or made equal at levels that the computer determines is optimal.
7. The method of any Claim 5 or 6 in which the parameters or settings are varied to levels that the computer determines is optimal to ensure that no areas are under or over-exposed.
8. The method of any preceding Claim 6 - 7 in which one or more cameras have parameters that are independent and hence not are synchronized or made equal with any other cameras.
9. The method of any preceding Claim in which the cameras include a master camera whose view overlaps the view from several of the other cameras.
10. The method of preceding Claim 9 in which the feed from the master camera determines white balance and/or exposure for some of the other cameras.
11. The method of preceding Claim 9 or 10 in which the feed from the master camera is not used to generate the photo-realistic 3D model of the scene.
12. The method of any preceding Claim in which the image feed from several cameras is processed by the computer to generate a texture for a specific surface or object in the scene.
13. The method of any preceding Claim in which the effects associated with real-time weather conditions are added to the 3D model by the computer
14. The method of any preceding Claim in which the 3D model shows part of the scene in real-time, and other parts are computer-generated images from an imagined historical view, such as tens, hundreds or thousands of years ago.
15. The method of Claim 14 in which the imagined historical view alters as the location of the display alters.
16. The method of Claim 14 - 15 in which the display is in a lift or elevator and as the lift moves, then the view moves forward in time, for example starting at the distant past at the ground floor and moving forward in time as the lift ascends or descends.
17. The method of Claim 14 - 16 in which the view at the top of the lift's position corresponds to the contemporary real-time view.
18. The method of Claim 14 - 17 in which the view at the top of the lift's position corresponds not to the contemporary real-time view, but a view from a number of years previously, to give an element of surprise when a person sees the actual view.
19. The method of any preceding Claim in which an elevator includes one or more of the displays showing the photo-realistic model.
20. The method of Claim 19 in which the display is formed as part of the elevator car door(s), viewable when inside the elevator.
21. The method of Claim 19 or 20 in which the display is formed as part of the landing door(s), viewable when waiting for an elevator.
22. The method of Claim 21 in which when the display is formed on the landing door panels, the panels provide two images: looking through the building (e.g. real time at the level of the landing stop) or looking along the elevator ride, synchronized with the selected elevator car's current position.
23. The method of Claim 19 - 22 in which the elevator doors and/or landing doors include large, flat panel displays, such as OLED or LCD.
24. The method of Claim 19 - 23 in which the landing door frame or nearby structures includes projector(s) which project the photo-realistic model onto the landing doors, viewable when waiting for an elevator.
25. The method of Claim 19 - 24 in which the elevator includes projector(s) which project the photo-realistic model onto the elevator doors, viewable when inside the elevator.
26. The method of Claim 19 - 25 in which the displays are used for any informational purpose, and not only for the photo-realistic 3D model.
27. The method of any preceding Claim in which the display can show the photo-realistic 3D model of the scene from a viewing position not occupied by any specific camera.
28. The method of any preceding Claim in which the display can show the photo-realistic 3D model of the scene moving between viewing positions not occupied by any specific camera.
29. The method of preceding Claim 28 in which the viewing position is set automatically depending on viewer's location, including height above ground.
30. The method of any preceding Claim in which the display is a display in an elevator in a building and the cameras are placed to provide a view from that building.
31. The method of any preceding Claim in which the display is a display on or near the doors of an elevator, and shows the view of the 3D model from the height at which the elevator is currently at.
32. The method of any preceding Claim in which the display is a display in an elevator in a building and the cameras are placed anywhere in the world to provide real-time changing views
from a different building, anywhere in the world, at the same height above ground as the elevator, as it moves up or down.
33. The method of any preceding Claim in which a viewing position is set manually by a viewer, such as by selecting a view from a viewer-selected location, and the viewer can then alter the direction (e.g. compass heading) from which the scene is viewed.
34. The method of any preceding Claim in which a viewing position is set manually by a viewer, such as by selecting a view from a viewer-selected location, and the viewer can then alter the height above ground that the scene is viewed from.
35. The method of any preceding Claim in which a display is a display on or near the doors of an elevator, and shows the view of the 3D model from the height at which the elevator is currently at.
36. The method of any preceding Claim in which a viewing position is independent of a viewer's location.
37. The method of any preceding Claim in which at least some of the cameras are vertically aligned.
38. The method of any preceding Claim in which at least some of the cameras are horizontally aligned.
39. The method of any preceding Claim in which at least some of the cameras are arranged in a matrix.
40. The method of any preceding Claim in which a video feed delivers high dynamic range images.
41. The method of any preceding Claim in which at least some of the cameras are fixed to a building and point out to the environment around the building.
42. The method of any preceding Claim in which at least some of the cameras are fixed inside a building and point to the environment inside the building.
43. The method of any preceding Claim in which at least some of the cameras are fixed to a vehicle and point out to the environment around the vehicle.
44. The method of any preceding Claim in which at least some of the cameras are fixed to a wearable computing device and point out to the environment around the person wearing the device.
45. The method of any preceding Claim in which virtual images, such as signage, are added to the 3D model.
46. The method of any preceding Claim in which a depth scanner in the elevator (e.g lidar, IR mapping, stereo camera) detects the presence of a single person traveling within the elevator and the system then works out their point of view and makes corresponding corrections to the picture perspective/parallax for the photo-realistic 3D model.
47. An elevator including a display showing images generated by a method as described above in Claims 1 - 46.
48. Elevator landing doors including a display showing images generated by a method as described above in Claims 1 - 46.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201580049568.5A CN107003600A (en) | 2014-09-15 | 2015-09-15 | Including the system for the multiple digital cameras for observing large scene |
EP15766437.6A EP3219100A1 (en) | 2014-09-15 | 2015-09-15 | A system comprising multiple digital cameras viewing a large scene |
US15/510,858 US10375382B2 (en) | 2014-09-15 | 2015-09-15 | System comprising multiple digital cameras viewing a large scene |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB201416254A GB201416254D0 (en) | 2014-09-15 | 2014-09-15 | LiftEye 2014 |
GB1416254.9 | 2014-09-15 | ||
GB201418606A GB201418606D0 (en) | 2014-10-20 | 2014-10-20 | Lifteye october 2014 |
GB1418606.8 | 2014-10-20 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016041953A1 true WO2016041953A1 (en) | 2016-03-24 |
Family
ID=54148487
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2015/071083 WO2016041953A1 (en) | 2014-09-15 | 2015-09-15 | A system comprising multiple digital cameras viewing a large scene |
Country Status (4)
Country | Link |
---|---|
US (1) | US10375382B2 (en) |
EP (1) | EP3219100A1 (en) |
CN (1) | CN107003600A (en) |
WO (1) | WO2016041953A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110235440A (en) * | 2017-01-31 | 2019-09-13 | 株式会社木村技研 | Optical projection system and projecting method |
US10742894B2 (en) | 2017-08-11 | 2020-08-11 | Ut-Battelle, Llc | Optical array for high-quality imaging in harsh environments |
WO2021248564A1 (en) * | 2020-06-08 | 2021-12-16 | 厦门理工学院 | Panoramic big data application monitoring and control system |
CN115802165A (en) * | 2023-02-10 | 2023-03-14 | 成都索贝数码科技股份有限公司 | Lens moving shooting method applied to live connection of different places and same scenes |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10204444B2 (en) * | 2016-04-28 | 2019-02-12 | Verizon Patent And Licensing Inc. | Methods and systems for creating and manipulating an individually-manipulable volumetric model of an object |
WO2018031441A1 (en) * | 2016-08-09 | 2018-02-15 | Contrast, Inc. | Real-time hdr video for vehicle control |
TWI633786B (en) * | 2016-12-15 | 2018-08-21 | 晶睿通訊股份有限公司 | Image analyzing method and camera |
EP3573339A4 (en) * | 2017-01-19 | 2020-08-05 | Sony Interactive Entertainment Inc. | Image delivery device |
CN108470360B (en) * | 2017-02-23 | 2022-06-17 | 钰立微电子股份有限公司 | Image device for generating depth map by using non-plane projection image and related method thereof |
CN108024071B (en) * | 2017-11-24 | 2022-03-08 | 腾讯数码(天津)有限公司 | Video content generation method, video content generation device, and storage medium |
US11417107B2 (en) * | 2018-02-19 | 2022-08-16 | Magna Electronics Inc. | Stationary vision system at vehicle roadway |
US11745983B2 (en) | 2018-08-08 | 2023-09-05 | Otis Elevator Company | Elevator system with LIDAR and/or RADAR sensor |
US10715714B2 (en) * | 2018-10-17 | 2020-07-14 | Verizon Patent And Licensing, Inc. | Machine learning-based device placement and configuration service |
CN109889787B (en) * | 2019-02-27 | 2021-02-02 | 浙江大华技术股份有限公司 | Video transmission method, device and system |
BR102019006465A2 (en) * | 2019-03-29 | 2020-10-13 | Globo Comunicação E Participações S.a. | SYSTEM AND METHOD OF CAPTURE AND IMAGE PROJECTION, AND, USE OF THE SYSTEM. |
CN110335300A (en) * | 2019-05-14 | 2019-10-15 | 广东康云科技有限公司 | Scene dynamics analogy method, system and storage medium based on video fusion |
US11095467B2 (en) * | 2019-08-16 | 2021-08-17 | Logitech Europe S.A. | Video conference system |
JP2021052293A (en) * | 2019-09-24 | 2021-04-01 | ソニー株式会社 | Information processing device, information processing method, and information processing program |
US11159713B2 (en) * | 2019-10-11 | 2021-10-26 | Varjo Technologies Oy | Imaging system and method of producing images |
US11734767B1 (en) | 2020-02-28 | 2023-08-22 | State Farm Mutual Automobile Insurance Company | Systems and methods for light detection and ranging (lidar) based generation of a homeowners insurance quote |
US12086861B1 (en) | 2020-04-27 | 2024-09-10 | State Farm Mutual Automobile Insurance Company | Systems and methods for commercial inventory mapping including a lidar-based virtual map |
KR102217759B1 (en) * | 2020-08-12 | 2021-02-22 | 주식회사 다지트 | Display device, elevator assembly including the same and method for image processing |
CN114430454A (en) * | 2020-10-28 | 2022-05-03 | 广东小天才科技有限公司 | Modeling method based on double cameras, wearable device, equipment and medium |
CN112486133A (en) * | 2020-12-09 | 2021-03-12 | 安徽三达信息科技有限公司 | Intelligent piece arranging and sorting system based on vertical form AGV |
US12073010B2 (en) | 2022-07-01 | 2024-08-27 | State Farm Mutual Automobile Insurance Company | VR environment for accident reconstruction |
US11790776B1 (en) | 2022-07-01 | 2023-10-17 | State Farm Mutual Automobile Insurance Company | Generating virtual reality (VR) alerts for challenging streets |
Family Cites Families (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06345370A (en) * | 1993-06-08 | 1994-12-20 | Mitsubishi Denki Bill Techno Service Kk | Door for elevator landing |
US5480305A (en) * | 1993-10-29 | 1996-01-02 | Southwest Research Institute | Weather simulation system |
US5850352A (en) * | 1995-03-31 | 1998-12-15 | The Regents Of The University Of California | Immersive video, including video hypermosaicing to generate from multiple video views of a scene a three-dimensional video mosaic from which diverse virtual video scene images are synthesized, including panoramic, scene interactive and stereoscopic images |
US5729471A (en) * | 1995-03-31 | 1998-03-17 | The Regents Of The University Of California | Machine dynamic selection of one video camera/image of a scene from multiple video cameras/images of the scene in accordance with a particular perspective on the scene, an object in the scene, or an event in the scene |
JPH11168754A (en) * | 1997-12-03 | 1999-06-22 | Mr System Kenkyusho:Kk | Image recording method, image database system, image recorder, and computer program storage medium |
US6674461B1 (en) | 1998-07-07 | 2004-01-06 | Matthew H. Klapman | Extended view morphing |
EP1297634A1 (en) * | 2000-06-09 | 2003-04-02 | iMove Inc. | Streaming panoramic video |
US6577333B2 (en) | 2000-12-12 | 2003-06-10 | Intel Corporation | Automatic multi-camera video composition |
FI112792B (en) * | 2001-08-14 | 2004-01-15 | Kone Corp | Elevator informative floor number display on the floor and in the basket |
US6940538B2 (en) | 2001-08-29 | 2005-09-06 | Sony Corporation | Extracting a depth map from known camera and model tracking data |
US7031525B2 (en) * | 2002-07-30 | 2006-04-18 | Mitsubishi Electric Research Laboratories, Inc. | Edge detection based on background change |
US7728853B2 (en) * | 2005-11-15 | 2010-06-01 | Honeywell International Inc. | Systems and methods for changing view perspective in 3-D graphical displays of buildings with stepped rotation |
US8384710B2 (en) * | 2007-06-07 | 2013-02-26 | Igt | Displaying and using 3D graphics on multiple displays provided for gaming environments |
CN101249914A (en) | 2008-04-17 | 2008-08-27 | 西子奥的斯电梯有限公司 | Elevator operation system, method and elevator operation information display system |
US8594467B2 (en) * | 2008-12-19 | 2013-11-26 | Microsoft Corporation | Interactive virtual display system for ubiquitous devices |
US9432591B2 (en) * | 2009-01-05 | 2016-08-30 | Duke University | Multiscale optical system having dynamic camera settings |
US20120188279A1 (en) * | 2009-09-29 | 2012-07-26 | Kent Demaine | Multi-Sensor Proximity-Based Immersion System and Method |
US10375287B2 (en) * | 2009-10-21 | 2019-08-06 | Disney Enterprises, Inc. | Object trail-based analysis and control of video |
CN102109751B (en) * | 2009-12-29 | 2013-08-07 | 财团法人工业技术研究院 | Miniaturized imaging module, three-dimensional display system using same and image arrangement method thereof |
US8913009B2 (en) * | 2010-02-03 | 2014-12-16 | Nintendo Co., Ltd. | Spatially-correlated multi-display human-machine interface |
KR20140026458A (en) | 2011-04-08 | 2014-03-05 | 파커비전, 인크. | Systems and methods of rf power transmission, modulation, and amplification |
JP2014514321A (en) | 2011-04-21 | 2014-06-19 | メルク・シャープ・アンド・ドーム・コーポレーション | Insulin-like growth factor 1 receptor inhibitor |
US9066024B2 (en) | 2011-06-08 | 2015-06-23 | Christopher C. Chang | Multi-camera system and method of calibrating the multi-camera system |
US9286711B2 (en) * | 2011-09-30 | 2016-03-15 | Microsoft Technology Licensing, Llc | Representing a location at a previous time period using an augmented reality display |
US9229231B2 (en) * | 2011-12-07 | 2016-01-05 | Microsoft Technology Licensing, Llc | Updating printed content with personalized virtual data |
WO2013139814A2 (en) * | 2012-03-19 | 2013-09-26 | Fittingbox | Model and method for producing photorealistic 3d models |
US20160330408A1 (en) * | 2015-04-13 | 2016-11-10 | Filippo Costanzo | Method for progressive generation, storage and delivery of synthesized view transitions in multiple viewpoints interactive fruition environments |
-
2015
- 2015-09-15 US US15/510,858 patent/US10375382B2/en active Active
- 2015-09-15 WO PCT/EP2015/071083 patent/WO2016041953A1/en active Application Filing
- 2015-09-15 CN CN201580049568.5A patent/CN107003600A/en active Pending
- 2015-09-15 EP EP15766437.6A patent/EP3219100A1/en not_active Withdrawn
Non-Patent Citations (3)
Title |
---|
CEVAHIR CIGLA ET AL: "Segmentation in multi-view video via color, depth and motion cues", 15TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING : ICIP 2008 ; SAN DIEGO, CALIFORNIA, USA, 12 - 15 OCTOBER 2008, IEEE, PISCATAWAY, NJ, USA, 12 October 2008 (2008-10-12), pages 2724 - 2727, XP031374604, ISBN: 978-1-4244-1765-0 * |
HANSUNG KIM ET AL: "A Real-Time 3D Modeling System Using Multiple Stereo Cameras for Free-Viewpoint Video Generation", 1 January 2006, IMAGE ANALYSIS AND RECOGNITION LECTURE NOTES IN COMPUTER SCIENCE;;LNCS, SPRINGER, BERLIN, DE, PAGE(S) 237 - 249, ISBN: 978-3-540-44894-5, XP019043784 * |
PAPADAKIS N ET AL: "Virtual Camera Synthesis for Soccer Game Replays", VISUAL MEDIA PRODUCTION (CVMP), 2010 CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, 17 November 2010 (2010-11-17), pages 97 - 106, XP031853451, ISBN: 978-1-4244-8872-8 * |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110235440A (en) * | 2017-01-31 | 2019-09-13 | 株式会社木村技研 | Optical projection system and projecting method |
CN110235440B (en) * | 2017-01-31 | 2022-08-23 | 株式会社木村技研 | Projection system and projection method |
US10742894B2 (en) | 2017-08-11 | 2020-08-11 | Ut-Battelle, Llc | Optical array for high-quality imaging in harsh environments |
US11601601B2 (en) | 2017-08-11 | 2023-03-07 | Ut-Battelle, Llc | Optical array for high-quality imaging in harsh environments |
WO2021248564A1 (en) * | 2020-06-08 | 2021-12-16 | 厦门理工学院 | Panoramic big data application monitoring and control system |
US11703820B2 (en) | 2020-06-08 | 2023-07-18 | Xiamen University Of Technology | Monitoring management and control system based on panoramic big data |
CN115802165A (en) * | 2023-02-10 | 2023-03-14 | 成都索贝数码科技股份有限公司 | Lens moving shooting method applied to live connection of different places and same scenes |
CN115802165B (en) * | 2023-02-10 | 2023-05-12 | 成都索贝数码科技股份有限公司 | Lens moving shooting method applied to live broadcast connection of different places and same scene |
Also Published As
Publication number | Publication date |
---|---|
US20170264890A1 (en) | 2017-09-14 |
EP3219100A1 (en) | 2017-09-20 |
CN107003600A (en) | 2017-08-01 |
US10375382B2 (en) | 2019-08-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10375382B2 (en) | System comprising multiple digital cameras viewing a large scene | |
US11727644B2 (en) | Immersive content production system with multiple targets | |
US9436076B2 (en) | Multi-projection system for extending visual element of main image | |
US10990342B2 (en) | Multi-viewpoint switched shooting system and method | |
US10853992B1 (en) | Systems and methods for displaying a virtual reality model | |
US20180343441A1 (en) | Perspective altering display system | |
CN102540464B (en) | Head-mounted display device which provides surround video | |
CN110866978A (en) | Camera synchronization method in real-time mixed reality video shooting | |
WO2019014285A1 (en) | Method, apparatus and system providing alternative reality environment | |
CA2913218A1 (en) | Systems and methods for a shared mixed reality experience | |
KR20070119018A (en) | Automatic scene modeling for the 3d camera and 3d video | |
BR122021006807B1 (en) | METHOD FOR INCREASED ENTERTAINMENT EXPERIENCES IN LAYERS | |
CN106125491A (en) | Many optical projection systems | |
CN111698391B (en) | Method for controlling real-time change of light parameters through simulated environment light parameters | |
CN113692734A (en) | System and method for acquiring and projecting images, and use of such a system | |
Takatori et al. | Large-scale projection-based immersive display: The design and implementation of largespace | |
Kim et al. | 3-d virtual studio for natural inter-“acting” | |
Ogi et al. | Usage of video avatar technology for immersive communication | |
Santos et al. | Display and rendering technologies for virtual and mixed reality design review | |
Bimber et al. | Digital illumination for augmented studios | |
CN105264438A (en) | Multi-projection system capable of refracting projection light of projection device | |
KR102654323B1 (en) | Apparatus, method adn system for three-dimensionally processing two dimension image in virtual production | |
CN118216136A (en) | Information processing apparatus, image processing method, and program | |
CN115966143A (en) | Background display device | |
CN117459702A (en) | Digital-real fusion naked eye 3D presentation method, system and related equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15766437 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15510858 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REEP | Request for entry into the european phase |
Ref document number: 2015766437 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2015766437 Country of ref document: EP |