WO2020194190A1 - Systems, apparatuses and methods for acquiring, processing and delivering stereophonic and panoramic images - Google Patents

Systems, apparatuses and methods for acquiring, processing and delivering stereophonic and panoramic images Download PDF

Info

Publication number
WO2020194190A1
WO2020194190A1 PCT/IB2020/052766 IB2020052766W WO2020194190A1 WO 2020194190 A1 WO2020194190 A1 WO 2020194190A1 IB 2020052766 W IB2020052766 W IB 2020052766W WO 2020194190 A1 WO2020194190 A1 WO 2020194190A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
live
panoramic
images
stereoscopic
Prior art date
Application number
PCT/IB2020/052766
Other languages
French (fr)
Inventor
Tomer MAROM
Shahar BIN-NUN
Rutha Aronson
Ilya Kottel
Original Assignee
Humaneyes Technologies Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Humaneyes Technologies Ltd. filed Critical Humaneyes Technologies Ltd.
Publication of WO2020194190A1 publication Critical patent/WO2020194190A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/111Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/167Synchronising or controlling image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/172Processing image signals image signals comprising non-image signal components, e.g. headers or format information
    • H04N13/178Metadata, e.g. disparity information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/194Transmission of image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/243Image signal generators using stereoscopic image cameras using three or more 2D image sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/282Image signal generators for generating image signals corresponding to three or more geometrical viewpoints, e.g. multi-view systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/286Image signal generators having separate monoscopic and stereoscopic modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/296Synchronisation thereof; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/21805Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4122Peripherals receiving signals from specially adapted client devices additional display device, e.g. video projector
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/4728End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for selecting a Region Of Interest [ROI], e.g. for requesting a higher resolution version of a selected region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/45Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/90Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2213/00Details of stereoscopic systems
    • H04N2213/001Constructional or mechanical details

Definitions

  • the present invention relates to the acquiring of live video images and delivering them as panoramic and/or stereoscopic video streams for display on client viewing devices, methods and systems that incorporate stitching and stereo- synthesizing modules, and imaging devices for acquiring multiple synchronized images that can be stitched and/or synthesized into panoramic and/or stereo video streams.
  • panoramic images include panoramic images of more than 180° in breadth and even up to 360°, i.e., surround images that can be used in‘virtual reality’ viewers or even on regular displays which can allow a viewer to shift their view within the panorama; such a view-shifting viewing feature is available, for example, on the popular video website Youtube.
  • Creating panoramic images from multiple images involves techniques such as stitching the images together and resolving any overlap between them.
  • Another example of complex images is the stereoscopic image, which creates or enhances the illusion of depth in an image by means of stereopsis for binocular vision.
  • This illusion of depth is what is commonly known as‘3D’ (short for three-dimensional) and is popular in various types of media.
  • 3D three-dimensional
  • a method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of remote clients comprises: (a) acquiring, using a camera apparatus comprising at least three imaging devices, a corresponding number of simultaneous live video images of a scene; (b) transmitting the simultaneous live video images from the camera apparatus to one or more remote servers; (c) creating, from the simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image using respective stitching and stereo image synthesizing modules, wherein (i) at least one of the simultaneous live video images is used in the creating of both the panoramic video image and the stereoscopic video image, (ii) at least one of the simultaneous live video images is used in the creating of the panoramic video image and not of the stereoscopic video image, and (iii) at least one of the simultaneous live video images is used in the creating of the stereoscopic image and not of the panoramic video image; and (d) simultaneously delivering, from the one or
  • an operation of the camera apparatus can be controlled from the one or more remote servers.
  • At least the simultaneous live video images used in creating the 3D stereoscopic video image can have an overlap therebetween.
  • the delivering can include delivering a single one of the live 2D panoramic video stream and the live 3D stereoscopic video stream to client viewing devices that are included in both the first and second pluralities of client viewing devices.
  • the inclusion of any given client viewing device in either or both of the first and second pluralities of client viewing devices can be based upon a capability of the client viewing device.
  • the inclusion of any given client viewing device in either or both of the first or second plurality of client viewing devices can be based upon a user selection.
  • a user selection can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
  • the user selection can be received before the delivering of the live video streams.
  • the user selection can be received during the delivering of the live video streams.
  • the user selection can be received via a client-side interface.
  • the user selection can be toggleable during the delivering of the live video streams.
  • the transmitting or receiving, and the creating and delivering can be performed in real time.
  • a method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients comprises: (a) receiving, by one or more servers and from a remote source, a plurality of simultaneous live video streams of a scene; (b) at the remote server(s), creating, from the plurality of received simultaneous live video streams, a live 2D panoramic video image and a live 3D stereoscopic video image of the scene using respective stitching and stereo image synthesizing modules; and (c) simultaneously streaming, from the one or more remote servers to client viewing devices in remote communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams, the simultaneous streaming being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the respective live video streams are synchronized.
  • each live video stream can be associated with a different respective view direction, the view directions and angles of view of the video streams providing overlap therebetween.
  • the 3D stereoscopic video is unstitched so that a left video stream thereof is from a first of the live video streams and a right video stream thereof is from a second of the live video streams, and (ii) the 2D panoramic image comprises a stitching between the first and second live video streams.
  • a method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of remote clients comprises: (a) acquiring, using a camera apparatus comprising a plurality of imaging devices, a corresponding plurality of simultaneous live video images of a scene; (b) transmitting the plurality of simultaneous live video images from the camera apparatus to one or more remote servers; (c) creating, from the plurality of simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules; and (d) simultaneously delivering, from the one or more remote servers to client viewing devices in communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the respective live
  • an operation of the camera apparatus can be controlled from the one or more remote servers.
  • the inclusion of any given client viewing device in either or both of the first and second pluralities of client viewing devices can be based upon a capability of the client viewing device. In some embodiments, the inclusion of any given client viewing device in either or both of the first or second plurality of client viewing devices can be based upon a user selection.
  • the delivering can include delivering a single one of the live panoramic video stream and the live 3D stereoscopic video stream to client viewing devices that are included in both the first and second pluralities of client viewing devices.
  • a user selection can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
  • a user selection can be received before the delivering of the live video streams. In some embodiments, a user selection can be received during the delivering of the live video streams. In some embodiments, a user selection can be received via a client-side interface. In some embodiments, the user selection can be toggleable during the delivering of the live video streams.
  • a location of a client viewing device can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
  • the content of the video images can be used in determining which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
  • a technical characteristic of one or both of the live video streams can be used in determining which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
  • the live panoramic video image and the live 3D stereoscopic video image are (i) stored in a non-transitory computer-readable storage medium at the one or more remote servers and (ii) made available for synchronized streaming on demand after the delivering of the live video streams has commenced. In some embodiments, it can be that the live panoramic video image and the live 3D stereoscopic video image are (i) stored in a non-transitory computer- readable storage medium at the one or more remote servers and (ii) made available for synchronized streaming on demand after the delivering of the live video streams has ended.
  • At least one live video image of the plurality of simultaneous live video images can be dewarped by the camera apparatus before they are transmitted to the one or more remote servers. In some embodiments, at least one live video image of the plurality of simultaneous live video images can be dewarped at the one or more remote servers.
  • the plurality of imaging devices has a corresponding plurality of respective optical axes, each of which can be angularly displaced from all of the others. In some embodiments, the plurality of imaging devices has a corresponding plurality of respective optical axes, of which at least a first two can be substantially parallel to each other and at least one other one can be parallel to but in an opposite direction to the at least first two.
  • the creating of the live panoramic video image by the stitching module can include asymmetrically cropping at least one of the simultaneous live video images.
  • the creating of at least one of the live panoramic video image and the live 3D stereoscopic image can include at least one of rotation and translation of one of the simultaneous live video images.
  • a system for delivering synchronized live video streams of panoramic and 3D stereoscopic images of a scene comprising: (a) respective stitching and stereo image synthesizing modules configured to create, from a plurality of live video images of a scene, a live panoramic video image and a live 3D stereoscopic video image; and (b) a delivery module configured to deliver, from one or more remote servers, the live panoramic video image and the live 3D stereoscopic video image as synchronized live video streams to a plurality of client viewing devices in communication with the one or more remote servers.
  • the system can be configured to receive the plurality of live video images of a scene from a camera apparatus comprising a plurality of imaging devices respectively configured to simultaneously acquire the plurality of live video images of the scene.
  • the system can be additionally configured to control the camera apparatus remotely from the one or more remote servers.
  • At least one of the stitching module and the stereo image synthesizing module can be located at the one or more remote servers.
  • the system can be configured to store the live panoramic video image and the live 3D stereoscopic video image in a non-transitory computer- readable storage medium at the one or more remote servers, so as to make said live panoramic video image and said live 3D stereoscopic video image available for synchronized streaming on demand at a later time.
  • the system can additionally comprise a dewarping module residing at the one or more remote servers for dewarping each of the plurality of live images.
  • the stitching module can be additionally configured to perform asymmetric cropping of at least one of the live video images.
  • At least one of said stitching module and said stereo image synthesizing module can be additionally configured to perform at least one of rotation and translation of one of the live video images.
  • the system can be configured to deliver said synchronized live video streams of panoramic and 3D stereoscopic images in real time.
  • the system can additionally comprise a toggling module configured to enable client toggling between a 3D stereoscopic display mode and a panoramic display mode.
  • a method for delivering synchronized live video streams of 2D panoramic and 3D stereoscopic images of a scene from one or more remote servers to a plurality of client devices.
  • the method comprises: (a) stitching a live 2D panoramic video image from a plurality of live video images of a scene received from a remote source; (b) synthesizing a live 3D stereoscopic video image from at least some of the plurality of live video images of the scene; and (c) delivering the live 2D panoramic video image and the live 3D stereoscopic video image as synchronized live video streams to the plurality of client viewing devices.
  • a method for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients.
  • the method comprises: (a) accessing a 2D panoramic video image and a 3D stereo video image, respectively stitched and synthesized from a plurality of simultaneously-captured video images of a scene, said video images of the scene having a first duration; (b) selecting from within the stitched 2D panoramic video image a first plurality of video segments for displaying in 2D panoramic mode and from within the synthesized 3D stereo video image a second plurality of video segments for displaying in 3D stereoscopic mode, wherein the selecting of each respective video segment includes analyzing at least one of its content and its composition, each one of the video segments corresponding to a different portion of the first duration; (c) creating a curated video stream that comprises the first and second plurality of video images; and (d) delivering the curated video stream, from the one or more servers to a plurality of remote client
  • a method for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients.
  • the method comprises: (a) accessing a plurality of simultaneously-captured, video images of a scene, said video images of the scene being having a first duration; (b) selecting a plurality of video segments from within said video images of the scene, each one of the video segments corresponding to a different portion of the first duration; (c) for each selected video segment, performing one video creation action selected from (i) stitching a 2D panoramic video segment and (ii) synthesizing a 3D stereo video segment, wherein the determination of which video creation action is performed for each respective selected video segment is based on an analysis of at least one of the segment’s content and the segment’s composition; (d) creating a curated video stream that comprises the stitched 2D panoramic video segments and the synthesized 3D stereo images; and (e) delivering the curated video stream, from the one or more servers to
  • the plurality of video images of the scene can be captured by imagers having respective optical axes angularly-displaced and/or laterally displaced from each other.
  • the 3D stereo images can be 360° 3D stereo images and the synthesizing includes stitching.
  • a method for delivering an enhanced video stream.
  • the method comprises: (a) acquiring a plurality of video images of a scene, said video images being simultaneously captured by imagers having angularly- displaced and/or laterally displaced optical axes; and (b) delivering, to a plurality of remote client devices, a live video stream comprising multiple time-differentiated segments, each segment comprising one of (i) a 2D panoramic video stream stitched from the plurality of video images captured during the period of time corresponding to the respective segment, and (ii) a 3D stereo video stream synthesized from at least some of the plurality of video images captured during the period of time corresponding to the respective segment.
  • the selection of which of the 2D panoramic video stream and the 3D stereo video stream is included in any given segment is based on the content and/or composition of the video images during the period of time corresponding to the respective segment, and (ii) the acquiring and delivering are done in real time so as to deliver a live video stream.
  • a method is disclosed according to embodiments for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients.
  • the method comprises: (a) accessing a plurality of simultaneously-captured, video images of a scene, said video images of the scene being having a first duration; (b) selecting a plurality of video segments from within the video images, each one of the video segments corresponding to a different portion of the first duration; (c) for each selected video segment, (i) stitching a corresponding 2D panoramic video segment and (ii) synthesizing a corresponding 3D stereo video segment; (d) creating a curated 2D panoramic video stream that comprises the stitched 2D panoramic video segments, and a curated 3D stereo video stream that comprises the synthesized 3D stereo video segments; and (e) delivering, from the one or more remote servers to client viewing devices in remote communication therewith, the curated 2D panoramic video stream and the curated 3D stereoscopic video stream, the delivering being such that: (i) the curated 2D panoramic video stream is delivered to a first plurality of client viewing devices and the curated 3D stereoscopic video stream is delivered to
  • the inclusion of any given client viewing device in either or both of the first or second plurality of client viewing devices can be based upon a user selection.
  • the delivering can include delivering a single one of the curated panoramic video stream and the curated 3D stereoscopic video stream to client viewing devices that are included in both the first and second pluralities of client viewing devices.
  • a user selection can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
  • the user selection can be received before the delivering of the curated video streams.
  • the user selection can be received during the delivering of the curated video streams.
  • the user selection can be toggleable during the delivering of the live video streams.
  • a method for delivering an enhanced video stream.
  • the method comprises: (a) acquiring a plurality of simultaneously- captured, video images of a scene; (b) initiating the delivery of a video stream from one or more servers to a plurality of remote client devices, such that the video stream delivered to each remote client is one of (i) a 2D panoramic video stream stitched from the plurality of video images and (ii) a 3D stereoscopic video stream synthesized from the plurality of video images; (c) determining a following-object in the video stream delivered to each remote client, the following-object including at least one of a person, thing or other visual element captured visually in at least one of the video images; and (d) delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
  • a method for delivering an enhanced video stream.
  • the method comprises: (a) acquiring a plurality of simultaneously- captured, video images of a scene, the video images including spatial audio captured by a plurality of microphones; (b) initiating the delivery of a video stream from one or more servers to a plurality of remote client devices, such that the video stream delivered to each remote client is one of (i) a 2D panoramic video stream stitched from the plurality of video images and comprising ambisonic audio processed from the captured spatial audio, and (ii) a 3D stereoscopic video stream synthesized from the plurality of video images and comprising ambisonic audio processed from the captured spatial audio; (c) determining a following-object in the video stream delivered to each remote client, the following object including a spatially-differentiated aural element captured in at least one of the video images; and (d) delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
  • the determining of the follow-object can include receiving a user input about the following-object at the one or more servers.
  • the determining of the follow-object can include analyzing at least one of (i) the overlapping video images, (ii) the 2D panoramic video stream and (iii) the 3D stereo video stream.
  • the determining of the follow-object can use stereoscopic parallax to identify an object.
  • the delivery of the alternative video stream can preempt the delivery of the video stream delivered to each remote client responsively to receipt of a user command or selection.
  • the video stream delivered to each remote client can be toggleably switchable with the alternative video stream.
  • the video stream delivered to each remote client can be delivered together with the alternative video stream such that the alternative video stream is displayable side-by-side or picture -in-picture together with video stream delivered to each remote client.
  • the video streams are live and synchronized with each other.
  • a method for delivering a reframed video stream.
  • the method comprises: (a) accessing a plurality of simultaneously- captured video images of a scene; (b) creating, from the plurality of simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules, the 2D panoramic video image and the 3D stereoscopic video image having respective first and second aspect ratios; and (c) delivering, from the one or more remote servers to client viewing devices in communication therewith, the live 2D panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) at least one of the 2D panoramic video stream and the 3D stereoscopic video stream is reframed for delivery in a third aspect ratio that
  • the respective live video streams can be delivered simultaneously and can be synchronized.
  • the accessing includes accessing at least three simultaneously-captured video images of the scene, at least one of which is used in both the panoramic stream and in the stereoscopic stream, at least one of which is used in the panoramic stream and not in the stereoscopic stream, and at least one of which is used in the stereoscopic stream and not in the panoramic stream.
  • a method for delivering simultaneous video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients.
  • the method comprises: (a) acquiring, by one or more servers, a plurality of video streams of a scene, the video streams having overlap therebetween; (b) at the one or more remote servers, creating, from the plurality of overlapping video streams of the scene, a 2D panoramic video stream, and a 3D stereoscopic video stream, using respective stitching and stereo image synthesizing modules; (c) initiating the synchronized delivery, from the one or more remote servers to a plurality of client devices, a live video stream comprising one of the live 2D panoramic video stream and the live 3D stereoscopic video stream; (e) receiving, at the one or more remote servers, a toggle command from a client; and (f) toggling the video stream delivered to the client so as to (i) replace the live 2D panoramic video stream with the live 3D stereo video stream, or (ii) to replace the live 3D
  • the 2D panoramic video stream has a 360° angle of view and the 3D stereo video stream has a 180° angle of view.
  • the 3D stereoscopic video is unstitched so that a left video stream thereof is from a first of the overlapping video streams and a right video stream thereof is from a second of the overlapping video streams, the 2D panoramic image comprising a stitching between the first and second live video streams.
  • the overlapping video images are live video images, and the creating and initiating delivery are performed in real time.
  • a camera apparatus for acquiring multiple simultaneous video images for creating therefrom 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces; and (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and a respective planar array of photodetectors defining a respective photodetector plane, each imaging device defining a respective optical axis orthogonal to said respective photodetector plane, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and the third imaging device is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel there
  • the angle of view of the second imaging device does not encompass any of the lens of the third camera.
  • the photodetector plane of the second imaging device can be not coplanar with, and displaced further outward from the second major face than, the photodetector plane of the third imaging device. In some embodiments, the photodetector plane of the second imaging device can be displaced further outward from the second major face than the photodetector plane of the third imaging device by at least 3 millimeters.
  • the imaging device apparatus additionally comprises a video-display module comprising a display screen installed on the first major face.
  • the first imaging device has an angle of view of at least 180°, and said angle of view does not encompass any of the display screen.
  • the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto. In some embodiments, the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto.
  • the camera apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices acquire the multiple simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the multiple simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the multiple simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via a display-control interface.
  • a second minor surface opposite the first minor surface can comprise a concave, textured finger-grip area extending across of a majority of the second minor surface in each of two orthogonal directions.
  • a camera apparatus for acquiring multiple simultaneous video images for creating 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces; (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and defining a respective optical axis, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, (ii) the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and (iii) ⁇ the third imaging device is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto; (c) a video-display module comprising a display screen installed on the first major face, the
  • a majority of the footprint of the mechanical support-stand interface on the first minor face falls within the borders of an orthographic projection of the first and second imaging devices on the first minor face.
  • the footprint of the mechanical support-stand interface on the first minor face can be center-aligned with a centerline of an orthographic projection of the first and second imaging devices on the first minor face on two orthogonal axes.
  • the apparatus can additionally comprise a first electronic status indicator on the first major face and a second electronic status indicator on the second major face.
  • the optical axis of the second imaging device can define a second direction that is opposite to the first direction or no more than 1 degree from being opposite thereto.
  • the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto.
  • the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto.
  • the apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices, acquire the simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via the display-control interface.
  • a second minor surface opposite the first minor surface can comprise a concave, textured finger-grip area extending across of a majority of the second minor surface in two orthogonal directions.
  • a camera apparatus for acquiring multiple simultaneous video images for creating 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces, and a bottom face characterized by the presence thereupon of a mechanical support- stand interface; (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and defining a respective optical axis, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, (ii) the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and (iii) the third imaging device is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto; and (c)
  • the first, second and third imaging devices are aligned with each other at the same height in the vertical direction within a tolerance of one -fifth of a diameter of any of the respective ultra wide-angle lenses.
  • the first, second and third imaging devices are aligned with each other at the same height in the vertical direction within a tolerance of one millimeter.
  • the first and second imaging devices can be aligned with each other in the x -dimension within a tolerance of one millimeter.
  • the mechanical support-stand interface can be aligned with the first and second imaging devices in the x-dimension within a tolerance of five millimeters.
  • the optical axis of the second imaging device can define a second direction that is opposite to the first direction or no more than 1 degree from being opposite thereto.
  • the optical axis of the third camera can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto.
  • the optical axis of the third camera can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto.
  • the apparatus can additionally comprise a first electronic status indicator on the first major face and a second electronic status indicator on the second major face.
  • the apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices, acquire the simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via the display-control interface.
  • a camera apparatus for acquiring multiple simultaneous video images for creating 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces; (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and a respective planar array of photodetectors defining a respective photodetector plane, each imaging device defining a respective optical axis orthogonal to said respective photodetector plane, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, (ii) the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and (iii) the third camera is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the
  • the photodetector plane of the second imaging device can be displaced further outward from the second major face than the photodetector plane of the third camera by at least 3 millimeters.
  • the second imaging device has an angle of view of at least 180°, and (ii) said angle of view does not encompass any of the lens of the third camera.
  • the second imaging device has an angle of view of at least 180°, and (ii) a 180° concentric portion of the angle of view does not encompass any of the lens of the third camera.
  • the apparatus can additionally comprise a video-display module comprising a display screen installed on the first major face.
  • the first imaging device has an angle of view of at least 180°, and (ii) said angle of view does not encompass any of the display screen.
  • the optical axis of the third camera can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto.
  • the optical axis of the third camera can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto.
  • the apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices, acquire the simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via a display-control interface.
  • a second minor surface opposite the first minor surface can comprise a concave, textured finger-grip area extending across of a majority of the second minor surface in each of two orthogonal directions.
  • Figs. 1A, IB and 1C show block diagrams of systems for delivering synchronized live video streams to a plurality of client viewing devices, according to embodiments of the present invention.
  • Fig. 2 shows block diagrams illustrating various examples of incorporating a dewarping module in a system for delivering synchronized live video streams, according to embodiments of the present invention.
  • Fig. 3 shows a block diagram of a system for delivering synchronized live video streams to first and second pluralities of client viewing devices, according to embodiments of the present invention.
  • Fig. 4 shows a block diagram of a system for delivering synchronized live video streams to first and second overlapping pluralities of client viewing devices, according to embodiments of the present invention.
  • Fig. 5A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention.
  • Fig. 5B shows a block diagram showing cloud-based sub modules of a stream determination module of a system for delivering synchronized live video streams, according to embodiments of the present invention.
  • Figs. 6A, 6B and 6C are schematic illustrations of a camera apparatus comprising multiple imaging devices capturing simultaneous overlapping video images of a 180° scene, according to embodiments of the present invention.
  • Figs. 6D, 6E and 6F are schematic illustrations of a camera apparatus comprising multiple imaging devices capturing simultaneous video images of a 360° scene, according to embodiments of the present invention.
  • Figs. 6G, 6H and 61 are respective top, front and rear views of a camera apparatus embodying the functionality of the camera apparatus of Fig. 6D.
  • Fig. 7A is a schematic illustration of a client viewing device comprising a screen offering a selection between a 3D stereoscopic video stream and a panoramic video stream, according to embodiments of the present invention.
  • Fig. 7B is a schematic illustration of a client viewing device comprising a screen, showing side-to-side panning of a panoramic image by a user.
  • Fig. 7C is a schematic illustration of a client viewing device comprising a virtual reality headset and controller offering a selection between a 3D stereoscopic video stream and a panoramic video stream, according to embodiments of the present invention.
  • Figs. 7D and 7E are schematic illustrations of user wearing a client viewing device comprising a virtual reality headset offering a toggle option for toggling between a 3D stereoscopic video stream and a panoramic video stream, according to embodiments of the present invention.
  • Figs. 8 and 9 show flow charts of methods for delivering simultaneous live video streams of panoramic and 3D stereoscopic images, according to embodiments of the present invention.
  • Fig. 10 shows a block diagram of a system for delivering synchronized live video streams to a plurality of client viewing devices, including a toggling module, according to embodiments of the present invention.
  • Fig. 11 shows a block diagram of a system for delivering synchronized live video streams to a plurality of client viewing devices, according to embodiments of the present invention.
  • Fig. 12A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention.
  • Fig. 12B shows a block diagram showing cloud-based sub modules of an object following module of a system for delivering synchronized live video streams, according to embodiments of the present invention.
  • Fig. 13A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention.
  • Fig. 13B shows a block diagram showing cloud-based sub modules of a curating module of a system for delivering synchronized live video streams, according to embodiments of the present invention according to embodiments of the present invention.
  • Fig. 14A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention.
  • Fig. 14B shows a block diagram showing cloud-based sub modules of a reframing module of a system for delivering synchronized live video streams, according to embodiments of the present invention according to embodiments of the present invention.
  • Fig. 15 shows a flow chart of a method for delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients according to embodiments of the present invention.
  • Fig. 16 shows a flow chart of a method for delivering synchronized live video streams of 2D panoramic and 3D stereoscopic images of a scene from one or more remote servers to a plurality of client devices according to embodiments of the present invention.
  • Figs. 17 and 18 show flow charts of methods of delivering curated video strea s including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients according to embodiments of the present invention.
  • Fig. 19 shows a flow chart of a method for delivering an enhanced video stream according to embodiments of the present invention.
  • Fig. 20 shows a flow chart of a method of delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients according to embodiments of the present invention.
  • Fig. 21 shows a flow chart of a method for delivering an enhanced video stream according to embodiments of the present invention.
  • Fig. 22 shows a flow chart of a method for delivering an enhanced video stream according to embodiments of the present invention.
  • Fig. 23 shows a flow chart of a method of delivering simultaneous video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients according to embodiments of the present invention.
  • Figs. 24A, 24B, 24C and 24D are, respectively, front, rear, bottom and top views of a camera apparatus comprising three imaging devices, according to embodiments of the present invention.
  • Fig. 25 is a bottom view of a camera apparatus comprising three imaging devices, showing the borders of an orthographic projection of two of the imaging devices on the bottom face of the apparatus, according to embodiments of the present invention.
  • Fig. 26 is a bottom view of a camera apparatus comprising three imaging devices, schematically showing the respective photodetector planes of the three imaging devices, according to embodiments of the present invention.
  • Fig. 27 is a bottom view of a camera apparatus comprising three imaging devices, schematically showing the respective angle of view of two of the imaging devices with respect to physical features on the two major surfaces of the apparatus, according to embodiments of the present invention.
  • subscripted reference numbers may be used to designate multiple separate appearances of elements of a single species, whether in a drawing or not; for example: 10i is a single appearance (out of a plurality of appearances) of element 10.
  • 10i is a single appearance (out of a plurality of appearances) of element 10.
  • the same elements can alternatively be referred to without subscript (e.g., 10 and not 10i) when not referring to a specific one of the multiple separate appearances, i.e., to the species in general.
  • module means any combination of electronic hardware, software and firmware necessary to perform the function associated with the module, e.g., a ‘stitching module’ includes hardware, software and firmware for ‘stitching’ images.
  • a ‘stitching module’ includes hardware, software and firmware for ‘stitching’ images.
  • imaging device ‘sensor’ and‘camera’ may be used interchangeably in this specification and in the appended claims, as well as in documents incorporated by reference, and all have the same meaning: a device for capturing digital images, including (at least) a CMOS or CCD device and a lens, and optionally any number of mechanical and electronic components and accessories.
  • the terms‘camera apparatus’ and‘imaging apparatus’ as used herein and in documents incorporated by reference have the same meaning as each other, are directed to an apparatus comprising multiple sensors/cameras/imaging devices.
  • an imaging device, sensor or camera can be a component of a camera apparatus or imaging apparatus, but not vice versa.
  • a camera apparatus comprising multiple imaging devices can acquire simultaneous video images of a scene.
  • Each of the imaging devices can have a wide angle of view, for example 180° or more.
  • a scene can encompass a field of view of 180° or more, or 360° or more.
  • the simultaneous video images can be overlapping, meaning that at least two of the video images overlap, or that there are multiple overlaps involving respective multiple pairs or groups of overlapping video images, e.g., at a minimum for any three images, image A overlaps with image B, and image B overlaps with image C. In some embodiments for three given images each image overlaps with the other two.
  • a camera apparatus has the potential to capture at least some overlapping images, but in actual operation only non-overlapping images are captured; in some of these cases the non-overlapping images are abutting one another, and in other cases they are nearly abutting but not quite.
  • the angles of view captured by one or more of the respective imaging devices is exactly 180°, or alternatively the angles of view of all of the imaging devices add up to exactly 360°, and in such cases the concept of‘overlapping’ as used herein is expanded to include adjacent or abutting images.
  • angles of view captured by one or more of the respective imaging devices is less than 180°, or alternatively the angles of view of at least some of the respective imaging devices do not overlap, and in such cases the concept of ‘stitching’ as used herein is expanded to include interpolating such that images can be considered‘stitched’ even though there is no actual overlap between them.
  • a stitching module can create a panoramic image from two or more of the overlapping (or abutting or nearly abutting) video images.
  • a 3D stereoscopic synthesizing module can create a 3D stereoscopic image from two or more of the overlapping video images, which can include at least in part the same images as the video images used in stitching the panoramic image.
  • Panoramic stitching and 3D stereoscopic synthesizing of multiple and overlapping images can be performed using any of the systems and methods described in pending international application PCT/IB2018/059771 filed on December 7, 2018, the contents of which are incorporated herein by reference in their entirety.
  • the aforementioned methods and systems are also disclosed in the following US patent applications: Ser. No. 62/596,112 filed on December 7, 2018, Ser. No. 62/681,667 filed on June 6, 2018, Ser. No. 62/700,095 filed on July 18, 2018, Ser. No. 62/700,104 filed on July 18, 2018, Ser. No. 62/700,571 filed on July 18, 2018, Ser. No. 62/700,580 filed on July 18, 2018, and Ser. No. 62/700,588 filed on July 18, 2018, the contents of all of which are incorporated herein by reference in their entirety.
  • either or both of the stitching module and the 3D stereoscopic synthesizing module are included in the camera apparatus, and in some embodiments, either or both of the stitching module and the 3D stereoscopic synthesizing module reside remotely, for example at one or more remote servers, which can be part of a server farm comprising one or more servers dedicated to the functions of stitching module and/or the 3D stereoscopic synthesizing module (and/or other modules which may be necessary for processing the video images, storing them, and/or delivering them to clients as video streams, where delivering them to clients can include selecting which of the video streams to deliver to any particular client).
  • the one or more servers can be‘in the cloud’, i.e., servers available from a commercial computing operation that includes any number of networked servers (including, without limitation, shared, dedicated and virtual servers) available to any number of unrelated customers.
  • remote servers are illustrated throughout as being‘in the cloud’ but - as described above - this is not necessarily the case in some embodiments.
  • the camera apparatus can include communications circuitry configured for uploading the overlapping video images to the one or more servers, and in other embodiments, the communications equipment is configured for uploading stitched panoramic images and synthesized 3D stereoscopic images to the one or more servers.
  • the phrase‘overlapping video images’ is used to indicate the video images captured by a camera apparatus before further processing such as dewarping or stitching or stereo-synthesizing, and the phrase will, in some cases, refer to video images that not actually overlapping but merely abutting one another or nearly abutting one another.
  • the one or more servers can include a delivery module for delivering the panoramic and 3D stereoscopic images (regardless of whether they were respectively stitched and synthesized at the one or more servers or at the camera apparatus) to client viewing devices for display thereupon.
  • FIG. 1A a block diagram shows an example of a system 100 for delivering synchronized live panoramic and stereoscopic video streams.
  • the system 100 of Fig. 1A includes a camera apparatus 150 that includes N imaging devices 155i .. 155 N , where N is an integer greater than 1.
  • the camera apparatus 150 is preferably operable to acquire, using the imaging devices 155i .. 155 N , N respective overlapping live video images llli .. 111 N .
  • the camera apparatus 150 also includes communications circuitry configured, inter alia, for uploading the overlapping video images llli .. 111 N to a remote server 170 using communications channel 201 which, like all other communications channels disclosed herein, can incorporate any combination of wired and wireless communications technologies and is preferably configured to transmit video images at the resolution enabled by the imaging devices 155i .. 155 N and at the throughput speed necessary for maintaining a substantially live and real-time transmission.
  • server is synonymous with the terms‘one or more servers,’‘one or more remote servers,’ and‘remote server,’ and we note that the specific distribution of loads and functions between multiple servers or multiple processors is not relevant to the present disclosure.
  • the system 100 additionally comprises a number of modules residing on one or more remote servers 170 in the cloud 199 - a stitching module 175 for stitching together overlapping live video images llli .. lll N that are transmitted to the server 170 by the camera apparatus 100, and a stereo image synthesizing module 176 for synthesizing 3D stereo images from the overlapping live video images llli .. 111 N -
  • storage medium 140 for example, a non-transitory, computer-readable storage medium
  • storage medium 140 can be resident at the cloud 199.
  • the stitching module 175 can be operative to transmit, i.e., download, a panoramic video image stitched from the overlapping live video images llli 11 I N to a plurality 195 of client viewing devices 190 as a live panoramic video stream 180 using communications channel 202.
  • the plurality 195 can include any number M of client viewing devices 190i .. 190 M in accordance with the operating capabilities of the server 170 and the communications channel 202 or 203.
  • the stereo image synthesizing module 176 can be operative to transmit, i.e., download, a 3D stereoscopic video image synthesized from the overlapping live video images llli ..
  • Fig. 1 A to the one or more remote servers 170, using the respective stitching module 175 and stereo image synthesizing module 176 so as to create therefrom respective panoramic and stereoscopic images and deliver said images as live panoramic video stream 180 and live 3D stereoscopic video stream 181 to the plurality 195 of client viewing devices 190 - is preferably performed without undue delays in processing time, such that the live streams 180, 181 are delivered in real time, i.e., within the amount of delay time acceptable in the industry for‘live’ and ‘real-time’ transmissions. It should be obvious to the skilled artisan that not all the components shown in Fig. 1 A are necessary for the proper operation of the system 100, nor does Fig. 1A illustrate every possible component and interface for such a system 100.
  • FIG. IB a second example of a system 100 for delivering synchronized live video streams is shown in a block diagram.
  • the system 100 of Fig. IB has similarities to the system 100 previously discussed with respect to Fig. 1A, and a number of differences:
  • Stitching module 175 and stereo image synthesizing module 176 in the example shown in Fig. IB, reside within the camera apparatus 150 and not in the cloud at the one or more servers 170 as was the case in the example of Fig. 1A;
  • the communications circuitry 157 of the camera apparatus 150 is configured to upload processed (pre-stitched/pre-synthesized) video images as live streams, i.e., a‘stitched’ live panoramic video stream 180 and a ‘synthesized’ 3D stereoscopic video stream 181, to the one or more servers
  • a delivery module 178 at the one or more servers 170 is configured to deliver the respective live video streams 180, 181 to the plurality 195 of client viewing devices 190.
  • the delivery module 178 can be a component of system 100.
  • Figs. 1A and IB illustrate every possible component and interface for such a system 100. It should also be obvious that certain aspects of the two figures can be combined - as a non-limiting example, it can be that a specific one of the respective stitching and stereo image synthesizing modules 175, 176 may be included in the camera apparatus 150 and the other one of the two modules resides in the cloud 199, i.e., at the one more remote servers 170.
  • FIG. 1C illustrates another example of a system 100 for delivering synchronized live panoramic and stereoscopic video streams.
  • system 100 includes key processing and client-interface components.
  • This is an example of a system 100 that does not necessarily include the camera apparatus 150 for acquiring the video images.
  • the video images to be respectively stitched and synthesized into panoramic and 3D stereoscopic streams 180, 181 can come to the server 170 from any source, or can, for example, be stored on the storage medium 140 residing at the server 170.
  • system components reside at servers 170 in the cloud 199, including stitching module 175 for stitching together overlapping video images, and a stereo image synthesizing module 176 for synthesizing 3D stereo images.
  • Storage medium 140 (for example, a non-transitory, computer-readable storage medium) can be resident at the cloud 199 in conjunction with and preferably in electronic communication with the server 170.
  • the stitching module 175 can be operative to transmit, i.e., download, a panoramic video image stitched from, e.g. , overlapping live or stored video images to a plurality 195 of client viewing devices 190 as a panoramic video stream 180 using communications channel 202.
  • the plurality 195 can include any number M of client viewing devices 190i .. 190 M in accordance with the operating capabilities of the server 170 and the communications channel 202 or 203.
  • the stereo image synthesizing module 176 can be operative to transmit, i.e., download, a 3D stereoscopic video image synthesized from, e.g., overlapping live or stored video images to the same plurality 195 of client viewing devices 190 as a live 3D stereoscopic video stream 181, using communications channel 203.
  • a 3D stereoscopic video image synthesized from, e.g., overlapping live or stored video images to the same plurality 195 of client viewing devices 190 as a live 3D stereoscopic video stream 181, using communications channel 203.
  • delivery module 178 the delivery of panoramic video stream 180 and 3D stereoscopic video stream 181 is handled by delivery module 178.
  • a dewarping module can be included as part of a system 100 for delivering synchronized live panoramic and stereoscopic video streams 180, 181, or otherwise can be available for use by components of the system.
  • a dewarping module 177 can reside in any one of various locations according to the specific embodiment. Fig. 2 shows a non-limiting set of examples for the location of dewarping modules 177. In Example A of Fig.
  • a dewarping module 177 is included in camera apparatus 150 which also includes stitching module 175 and stereo image synthesizing module 176.
  • the dewarping module 177 can be configured to pre-process the acquired live overlapping video images (e.g., llli .. 111 N of Fig. 1A) for the common use of the stitching module 175 and the stereo image synthesizing module 176.
  • Example B of Fig. 2 shows an embodiment in which the dewarping module 177 is included in camera apparatus 150 but stitching module 175 and stereo image synthesizing module 176 are not included in the camera apparatus 150.
  • the communications circuitry 157 of the camera apparatus 150 are preferably configured to upload dewarped video images and not the raw overlapping live video images llli .. 111N as discussed with respect to Fig. 1 A.
  • Example C of Fig. 2 shows another embodiment, in which the dewarping module 177 resides in the cloud, e.g., at remote server 170, along with stitching module 175 and stereo image synthesizing module 176.
  • Example D of Fig. 2 shows yet another embodiment in which each of the stitching module 175 and stereo image synthesizing module 176 include a dewarping module 177.
  • Example D can be suitable to embodiments in which one of the stitching module 175 and stereo image synthesizing module 176 is located in a camera apparatus 150 and the other of the two modules resides at the remote servers 170, although the implementation of Example D is not limited to such embodiments.
  • Fig 3 illustrates yet another embodiment of a system 100 for delivering synchronized live panoramic and stereoscopic video streams 180, 181.
  • a comparison with Fig. 1A will reveal that the difference between the block diagrams of Fig. 3 and Fig. 1A is that in Fig. 1A, the two live video streams 180, 181 are simply shown as being delivered to a plurality 195 of client viewing devices 190 (i.e., the M devices 190i .. 190M), while in Fig. 3 the following distinction is made: the live panoramic video image 180 is streamed (i.e., delivered) to a first plurality 191 of client viewing devices 190 (specifically the X devices 190i ..
  • a first overlapping plurality 193 of client viewing devices 190 can overlap with a second overlapping plurality 194 of client viewing devices 190, wherein the first overlapping plurality 193 includes the client viewing devices 190i .. 190x and the second overlapping plurality includes the client viewing devices 190x .. 190M- The client viewing devices 190i ..
  • the client viewing devices 190x +i .. 190M of the second plurality 194 receive‘only’ the live 3D stereoscopic video image stream 181; client viewing device 190x is included in both the first and second overlapping pluralities 193, 194 and receives both streams 180, 181.
  • both streams are delivered to the devices 190 in the overlap of the pluralities 193, 194, only one video stream can viewed/displayed at a time, and a selection or determination of which stream is viewed/displayed will be discussed below.
  • the overlap can include more than one client viewing device 190 that receives both live video streams 180, 181.
  • An embodiment in which the overlap includes ‘all’ of the client viewing devices, i.e., in which all of the client viewing devices 190i ... 190 M simultaneously receive both live video streams 180, 181 has already been illustrated in Fig. 1A.
  • a single client viewing device 190 can simultaneously receive, i.e., have streamed or delivered to it, two synchronized live video streams 180, 181. Any selection or switching/toggling between the two streams may be handled by or at the client viewing device 190.
  • This can be desirable, for example, in implementations wherein it is desired to have a switching or selecting between the two streams 180, 181 be as instantaneous as possible for the receiver of both streams, and communications protocols or other communications ‘overhead’ makes it less desirable to switch between streams‘at the source’, i.e., at the one or more servers 170.
  • only one stream is delivered to each client viewing device, and the determination or selection of which stream goes to which client viewing device is not made, at least not entirely, by or at the client viewing device 190.
  • a selection or determination is made by or on behalf of the system as to which one of two live synchronized video streams (panoramic or 3D stereo) is streamed (‘streamed’ and‘delivered’ are used interchangeably in this disclosure) to any given client viewing device.
  • a selection is made by a user as to which one of the two live synchronized video streams is streamed.
  • both streams can be delivered to a single client viewing device, and a selection is made, generally by a user, as to which of the two streams is displayed by the client viewing device.
  • a stream determination module 215 is shown residing at the one or more remote servers 170 in the cloud 199.
  • the dashed lines of the stitching module 175 and the stereo image synthesizing module 176 inform us that either or both of these two modules can reside at the one or more remote servers 170 (as in the example of Fig. 1A) or at the camera apparatus 150 (as in the example of Fig. IB).
  • the stream determination module 215 can be operative to make a determination or selection, or respond to a user selection, of which live video stream (panoramic or 3D stereo) is delivered to any given client viewing device 190 at any given time (although in some embodiments the determination or selection can be permanent or ‘until further notice’). Any of the sub-modules of the stream determination module 215 can determine in which plurality 191, 192 of Fig. 3 or 193, 194 or Fig. 4) a given client viewing device 190 is included.
  • the stream determination module 215 can include a user-selection based stream determination sub-module 216.
  • a user selection can determine which live video stream is delivered to a given client viewing device 190.
  • a user selection can determine which of the two streams simultaneously delivered to a given client viewing device 190 is actually displayed by the client viewing device 190.
  • a user selection can be received via a client- side interface, e.g., on the client viewing device itself, as will be discussed later in connection with Figs. 7A and 7C.
  • the user selection can be received before the initiation of a streaming event that includes delivering live video streams, or a user selection may be received during the delivering of the live video streams.
  • the selection can be toggled (switched back and forth) during the delivering of the live video streams, as will be discussed in connection with Figs. 7D, 7E and 10.
  • the stream determination module 215 can additionally or alternatively include a client-location based stream determination submodule 217.
  • the determination of which live stream is delivered to a given client viewing device is made on the basis of the client location at any resolution - by country, by region, or even by where in a building or room the client is located.
  • the selected live stream can be switched from panoramic to 3D stereo or the reverse if and when the client location changes.
  • the stream determination module 215 can additionally or alternatively include a client-history based stream determination submodule 218.
  • a client-history based stream determination submodule 218 can make a determination based on client viewing history - for example, if a client always chooses the 3D stereo stream, it can be desirable (e.g., for purposes of better customer service or reduced communications overhead) to simply stream according to the client’s historical choice.
  • a client can optionally be offered the option to switch streams.
  • the stream determination module 215 can additionally or alternatively include a client viewing device capability-based stream determination submodule 218.
  • it can be desirable to check, e.g., through a communications channel such as communications channels 202 or 203 what the technical capabilities (panoramic vs. stereo) of a client viewing device 190 are, i.e., by receiving the specification upon electronic request or looking up a model specification in a database, and to pre determine the choice of live streams accordingly.
  • the stream determination module 215 can additionally alternatively be operable to make a determination as to which live stream to deliver on the basis of a technical characteristic of the content of the video images (for example, but not exhaustively, the aspect ratio or field of view of the video images, the resolution of the video images, and the calibration status of the imaging devices used to acquire the video images).
  • the basis for the determination can be the nature of the content of the video images and which format is more appropriate, e.g., provides the best user experience.
  • FIGs. 5 A and 5B are not intended to show an exhaustive inventory of the components of a system for creating and delivering panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system.
  • the skilled artisan will understand that the cloud-residing modules and components can be combined with any one or more of the components, features and communications arrangements of any of the systems 100 illustrated in figures accompanying this disclosure.
  • FIG. 6A-6F a number of examples of camera apparatuses 150 for simultaneous acquisition of overlapping video images are schematically illustrated.
  • camera apparatus 150 includes two imaging devices 155i, 155 2 with respective optical axes indicated by lines 401i, 401 2 -
  • the camera apparatus 150 of Fig. 6A is operative to simultaneously acquire overlapping video images of scene 420 which has a combined field of view of more than 180° using the two imaging devices 155i, 155 2 .
  • camera apparatus 150 includes three imaging devices 155i, 155 2 , 1553, with respective optical axes indicated by lines 401r, 4012 , 4013.
  • the camera apparatus 150 of Fig. 6B is operative to simultaneously acquire overlapping video images of scene 420 which has a field of view of more than 180°.
  • video images acquired by imaging devices 155i and 155 3 are used for synthesizing a 3D stereo video image for streaming, and images acquired by all three imaging devices 155i, 155 2 , 155 3 are used for stitching a panoramic video image (and in other embodiments images from the same two image devices 155i and 155 3 are used for creating both video streams).
  • camera apparatus 150 includes two imaging devices 155i, 155 2 with respective optical axes indicated by lines 401i, 401 2 -
  • the camera apparatus 150 of Fig. 6C is operative to simultaneously acquire overlapping video images of scene 420 which has a combined field of view of more than 180° using the two imaging devices 155i, 155 2 .
  • camera apparatus 150 of Fig. 6C includes two additional imaging devices 156 which are not used for the acquisition of images for creating synchronized panoramic and 3D stereo images.
  • camera apparatus 150 includes two imaging devices 155i, 155 3 on one side of the camera apparatus 150, with respective optical axes indicated by lines 401r, 4013.
  • the camera apparatus includes a third imaging device 1552 on the opposite side of the camera, with respective optical axis indicated by line 401 2.
  • the camera apparatus 150 of Fig. 6D is operative to simultaneously acquire overlapping video images of surrounding scene 421-422 which has a combined field of view of 360°.
  • video images acquired by imaging devices 155r and 155 3 are used for synthesizing a 3D stereo video image for streaming, and images acquired by all three imaging devices 155i, 155 2 , 155 3 are used for stitching a 360° panoramic video image (and in other embodiments images from the same two image devices 155i and 155 3 are used for creating both the 3D stereo stream and a 180° panoramic video stream).
  • optical axes 401i, 401 3 are substantially parallel to each other (substantially meaning within plus/minus 1 ° or within plus/minus 5°) and optical axis 401 2 is in the opposing direction.
  • camera apparatus 150 includes two imaging devices 155i, 155 3 on one side of the camera apparatus 150, with respective optical axes indicated by lines 401i, 401 3 angularly displaced from each other.
  • the camera apparatus includes a third imaging device 155 2 on the opposite side of the camera, with respective optical axis indicated by line 401 2 , also angularly displaced from the other two optical axes indicated by lines 401i, 401 3.
  • the camera apparatus 150 of Fig. 6E is operative to simultaneously acquire overlapping video images of surrounding scene 421-422 which has a combined field of view of 360°.
  • video images acquired by imaging devices 155i and 155 3 are used for synthesizing a 3D stereo video image for streaming, and images acquired by all three imaging devices 155i, 155 2 , 155 3 are used for stitching a 360° panoramic video image (and in other embodiments images from the same two image devices 155i and 155 3 are used for creating both the 3D stereo stream and a 180° panoramic video stream).
  • camera apparatus 150 includes four imaging devices 155i, 155 2 , 155 3 , 155 4 with respective optical axes indicated by lines 401i, 401 2 , 401 3 , 401 4 all of which are angularly displaced from each other.
  • the camera apparatus 150 of Fig. 6F is operative to simultaneously acquire overlapping video images of surrounding scene 421-422 which has a combined field of view of 360°.
  • video images acquired by all four imaging devices 155i, 155 2 , 155 3 , 155 4 are used both for stitching a 360° panoramic video image and synthesizing a 360° 3D stereoscopic video image.
  • images acquired by imaging devices 155i, 155 2 are used for synthesizing a 180° 3D stereoscopic video image while all four imaging devices 155i, 155 2 , 155 3 , 155 4 are used for stitching a 360° panoramic video image.
  • Figs. 6G, 6H and 61 are respective top, front and rear views of a camera apparatus 150 embodying the functionality of the camera apparatus of Fig. 6D.
  • a camera apparatus 150 can include a viewscreen 154.
  • stereophonic and/or panoramic video images can be created using images acquired by any number of imaging devices. For example, at a professional sports event, dozens of imaging devices may be spread around a stadium so as to provide high-resolution imaging from many different angles.
  • a number of examples of client viewing devices 190 and indications of their use for viewing live synchronized panoramic and/or 3D stereoscopic video image streams are schematically illustrated.
  • client viewing device 190 comprises a conventional computer display monitor adapted to receive video streams in data communication with the one or more servers 170, e.g., through wired and/or wireless internet-based connections 202, 203.
  • a 180° scene 420 is displayed on the screen 190.
  • An onscreen widget 95 as a non-limiting illustrative example of a client-side control element, allows a user to select between a 3D stereoscopic image stream and a 180° panoramic image stream.
  • selecting one of the two live streams can be to determine which live stream (of panoramic and stereoscopic) is delivered to the client viewing device 190 and in other embodiments the selecting can be to determine which of the two delivered streams is to be displayed on the client viewing device 190.
  • selecting a live stream can be through a portal.
  • a portal can be a public portal such as, for example, YouTube, where each stream is accessible through a link to its own respective internet address.
  • a portal can be accessed through a single dedicated link to an internet address, at which address the user is allowed to select one of the live streams and/or toggle between them.
  • a single portal can be accessed as a single dedicated link to an internet address, at which the user is automatically directed to one stream or the other, based on criteria discussed earlier such as client location or viewing history, or characteristics of the video content or of the client viewing device.
  • a user can choose to pay for one video stream or the other, or for both interchangeably.
  • a surrounding scene 421-422 is viewed as a 360° panoramic stream on a conventional touch screen monitor, where a user can pan back and forth with her finger 90, as indicated by finger movement lines 91, to view different sections of the surrounding scene 421-422.
  • the panning can be done by using a pointing device such as a mouse (not shown). While it might be more‘ideal’ to watch such images on a‘virtual reality’ device, the specific panning capability illustrated in Fig. 7B can be used on conventional computer monitor screens.
  • a client viewing device 190 comprises a virtual reality (VR) headset with a handheld controller 194 (which can alternatively comprise a smartphone or tablet running an appropriate app).
  • VR virtual reality
  • An onscreen (i.e., on the controller screen) widget 95 allows a user to select between a 3D stereoscopic image stream and a panoramic image stream.
  • Figs. 7D and 7E illustrate a client viewing device 190 comprising a VR headset worn by user 90.
  • the user 90 can pan or navigate the 360° viewing angle of a live video stream of the surrounding scene 421-422.
  • the user 90 looks in a first direction (indicated by arrow 93) in Fig. 7D and sees scene 421, then looks in the opposing direction in Fig. 7E and sees scene 422.
  • Toggle widget 94 appears on the VR screen 195 to enable the user 90 to toggle between the synchronized live video streams.
  • a flowchart of a method for delivering simultaneous live video streams of panoramic and 3D stereoscopic images to a plurality of clients.
  • the method includes four steps, as follows:
  • Step SOI acquiring, using a camera apparatus 150 comprising a plurality of imaging devices 155, a corresponding plurality of simultaneous live video images 111 of a scene 420, or 421-422. Any of the camera apparatuses 150 disclosed in Figs. 6A- F would be suitable for carrying out the method.
  • Step S02 transmitting the plurality of live video images 111 from the camera apparatus 150 to one or more remote servers 170 using communications circuitry 157 of the camera apparatus 150 and communications channel 201.
  • Step S03 creating, from the plurality of live video images 111, a live panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules 175, 176 that reside at the one or more servers 190.
  • creating the live panoramic video image includes using all N of the corresponding plurality of the simultaneous live video images 155
  • creating the live 3D stereoscopic video image includes using a proper subset of said corresponding plurality.
  • Step S04 simultaneously delivering, from the one or more remote servers 170 to client viewing devices 190 in communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams 180, 181, the delivering being such that: (i) the live panoramic video stream 180 is delivered to a first plurality 191 or 193 of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality 192 or 194 of client viewing devices 190.
  • the two pluralities 191, 192 do not overlap, and in some embodiments the two pluralities 193, 294 overlap such that a group of client viewing devices 190 in the overlap have both live streams delivered to them.
  • the live video streams are synchronized with each other to enable smooth or seamless switching (toggling) between them.
  • not all steps of the method are performed.
  • FIG. 9 a flowchart of another method is presented for delivering simultaneous live video streams of panoramic and 3D stereoscopic images to a plurality of clients.
  • the method includes four steps, as follows:
  • Step Sll acquiring, using a camera apparatus 150 comprising a plurality of imaging devices 155, a corresponding plurality of simultaneous live video images 111 of a scene 420, or 421-422. Any of the camera apparatuses 150 disclosed in Figs. 6A- F would be suitable for carrying out the method.
  • Step S12 creating, from the plurality of live video images 111, a live panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules 175, 176 that reside at the camera apparatus 150.
  • creating the live panoramic video image includes using allN of the corresponding plurality of the simultaneous live video images 111
  • creating the live 3D stereoscopic video image includes using a proper subset of said corresponding plurality.
  • Step S13 transmitting the live panoramic video image and the live 3D stereoscopic video image from the camera apparatus 150 to one or more remote servers 170 using communications circuitry 157 of the camera apparatus 150 and communications channel 201.
  • Step S14 simultaneously delivering, from the one or more remote servers 170 to client viewing devices 190 in communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams 180, 181, the delivering being such that: (i) the live panoramic video stream 180 is delivered to a first plurality 191 or 193 of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality 192 or 194 of client viewing devices 190.
  • the two pluralities 191, 192 do not overlap, and in some embodiments the two pluralities 193, 294 overlap such that a group of client viewing devices 190 in the overlap have both live streams delivered to them.
  • the live video streams are synchronized with each other to enable smooth or seamless switching (toggling) between them.
  • not all steps of the method are performed. Steps from multiple methods can be combined into a composite method.
  • either of the methods can additionally include storage and retrieval steps, in which uploaded video images 111 or uploaded stitched and synthesized panoramic and stereoscopic images, respectively, are stored for later, e.g., on-demand retrieval by a client.
  • the system 100 includes panoramic-image video module 175 and stereo image video module 176, respectively configured to deliver the 2D panoramic video image and the 3D stereoscopic video image as synchronized video streams 180, 181 to a plurality 195 of client viewing devices 190 for display thereupon.
  • the panoramic-image and stereo-image video modules 175, 176 are further configured, respectively, to stitch said panoramic video image and synthesize said 3D stereoscopic video image from a plurality of simultaneously acquired video images 111.
  • the panoramic-image and stereo-image video modules 175, 176 reside at one or more remote servers 170 that are in communication with the plurality 195 of client viewing devices 190.
  • the system 100 of Fig. 10 additionally comprises a toggling module 168 for toggling between the panoramic and stereoscopic display modes if supported by a client viewing device 190.
  • the toggling can be responsive to a user selection received via a client-side interface, e.g., the toggling widget 94 of Figs. 7D and 7E.
  • the toggling can be triggered by a content analysis module 169, which is preferably configured to toggle between the display modes based on content analysis.
  • the content analysis module 169 can analyze the video images (at any stage of processing), and determine that a 3D stereoscopic stream is more appropriate or that a panoramic stream is more appropriate.
  • system 100 can include key processing and client-interface components.
  • the system 100 in in communication (e.g., by communications means 201) with a camera apparatus 150 for receiving video images 111.
  • the camera apparatus 150 e.g., a user of the camera apparatus
  • the camera apparatus 150 can initiate the communication with the one or more servers 170 for uploading the live video images 111, for processing by the stitching module 175 and stereo image synthesizing module 176, and delivery of the created panoramic video stream 180 and 3D stereoscopic video stream 181 to a plurality 195 of client viewing devices 190.
  • the camera apparatus 150 can be used as a‘slave’ device, i.e., where at least one of its operations (e.g., starting to acquire images, uploading to the servers 170, etc.) is controlled from the cloud 199, i.e., the one or more servers 170.
  • the system 100 with or without image processing (e.g., stitching and stereo synthesizing) can be offered as a commercial delivery platform for streaming 2D panoramic 180 and 3D stereo 181 streams for third parties initiating live (or delayed) multi-stream video capture, which is facilitated by establishing a communications connection between third party camera apparatus 150 and the one or more cloud-based servers 170 of the commercial delivery platform.
  • uploaded video images 111 and/or‘processed’ 2D panoramic and 3D stereo video images can be stored using non-transient storage media 140 which resides at the one or more servers 170.
  • Communications means 201, 202 and 203 shown illustrated in Fig. 11 can use any wired or wireless technology known in the art suitable for transmission of high-resolution video images.
  • a system 100 as disclosed herein for delivering 2D panoramic and 3D stereoscopic video streams can additionally or alternatively deliver enhanced and/or curated video streams to remote clients and remote client viewing devices.
  • an alternative video stream can be created by‘following’ a selected object captured in a scene.
  • This can include an alternative video stream in which the followed object is at the center or in the foreground, which requires a different composition than a typical content-agnostic stitched panorama or synthesized 3D image.
  • the original live video images can be of a concert.
  • the panoramic image stitched from the video images allow a remote client to experience the concert in 360°‘virtual reality’.
  • the stereo image created allow a remote client to see the band in the apparent or simulated 3D created by combining images with stereo parallax, as is known in the art.
  • an alternative image stream can be created out of the already-extant video images, with an alternative focus, such on the band’s guitarist, or drummer. This may require, in some cases, reducing the angle-of-view of the resulting panoramic stream, or compromising on the apparent stereo‘depth’ in the 3D stream. Despite these possible compromises in the quality of the‘immersive’ experience, providing clients with an opportunity to receive such alternative video streams with object-following can greatly enhance the user experience.
  • a 2D panoramic/3D stereoscopic delivery system 100 can provide the alternative video stream(s) in any number of ways: (i) instead of the ‘general’ or ‘default’ video streams delivered; (ii) toggleable between the‘regular’ live streams and the alternative ones; or (iii) jointly delivered, e.g., side-by-side or picture-in-picture.
  • an object-following module 245 is shown residing at the one or more remote servers 170 in the cloud 199.
  • Figs. 12A and 12B are not intended to show an exhaustive inventory of the components of a system for creating and delivering panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system.
  • the object-following module 245 includes a number of sub- modules.
  • Visual object analysis submodule 246 can be provided for identifying a visual object and optionally offering to clients the ability to‘follow’ the visual object.
  • Aural object analysis submodule 247 can provide the same function as the visual object analysis submodule 246 but for‘sound objects.’
  • Multi-stream video capture can include multi-channel audio capture as well.
  • the four microphones 405 of the camera apparatus 150 of Fig. 6F can capture four-channel surround sound; other numbers of microphones can be used to provide different levels of audio resolution.
  • the captured multi-channel audio can be processed so as to form an ambisonic soundtrack to the immersive video streams.
  • Ambisonics as noted by Wikipedia, is a full-sphere isotropic surround sound format.
  • the aural object analysis submodule 247 is preferably configured to perform audio processing that can identify‘sound objects’ that can then be‘followed’ in alternative video streams.
  • audio processing can identify‘sound objects’ that can then be‘followed’ in alternative video streams.
  • the vocal track might be identified as an aural object and thus an alternative video stream may remain focused on the source of the sound (the singer) as it moves around the stage.
  • the source of sound need not be one that moves about, and a static source of sound (e.g., the drummer) can also be identified as an aural object for object-following.
  • a user selection submodule 248 can be used to solicit and receive a user designation of a visual or aural following-object rather than use a following-object identified by the visual object analysis submodule 246 or the aural object analysis submodule 247.
  • a delivery submodule 248 can be used to provide the toggling and joint display (side-by-side and picture-in-picture) functionalities described earlier.
  • the object-following module 245 as described herein is based on a ‘post processing’ approach.
  • the enhancing of the video streams is based on visual and/or aural analysis, and/or on client inputs (e.g., using user-selection submodule 248).
  • an alternative video stream can be created by curating‘highlight’ and segments from video streams.
  • Use of the curating module can include an alternative video stream in which, for example, only highlights of captured and/or processed video is shown.
  • a curated video stream can be shorter in duration than the original source video images, and can include a mix of 2D panoramic and 3D stereoscopic segments.
  • a curated video stream can be the same duration as the original source video images, and can include a mix of 2D panoramic and 3D stereoscopic segments.
  • a curated video stream shorter in duration than the original source video images can include either all 2D panoramic segments or all 3D stereo segments.
  • Clients can be offered a choice to receive delivery of any one of such curated video streams in addition or instead of the‘regular’ full 2D panoramic and/or 3D stereo video streams.
  • the segments typically do not overlap, i.e., the same footage is typically not included twice (in two formats) in the same curated video stream.
  • both all -2D panoramic and all-3D stereo curated streams are offered for delivery, the streams are typically synchronized such that at any point a client can toggle from one curated stream to the other without loss or duplication of segments or portions of segments.
  • a curating module 255 is shown residing at the one or more remote servers 170 in the cloud 199.
  • FIGS. 13A and 13B are not intended to show an exhaustive inventory of the components of a system for creating and delivering panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system.
  • the skilled artisan will understand that the various system modules and components can be combined with any one or more of the components, features and communications arrangements of any of the systems 100 illustrated in figures accompanying this disclosure.
  • the curating module 255 includes a number of sub-modules. Segmentation submodule 256 provides the functionality of identifying segments within either the source (originally acquired) video images or the processed (2D panoramic and/or 3D stereo) video images.
  • the term‘segments’ as used herein can be used interchangeably with‘scenes’; machine-based techniques for identification of scenes from video images are known in the art.
  • a segment analysis submodule 257 can be provided to determine whether any given segment is preferably included in a curated stream as a 2D panoramic segment or as a 3D stereoscopic segment. This can be determined by the content and/or composition of a segment. To use once again the example of a concert, a segment heavy on the vocals might be better served by a 3D stereo conversion of the lead singer front and center.
  • An editing submodule 258 can be provided for editing the video segments into one or more curated video streams.
  • a delivery submodule 259 can be used for providing the delivery interface with remote client viewing devices.
  • a refraining module 265 is shown residing at the one or more remote servers 170 in the cloud 199.
  • Figs. 14A and 14B are not intended to show an exhaustive inventory of the components of a system for creating and delivering reframed panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system.
  • the term‘reframing’ as used herein means changing the aspect ratio of a video image or stream.
  • a desirable aspect ratio can be one that is compatible with display devices, such as, and not exhaustively, typical commercial video aspect ratios (width:height) of 1.78:1, 2:1, or 2.4:1.
  • display devices such as, and not exhaustively, typical commercial video aspect ratios (width:height) of 1.78:1, 2:1, or 2.4:1.
  • the various system modules and components can be combined with any one or more of the components, features and communications arrangements of any of the systems 100 illustrated in figures accompanying this disclosure.
  • the refraining module 265 includes a number of sub-modules.
  • Frame analysis submodule 266 provides the functionality of analyzing frames or segments within either the source (originally acquired) video images or the processed (2D panoramic and/or 3D stereo) video images.
  • this submodule is effective to analyze an image and determine an optional aspect ratio. Additionally or alternatively, this analysis can determine how the image or stream should be cropped so as to achieve a target aspect ratio.
  • reframing can be‘naive’ and simply retain a centered portion of an image.
  • the submodule may determine that the‘action’ in a scene is taking place in the left side of the original acquired video image and therefore reframing the image, non-centered, to include all of the left side and none of the right side, is preferable.
  • a user selection submodule 265 can be used to solicit and/or receive a user preference for the aspect ratio of a reframed video stream. Alternatively or additionally, this submodule can be used to solicit and/or receive a user preference for how the video stream is to be reframed. In an example, a user is provided with an onscreen capability to select where the image or stream is cropped, e.g., centered or off to one side (or top or bottom) of the original image.
  • a delivery submodule 268 can be used to provide the toggling functionality, i.e., to toggle between different aspect ratios, including, without limitation, the original and target aspect ratios.
  • Stream determination module 215, toggling module 168 (including content analysis module 169), object-following module 245, ambisonic processing module 250, curating module 255, and refraining module 265 are all configured to apply any relevant rules and/or other overhead inputs (e.g. image metadata) for determining the content and/or format of any given video stream to any given client viewing device 190.
  • rules and/or other overhead inputs are applied after the processing and stitching, etc., of acquired video images into panoramic and stereoscopic images, and are applied only for configuring the video streams delivered to the client viewing devices 190.
  • the stitching module 175 and stereo image synthesizing module 215 do not receive rules and/or other inputs other than the acquired (and optionally dewarped) video images for the purpose of stitching and synthesizing.
  • a flowchart of a method for delivering simultaneous live video streams 180, 181 of 2D panoramic and 3D stereoscopic images to a plurality of clients.
  • the method includes three steps, as follows:
  • Step S21 receiving, by one or more servers 170 and from a remote source, a plurality of live video streams 111 of a scene.
  • the video streams 111 have overlap therebetween.
  • Step S22 creating, at the remote server(s) 170, from the plurality of received live video streams 111, a live 2D panoramic video image and a live 3D stereoscopic video image of the scene using respective stitching and stereo image synthesizing modules 175, 176.
  • Step S23 simultaneously streaming, from the one or more remote servers 170 to client viewing devices 190 in remote communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams 180, 181, the simultaneous streaming being such that: (i) the live panoramic video stream 180 is delivered to a first plurality 191 of client viewing devices 190 and the live 3D stereoscopic video stream 181 is delivered to a second plurality 192 of client viewing devices, and (ii) the respective live video streams are synchronized.
  • the live panoramic video stream 180 is delivered instead to a first overlapping plurality 193 of client viewing devices 190 and the live 3D stereoscopic video stream 181 is delivered instead to a second overlapping plurality 194 of client viewing devices.
  • a flowchart of a method for delivering synchronized live video streams 180, 181 of 2D panoramic and 3D stereoscopic images of a scene from one or more remote servers 170 to a plurality of client devices 190.
  • the method includes three steps, as follows:
  • Step S31 stitching a live 2D panoramic video image from a plurality of live video images 111 of a scene received from a remote source.
  • Step S32 synthesizing a live 3D stereoscopic video image from at least some of the plurality of live video images 111 of the scene.
  • Step S33 delivering the 2D live panoramic video image and the live 3D stereoscopic video image as synchronized live video streams 180, 181 to the plurality of client viewing devices 190.
  • Step S41 accessing a 2D panoramic video image and a 3D stereo video image, stitched and synthesized from a plurality of video images 111.
  • Step S42 selecting respective first and second pluralities of video segments from the 2D panoramic video image and the 3D stereo image.
  • Step S43 creating a curated video stream that comprises the first and second plurality of video images.
  • Step S44 delivering the curated video stream to a plurality of remote client viewing devices 190, so as to deliver both 2D panoramic video segments and 3D stereo video segments.
  • a flowchart of a method for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers 170 to a plurality 195 of remote clients 190.
  • the method includes five steps, as follows:
  • Step S51 accessing a plurality of simultaneously-captured video images 111 of a scene, having a first duration.
  • Step S52 selecting video segments from within the video images, each corresponding to a different portion of the first duration.
  • Step S53 for each selected video segment either stitching a 2D panoramic video segment or synthesizing a 3D stereo video segment, based on an analysis of the segment’s content and/or composition.
  • Step S54 creating a curated video stream that comprises the stitched 2D panoramic video segments and the synthesized 3D stereo images.
  • Step S55 delivering the curated video stream to a plurality 195 of remote client viewing devices 190.
  • Fig. 19 a flowchart of a method is presented for delivering an enhanced video stream. The method includes two steps, as follows:
  • Step S61 acquiring a plurality of simultaneously-captured video images of a scene.
  • Step S62 delivering a live video stream comprising multiple time-differentiated segments, each segment comprising either a 2D panoramic video stream or 3D stereo video stream.
  • the selection of which of the 2D panoramic video stream and the 3D stereo video stream is included/synopsized in any given segment is based on the content and/or composition of the video images.
  • the acquiring and delivering of Steps S61 and S62 are done in real time so as to deliver a live video stream.
  • a flowchart of a method for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients.
  • the method includes five steps, as follows:
  • Step S71 accessing a plurality of simultaneously-captured, video images 111 of a scene, the video images 111 having a first duration.
  • Step S72 selecting video segments from within the video images 111, each segment corresponding to a different portion of the first duration.
  • Step S73 for each selected video segment both stitching a corresponding 2D panoramic video segment and synthesizing a corresponding 3D stereo video segment.
  • Step S74 creating a curated 2D panoramic video stream from the stitched 2D panoramic video segments, and a curated 3D stereo video stream from the synthesized 3D stereo video segments.
  • Step S75 delivering the curated 2D panoramic video stream to a first plurality 191 of remote client viewing devices 190, and the curated 3D stereo video stream to a second plurality 192 of remote client viewing devices 190.
  • the curated panoramic video stream 180 is delivered instead to a first overlapping plurality 193 of client viewing devices 190 and the curated 3D stereoscopic video stream 181 is delivered instead to a second overlapping plurality 194 of client viewing devices.
  • a flowchart of a method is presented for delivering an enhanced video stream.
  • the method includes four steps, as follows:
  • Step S81 acquiring a plurality of simultaneously-captured, video images 111 of a scene.
  • Step S82 initiating the delivery of a video stream, such that the video stream delivered to each remote client is either a 2D panoramic video stream stitched from the plurality of video images 111 or a 3D stereoscopic video stream synthesized from the plurality of video images 111.
  • Step S83 determining a following-object in the delivered video stream, including at least one of a person, thing or other visual element captured visually in at least one of the video images.
  • Step S84 delivering an alternative video stream to at least one remote client device 190 based on the determination of the following-object.
  • a flowchart of a method is presented for delivering an enhanced video stream.
  • the method includes four steps, as follows:
  • Step S91 acquiring a plurality of simultaneously-captured, video images 111 of a scene, including spatial audio captured by a plurality of microphones 405.
  • Step S92 initiating the delivery of a video stream, such that the video stream delivered to each remote client is either a 2D panoramic video stream stitched from the plurality of video images 111 and comprising ambisonic audio processed from the captured spatial audio or a 3D stereoscopic video stream synthesized from the plurality of video images 111 and comprising ambisonic audio processed from the captured spatial audio.
  • Step S93 determining a following-object in the delivered video stream, including a spatially-differentiated aural element captured in at least one of the video images.
  • Step S94 delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
  • a flowchart of a method is presented for delivering simultaneous video streams 180, 181 of 2D panoramic and 3D stereoscopic images to a plurality of clients.
  • the method includes five steps, as follows:
  • Step S101 acquiring, by one or more servers 170 (e.g., in the cloud 199), a plurality of video streams 111 of a scene, the video streams having overlap therebetween.
  • Step S102 at the one or more remote servers 170, creating, from the plurality of overlapping video streams 111 of the scene, a 2D panoramic video stream, and a 3D stereoscopic video stream, using respective stitching and stereo image synthesizing modules 175, 176.
  • Step S103 initiating the synchronized delivery, from the one or more remote servers 170 to a plurality 195 of client devices 190, a live video stream comprising one of the live 2D panoramic video stream 180 and the live 3D stereoscopic video stream
  • Step S104 receiving, at the one or more remote servers 170, a toggle command from a client.
  • Step S105 toggling the video stream delivered to the client so as to either replace the live 2D panoramic video stream 180 with the live 3D stereo video stream 181, or to replace the live 3D stereo video stream 181 with the live 2D panoramic video stream
  • a camera apparatus comprising multiple imaging devices can acquire simultaneous video images of a scene.
  • Each of the imaging devices can have a wide angle of view, for example 180° or more.
  • a scene can encompass a field of view of 180° or more, or 360°.
  • the simultaneous video images can be overlapping, meaning that at least two of the video images overlap or, equivalently, at least abut or nearly abut.
  • one of the video images can overlap with each of the two other video images.
  • the angles of view captured by one or more of the respective imaging devices is exactly 180°, or alternatively the angles of view of two of the imaging devices add up to exactly 360°, and in such cases the concept of ‘overlapping’ as used herein is expanded to include adjacent or abutting images.
  • angles of view captured by one or more of the respective imaging devices is less than 180°, or alternatively the angles of view of at least some of the respective imaging devices do not overlap, and in such cases the concept of‘stitching’ as used herein is expanded to include interpolating such that images can be considered‘stitched’ even though there is no actual overlap between them.
  • the camera apparatus can include communications circuitry configured for uploading the overlapping video images to the one or more servers, and in other embodiments, the communications equipment is configured for uploading stitched panoramic images and synthesized 3D stereoscopic images to the one or more servers.
  • communications circuitry can include integral or‘built-in’ communications circuitry, or to modules that are easily attachable to a camera apparatus, e.g., where the camera apparatus is equipped with an appropriate connector.
  • Figures 24A, 24B, 24C and 24D show, respectively, front, rear, bottom and top projection views of a camera apparatus 500 capable of acquiring multiple simultaneous video images from which 180° stereo and 360° panoramic video streams (or files) can be created.
  • a camera apparatus 500 is preferably constructed as a‘unibody’ device with a rigid frame 501.
  • Unibody in this context means that the device has no moving parts for positioning/repositioning individual imaging devices, and no folding mechanisms that place one or more individual imaging devices in a position more or less suitable for imaging in a specific mode (stereo or panoramic).
  • the rigid frame 501 can of course be constructed from multiple components, and it can also have openings with moveable ore removeable covers for connections and controls.
  • the camera apparatus 500 illustrated in Figs. 24A-24D has two‘major faces’ - the front and rear faces 540i, 540 2 - and four minor faces, which include the top 546 and the bottom 545.
  • the shape of the camera apparatus 500 in all of the figures consistently show, for clarity purposes, a single example of a product design, but the product can have any rectangular/prismatic shape (including, for example, rectangles with or without cut or rounded corners, or with rounded sides or edges). Nonetheless, it can be desirable to provide a bottom face 545 that allows the camera apparatus 500 to be placed on a flat surface unattended without support, whether for storage or for operation.
  • Camera apparatus 500 comprises three imaging devices 520.
  • a dual-mode device having only two imaging devices would necessarily require mechanical manipulation of one or both imaging devices in order to orient them properly for a chosen mode.
  • a device may have more than three imaging devices, but three is the practical minimum number of imaging devices for having dual mode capabilities in a rigid design.
  • Each one of the imaging devices 520 is a wide-angle camera, meaning that it has a respective wide-angle lens 530 installed thereupon.
  • a wide-angle lens 530 can comprise a fisheye lens, which advantageously has an angle-of-view of at least 180°, i.e., a hemispheric or greater angle-or-view. All three imaging devices 520 can be operable to acquire images at least when the camera is in the orientation shown in Figs. 24A-D.
  • a camera apparatus 500 and respective imaging devices 520 are an example of, and are respectively equivalent to, the camera apparatus 150 and respective imaging devices 155 discussed hereinabove.
  • This equivalence includes, without limitation, features and functionalities.
  • a first imaging device 520i is installed on a first major face 540i while the other two imaging devices 520 2 and 520r are installed on the second major face 540 2 .
  • First and second imaging devices 520i and 520 2 point in opposite directions and are therefore useful for jointly and simultaneously capturing images of a 360° scene that can be stitched into 360° panoramic images.
  • Each of these two imaging devices 520i, 520 2 has a respective optical axis 401i, 401 2 which are: directly opposite each other and therefore collinear; or in exactly opposing directions but not collinear, and therefore parallel; or in nearly opposing directions, meaning within ⁇ 10° of rotation from being opposite, or within ⁇ 5° of rotation from being opposite, or within ⁇ 1° of rotation from being opposite.
  • the terminology of being‘within ⁇ x°’ of opposite or parallel indicates a range, meaning anywhere in the range from -x° to +x° away from being opposite or parallel, as the case may be, and in every case means‘from -x° to +x° inclusive’.
  • a display screen 510 is installed on the first major face 540i.
  • the display screen 510 can be flush with the surface of the first major face 540i, or slightly sunken, or otherwise displaced from being flush based on aesthetic or ergonomic considerations. It is preferably a touchscreen so as to facilitate receiving user inputs.
  • a display-control interface 515 can be provided on the first major face 540i.
  • Display-control interface 515 can be, as per the example illustrated in Fig. 24B, a button.
  • the display screen 510 is operable to display and/or play back 180° or 360° images in response to an input received via either the display screen 510 (if a touchscreen) or the display-control interface 515.
  • a video display module may include a display screen 510, display-control interface 515, and any required electronics or software for performing the video-display functions.
  • the first major face 540i additionally includes an electronic status indicator 516i which can be operable to show on/off status, or to indicate that filming or recording or transmitting is in progress. While illustrated in Fig. 24B as a simple LED display, the electronic status indicator 516i can include any kind of electronic display such as, for example, a small LCD screen, or multiple LED’s. Either of the two major faces 540i, 540 2 may be flat or rounded, and/or may optionally include raised or sunken surfaces for aesthetic or ergonomic purposes.
  • the first major face 540i is identified in the figures as the‘front’ of the camera apparatus 500 (and second major face 540 2 as the‘rear’ face), but this is only for convenience and it has no meaning with respect to design or function, and especially in light of the fact that the camera apparatus 500 is operative to acquire images in any and all directions.
  • Two imaging devices 520 2 , 520 3 are installed on the second major face 540 2 .
  • a second electronic status indicator 516 2 can also be installed on the second major face 540 2 .
  • the second and third imaging devices 520 2 , 520 3 face in the same direction as each other, meaning that their respective optical axes 401 2 , 401 3 are parallel to each other, or within ⁇ 5° of parallel or within ⁇ 2° of parallel or within ⁇ 1° of parallel.
  • Images acquired simultaneously by the second and third imaging devices 520 2 , 520 3 can be used together in synthesizing stereo video images, e.g., 180° stereo video images for viewing on a 3D viewing device.
  • the second and third imaging devices 520 2 , 52 3 are preferably installed at the same height HLENS as each other, where height HLENS is the height in the y-dimension to the centerline CLLENS as shown in Fig. 24A, as measured in the y-dimension when the camera apparatus 500 is placed on a flat surface, i.e., with bottom face 545 on a flat surface.
  • the first imaging device 520i is also preferably installed at the same height HLENS as shown in Fig. 24B, and in particular at the same height of the second imaging device 520 2 with which it is‘back-to-back’.
  • the expression“same height” can mean that the imaging devices 520 are exactly at the same height, or have respective heights that are different from each other by no more than 50% of a diameter 908 of any one of the three respective lenses 530i, 530 2 , 530 3 (or the smallest of the respective diameters 908 if the lenses 530 don’t all have the same diameter), or by no more than 20% of the diameter 908, or no more than 2 millimeters, or no more than 1 millimeter.
  • Fig. 24D shows an example of a top face 546 of the camera apparatus 500.
  • Top face 546 is ergonomically enhanced with a concave, textured finger-grip area 547.
  • Finger-grip area 547 preferably extends in each direction across of a majority of the top face 546 so as to provide as substantial an area as possible on what is generally a small device.
  • the finger-grip area 547 can be at least partly textured, as illustrated, with area or areas that are less smooth to the touch than the surrounding surface.
  • the texture pattern illustrated in Fig. 24D is non-limiting and merely shows one possible example. In some embodiments which are not illustrated, the finger-grip area can be flat or convex, or concave but untextured.
  • Fig. 24C is a bottom view of the camera apparatus 500, showing a mechanical support-stand interface 550 installed on the bottom face 545, one of the minor faces of the camera apparatus 500.
  • the interface 550 can include, for example, a threaded socket for attaching a tripod with a matching threaded connecting element. According to embodiments, it can be desirable for the support interface 550 to be closer to the second and third imaging devices 520 2 , 520 3 than to the first imaging device 520i.
  • the interface 550 is preferably installed along a longitudinal centerline CLx which represents the‘middle’ of the camera apparatus 500 in the -dimension.
  • the ‘middle’ can be a line drawn along a central longitudinal axis of the bottom face 545, or can be a line indicating a center of balance that takes into account the weight of the various elements installed on both major faces 540 of the rigid body 501 of the camera apparatus 500. It can be desirable for the interface 550 to be positioned along the x- dimension centerline CLx and centered on the optical axes 401i, 401 2 of the first and second imaging devices 520i, 520 2 , as is illustrated in Fig. 24C.
  • this position of the interface 550 provides the greatest angular stability for capturing image for a 360° panorama, i.e., ensuring that the two imaging devices 520i, 520 2 tasked with acquiring the images for the 360° are held stably at an optimal angle for capturing images.
  • the interface 550 is not located exactly at the intersection of the centerline CLx and optical axes 401i, 401 2 , but is within the footprint of the two imaging devices 520i, 520 2 .
  • Fig. 25 shows an orthogonal projection 590 of first and second imaging devices 520i, 520 2 (including respective lenses 530i, 530 2 ) on the bottom face 545 of the camera apparatus 500.
  • the orthogonal projection 590 is a projection of the combined footprint of the two imaging devices 520i, 520 2 on the bottom face 545.
  • the mechanical support-stand interface 550, or at least a majority of the interface 550, is preferably located within the orthogonal projection 590.
  • the interface 550 can be substantially ( ⁇ 3 mm, or ⁇ 2 mm, or ⁇ 1 mm) centered around the‘intersection’ of the v-dimcnsion centerline CLx with the optical axes 401i, 401 2 , which are represented in Fig. 25 as optical centerline CLOPT.
  • the internal workings of a digital camera apparatus can include a planar array of photodetectors. Such a planar array can define a respective photodetector plane for each imaging device. This is illustrated in Fig. 26, where first, second and third imaging devices 520i, 520 2 , 520 3 are associated with respective photodetector planes Pi, P2, P3 ⁇ As discussed hereinabove with respect to Fig. 24C, the optical axes 401i, 401 2 of the first and second imaging devices 520i, 520 2 are directly opposite to each other (collinear), or within ⁇ 10° of rotation from being opposite to each other, or within ⁇ 5° of rotation from being opposite to each other, or within ⁇ 1° of rotation from being opposite to each other.
  • Each imaging device’s respective optical axis 401 is perpendicular to the given camera’s photodetector plane P. Therefore, the photodetector planes Pi, P2 associated with the first and second imaging devices 520i, 520 2 are correspondingly parallel to each other or within ⁇ 10° of being parallel, or within ⁇ 5°, or within ⁇ 1 °. As also discussed above with respect to Fig.
  • respective optical axes 401 2 , 401 3 of the second and third imaging devices 520 2 , 520 3 are parallel to each other, or within ⁇ 5° of parallel or within ⁇ 2° of parallel or within ⁇ 1° of parallel, and therefore the respective photodetector planes P2, P3 are parallel to each other, or within ⁇ 5° of parallel or within ⁇ 2° of parallel or within ⁇ 1° of parallel.
  • respective photodetector planes P2, P3 of the second and third imaging devices 520 2 , 520 3 are not co-planar, as the second imaging device 520 2 extends further from rigid body 501 than does the third imaging device 520 3 .
  • the second imaging device 520 2 extends at least 2 mm and no more than 10 mm further from the second major face 540 2 of the rigid body 501 than does the third imaging device 520 3 , or at least 3 mm and not more than 9 mm, or at least 4 mm and not more than 8 mm.
  • each imaging device 520 has an angle of view a limited by its optics such as, for example, its respective wide-angle lens 530.
  • the second imaging device 520 2 preferably has an angle of view of at least 180°.
  • a desirable outcome of the photodetector planes P2, P3 associated with the second and third imaging devices 520 2 , 520 3 not being co-planar, and of the second imaging device 520 2 extending further from the second face 540 2 of the rigid body 501 than does the third imaging device 520 3 is that the third imaging device 520 3 and its wide-angle lens 530 3 are outside the (at least) 180° angle of view of the second imaging device 520 2 .
  • the third imaging device 520 3 and its wide-angle lens 530 3 are outside the (at least) 180° angle of view of the second imaging device 520 2 .
  • the dashed line marked AOV2 (signifying the limits of the angle-of-view 0.2 of the second imaging device 520 2 ) can be seen to pass above the wide-angle lens 530 3 of the third imaging device 520 3 .
  • the first imaging device 520i also has an angle-of-view ai of at least 180° and the corresponding AOVi line passes well away from the display screen 510 and any other features of first major face 540i. The resulting effect is that panoramic images acquired using the first and second imaging devices 520i, 520 2 can be produced without capturing any extraneous parts of the camera apparatus 500 as per the examples described in this paragraph.
  • the camera apparatus 500 can provide, inter alia, the following image-capture modes: (i) using only the first and second imaging devices 520i, 520 2 for capturing images to be used in creating (stitching) 360° panoramic images; (ii) using only the second and third imaging devices 520 2 , 520 3 for capturing images to be used in creating 180° 3D stereo images; and (iii) using all three imaging devices 520i .. 520 3 for capturing images to be used for creating synchronized panoramic and stereo images, where images acquired by the first and second imaging devices 520i, 520 2 are used for creating the panoramic images, and images acquired by the second and third imaging devices 520 2 , 520 3 can be simultaneously used for creating the stereo images.
  • the image-capture mode can be user-selected using the display screen 510 (if a touchscreen) and/or the display-control interface 515.
  • each of the verbs, “comprise”, “include” and “have”, and conjugates thereof, are used to indicate that the object or objects of the verb are not necessarily a complete listing of members, components, elements or parts of the subject or subjects of the verb.
  • the singular form “a”, “an” and “the” include plural references unless the context clearly dictates otherwise.
  • the term “a marking” or “at least one marking” may include a plurality of markings.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Library & Information Science (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Studio Devices (AREA)
  • Stereoscopic And Panoramic Photography (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

Methods of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of remote clients comprise acquiring, using a camera apparatus comprising at least three imaging devices, a corresponding number of simultaneous live video images of a scene; transmitting the simultaneous live video images from the camera apparatus to one or more remote servers; creating, from the simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image using respective stitching and stereo image synthesizing modules; and simultaneously delivering, from the one or more remote servers to client viewing devices in communication therewith, the live 2D panoramic video image and the live 3D stereoscopic video image as respective live video streams.

Description

SYSTEMS, APPARATUSES AND METHODS FOR ACQUIRING, PROCESSING AND DELIVERING STEREOPHONIC AND PANORAMIC IMAGES
CROSS-REFERENCE TO RELATED APPLICATIONS
This patent application claim the benefit of U.S. Provisional Patent Application No. 62823409 filed on March 25, 2019, and of U.S. Provisional Patent Application No. 62893497 filed on August 29, 2019, both of which are incorporated herein by reference in their entirety.
FIELD OF THE INVENTION
The present invention relates to the acquiring of live video images and delivering them as panoramic and/or stereoscopic video streams for display on client viewing devices, methods and systems that incorporate stitching and stereo- synthesizing modules, and imaging devices for acquiring multiple synchronized images that can be stitched and/or synthesized into panoramic and/or stereo video streams.
BACKGROUND This section is intended to introduce the reader to various aspects of art that may be related to various aspects of the present invention, which are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present invention. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art.
There is ever-increasing consumer demand for handheld electronic devices that can perform functions that previously required larger devices and expensive software to accomplish. In the field of photography, there is interest in being able to produce more complex images beyond the ordinary digital still and video images that have become known for many years. Such complex images include panoramic images of more than 180° in breadth and even up to 360°, i.e., surround images that can be used in‘virtual reality’ viewers or even on regular displays which can allow a viewer to shift their view within the panorama; such a view-shifting viewing feature is available, for example, on the popular video website Youtube. Creating panoramic images from multiple images involves techniques such as stitching the images together and resolving any overlap between them. Another example of complex images is the stereoscopic image, which creates or enhances the illusion of depth in an image by means of stereopsis for binocular vision. This illusion of depth is what is commonly known as‘3D’ (short for three-dimensional) and is popular in various types of media. There is a need for a device that can capture images that can be used to create complex images easily, including without an a priori decision as to whether the acquire images are to be used in panoramic or stereoscopic complex images.
There is also a need for a system to acquire or process complex images and stream them so that users can receive simultaneous panoramic and stereoscopic video streams for selection, and/or have the ability to select curated, enhanced and/or reframed video streams.
SUMMARY
According to embodiments of the invention, a method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of remote clients comprises: (a) acquiring, using a camera apparatus comprising at least three imaging devices, a corresponding number of simultaneous live video images of a scene; (b) transmitting the simultaneous live video images from the camera apparatus to one or more remote servers; (c) creating, from the simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image using respective stitching and stereo image synthesizing modules, wherein (i) at least one of the simultaneous live video images is used in the creating of both the panoramic video image and the stereoscopic video image, (ii) at least one of the simultaneous live video images is used in the creating of the panoramic video image and not of the stereoscopic video image, and (iii) at least one of the simultaneous live video images is used in the creating of the stereoscopic image and not of the panoramic video image; and (d) simultaneously delivering, from the one or more remote servers to client viewing devices in communication therewith, the live 2D panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live 2D panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the respective live video streams are synchronized.
In some embodiments, an operation of the camera apparatus can be controlled from the one or more remote servers.
In some embodiments, at least the simultaneous live video images used in creating the 3D stereoscopic video image can have an overlap therebetween.
In some embodiments, the delivering can include delivering a single one of the live 2D panoramic video stream and the live 3D stereoscopic video stream to client viewing devices that are included in both the first and second pluralities of client viewing devices.
In some embodiments, the inclusion of any given client viewing device in either or both of the first and second pluralities of client viewing devices can be based upon a capability of the client viewing device.
In some embodiments, the inclusion of any given client viewing device in either or both of the first or second plurality of client viewing devices can be based upon a user selection. In some embodiments, a user selection can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device. In some embodiments involving a user selection, the user selection can be received before the delivering of the live video streams. In some embodiments involving a user selection, the user selection can be received during the delivering of the live video streams. In some embodiments involving a user selection, the user selection can be received via a client-side interface. In some embodiments involving a user selection, the user selection can be toggleable during the delivering of the live video streams.
In some embodiments, the transmitting or receiving, and the creating and delivering can be performed in real time.
According to embodiments of the invention, a method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients comprises: (a) receiving, by one or more servers and from a remote source, a plurality of simultaneous live video streams of a scene; (b) at the remote server(s), creating, from the plurality of received simultaneous live video streams, a live 2D panoramic video image and a live 3D stereoscopic video image of the scene using respective stitching and stereo image synthesizing modules; and (c) simultaneously streaming, from the one or more remote servers to client viewing devices in remote communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams, the simultaneous streaming being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the respective live video streams are synchronized.
In some embodiments, each live video stream can be associated with a different respective view direction, the view directions and angles of view of the video streams providing overlap therebetween.
In some embodiments, it can be that (i) the 3D stereoscopic video is unstitched so that a left video stream thereof is from a first of the live video streams and a right video stream thereof is from a second of the live video streams, and (ii) the 2D panoramic image comprises a stitching between the first and second live video streams.
According to embodiments of the invention, a method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of remote clients comprises: (a) acquiring, using a camera apparatus comprising a plurality of imaging devices, a corresponding plurality of simultaneous live video images of a scene; (b) transmitting the plurality of simultaneous live video images from the camera apparatus to one or more remote servers; (c) creating, from the plurality of simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules; and (d) simultaneously delivering, from the one or more remote servers to client viewing devices in communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the respective live video streams are synchronized.
In some embodiments, an operation of the camera apparatus can be controlled from the one or more remote servers.
In some embodiments, it can be that more of the simultaneous live video images are used in creating the live panoramic video image than in creating the live 3D stereoscopic video image, and at least the simultaneous live video images used in creating the 3D stereoscopic video image have an overlap therebetween.
In some embodiments, the inclusion of any given client viewing device in either or both of the first and second pluralities of client viewing devices can be based upon a capability of the client viewing device. In some embodiments, the inclusion of any given client viewing device in either or both of the first or second plurality of client viewing devices can be based upon a user selection.
In some embodiments, the delivering can include delivering a single one of the live panoramic video stream and the live 3D stereoscopic video stream to client viewing devices that are included in both the first and second pluralities of client viewing devices. In some such embodiments, a user selection can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
In some embodiments, a user selection can be received before the delivering of the live video streams. In some embodiments, a user selection can be received during the delivering of the live video streams. In some embodiments, a user selection can be received via a client-side interface. In some embodiments, the user selection can be toggleable during the delivering of the live video streams.
In some embodiments, a location of a client viewing device can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device. In some embodiments, the content of the video images can be used in determining which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device. In some embodiments, a technical characteristic of one or both of the live video streams can be used in determining which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
In some embodiments, it can be that the live panoramic video image and the live 3D stereoscopic video image are (i) stored in a non-transitory computer-readable storage medium at the one or more remote servers and (ii) made available for synchronized streaming on demand after the delivering of the live video streams has commenced. In some embodiments, it can be that the live panoramic video image and the live 3D stereoscopic video image are (i) stored in a non-transitory computer- readable storage medium at the one or more remote servers and (ii) made available for synchronized streaming on demand after the delivering of the live video streams has ended.
In some embodiments, at least one live video image of the plurality of simultaneous live video images can be dewarped by the camera apparatus before they are transmitted to the one or more remote servers. In some embodiments, at least one live video image of the plurality of simultaneous live video images can be dewarped at the one or more remote servers.
In some embodiments, the plurality of imaging devices has a corresponding plurality of respective optical axes, each of which can be angularly displaced from all of the others. In some embodiments, the plurality of imaging devices has a corresponding plurality of respective optical axes, of which at least a first two can be substantially parallel to each other and at least one other one can be parallel to but in an opposite direction to the at least first two.
In some embodiments, the creating of the live panoramic video image by the stitching module can include asymmetrically cropping at least one of the simultaneous live video images.
In some embodiments, the creating of at least one of the live panoramic video image and the live 3D stereoscopic image can include at least one of rotation and translation of one of the simultaneous live video images.
In some embodiments, the transmitting or receiving, and the creating and delivering can be performed in real time. A system for delivering synchronized live video streams of panoramic and 3D stereoscopic images of a scene according to embodiments of the present invention is disclosed, the system comprising: (a) respective stitching and stereo image synthesizing modules configured to create, from a plurality of live video images of a scene, a live panoramic video image and a live 3D stereoscopic video image; and (b) a delivery module configured to deliver, from one or more remote servers, the live panoramic video image and the live 3D stereoscopic video image as synchronized live video streams to a plurality of client viewing devices in communication with the one or more remote servers.
In some embodiments, the system can be configured to receive the plurality of live video images of a scene from a camera apparatus comprising a plurality of imaging devices respectively configured to simultaneously acquire the plurality of live video images of the scene. In some such embodiments, the system can be additionally configured to control the camera apparatus remotely from the one or more remote servers.
In some embodiments, at least one of the stitching module and the stereo image synthesizing module can be located at the one or more remote servers.
In some embodiments, the system can be configured to store the live panoramic video image and the live 3D stereoscopic video image in a non-transitory computer- readable storage medium at the one or more remote servers, so as to make said live panoramic video image and said live 3D stereoscopic video image available for synchronized streaming on demand at a later time.
In some embodiments, the system can additionally comprise a dewarping module residing at the one or more remote servers for dewarping each of the plurality of live images.
In some embodiments, the stitching module can be additionally configured to perform asymmetric cropping of at least one of the live video images.
In some embodiments, at least one of said stitching module and said stereo image synthesizing module can be additionally configured to perform at least one of rotation and translation of one of the live video images. In some embodiments, the system can be configured to deliver said synchronized live video streams of panoramic and 3D stereoscopic images in real time.
In some embodiments, the system can additionally comprise a toggling module configured to enable client toggling between a 3D stereoscopic display mode and a panoramic display mode.
A method is disclosed, according to embodiments, for delivering synchronized live video streams of 2D panoramic and 3D stereoscopic images of a scene from one or more remote servers to a plurality of client devices. The method comprises: (a) stitching a live 2D panoramic video image from a plurality of live video images of a scene received from a remote source; (b) synthesizing a live 3D stereoscopic video image from at least some of the plurality of live video images of the scene; and (c) delivering the live 2D panoramic video image and the live 3D stereoscopic video image as synchronized live video streams to the plurality of client viewing devices.
A method is disclosed, according to embodiments, for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients. The method comprises: (a) accessing a 2D panoramic video image and a 3D stereo video image, respectively stitched and synthesized from a plurality of simultaneously-captured video images of a scene, said video images of the scene having a first duration; (b) selecting from within the stitched 2D panoramic video image a first plurality of video segments for displaying in 2D panoramic mode and from within the synthesized 3D stereo video image a second plurality of video segments for displaying in 3D stereoscopic mode, wherein the selecting of each respective video segment includes analyzing at least one of its content and its composition, each one of the video segments corresponding to a different portion of the first duration; (c) creating a curated video stream that comprises the first and second plurality of video images; and (d) delivering the curated video stream, from the one or more servers to a plurality of remote client viewing devices in remote communication therewith, so as to deliver both 2D panoramic video segments and 3D stereo video segments to the client viewing devices.
A method is disclosed, according to embodiments, for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients. The method comprises: (a) accessing a plurality of simultaneously-captured, video images of a scene, said video images of the scene being having a first duration; (b) selecting a plurality of video segments from within said video images of the scene, each one of the video segments corresponding to a different portion of the first duration; (c) for each selected video segment, performing one video creation action selected from (i) stitching a 2D panoramic video segment and (ii) synthesizing a 3D stereo video segment, wherein the determination of which video creation action is performed for each respective selected video segment is based on an analysis of at least one of the segment’s content and the segment’s composition; (d) creating a curated video stream that comprises the stitched 2D panoramic video segments and the synthesized 3D stereo images; and (e) delivering the curated video stream, from the one or more servers to a plurality of remote client viewing devices in remote communication therewith.
In some embodiments, the plurality of video images of the scene can be captured by imagers having respective optical axes angularly-displaced and/or laterally displaced from each other.
In some embodiments, the 3D stereo images can be 360° 3D stereo images and the synthesizing includes stitching.
A method is disclosed according to embodiments for delivering an enhanced video stream. The method comprises: (a) acquiring a plurality of video images of a scene, said video images being simultaneously captured by imagers having angularly- displaced and/or laterally displaced optical axes; and (b) delivering, to a plurality of remote client devices, a live video stream comprising multiple time-differentiated segments, each segment comprising one of (i) a 2D panoramic video stream stitched from the plurality of video images captured during the period of time corresponding to the respective segment, and (ii) a 3D stereo video stream synthesized from at least some of the plurality of video images captured during the period of time corresponding to the respective segment. According to the method, (i) the selection of which of the 2D panoramic video stream and the 3D stereo video stream is included in any given segment is based on the content and/or composition of the video images during the period of time corresponding to the respective segment, and (ii) the acquiring and delivering are done in real time so as to deliver a live video stream. A method is disclosed according to embodiments for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients. The method comprises: (a) accessing a plurality of simultaneously-captured, video images of a scene, said video images of the scene being having a first duration; (b) selecting a plurality of video segments from within the video images, each one of the video segments corresponding to a different portion of the first duration; (c) for each selected video segment, (i) stitching a corresponding 2D panoramic video segment and (ii) synthesizing a corresponding 3D stereo video segment; (d) creating a curated 2D panoramic video stream that comprises the stitched 2D panoramic video segments, and a curated 3D stereo video stream that comprises the synthesized 3D stereo video segments; and (e) delivering, from the one or more remote servers to client viewing devices in remote communication therewith, the curated 2D panoramic video stream and the curated 3D stereoscopic video stream, the delivering being such that: (i) the curated 2D panoramic video stream is delivered to a first plurality of client viewing devices and the curated 3D stereoscopic video stream is delivered to a second plurality of client viewing devices.
In some embodiments, the inclusion of any given client viewing device in either or both of the first or second plurality of client viewing devices can be based upon a user selection.
In some embodiments, the delivering can include delivering a single one of the curated panoramic video stream and the curated 3D stereoscopic video stream to client viewing devices that are included in both the first and second pluralities of client viewing devices. In some such embodiments, a user selection can determine which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device. In some such embodiments, the user selection can be received before the delivering of the curated video streams. In some such embodiments, the user selection can be received during the delivering of the curated video streams. In some such embodiments, the user selection can be toggleable during the delivering of the live video streams.
A method is disclosed, according to embodiments, for delivering an enhanced video stream. The method comprises: (a) acquiring a plurality of simultaneously- captured, video images of a scene; (b) initiating the delivery of a video stream from one or more servers to a plurality of remote client devices, such that the video stream delivered to each remote client is one of (i) a 2D panoramic video stream stitched from the plurality of video images and (ii) a 3D stereoscopic video stream synthesized from the plurality of video images; (c) determining a following-object in the video stream delivered to each remote client, the following-object including at least one of a person, thing or other visual element captured visually in at least one of the video images; and (d) delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
A method is disclosed, according to embodiments, for delivering an enhanced video stream. The method comprises: (a) acquiring a plurality of simultaneously- captured, video images of a scene, the video images including spatial audio captured by a plurality of microphones; (b) initiating the delivery of a video stream from one or more servers to a plurality of remote client devices, such that the video stream delivered to each remote client is one of (i) a 2D panoramic video stream stitched from the plurality of video images and comprising ambisonic audio processed from the captured spatial audio, and (ii) a 3D stereoscopic video stream synthesized from the plurality of video images and comprising ambisonic audio processed from the captured spatial audio; (c) determining a following-object in the video stream delivered to each remote client, the following object including a spatially-differentiated aural element captured in at least one of the video images; and (d) delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
In some embodiments, the determining of the follow-object can include receiving a user input about the following-object at the one or more servers.
In some embodiments, the determining of the follow-object can include analyzing at least one of (i) the overlapping video images, (ii) the 2D panoramic video stream and (iii) the 3D stereo video stream.
In some embodiments, the determining of the follow-object can use stereoscopic parallax to identify an object.
In some embodiments, the delivery of the alternative video stream can preempt the delivery of the video stream delivered to each remote client responsively to receipt of a user command or selection. In some embodiments, the video stream delivered to each remote client can be toggleably switchable with the alternative video stream.
In some embodiments, the video stream delivered to each remote client can be delivered together with the alternative video stream such that the alternative video stream is displayable side-by-side or picture -in-picture together with video stream delivered to each remote client.
In some embodiments, it can be that the video streams are live and synchronized with each other.
A method is disclosed, according to embodiments, for delivering a reframed video stream. The method comprises: (a) accessing a plurality of simultaneously- captured video images of a scene; (b) creating, from the plurality of simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules, the 2D panoramic video image and the 3D stereoscopic video image having respective first and second aspect ratios; and (c) delivering, from the one or more remote servers to client viewing devices in communication therewith, the live 2D panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) at least one of the 2D panoramic video stream and the 3D stereoscopic video stream is reframed for delivery in a third aspect ratio that is different than the first and second aspect ratios.
In some embodiments, the respective live video streams can be delivered simultaneously and can be synchronized.
In some embodiments, it can be that the accessing includes accessing at least three simultaneously-captured video images of the scene, at least one of which is used in both the panoramic stream and in the stereoscopic stream, at least one of which is used in the panoramic stream and not in the stereoscopic stream, and at least one of which is used in the stereoscopic stream and not in the panoramic stream.
A method is disclosed, according to embodiments, for delivering simultaneous video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients. The method comprises: (a) acquiring, by one or more servers, a plurality of video streams of a scene, the video streams having overlap therebetween; (b) at the one or more remote servers, creating, from the plurality of overlapping video streams of the scene, a 2D panoramic video stream, and a 3D stereoscopic video stream, using respective stitching and stereo image synthesizing modules; (c) initiating the synchronized delivery, from the one or more remote servers to a plurality of client devices, a live video stream comprising one of the live 2D panoramic video stream and the live 3D stereoscopic video stream; (e) receiving, at the one or more remote servers, a toggle command from a client; and (f) toggling the video stream delivered to the client so as to (i) replace the live 2D panoramic video stream with the live 3D stereo video stream, or (ii) to replace the live 3D stereo video stream with the live 2D panoramic video stream.
In some embodiments, it can be that the 2D panoramic video stream has a 360° angle of view and the 3D stereo video stream has a 180° angle of view.
In some embodiments, it can be that the 3D stereoscopic video is unstitched so that a left video stream thereof is from a first of the overlapping video streams and a right video stream thereof is from a second of the overlapping video streams, the 2D panoramic image comprising a stitching between the first and second live video streams.
In some embodiments, it can be that the overlapping video images are live video images, and the creating and initiating delivery are performed in real time.
According to embodiments a camera apparatus for acquiring multiple simultaneous video images for creating therefrom 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces; and (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and a respective planar array of photodetectors defining a respective photodetector plane, each imaging device defining a respective optical axis orthogonal to said respective photodetector plane, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and the third imaging device is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto, the second imaging device has an angle of view of at least 180°, and (iii) a 180° concentric portion of the second-imaging-device angle of view does not encompass any of the lens of the third camera.
In some embodiments, it can be that the angle of view of the second imaging device does not encompass any of the lens of the third camera.
In some embodiments, the photodetector plane of the second imaging device can be not coplanar with, and displaced further outward from the second major face than, the photodetector plane of the third imaging device. In some embodiments, the photodetector plane of the second imaging device can be displaced further outward from the second major face than the photodetector plane of the third imaging device by at least 3 millimeters.
In some embodiments, the imaging device apparatus additionally comprises a video-display module comprising a display screen installed on the first major face.
In some embodiments, the first imaging device has an angle of view of at least 180°, and said angle of view does not encompass any of the display screen.
In some embodiments, the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto. In some embodiments, the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto.
In some embodiments, the camera apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices acquire the multiple simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the multiple simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the multiple simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via a display-control interface.
In some embodiments, a second minor surface opposite the first minor surface can comprise a concave, textured finger-grip area extending across of a majority of the second minor surface in each of two orthogonal directions.
According to embodiments of the invention, a camera apparatus for acquiring multiple simultaneous video images for creating 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces; (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and defining a respective optical axis, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, (ii) the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and (iii) \the third imaging device is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto; (c) a video-display module comprising a display screen installed on the first major face, the display screen being operable to display and/or play back 180° or 360° images in response to an input received via the display screen and/or via a display- control interface; and (d) a mechanical support-stand interface installed on a first minor face of the rigid body, disposed closer to each of the first and second imaging devices than to the third imaging device.
In some embodiments, it can be that a majority of the footprint of the mechanical support-stand interface on the first minor face falls within the borders of an orthographic projection of the first and second imaging devices on the first minor face.
In some embodiments, it can be that all of the footprint of the mechanical support-stand interface on the first minor face falls within the borders of an orthographic projection of the first and second imaging devices on the first minor face.
In some embodiments, the footprint of the mechanical support-stand interface on the first minor face can be center-aligned with a centerline of an orthographic projection of the first and second imaging devices on the first minor face on two orthogonal axes.
In some embodiments, the apparatus can additionally comprise a first electronic status indicator on the first major face and a second electronic status indicator on the second major face.
In some embodiments, the optical axis of the second imaging device can define a second direction that is opposite to the first direction or no more than 1 degree from being opposite thereto.
In some embodiments, the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto.
In some embodiments, the optical axis of the third imaging device can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto. In some embodiments, the apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices, acquire the simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via the display-control interface.
In some embodiments, a second minor surface opposite the first minor surface can comprise a concave, textured finger-grip area extending across of a majority of the second minor surface in two orthogonal directions.
According to embodiments of the invention, a camera apparatus for acquiring multiple simultaneous video images for creating 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces, and a bottom face characterized by the presence thereupon of a mechanical support- stand interface; (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and defining a respective optical axis, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, (ii) the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and (iii) the third imaging device is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto; and (c) a video-display module comprising a display screen installed on the first major face, the display screen being operable to display and/or play back 180° or 360° images in response to an input received via the display screen and/or via a display-control interface, wherein: (i) the bottom face defines an x-z plane, oriented such that a long centerline CLx of the bottom face has only an x-axis dimension, (ii) a y-axis is orthogonal to the x-z plane and defines a vertical direction, and (iii) when the camera apparatus is standing unattended on a flat surface on its bottom face, the first, second and third imaging devices are aligned with each other at the same height in the vertical direction within a tolerance of one -half of a diameter of any of the respective ultra wide-angle lenses.
In some embodiments, it can be that when the camera apparatus is standing unattended on a flat surface on its bottom face, the first, second and third imaging devices are aligned with each other at the same height in the vertical direction within a tolerance of one -fifth of a diameter of any of the respective ultra wide-angle lenses.
In some embodiments, it can be that when the camera apparatus is standing unattended on a flat surface on its bottom face, the first, second and third imaging devices are aligned with each other at the same height in the vertical direction within a tolerance of one millimeter.
In some embodiments, the first and second imaging devices can be aligned with each other in the x -dimension within a tolerance of one millimeter.
In some embodiments, the mechanical support-stand interface can be aligned with the first and second imaging devices in the x-dimension within a tolerance of five millimeters. In some embodiments, the optical axis of the second imaging device can define a second direction that is opposite to the first direction or no more than 1 degree from being opposite thereto.
In some embodiments, the optical axis of the third camera can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto.
In some embodiments, the optical axis of the third camera can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto.
In some embodiments, the apparatus can additionally comprise a first electronic status indicator on the first major face and a second electronic status indicator on the second major face.
In some embodiments, the apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices, acquire the simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via the display-control interface.
According to embodiments of the present invention, a camera apparatus for acquiring multiple simultaneous video images for creating 180° stereo and 360° panoramic video streams comprises: (a) a rigid body comprising first and second major faces; (b) first, second and third imaging devices fixedly installed in the rigid body, each imaging device comprising a respective ultra wide-angle lens and a respective planar array of photodetectors defining a respective photodetector plane, each imaging device defining a respective optical axis orthogonal to said respective photodetector plane, wherein: (i) the first imaging device is disposed such that its lens extends from the first major face and its optical axis defines a first direction, (ii) the second imaging device is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and (iii) the third camera is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto, wherein the respective photodetector planes of the first and second imaging devices are not coplanar.
In some embodiments, the photodetector plane of the second imaging device can be displaced further outward from the second major face than the photodetector plane of the third camera by at least 3 millimeters.
In some embodiments, it can be that) the second imaging device has an angle of view of at least 180°, and (ii) said angle of view does not encompass any of the lens of the third camera.
In some embodiments, it can be that (i) the second imaging device has an angle of view of at least 180°, and (ii) a 180° concentric portion of the angle of view does not encompass any of the lens of the third camera.
In some embodiments, the apparatus can additionally comprise a video-display module comprising a display screen installed on the first major face.
In some embodiments, it can be that (i) the first imaging device has an angle of view of at least 180°, and (ii) said angle of view does not encompass any of the display screen.
In some embodiments, the optical axis of the third camera can define a third direction that is parallel to the second direction or within 5 degrees of being parallel thereto.
In some embodiments, the optical axis of the third camera can define a third direction that is parallel to the second direction or within 1 degree of being parallel thereto.
In some embodiments, the apparatus can be configured to acquire multiple simultaneous video images in a first mode in which the first and second imaging devices, acquire the simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third imaging devices acquire the simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third imaging devices acquire the simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via a display-control interface. In some embodiments, a second minor surface opposite the first minor surface can comprise a concave, textured finger-grip area extending across of a majority of the second minor surface in each of two orthogonal directions.
BRIEF DESCRIPTION OF THE DRAWINGS The invention will now be described further, by way of example, with reference to the accompanying drawings, in which the dimensions of components and features shown in the figures are chosen for convenience and clarity of presentation and not necessarily to scale. Also, in some drawings the relative sizes of objects, and the relative distances between objects, may be exaggeratedly large or small for the sake of convenience and clarity of presentation. In the drawings:
Figs. 1A, IB and 1C show block diagrams of systems for delivering synchronized live video streams to a plurality of client viewing devices, according to embodiments of the present invention.
Fig. 2 shows block diagrams illustrating various examples of incorporating a dewarping module in a system for delivering synchronized live video streams, according to embodiments of the present invention.
Fig. 3 shows a block diagram of a system for delivering synchronized live video streams to first and second pluralities of client viewing devices, according to embodiments of the present invention. Fig. 4 shows a block diagram of a system for delivering synchronized live video streams to first and second overlapping pluralities of client viewing devices, according to embodiments of the present invention.
Fig. 5A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention. Fig. 5B shows a block diagram showing cloud-based sub modules of a stream determination module of a system for delivering synchronized live video streams, according to embodiments of the present invention.
Figs. 6A, 6B and 6C are schematic illustrations of a camera apparatus comprising multiple imaging devices capturing simultaneous overlapping video images of a 180° scene, according to embodiments of the present invention.
Figs. 6D, 6E and 6F are schematic illustrations of a camera apparatus comprising multiple imaging devices capturing simultaneous video images of a 360° scene, according to embodiments of the present invention.
Figs. 6G, 6H and 61 are respective top, front and rear views of a camera apparatus embodying the functionality of the camera apparatus of Fig. 6D.
Fig. 7A is a schematic illustration of a client viewing device comprising a screen offering a selection between a 3D stereoscopic video stream and a panoramic video stream, according to embodiments of the present invention.
Fig. 7B is a schematic illustration of a client viewing device comprising a screen, showing side-to-side panning of a panoramic image by a user.
Fig. 7C is a schematic illustration of a client viewing device comprising a virtual reality headset and controller offering a selection between a 3D stereoscopic video stream and a panoramic video stream, according to embodiments of the present invention.
Figs. 7D and 7E are schematic illustrations of user wearing a client viewing device comprising a virtual reality headset offering a toggle option for toggling between a 3D stereoscopic video stream and a panoramic video stream, according to embodiments of the present invention.
Figs. 8 and 9 show flow charts of methods for delivering simultaneous live video streams of panoramic and 3D stereoscopic images, according to embodiments of the present invention.
Fig. 10 shows a block diagram of a system for delivering synchronized live video streams to a plurality of client viewing devices, including a toggling module, according to embodiments of the present invention. Fig. 11 shows a block diagram of a system for delivering synchronized live video streams to a plurality of client viewing devices, according to embodiments of the present invention.
Fig. 12A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention.
Fig. 12B shows a block diagram showing cloud-based sub modules of an object following module of a system for delivering synchronized live video streams, according to embodiments of the present invention.
Fig. 13A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention.
Fig. 13B shows a block diagram showing cloud-based sub modules of a curating module of a system for delivering synchronized live video streams, according to embodiments of the present invention according to embodiments of the present invention.
Fig. 14A shows a block diagram showing cloud-based modules of a system for delivering synchronized live video streams, according to embodiments of the present invention.
Fig. 14B shows a block diagram showing cloud-based sub modules of a reframing module of a system for delivering synchronized live video streams, according to embodiments of the present invention according to embodiments of the present invention.
Fig. 15 shows a flow chart of a method for delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients according to embodiments of the present invention.
Fig. 16 shows a flow chart of a method for delivering synchronized live video streams of 2D panoramic and 3D stereoscopic images of a scene from one or more remote servers to a plurality of client devices according to embodiments of the present invention. Figs. 17 and 18 show flow charts of methods of delivering curated video strea s including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients according to embodiments of the present invention.
Fig. 19 shows a flow chart of a method for delivering an enhanced video stream according to embodiments of the present invention.
Fig. 20 shows a flow chart of a method of delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients according to embodiments of the present invention.
Fig. 21 shows a flow chart of a method for delivering an enhanced video stream according to embodiments of the present invention.
Fig. 22 shows a flow chart of a method for delivering an enhanced video stream according to embodiments of the present invention.
Fig. 23 shows a flow chart of a method of delivering simultaneous video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients according to embodiments of the present invention.
Figs. 24A, 24B, 24C and 24D are, respectively, front, rear, bottom and top views of a camera apparatus comprising three imaging devices, according to embodiments of the present invention.
Fig. 25 is a bottom view of a camera apparatus comprising three imaging devices, showing the borders of an orthographic projection of two of the imaging devices on the bottom face of the apparatus, according to embodiments of the present invention.
Fig. 26 is a bottom view of a camera apparatus comprising three imaging devices, schematically showing the respective photodetector planes of the three imaging devices, according to embodiments of the present invention.
Fig. 27 is a bottom view of a camera apparatus comprising three imaging devices, schematically showing the respective angle of view of two of the imaging devices with respect to physical features on the two major surfaces of the apparatus, according to embodiments of the present invention. DETAILED DESCRIPTION OF THE ILLUSTRATED EMBODIMENTS
The invention is herein described, by way of example only, with reference to the accompanying drawings. With specific reference now to the drawings in detail, it is stressed that the particulars shown are by way of example and for purposes of illustrative discussion of the preferred embodiments of the present invention only, and are presented in the cause of providing what is believed to be the most useful and readily understood description of the principles and conceptual aspects of the invention. In this regard, no attempt is made to show structural details of the invention in more detail than is necessary for a fundamental understanding of the invention, the description taken with the drawings making apparent to those skilled in the art how the several forms of the invention may be embodied in practice. Throughout the drawings, like-referenced characters are generally used to designate like elements.
Note: Throughout this disclosure, subscripted reference numbers (e.g., 10i or 10A) may be used to designate multiple separate appearances of elements of a single species, whether in a drawing or not; for example: 10i is a single appearance (out of a plurality of appearances) of element 10. The same elements can alternatively be referred to without subscript (e.g., 10 and not 10i) when not referring to a specific one of the multiple separate appearances, i.e., to the species in general.
For convenience, in the context of the description herein, various terms are presented here. To the extent that definitions are provided, explicitly or implicitly, here or elsewhere in this application, such definitions are understood to be consistent with the usage of the defined terms by those of skill in the pertinent art(s). Furthermore, such definitions are to be construed in the broadest possible sense consistent with such usage.
The term ‘module’ as used herein means any combination of electronic hardware, software and firmware necessary to perform the function associated with the module, e.g., a ‘stitching module’ includes hardware, software and firmware for ‘stitching’ images. The myriad technologies available in each of the three technology categories of hardware, software and firmware are well known in their specific areas of technical endeavor, and the present invention is intended to encompass the use of any such technologies, whether known and commonly in use at the time of the invention or developed and made available during the life of any patent ensuing from this disclosure. The terms‘imaging device’ ,‘sensor’ and‘camera’ may be used interchangeably in this specification and in the appended claims, as well as in documents incorporated by reference, and all have the same meaning: a device for capturing digital images, including (at least) a CMOS or CCD device and a lens, and optionally any number of mechanical and electronic components and accessories. In contrast, the terms‘camera apparatus’ and‘imaging apparatus’ as used herein and in documents incorporated by reference have the same meaning as each other, are directed to an apparatus comprising multiple sensors/cameras/imaging devices. In this parlance, an imaging device, sensor or camera can be a component of a camera apparatus or imaging apparatus, but not vice versa.
According to some embodiments of the invention, a camera apparatus comprising multiple imaging devices can acquire simultaneous video images of a scene. Each of the imaging devices can have a wide angle of view, for example 180° or more. A scene can encompass a field of view of 180° or more, or 360° or more. The simultaneous video images can be overlapping, meaning that at least two of the video images overlap, or that there are multiple overlaps involving respective multiple pairs or groups of overlapping video images, e.g., at a minimum for any three images, image A overlaps with image B, and image B overlaps with image C. In some embodiments for three given images each image overlaps with the other two. In some embodiments a camera apparatus has the potential to capture at least some overlapping images, but in actual operation only non-overlapping images are captured; in some of these cases the non-overlapping images are abutting one another, and in other cases they are nearly abutting but not quite. In some embodiments the angles of view captured by one or more of the respective imaging devices is exactly 180°, or alternatively the angles of view of all of the imaging devices add up to exactly 360°, and in such cases the concept of‘overlapping’ as used herein is expanded to include adjacent or abutting images. In some embodiments the angles of view captured by one or more of the respective imaging devices is less than 180°, or alternatively the angles of view of at least some of the respective imaging devices do not overlap, and in such cases the concept of ‘stitching’ as used herein is expanded to include interpolating such that images can be considered‘stitched’ even though there is no actual overlap between them.
As the images are acquired by the camera apparatus or with minimal delay of no more than a few milliseconds or a few tens of milliseconds or a few hundreds of milliseconds. A stitching module can create a panoramic image from two or more of the overlapping (or abutting or nearly abutting) video images. At the same time, a 3D stereoscopic synthesizing module can create a 3D stereoscopic image from two or more of the overlapping video images, which can include at least in part the same images as the video images used in stitching the panoramic image.
Panoramic stitching and 3D stereoscopic synthesizing of multiple and overlapping images, including asymmetric cropping, in-camera calibration, image acquisition and manipulation, and related functions can be performed using any of the systems and methods described in pending international application PCT/IB2018/059771 filed on December 7, 2018, the contents of which are incorporated herein by reference in their entirety. The aforementioned methods and systems are also disclosed in the following US patent applications: Ser. No. 62/596,112 filed on December 7, 2018, Ser. No. 62/681,667 filed on June 6, 2018, Ser. No. 62/700,095 filed on July 18, 2018, Ser. No. 62/700,104 filed on July 18, 2018, Ser. No. 62/700,571 filed on July 18, 2018, Ser. No. 62/700,580 filed on July 18, 2018, and Ser. No. 62/700,588 filed on July 18, 2018, the contents of all of which are incorporated herein by reference in their entirety.
In some embodiments, either or both of the stitching module and the 3D stereoscopic synthesizing module are included in the camera apparatus, and in some embodiments, either or both of the stitching module and the 3D stereoscopic synthesizing module reside remotely, for example at one or more remote servers, which can be part of a server farm comprising one or more servers dedicated to the functions of stitching module and/or the 3D stereoscopic synthesizing module (and/or other modules which may be necessary for processing the video images, storing them, and/or delivering them to clients as video streams, where delivering them to clients can include selecting which of the video streams to deliver to any particular client). Alternatively, the one or more servers can be‘in the cloud’, i.e., servers available from a commercial computing operation that includes any number of networked servers (including, without limitation, shared, dedicated and virtual servers) available to any number of unrelated customers. For convenience, remote servers are illustrated throughout as being‘in the cloud’ but - as described above - this is not necessarily the case in some embodiments. In some embodiments, the camera apparatus can include communications circuitry configured for uploading the overlapping video images to the one or more servers, and in other embodiments, the communications equipment is configured for uploading stitched panoramic images and synthesized 3D stereoscopic images to the one or more servers. The phrase‘overlapping video images’ is used to indicate the video images captured by a camera apparatus before further processing such as dewarping or stitching or stereo-synthesizing, and the phrase will, in some cases, refer to video images that not actually overlapping but merely abutting one another or nearly abutting one another. The one or more servers can include a delivery module for delivering the panoramic and 3D stereoscopic images (regardless of whether they were respectively stitched and synthesized at the one or more servers or at the camera apparatus) to client viewing devices for display thereupon.
In this disclosure and in the claims appended thereto, the following terminology
Referring to the figures, and in particular to Fig. 1A, a block diagram shows an example of a system 100 for delivering synchronized live panoramic and stereoscopic video streams. It should be noted that throughout this disclosure, the term‘system’ if used without reference number applies to a system 100 for delivering synchronized live panoramic and stereoscopic video streams according to any one or more of the embodiments disclosed herein. The system 100 of Fig. 1A includes a camera apparatus 150 that includes N imaging devices 155i .. 155N, where N is an integer greater than 1. The camera apparatus 150 is preferably operable to acquire, using the imaging devices 155i .. 155N, N respective overlapping live video images llli .. 111N. The camera apparatus 150 also includes communications circuitry configured, inter alia, for uploading the overlapping video images llli .. 111N to a remote server 170 using communications channel 201 which, like all other communications channels disclosed herein, can incorporate any combination of wired and wireless communications technologies and is preferably configured to transmit video images at the resolution enabled by the imaging devices 155i .. 155N and at the throughput speed necessary for maintaining a substantially live and real-time transmission. When the term‘server’ is used herein, it is synonymous with the terms‘one or more servers,’‘one or more remote servers,’ and‘remote server,’ and we note that the specific distribution of loads and functions between multiple servers or multiple processors is not relevant to the present disclosure. According to embodiments, the system 100 additionally comprises a number of modules residing on one or more remote servers 170 in the cloud 199 - a stitching module 175 for stitching together overlapping live video images llli .. lllN that are transmitted to the server 170 by the camera apparatus 100, and a stereo image synthesizing module 176 for synthesizing 3D stereo images from the overlapping live video images llli .. 111N- In conjunction with the server 170, storage medium 140 (for example, a non-transitory, computer-readable storage medium) can be resident at the cloud 199.
As shown in Fig. 1A, the stitching module 175 can be operative to transmit, i.e., download, a panoramic video image stitched from the overlapping live video images llli 11 IN to a plurality 195 of client viewing devices 190 as a live panoramic video stream 180 using communications channel 202. The plurality 195 can include any number M of client viewing devices 190i .. 190M in accordance with the operating capabilities of the server 170 and the communications channel 202 or 203. The stereo image synthesizing module 176 can be operative to transmit, i.e., download, a 3D stereoscopic video image synthesized from the overlapping live video images llli .. lllN to the same plurality 195 of client viewing devices 190 as a live 3D stereoscopic video stream 181, using communications channel 203. It is important to note that the entire process embodied in the system of Fig. 1A - including, without limitation, acquiring the plurality of images overlapping live video images llli HIN, uploading the overlapping live video images llli .. 111N to the one or more remote servers 170, using the respective stitching module 175 and stereo image synthesizing module 176 so as to create therefrom respective panoramic and stereoscopic images and deliver said images as live panoramic video stream 180 and live 3D stereoscopic video stream 181 to the plurality 195 of client viewing devices 190 - is preferably performed without undue delays in processing time, such that the live streams 180, 181 are delivered in real time, i.e., within the amount of delay time acceptable in the industry for‘live’ and ‘real-time’ transmissions. It should be obvious to the skilled artisan that not all the components shown in Fig. 1 A are necessary for the proper operation of the system 100, nor does Fig. 1A illustrate every possible component and interface for such a system 100.
Referring now to Fig. IB, a second example of a system 100 for delivering synchronized live video streams is shown in a block diagram. The system 100 of Fig. IB has similarities to the system 100 previously discussed with respect to Fig. 1A, and a number of differences:
(1) Stitching module 175 and stereo image synthesizing module 176, in the example shown in Fig. IB, reside within the camera apparatus 150 and not in the cloud at the one or more servers 170 as was the case in the example of Fig. 1A;
(2) The communications circuitry 157 of the camera apparatus 150 is configured to upload processed (pre-stitched/pre-synthesized) video images as live streams, i.e., a‘stitched’ live panoramic video stream 180 and a ‘synthesized’ 3D stereoscopic video stream 181, to the one or more servers
170;
(3) A delivery module 178 at the one or more servers 170 is configured to deliver the respective live video streams 180, 181 to the plurality 195 of client viewing devices 190. Although not illustrated as being a component of system 100, in some embodiments the delivery module 178 can be a component of system 100.
It should be obvious to the skilled artisan that not all the components shown in Figs. 1A and IB are necessary for the proper operation of the system 100, nor do Figs. 1A and IB illustrate every possible component and interface for such a system 100. It should also be obvious that certain aspects of the two figures can be combined - as a non-limiting example, it can be that a specific one of the respective stitching and stereo image synthesizing modules 175, 176 may be included in the camera apparatus 150 and the other one of the two modules resides in the cloud 199, i.e., at the one more remote servers 170.
The block diagram of Fig. 1C illustrates another example of a system 100 for delivering synchronized live panoramic and stereoscopic video streams. In the illustrated embodiments, system 100 includes key processing and client-interface components. This is an example of a system 100 that does not necessarily include the camera apparatus 150 for acquiring the video images. Thus, the video images to be respectively stitched and synthesized into panoramic and 3D stereoscopic streams 180, 181 can come to the server 170 from any source, or can, for example, be stored on the storage medium 140 residing at the server 170. In the example of Fig. 1C, system components reside at servers 170 in the cloud 199, including stitching module 175 for stitching together overlapping video images, and a stereo image synthesizing module 176 for synthesizing 3D stereo images. Storage medium 140 (for example, a non-transitory, computer-readable storage medium) can be resident at the cloud 199 in conjunction with and preferably in electronic communication with the server 170.
As shown in Fig. 1C, the stitching module 175 can be operative to transmit, i.e., download, a panoramic video image stitched from, e.g. , overlapping live or stored video images to a plurality 195 of client viewing devices 190 as a panoramic video stream 180 using communications channel 202. The plurality 195 can include any number M of client viewing devices 190i .. 190M in accordance with the operating capabilities of the server 170 and the communications channel 202 or 203. The stereo image synthesizing module 176 can be operative to transmit, i.e., download, a 3D stereoscopic video image synthesized from, e.g., overlapping live or stored video images to the same plurality 195 of client viewing devices 190 as a live 3D stereoscopic video stream 181, using communications channel 203. In some embodiments, the delivery of panoramic video stream 180 and 3D stereoscopic video stream 181 is handled by delivery module 178.
As is known in the art, images acquired by imaging devices having ultra-wide angle lenses such as fish-eye lenses are commonly dewarped to remove the distortive effects of the lenses, before being further processed into complex images such as panoramic images or stereoscopic images (or even before being viewed as simplex images). A dewarping module can be included as part of a system 100 for delivering synchronized live panoramic and stereoscopic video streams 180, 181, or otherwise can be available for use by components of the system. A dewarping module 177 can reside in any one of various locations according to the specific embodiment. Fig. 2 shows a non-limiting set of examples for the location of dewarping modules 177. In Example A of Fig. 2, a dewarping module 177 is included in camera apparatus 150 which also includes stitching module 175 and stereo image synthesizing module 176. According to Example A, the dewarping module 177 can be configured to pre-process the acquired live overlapping video images (e.g., llli .. 111N of Fig. 1A) for the common use of the stitching module 175 and the stereo image synthesizing module 176. Example B of Fig. 2 shows an embodiment in which the dewarping module 177 is included in camera apparatus 150 but stitching module 175 and stereo image synthesizing module 176 are not included in the camera apparatus 150. In such an embodiment, the communications circuitry 157 of the camera apparatus 150 are preferably configured to upload dewarped video images and not the raw overlapping live video images llli .. 111N as discussed with respect to Fig. 1 A. Example C of Fig. 2 shows another embodiment, in which the dewarping module 177 resides in the cloud, e.g., at remote server 170, along with stitching module 175 and stereo image synthesizing module 176. Example D of Fig. 2 shows yet another embodiment in which each of the stitching module 175 and stereo image synthesizing module 176 include a dewarping module 177. This arrangement can be suitable to embodiments in which one of the stitching module 175 and stereo image synthesizing module 176 is located in a camera apparatus 150 and the other of the two modules resides at the remote servers 170, although the implementation of Example D is not limited to such embodiments.
Fig 3 illustrates yet another embodiment of a system 100 for delivering synchronized live panoramic and stereoscopic video streams 180, 181. A comparison with Fig. 1A will reveal that the difference between the block diagrams of Fig. 3 and Fig. 1A is that in Fig. 1A, the two live video streams 180, 181 are simply shown as being delivered to a plurality 195 of client viewing devices 190 (i.e., the M devices 190i .. 190M), while in Fig. 3 the following distinction is made: the live panoramic video image 180 is streamed (i.e., delivered) to a first plurality 191 of client viewing devices 190 (specifically the X devices 190i .. 190x), and the live 3D stereoscopic video image 181 is streamed (i.e., delivered) to a second plurality 192 of client viewing devices 190 (specifically the M-X devices 190x+i .. 190M), where X is an integer greater than 1 and M- X is also greater than 1. As shown in Fig. 4, a first overlapping plurality 193 of client viewing devices 190 can overlap with a second overlapping plurality 194 of client viewing devices 190, wherein the first overlapping plurality 193 includes the client viewing devices 190i .. 190x and the second overlapping plurality includes the client viewing devices 190x .. 190M- The client viewing devices 190i .. 190x-i of the first plurality 193 receive ‘only’ the live panoramic video image stream 180, the client viewing devices 190x+i .. 190M of the second plurality 194 receive‘only’ the live 3D stereoscopic video image stream 181; client viewing device 190x is included in both the first and second overlapping pluralities 193, 194 and receives both streams 180, 181. Although both streams are delivered to the devices 190 in the overlap of the pluralities 193, 194, only one video stream can viewed/displayed at a time, and a selection or determination of which stream is viewed/displayed will be discussed below. In other embodiments, the overlap can include more than one client viewing device 190 that receives both live video streams 180, 181. An embodiment in which the overlap includes ‘all’ of the client viewing devices, i.e., in which all of the client viewing devices 190i ... 190M simultaneously receive both live video streams 180, 181 has already been illustrated in Fig. 1A.
In some embodiments, as mentioned previously, a single client viewing device 190 can simultaneously receive, i.e., have streamed or delivered to it, two synchronized live video streams 180, 181. Any selection or switching/toggling between the two streams may be handled by or at the client viewing device 190. This can be desirable, for example, in implementations wherein it is desired to have a switching or selecting between the two streams 180, 181 be as instantaneous as possible for the receiver of both streams, and communications protocols or other communications ‘overhead’ makes it less desirable to switch between streams‘at the source’, i.e., at the one or more servers 170. In other implementations, (e.g., as shown in Fig. 3) only one stream is delivered to each client viewing device, and the determination or selection of which stream goes to which client viewing device is not made, at least not entirely, by or at the client viewing device 190.
For purposes of clarification, it should be pointed out that in some embodiments, a selection or determination is made by or on behalf of the system as to which one of two live synchronized video streams (panoramic or 3D stereo) is streamed (‘streamed’ and‘delivered’ are used interchangeably in this disclosure) to any given client viewing device. In some embodiments, a selection is made by a user as to which one of the two live synchronized video streams is streamed. In some embodiments, both streams can be delivered to a single client viewing device, and a selection is made, generally by a user, as to which of the two streams is displayed by the client viewing device.
Referring now to Figs. 5A and 5B, a stream determination module 215 is shown residing at the one or more remote servers 170 in the cloud 199. The dashed lines of the stitching module 175 and the stereo image synthesizing module 176 inform us that either or both of these two modules can reside at the one or more remote servers 170 (as in the example of Fig. 1A) or at the camera apparatus 150 (as in the example of Fig. IB).
The stream determination module 215 can be operative to make a determination or selection, or respond to a user selection, of which live video stream (panoramic or 3D stereo) is delivered to any given client viewing device 190 at any given time (although in some embodiments the determination or selection can be permanent or ‘until further notice’). Any of the sub-modules of the stream determination module 215 can determine in which plurality 191, 192 of Fig. 3 or 193, 194 or Fig. 4) a given client viewing device 190 is included.
The stream determination module 215 can include a user-selection based stream determination sub-module 216. In some embodiments a user selection can determine which live video stream is delivered to a given client viewing device 190. In other embodiments, a user selection can determine which of the two streams simultaneously delivered to a given client viewing device 190 is actually displayed by the client viewing device 190. In some embodiments, a user selection can be received via a client- side interface, e.g., on the client viewing device itself, as will be discussed later in connection with Figs. 7A and 7C. The user selection can be received before the initiation of a streaming event that includes delivering live video streams, or a user selection may be received during the delivering of the live video streams. In some embodiments, the selection can be toggled (switched back and forth) during the delivering of the live video streams, as will be discussed in connection with Figs. 7D, 7E and 10.
The stream determination module 215 can additionally or alternatively include a client-location based stream determination submodule 217. In some embodiments, the determination of which live stream is delivered to a given client viewing device is made on the basis of the client location at any resolution - by country, by region, or even by where in a building or room the client is located. In some embodiments, the selected live stream can be switched from panoramic to 3D stereo or the reverse if and when the client location changes.
The stream determination module 215 can additionally or alternatively include a client-history based stream determination submodule 218. In embodiments, such a submodule can make a determination based on client viewing history - for example, if a client always chooses the 3D stereo stream, it can be desirable (e.g., for purposes of better customer service or reduced communications overhead) to simply stream according to the client’s historical choice. In some such embodiments, a client can optionally be offered the option to switch streams.
The stream determination module 215 can additionally or alternatively include a client viewing device capability-based stream determination submodule 218. In some embodiments, it can be desirable to check, e.g., through a communications channel such as communications channels 202 or 203 what the technical capabilities (panoramic vs. stereo) of a client viewing device 190 are, i.e., by receiving the specification upon electronic request or looking up a model specification in a database, and to pre determine the choice of live streams accordingly.
The stream determination module 215 can additionally alternatively be operable to make a determination as to which live stream to deliver on the basis of a technical characteristic of the content of the video images (for example, but not exhaustively, the aspect ratio or field of view of the video images, the resolution of the video images, and the calibration status of the imaging devices used to acquire the video images). In some embodiments, the basis for the determination can be the nature of the content of the video images and which format is more appropriate, e.g., provides the best user experience.
The block diagrams of Figs. 5 A and 5B are not intended to show an exhaustive inventory of the components of a system for creating and delivering panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system. The skilled artisan will understand that the cloud-residing modules and components can be combined with any one or more of the components, features and communications arrangements of any of the systems 100 illustrated in figures accompanying this disclosure.
Referring now to Fig. 6A-6F, a number of examples of camera apparatuses 150 for simultaneous acquisition of overlapping video images are schematically illustrated.
In Fig. 6A, camera apparatus 150 includes two imaging devices 155i, 1552 with respective optical axes indicated by lines 401i, 4012- The camera apparatus 150 of Fig. 6A is operative to simultaneously acquire overlapping video images of scene 420 which has a combined field of view of more than 180° using the two imaging devices 155i, 1552.
In Fig. 6B, camera apparatus 150 includes three imaging devices 155i, 1552, 1553, with respective optical axes indicated by lines 401r, 4012 , 4013. The camera apparatus 150 of Fig. 6B is operative to simultaneously acquire overlapping video images of scene 420 which has a field of view of more than 180°. In a preferred mode of operation, video images acquired by imaging devices 155i and 1553 are used for synthesizing a 3D stereo video image for streaming, and images acquired by all three imaging devices 155i, 1552, 1553 are used for stitching a panoramic video image (and in other embodiments images from the same two image devices 155i and 1553 are used for creating both video streams).
In Fig. 6C, camera apparatus 150 includes two imaging devices 155i, 1552 with respective optical axes indicated by lines 401i, 4012- The camera apparatus 150 of Fig. 6C is operative to simultaneously acquire overlapping video images of scene 420 which has a combined field of view of more than 180° using the two imaging devices 155i, 1552. In addition, camera apparatus 150 of Fig. 6C includes two additional imaging devices 156 which are not used for the acquisition of images for creating synchronized panoramic and 3D stereo images.
In all of Figs. 6A-C, all of the respective optical axes 410 are angularly displaced from each other.
In Fig. 6D, camera apparatus 150 includes two imaging devices 155i, 1553 on one side of the camera apparatus 150, with respective optical axes indicated by lines 401r, 4013. In addition, the camera apparatus includes a third imaging device 1552 on the opposite side of the camera, with respective optical axis indicated by line 4012. The camera apparatus 150 of Fig. 6D is operative to simultaneously acquire overlapping video images of surrounding scene 421-422 which has a combined field of view of 360°. In a preferred mode of operation, video images acquired by imaging devices 155r and 1553 are used for synthesizing a 3D stereo video image for streaming, and images acquired by all three imaging devices 155i, 1552, 1553 are used for stitching a 360° panoramic video image (and in other embodiments images from the same two image devices 155i and 1553 are used for creating both the 3D stereo stream and a 180° panoramic video stream). In Fig. 6D, optical axes 401i, 4013 are substantially parallel to each other (substantially meaning within plus/minus 1 ° or within plus/minus 5°) and optical axis 4012 is in the opposing direction.
In Fig. 6E, camera apparatus 150 includes two imaging devices 155i, 1553 on one side of the camera apparatus 150, with respective optical axes indicated by lines 401i, 4013 angularly displaced from each other. In addition, the camera apparatus includes a third imaging device 1552 on the opposite side of the camera, with respective optical axis indicated by line 4012, also angularly displaced from the other two optical axes indicated by lines 401i, 4013. The camera apparatus 150 of Fig. 6E is operative to simultaneously acquire overlapping video images of surrounding scene 421-422 which has a combined field of view of 360°. In a preferred mode of operation, video images acquired by imaging devices 155i and 1553 are used for synthesizing a 3D stereo video image for streaming, and images acquired by all three imaging devices 155i, 1552, 1553 are used for stitching a 360° panoramic video image (and in other embodiments images from the same two image devices 155i and 1553 are used for creating both the 3D stereo stream and a 180° panoramic video stream).
In Fig. 6F, camera apparatus 150 includes four imaging devices 155i, 1552, 1553, 1554 with respective optical axes indicated by lines 401i, 4012, 4013, 4014 all of which are angularly displaced from each other. The camera apparatus 150 of Fig. 6F is operative to simultaneously acquire overlapping video images of surrounding scene 421-422 which has a combined field of view of 360°. In a first preferred mode of operation, video images acquired by all four imaging devices 155i, 1552, 1553, 1554 are used both for stitching a 360° panoramic video image and synthesizing a 360° 3D stereoscopic video image. In a second preferred mode of operation, images acquired by imaging devices 155i, 1552 are used for synthesizing a 180° 3D stereoscopic video image while all four imaging devices 155i, 1552, 1553, 1554 are used for stitching a 360° panoramic video image.
Figs. 6G, 6H and 61 are respective top, front and rear views of a camera apparatus 150 embodying the functionality of the camera apparatus of Fig. 6D. As can be seen in Fig. 61, a camera apparatus 150 can include a viewscreen 154.
It will be obvious to a skilled artisan that stereophonic and/or panoramic video images can be created using images acquired by any number of imaging devices. For example, at a professional sports event, dozens of imaging devices may be spread around a stadium so as to provide high-resolution imaging from many different angles.
Referring now to Fig. 7A-7E, a number of examples of client viewing devices 190 and indications of their use for viewing live synchronized panoramic and/or 3D stereoscopic video image streams are schematically illustrated.
In Fig. 7A, client viewing device 190 comprises a conventional computer display monitor adapted to receive video streams in data communication with the one or more servers 170, e.g., through wired and/or wireless internet-based connections 202, 203. A 180° scene 420 is displayed on the screen 190. An onscreen widget 95, as a non-limiting illustrative example of a client-side control element, allows a user to select between a 3D stereoscopic image stream and a 180° panoramic image stream.
As discussed earlier, in some embodiments, selecting one of the two live streams can be to determine which live stream (of panoramic and stereoscopic) is delivered to the client viewing device 190 and in other embodiments the selecting can be to determine which of the two delivered streams is to be displayed on the client viewing device 190.
In some embodiments, selecting a live stream can be through a portal. A portal can be a public portal such as, for example, YouTube, where each stream is accessible through a link to its own respective internet address. Alternatively, a portal can be accessed through a single dedicated link to an internet address, at which address the user is allowed to select one of the live streams and/or toggle between them. In other embodiments, a single portal can be accessed as a single dedicated link to an internet address, at which the user is automatically directed to one stream or the other, based on criteria discussed earlier such as client location or viewing history, or characteristics of the video content or of the client viewing device. In some embodiments, a user can choose to pay for one video stream or the other, or for both interchangeably.
In Fig. 7B, a surrounding scene 421-422 is viewed as a 360° panoramic stream on a conventional touch screen monitor, where a user can pan back and forth with her finger 90, as indicated by finger movement lines 91, to view different sections of the surrounding scene 421-422. Alternatively the panning can be done by using a pointing device such as a mouse (not shown). While it might be more‘ideal’ to watch such images on a‘virtual reality’ device, the specific panning capability illustrated in Fig. 7B can be used on conventional computer monitor screens.
In Fig. 7C, a client viewing device 190 comprises a virtual reality (VR) headset with a handheld controller 194 (which can alternatively comprise a smartphone or tablet running an appropriate app). An onscreen (i.e., on the controller screen) widget 95, as a non-limiting illustrative example of a client-side control element, allows a user to select between a 3D stereoscopic image stream and a panoramic image stream.
Figs. 7D and 7E illustrate a client viewing device 190 comprising a VR headset worn by user 90. With movements 92 of his head, the user 90 can pan or navigate the 360° viewing angle of a live video stream of the surrounding scene 421-422. As illustrated, the user 90 looks in a first direction (indicated by arrow 93) in Fig. 7D and sees scene 421, then looks in the opposing direction in Fig. 7E and sees scene 422. Toggle widget 94 appears on the VR screen 195 to enable the user 90 to toggle between the synchronized live video streams.
Referring now to Fig. 8, a flowchart of a method is presented for delivering simultaneous live video streams of panoramic and 3D stereoscopic images to a plurality of clients. The method includes four steps, as follows:
Step SOI acquiring, using a camera apparatus 150 comprising a plurality of imaging devices 155, a corresponding plurality of simultaneous live video images 111 of a scene 420, or 421-422. Any of the camera apparatuses 150 disclosed in Figs. 6A- F would be suitable for carrying out the method.
Step S02 transmitting the plurality of live video images 111 from the camera apparatus 150 to one or more remote servers 170 using communications circuitry 157 of the camera apparatus 150 and communications channel 201.
Step S03 creating, from the plurality of live video images 111, a live panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules 175, 176 that reside at the one or more servers 190. In some embodiments, creating the live panoramic video image includes using all N of the corresponding plurality of the simultaneous live video images 155, and creating the live 3D stereoscopic video image includes using a proper subset of said corresponding plurality. Step S04 simultaneously delivering, from the one or more remote servers 170 to client viewing devices 190 in communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams 180, 181, the delivering being such that: (i) the live panoramic video stream 180 is delivered to a first plurality 191 or 193 of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality 192 or 194 of client viewing devices 190. In some embodiments the two pluralities 191, 192 do not overlap, and in some embodiments the two pluralities 193, 294 overlap such that a group of client viewing devices 190 in the overlap have both live streams delivered to them. The live video streams are synchronized with each other to enable smooth or seamless switching (toggling) between them.
In some embodiments, not all steps of the method are performed.
Referring now to Fig. 9, a flowchart of another method is presented for delivering simultaneous live video streams of panoramic and 3D stereoscopic images to a plurality of clients. The method includes four steps, as follows:
Step Sll acquiring, using a camera apparatus 150 comprising a plurality of imaging devices 155, a corresponding plurality of simultaneous live video images 111 of a scene 420, or 421-422. Any of the camera apparatuses 150 disclosed in Figs. 6A- F would be suitable for carrying out the method.
Step S12 creating, from the plurality of live video images 111, a live panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules 175, 176 that reside at the camera apparatus 150. In some embodiments, creating the live panoramic video image includes using allN of the corresponding plurality of the simultaneous live video images 111, and creating the live 3D stereoscopic video image includes using a proper subset of said corresponding plurality.
Step S13 transmitting the live panoramic video image and the live 3D stereoscopic video image from the camera apparatus 150 to one or more remote servers 170 using communications circuitry 157 of the camera apparatus 150 and communications channel 201. Step S14 simultaneously delivering, from the one or more remote servers 170 to client viewing devices 190 in communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams 180, 181, the delivering being such that: (i) the live panoramic video stream 180 is delivered to a first plurality 191 or 193 of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality 192 or 194 of client viewing devices 190. In some embodiments the two pluralities 191, 192 do not overlap, and in some embodiments the two pluralities 193, 294 overlap such that a group of client viewing devices 190 in the overlap have both live streams delivered to them. The live video streams are synchronized with each other to enable smooth or seamless switching (toggling) between them.
In some embodiments, not all steps of the method are performed. Steps from multiple methods can be combined into a composite method.
In some embodiments, either of the methods can additionally include storage and retrieval steps, in which uploaded video images 111 or uploaded stitched and synthesized panoramic and stereoscopic images, respectively, are stored for later, e.g., on-demand retrieval by a client.
Reference is made to Fig. 10, where another example of a system 100 for delivering synchronized video streams of panoramic and 3D stereoscopic images is illustrated. The system 100 includes panoramic-image video module 175 and stereo image video module 176, respectively configured to deliver the 2D panoramic video image and the 3D stereoscopic video image as synchronized video streams 180, 181 to a plurality 195 of client viewing devices 190 for display thereupon. In some embodiments, the panoramic-image and stereo-image video modules 175, 176 are further configured, respectively, to stitch said panoramic video image and synthesize said 3D stereoscopic video image from a plurality of simultaneously acquired video images 111. In some embodiments, the panoramic-image and stereo-image video modules 175, 176 reside at one or more remote servers 170 that are in communication with the plurality 195 of client viewing devices 190.
The system 100 of Fig. 10 additionally comprises a toggling module 168 for toggling between the panoramic and stereoscopic display modes if supported by a client viewing device 190. The toggling can be responsive to a user selection received via a client-side interface, e.g., the toggling widget 94 of Figs. 7D and 7E. Alternatively or additionally, the toggling can be triggered by a content analysis module 169, which is preferably configured to toggle between the display modes based on content analysis. For example, the content analysis module 169 can analyze the video images (at any stage of processing), and determine that a 3D stereoscopic stream is more appropriate or that a panoramic stream is more appropriate.
We refer now to Fig. 11, which shows a block diagram illustrating another example of a system 100 for delivering synchronized live panoramic and stereoscopic video streams. As shown, system 100 can include key processing and client-interface components. In this example, the system 100 in in communication (e.g., by communications means 201) with a camera apparatus 150 for receiving video images 111. In some embodiments, the camera apparatus 150 (e.g., a user of the camera apparatus) can initiate the communication with the one or more servers 170 for uploading the live video images 111, for processing by the stitching module 175 and stereo image synthesizing module 176, and delivery of the created panoramic video stream 180 and 3D stereoscopic video stream 181 to a plurality 195 of client viewing devices 190. In other embodiments, the camera apparatus 150 can be used as a‘slave’ device, i.e., where at least one of its operations (e.g., starting to acquire images, uploading to the servers 170, etc.) is controlled from the cloud 199, i.e., the one or more servers 170. In still other embodiments, the system 100, with or without image processing (e.g., stitching and stereo synthesizing) can be offered as a commercial delivery platform for streaming 2D panoramic 180 and 3D stereo 181 streams for third parties initiating live (or delayed) multi-stream video capture, which is facilitated by establishing a communications connection between third party camera apparatus 150 and the one or more cloud-based servers 170 of the commercial delivery platform. In some embodiments, uploaded video images 111 and/or‘processed’ 2D panoramic and 3D stereo video images can be stored using non-transient storage media 140 which resides at the one or more servers 170. Communications means 201, 202 and 203 shown illustrated in Fig. 11 can use any wired or wireless technology known in the art suitable for transmission of high-resolution video images.
In some embodiments, a system 100 as disclosed herein for delivering 2D panoramic and 3D stereoscopic video streams can additionally or alternatively deliver enhanced and/or curated video streams to remote clients and remote client viewing devices.
In a first example of enhanced video streams, an alternative video stream can be created by‘following’ a selected object captured in a scene. This can include an alternative video stream in which the followed object is at the center or in the foreground, which requires a different composition than a typical content-agnostic stitched panorama or synthesized 3D image. As an example, the original live video images can be of a concert. The panoramic image stitched from the video images allow a remote client to experience the concert in 360°‘virtual reality’. The stereo image created allow a remote client to see the band in the apparent or simulated 3D created by combining images with stereo parallax, as is known in the art. While the content- agnostic results might show the lead singer upfront throughout the production, an alternative image stream can be created out of the already-extant video images, with an alternative focus, such on the band’s guitarist, or drummer. This may require, in some cases, reducing the angle-of-view of the resulting panoramic stream, or compromising on the apparent stereo‘depth’ in the 3D stream. Despite these possible compromises in the quality of the‘immersive’ experience, providing clients with an opportunity to receive such alternative video streams with object-following can greatly enhance the user experience. A 2D panoramic/3D stereoscopic delivery system 100 can provide the alternative video stream(s) in any number of ways: (i) instead of the ‘general’ or ‘default’ video streams delivered; (ii) toggleable between the‘regular’ live streams and the alternative ones; or (iii) jointly delivered, e.g., side-by-side or picture-in-picture.
Referring now to Figs. 12A and 12B, an object-following module 245 is shown residing at the one or more remote servers 170 in the cloud 199.
The block diagrams of Figs. 12A and 12B are not intended to show an exhaustive inventory of the components of a system for creating and delivering panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system. The skilled artisan will understand that the various system modules and components can be combined with any one or more of the components, features and communications arrangements of any of the systems 100 illustrated in figures accompanying this disclosure. In embodiments, the object-following module 245 includes a number of sub- modules. Visual object analysis submodule 246 can be provided for identifying a visual object and optionally offering to clients the ability to‘follow’ the visual object. In any one set of video images there can be a number of such visual objects; the example discussed earlier suggested (1) a guitarist and (2) a drummer with respect to a video stream of a concert. Aural object analysis submodule 247 can provide the same function as the visual object analysis submodule 246 but for‘sound objects.’ Multi-stream video capture can include multi-channel audio capture as well. For example, the four microphones 405 of the camera apparatus 150 of Fig. 6F can capture four-channel surround sound; other numbers of microphones can be used to provide different levels of audio resolution. The captured multi-channel audio can be processed so as to form an ambisonic soundtrack to the immersive video streams. Ambisonics, as noted by Wikipedia, is a full-sphere isotropic surround sound format. Similar to the visual object analysis submodule 246 employing image processing techniques to identify prominent or interesting visual objects, the aural object analysis submodule 247 is preferably configured to perform audio processing that can identify‘sound objects’ that can then be‘followed’ in alternative video streams. Using the previous example of the concert, it could be that the lead singer is the type who prances about the stage with a handheld microphone rather than staying center-stage; the vocal track might be identified as an aural object and thus an alternative video stream may remain focused on the source of the sound (the singer) as it moves around the stage. Of course, the source of sound need not be one that moves about, and a static source of sound (e.g., the drummer) can also be identified as an aural object for object-following. A user selection submodule 248 can be used to solicit and receive a user designation of a visual or aural following-object rather than use a following-object identified by the visual object analysis submodule 246 or the aural object analysis submodule 247. A delivery submodule 248 can be used to provide the toggling and joint display (side-by-side and picture-in-picture) functionalities described earlier.
The object-following module 245 as described herein is based on a ‘post processing’ approach. In other words, rather than the system receiving object-specific or location-specific data together with the video stream and using such information in the stitching and processing of the panoramic and stereoscopic images, the enhancing of the video streams is based on visual and/or aural analysis, and/or on client inputs (e.g., using user-selection submodule 248).
In a second example of enhanced video streams, an alternative video stream can be created by curating‘highlight’ and segments from video streams. Use of the curating module can include an alternative video stream in which, for example, only highlights of captured and/or processed video is shown. In one example, a curated video stream can be shorter in duration than the original source video images, and can include a mix of 2D panoramic and 3D stereoscopic segments. In another example, a curated video stream can be the same duration as the original source video images, and can include a mix of 2D panoramic and 3D stereoscopic segments. In a further example, a curated video stream shorter in duration than the original source video images can include either all 2D panoramic segments or all 3D stereo segments. Clients can be offered a choice to receive delivery of any one of such curated video streams in addition or instead of the‘regular’ full 2D panoramic and/or 3D stereo video streams. In embodiments in which both 2D panoramic and 3D stereo segments are included in a curated video stream, the segments typically do not overlap, i.e., the same footage is typically not included twice (in two formats) in the same curated video stream. In embodiments in which both all -2D panoramic and all-3D stereo curated streams are offered for delivery, the streams are typically synchronized such that at any point a client can toggle from one curated stream to the other without loss or duplication of segments or portions of segments.
Referring now to Figs. 13A and 13B, a curating module 255 is shown residing at the one or more remote servers 170 in the cloud 199.
The block diagrams of Figs. 13A and 13B are not intended to show an exhaustive inventory of the components of a system for creating and delivering panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system. The skilled artisan will understand that the various system modules and components can be combined with any one or more of the components, features and communications arrangements of any of the systems 100 illustrated in figures accompanying this disclosure.
In embodiments, the curating module 255 includes a number of sub-modules. Segmentation submodule 256 provides the functionality of identifying segments within either the source (originally acquired) video images or the processed (2D panoramic and/or 3D stereo) video images. The term‘segments’ as used herein can be used interchangeably with‘scenes’; machine-based techniques for identification of scenes from video images are known in the art. (The term‘segment’ was preferred here for convenience because the word‘scene’ so as to avoid confusion with the use of the word elsewhere in this disclosure in the context of e.g.,‘acquiring simultaneous images of a scene,’ which would be referring to an entire video session and not just a portion thereof.) A segment analysis submodule 257 can be provided to determine whether any given segment is preferably included in a curated stream as a 2D panoramic segment or as a 3D stereoscopic segment. This can be determined by the content and/or composition of a segment. To use once again the example of a concert, a segment heavy on the vocals might be better served by a 3D stereo conversion of the lead singer front and center. Alternatively, a segment heavy on the instrumentals might be better served by a full panoramic display. An editing submodule 258 can be provided for editing the video segments into one or more curated video streams. A delivery submodule 259 can be used for providing the delivery interface with remote client viewing devices.
Referring now to Figs. 14A and 14B, a refraining module 265 is shown residing at the one or more remote servers 170 in the cloud 199.
The block diagrams of Figs. 14A and 14B are not intended to show an exhaustive inventory of the components of a system for creating and delivering reframed panoramic and stereoscopic video streams, but rather to highlight certain aspects of the system. The term‘reframing’ as used herein means changing the aspect ratio of a video image or stream. For example, a desirable aspect ratio can be one that is compatible with display devices, such as, and not exhaustively, typical commercial video aspect ratios (width:height) of 1.78:1, 2:1, or 2.4:1. The skilled artisan will understand that the various system modules and components can be combined with any one or more of the components, features and communications arrangements of any of the systems 100 illustrated in figures accompanying this disclosure.
In embodiments, the refraining module 265 includes a number of sub-modules. Frame analysis submodule 266 provides the functionality of analyzing frames or segments within either the source (originally acquired) video images or the processed (2D panoramic and/or 3D stereo) video images. In some embodiments, this submodule is effective to analyze an image and determine an optional aspect ratio. Additionally or alternatively, this analysis can determine how the image or stream should be cropped so as to achieve a target aspect ratio. In a first example, reframing can be‘naive’ and simply retain a centered portion of an image. In a second example, the submodule may determine that the‘action’ in a scene is taking place in the left side of the original acquired video image and therefore reframing the image, non-centered, to include all of the left side and none of the right side, is preferable. A user selection submodule 265 can be used to solicit and/or receive a user preference for the aspect ratio of a reframed video stream. Alternatively or additionally, this submodule can be used to solicit and/or receive a user preference for how the video stream is to be reframed. In an example, a user is provided with an onscreen capability to select where the image or stream is cropped, e.g., centered or off to one side (or top or bottom) of the original image. A delivery submodule 268 can be used to provide the toggling functionality, i.e., to toggle between different aspect ratios, including, without limitation, the original and target aspect ratios.
It will be obvious to the skilled practitioner that the discussions of Figs. 1A, 3 and 4 with respect to the various embodiments of delivering video streams (i) to a plurality 195 of client viewing devices 190, (ii) to respective first and second pluralities 191, 192 of client viewing devices 190, and (iii) to first and second overlapping pluralities 193, 194 of client viewing devices 190 are also applicable to the delivery of enhanced, curated, alternative and/or reframed video streams as detailed in the foregoing discussions, such that enhanced, curated and alternative video streams can be combined with any such delivery embodiments (and stream selection embodiments) within the scope of the present invention.
Stream determination module 215, toggling module 168 (including content analysis module 169), object-following module 245, ambisonic processing module 250, curating module 255, and refraining module 265 are all configured to apply any relevant rules and/or other overhead inputs (e.g. image metadata) for determining the content and/or format of any given video stream to any given client viewing device 190. With respect to every one of the modules listed in the preceding sentence, rules and/or other overhead inputs are applied after the processing and stitching, etc., of acquired video images into panoramic and stereoscopic images, and are applied only for configuring the video streams delivered to the client viewing devices 190. The stitching module 175 and stereo image synthesizing module 215 do not receive rules and/or other inputs other than the acquired (and optionally dewarped) video images for the purpose of stitching and synthesizing.
Referring now to Fig. 22A, a flowchart of a method is presented for delivering simultaneous live video streams 180, 181 of 2D panoramic and 3D stereoscopic images to a plurality of clients. The method includes three steps, as follows:
Step S21 receiving, by one or more servers 170 and from a remote source, a plurality of live video streams 111 of a scene. In some embodiments, the video streams 111 have overlap therebetween.
Step S22 creating, at the remote server(s) 170, from the plurality of received live video streams 111, a live 2D panoramic video image and a live 3D stereoscopic video image of the scene using respective stitching and stereo image synthesizing modules 175, 176.
Step S23 simultaneously streaming, from the one or more remote servers 170 to client viewing devices 190 in remote communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams 180, 181, the simultaneous streaming being such that: (i) the live panoramic video stream 180 is delivered to a first plurality 191 of client viewing devices 190 and the live 3D stereoscopic video stream 181 is delivered to a second plurality 192 of client viewing devices, and (ii) the respective live video streams are synchronized. In some embodiments, the live panoramic video stream 180 is delivered instead to a first overlapping plurality 193 of client viewing devices 190 and the live 3D stereoscopic video stream 181 is delivered instead to a second overlapping plurality 194 of client viewing devices.
Referring now to Fig. 16, a flowchart of a method is presented for delivering synchronized live video streams 180, 181 of 2D panoramic and 3D stereoscopic images of a scene from one or more remote servers 170 to a plurality of client devices 190. The method includes three steps, as follows:
Step S31 stitching a live 2D panoramic video image from a plurality of live video images 111 of a scene received from a remote source. Step S32 synthesizing a live 3D stereoscopic video image from at least some of the plurality of live video images 111 of the scene.
Step S33 delivering the 2D live panoramic video image and the live 3D stereoscopic video image as synchronized live video streams 180, 181 to the plurality of client viewing devices 190.
Referring now to Fig. 17, a flowchart of a method is presented delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers 170 to a plurality of remote clients. The method includes four steps, as follows: Step S41 accessing a 2D panoramic video image and a 3D stereo video image, stitched and synthesized from a plurality of video images 111.
Step S42 selecting respective first and second pluralities of video segments from the 2D panoramic video image and the 3D stereo image.
Step S43 creating a curated video stream that comprises the first and second plurality of video images.
Step S44 delivering the curated video stream to a plurality of remote client viewing devices 190, so as to deliver both 2D panoramic video segments and 3D stereo video segments.
Referring now to Fig. 18, a flowchart of a method is presented for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers 170 to a plurality 195 of remote clients 190. The method includes five steps, as follows:
Step S51 accessing a plurality of simultaneously-captured video images 111 of a scene, having a first duration. Step S52 selecting video segments from within the video images, each corresponding to a different portion of the first duration.
Step S53 for each selected video segment, either stitching a 2D panoramic video segment or synthesizing a 3D stereo video segment, based on an analysis of the segment’s content and/or composition. Step S54 creating a curated video stream that comprises the stitched 2D panoramic video segments and the synthesized 3D stereo images.
Step S55 delivering the curated video stream to a plurality 195 of remote client viewing devices 190. Referring now to Fig. 19, a flowchart of a method is presented for delivering an enhanced video stream. The method includes two steps, as follows:
Step S61 acquiring a plurality of simultaneously-captured video images of a scene.
Step S62 delivering a live video stream comprising multiple time-differentiated segments, each segment comprising either a 2D panoramic video stream or 3D stereo video stream.
According to the method, the selection of which of the 2D panoramic video stream and the 3D stereo video stream is included/synopsized in any given segment is based on the content and/or composition of the video images. The acquiring and delivering of Steps S61 and S62 are done in real time so as to deliver a live video stream.
Referring now to Fig. 20, a flowchart of a method is presented for delivering curated video streams including panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients. The method includes five steps, as follows:
Step S71 accessing a plurality of simultaneously-captured, video images 111 of a scene, the video images 111 having a first duration.
Step S72 selecting video segments from within the video images 111, each segment corresponding to a different portion of the first duration.
Step S73 for each selected video segment, both stitching a corresponding 2D panoramic video segment and synthesizing a corresponding 3D stereo video segment. Step S74 creating a curated 2D panoramic video stream from the stitched 2D panoramic video segments, and a curated 3D stereo video stream from the synthesized 3D stereo video segments.
Step S75 delivering the curated 2D panoramic video stream to a first plurality 191 of remote client viewing devices 190, and the curated 3D stereo video stream to a second plurality 192 of remote client viewing devices 190. In some embodiments, the curated panoramic video stream 180 is delivered instead to a first overlapping plurality 193 of client viewing devices 190 and the curated 3D stereoscopic video stream 181 is delivered instead to a second overlapping plurality 194 of client viewing devices.
Referring now to Fig. 21, a flowchart of a method is presented for delivering an enhanced video stream. The method includes four steps, as follows:
Step S81 acquiring a plurality of simultaneously-captured, video images 111 of a scene.
Step S82 initiating the delivery of a video stream, such that the video stream delivered to each remote client is either a 2D panoramic video stream stitched from the plurality of video images 111 or a 3D stereoscopic video stream synthesized from the plurality of video images 111.
Step S83 determining a following-object in the delivered video stream, including at least one of a person, thing or other visual element captured visually in at least one of the video images.
Step S84 delivering an alternative video stream to at least one remote client device 190 based on the determination of the following-object.
Referring now to Fig. 22, a flowchart of a method is presented for delivering an enhanced video stream. The method includes four steps, as follows:
Step S91 acquiring a plurality of simultaneously-captured, video images 111 of a scene, including spatial audio captured by a plurality of microphones 405.
Step S92 initiating the delivery of a video stream, such that the video stream delivered to each remote client is either a 2D panoramic video stream stitched from the plurality of video images 111 and comprising ambisonic audio processed from the captured spatial audio or a 3D stereoscopic video stream synthesized from the plurality of video images 111 and comprising ambisonic audio processed from the captured spatial audio.
Step S93 determining a following-object in the delivered video stream, including a spatially-differentiated aural element captured in at least one of the video images. Step S94 delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
Referring now to Fig. 23, a flowchart of a method is presented for delivering simultaneous video streams 180, 181 of 2D panoramic and 3D stereoscopic images to a plurality of clients. The method includes five steps, as follows:
Step S101 acquiring, by one or more servers 170 (e.g., in the cloud 199), a plurality of video streams 111 of a scene, the video streams having overlap therebetween.
Step S102 at the one or more remote servers 170, creating, from the plurality of overlapping video streams 111 of the scene, a 2D panoramic video stream, and a 3D stereoscopic video stream, using respective stitching and stereo image synthesizing modules 175, 176.
Step S103 initiating the synchronized delivery, from the one or more remote servers 170 to a plurality 195 of client devices 190, a live video stream comprising one of the live 2D panoramic video stream 180 and the live 3D stereoscopic video stream
181.
Step S104 receiving, at the one or more remote servers 170, a toggle command from a client.
Step S105 toggling the video stream delivered to the client so as to either replace the live 2D panoramic video stream 180 with the live 3D stereo video stream 181, or to replace the live 3D stereo video stream 181 with the live 2D panoramic video stream
180.
Any of the methods and method steps illustrated in the flowcharts of Figs. 8, 9 and 15-23 can be combined in a useful way and such combinations are wholly within the scope of the present invention. In some embodiments, not all steps of any given method are performed.
According to some embodiments of the invention, a camera apparatus comprising multiple imaging devices can acquire simultaneous video images of a scene. Each of the imaging devices can have a wide angle of view, for example 180° or more. A scene can encompass a field of view of 180° or more, or 360°. The simultaneous video images can be overlapping, meaning that at least two of the video images overlap or, equivalently, at least abut or nearly abut. In some embodiments one of the video images can overlap with each of the two other video images. In some embodiments the angles of view captured by one or more of the respective imaging devices is exactly 180°, or alternatively the angles of view of two of the imaging devices add up to exactly 360°, and in such cases the concept of ‘overlapping’ as used herein is expanded to include adjacent or abutting images. In some embodiments the angles of view captured by one or more of the respective imaging devices is less than 180°, or alternatively the angles of view of at least some of the respective imaging devices do not overlap, and in such cases the concept of‘stitching’ as used herein is expanded to include interpolating such that images can be considered‘stitched’ even though there is no actual overlap between them.
In some embodiments, the camera apparatus can include communications circuitry configured for uploading the overlapping video images to the one or more servers, and in other embodiments, the communications equipment is configured for uploading stitched panoramic images and synthesized 3D stereoscopic images to the one or more servers. Any reference herein to communications circuitry can include integral or‘built-in’ communications circuitry, or to modules that are easily attachable to a camera apparatus, e.g., where the camera apparatus is equipped with an appropriate connector.
We now refer to Figures 24A, 24B, 24C and 24D, which show, respectively, front, rear, bottom and top projection views of a camera apparatus 500 capable of acquiring multiple simultaneous video images from which 180° stereo and 360° panoramic video streams (or files) can be created.
A camera apparatus 500 according to embodiments is preferably constructed as a‘unibody’ device with a rigid frame 501. Unibody in this context means that the device has no moving parts for positioning/repositioning individual imaging devices, and no folding mechanisms that place one or more individual imaging devices in a position more or less suitable for imaging in a specific mode (stereo or panoramic). The rigid frame 501 can of course be constructed from multiple components, and it can also have openings with moveable ore removeable covers for connections and controls. The camera apparatus 500 illustrated in Figs. 24A-24D has two‘major faces’ - the front and rear faces 540i, 5402 - and four minor faces, which include the top 546 and the bottom 545. The shape of the camera apparatus 500 in all of the figures consistently show, for clarity purposes, a single example of a product design, but the product can have any rectangular/prismatic shape (including, for example, rectangles with or without cut or rounded corners, or with rounded sides or edges). Nonetheless, it can be desirable to provide a bottom face 545 that allows the camera apparatus 500 to be placed on a flat surface unattended without support, whether for storage or for operation. Camera apparatus 500 comprises three imaging devices 520. A dual-mode device having only two imaging devices would necessarily require mechanical manipulation of one or both imaging devices in order to orient them properly for a chosen mode. A device may have more than three imaging devices, but three is the practical minimum number of imaging devices for having dual mode capabilities in a rigid design. Each one of the imaging devices 520 is a wide-angle camera, meaning that it has a respective wide-angle lens 530 installed thereupon. A wide-angle lens 530 can comprise a fisheye lens, which advantageously has an angle-of-view of at least 180°, i.e., a hemispheric or greater angle-or-view. All three imaging devices 520 can be operable to acquire images at least when the camera is in the orientation shown in Figs. 24A-D.
With respect to the embodiments disclosed herein, a camera apparatus 500 and respective imaging devices 520 are an example of, and are respectively equivalent to, the camera apparatus 150 and respective imaging devices 155 discussed hereinabove. This equivalence includes, without limitation, features and functionalities. A first imaging device 520i is installed on a first major face 540i while the other two imaging devices 5202 and 520r are installed on the second major face 5402. First and second imaging devices 520i and 5202 point in opposite directions and are therefore useful for jointly and simultaneously capturing images of a 360° scene that can be stitched into 360° panoramic images. Each of these two imaging devices 520i, 5202 has a respective optical axis 401i, 4012 which are: directly opposite each other and therefore collinear; or in exactly opposing directions but not collinear, and therefore parallel; or in nearly opposing directions, meaning within ±10° of rotation from being opposite, or within ±5° of rotation from being opposite, or within ±1° of rotation from being opposite. The terminology of being‘within ± x°’ of opposite or parallel indicates a range, meaning anywhere in the range from -x° to +x° away from being opposite or parallel, as the case may be, and in every case means‘from -x° to +x° inclusive’.
A display screen 510 is installed on the first major face 540i. The display screen 510 can be flush with the surface of the first major face 540i, or slightly sunken, or otherwise displaced from being flush based on aesthetic or ergonomic considerations. It is preferably a touchscreen so as to facilitate receiving user inputs. Regardless of whether display screen 510 is a touchscreen, a display-control interface 515 can be provided on the first major face 540i. Display-control interface 515 can be, as per the example illustrated in Fig. 24B, a button. The display screen 510 is operable to display and/or play back 180° or 360° images in response to an input received via either the display screen 510 (if a touchscreen) or the display-control interface 515. A video display module may include a display screen 510, display-control interface 515, and any required electronics or software for performing the video-display functions.
The first major face 540i additionally includes an electronic status indicator 516i which can be operable to show on/off status, or to indicate that filming or recording or transmitting is in progress. While illustrated in Fig. 24B as a simple LED display, the electronic status indicator 516i can include any kind of electronic display such as, for example, a small LCD screen, or multiple LED’s. Either of the two major faces 540i, 5402 may be flat or rounded, and/or may optionally include raised or sunken surfaces for aesthetic or ergonomic purposes. The first major face 540i is identified in the figures as the‘front’ of the camera apparatus 500 (and second major face 5402 as the‘rear’ face), but this is only for convenience and it has no meaning with respect to design or function, and especially in light of the fact that the camera apparatus 500 is operative to acquire images in any and all directions.
Two imaging devices 5202, 5203 are installed on the second major face 5402. A second electronic status indicator 5162 can also be installed on the second major face 5402. The second and third imaging devices 5202, 5203 face in the same direction as each other, meaning that their respective optical axes 4012, 4013 are parallel to each other, or within ±5° of parallel or within ±2° of parallel or within ±1° of parallel. Images acquired simultaneously by the second and third imaging devices 5202, 5203 can be used together in synthesizing stereo video images, e.g., 180° stereo video images for viewing on a 3D viewing device.
The second and third imaging devices 5202, 52 3 are preferably installed at the same height HLENS as each other, where height HLENS is the height in the y-dimension to the centerline CLLENS as shown in Fig. 24A, as measured in the y-dimension when the camera apparatus 500 is placed on a flat surface, i.e., with bottom face 545 on a flat surface. The first imaging device 520i is also preferably installed at the same height HLENS as shown in Fig. 24B, and in particular at the same height of the second imaging device 5202 with which it is‘back-to-back’. The expression“same height” can mean that the imaging devices 520 are exactly at the same height, or have respective heights that are different from each other by no more than 50% of a diameter 908 of any one of the three respective lenses 530i, 5302, 5303 (or the smallest of the respective diameters 908 if the lenses 530 don’t all have the same diameter), or by no more than 20% of the diameter 908, or no more than 2 millimeters, or no more than 1 millimeter.
Fig. 24D shows an example of a top face 546 of the camera apparatus 500. Top face 546 is ergonomically enhanced with a concave, textured finger-grip area 547. Finger-grip area 547 preferably extends in each direction across of a majority of the top face 546 so as to provide as substantial an area as possible on what is generally a small device. To improve a user’s grip, the finger-grip area 547 can be at least partly textured, as illustrated, with area or areas that are less smooth to the touch than the surrounding surface. The texture pattern illustrated in Fig. 24D is non-limiting and merely shows one possible example. In some embodiments which are not illustrated, the finger-grip area can be flat or convex, or concave but untextured.
Fig. 24C is a bottom view of the camera apparatus 500, showing a mechanical support-stand interface 550 installed on the bottom face 545, one of the minor faces of the camera apparatus 500. The interface 550 can include, for example, a threaded socket for attaching a tripod with a matching threaded connecting element. According to embodiments, it can be desirable for the support interface 550 to be closer to the second and third imaging devices 5202, 5203 than to the first imaging device 520i. In some embodiments, the interface 550 is preferably installed along a longitudinal centerline CLx which represents the‘middle’ of the camera apparatus 500 in the -dimension. The ‘middle’ can be a line drawn along a central longitudinal axis of the bottom face 545, or can be a line indicating a center of balance that takes into account the weight of the various elements installed on both major faces 540 of the rigid body 501 of the camera apparatus 500. It can be desirable for the interface 550 to be positioned along the x- dimension centerline CLx and centered on the optical axes 401i, 4012 of the first and second imaging devices 520i, 5202, as is illustrated in Fig. 24C. In an example, this position of the interface 550 provides the greatest angular stability for capturing image for a 360° panorama, i.e., ensuring that the two imaging devices 520i, 5202 tasked with acquiring the images for the 360° are held stably at an optimal angle for capturing images.
In some embodiments, the interface 550 is not located exactly at the intersection of the centerline CLx and optical axes 401i, 4012, but is within the footprint of the two imaging devices 520i, 5202. Fig. 25 shows an orthogonal projection 590 of first and second imaging devices 520i, 5202 (including respective lenses 530i, 5302) on the bottom face 545 of the camera apparatus 500. The orthogonal projection 590 is a projection of the combined footprint of the two imaging devices 520i, 5202 on the bottom face 545. The mechanical support-stand interface 550, or at least a majority of the interface 550, is preferably located within the orthogonal projection 590. As described in the previous paragraph, it can be desirable for the interface 550 to be substantially (±3 mm, or ±2 mm, or ±1 mm) centered around the‘intersection’ of the v-dimcnsion centerline CLx with the optical axes 401i, 4012, which are represented in Fig. 25 as optical centerline CLOPT.
The internal workings of a digital camera apparatus can include a planar array of photodetectors. Such a planar array can define a respective photodetector plane for each imaging device. This is illustrated in Fig. 26, where first, second and third imaging devices 520i, 5202, 5203 are associated with respective photodetector planes Pi, P2, P3· As discussed hereinabove with respect to Fig. 24C, the optical axes 401i, 4012 of the first and second imaging devices 520i, 5202 are directly opposite to each other (collinear), or within ±10° of rotation from being opposite to each other, or within ±5° of rotation from being opposite to each other, or within ±1° of rotation from being opposite to each other. Each imaging device’s respective optical axis 401 is perpendicular to the given camera’s photodetector plane P. Therefore, the photodetector planes Pi, P2 associated with the first and second imaging devices 520i, 5202 are correspondingly parallel to each other or within ±10° of being parallel, or within ±5°, or within ±1 °. As also discussed above with respect to Fig. 24C, respective optical axes 4012, 4013 of the second and third imaging devices 5202, 5203 are parallel to each other, or within ±5° of parallel or within ±2° of parallel or within ±1° of parallel, and therefore the respective photodetector planes P2, P3 are parallel to each other, or within ±5° of parallel or within ±2° of parallel or within ±1° of parallel. According to embodiments, respective photodetector planes P2, P3 of the second and third imaging devices 5202, 5203 are not co-planar, as the second imaging device 5202 extends further from rigid body 501 than does the third imaging device 5203. In embodiments, the second imaging device 5202 extends at least 2 mm and no more than 10 mm further from the second major face 5402 of the rigid body 501 than does the third imaging device 5203, or at least 3 mm and not more than 9 mm, or at least 4 mm and not more than 8 mm.
As illustrated schematically in Fig. 27, each imaging device 520 has an angle of view a limited by its optics such as, for example, its respective wide-angle lens 530. The second imaging device 5202 preferably has an angle of view
Figure imgf000059_0001
of at least 180°. A desirable outcome of the photodetector planes P2, P3 associated with the second and third imaging devices 5202, 5203 not being co-planar, and of the second imaging device 5202 extending further from the second face 5402 of the rigid body 501 than does the third imaging device 5203, is that the third imaging device 5203 and its wide-angle lens 5303 are outside the (at least) 180° angle of view
Figure imgf000059_0002
of the second imaging device 5202. In Fig. 27, the dashed line marked AOV2 (signifying the limits of the angle-of-view 0.2 of the second imaging device 5202) can be seen to pass above the wide-angle lens 5303 of the third imaging device 5203. On the other major face 540i, the first imaging device 520i also has an angle-of-view ai of at least 180° and the corresponding AOVi line passes well away from the display screen 510 and any other features of first major face 540i. The resulting effect is that panoramic images acquired using the first and second imaging devices 520i, 5202 can be produced without capturing any extraneous parts of the camera apparatus 500 as per the examples described in this paragraph.
In embodiments, the camera apparatus 500 can provide, inter alia, the following image-capture modes: (i) using only the first and second imaging devices 520i, 5202 for capturing images to be used in creating (stitching) 360° panoramic images; (ii) using only the second and third imaging devices 5202, 5203 for capturing images to be used in creating 180° 3D stereo images; and (iii) using all three imaging devices 520i .. 5203 for capturing images to be used for creating synchronized panoramic and stereo images, where images acquired by the first and second imaging devices 520i, 5202 are used for creating the panoramic images, and images acquired by the second and third imaging devices 5202, 5203 can be simultaneously used for creating the stereo images. The image-capture mode can be user-selected using the display screen 510 (if a touchscreen) and/or the display-control interface 515.
The present invention has been described using detailed descriptions of embodiments thereof that are provided by way of example and are not intended to limit the scope of the invention. The described embodiments comprise different features, not all of which are required in all embodiments of the invention. Some embodiments of the present invention utilize only some of the features or possible combinations of the features. Variations of embodiments of the present invention that are described and embodiments of the present invention comprising different combinations of features noted in the described embodiments will occur to persons skilled in the art to which the invention pertains.
In the description and claims of the present disclosure, each of the verbs, "comprise", "include" and "have", and conjugates thereof, are used to indicate that the object or objects of the verb are not necessarily a complete listing of members, components, elements or parts of the subject or subjects of the verb. As used herein, the singular form "a", "an" and "the" include plural references unless the context clearly dictates otherwise. For example, the term "a marking" or "at least one marking" may include a plurality of markings.

Claims

1. A method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of remote clients, the method comprising: a. acquiring, using a camera apparatus comprising at least three imaging devices, a corresponding number of simultaneous live video images of a scene; b. transmitting the simultaneous live video images from the camera apparatus to one or more remote servers; c. creating, from the simultaneous live video images, a live 2D panoramic video image and a live 3D stereoscopic video image using respective stitching and stereo image synthesizing modules, wherein (i) at least one of the
simultaneous live video images is used in the creating of both the panoramic video image and the stereoscopic video image, (ii) at least one of the simultaneous live video images is used in the creating of the panoramic video image and not of the stereoscopic video image, and (iii) at least one of the simultaneous live video images is used in the creating of the stereoscopic image and not of the panoramic video image; and d. simultaneously delivering, from the one or more remote servers to client viewing devices in communication therewith, the live 2D panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live 2D panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the respective live video streams are synchronized.
2. The method of claim 1, wherein an operation of the camera apparatus is controlled from the one or more remote servers.
3. The method of either one of claims 1 or 2, wherein the inclusion of any given client viewing device in either or both of the first or second plurality of client viewing devices is based upon a user selection.
4. The method of any one of the preceding claims, wherein the user selection is toggleable during the delivering of the live video streams.
5. The method of any one of the preceding claims, wherein the transmitting or
receiving, and the creating and delivering are performed in real time.
6. A method of delivering curated video streams including 2D panoramic and 3D stereoscopic images from one or more servers to a plurality of remote clients, the method comprising: a. accessing a 2D panoramic video image and a 3D stereo video image,
respectively stitched and synthesized from a plurality of simultaneously- captured video images of a scene, said video images of the scene having a first duration; b. selecting from within the stitched 2D panoramic video image a first plurality of video segments for displaying in 2D panoramic mode and from within the synthesized 3D stereo video image a second plurality of video segments for displaying in 3D stereoscopic mode, wherein the selecting of each respective video segment includes analyzing at least one of its content and its composition, each one of the video segments corresponding to a different portion of the first duration; c. creating a curated video stream that comprises the first and second plurality of video images; and d. delivering the curated video stream, from the one or more servers to a plurality of remote client viewing devices in remote communication therewith, so as to deliver both 2D panoramic video segments and 3D stereo video segments to the client viewing devices.
7. The method of claim 6, wherein the plurality of video images of the scene are captured by imagers having respective optical axes angularly-displaced and/or laterally displaced from each other.
8. A method of delivering curated video streams including panoramic and 3D
stereoscopic images from one or more servers to a plurality of remote clients, the method comprising: a. accessing a plurality of simultaneously-captured, video images of a scene, said video images of the scene a first duration; b. selecting a plurality of video segments from within said video images of the scene, each one of the video segments corresponding to a different portion of the first duration; c. for each selected video segment, performing one video-creation action
selected from (i) stitching a 2D panoramic video segment and (ii) synthesizing a 3D stereo video segment, wherein the determination of which video-creation action is performed for each respective selected video segment is based on an analysis of at least one of the segment’ s content and the segment’ s composition; d. creating a curated video stream that comprises the stitched 2D panoramic video segments and the synthesized 3D stereo images; and e. delivering the curated video stream, from the one or more servers to a plurality of remote client viewing devices in remote communication therewith.
9. The method of claim 8, wherein the plurality of video images of the scene are captured by imagers having respective optical axes angularly-displaced and/or laterally displaced from each other.
10. A method for delivering an enhanced video stream, comprising: a. acquiring a plurality of video images of a scene, said video images being simultaneously captured by imagers having angularly-displaced and/or laterally displaced optical axes; and b. delivering, to a plurality of remote client devices, a video stream comprising multiple time-differentiated segments, each segment comprising one of (i) a 2D panoramic video stream stitched from the plurality of video images captured during the period of time corresponding to the respective segment, and (ii) a 3D stereo video stream synthesized from at least some of the plurality of video images captured during the period of time corresponding to the respective segment, wherein the selection of which of the 2D panoramic video stream and the 3D stereo video stream is included in any given segment is based on the content and/or composition of the video images during the period of time corresponding to the respective segment.
11. The method of claim 10, wherein the acquiring and delivering are done in real time such that the delivered video stream is a live video stream.
12. A method of delivering curated video streams including panoramic and 3D
stereoscopic images from one or more servers to a plurality of remote clients, the method comprising: a. accessing a plurality of simultaneously-captured video images of a scene, said video images of the scene a first duration; b. selecting a plurality of video segments from within said video images, each one of the video segments corresponding to a different portion of the first duration; c. for each selected video segment, (i) stitching a corresponding 2D panoramic video segment and (ii) synthesizing a corresponding 3D stereo video segment; d. creating a curated 2D panoramic video stream that comprises the stitched 2D panoramic video segments, and a curated 3D stereo video stream that comprises the synthesized 3D stereo video segments; and e. delivering, from the one or more remote servers to client viewing devices in remote communication therewith, the curated 2D panoramic video stream and the curated 3D stereoscopic video stream, the delivering being such that: (i) the curated 2D panoramic video stream is delivered to a first plurality of client viewing devices and the curated 3D stereoscopic video stream is delivered to a second plurality of client viewing devices.
13. The method of claim 12, wherein a user selection determines which one of the two live video streams is delivered to the respective client viewing device and/or is displayed on the respective client viewing device.
14. The method of either one of claims 12 or 13, wherein the user selection is received during the delivering of the curated video streams.
15. The method of any one of claims 11 to 14, wherein the user selection is toggleable during the delivering of the live video streams.
16. A method for delivering an enhanced video stream, comprising:
a. acquiring a plurality of simultaneously-captured, video images of a scene; b. initiating the delivery of a video stream from one or more servers to a plurality of remote client devices, such that the video stream delivered to each remote client is one of (i) a 2D panoramic video stream stitched from the plurality of video images and (ii) a 3D stereoscopic video stream synthesized from the plurality of video images; c. determining a following-object in the video stream delivered to each remote client, the following-object including at least one of a person, thing or other visual element captured visually in at least one of the video images; and d. delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
17. A method for delivering an enhanced video stream, comprising: a. acquiring a plurality of simultaneously-captured video images of a scene, the video images including spatial audio captured by a plurality of microphones; b. initiating the delivery of a video stream from one or more servers to a plurality of remote client devices, such that the video stream delivered to each remote client is one of (i) a 2D panoramic video stream stitched from the plurality of video images and comprising ambisonic audio processed from the captured spatial audio, and (ii) a 3D stereoscopic video stream synthesized from the plurality of video images and comprising ambisonic audio processed from the captured spatial audio; c. determining a following-object in the video stream delivered to each remote client, the following object including a spatially-differentiated aural element captured in at least one of the video images; and
d. delivering an alternative video stream to at least one remote client device based on the determination of the following-object.
18. The method of either one of claims 16 or 17, wherein the determining of the follow-object includes receiving a user input about the following-object at the one or more servers.
19. The method of either one of claims 16 or 17, wherein the determining of the
follow-object includes analyzing at least one of (i) the video images, (ii) the 2D panoramic video stream and (iii) the 3D stereo video stream.
20. The method of claim 16, wherein the determining of the follow-object uses
stereoscopic parallax to identify an object.
21. The method of any one of claims 16 to 20, wherein the delivery of the alternative video stream preempts the delivery of the video stream delivered to each remote client responsively to receipt of a user command or selection.
22. The method of any one of claims 16 to 21, wherein the video stream delivered to each remote client is toggleably switchable with the alternative video stream.
23. The method of any one of claims 16 to 22, wherein the video stream delivered to each remote client is delivered together with the alternative video stream such that the alternative video stream is displayable side-by-side or picture-in-picture together with video stream delivered to each remote client.
24. The method of any one of claims 16 to 23, wherein the video streams are live and synchronized with each other.
25. A method for delivering a reframed video stream, comprising: a. accessing a plurality of simultaneously-captured video images of a scene; b. creating, from the plurality of simultaneous live video images, a live 2D
panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules, the 2D panoramic video image and the 3D stereoscopic video image having respective first and second aspect ratios; and c. delivering, from the one or more remote servers to client viewing devices in communication therewith, the live 2D panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) at least one of the 2D panoramic video stream and the 3D stereoscopic video stream is reframed for delivery in a third aspect ratio that is different than the first and second aspect ratios.
26. The method of claim 25, wherein the respective live video streams are delivered simultaneously and are synchronized.
27. The method of either one of claims 25 or 26, wherein the accessing includes
accessing at least three simultaneously-captured video images of the scene, at least one of which is used in both the panoramic stream and in the stereoscopic stream, at least one of which is used in the panoramic stream and not in the stereoscopic stream, and at least one of which is used in the stereoscopic stream and not in the panoramic stream.
28. A method of delivering simultaneous live video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients, the method comprising: a. receiving, by at least one server and from a remote source, a plurality of
simultaneous live video streams of a scene; b. at the at least one server, creating, from the plurality of received simultaneous live video streams, a live 2D panoramic video image and a live 3D
stereoscopic video image of the scene using respective stitching and stereo image synthesizing modules; and c. simultaneously streaming, from the one or more servers to client viewing devices in remote communication therewith, the live 2D panoramic video image and the live 3D stereoscopic video image as respective live video streams, the simultaneous streaming being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the streamed respective live video streams are synchronized.
29. The method of claim 28, wherein each live video stream is associated with a different respective view direction, the view directions and angles of view of the video streams providing overlap therebetween.
30. The method of either one of claims 28 or 29, wherein (i) the 3D stereoscopic video is unstitched so that a left video stream thereof is from a first of the live video streams and a right video stream thereof is from a second of the live video streams, and (ii) the 2D panoramic image comprises a stitching between the first and second live video streams.
31. A method of delivering simultaneous live video streams of 360° 2D panoramic and 180° 3D stereoscopic images to a plurality of remote clients, the method comprising: a. acquiring, using a camera apparatus comprising a plurality of imaging devices, a corresponding plurality of simultaneous live video images of a scene; b. transmitting the plurality of simultaneous live video images from the camera apparatus to one or more remote servers;
c. creating, from the plurality of simultaneous live video images, a live 2D
panoramic video image and a live 3D stereoscopic video image, using respective stitching and stereo image synthesizing modules; and d. simultaneously delivering, from the one or more remote servers to client viewing devices in communication therewith, the live panoramic video image and the live 3D stereoscopic video image as respective live video streams, the delivering being such that: (i) the live panoramic video stream is delivered to a first plurality of client viewing devices and the live 3D stereoscopic video stream is delivered to a second plurality of client viewing devices, and (ii) the respective live video streams are synchronized.
32. The method of claim 31, wherein an operation of the camera apparatus is
controlled from the one or more remote servers.
33. The method of either one of claims 31 or 32, wherein (i) more of the simultaneous live video images are used in creating the live panoramic video image than in creating the live 3D stereoscopic video image and (ii) at least the simultaneous live video images used in creating the 3D stereoscopic video image have an overlap therebetween.
34. The method of any one of claims 31 to 33, wherein the plurality of imaging
devices has a corresponding plurality of respective optical axes, of which at least a first two are substantially parallel to each other and at least one other one is parallel to but in an opposite direction to the at least first two.
35. The method of any one of claims 31 to 34, wherein the transmitting or receiving, and the creating and delivering are performed in real time.
36. A system for delivering synchronized live video streams of panoramic and 3D stereoscopic images of a scene, the system comprising: a. respective stitching and stereo image synthesizing modules configured to create, from a plurality of live video images of a scene, a live panoramic video image and a live 3D stereoscopic video image; and a delivery module configured to deliver, from one or more remote servers, the live panoramic video image and the live 3D stereoscopic video image as synchronized live video streams to a plurality of client viewing devices in communication with the one or more remote servers.
37. The system of claim 36, additionally configured to control the camera apparatus remotely from the one or more remote servers.
38. The system of either one of claims 36 or 37, additionally comprising a dewarping module residing at the one or more remote servers for dewarping each of the plurality of simultaneous live images.
39. The system of any one of claims 36 to 38, configured to deliver said synchronized live video streams of panoramic and 3D stereoscopic images in real time.
40. A method for delivering synchronized live video streams of 2D panoramic and 3D stereoscopic images of a scene from one or more remote servers to a plurality of client devices, the method comprising: a. stitching a live 2D panoramic video image from a plurality of live video
images of a scene received from a remote source; b. synthesizing a live 3D stereoscopic video image from at least some of the plurality of live video images of the scene; and c. delivering the live 2D panoramic video image and the live 3D stereoscopic video image as synchronized live video streams to the plurality of client viewing devices
41. A method of delivering simultaneous video streams of 2D panoramic and 3D stereoscopic images to a plurality of clients, the method comprising: a. acquiring, by one or more servers, a plurality of video streams of a scene, the video streams having overlap therebetween; b. at the one or more remote servers, creating, from the plurality of overlapping video streams of the scene, a 2D panoramic video stream, and a 3D stereoscopic video stream, using respective stitching and stereo image synthesizing modules; c. initiating the synchronized delivery, from the one or more remote servers to a plurality of client devices, a live video stream comprising one of the live 2D panoramic video stream and the live 3D stereoscopic video stream; d. receiving, at the one or more remote servers, a toggle command from a client; and e. toggling the video stream delivered to the client so as to (i) replace the live 2D panoramic video stream with the live 3D stereo video stream, or (ii) to replace the live 3D stereo video stream with the live 2D panoramic video stream.
42. The method of claim 41, wherein the 2D panoramic video stream has a 360° angle of view and the 3D stereo video stream has a 180° angle of view.
43. The method of either one of claims 41 or 42, wherein the 3D stereoscopic video is unstitched so that a left video stream thereof is from a first of the overlapping video streams and a right video stream thereof is from a second of the overlapping video streams, the 2D panoramic image comprising a stitching between the first and second live video streams.
44. An imaging apparatus for acquiring multiple simultaneous video images for creating therefrom 180° stereo and 360° panoramic video streams, the imaging apparatus comprising: a. a rigid body comprising first and second major faces; and b. first, second and third cameras fixedly installed in the rigid body, each
camera comprising a respective ultra wide-angle lens and a respective planar array of photodetectors defining a respective photodetector plane, each camera defining a respective optical axis orthogonal to said respective photodetector plane, wherein: i. the first camera is disposed such that its lens extends from the first major face and its optical axis defines a first direction, the second camera is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and the third camera is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto, ii. the second camera has an angle of view of at least 180°, and iii. a 180° concentric portion of the second-camera angle of view does not encompass any of the lens of the third camera.
45. The imaging apparatus of claim 44, wherein the angle of view of the second
camera does not encompass any of the lens of the third camera.
46. The imaging apparatus of either one of claims 44 or 45, wherein the photodetector plane of the second camera is not coplanar with, and is displaced further outward from the second major face than, the photodetector plane of the third camera.
47. The imaging apparatus of claim 46, wherein the photodetector plane of the second camera is displaced further outward from the second major face than the photodetector plane of the third camera by at least 3 millimeters.
48. The imaging apparatus of any one of claims 44 to 47, configured to acquire multiple simultaneous video images in a first mode in which the first and second cameras acquire the multiple simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third cameras acquire the multiple simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third cameras acquire the multiple simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via a display-control interface.
49. An imaging apparatus for acquiring multiple simultaneous video images for
creating 180° stereo and 360° panoramic video streams, the imaging apparatus comprising: a. a rigid body comprising first and second major faces; b. first, second and third cameras fixedly installed in the rigid body, each camera comprising a respective ultra wide-angle lens and defining a respective optical axis, wherein:
i. the first camera is disposed such that its lens extends from the first major face and its optical axis defines a first direction, ii. the second camera is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and
iii. the third camera is disposed such that its lens extends from the second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto; c. a video-display module comprising a display screen installed on the first major face, the display screen being operable to display and/or play back 180° or 360° images in response to an input received via the display screen and/or via a display -control interface; and d. a mechanical support-stand interface installed on a first minor face of the rigid body, disposed closer to each of the first and second cameras than to the third camera.
50. The imaging apparatus of claim 49, wherein a majority of the footprint of the mechanical support-stand interface on the first minor face falls within the borders of an orthographic projection of the first and second cameras on the first minor face.
51. The imaging apparatus of either one of claims 49 or 50, wherein all of the
footprint of the mechanical support-stand interface on the first minor face falls within the borders of an orthographic projection of the first and second cameras on the first minor face.
52. The imaging apparatus of any one of claims 49 to 51 , additionally comprising a first electronic status indicator on the first major face and a second electronic status indicator on the second major face.
53. The imaging apparatus of any one of claims 49 to 52, configured to acquire
multiple simultaneous video images in a first mode in which the first and second cameras acquire the multiple simultaneous video images for stitching into a 360° panoramic video image, in a second mode in which the second and third cameras acquire the multiple simultaneous video images for synthesizing into a 180° stereo video image, or in a third mode in which the first, second and third cameras acquire the multiple simultaneous video images for creating synchronized 180° stereo and 360° panoramic video streams, the configuration being set in response to an input received via the display screen and/or via the display-control interface.
54. An imaging apparatus for acquiring multiple simultaneous video images for
creating 180° stereo and 360° panoramic video streams, the imaging apparatus comprising: a. a rigid body comprising first and second major faces and a bottom face characterized by the presence thereupon of a mechanical support-stand interface; b. first, second and third cameras fixedly installed in the rigid body, each camera comprising a respective ultra wide-angle lens and defining a respective optical axis, wherein: i. the first camera is disposed such that its lens extends from the first major face and its optical axis defines a first direction, ii. the second camera is disposed such that its lens extends from the second major face and its optical axis defines a second direction that is opposite to the first direction or no more than 5 degrees from being opposite thereto, and iii. the third camera is disposed such that its lens extends from the
second major face and its optical axis defines a third direction that is parallel to the second direction or within 10 degrees of being parallel thereto; and c. a video-display module comprising a display screen installed on the first major face, the display screen being operable to display and/or play back 180° or 360° images in response to an input received via the display screen and/or via a display-control interface,
wherein: i. the bottom face defines an x-z plane, oriented such that a long centerline of the bottom face has only an x-axis dimension, ii. a y-axis is orthogonal to the x-z plane and defines a vertical
direction, and iii. when the imaging apparatus is standing unattended on a flat surface on its bottom face, the first, second and third cameras are aligned with each other at the same height in the vertical direction within a tolerance of one-half of a diameter of any of the respective ultra wide-angle lenses.
55. The imaging apparatus of claim 54, wherein when the imaging apparatus is
standing unattended on a flat surface on its bottom face, the first, second and third cameras are aligned with each other at the same height in the vertical direction within a tolerance of one millimeter.
PCT/IB2020/052766 2019-03-25 2020-03-24 Systems, apparatuses and methods for acquiring, processing and delivering stereophonic and panoramic images WO2020194190A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201962823409P 2019-03-25 2019-03-25
US62/823,409 2019-03-25
US201962893497P 2019-08-29 2019-08-29
US62/893,497 2019-08-29

Publications (1)

Publication Number Publication Date
WO2020194190A1 true WO2020194190A1 (en) 2020-10-01

Family

ID=71474902

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2020/052766 WO2020194190A1 (en) 2019-03-25 2020-03-24 Systems, apparatuses and methods for acquiring, processing and delivering stereophonic and panoramic images

Country Status (2)

Country Link
CN (3) CN211018941U (en)
WO (1) WO2020194190A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160088282A1 (en) * 2014-09-22 2016-03-24 Samsung Electronics Company, Ltd. Transmission of three-dimensional video
US20160191815A1 (en) * 2014-07-25 2016-06-30 Jaunt Inc. Camera array removing lens distortion
US20180225805A1 (en) * 2017-02-09 2018-08-09 Jaunt Inc. Generating Virtual Reality Content Based on Corrections to Stitching Errors
WO2018146376A1 (en) * 2017-02-13 2018-08-16 Nokia Technologies Oy An apparatus, a method and a computer program for video coding and decoding
US20180342043A1 (en) * 2017-05-23 2018-11-29 Nokia Technologies Oy Auto Scene Adjustments For Multi Camera Virtual Reality Streaming
US20190019299A1 (en) * 2016-01-03 2019-01-17 Humaneyes Technologies Ltd. Adaptive stitching of frames in the process of creating a panoramic frame

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160191815A1 (en) * 2014-07-25 2016-06-30 Jaunt Inc. Camera array removing lens distortion
US20160088282A1 (en) * 2014-09-22 2016-03-24 Samsung Electronics Company, Ltd. Transmission of three-dimensional video
US20190019299A1 (en) * 2016-01-03 2019-01-17 Humaneyes Technologies Ltd. Adaptive stitching of frames in the process of creating a panoramic frame
US20180225805A1 (en) * 2017-02-09 2018-08-09 Jaunt Inc. Generating Virtual Reality Content Based on Corrections to Stitching Errors
WO2018146376A1 (en) * 2017-02-13 2018-08-16 Nokia Technologies Oy An apparatus, a method and a computer program for video coding and decoding
US20180342043A1 (en) * 2017-05-23 2018-11-29 Nokia Technologies Oy Auto Scene Adjustments For Multi Camera Virtual Reality Streaming

Also Published As

Publication number Publication date
CN211018942U (en) 2020-07-14
CN211018941U (en) 2020-07-14
CN211019015U (en) 2020-07-14

Similar Documents

Publication Publication Date Title
US10440322B2 (en) Automated configuration of behavior of a telepresence system based on spatial detection of telepresence components
US10171771B2 (en) Camera system for video conference endpoints
CA2949005C (en) Method and system for low cost television production
US20150124171A1 (en) Multiple vantage point viewing platform and user interface
US7224382B2 (en) Immersive imaging system
EP2352290B1 (en) Method and apparatus for matching audio and video signals during a videoconference
US20100013738A1 (en) Image capture and display configuration
US20180227501A1 (en) Multiple vantage point viewing platform and user interface
US10296281B2 (en) Handheld multi vantage point player
US10156898B2 (en) Multi vantage point player with wearable display
WO2012100114A2 (en) Multiple viewpoint electronic media system
US20220264055A1 (en) Video Conference Apparatus, Video Conference Method and Computer Program Using a Spatial Virtual Reality Environment
JP6628343B2 (en) Apparatus and related methods
US10732706B2 (en) Provision of virtual reality content
US10664225B2 (en) Multi vantage point audio player
WO2017118662A1 (en) Spherical virtual reality camera
CN110730340B (en) Virtual audience display method, system and storage medium based on lens transformation
US20150304724A1 (en) Multi vantage point player
US20090153550A1 (en) Virtual object rendering system and method
US20110149038A1 (en) Video processing apparatus capable of reproducing video content including a plurality of videos and control method therefor
WO2018027067A1 (en) Methods and systems for panoramic video with collaborative live streaming
WO2020194190A1 (en) Systems, apparatuses and methods for acquiring, processing and delivering stereophonic and panoramic images
KR20180092411A (en) Method and apparatus for transmiting multiple video
KR20210064654A (en) System for playing three dimension image of 360 degrees
JP7472483B2 (en) System, method, device, and program for observing and noticing with a super-wide 360-degree camera

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20777198

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20777198

Country of ref document: EP

Kind code of ref document: A1