US11930290B2 - Panoramic picture in picture video - Google Patents
Panoramic picture in picture video Download PDFInfo
- Publication number
- US11930290B2 US11930290B2 US16/963,738 US201916963738A US11930290B2 US 11930290 B2 US11930290 B2 US 11930290B2 US 201916963738 A US201916963738 A US 201916963738A US 11930290 B2 US11930290 B2 US 11930290B2
- Authority
- US
- United States
- Prior art keywords
- video stream
- immersive
- production
- video
- end user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000004519 manufacturing process Methods 0.000 claims abstract description 62
- 238000000034 method Methods 0.000 claims abstract description 26
- 238000003860 storage Methods 0.000 claims description 11
- 230000008859 change Effects 0.000 claims description 6
- 238000004091 panning Methods 0.000 claims 1
- 238000012545 processing Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000012546 transfer Methods 0.000 description 4
- 238000013461 design Methods 0.000 description 3
- 239000000835 fiber Substances 0.000 description 3
- 238000002370 liquid polymer infiltration Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 235000006732 Torreya nucifera Nutrition 0.000 description 2
- 244000111306 Torreya nucifera Species 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000001276 controlling effect Effects 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 239000013307 optical fiber Substances 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 241000699670 Mus sp. Species 0.000 description 1
- 239000000872 buffer Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 238000010348 incorporation Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000001902 propagating effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
- 238000010922 spray-dried dispersion Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/265—Mixing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/445—Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
- H04N5/45—Picture in picture, e.g. displaying simultaneously another television channel in a region of the screen
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/21805—Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
- H04N21/4316—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/485—End-user interface for client configuration
- H04N21/4858—End-user interface for client configuration for modifying screen layout parameters, e.g. fonts, size of the windows
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/695—Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2624—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects for obtaining an image which is composed of whole input images, e.g. splitscreen
Definitions
- PIP Picture-in-Picture
- IP internet protocol
- OTA digital Over-the-Air
- FIG. 1 illustrates an example information handling device.
- FIG. 2 illustrates an example information handling system.
- FIG. 3 illustrates an example embodiment
- FIG. 4 illustrates an example embodiment
- a panoramic video environment is defined as a virtual experience, viewed on personal computers, mobile devices, virtual reality headsets or “goggles,” “smart” televisions, etc., where the user views a panorama from a first person perspective relative to the camera that is capturing the video.
- the “view” may be a two-dimensional “portal” into the panoramic environment, or a stereoscopic view (three dimensional view).
- the panorama is typically 3600 in one plane (azimuth), and varying from 50°-360° in the orthogonal plane (altitude).
- the user will typically have directional control in the view portal, including the ability to pan, tilt, and zoom from the perspective of the camera
- Panoramic or immersive video is becoming increasingly popular, particularly on social media web sites such as Facebook®, YouTube®, and Twitter®.
- Facebook is a registered trademark of Facebook, Inc.
- YouTube® is a registered trademark of Google, Inc.
- Twitter® is a registered trademark Twitter, Inc.
- panoramic video permits the user to pan, tilt, and zoom (PTZ) in the immersive environment, thus creating a personalized, first-person experience in contrast with the curated broadcast view.
- Embodiments provide that, while viewing the stream, the user may pan, tilt, and zoom the view in an immersive environment. It should be understood that the user would not be controlling the camera capturing the stream. Rather, because the video stream is captured as a panoramic view, the user is able to manipulate the view within the stream. In other words, while the video stream is captured in panoramic, a user only sees a portion of the panoramic stream at a single time, referred to herein as a view. Therefore, the user is able to provide commands (e.g., tilt, pan, zoom, etc.) to see a different portion, or view, of the panoramic stream.
- commands e.g., tilt, pan, zoom, etc.
- Broadcasters routinely create PIP experiences during production, having multiple camera feeds from numerous angles, vantage points, and points-of-view, etc. However, there is no way in which a broadcast or live stream viewer could interact with any of the video sources that comprise the broadcast.
- the camera views are deterministic, being specified at the time of production.
- An embodiment provides a method comprising: capturing, using two or more cameras, at least one of which is panoramic, video from high value, interesting, or unusual vantage points (desirable vantage points), and creating a multiplexed stream, for consumption on a suitably-equipped television, personal computer (PC), mobile phone, or head-mount display (HMD), allowing the user/viewer the ability to interact with at least one of the camera views in such a way so as to change their view from the perspective of that camera.
- PC personal computer
- HMD head-mount display
- the immersive cameras may be connected to a workstation which contains a frame grabber which ingests the video frames from two or more cameras simultaneously and transfers them to a graphics processing unit (GPU) for processing.
- a workstation which contains a frame grabber which ingests the video frames from two or more cameras simultaneously and transfers them to a graphics processing unit (GPU) for processing.
- GPU graphics processing unit
- Various video pipeline operations are performed on the GPU, including debayering (demosaicing), noise reduction, color correction, and the like.
- debayering demosaicing
- noise reduction noise reduction
- color correction color correction
- the video images are mathematically transformed to yield an industry-standard projection, such as equirectangular. This procedure is performed on a frame-by-frame basis, typically at the rate of thirty to sixty frames per second.
- the camera views ingested by the workstation may be processed individually, or composited to form a single projection as taught in U.S. Provisional Application No. 62/571,876 entitled “Creating Multi-Camera Panoramic Projections”, filed on Oct. 13, 2017, attached as an Appendix, the contents of which are fully incorporated by reference herein.
- the final projection may then be encoded with an industry standard codec, such as H.264 or HEVC (High Efficiency Video Coding).
- H.264 or MPEG-4 Part 10, Advanced Video Coding (MPEG-4 AVC) is a block-oriented motion-compensation-based video compression standard.
- H.264 It is a commonly used format for the recording, compression, and distribution of video content.
- the encoded video frames are streamed via the workstation network interface card.
- SDI Serial Digital Interface
- SMPTE Society of Motion Picture and Television Engineers
- video frames may be composited with the production video, during the production pipeline operations, such that a PIP view of the immersive camera is shown as an overlay on the video production.
- one or more cameras are used to capture the live event. These cameras feed into a production switcher/router (backhaul) which allows the event producer, in real time, to choose which cameras are “aired”.
- the production signal proceeds through the pipeline, where graphics, scores, game clocks, and the like are overlayed (“keyed”) onto the video.
- the video is packaged for transport.
- a DVB Digital Video Broadcasting
- MPEG-2 TS Transport Stream based upon standards set by the MPEG (Moving Picture Expert Group) is often employed.
- the video signal is typically encoded with H.264 or HEVC, then relayed via Real Time Messaging Protocol (RTMP) or some other suitable transport stream.
- RTMP Real Time Messaging Protocol
- An embodiment provides that injecting a multiplexer, at the broadcast site, after the encoding process, allows us to combine multiple elementary streams.
- the live production stream may be combined with the immersive camera streams, yielding a single transport stream directed to the cloud for consumption.
- This transport stream thus, contains multiple elementary streams—one for the production video stream, and one or more for each of the immersive camera streams.
- An OTT (Over the Top) provider's contribution encoder ingests the transport stream, and then disseminates to a scalable number of viewers, using various protocols (e.g, HLS, MPEG-DASH, and the like), using resolutions and codecs appropriate for the end-users' devices, which may include personal computers, mobile phones, head-mounted displays, and the like.
- various protocols e.g, HLS, MPEG-DASH, and the like
- resolutions and codecs appropriate for the end-users' devices, which may include personal computers, mobile phones, head-mounted displays, and the like.
- End users watching the streamed production in a streaming player application have the ability, by clicking on an icon or some type of user-interface (UI) indicator, or the PIP itself, to change their view according to one of the following non-limiting scenarios:
- UI user-interface
- the production stream and immersive streams are not multiplexed on site, but on cloud-based servers.
- the broadcaster encodes and streams the production feed independently of the immersive camera feed(s).
- the cloud-based contribution encoder multiplexes the multiple streams and makes the resultant stream available for mass consumption.
- the PIP as seen by the user/viewer, may be a window drawn by the software player application, rather than a part of the broadcast stream.
- the PIP window may be static or may move dynamically, or it may be moved and/or resized by the user. This additional flexibility adds to the overall user-controlled experience.
- the user may elect to “auto track” a device.
- positional data for objects of interest e.g., a player
- the viewing application automatically adjust the gaze direction per the location of the object of interest in the panorama as taught in U.S. Patent Pub. No. 2016/0173775.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises at least one executable instruction for implementing the specified logical function(s).
- the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- FIG. 1 includes a system on a chip design found for example in tablet or other mobile computing platforms.
- Software and processor(s) are combined in a single chip 110 .
- Processors comprise internal arithmetic units, registers, cache memory, busses, I/O ports, etc., as is well known in the art. Internal busses and the like depend on different vendors, but essentially all the peripheral devices ( 120 ) may attach to a single chip 110 .
- the circuitry 100 combines the processor, memory control, and I/O controller hub all into a single chip 110 .
- systems 100 of this type do not typically use SATA or PCI or LPC. Common interfaces, for example, include SDIO and I2C.
- power management chip(s) 130 e.g., a battery management unit, BMU, which manage power as supplied, for example, via a rechargeable battery 140 , which may be recharged by a connection to a power source (not shown).
- BMU battery management unit
- a single chip, such as 110 is used to supply BIOS like functionality and DRAM memory.
- System 100 typically includes one or more of a WWAN transceiver 150 and a WLAN transceiver 160 for connecting to various networks, such as telecommunications networks and wireless Internet devices, e.g., access points. Additionally, devices 120 are commonly included, e.g., an image sensor such as a camera. System 100 often includes a touch screen 170 for data input and display/rendering. System 100 also typically includes various memory devices, for example flash memory 180 and SDRAM 190 .
- FIG. 2 depicts a block diagram of another example of information handling device circuits, circuitry or components.
- the example depicted in FIG. 2 may correspond to computing systems such as the THINKPAD® series of personal computers sold by Lenovo (US) Inc. of Morrisville, NC, or other devices.
- embodiments may include other features or only some of the features of the example illustrated in FIG. 2 .
- FIG. 2 includes a so-called chipset 210 (a group of integrated circuits, or chips, that work together, chipsets) with an architecture that may vary depending on manufacturer (for example, INTEL, AMD, ARM, etc.).
- INTEL is a registered trademark of Intel Corporation in the United States and other countries.
- AMD is a registered trademark of Advanced Micro Devices, Inc. in the United States and other countries.
- ARM is an unregistered trademark of ARM Holdings plc in the United States and other countries.
- the architecture of the chipset 210 includes a core and memory control group 220 and an I/O controller hub 250 that exchanges information (for example, data, signals, commands, etc.) via a direct management interface (DMI) 242 or a link controller 244 .
- DMI direct management interface
- the DMI 242 is a chip-to-chip interface (sometimes referred to as being a link between a “northbridge” and a “southbridge”).
- the core and memory control group 220 include one or more processors 222 (for example, single or multi-core) and a memory controller hub 226 that exchange information via a front side bus (FSB) 224 ; noting that components of the group 220 may be integrated in a chip that supplants the conventional “northbridge” style architecture.
- processors 222 comprise internal arithmetic units, registers, cache memory, busses, I/O ports, etc., as is well known in the art.
- the memory controller hub 226 interfaces with memory 240 (for example, to provide support for a type of RAM that may be referred to as “system memory” or “memory”).
- the memory controller hub 226 further includes a low voltage differential signaling (LVDS) interface 232 for a display device 292 (for example, a CRT, a flat panel, touch screen, etc.).
- a block 238 includes some technologies that may be supported via the LVDS interface 232 (for example, serial digital video, HDMI/DVI, display port).
- the memory controller hub 226 also includes a PCI-express interface (PCI-E) 234 that may support discrete graphics 236 .
- PCI-E PCI-express interface
- the I/O hub controller 250 includes a SATA interface 251 (for example, for HDDs, SDDs, etc., 280 ), a PCI-E interface 252 (for example, for wireless connections 282 ), a USB interface 253 (for example, for devices 284 such as a digitizer, keyboard, mice, cameras, phones, microphones, storage, other connected devices, etc.), a network interface 254 (for example, LAN), a GPIO interface 255 , a LPC interface 270 (for ASICs 271 , a TPM 272 , a super I/O 273 , a firmware hub 274 , BIOS support 275 as well as various types of memory 276 such as ROM 277 , Flash 278 , and NVRAM 279 ) a power management interface 261 , a clock generator interface 262 , an audio interface 263 (for example, for speakers 294 ), a TCO interface 264 , a system management bus interface 265 , and
- the system upon power on, may be configured to execute boot code 290 for the BIOS 268 , as stored within the SPI Flash 266 , and thereafter processes data under the control of one or more operating systems and application software (for example, stored in system memory 240 ).
- An operating system may be stored in any of a variety of locations and accessed, for example, according to instructions of the BIOS 268 .
- a device may include fewer or more features than shown in the system of FIG. 2 .
- Information handling device circuitry may be used in devices such as tablets, smart phones, smart speakers, personal computer devices generally, and/or electronic devices which enable users to communicate with a digital assistant.
- the circuitry outlined in FIG. 1 may be implemented in a tablet or smart phone embodiment
- the circuitry outlined in FIG. 2 may be implemented in a personal computer embodiment.
- FIG. 3 provided is a non-limiting example embodiment in the context of broadcasting a soccer match.
- two opposing nets in which are located two inward-facing immersive cameras ( 1 B).
- traditional broadcast cameras 1 A
- Traditional broadcast cameras are used to capture a side view of the entire field, and wireless cameras are used for on-the-field and sideline shots where a mobile point-of-view (POV) is advantageous.
- the traditional cameras utilize SDI technology to transmit signals of varying resolutions to the production backhaul at ( 2 ).
- the backhaul at ( 2 ) may consist of numerous routers and switching devices though only one SDI SWITCHER ROUTER is shown in FIG. 3 for clarity.
- the matrix of cameras ( 1 A and 1 B) is available to the broadcast producer and production staff.
- the production pipeline ( 3 ) is used to augment the camera selection with replays, audio, transitions, graphic overlays such as the score and game clock, and the like. It is in this stage, in at least one embodiment, that one or more PIP windows may be drawn, overlaying the primary broadcast video. This may be done via a ganging of powerful graphics workstations, graphics processors, and software, with operators managing the workflow.
- One non-limiting example of such a system is the CHYRONHEGO® LIVE COMPOSITOR system.
- CHYRONHEGO is a registered trademark of ChyronHego Corporation of New York. Other such systems are available and would be equally useful.
- each immersive camera ( 1 B) captures a horizontal field of view (HFOV) and vertical field of view (VFOV) of 180°—a full hemisphere.
- HFOV horizontal field of view
- VFOV vertical field of view
- An example of such a cameras might be an IO Industries Flare 48M30CCX-A with a short focal length “fisheye” lens, such as the Canon 8-15 mm f4 USM L.
- the cameras may communicate with a remote workstation ( 8 ), typically located in a production truck that can be 1-10 km from the location of the cameras.
- the remote workstation ( 8 ) is a custom-built unit from Next Computing, Inc. comprising an ASUS® X99-E Mainboard with an Intel® Xeon-E5 Processor having 16 GB DDR4-2133 MHz-non ECC RAM with an 850 W Power Supply running Microsoft Windows 10-Pro 64 Bit operating system.
- the frame grabber is a Komodo Kaya FXP, manufactured by Kaya Instruments.
- ASUS® is a registered trademark of Asustek Computer Incorporation of Taiwan.
- INTEL® is a registered trademark of Intel Corporation of Delaware.
- the communication of the multiplexed video may be accomplished by utilizing one or more 10 Gbs fiber optic links, as is common in the industry.
- the fiber links connect to a frame grabber card ( 8 . 1 ) located in the remote workstation ( 8 ).
- the frame grabber is a Komodo Kaya FXP, manufactured by Kaya Instruments.
- the frame grabber ( 8 . 1 ) driver along with custom software, allows for the direct transfer of video frames from the frame grabber memory to the graphics processing unit (GPU— 8 . 3 .
- the GPU ( 8 . 3 ) is specifically designed to process video frames at high rates, whereas the workstation central processing unit (CPU) is designed for more general purpose tasks.
- the objective is to transfer video frames from the frame grabber ( 8 . 1 ) to the GPU ( 8 . 3 ) with as little CPU intervention as possible, as this would degrade performance.
- One embodiment describes having only two cameras connected to the workstation. This is a non-limiting embodiment, since the workstation may contain multiple frame grabbers, each connected to multiple cameras.
- the practical limitations to the number of cameras are dictated by the state of the art of both computers and their buses.
- the standard bus interface is PCI (Peripheral Connect Interconnect), and its derivatives—PCIe (PCI-Express).
- PCI bus standard continues to evolve, with each iteration increasing the number of “lanes” resulting in a greater number of Gigabytes per second, thus facilitating the transfer of video frames at a greater rate.
- Video frames are pushed to the GPU ( 8 . 3 ) memory, they can be manipulated. Typical operations are debayering (demosaicing), denoising, white balance adjustment, the application of 3D LUTs (Look-Up Tables) used for color correction, and the like. These video pipeline operations are required to improve the video quality for adherence to professional video standards. These video pipeline operations are performed in custom software optimized for GPU processing.
- the present example embodiment uses GPUs manufactured by NVIDIA company, specifically the GeForce 1080 Ti, a registered trademark of NVIDIA company, and employs software and libraries written in CUDA—NVIDIA's proprietary processing language CUDAV, a registered trademark of NVIDIA company. This language harnesses the massively parallel processing architecture of the GPU.
- the next operations on the GPU ( 8 . 3 ) consist of transformational operations. These transformational operations may be used to create a composited projection that may be encoded and streamed as is taught in U.S. Provisional Patent Application No. 62/571,876 entitled “Creating Multi-Camera Panoramic Projections”.
- the captured circular frames are transformed in software into an equirectangular, or a portion of an equirectangular, projection.
- the circle will map (transform) into one half of an equirectangular projection.
- the mathematical transform maps pixels in the source video frame (oval) to the equirectangular video frame.
- each video frame is encoded.
- Video may be encoded using the H.264 or HEVC codecs for video transmission. Other codecs may be used to obtain desirable quality. This is performed on ASICs built in to the GPU.
- the video frames may be sent to the NIC (network interface card) ( 8 . 4 ), being converted to an Internet Protocol (IP) stream and packaged suitably for transport to the internet ( 11 ), typically being connected via a CAT6 patch cable, fiber optic link, or wireless transmission.
- IP Internet Protocol
- the immersive cameras views are also available for “airing” in production.
- An operator located at the workstation ( 8 ) may utilize a joystick or other means to create unique PTZ views which are then converted to standard SDI broadcast signals via a SDI output computer card ( 8 . 2 ) within the workstation.
- the number of unique SDI outputs from the workstation is a practical limitation of the type of SDI card ( 8 . 2 ) employed, as well as the processing power of the workstation ( 8 ), and is not correlated with the number of cameras. In principal, the number of unique PTZ views from each immersive camera is essentially infinite.
- SDI video frames from the immersive cameras may be used in the broadcast production to create a PIP window, using video compositing techniques well known in the industry.
- the traditional cameras are capturing the action on the entire field, with occasional “close-ups” of the players, while the immersive cameras are capturing video in both nets.
- the broadcast video could then show the field view as the primary view, with two PIPs of each net.
- the user may click on either PIP and activate the secondary or tertiary stream, thus placing them in a first person perspective from the vantage point of that camera, allowing them to PTZ.
- the various streams are routed to a multiplexer ( 10 )—the “aired” video stream as well one or more immersive camera streams.
- the multiplexer (MUX) is a workstation or appliance that is capable of ingesting multiple simultaneous elementary streams, each with audio and video, and creating a transport stream containing a plurality of elementary streams.
- the MUX ( 10 ) has FIFO buffers for ingesting and synchronizing the incoming IP packets. Typically, there is a latency incurred at this point.
- the transport stream is then directed to an OTT (over the top) provider ( 11 . 1 ) that has the capability of transcoding the contribution stream and disseminating it to a large audience.
- OTT over the top
- One concern with combing multiple streams is the resultant bandwidth requirement.
- the MUX ( 10 ) is cloud-based.
- the PIP functionality may be dynamically constructed in software, rather than in broadcast production. This is a preferred embodiment for several reasons. First, it lowers broadcast costs and complexities. Secondly, it affords the rights owner(s) the opportunity to upcharge and monetize the immersive camera streams. Thirdly, it provides the OTT operator an opportunity to “strip” away the immersive streams for the case of streaming to users with insufficient bandwidth, or to create separate program streams. Fourthly, it enhances the end-user/viewer experience by allowing the PIP window(s) on the user's player software application to be dynamic and user-configurable.
- one user may choose not to view PIPs, and can turn this functionality off such that the OTT stream will be unbundled and contain only the streamed broadcast.
- Another user may want to resize their PIP(s) because the immersive camera views are more interesting than the broadcast stream.
- a custom application on a STB or smart television would allow the selection of the PIP(s), as well as the navigation in the immersive video, via a remote control or mobile device used as a proxy for controlling the STB or smart television.
- FIG. 4 a non-limiting example embodiment is provided.
- This upper figure depicts the broadcast output ( 1 . 1 ) which may be streamed to the internet for viewing. It should be understood that this is merely a captured frame of the continuous video program.
- the lower figure ( 1 . 2 ) depicts the same production output overlayed with an immersive PIP window.
- the immersive PIP video frames update synchronously with the production video frames.
- the user may select the PIP window, thereby entering into a mode whereby they may elect to change their PTZ.
- one PIP window is shown.
- the number of PIP windows corresponds to the number of immersive cameras as well the available bandwidth in which to stream the multiple steams.
- FIG. 1 and FIG. 2 illustrate a non-limiting examples of such devices and components thereof.
- aspects may be embodied as a system, method or device program product. Accordingly, aspects may take the form of an entirely hardware embodiment or an embodiment including software that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects may take the form of a device program product embodied in one or more device readable medium(s) having device readable program code embodied therewith.
- the non-signal medium may be a storage medium.
- a storage medium may be, for example, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a storage medium would include the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
- a storage medium is a non-transitory storage medium, inclusive of all storage media other than signal or propagating media.
- Program code embodied on a storage medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, et cetera, or any suitable combination of the foregoing.
- Program code for carrying out operations may be written in any combination of one or more programming languages.
- the program code may execute entirely on a single device, partly on a single device, as a stand-alone software package, partly on single device and partly on another device, or entirely on the other device.
- the devices may be connected through any type of connection or network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made through other devices (for example, through the Internet using an Internet Service Provider) or through a hard wire connection, such as over a USB connection.
- LAN local area network
- WAN wide area network
- the program instructions may also be stored in a device readable medium that can direct a device to function in a particular manner, such that the instructions stored in the device readable medium produce an article of manufacture including instructions which implement the functions/acts specified.
- the program instructions may also be loaded onto a device to cause a series of operational steps to be performed on the device to produce a device implemented process such that the instructions which execute on the device provide processes for implementing the functions/acts specified.
Abstract
Description
-
- (I) The immersive view may become the primary view, allowing the user to interact as described above. The production view may become the PIP;
- (II) The immersive view may populate the entire screen without displaying a PIP window. A keystroke or clicking on an icon might return the user to the primary view; or
- (III) The immersive PIP window may be selected from amongst several choices based upon the user's preferences. As taught in U.S. Patent Pub. No. 2016/0173775, positional data of objects of interest may be packaged as metadata with each immersive camera video frame. In cases where such metadata is available, that data may be used to provide PIP alternatives for the user. For example, one user may prefer to watch the goalies, while another may prefer to follow the ball.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/963,738 US11930290B2 (en) | 2018-01-22 | 2019-01-21 | Panoramic picture in picture video |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201862620218P | 2018-01-22 | 2018-01-22 | |
US16/963,738 US11930290B2 (en) | 2018-01-22 | 2019-01-21 | Panoramic picture in picture video |
PCT/US2019/014417 WO2019144076A1 (en) | 2018-01-22 | 2019-01-21 | Panoramic picture in picture video |
Publications (2)
Publication Number | Publication Date |
---|---|
US20210084254A1 US20210084254A1 (en) | 2021-03-18 |
US11930290B2 true US11930290B2 (en) | 2024-03-12 |
Family
ID=67301204
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/963,738 Active US11930290B2 (en) | 2018-01-22 | 2019-01-21 | Panoramic picture in picture video |
Country Status (2)
Country | Link |
---|---|
US (1) | US11930290B2 (en) |
WO (1) | WO2019144076A1 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110446056A (en) * | 2019-07-30 | 2019-11-12 | 咪咕文化科技有限公司 | A kind of method for processing video frequency, equipment, system and computer readable storage medium |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020147991A1 (en) * | 2001-04-10 | 2002-10-10 | Furlan John L. W. | Transmission of panoramic video via existing video infrastructure |
US7496277B2 (en) * | 2003-06-02 | 2009-02-24 | Disney Enterprises, Inc. | System and method of programmatic window control for consumer video players |
US8970666B2 (en) * | 2011-09-16 | 2015-03-03 | Disney Enterprises, Inc. | Low scale production system and method |
US20150289032A1 (en) * | 2014-04-03 | 2015-10-08 | Nbcuniversal Media, Llc | Main and immersive video coordination system and method |
US9185361B2 (en) * | 2008-07-29 | 2015-11-10 | Gerald Curry | Camera-based tracking and position determination for sporting events using event information and intelligence data extracted in real-time from position information |
US9781356B1 (en) * | 2013-12-16 | 2017-10-03 | Amazon Technologies, Inc. | Panoramic video viewer |
US20180025751A1 (en) * | 2016-07-22 | 2018-01-25 | Zeality Inc. | Methods and System for Customizing Immersive Media Content |
US10277813B1 (en) * | 2015-06-25 | 2019-04-30 | Amazon Technologies, Inc. | Remote immersive user experience from panoramic video |
US20190394375A1 (en) * | 2017-01-30 | 2019-12-26 | Evs Broadcast Equipment Sa | Video Production Apparatus and Method for Video Production |
-
2019
- 2019-01-21 WO PCT/US2019/014417 patent/WO2019144076A1/en active Application Filing
- 2019-01-21 US US16/963,738 patent/US11930290B2/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020147991A1 (en) * | 2001-04-10 | 2002-10-10 | Furlan John L. W. | Transmission of panoramic video via existing video infrastructure |
US7496277B2 (en) * | 2003-06-02 | 2009-02-24 | Disney Enterprises, Inc. | System and method of programmatic window control for consumer video players |
US9185361B2 (en) * | 2008-07-29 | 2015-11-10 | Gerald Curry | Camera-based tracking and position determination for sporting events using event information and intelligence data extracted in real-time from position information |
US8970666B2 (en) * | 2011-09-16 | 2015-03-03 | Disney Enterprises, Inc. | Low scale production system and method |
US9781356B1 (en) * | 2013-12-16 | 2017-10-03 | Amazon Technologies, Inc. | Panoramic video viewer |
US20150289032A1 (en) * | 2014-04-03 | 2015-10-08 | Nbcuniversal Media, Llc | Main and immersive video coordination system and method |
US10277813B1 (en) * | 2015-06-25 | 2019-04-30 | Amazon Technologies, Inc. | Remote immersive user experience from panoramic video |
US20180025751A1 (en) * | 2016-07-22 | 2018-01-25 | Zeality Inc. | Methods and System for Customizing Immersive Media Content |
US20190394375A1 (en) * | 2017-01-30 | 2019-12-26 | Evs Broadcast Equipment Sa | Video Production Apparatus and Method for Video Production |
Non-Patent Citations (1)
Title |
---|
Young, Lee W., International Search Report for PCT Application Serial No. PCT/US2019/014417, dated Mar. 25, 2019, 2 pages, Alexandria, Virginia. |
Also Published As
Publication number | Publication date |
---|---|
US20210084254A1 (en) | 2021-03-18 |
WO2019144076A1 (en) | 2019-07-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7029562B2 (en) | Equipment and methods for providing and displaying content | |
US10021301B2 (en) | Omnidirectional camera with multiple processors and/or multiple sensors connected to each processor | |
US9843725B2 (en) | Omnidirectional camera with multiple processors and/or multiple sensors connected to each processor | |
KR20190121867A (en) | Method and apparatus for packaging and streaming virtual reality media content | |
US10623636B2 (en) | Generating objects in real time panoramic video | |
CN108289231B (en) | Integrated panoramic player | |
US10511767B2 (en) | Information processing device, information processing method, and program | |
JP7177034B2 (en) | Method, apparatus and stream for formatting immersive video for legacy and immersive rendering devices | |
KR20180038256A (en) | Method, and system for compensating delay of virtural reality stream | |
US11889228B2 (en) | Conference device with multi-videostream capability | |
US10638029B2 (en) | Shared experiences in panoramic video | |
US11490129B2 (en) | Creating multi-camera panoramic projections | |
US11930290B2 (en) | Panoramic picture in picture video | |
US10764655B2 (en) | Main and immersive video coordination system and method | |
WO2021032105A1 (en) | Code stream processing method and device, first terminal, second terminal and storage medium | |
US11792380B2 (en) | Video transmission method, video processing device, and video generating system for virtual reality | |
US11863902B2 (en) | Techniques for enabling high fidelity magnification of video | |
US10264241B2 (en) | Complimentary video content | |
US20230222754A1 (en) | Interactive video playback techniques to enable high fidelity magnification | |
US11451745B2 (en) | Conference device with multi-videostream control | |
US20210195300A1 (en) | Selection of animated viewing angle in an immersive virtual environment | |
Macq et al. | Application Scenarios and Deployment Domains |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: C360 TECHNOLOGIES, INC., PENNSYLVANIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LOWRY, BRIAN C.;TOMKO, JOSEPH B.;WIMER, EVAN A.;SIGNING DATES FROM 20200716 TO 20200720;REEL/FRAME:053269/0078 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
AS | Assignment |
Owner name: COSM, INC., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:C360 TECHNOLOGIES, INC.;REEL/FRAME:065013/0565 Effective date: 20230920 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |