WO2013144049A1 - Method and apparatus for encoding a selected spatial portion of a video stream - Google Patents

Method and apparatus for encoding a selected spatial portion of a video stream Download PDF

Info

Publication number
WO2013144049A1
WO2013144049A1 PCT/EP2013/056189 EP2013056189W WO2013144049A1 WO 2013144049 A1 WO2013144049 A1 WO 2013144049A1 EP 2013056189 W EP2013056189 W EP 2013056189W WO 2013144049 A1 WO2013144049 A1 WO 2013144049A1
Authority
WO
WIPO (PCT)
Prior art keywords
video stream
motion
encoding
spatial portion
selected spatial
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/EP2013/056189
Other languages
English (en)
French (fr)
Inventor
Patrice Rondao Alface
Jean-François Macq
Nico Verzijp
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Alcatel Lucent SAS
Original Assignee
Alcatel Lucent SAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alcatel Lucent SAS filed Critical Alcatel Lucent SAS
Priority to JP2015502259A priority Critical patent/JP6121518B2/ja
Priority to CN201380018231.9A priority patent/CN104365095B/zh
Priority to KR1020147030447A priority patent/KR20150003776A/ko
Priority to US14/389,033 priority patent/US20150117524A1/en
Publication of WO2013144049A1 publication Critical patent/WO2013144049A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/109Selection of coding mode or of prediction mode among a plurality of temporal predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • H04N19/139Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/187Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/40Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process

Definitions

  • the present invention relates to the field of video stream encoding .
  • pan/tilt/ zoom (PTZ) functions result in a loss of quality or resolution of the enlarged image portion.
  • Known server-side pan/tilt/ zoom (PTZ) functions result in a massive amount of computation at the server side.
  • the method comprising: obtaining picture element information pertaining to the selected spatial portion; obtaining encoding hints derived from a complementary spatial portion of the original video stream that is peripheral to the selected spatial portion; and encoding the selected spatial portion with use of the encoding hints.
  • the encoding process for the selected portion can be made more efficient by taking into account information about the bigger picture, in particular, features of the original video stream appearing in a region outside the selected portion, but close enough to be of a nature to influence the coding of subsequent frames.
  • Relevant peripheral features are preferably selected on the basis of the motion of these features, and of the "motion" of the selected portion; i.e., the motion of the image that is induced by any panning, tilting, or zooming with respect to the original video stream.
  • the method according to the present invention further comprises obtaining the original video stream; obtaining selection information representing the selected spatial portion of the stand-alone video-stream; and extracting the picture element information pertaining to the selected spatial portion from the original video stream in accordance with the selection information.
  • the information of the original video stream is used to generate the reduced video, such that the quality (e.g., resolution) of the original video stream can be retained to the greatest possible extent in the reduced video stream.
  • the method according to the present invention further comprises obtaining motion and feature information pertaining to the original video stream;
  • the motion adapter is configured to identify the relevant features by comparing a motion vector of a candidate feature to a motion vector
  • peripheral to the selected portion can be detected as moving into the region of interest, in which case their appearance can be anticipated, and new values may be selected for variable encoding parameters so as to optimally deal with the anticipated presence of that feature.
  • the encoding is performed by means of a scalable video codec.
  • the original video stream is provided as the base layer for the encoding.
  • the scalable video codec is an H.264 SVC codec.
  • the obtaining of the picture element information pertaining to the selected spatial portion comprises
  • the original video stream and/or the selected spatial portion may originally be available as encoded streams only. In that case, judicious transcoding would be more efficient than decoding and recoding of the raw video stream.
  • a computer program adapted to perform, when executed, the method according to embodiments of the present invention.
  • a computer readable storage medium comprising instructions to cause a data processing apparatus to carry out steps of the method according to embodiments of the present invention.
  • an apparatus for encoding a selected spatial portion of an original video stream as a stand-alone video stream comprising: a video input interface for receiving the original video stream; a selection information input interface for receiving selection information representing the selected spatial portion; a motion and feature
  • a panorama reframer operatively coupled to the video input interface and the selection information input interface, the panorama reframer being configured to extract picture element information pertaining to the selected spatial portion from the original video stream in accordance with the selection information;
  • a motion adapter operatively coupled to the motion and feature information input
  • the motion adapter being configured to identify relevant
  • a hinted encoder operatively coupled to the panorama reframer and the motion adapter, the hinted encoder being configured to encode the selected spatial portion with use of the encoding hints.
  • the hinted encoder is configured to encode the selected spatial portion by means of a scalable video codec.
  • the hinted encoder is configured to provide the original video stream as the base layer for the encoding.
  • the scalable video codec is an H.264 SVC codec.
  • the video input interface is configured to obtain the video stream as an encoded video stream, and wherein the hinted encoder is configured to transcode the encoded video stream.
  • the invention further provides a system for encoding a selected spatial portion of an original video stream as a stand-alone video stream, the system comprising a feature analyzer configured to extract motion information pertaining to features in the peripheral portion, and a number of apparatus according to any of claims 9-13, the feature analyzer being coupled to respective motion and feature information interfaces of the apparatus.
  • Figure 1 provides a flow chart of a method according to an embodiment of the present invention
  • Figure 2 provides a schematic illustration of an apparatus and a system according to embodiments of the present invention
  • Figure 3 illustrates the motion vector comparison performed in embodiments of the present invention.
  • the invention is related to a system composed of a proxy serving multiple users which can request a personalized region of interest (Rol), in a pan-tilt-zoom (PTZ)
  • Rol personalized region of interest
  • PTZ pan-tilt-zoom
  • the invention is further related to the optimization of the computational cost at the proxy side of the video encoding of these personalized video streams from the same video panorama source.
  • the term "panorama” is generally used herein to designate a high-resolution video resolution, typically above 4000 pixels wide, which may contain a cylindrical or spherical mapping of a wide-angle view or representation of a physical space as a video.
  • the panorama may be composed from the different video sources which are integrated and fused together.
  • Current solutions perform user requested cropping and PTZ operations in order to generate the desired Rol frame at time t in raw data and encode it using a H.264, or WebM alike codec with motion compensation and/or intra coding. Unfortunately, using this technique, motion
  • Rol-relative motion can then be computed from these panorama motion vectors by compensating the possible Rol user-requested displacement and resolution changes.
  • these compensated motion vectors are sent as hints to the encoder devoted to the coding of that Rol. The encoder can refine this motion vector "hint" if necessary.
  • Figure 1 provides a flow chart of a method according to an embodiment of the present invention.
  • the skilled person shall appreciate that the various illustrated steps are not necessarily performed by a single entity.
  • steps shown as taking place in parallel may be conducted
  • the original video stream is obtained 100 and used on the one hand to extract the relevant picture element information for the selected region 130, and on the other hand to extract motion and feature information 150.
  • a region of interest Rol
  • the selection action may be limited to panning, zooming, and tilting with respect to the original video stream, wherein the aspect ratio of the selected region is preferably limited to a fixed aspect ratio associated with the viewing device (e.g., 3x4 or 16x9) .
  • Picture element information is meant to comprise any form of representation of the video images within the selected region. This may be a stream of uncompressed video images, or an encoded video stream. Embodiments of the method according to the invention present the advantage are that motion is computed at highest
  • Embodiments of the method according to the invention present the further advantage that overlapping Rols or aligned Rols for different users do not require a duplication of the motion estimation effort for their respective encoders as motion data is pre-computed on the panorama .
  • Encoding hints are derived 160 from the extracted motion and feature information. While the motion and feature
  • the encoding hint derivation selects those features that are relevant for improving the encoding performance, as will be described in more detail below.
  • the encoder obtains the picture element information
  • the encoder encodes 180 the selected portion of the video stream.
  • the encoding hints are bits of information derived from the non-selected portion of the original video stream, used in addition to the information about the selected portion, to improve the encoding of the latter.
  • methods according to embodiments of the present invention track navigation and zoom commands from the client (at step 120) , and then scale the global motion data in response to zoom commands and add Rol motion in response to panning or tilting commands (at step 160) .
  • the encoding (at step 180) takes place on the basis of the obtained set of motion vectors, which are used either as hints as or actual motion vectors: If the reference frame area is not available in the encoder, a predicted motion vector and a search window range are sent to the encoder or an intra prediction mode.
  • the aforementioned method may in general be carried out by an apparatus for encoding a selected spatial portion of an original video stream as a stand-alone video stream, the apparatus comprising means for obtaining picture element information 140 pertaining to the selected spatial portion; means for obtaining encoding hints 170 derived from a complementary spatial portion of the original video stream that is peripheral to the selected spatial portion; and means for encoding the selected spatial portion 180 with use of the encoding hints.
  • the apparatus may further comprise means for obtaining the original video stream 100; means for obtaining selection information 120 representing the selected spatial portion of the stand-alone video-stream; and means for extracting the picture element information 130 pertaining to the selected spatial portion from the original video stream in accordance with the selection information.
  • the apparatus may further comprise means for obtaining motion and feature information 150 pertaining to the original video stream; means for identifying relevant features pertaining to a region that is peripheral to the selected spatial portion in the motion and feature information; and means for deriving the encoding hints 160 from the identified relevant features.
  • the means for encoding 180 may operate with a scalable video codec. More specifically, the means for encoding 180 may provide the original video stream as the base layer for the encoding. Also more specifically, the scalable video codec may be an H.264 SVC codec. The means for obtaining the picture element information 140 pertaining to the selected spatial portion, may be adapted to obtain an encoded video stream, and the encoding means 180 may be adapted to transcode the encoded video stream.
  • Figure 2 provides a schematic illustration of an apparatus and a system according to embodiments of the present invention.
  • the apparatus 200 comprises a video input interface 211 for receiving said original video stream 199, a selection information input interface 232 for receiving selection information representing said selected spatial portion, a motion and feature information input interface 221 for receiving feature information derived from said original video stream, a panorama reframer 211, operatively coupled to said video input interface 211 and said selection information input interface 232, said panorama reframer 211 being configured to extract picture element information pertaining to said selected spatial portion from said original video stream in accordance with said selection information, a motion adapter 220, operatively coupled to said motion and feature information input interface 221 and said selection information input interface 232, said motion adapter 220 being configured to identify relevant features pertaining to a region that is peripheral to said selected spatial portion in said motion and feature information and to derive encoding hints from said identified relevant features, and a hinted encoder 230, operatively coupled to said panorama reframer 210 and said motion adapter 220, said hinted encoder 230 being configured to encode said selected spatial portion with use of said
  • the apparatus 200 further has an output interface 231 for sending the encoded selected video stream to a client 300 over a network 250.
  • the illustrated network 250 may consist of one or more network links. It typically includes an access link on the client side.
  • interface designates the necessary hardware and software required to establish data communication
  • An access interface may for instance include an interface for an access link such as an xDSL, xPON, WMAN, or 3G link.
  • a LAN interface may for instance include an interface for one or more of an IEEE 802.3 "Ethernet" link, an IEEE 802.11
  • a PAN interface may for instance include a USB interface or a Bluetooth interface.
  • Higher- layer protocols for communication over several network segments are preferably protocols from the TCP/IP protocol suite .
  • the client 300 comprises a standard decoder 310 for decoding the received video stream.
  • the client 300 further comprises means 320 to allow the end user to select a region of interest, and to feed the resulting selection information back to the apparatus 200.
  • a system according to the invention may comprise a motion and feature analyzer 240 operating on the original video stream 199.
  • this motion and feature analyzer 240 is centralized, or located "in the cloud", and offers its output to one or several apparatus 200 according to the invention.
  • Adaptive Rol Encoder 200 comprises a motion adaptation module 220, a panorama video reframing module 210 and a hinted encoder 230 delivering a client 300 as represented in Fig. 2.
  • the panorama motion analyzer 240 receives a raw or encoded panorama video stream 199 and delivers motion information to the motion adaptation module 220 of the N Adaptive Rol encoders 200. If the panorama analyzer 240 performs its task in the compressed domain, the reframing module 210 shall allow for the necessary partial decoding. If the encoder 230 is an MPEG-7 compliant transcoder, hinting information may be provided as MPEG-7 transcoding hints to the encoder 230. Such a transcoder is known from Peter M. Kuhn, Teruhiko Suzuki and Anthony Vetro, "MPEG-7 Transcoding Hints for Reduced Complexity and Improved Quality, " in Proceeding of PacketVideo '01, 2001.
  • Each motion adaptation module 220 receives motion and feature information from the panorama motion analyzer 240 and the Rol position change requests from its client 300. Based on this information, the module 220 outputs hints to its connected hinted encoder 230.
  • the panorama reframing module 210 also reads the Rol request and re-samples and crops the requested region of the panorama and prepares it for the hinted encoder 230 in the requested position and resolution. If the panorama region needs to be mapped in 2D (e.g. cylindrical map for spherical video) then this is also done in the panorama reframing module 210.
  • Each hinted encoder 230 receives hints from its motion adaptation module 220 as well as raw cropped video stream at the requested resolution and position.
  • the hints are
  • the hints can be composed of a direct motion vector, a mode decision (motion partition) , a predicted motion vector on which the search window is to be centered as well as the size of this search window, an intra mode prediction, a skip decision, etc.
  • a mode decision motion partition
  • An example of this can be found in Ralph A.
  • C C r + C m with C m > C r (typically C m » C r ) .
  • the computational complexity of the panorama analysis (step 150 of Figure 1; the function of the analyzer 240 in Figure 2) is denoted as C p with C p > C.
  • serving N clients results in a complexity of only C p + N*C r , instead of the classical linearity in function of ( C p + C r ) ⁇ Adding one client only adds an increment of Cr which is much smaller than the term C m + C r in the traditional case .
  • the motion information computed by the panorama motion and feature analyzer 240 needs to be adapted to the client request by the motion adaptation module 220. This requires detecting whether the available motion vectors are pointing to panorama areas that are available as reference frames in the hinted encoder 230. Such detection is represented on Figure 3, where the panorama raw video frames motion and feature information are depicted together with one example of Rol request for M frames. Each panorama frame n(t) at time t is analyzed so as to produce for each pixel feature information and motion information in the panorama motion and feature analysis module 240.
  • Feature information is composed of e.g. edge position and orientation information, a segmentation of the panorama together with SIFT
  • motion information is represented as one motion vector for each of the M previous frames n(t-l)... ⁇ ( t-M) to serve as reference frames.
  • the motion adaptation module 220 receives the Ro l request and computes the Ro l position changes in the panorama for the previous M reference frames used by the hinted encoder 230.
  • the Ro l motion changes are represented by the vectors denoted as r in Figure 3, where the current Ro l macroblocks are represented as a regular grid and previous Ro l frame positions are represented as dashed rectangular regions (with width w(t) and height h(t) ) .
  • the motion adaptation module then checks for every other motion adaptation module.
  • the Rol request from the client 300 consists in a Rol
  • This information is used by the motion adaptation module 220 and by the panorama reframing module 210.
  • This module 210 This module 210
  • these clients 300 can be aggregated as virtual super-clients. This is for instance possible if the Rol displacement is guided by a script (e.g. tracking of an object or person or
  • the hinted encoder uses a scalable encoding technique, such as the scalable extension of H.264, SVC. This enables adaptation of the stream to the needs of the client 300 while still reducing the scalable encoding technique
  • the panorama motion analyzer 240 then specifically outputs multi- resolution motion and feature information so as to enable this.
  • the method according to the invention includes an additional step to estimate whether it is more efficient to use the motion analysis on the full panorama than to encode each Rol separately without hints.
  • a computation is performed to determine the minimum number N p of clients for which the panorama motion analysis (step 150) cost is compensated by the higher scalability.
  • N p is given by the following formula:
  • processors may be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software.
  • the functions may be provided by a single dedicated processor, by a single shared processor, or by a plurality of
  • controller should not be construed to refer exclusively to hardware capable of executing software, and may implicitly include, without limitation, digital signal processor (DSP) hardware, network processor, application specific integrated circuit (ASIC), field programmable gate array (FPGA), read only memory (ROM) for storing software, random access memory (RAM), and non volatile storage.
  • DSP digital signal processor
  • ASIC application specific integrated circuit
  • FPGA field programmable gate array
  • ROM read only memory
  • RAM random access memory
  • any switches shown in the FIGS are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even manually, the particular technique being selectable by the implementer as more specifically understood from the context.
  • program storage devices e.g., digital data storage media, which are machine or computer readable and encode machine-executable or computer-executable
  • the program storage devices may be, e.g., digital memories, magnetic storage media such as a magnetic disks and magnetic tapes, hard drives, or optically readable digital data storage media.
  • the embodiments are also intended to cover computers programmed to perform said steps of the above- described methods.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
PCT/EP2013/056189 2012-03-30 2013-03-25 Method and apparatus for encoding a selected spatial portion of a video stream Ceased WO2013144049A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
JP2015502259A JP6121518B2 (ja) 2012-03-30 2013-03-25 ビデオストリームの選択された空間部分を符号化するための方法および装置
CN201380018231.9A CN104365095B (zh) 2012-03-30 2013-03-25 用于对视频流的选定空间部分进行编码的方法和装置
KR1020147030447A KR20150003776A (ko) 2012-03-30 2013-03-25 비디오 스트림의 선택된 공간 부분을 인코딩하는 방법 및 장치
US14/389,033 US20150117524A1 (en) 2012-03-30 2013-03-25 Method and apparatus for encoding a selected spatial portion of a video stream

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP20120305389 EP2645713A1 (en) 2012-03-30 2012-03-30 Method and apparatus for encoding a selected spatial portion of a video stream
EP12305389.4 2012-03-30

Publications (1)

Publication Number Publication Date
WO2013144049A1 true WO2013144049A1 (en) 2013-10-03

Family

ID=47988985

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2013/056189 Ceased WO2013144049A1 (en) 2012-03-30 2013-03-25 Method and apparatus for encoding a selected spatial portion of a video stream

Country Status (6)

Country Link
US (1) US20150117524A1 (enExample)
EP (1) EP2645713A1 (enExample)
JP (1) JP6121518B2 (enExample)
KR (1) KR20150003776A (enExample)
CN (1) CN104365095B (enExample)
WO (1) WO2013144049A1 (enExample)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2961182A1 (en) * 2014-06-27 2015-12-30 Alcatel Lucent Method, system and device for navigating in ultra high resolution video content by a client device

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150296215A1 (en) * 2014-04-11 2015-10-15 Microsoft Corporation Frame encoding using hints
EP2961183A1 (en) * 2014-06-27 2015-12-30 Alcatel Lucent Method, system and related selection device for navigating in ultra high resolution video content
US9769494B2 (en) * 2014-08-01 2017-09-19 Ati Technologies Ulc Adaptive search window positioning for video encoding
US9917877B2 (en) 2014-10-20 2018-03-13 Google Llc Streaming the visible parts of a spherical video
GB2545999A (en) * 2014-10-20 2017-07-05 Google Inc Continuous prediction domain
US9918082B2 (en) 2014-10-20 2018-03-13 Google Llc Continuous prediction domain
US9918094B2 (en) 2014-10-20 2018-03-13 Google Llc Compressing and representing multi-view video
US10749923B2 (en) * 2015-06-08 2020-08-18 Apple Inc. Contextual video content adaptation based on target device
JP6559337B2 (ja) * 2015-09-23 2019-08-14 ノキア テクノロジーズ オーユー 360度パノラマビデオの符号化方法、符号化装置、及びコンピュータプログラム
GB2598255B (en) * 2016-02-17 2022-07-27 V Nova Int Ltd Physical adapter, signal processing equipment, methods and computer programs
US10334224B2 (en) * 2016-02-19 2019-06-25 Alcacruz Inc. Systems and method for GPU based virtual reality video streaming server
EP3249928A1 (en) 2016-05-23 2017-11-29 Thomson Licensing Method, apparatus and stream of formatting an immersive video for legacy and immersive rendering devices
EP3301915A1 (en) 2016-09-30 2018-04-04 Thomson Licensing Method and apparatus for omnidirectional video coding with adaptive intra most probable modes
EP3306937A1 (en) 2016-10-05 2018-04-11 Thomson Licensing Method and apparatus for encoding and decoding a video
CN107945101B (zh) * 2016-10-13 2021-01-29 华为技术有限公司 图像处理方法和装置
WO2018117706A1 (ko) * 2016-12-22 2018-06-28 주식회사 케이티 비디오 신호 처리 방법 및 장치
TWI780682B (zh) 2017-04-21 2022-10-11 美商時美媒體公司 用於藉由預測運動向量及/或快取重複運動向量的玩家輸入運動補償的系統及方法
US10225564B2 (en) 2017-04-21 2019-03-05 Zenimax Media Inc Systems and methods for rendering and pre-encoded load estimation based encoder hinting
CN108933920B (zh) * 2017-05-25 2023-02-17 中兴通讯股份有限公司 一种视频画面的输出、查看方法及装置
CN107396081B (zh) * 2017-06-19 2019-04-12 深圳市铂岩科技有限公司 针对全景视频的优化编码方法及装置
CN108307219B (zh) * 2017-11-07 2020-12-01 深圳市佳创视讯技术股份有限公司 一种高清电视全景视频智能展示方法
CN109121000A (zh) * 2018-08-27 2019-01-01 北京优酷科技有限公司 一种视频处理方法及客户端
CN109168032B (zh) * 2018-11-12 2021-08-27 广州酷狗计算机科技有限公司 视频数据的处理方法、终端、服务器及存储介质
US12316973B2 (en) 2021-12-13 2025-05-27 Be More Colorful, LLC Method to use 360-degree cameras

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110051808A1 (en) * 2009-08-31 2011-03-03 iAd Gesellschaft fur informatik, Automatisierung und Datenverarbeitung Method and system for transcoding regions of interests in video surveillance
US20110199389A1 (en) * 2008-12-19 2011-08-18 Microsoft Corporation Interactive virtual display system for ubiquitous devices

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3300228B2 (ja) * 1996-05-02 2002-07-08 株式会社メガチップス 画像通信システム
US6249613B1 (en) * 1997-03-31 2001-06-19 Sharp Laboratories Of America, Inc. Mosaic generation and sprite-based coding with automatic foreground and background separation
US6982764B1 (en) * 2000-05-25 2006-01-03 Northrop Grumman Corporation Image enhancement
FI114679B (fi) * 2002-04-29 2004-11-30 Nokia Corp Satunnaisaloituspisteet videokoodauksessa
US20100002070A1 (en) * 2004-04-30 2010-01-07 Grandeye Ltd. Method and System of Simultaneously Displaying Multiple Views for Video Surveillance
JP4305904B2 (ja) * 2003-09-22 2009-07-29 Kddi株式会社 部分画像符号化装置
US8427538B2 (en) * 2004-04-30 2013-04-23 Oncam Grandeye Multiple view and multiple object processing in wide-angle video camera
AU2005200713A1 (en) * 2005-02-17 2006-08-31 Canon Kabushiki Kaisha Motion estimation in a video sequence
US20080151049A1 (en) * 2006-12-14 2008-06-26 Mccubbrey David L Gaming surveillance system and method of extracting metadata from multiple synchronized cameras
US8498495B2 (en) * 2007-01-26 2013-07-30 Telefonaktiebolaget Lm Ericsson (Publ) Border region processing in images
US20090300692A1 (en) * 2008-06-02 2009-12-03 Mavlankar Aditya A Systems and methods for video streaming and display
US10440329B2 (en) * 2009-05-22 2019-10-08 Immersive Media Company Hybrid media viewing application including a region of interest within a wide field of view
IL199763B (en) * 2009-07-08 2018-07-31 Elbit Systems Ltd Automatic contractual system and method for observation
JP5089658B2 (ja) * 2009-07-16 2012-12-05 株式会社Gnzo 送信装置及び送信方法
JP5560009B2 (ja) * 2009-09-07 2014-07-23 株式会社日立国際電気 動画像符号化装置
TWI420906B (zh) * 2010-10-13 2013-12-21 Ind Tech Res Inst 興趣區域之追蹤系統與方法及電腦程式產品

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110199389A1 (en) * 2008-12-19 2011-08-18 Microsoft Corporation Interactive virtual display system for ubiquitous devices
US20110051808A1 (en) * 2009-08-31 2011-03-03 iAd Gesellschaft fur informatik, Automatisierung und Datenverarbeitung Method and system for transcoding regions of interests in video surveillance

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
ADITYA MAVLANKAR ET AL: "Region-of-interest prediction for interactively streaming regions of high resolution video", PACKET VIDEO 2007, IEEE, PI, 1 November 2007 (2007-11-01), pages 68 - 77, XP031170600, ISBN: 978-1-4244-0980-8 *
ADITYA MAVLANKAR ET AL: "Spatial-Random-Access-Enabled Video Coding for Interactive Virtual Pan/Tilt/Zoom Functionality", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 21, no. 5, 1 May 2011 (2011-05-01), pages 577 - 588, XP011321299, ISSN: 1051-8215, DOI: 10.1109/TCSVT.2011.2129170 *
ICHIMURA D ET AL: "A Tool for Interactive ROI Scalability", 74. MPEG MEETING; 17-10-2005 - 21-10-2005; NICE; (MOTION PICTUREEXPERT GROUP OR ISO/IEC JTC1/SC29/WG11),, no. M12501, 12 October 2005 (2005-10-12), XP030041171, ISSN: 0000-0243 *
PATRICE RONDAO ALFACE ET AL: "Interactive Omnidirectional Video Delivery: A Bandwidth-Effective Approach", BELL LABS TECHNICAL JOURNAL, vol. 16, no. 4, 1 March 2012 (2012-03-01), pages 135 - 147, XP055034585, ISSN: 1089-7089, DOI: 10.1002/bltj.20538 *
S. HEYMANN ET AL: "Representation, Coding and Interactive Rendering of High-Resolution Panoramic Images and Video using MPEG-4", PROC. PANORAMIC PHOTOGRAMMETRY WORKSHOP (PPW), 28 February 2005 (2005-02-28), Berlin, Germany, XP055034771 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2961182A1 (en) * 2014-06-27 2015-12-30 Alcatel Lucent Method, system and device for navigating in ultra high resolution video content by a client device

Also Published As

Publication number Publication date
KR20150003776A (ko) 2015-01-09
CN104365095A (zh) 2015-02-18
EP2645713A1 (en) 2013-10-02
CN104365095B (zh) 2018-04-27
US20150117524A1 (en) 2015-04-30
JP2015515201A (ja) 2015-05-21
JP6121518B2 (ja) 2017-04-26

Similar Documents

Publication Publication Date Title
EP2645713A1 (en) Method and apparatus for encoding a selected spatial portion of a video stream
US20220150488A1 (en) Encoder, decoder, encoding method, and decoding method
Gaddam et al. Tiling in interactive panoramic video: Approaches and evaluation
US11895316B2 (en) Encoder, decoder, encoding method, and decoding method
US10616592B2 (en) Method and encoder for encoding a video stream in a video coding format supporting auxiliary frames
US20190273931A1 (en) Encoder, decoder, encoding method, and decoding method
US20190273922A1 (en) Encoder, decoder, encoding method, and decoding method
de la Fuente et al. Delay impact on MPEG OMAF’s tile-based viewport-dependent 360 video streaming
US11463724B2 (en) Encoder, decoder, encoding method, and decoding method
US12309370B2 (en) Encoder, encoding method, decoder, and decoding method
US11616977B2 (en) Encoder, decoder, encoding method, and decoding method
CN112703737A (zh) 多向视频流的可扩展性
US11202093B2 (en) Encoder, decoder, encoding method, and decoding method
US20250024029A1 (en) Encoder, encoding method, decoder, and decoding method
US20190141335A1 (en) Decoder, encoder, decoding method, and encoding method
US11729398B2 (en) Encoder, encoding method, decoder, and decoding method
US11240530B2 (en) Encoder, decoder, encoding method, and decoding method
US20240406414A1 (en) Encoder, encoding method, decoder, and decoding method
US20230262252A1 (en) Encoder, decoder, encoding method, and decoding method
CA3057924A1 (en) System and method to optimize the size of a video recording or video transmission by identifying and recording a region of interest in a higher definition than the rest of the image that is saved or transmitted in a lower definition format
Yu et al. Convolutional neural network for intermediate view enhancement in multiview streaming
US11716470B2 (en) Encoder, decoder, encoding method, and decoding method
TWI901568B (zh) 編碼裝置及解碼裝置
CN119182922A (zh) 视频处理方法及装置
HK1184299A (en) Video bit stream transmission system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13711686

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2015502259

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 14389033

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 20147030447

Country of ref document: KR

Kind code of ref document: A

122 Ep: pct application non-entry in european phase

Ref document number: 13711686

Country of ref document: EP

Kind code of ref document: A1