WO2020057249A1 - 图像处理方法、装置、系统、网络设备、终端及存储介质 - Google Patents
图像处理方法、装置、系统、网络设备、终端及存储介质 Download PDFInfo
- Publication number
- WO2020057249A1 WO2020057249A1 PCT/CN2019/097355 CN2019097355W WO2020057249A1 WO 2020057249 A1 WO2020057249 A1 WO 2020057249A1 CN 2019097355 W CN2019097355 W CN 2019097355W WO 2020057249 A1 WO2020057249 A1 WO 2020057249A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- interest
- video
- image
- information
- image processing
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 51
- 230000015572 biosynthetic process Effects 0.000 claims abstract description 26
- 238000003786 synthesis reaction Methods 0.000 claims abstract description 26
- 239000002131 composite material Substances 0.000 claims description 44
- 238000012545 processing Methods 0.000 claims description 31
- 238000004891 communication Methods 0.000 claims description 19
- 239000000203 mixture Substances 0.000 claims description 18
- 238000004590 computer program Methods 0.000 claims description 15
- 230000004927 fusion Effects 0.000 claims description 13
- 238000013467 fragmentation Methods 0.000 claims description 5
- 238000006062 fragmentation reaction Methods 0.000 claims description 5
- 238000004458 analytical method Methods 0.000 claims description 4
- 230000000153 supplemental effect Effects 0.000 claims description 4
- 238000000034 method Methods 0.000 abstract description 26
- 230000008569 process Effects 0.000 abstract description 11
- 238000010586 diagram Methods 0.000 description 18
- 239000012634 fragment Substances 0.000 description 15
- 238000005516 engineering process Methods 0.000 description 8
- 230000009286 beneficial effect Effects 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 238000001914 filtration Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000007500 overflow downdraw method Methods 0.000 description 2
- 230000002194 synthesizing effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/4728—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for selecting a Region Of Interest [ROI], e.g. for requesting a higher resolution version of a selected region
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/119—Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/167—Position within a video image, e.g. region of interest [ROI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/184—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/234345—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements the reformatting operation being performed only on part of the stream, e.g. a region of the image or a time segment
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
- H04N21/440245—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display the reformatting operation being performed only on part of the stream, e.g. a region of the image or a time segment
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/816—Monomedia components thereof involving special video data, e.g 3D video
Definitions
- the embodiments of the present invention relate to, but are not limited to, the field of image coding and decoding technologies, and in particular, to but not limited to an image processing method, device, system, network device, terminal, and storage medium.
- Video applications are gradually developing from single-view, low-resolution, low-bit-rate to multi-view, high-resolution, and high-bit-rate, to provide users with new video content types and video presentation characteristics, and better presence And viewing experience.
- 360-degree panoramic video (hereinafter referred to as panoramic video) is a new type of video content. Users can choose a certain angle to watch according to their subjective needs, thereby achieving 360-degree full-viewing. Although the current network performance and hardware processing performance are high, but the dramatic increase in the number of users and the huge amount of panoramic video data, it is still necessary to reduce the network and hardware resource occupation on the premise of ensuring the user's viewing experience.
- the Region of Interest (ROI) technology can capture and display panoramic videos according to user preferences. It is not necessary to process all panoramic videos. However, in related technologies, there is usually only one ROI, which can only display the panorama in a limited manner. Part of the images in the video cannot meet the needs of users to watch multiple ROIs. Therefore, how to implement coding when there are multiple ROIs to indicate the composite display of each ROI needs to be solved urgently.
- ROI Region of Interest
- the image processing method, device, system, network device, terminal, and storage medium provided by the embodiments of the present invention mainly solve the technical problem of how to implement encoding when there are multiple ROIs.
- an embodiment of the present invention provides an image processing method, including:
- An embodiment of the present invention further provides an image processing method, including:
- An embodiment of the present invention further provides an image processing method, including:
- the network side obtains composite instruction information for indicating a composite display manner of each of the regions of interest in a video image, generates a media stream of the video image based on the composite instruction information, and sends the media stream to a target node;
- the target node receives the media stream, parses from the media stream to obtain synthesis instruction information of a region of interest, and controls playback display of a video stream in the media stream according to the synthesis instruction information.
- An embodiment of the present invention further provides an image processing apparatus, including:
- An acquisition module configured to acquire synthesis instruction information used to indicate a synthesis display manner of each of the regions of interest in a video image
- a processing module configured to generate a media stream of the video image based on the composition instruction information.
- An embodiment of the present invention further provides an image processing apparatus, including:
- a receiving module for receiving a video stream of video images and description data
- An analysis module configured to parse and obtain synthetic indication information of a region of interest from the description data
- a control module configured to control a composite playback display of an image of a region of interest in the video stream according to the composite instruction information.
- An embodiment of the present invention further provides an image processing system, which includes the two image processing devices described above.
- An embodiment of the present invention further provides a network device, including a first processor, a first memory, and a first communication bus;
- the first communication bus is configured to implement connection and communication between a first processor and a first memory
- the first processor is configured to execute one or more computer programs stored in a first memory to implement the steps of the image processing method according to any one of the preceding items.
- An embodiment of the present invention further provides a terminal including a second processor, a second memory, and a second communication bus;
- the second communication bus is configured to implement connection and communication between a second processor and a second memory
- the second processor is configured to execute one or more computer programs stored in a second memory to implement the steps of the image processing method as described above.
- An embodiment of the present invention further provides a storage medium, where the storage medium stores one or more programs, and the one or more programs can be executed by one or more processors to implement the image processing method described above. step.
- FIG. 1 is a schematic flowchart of an image processing method according to Embodiment 1 of the present invention
- FIG. 2 is a first schematic diagram of ROI image stitching indication according to the first embodiment of the present invention
- FIG. 3 is a second schematic diagram of ROI image stitching indication according to the first embodiment of the present invention.
- FIG. 4 is a third schematic diagram of a ROI image stitching indication according to the first embodiment of the present invention.
- FIG. 5 is a fourth schematic diagram of ROI image stitching indication according to the first embodiment of the present invention.
- FIG. 6 is a fifth schematic diagram of ROI image stitching indication according to the first embodiment of the present invention.
- FIG. 7 is a first schematic diagram of a ROI image fusion indication according to the first embodiment of the present invention.
- FIG. 8 is a second schematic diagram of a ROI image fusion indication according to the first embodiment of the present invention.
- FIG. 9 is a schematic diagram of overlapping regions of a ROI image according to the first embodiment of the present invention.
- FIG. 10 is a schematic diagram of an ROI image nesting instruction according to the first embodiment of the present invention.
- FIG. 11 is a schematic diagram of processing a transparent channel of an ROI image according to the first embodiment of the present invention.
- FIG. 12 is a schematic diagram of an ROI image coordinate position according to the first embodiment of the present invention.
- FIG. 13 is a first schematic diagram of generating a ROI image video stream according to the first embodiment of the present invention.
- FIG. 14 is a second schematic diagram of generating a ROI image video stream according to the first embodiment of the present invention.
- FIG. 15 is a schematic flowchart of an image processing method according to a second embodiment of the present invention.
- FIG. 16 is a schematic flowchart of an image processing method according to a third embodiment of the present invention.
- FIG. 17 is a schematic structural diagram of an image processing apparatus according to a fourth embodiment of the present invention.
- FIG. 18 is a schematic structural diagram of an image processing apparatus according to Embodiment 5 of the present invention.
- FIG. 19 is a schematic structural diagram of an image processing system according to a sixth embodiment of the present invention.
- FIG. 20 is a schematic structural diagram of a network device according to Embodiment 7 of the present invention.
- FIG. 21 is a schematic structural diagram of a terminal according to an eighth embodiment of the present invention.
- an embodiment of the present invention provides an image processing method, which is mainly applied to network-side devices, encoders, etc. However, it is not limited to devices such as servers and base stations. See FIG. 1, which includes the following steps:
- composition instruction information used to indicate a composite display manner between regions of interest in a video image.
- composition instruction information is obtained, which is used to indicate a composite display manner between ROIs in the video image. It should be understood that when the video image does not exist or the ROI is not divided, there is no process of obtaining the synthesis instruction information. Only when there are multiple ROIs, corresponding synthetic instruction information is acquired. In the case where there is only one ROI, this solution can also be adopted to control the display of the one ROI.
- the corresponding synthesis instruction information is acquired to indicate the synthesis display mode of the ROI.
- ROI including but not limited to:
- the video image can be analyzed in advance through image processing, ROI recognition and other technologies, and then the specific content or specific spatial position in the panoramic video can be divided through the analysis results to form different ROIs.
- a camera is used to track and shoot the ball's motion trajectory as an ROI; or to identify and track a specific target (such as a player) in the captured video image using ROI recognition technology to form an ROI .
- the user's area of interest information is collected, and specific content or specific spatial position in the panoramic video is automatically divided according to this information, thereby forming different areas of interest.
- the user selects the area of interest while watching the video image.
- a media stream of the video image is generated based on the obtained composite instruction information. That is, the synthesis instruction information is encoded and written into a code stream of the video image, thereby generating a media stream of the video image.
- the media stream can be decoded, and at least the ROIs in the video image can be synthesized and displayed for playback.
- At least one of the following instruction information is included:
- First instruction information for instructing mosaic display of regions of interest second instruction information for instructing fusion display of regions of interest, third instruction information for instructing nested display of regions of interest, use
- the fourth instruction information instructs zooming display of the region of interest
- the fifth instruction information instructs rotation display of the region of interest
- the sixth instruction information instructs clipping and display of the region of interest.
- the first indication information is used to instruct stitching of each ROI, so-called stitching, that is, two ROIs are adjacent and do not overlap.
- stitching that is, two ROIs are adjacent and do not overlap.
- areas A, B, C, and D are the four regions of interest in the video image. They have the same size and can be stitched together according to where they appear in the panorama.
- the areas A, B, C, and D may be spliced together at random positions or specified positions.
- the sizes of the A, B, C, and D regions may be inconsistent.
- the positions of the A, B, C, and D regions can be randomly arranged, and the sizes are also inconsistent.
- the A, B, C, and D regions can be formed into a non-rectangular arbitrary shape after being spliced.
- the second instruction information is used to instruct fusion of each ROI.
- the fusion causes a partially overlapping area between the two ROIs, but does not completely superimpose one of the ROIs on the other ROI.
- areas A, B, C, and D are four areas of interest in a video image, and they are overlapped and fused together in a specific range of areas.
- the composite display mode of the four regions of interest may be direct coverage of pixels in a fixed coverage order, and the order of superimposition is A ⁇ B ⁇ C ⁇ D. Therefore, the last covered D Covered by the other three ROIs.
- the pixel values of the overlapping partial regions generated by the fusion may be processed in the following manner: as shown in FIG. 9, the pixels of the overlapping partial regions of the four different ROI regions are calculated to generate new pixel values. For example, the average value of all pixels, or different weights for pixels in different regions, or new pixel values calculated according to the feature matching method, to obtain a natural image fusion effect.
- the feature matching method calculates new pixel values, and is usually applied to network-side devices with strong video processing capabilities to obtain the best fusion effect. It is also theoretically applicable to the terminal side, but it has higher requirements for terminal performance.
- the third type of instruction information is used to instruct nested display of the ROI, and the nested display is to completely overlap one ROI on the other.
- areas A and B are two areas of interest in the video image. B is completely overlapped on A and nested together.
- the nesting position can be set according to actual needs, such as the size of the image frame. , Overlay the relatively small ROI of the image frame on the relatively large ROI, or customize it according to the user.
- the fourth type of indication information is used to instruct the ROI to be scaled, and the scale is to change the size of the image.
- Including the zoom ratio value for example, when the zoom ratio value is 2, it can indicate that the diagonal length of the ROI is doubled to the original.
- the fifth type of instruction information is used to instruct the ROI to be rotated, including a rotation type and a rotation angle.
- the rotation type includes, but is not limited to, horizontal rotation and vertical rotation.
- the sixth instruction information is used to instruct the region of interest to be displayed.
- regions A and B are two regions of interest in the video image.
- the circular region in region B is intercepted, which can be implemented by using Alpha transparent channels.
- the extracted B and A may be nested to synthesize the image.
- the H.264 / AVC standard and the H.265 / HEVC (High Efficiency Video Coding) standard can be used to encode a video image.
- the obtained composite instruction information is written into a code stream of a video image.
- feature information of a corresponding ROI in the video image may also be obtained, and a media stream of the video image is generated based on the obtained composite instruction information and the feature information. That is, the above composite instruction information and feature information are simultaneously written into a code stream of a video image.
- the generated media stream includes at least the following two parts: description data and video stream.
- description data is mainly used to instruct decoding of a video stream to implement playback of a video image.
- the description data may include at least one of the following information: for example, time synchronization information, text information, and other related information.
- the description data as a part of the video image is optional and has the following two forms: First, it can be encoded with the video stream in the form of a code stream, that is, part of the data in the video stream; also It can be encoded separately from the video stream and separated from the video stream.
- the ROI characteristic information includes position information and / or coding quality indication information; wherein the position information includes coordinate information of a specific position of the ROI, and a length value and a width value of the ROI.
- the specific position may be a position of any one of the four corners of the ROI region, such as a pixel point in the upper left corner and a pixel point in the lower right corner; or the position of the center point of the ROI region.
- the encoding quality indication information can be the encoding quality level used in the encoding process. Different encoding quality indication information characterizes different encoding quality levels. After encoding through different encoding quality levels, the image quality produced is also different.
- the encoding quality indication information may be “1”, “2”, “3”, “4", “5", “6", and different values represent different encoding quality levels.
- the encoding quality indication information is “1” ", It means that low-quality coding is used; in contrast, when the coding quality indication information is” 2 ", it means that the medium-quality coding is better than" 1 "; the larger the value, the coding quality increases in order.
- the ROI position information may also be characterized as follows: Referring to FIG. 12, the upper side of the ROI region 121 is located at the 300th line of the video image, the lower side is located at the 600th line of the video image, and the left side The edge is in column 500 of the video image, and the right edge is in column 800 of the video image. That is, the position information of the ROI area is identified by its rank. For the image area of 1920 * 1080, the pixel position of the upper left corner is (0,0), and the pixel position of the lower right corner is (1919,1079).
- a Cartesian coordinate system may be used, and other non-Cartesian curvilinear coordinate systems may be used, such as a cylindrical surface, a spherical surface, or a polar coordinate system.
- the length value of the ROI is based on the above FIG. 12.
- the point can be used as the width value of the ROI. The opposite is also possible.
- table_id the identifier of the table
- roi_num contains the number of regions of interest
- roi_width width of the region of interest
- roi_height height of the region of interest
- roi_quality quality information of the region of interest
- relation_type synthetic indication information of the region of interest, 0 for stitching, 1 for embedding, and 2 for fusion;
- rotation rotation angle of the region of interest
- alpha_flag transparent channel identifier, 0 for the absence of transparent channel information, 1 for the presence of transparent channel information;
- alpha_info transparent channel information, combined with the region of interest (intercepted) to generate a new image
- filter_info When the relation_type is a fusion method, it can indicate the filtering method of the fusion region, such as the mean, median, etc .;
- user_data () user information.
- the above roi_info_table containing the ROI synthesis instruction information and characteristic information is written into the description data of the video image.
- the description data optionally, it includes at least one of the following: Supplemental Enhancement Information (SEI), video availability information (VideoUsability Information, VUI), system layer media attribute description unit.
- SEI Supplemental Enhancement Information
- VUI VideoUsability Information
- the roi_info_table is written into the supplemental enhancement information in the video bitstream.
- a specific example can be the structure shown in Table 2 below.
- roi_info_table contains relevant information (synthesis instruction information, feature information, etc.) of the corresponding ROI, which is written into the supplemental enhancement information, and the information whose identification information is ROI_INFO can be obtained from the SEI information.
- the information corresponding to the ROI_INFO is used as the identification information of the SEI information.
- roi_info_flag in Table 3 When roi_info_flag in Table 3 is equal to 1, it indicates that there is subsequent ROI information.
- roi_info_table () is also the roi_info_table data structure in Table 1 above, and contains ROI related information.
- the region of interest information whose identification information is roi_info_flag is 1 can be obtained from the VUI information.
- system layer media attribute description unit includes, but is not limited to, a descriptor for a transport stream, a data unit in a file format (for example, in a Box), and media description information for a transport stream (for example, a media presentation description) (Media, Presentation, Description, MPD).
- the ROI synthesis instruction information and feature information are written into the SEI, and it can further be combined with its temporal motion-constrained tile sets (MCTS).
- MCTS temporal motion-constrained tile sets
- the relevant ROI information and the use of H .265 / HEVC standard time-domain motion limited tile set combination can be combined with the tiles tightly.
- Combining the ROI synthesis instruction information with the tiles tightly can flexibly extract the required tile data without adding separate codec ROI data, which can meet the different needs of users and is more conducive to interacting with users in the application. As shown in Table 4 below.
- roi_info_flag 0 indicates that there is no related information of the region of interest, and 1 indicates that there is related information of the region of interest.
- roi_info An example of roi_info is shown in Table 5 below.
- roi_num contains the number of regions of interest
- roi_width width of the region of interest
- roi_height height of the region of interest
- roi_quality quality information of the region of interest
- relation_type the relation of the region of interest, 0 for stitching, 1 for embedding, 2 for fusion;
- rotation rotation angle of the region of interest
- alpha_flag transparent channel identifier, 0 for the absence of transparent channel information, 1 for the presence of transparent channel information;
- filter_info When the relation_type is a fusion method, it can indicate the filtering method of the fusion region, such as the mean, median, etc.
- video image data is included.
- the process of generating the video stream includes: obtaining a region of interest of the video image, dividing the associated image of each region of interest in the same image frame into at least one fragmentation unit and independently encoding to generate a first segment of the video image.
- a video stream is included.
- a first frame image of a video image is acquired, and an associated image of each ROI in the first frame image is determined. It is assumed that there are two ROIs in the video image, namely ROI131 and ROI132, and Assume that there is an associated image A1 of ROI131 and an associated image B1 of ROI132. At this time, the associated image A1 of ROI131 is divided into at least one fragment unit for independent encoding, and the associated image B1 of ROI132 is divided into at least one fragment unit.
- the related image A1 is divided into one fragment unit a11 for independent encoding
- the related image B1 is divided into two fragment units, which are separately encoded for b11 and b12.
- any existing encoding method can be used for encoding, which can be independent encoding or non-independent encoding.
- the finally generated first video stream includes at least all independently coded slice units of each ROI.
- the user when the user only needs to watch the ROI image, he can only extract the fragment units corresponding to the ROI in the first video stream (without extracting all the fragment units) and independently decode the fragment unit. , Without relying on other fragments to complete decoding, reducing the requirements on the receiving end decoding performance.
- ROI-related images in the video image can be encoded, and other regions except the associated image are not encoded, or the associated image and other regions are encoded separately.
- the slice unit includes slice slices of the H.264 / AVC standard, tile tiles of the H.265 / HEVC standard, and the like.
- the video stream in the media stream it may also be a second video stream, and the process of generating the second video stream is as follows: after synthesizing each associated image according to the composition instruction information, the image is processed as a frame to be processed, and The frame division is encoded in at least one slice unit to generate a second video stream of the region of interest.
- the second video stream is an associated image (C1 and D1, respectively) of the ROI (including ROI141 and ROI142 in FIG. 14) in the same image frame of the video image.
- the instruction information is synthesized. It is assumed here that stitching is combined, and the synthesized image is regarded as a to-be-processed image frame E1, and then the to-be-processed image frame E1 is divided into at least one fragment unit (such as e11) for encoding; encoding
- the method can adopt independent encoding, non-independent encoding, or other encoding methods.
- the other image frames in the video image are also processed in the above manner, and each image frame may be processed in parallel or serial manner. Thereby, the second video stream is generated.
- the second video stream can be processed by commonly used decoding methods.
- the synthesized ROI image can be obtained directly, and there is no need to combine the ROI-related images.
- This encoding method is beneficial to reduce the processing on the decoding side. Load to improve decoding efficiency. But it needs to be synthesized first when encoding.
- the network side or the encoding end may generate the above two types of video streams for the same video image.
- the generated media stream can be stored or sent to the corresponding target node. For example, when receiving a video image acquisition request from a target node, sending the media stream to the target node is triggered.
- sending the media stream to the target node is triggered.
- the identification information of the acquired content indicated by the acquisition request is parsed, and the media stream is sent to the target node according to the identification information.
- the server receives a request for a video image from the terminal, and according to the request, the media stream of the video image ( Including the first video stream and the description data) to the terminal.
- the terminal can decode the media stream to fully play the video image.
- the terminal can also decode the media stream, extract the segmentable unit data of the region of interest, and combine the description data to play and display the image of the region of interest.
- the fragment unit (without decoding operation) of the region of interest in the first video stream and the description data are extracted and sent to the target node; for example, the server side may receive the terminal pair
- the server finds the data corresponding to the region of interest that can independently encode the fragment unit according to the request information. After extraction, it adds the relevant information of the region of interest (synthesis instruction information and feature information) or the modified interest. Area information, generate a new code stream and send it to the terminal. Avoid sending all code streams to the terminal, reducing network bandwidth occupation and transmission delay.
- the second video stream and the description data are sent to the target node.
- the server can also choose to send the second video stream of the video image and the description data to the terminal according to the request sent by the terminal, and the terminal can directly obtain the synthesized ROI image after decoding it.
- the synthesis instruction information synthesizes the ROI, which is beneficial to reducing terminal resource occupation and improving terminal processing efficiency.
- the video image may be a 360-degree panoramic video, a stereo video, or the like.
- the relevant information of the ROI can be applied to the left and right fields of view at the same time.
- the image processing method provided by the embodiment of the present invention by writing the composite instruction information into a video image code stream, is used to indicate the composite display of the ROI image in the video image, thereby realizing the encoding process in which multiple ROIs exist in the video image. It meets the user's viewing requirements for viewing multiple ROI images at the same time.
- the decoding end can be decoded independently without relying on other fragments to achieve decoding.
- an embodiment of the present invention provides an image processing method, which is mainly applied to a terminal, a decoder, and the like, including but not limited to a mobile phone, a personal computer, and the like.
- the image processing method includes the following steps:
- the synthetic instruction information of the region of interest information is extracted.
- characteristic information of the ROI may also be obtained from the description data, including position information and coding quality indication information.
- ROI image data that is, video stream data is obtained.
- S153 Control the composite playback display of the image of the region of interest in the video stream according to the composite instruction information.
- the ROI image is synthesized and displayed according to the composition instruction information.
- the method before receiving the video stream of the video image and the description data, the method further includes sending an acquisition request to a network side (or an encoding end), and the acquisition request may further include identification information for indicating acquisition content. To get different video streams.
- the identification information when the identification information is set to the first identification, it can be used to instruct the acquisition of the first video stream and description data of the corresponding video image; when the identification information is set to the second identification, it can be used to instruct the first acquisition of the corresponding video image The fragmentation unit and description data of the region of interest in the video stream; when the identification information is set to the third identification, it can be used to instruct the second video stream and description data to obtain the corresponding video image.
- the media streams received by the network side are different, and there are corresponding differences in subsequent processing.
- the identification information in the acquisition request is the first identification
- the first video stream and description data of the corresponding video image will be obtained.
- the first video stream and description data can be decoded to obtain the complete image of the video image.
- the independently-encoded slice unit data of the ROI image in the first video stream is extracted, and the ROI image is synthesized and displayed according to the ROI synthesis instruction information in the description data.
- the terminal can directly perform the decoding operation on the independently decodable ROI of the ROI. After synthesizing the ROI image with the composition instruction information in the description data, the display is displayed.
- the terminal can directly decode it using conventional decoding methods to obtain a synthesized ROI image, and then perform Play display.
- the acquisition request is not limited to including identification information for indicating the acquisition of content, but also includes other necessary information, such as the address information of the local end and the opposite end, identification information of the video image requested to be acquired, and verification information.
- an embodiment of the present invention provides an image processing method, which is mainly applied to a system including a network side and a terminal side.
- the image processing method mainly includes the following steps:
- the network side obtains composite instruction information for indicating a composite display manner of each region of interest in the video image.
- the network side generates a media stream of the video image based on the synthesis instruction information.
- S163 The network side sends the media stream to the target node.
- the target node receives the media stream.
- the target node parses the composite indication information of the region of interest from the media stream.
- the target node controls the playback display of the video stream in the media stream according to the composite instruction information.
- Embodiment 1 For details, refer to related descriptions in Embodiment 1 and / or Embodiment 2, and details are not described herein again.
- the media stream generated by the network side and the media stream sent by the network side to the target node may be the same or different.
- the network side may flexibly select a video stream to be sent to the target node according to the acquisition request of the target node, instead of a specific video stream. Therefore, the media stream generated by the network side can be used as the first media stream, and the media stream sent to the target node can be used as the second media stream, so as to facilitate differentiation.
- Embodiment 4 is a diagrammatic representation of Embodiment 4:
- the embodiment of the present invention provides an image processing apparatus, which is used to implement the steps of the image processing method according to the first embodiment.
- the image processing apparatus includes:
- the obtaining module 171 is configured to obtain composite instruction information for indicating a composite display manner of each of the regions of interest in the video image; and the processing module 172 is configured to generate a media stream of the video image based on the composite instruction information.
- the processing module 172 is configured to generate a media stream of the video image based on the composite instruction information.
- Embodiment 5 is a diagrammatic representation of Embodiment 5:
- the embodiment of the present invention provides an image processing apparatus, which is used to implement the steps of the image processing method according to the second embodiment.
- the image processing apparatus includes:
- a receiving module 181, configured to receive a video stream of video images and description data
- An analysis module 182 configured to parse and obtain synthesis indication information of a region of interest from the description data
- the control module 183 is configured to control the playback display of the image of the region of interest in the video stream according to the composition instruction information.
- Embodiment 6 is a diagrammatic representation of Embodiment 6
- an embodiment of the present invention provides an image processing system including an image processing device 191 as the fourth embodiment and an image processing device 192 as described in the fifth embodiment, as shown in FIG. 19.
- the image processing system is used to implement the image processing method described in the third embodiment.
- Embodiment 7 is a diagrammatic representation of Embodiment 7:
- the embodiment of the present invention provides a network device. Referring to FIG. 20, it includes a first processor 201, a first memory 202, and a first communication bus 203.
- the first communication bus 203 is used to implement connection and communication between the first processor 201 and the first memory 202.
- the first processor 201 is configured to execute one or more computer programs stored in the first memory 202 to implement the steps of the image processing method described in the first embodiment. For details, refer to the description in the first embodiment, and details are not described herein again.
- Embodiment 8 is a diagrammatic representation of Embodiment 8
- the embodiment of the present invention provides a terminal.
- the terminal includes a second processor 211, a second memory, and a second communication bus 213.
- the second communication bus 213 is used to implement connection and communication between the second processor 211 and the second memory 212.
- the second processor 211 is configured to execute one or more computer programs stored in the second memory 212 to implement the steps of the image processing method described in the second embodiment. For details, refer to the description in the second embodiment, and details are not described herein again.
- the embodiment of the present invention provides a storage medium.
- the storage medium may be a computer-readable storage medium.
- the storage medium stores one or more computer programs, and the one or more computers.
- the program may be executed by one or more processors to implement the steps of the image processing method as described in the first embodiment or the second embodiment.
- the storage medium includes volatile or nonvolatile, removable or non-removable implemented in any method or technology used to store information such as computer-readable instructions, data structures, computer program modules or other data.
- Storage media include, but are not limited to, RAM (Random Access Memory, Random Access Memory), ROM (Read-Only Memory, Read-Only Memory), EEPROM (Electrically Erasable Programmable Read Only Memory) Or other memory technology, CD-ROM (Compact Disc Read-Only Memory), digital versatile disc (DVD) or other optical disc storage, magnetic box, magnetic tape, disk storage or other magnetic storage device, or can be used Any other medium for storing the desired information and accessible by the computer.
- This embodiment also provides a computer program (also referred to as computer software), which can be distributed on a computer-readable medium and executed by a computable device to implement the above-mentioned first embodiment and / or the second embodiment. At least one step of the image processing method; and in some cases, at least one step shown or described may be performed in an order different from that described in the above embodiments.
- a computer program also referred to as computer software
- This embodiment also provides a computer program product including a computer-readable device, where the computer-readable device stores the computer program as shown above.
- the computer-readable device in this embodiment may include a computer-readable storage medium as shown above.
- composition instruction information for indicating a composition display manner between regions of interest in a video image, based on the composition instruction Information to generate the media stream of the video image; that is, to write the composite instruction information into the code stream of the video image, to realize the encoding process of the video image when there are multiple (at least two) ROIs, and during video playback, the Based on the composite instruction information, the composite display and playback of each ROI can be controlled, which can meet the user's need to watch multiple ROIs at the same time.
- the technical effects including but not limited to the above can be achieved.
- a communication medium typically contains computer-readable instructions, data structures, computer program modules, or other data in a modulated data signal such as a carrier wave or other transmission mechanism, and may include any information delivery medium. Therefore, the present invention is not limited to any specific combination of hardware and software.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
Claims (21)
- 一种图像处理方法,其特征在于,包括:获取用于指示视频图像中感兴趣区域之间的合成显示方式的合成指示信息;基于所述合成指示信息生成所述视频图像的媒体流。
- 如权利要求1所述的图像处理方法,其特征在于,所述图像处理方法还包括:获取各所述感兴趣区域的特征信息;所述基于所述合成指示信息生成所述视频图像的媒体流包括:基于所述合成指示信息以及所述特征信息,生成所述视频图像的媒体流。
- 如权利要求1所述的图像处理方法,其特征在于,所述合成指示信息包括如下至少一种:用于指示将所述感兴趣区域进行拼接显示的第一指示信息、用于指示将所述感兴趣区域进行融合显示的第二指示信息、用于指示将所述感兴趣区域进行嵌套显示的第三指示信息、用于指示将所述感兴趣区域进行缩放显示的第四指示信息、用于指示将所述感兴趣区域进行旋转显示的第五指示信息、用于指示将所述感兴趣区域进行截取显示的第六指示信息。
- 如权利要求2所述的图像处理方法,其特征在于,所述特征信息包括位置信息和/或编码质量指示信息;所述位置信息包括所述感兴趣区域特定位置的坐标信息、以及所述感兴趣区域的长度值和宽度值。
- 如权利要求2所述的图像处理方法,其特征在于,所述媒体流包括描述数据;所述基于所述合成指示信息以及所述特征信息,生成所述视频图像的媒体流包括:将所述合成指示信息以及所述特征信息写入到所述描述数据中,以生成所述媒体流。
- 如权利要求5所述的图像处理方法,其特征在于,所述描述数据包括如下至少一种:补充增强信息、视频可用性信息、系统层媒体属性描述单元。
- 如权利要求5所述的图像处理方法,其特征在于,所述媒体流还包括视频流,所述图像处理方法还包括:获取所述视频图像的所述感兴趣区域,将同一图像帧中各所述感兴趣区域的关联图像,划分在至少一个分片单元中进行独立编码,以生成所述视频图像的第一视频流。
- 如权利要求7所述的图像处理方法,其特征在于,所述图像处理方法还包括:将各所述关联图像按照所述合成指示信息进行合成后作为一张待处理图像帧,将所述待处理图像帧划分在至少一个分片单元中进行编码,以生成所述感兴趣区域的第二视频流。
- 如权利要求8所述的图像处理方法,其特征在于,图像处理方法还包括:将媒体流进行存储或发送给目标节点。
- 如权利要求9所述的图像处理方法,其特征在于,所述将所述媒体流发送给目标节点之前,还包括:接收到所述目标节点对所述视频图像的获取请求。
- 如权利要求10所述的图像处理方法,其特征在于,所述将所述媒体流发送给目标节点包括:解析所述获取请求所指示获取内容的标识信息,根据所述标识信息将所述媒体流发送给所述目标节点。
- 如权利要求11所述的图像处理方法,其特征在于,所述根据所述标识信息将所述媒体流发送给所述目标节点包括:在所述标识信息为第一标识时,将所述第一视频流以及所述描述数据发送给所述目标节点;在所述标识信息为第二标识时,提取所述第一视频流中感兴趣区域的分片单元以及所述描述数据,发送给所述目标节点;在所述标识信息为第三标识时,将所述第二视频流以及所述描述数据发送给所述目标节点。
- 如权利要求1-12任一项所述的图像处理方法,其特征在于,所述视频图像为全景视频图像。
- 一种图像处理方法,其特征在于,包括:接收视频图像的视频流以及描述数据;从所述描述数据中解析得到感兴趣区域的合成指示信息;根据所述合成指示信息控制所述视频流中感兴趣区域图像的合成播放显示。
- 一种图像处理方法,其特征在于,包括:网络侧获取用于指示视频图像中各所述感兴趣区域的合成显示方式的合成指示信息,基于所述合成指示信息生成所述视频图像的媒体流,并将媒体流发送给目标节点;所述目标节点接收所述媒体流,从所述媒体流中解析得到感兴趣区域的合成指示信息,根据所述合成指示信息控制所述媒体流中视频流的播放显示。
- 一种图像处理装置,其特征在于,包括:获取模块,用于获取用于指示视频图像中各所述感兴趣区域的合成显示方式的合成指示信息;处理模块,用于基于所述合成指示信息生成所述视频图像的媒体流。
- 一种图像处理装置,其特征在于,包括:接收模块,用于接收视频图像的视频流以及描述数据;解析模块,用于从所述描述数据中解析得到感兴趣区域的合成指示信息;控制模块,用于根据所述合成指示信息控制所述视频流中感兴趣区域图像的合成播放显示。
- 一种图像处理系统,其特征在于,包括:如权利要求16所述的图像处理装置以及如权利要求17所述的图像处理装置。
- 一种网络设备,其特征在于,包括第一处理器、第一存储器及第一通信总线;所述第一通信总线用于实现第一处理器和第一存储器之间的连接通信;所述第一处理器用于执行第一存储器中存储的一个或者多个计算机程序,以实现如权利要求1至13中任一项所述的图像处理方法的步骤。
- 一种终端,其特征在于,包括:第二处理器、第二存储器及第二通信总线;所述第二通信总线用于实现第二处理器和第二存储器之间的连接通信;所述第二处理器用于执行第二存储器中存储的一个或者多个计算机程序,以实现如权利要求14中所述的图像处理方法的步骤。
- 一种存储介质,其特征在于,所述存储介质存储有一个或者多个计算机程序,所述一个或者多个计算机程序可被一个或者多个处理器执行,以实现如权利要求1至13中任一项,或如权利要求14中所述的图片处理方法的步骤。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020217011376A KR102649812B1 (ko) | 2018-09-19 | 2019-07-23 | 이미지 처리 방법, 장치, 시스템, 네트워크 기기, 단말기 및 저장 매체 |
EP19862435.5A EP3855750A4 (en) | 2018-09-19 | 2019-07-23 | IMAGE PROCESSING PROCESS, APPARATUS AND SYSTEM, NETWORK DEVICE, TERMINAL AND RECORDING MEDIA |
JP2021515166A JP7425788B2 (ja) | 2018-09-19 | 2019-07-23 | 画像処理方法、装置、システム、ネットワーク機器、端末及びコンピュータプログラム |
US17/276,572 US12058446B2 (en) | 2018-09-19 | 2019-07-23 | Method and system of image processing with composition and display of regions of interest |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811095593.9 | 2018-09-19 | ||
CN201811095593.9A CN110933461B (zh) | 2018-09-19 | 2018-09-19 | 图像处理方法、装置、系统、网络设备、终端及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020057249A1 true WO2020057249A1 (zh) | 2020-03-26 |
Family
ID=69856069
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/097355 WO2020057249A1 (zh) | 2018-09-19 | 2019-07-23 | 图像处理方法、装置、系统、网络设备、终端及存储介质 |
Country Status (6)
Country | Link |
---|---|
US (1) | US12058446B2 (zh) |
EP (1) | EP3855750A4 (zh) |
JP (1) | JP7425788B2 (zh) |
KR (1) | KR102649812B1 (zh) |
CN (2) | CN115883882A (zh) |
WO (1) | WO2020057249A1 (zh) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113965749A (zh) * | 2020-12-14 | 2022-01-21 | 深圳市云数链科技有限公司 | 静态摄像机视频传输方法及系统 |
CN113206853B (zh) * | 2021-05-08 | 2022-07-29 | 杭州当虹科技股份有限公司 | 一种视频批改结果保存改进方法 |
CN113573059B (zh) * | 2021-09-23 | 2022-03-01 | 中兴通讯股份有限公司 | 图像显示方法、装置、存储介质及电子装置 |
CN118101875A (zh) * | 2022-11-22 | 2024-05-28 | 中兴通讯股份有限公司 | 多媒体处理方法、媒体服务器和存储介质 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006352539A (ja) * | 2005-06-16 | 2006-12-28 | Sharp Corp | 広視野映像システム |
CN1889686A (zh) * | 2006-07-14 | 2007-01-03 | 北京时越网络技术有限公司 | 一种同时显示多路视频信息的方法 |
CN101521745A (zh) * | 2009-04-14 | 2009-09-02 | 王广生 | 一组多镜头光心重合式全方位摄像装置及全景摄像、转播的方法 |
CN102265626A (zh) * | 2008-12-22 | 2011-11-30 | 韩国电子通信研究院 | 传送有关立体图像的数据的方法、重放立体图像的方法、和创建立体图像的文件的方法 |
CN105578204A (zh) * | 2014-10-14 | 2016-05-11 | 青岛海信电器股份有限公司 | 一种多视频数据显示的方法及装置 |
CN106331732A (zh) * | 2016-09-26 | 2017-01-11 | 北京疯景科技有限公司 | 生成、展现全景内容的方法及装置 |
CN108322727A (zh) * | 2018-02-28 | 2018-07-24 | 北京搜狐新媒体信息技术有限公司 | 一种全景视频传输方法及装置 |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000165641A (ja) * | 1998-11-24 | 2000-06-16 | Matsushita Electric Ind Co Ltd | 画像処理方法,画像処理装置およびデータ記憶媒体 |
US8184153B2 (en) * | 2005-09-26 | 2012-05-22 | Electronics And Telecommunications Research Institute | Method and apparatus for defining and reconstructing ROIs in scalable video coding |
KR101255226B1 (ko) * | 2005-09-26 | 2013-04-16 | 한국과학기술원 | 스케일러블 비디오 코딩에서 다중 roi 설정, 복원을위한 장치 및 방법 |
US9691098B2 (en) * | 2006-07-07 | 2017-06-27 | Joseph R. Dollens | Method and system for managing and displaying product images with cloud computing |
JP5194679B2 (ja) * | 2007-09-26 | 2013-05-08 | 日産自動車株式会社 | 車両用周辺監視装置および映像表示方法 |
US20120140067A1 (en) * | 2010-12-07 | 2012-06-07 | Scott Crossen | High Definition Imaging Over Legacy Surveillance and Lower Bandwidth Systems |
JP5870835B2 (ja) * | 2012-04-27 | 2016-03-01 | 富士通株式会社 | 動画像処理装置、動画像処理方法および動画像処理プログラム |
US9827487B2 (en) * | 2012-05-14 | 2017-11-28 | Sphero, Inc. | Interactive augmented reality using a self-propelled device |
US9497405B2 (en) * | 2012-07-17 | 2016-11-15 | Nec Display Solutions, Ltd. | Display device for displaying videos side by side without overlapping each other and method for the same |
JP6141084B2 (ja) * | 2013-04-19 | 2017-06-07 | キヤノン株式会社 | 撮像装置 |
CN103402087A (zh) * | 2013-07-23 | 2013-11-20 | 北京大学 | 一种基于可分级位流的视频编解码方法 |
WO2015197815A1 (en) * | 2014-06-27 | 2015-12-30 | Koninklijke Kpn N.V. | Determining a region of interest on the basis of a hevc-tiled video stream |
JP2016048839A (ja) | 2014-08-27 | 2016-04-07 | 株式会社小糸製作所 | 電子制御ユニットおよび車両用映像システム |
-
2018
- 2018-09-19 CN CN202310035569.0A patent/CN115883882A/zh active Pending
- 2018-09-19 CN CN201811095593.9A patent/CN110933461B/zh active Active
-
2019
- 2019-07-23 JP JP2021515166A patent/JP7425788B2/ja active Active
- 2019-07-23 KR KR1020217011376A patent/KR102649812B1/ko active IP Right Grant
- 2019-07-23 WO PCT/CN2019/097355 patent/WO2020057249A1/zh unknown
- 2019-07-23 US US17/276,572 patent/US12058446B2/en active Active
- 2019-07-23 EP EP19862435.5A patent/EP3855750A4/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006352539A (ja) * | 2005-06-16 | 2006-12-28 | Sharp Corp | 広視野映像システム |
CN1889686A (zh) * | 2006-07-14 | 2007-01-03 | 北京时越网络技术有限公司 | 一种同时显示多路视频信息的方法 |
CN102265626A (zh) * | 2008-12-22 | 2011-11-30 | 韩国电子通信研究院 | 传送有关立体图像的数据的方法、重放立体图像的方法、和创建立体图像的文件的方法 |
CN101521745A (zh) * | 2009-04-14 | 2009-09-02 | 王广生 | 一组多镜头光心重合式全方位摄像装置及全景摄像、转播的方法 |
CN105578204A (zh) * | 2014-10-14 | 2016-05-11 | 青岛海信电器股份有限公司 | 一种多视频数据显示的方法及装置 |
CN106331732A (zh) * | 2016-09-26 | 2017-01-11 | 北京疯景科技有限公司 | 生成、展现全景内容的方法及装置 |
CN108322727A (zh) * | 2018-02-28 | 2018-07-24 | 北京搜狐新媒体信息技术有限公司 | 一种全景视频传输方法及装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3855750A4 * |
Also Published As
Publication number | Publication date |
---|---|
EP3855750A4 (en) | 2021-10-20 |
CN110933461B (zh) | 2022-12-30 |
KR20210059759A (ko) | 2021-05-25 |
EP3855750A1 (en) | 2021-07-28 |
CN115883882A (zh) | 2023-03-31 |
JP7425788B2 (ja) | 2024-01-31 |
CN110933461A (zh) | 2020-03-27 |
US12058446B2 (en) | 2024-08-06 |
KR102649812B1 (ko) | 2024-03-21 |
JP2022501902A (ja) | 2022-01-06 |
US20220053127A1 (en) | 2022-02-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11109013B2 (en) | Method of transmitting 360-degree video, method of receiving 360-degree video, device for transmitting 360-degree video, and device for receiving 360-degree video | |
US11284124B2 (en) | Spatially tiled omnidirectional video streaming | |
CN108476324B (zh) | 增强视频流的视频帧中的感兴趣区域的方法、计算机和介质 | |
KR102357137B1 (ko) | 이미지 처리 방법, 단말기, 및 서버 | |
WO2020057249A1 (zh) | 图像处理方法、装置、系统、网络设备、终端及存储介质 | |
CN110121734B (zh) | 一种信息的处理方法及装置 | |
CN112204993B (zh) | 使用重叠的被分区的分段的自适应全景视频流式传输 | |
US20200112710A1 (en) | Method and device for transmitting and receiving 360-degree video on basis of quality | |
US10757463B2 (en) | Information processing apparatus and information processing method | |
US11694303B2 (en) | Method and apparatus for providing 360 stitching workflow and parameter | |
JP2022133439A (ja) | メディアコンテンツを送信するための方法、装置及びコンピュータプログラム | |
US20180176650A1 (en) | Information processing apparatus and information processing method | |
WO2019137313A1 (zh) | 一种媒体信息的处理方法及装置 | |
CA3018600C (en) | Method, apparatus and stream of formatting an immersive video for legacy and immersive rendering devices | |
TW201841499A (zh) | 用於軌道合成的方法以及裝置 | |
CN110637463B (zh) | 360度视频处理方法 | |
KR102499900B1 (ko) | 고해상도 영상의 스트리밍을 위한 영상 전송 장치와 영상 재생 장치 및 그 동작 방법 | |
WO2023194648A1 (en) | A method, an apparatus and a computer program product for media streaming of immersive media |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19862435 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021515166 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20217011376 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2019862435 Country of ref document: EP Effective date: 20210419 |