CN110858896A - VR image processing method - Google Patents
VR image processing method Download PDFInfo
- Publication number
- CN110858896A CN110858896A CN201810975482.0A CN201810975482A CN110858896A CN 110858896 A CN110858896 A CN 110858896A CN 201810975482 A CN201810975482 A CN 201810975482A CN 110858896 A CN110858896 A CN 110858896A
- Authority
- CN
- China
- Prior art keywords
- image
- focus area
- processing method
- image processing
- profile
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/277—Analysis of motion involving stochastic approaches, e.g. using Kalman filters
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Image Processing (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The application discloses a VR image processing method, which comprises the following steps: determining an image needing to be processed; selecting at least one focus area; determining a focus area profile of the focus area; performing fusion processing on the focus area and the background image according to the contour of the focus area; and transmitting the fused image to a VR head display. The application has reached the high definition part of guaranteeing to transmit the image and partly links up naturally with non-high definition, the technological effect of effectual promotion user experience's comfort level.
Description
Technical Field
The application relates to the technical field of VR image transmission, in particular to a VR image processing method.
Background
The requirement for resolution in image transmission is higher and higher nowadays, especially in the VR image transmission field, because VR needs binocular output, there is a demand for double-size image output in transmission. According to the feasible scheme at present, if the transmitting end uses X264 coding compression and then uses h264 decoding at the receiving end, the compressed video stream data can be transmitted in real time, and the network transmission data volume can be effectively reduced. However, in the field of network transmission, there are situations that high-definition and ultra-high-definition videos are transmitted to cause delay aggravation and the possibility that the videos cannot be observed instantly, particularly, the transmission pressure of the 4K or 8K videos is higher, the higher the compression ratio of the corresponding video coding field is, the lower the definition is, the possibility that the videos can be transmitted instantly and watched by a client cannot be met, and the requirement of VR binocular output panoramic high-definition images cannot be met.
Disclosure of Invention
The application aims to provide a VR image processing method, which ensures that a high-definition part and a non-high-definition part of a transmission image are naturally linked, and effectively improves the comfort level of user experience.
In order to achieve the above object, the present application provides a VR image processing method, including the steps of: determining an image needing to be processed; selecting at least one focus area; determining a focus area profile of the focus area; performing fusion processing on the focus area and the background image according to the contour of the focus area; and transmitting the fused image to a VR head display.
Preferably, the focal region profile is determined using the Camshift algorithm.
Preferably, the method for determining the profile of the focal region specifically comprises: determining a first edge profile; determining a second edge profile; and storing the first edge outline and the second edge outline into a focus object file library.
Preferably, the second edge profile is determined from the first edge profile inward contraction calculation.
Preferably, the inward shrinkage calculation formula of each pixel on the first edge profile is as follows: the shrunk pixel value is (first contour pixel value-center point) × shrinkage factor.
Preferably, the shrinkage factor is 0.6 to 0.9.
Preferably, the shrinkage factor is 0.8.
Preferably, the transmission method of the fused image is as follows: transmitting the focus area in high definition and synchronizing the background image in non-high definition.
The beneficial effect that this application realized is as follows:
(1) the high-definition part and the non-high-definition part of the transmission image are naturally connected, and the comfort level of user experience is effectively improved.
(2) The technical effects of reducing the data volume of network transmission, avoiding transmission delay and meeting the requirement that a client watches a high-definition video can be achieved.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings needed to be used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments described in the present application, and other drawings can be obtained by those skilled in the art according to the drawings.
FIG. 1 is a flow diagram of one embodiment of a VR image processing method;
FIG. 2 is a flow chart of one embodiment of a method for determining a focal region profile.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
As shown in fig. 1, the present invention provides a VR image processing method, including the following steps:
s110: and playing the video and determining the images needing to be processed in the video.
Specifically, the staff plays the video through the video playing tool, in the playing process, the staff pauses the video at a position where a focus needs to be set, and the frame of image displayed on the video playing tool after pausing is the image needing to be processed.
S120: at least one focal region is selected from the image to be processed.
Specifically, the staff performs framing on at least one region in the image by using the mouse, the framed region is a focus region, and after the focus region is determined, S130 is performed.
S130: a focal region profile of the focal region is determined.
Further: as an embodiment, the focus area profile is determined using the CamShift algorithm (continuous adaptive mean-Shift algorithm).
Specifically, a Camshift algorithm (continuous Adaptive Mean-Shift algorithm) is an improved algorithm for the Meansshift algorithm, and can adjust the size of a Search Window in real time along with the change of the size of a tracking target in the tracking process, make all frames of a video image into the Meansshift algorithm, use the result of the previous frame (i.e. the centroid and size of the Search Window) as the initial value of the Search Window of the Meansshift algorithm of the next frame to find the optimal iteration result, use the centroid in the optimal iteration result in the zero-order moment M00 and the first-order moment M10 of the Window, and calculate the size of the zero-order moment in M01, where the zero-order moment is the integral of all pixels in the Search Window, and the integral of all pixels in the Search Window is equal to the size of the Search Window, and the size of the Search Window is the focus area.
Further, the operations of Camshift and MeanShift are performed on the back projection image, a histogram of the H component of the target region (i.e., the focal region) is generated through back projection calculation, and the pixel value of each pixel point on the target region is replaced by the value of the histogram bin corresponding to the bin where the current pixel value is located. After reverse projection, most pixel values of a search window of a target area are normalized to be a maximum value of 255, a threshold value is determined according to a test result in use, if the calculated zero-order moment is larger than the threshold value, the target is paved on the whole search window, whether the target area exists in an area outside the search window needs to be judged, and if the target area exists, the size of the search window needs to be increased; if the zeroth order moment is less than the threshold, the search window needs to be reduced in size. Therefore, when the size of the target changes, the size of the search window can be adjusted in a self-adaptive mode through the Camshift algorithm, the target area is tracked, and the outline of the focus area is obtained.
Further, as shown in fig. 2, as another embodiment, the method for determining the focal region profile includes the following steps:
s210: a first edge profile is determined.
Specifically, the worker uses a mouse to draw a first edge contour on the image to be processed, and the area surrounded by the first edge contour is larger than the focus area to be framed.
S220: a second edge profile is determined.
Specifically, a second edge profile is computed from the first edge profile inward scaling.
Specifically, all paths of all pixel points on the track are calculated according to the moving track of the mouse when the staff draws the first edge profile. And calculating the average value of all paths to obtain the central point. And (3) inwards shrinking all the pixel points to calculate to obtain a second edge profile, wherein the calculation formula is as follows:
pixel value after shrinkage (first contour pixel value-center point) x shrinkage factor
The value range of the shrinkage coefficient is 0.6-0.9, the value of the shrinkage coefficient is a value obtained by a user according to real-time test, and preferably, the value of the coefficient is 0.8. The connecting line of the shrunk pixel values of all the pixels on the image to be processed is the second edge contour.
And S230, storing the first edge contour and the second edge contour into a focus object file library.
Specifically, the data such as the image size, the focus area, the first edge profile, the second edge profile, and the like of each frame of image are transmitted to the focus object file library for storage.
S140: and carrying out fusion processing on the focus area and the background image according to the contour of the focus area.
When the video is played, the computer reads the data of the focus object, the first edge contour, the second edge contour, the image size and the like corresponding to each frame of image in the played video from the focus object file library. And synchronously transmitting the definition data, the first edge profile, the second edge profile and the video data to the VR client. After receiving the data, the VR client renders a non-high-definition image as a background image by using the original image, then renders a layer of high-definition image of a focus area in the corresponding focus area on the background image, and gradually fuses to the second edge outline according to the first edge outline to form a fused area, so that the processed image is ensured to have comfortable visual sense when displayed. After the fusion process of the focus area and the background image is completed, S150 is performed.
S150: and transmitting the fused image to a VR head display.
Further, the transmission method of the fused image is as follows: transmitting the focus area in high definition and synchronizing the background image in non-high definition.
The beneficial effect that this application realized is as follows:
(1) the high-definition part and the non-high-definition part of the transmission image are naturally connected, and the comfort level of user experience is effectively improved.
(2) The technical effects of reducing the data volume of network transmission, avoiding transmission delay and meeting the requirement that a client watches a high-definition video can be achieved.
While the preferred embodiments of the present application have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all alterations and modifications as fall within the scope of the application. It will be apparent to those skilled in the art that various changes and modifications may be made in the present application without departing from the spirit and scope of the application. Thus, if such modifications and variations of the present application fall within the scope of the claims of the present application and their equivalents, the present application is intended to include such modifications and variations as well.
Claims (8)
1. A VR image processing method is characterized by comprising the following steps:
determining an image needing to be processed;
selecting at least one focus area;
determining a focus area profile of the focus area;
performing fusion processing on the focus area and the background image according to the contour of the focus area;
and transmitting the fused image to a VR head display.
2. The VR image processing method of claim 1, wherein the focus area profile is determined using a CamShift algorithm.
3. The VR image processing method of claim 1, wherein the method of determining the focus area profile is specifically:
determining a first edge profile;
determining a second edge profile;
and storing the first edge outline and the second edge outline into a focus object file library.
4. The VR image processing method of claim 3, wherein the second edge profile is determined from a first edge profile inward scaling calculation.
5. The VR image processing method of claim 4, wherein the inward shrinkage for each pixel on the first edge profile is calculated by: the shrunk pixel value is (first contour pixel value-center point) × shrinkage factor.
6. The VR image processing method of claim 5 wherein the shrinkage factor is 0.6 to 0.9.
7. The VR image processing method of claim 6, wherein the contraction coefficient is 0.8.
8. The VR image processing method of claim 1, wherein the fused image is transmitted by: transmitting the focus area in high definition and synchronizing the background image in non-high definition.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810975482.0A CN110858896B (en) | 2018-08-24 | 2018-08-24 | VR image processing method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810975482.0A CN110858896B (en) | 2018-08-24 | 2018-08-24 | VR image processing method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110858896A true CN110858896A (en) | 2020-03-03 |
CN110858896B CN110858896B (en) | 2021-06-08 |
Family
ID=69636438
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810975482.0A Active CN110858896B (en) | 2018-08-24 | 2018-08-24 | VR image processing method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110858896B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103249349A (en) * | 2010-12-02 | 2013-08-14 | 奥林巴斯株式会社 | Endoscopic image processing apparatus and program |
CN106446894A (en) * | 2016-09-27 | 2017-02-22 | 广东技术师范学院 | Method for recognizing position of spherical object based on contour |
CN106959964A (en) * | 2016-01-12 | 2017-07-18 | 阿里巴巴集团控股有限公司 | Interface background display methods |
WO2017168229A1 (en) * | 2016-04-01 | 2017-10-05 | Linear Albra Technologies Limited | Systems and methods for head-mounted display adapted to human visual mechanism |
CN108848389A (en) * | 2018-07-27 | 2018-11-20 | 恒信东方文化股份有限公司 | A kind of panoramic video processing method, apparatus and system |
-
2018
- 2018-08-24 CN CN201810975482.0A patent/CN110858896B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103249349A (en) * | 2010-12-02 | 2013-08-14 | 奥林巴斯株式会社 | Endoscopic image processing apparatus and program |
CN106959964A (en) * | 2016-01-12 | 2017-07-18 | 阿里巴巴集团控股有限公司 | Interface background display methods |
WO2017168229A1 (en) * | 2016-04-01 | 2017-10-05 | Linear Albra Technologies Limited | Systems and methods for head-mounted display adapted to human visual mechanism |
CN106446894A (en) * | 2016-09-27 | 2017-02-22 | 广东技术师范学院 | Method for recognizing position of spherical object based on contour |
CN108848389A (en) * | 2018-07-27 | 2018-11-20 | 恒信东方文化股份有限公司 | A kind of panoramic video processing method, apparatus and system |
Also Published As
Publication number | Publication date |
---|---|
CN110858896B (en) | 2021-06-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10462446B2 (en) | Processing a depth map for an image | |
US8983175B2 (en) | Video processing method and device for depth extraction | |
US10977809B2 (en) | Detecting motion dragging artifacts for dynamic adjustment of frame rate conversion settings | |
CN107211193B (en) | Intelligent adaptive video streaming method and system driven by perception experience quality estimation | |
CN101682794B (en) | Method, apparatus and system for processing depth-related information | |
JP4153202B2 (en) | Video encoding device | |
EP1274043A2 (en) | Image signal coding method and apparatus | |
CN103402098B (en) | A kind of video frame interpolation method based on image interpolation | |
US20110080464A1 (en) | Method and a device for filling occluded areas of a depth or disparity map estimated from at least two images | |
US20120162451A1 (en) | Digital image stabilization | |
KR100524077B1 (en) | Apparatus and method of temporal smoothing for intermediate image generation | |
CN105611116B (en) | A kind of global motion vector method of estimation and monitor video digital image stabilization method and device | |
KR102182697B1 (en) | Apparatus and method for processing image | |
KR20130094626A (en) | Method and apparatus for stereo matching | |
CN105721767B (en) | The method for handling video flowing | |
JP2011142663A (en) | Method and apparatus for efficient video processing | |
CN110620924A (en) | Method and device for processing coded data, computer equipment and storage medium | |
US20040013196A1 (en) | Quantization control system for video coding | |
JP2004023288A (en) | Preprocessing system for moving image encoding | |
CN110858896B (en) | VR image processing method | |
JP3674186B2 (en) | Image information conversion apparatus and method | |
Ling et al. | Quality assessment for synthesized view based on variable-length context tree | |
KR20120019662A (en) | Pre-processing method of a moving picture to be transmitted and system transmitting a moving picture using thereof | |
JP6010625B2 (en) | Distortion / quality measurement | |
Zafarifar et al. | Blue sky detection for picture quality enhancement |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
CB02 | Change of applicant information |
Address after: 100007 204, 2nd floor, building 3, No.2, zanjingguan Hutong, Dongcheng District, Beijing Applicant after: Oriental Dream Virtual Reality Technology Co., Ltd Address before: 100097 Beijing city Haidian District landianchang Road No. 25 11-20 Applicant before: BEIJING HENGXIN RAINBOW INFORMATION TECHNOLOGY Co.,Ltd. |
|
CB02 | Change of applicant information | ||
GR01 | Patent grant | ||
GR01 | Patent grant |