CN110858896A - VR image processing method - Google Patents

VR image processing method Download PDF

Info

Publication number
CN110858896A
CN110858896A CN201810975482.0A CN201810975482A CN110858896A CN 110858896 A CN110858896 A CN 110858896A CN 201810975482 A CN201810975482 A CN 201810975482A CN 110858896 A CN110858896 A CN 110858896A
Authority
CN
China
Prior art keywords
image
focus area
processing method
image processing
profile
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810975482.0A
Other languages
Chinese (zh)
Other versions
CN110858896B (en
Inventor
孟宪民
李小波
赵德贤
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BEIJING HENGXIN CAIHONG INFORMATION TECHNOLOGY Co Ltd
Original Assignee
BEIJING HENGXIN CAIHONG INFORMATION TECHNOLOGY Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BEIJING HENGXIN CAIHONG INFORMATION TECHNOLOGY Co Ltd filed Critical BEIJING HENGXIN CAIHONG INFORMATION TECHNOLOGY Co Ltd
Priority to CN201810975482.0A priority Critical patent/CN110858896B/en
Publication of CN110858896A publication Critical patent/CN110858896A/en
Application granted granted Critical
Publication of CN110858896B publication Critical patent/CN110858896B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/277Analysis of motion involving stochastic approaches, e.g. using Kalman filters

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The application discloses a VR image processing method, which comprises the following steps: determining an image needing to be processed; selecting at least one focus area; determining a focus area profile of the focus area; performing fusion processing on the focus area and the background image according to the contour of the focus area; and transmitting the fused image to a VR head display. The application has reached the high definition part of guaranteeing to transmit the image and partly links up naturally with non-high definition, the technological effect of effectual promotion user experience's comfort level.

Description

VR image processing method
Technical Field
The application relates to the technical field of VR image transmission, in particular to a VR image processing method.
Background
The requirement for resolution in image transmission is higher and higher nowadays, especially in the VR image transmission field, because VR needs binocular output, there is a demand for double-size image output in transmission. According to the feasible scheme at present, if the transmitting end uses X264 coding compression and then uses h264 decoding at the receiving end, the compressed video stream data can be transmitted in real time, and the network transmission data volume can be effectively reduced. However, in the field of network transmission, there are situations that high-definition and ultra-high-definition videos are transmitted to cause delay aggravation and the possibility that the videos cannot be observed instantly, particularly, the transmission pressure of the 4K or 8K videos is higher, the higher the compression ratio of the corresponding video coding field is, the lower the definition is, the possibility that the videos can be transmitted instantly and watched by a client cannot be met, and the requirement of VR binocular output panoramic high-definition images cannot be met.
Disclosure of Invention
The application aims to provide a VR image processing method, which ensures that a high-definition part and a non-high-definition part of a transmission image are naturally linked, and effectively improves the comfort level of user experience.
In order to achieve the above object, the present application provides a VR image processing method, including the steps of: determining an image needing to be processed; selecting at least one focus area; determining a focus area profile of the focus area; performing fusion processing on the focus area and the background image according to the contour of the focus area; and transmitting the fused image to a VR head display.
Preferably, the focal region profile is determined using the Camshift algorithm.
Preferably, the method for determining the profile of the focal region specifically comprises: determining a first edge profile; determining a second edge profile; and storing the first edge outline and the second edge outline into a focus object file library.
Preferably, the second edge profile is determined from the first edge profile inward contraction calculation.
Preferably, the inward shrinkage calculation formula of each pixel on the first edge profile is as follows: the shrunk pixel value is (first contour pixel value-center point) × shrinkage factor.
Preferably, the shrinkage factor is 0.6 to 0.9.
Preferably, the shrinkage factor is 0.8.
Preferably, the transmission method of the fused image is as follows: transmitting the focus area in high definition and synchronizing the background image in non-high definition.
The beneficial effect that this application realized is as follows:
(1) the high-definition part and the non-high-definition part of the transmission image are naturally connected, and the comfort level of user experience is effectively improved.
(2) The technical effects of reducing the data volume of network transmission, avoiding transmission delay and meeting the requirement that a client watches a high-definition video can be achieved.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings needed to be used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments described in the present application, and other drawings can be obtained by those skilled in the art according to the drawings.
FIG. 1 is a flow diagram of one embodiment of a VR image processing method;
FIG. 2 is a flow chart of one embodiment of a method for determining a focal region profile.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
As shown in fig. 1, the present invention provides a VR image processing method, including the following steps:
s110: and playing the video and determining the images needing to be processed in the video.
Specifically, the staff plays the video through the video playing tool, in the playing process, the staff pauses the video at a position where a focus needs to be set, and the frame of image displayed on the video playing tool after pausing is the image needing to be processed.
S120: at least one focal region is selected from the image to be processed.
Specifically, the staff performs framing on at least one region in the image by using the mouse, the framed region is a focus region, and after the focus region is determined, S130 is performed.
S130: a focal region profile of the focal region is determined.
Further: as an embodiment, the focus area profile is determined using the CamShift algorithm (continuous adaptive mean-Shift algorithm).
Specifically, a Camshift algorithm (continuous Adaptive Mean-Shift algorithm) is an improved algorithm for the Meansshift algorithm, and can adjust the size of a Search Window in real time along with the change of the size of a tracking target in the tracking process, make all frames of a video image into the Meansshift algorithm, use the result of the previous frame (i.e. the centroid and size of the Search Window) as the initial value of the Search Window of the Meansshift algorithm of the next frame to find the optimal iteration result, use the centroid in the optimal iteration result in the zero-order moment M00 and the first-order moment M10 of the Window, and calculate the size of the zero-order moment in M01, where the zero-order moment is the integral of all pixels in the Search Window, and the integral of all pixels in the Search Window is equal to the size of the Search Window, and the size of the Search Window is the focus area.
Further, the operations of Camshift and MeanShift are performed on the back projection image, a histogram of the H component of the target region (i.e., the focal region) is generated through back projection calculation, and the pixel value of each pixel point on the target region is replaced by the value of the histogram bin corresponding to the bin where the current pixel value is located. After reverse projection, most pixel values of a search window of a target area are normalized to be a maximum value of 255, a threshold value is determined according to a test result in use, if the calculated zero-order moment is larger than the threshold value, the target is paved on the whole search window, whether the target area exists in an area outside the search window needs to be judged, and if the target area exists, the size of the search window needs to be increased; if the zeroth order moment is less than the threshold, the search window needs to be reduced in size. Therefore, when the size of the target changes, the size of the search window can be adjusted in a self-adaptive mode through the Camshift algorithm, the target area is tracked, and the outline of the focus area is obtained.
Further, as shown in fig. 2, as another embodiment, the method for determining the focal region profile includes the following steps:
s210: a first edge profile is determined.
Specifically, the worker uses a mouse to draw a first edge contour on the image to be processed, and the area surrounded by the first edge contour is larger than the focus area to be framed.
S220: a second edge profile is determined.
Specifically, a second edge profile is computed from the first edge profile inward scaling.
Specifically, all paths of all pixel points on the track are calculated according to the moving track of the mouse when the staff draws the first edge profile. And calculating the average value of all paths to obtain the central point. And (3) inwards shrinking all the pixel points to calculate to obtain a second edge profile, wherein the calculation formula is as follows:
pixel value after shrinkage (first contour pixel value-center point) x shrinkage factor
The value range of the shrinkage coefficient is 0.6-0.9, the value of the shrinkage coefficient is a value obtained by a user according to real-time test, and preferably, the value of the coefficient is 0.8. The connecting line of the shrunk pixel values of all the pixels on the image to be processed is the second edge contour.
And S230, storing the first edge contour and the second edge contour into a focus object file library.
Specifically, the data such as the image size, the focus area, the first edge profile, the second edge profile, and the like of each frame of image are transmitted to the focus object file library for storage.
S140: and carrying out fusion processing on the focus area and the background image according to the contour of the focus area.
When the video is played, the computer reads the data of the focus object, the first edge contour, the second edge contour, the image size and the like corresponding to each frame of image in the played video from the focus object file library. And synchronously transmitting the definition data, the first edge profile, the second edge profile and the video data to the VR client. After receiving the data, the VR client renders a non-high-definition image as a background image by using the original image, then renders a layer of high-definition image of a focus area in the corresponding focus area on the background image, and gradually fuses to the second edge outline according to the first edge outline to form a fused area, so that the processed image is ensured to have comfortable visual sense when displayed. After the fusion process of the focus area and the background image is completed, S150 is performed.
S150: and transmitting the fused image to a VR head display.
Further, the transmission method of the fused image is as follows: transmitting the focus area in high definition and synchronizing the background image in non-high definition.
The beneficial effect that this application realized is as follows:
(1) the high-definition part and the non-high-definition part of the transmission image are naturally connected, and the comfort level of user experience is effectively improved.
(2) The technical effects of reducing the data volume of network transmission, avoiding transmission delay and meeting the requirement that a client watches a high-definition video can be achieved.
While the preferred embodiments of the present application have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all alterations and modifications as fall within the scope of the application. It will be apparent to those skilled in the art that various changes and modifications may be made in the present application without departing from the spirit and scope of the application. Thus, if such modifications and variations of the present application fall within the scope of the claims of the present application and their equivalents, the present application is intended to include such modifications and variations as well.

Claims (8)

1. A VR image processing method is characterized by comprising the following steps:
determining an image needing to be processed;
selecting at least one focus area;
determining a focus area profile of the focus area;
performing fusion processing on the focus area and the background image according to the contour of the focus area;
and transmitting the fused image to a VR head display.
2. The VR image processing method of claim 1, wherein the focus area profile is determined using a CamShift algorithm.
3. The VR image processing method of claim 1, wherein the method of determining the focus area profile is specifically:
determining a first edge profile;
determining a second edge profile;
and storing the first edge outline and the second edge outline into a focus object file library.
4. The VR image processing method of claim 3, wherein the second edge profile is determined from a first edge profile inward scaling calculation.
5. The VR image processing method of claim 4, wherein the inward shrinkage for each pixel on the first edge profile is calculated by: the shrunk pixel value is (first contour pixel value-center point) × shrinkage factor.
6. The VR image processing method of claim 5 wherein the shrinkage factor is 0.6 to 0.9.
7. The VR image processing method of claim 6, wherein the contraction coefficient is 0.8.
8. The VR image processing method of claim 1, wherein the fused image is transmitted by: transmitting the focus area in high definition and synchronizing the background image in non-high definition.
CN201810975482.0A 2018-08-24 2018-08-24 VR image processing method Active CN110858896B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810975482.0A CN110858896B (en) 2018-08-24 2018-08-24 VR image processing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810975482.0A CN110858896B (en) 2018-08-24 2018-08-24 VR image processing method

Publications (2)

Publication Number Publication Date
CN110858896A true CN110858896A (en) 2020-03-03
CN110858896B CN110858896B (en) 2021-06-08

Family

ID=69636438

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810975482.0A Active CN110858896B (en) 2018-08-24 2018-08-24 VR image processing method

Country Status (1)

Country Link
CN (1) CN110858896B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103249349A (en) * 2010-12-02 2013-08-14 奥林巴斯株式会社 Endoscopic image processing apparatus and program
CN106446894A (en) * 2016-09-27 2017-02-22 广东技术师范学院 Method for recognizing position of spherical object based on contour
CN106959964A (en) * 2016-01-12 2017-07-18 阿里巴巴集团控股有限公司 Interface background display methods
WO2017168229A1 (en) * 2016-04-01 2017-10-05 Linear Albra Technologies Limited Systems and methods for head-mounted display adapted to human visual mechanism
CN108848389A (en) * 2018-07-27 2018-11-20 恒信东方文化股份有限公司 A kind of panoramic video processing method, apparatus and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103249349A (en) * 2010-12-02 2013-08-14 奥林巴斯株式会社 Endoscopic image processing apparatus and program
CN106959964A (en) * 2016-01-12 2017-07-18 阿里巴巴集团控股有限公司 Interface background display methods
WO2017168229A1 (en) * 2016-04-01 2017-10-05 Linear Albra Technologies Limited Systems and methods for head-mounted display adapted to human visual mechanism
CN106446894A (en) * 2016-09-27 2017-02-22 广东技术师范学院 Method for recognizing position of spherical object based on contour
CN108848389A (en) * 2018-07-27 2018-11-20 恒信东方文化股份有限公司 A kind of panoramic video processing method, apparatus and system

Also Published As

Publication number Publication date
CN110858896B (en) 2021-06-08

Similar Documents

Publication Publication Date Title
US10462446B2 (en) Processing a depth map for an image
US8983175B2 (en) Video processing method and device for depth extraction
US10977809B2 (en) Detecting motion dragging artifacts for dynamic adjustment of frame rate conversion settings
CN107211193B (en) Intelligent adaptive video streaming method and system driven by perception experience quality estimation
CN101682794B (en) Method, apparatus and system for processing depth-related information
JP4153202B2 (en) Video encoding device
EP1274043A2 (en) Image signal coding method and apparatus
CN103402098B (en) A kind of video frame interpolation method based on image interpolation
US20110080464A1 (en) Method and a device for filling occluded areas of a depth or disparity map estimated from at least two images
US20120162451A1 (en) Digital image stabilization
KR100524077B1 (en) Apparatus and method of temporal smoothing for intermediate image generation
CN105611116B (en) A kind of global motion vector method of estimation and monitor video digital image stabilization method and device
KR102182697B1 (en) Apparatus and method for processing image
KR20130094626A (en) Method and apparatus for stereo matching
CN105721767B (en) The method for handling video flowing
JP2011142663A (en) Method and apparatus for efficient video processing
CN110620924A (en) Method and device for processing coded data, computer equipment and storage medium
US20040013196A1 (en) Quantization control system for video coding
JP2004023288A (en) Preprocessing system for moving image encoding
CN110858896B (en) VR image processing method
JP3674186B2 (en) Image information conversion apparatus and method
Ling et al. Quality assessment for synthesized view based on variable-length context tree
KR20120019662A (en) Pre-processing method of a moving picture to be transmitted and system transmitting a moving picture using thereof
JP6010625B2 (en) Distortion / quality measurement
Zafarifar et al. Blue sky detection for picture quality enhancement

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Address after: 100007 204, 2nd floor, building 3, No.2, zanjingguan Hutong, Dongcheng District, Beijing

Applicant after: Oriental Dream Virtual Reality Technology Co., Ltd

Address before: 100097 Beijing city Haidian District landianchang Road No. 25 11-20

Applicant before: BEIJING HENGXIN RAINBOW INFORMATION TECHNOLOGY Co.,Ltd.

CB02 Change of applicant information
GR01 Patent grant
GR01 Patent grant