CN106127680A - A kind of 720 degree of panoramic video fast browsing methods - Google Patents

A kind of 720 degree of panoramic video fast browsing methods Download PDF

Info

Publication number
CN106127680A
CN106127680A CN201610496238.7A CN201610496238A CN106127680A CN 106127680 A CN106127680 A CN 106127680A CN 201610496238 A CN201610496238 A CN 201610496238A CN 106127680 A CN106127680 A CN 106127680A
Authority
CN
China
Prior art keywords
video
frame
view
camera lens
degree
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610496238.7A
Other languages
Chinese (zh)
Other versions
CN106127680B (en
Inventor
罗文峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Youxiang Computing Technology Co Ltd
Original Assignee
Shenzhen Youxiang Computing Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Youxiang Computing Technology Co Ltd filed Critical Shenzhen Youxiang Computing Technology Co Ltd
Priority to CN201610496238.7A priority Critical patent/CN106127680B/en
Publication of CN106127680A publication Critical patent/CN106127680A/en
Application granted granted Critical
Publication of CN106127680B publication Critical patent/CN106127680B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/60Rotation of a whole image or part thereof
    • G06T3/608Skewing or deskewing, e.g. by two-pass or three-pass rotation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/74Browsing; Visualisation therefor
    • G06F16/743Browsing; Visualisation therefor a collection of video files or sequences
    • G06T3/04
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4038Scaling the whole image or part thereof for image mosaicing, i.e. plane images composed of plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/49Segmenting video sequences, i.e. computational techniques such as parsing or cutting the sequence, low-level clustering or determining units such as shots or scenes

Abstract

The invention discloses a kind of 720 degree of panoramic video fast browsing methods, first pass through back projection method 720 degree of full-view video images are reconstructed, obtain view corresponding on each direction of visual lines of sphere view space, then camera lens length is judged by the absolute brightness frame difference of adjacent image frame in calculating video sequence, then key frame is extracted, it is achieved panoramic video fast browsing.The present invention can be quickly generated virtual scene see-through view on different sight direction, and analogue camera rotates and zoom motion effectively on all directions view, improve the surfing of virtual scene, it is possible to well meet this specific application area of virtual reality system.

Description

A kind of 720 degree of panoramic video fast browsing methods
Technical field
The invention belongs to technical field of image processing, relate to video panorama image and process, refer in particular to a kind of 720 degree of panoramic videos Fast browsing method.
Background technology
Along with the development of information technology, the scene information obtaining wide viewing angle scope is required more and more higher by people, and tradition is taken the photograph Image method can only obtain the picture frame of limited angular range, and image mosaic technology produces for solving this difficult problem and sends out rapidly Exhibition.It is the complete image by two width of phase mutual overlay information or several pictures are spliced into a ultra-wide visual angle, to reach To reducing picture redundancy and obtaining the purpose of more wide viewing angle information, wherein the generation of panoramic picture is also image mosaic technology One typical case's application.
720 degree of panoramic videos are a kind of sequence of video images based on Sphere Measurement Model, can realize horizontal 360-degree with vertical Looking around of 360 degree of visual angle directions browses.When browsing, need according to current gaze direction and FOV (Field of View) spherical video figure As carrying out back project, to obtain the plane perspective image meeting human eye vision custom.In this way, can be the most real Existing analogue camera rotary motion and zoom motion, change the ken.
Need to take a substantial amount of time and energy for having access to and play back massive video data, traditional dilatory browsing method Easily ignoring the unexpected abnormality event in the short time, the search of long-time video data is unfavorable for the extraction of effective information.Therefore Needing panoramic video has been carried out further process, it is achieved the fast browsing to panoramic video, its core work is to original The extraction of critical sequences in the segmentation of video and video.
Currently, Video segmentation and extraction method of key frame are broadly divided into four classes:
One is simple generating algorithm, by video sequence being carried out the method waiting time uniform sampling to extract key frame, But this method is due to amount of video information change difference in the short time, easily occur that key-frame extraction too much or represents asking of deficiency Topic;
Two is the generation method of view-based access control model information, according to the visual informations such as the color in video, shape, texture, application Various video processing technique carry out the operations such as scene clustering, shot detection, key-frame extraction, ultimately generate video skimming, this The method of view-based access control model feature is significantly improved in simple generating algorithm, but have ignored the audio frequency in original video, word The information such as curtain;
Three is the generation method merging multiple features, as used the appearance of VIP in face recognition technology detection news, Utilize the wonderful etc. in audio signal processing technique detection sports video, in conjunction with feature and other image procossing skills of video self The art multiple Feature Fusion to video, algorithm process process is more complicated;
Four is the generation method semantic based on video syntax, seeks between camera lens and camera lens, knot between scene and scene Structure rule, forms video frequency abstract based on this.
In sum, for different video type and purposes, different to video fast browsing processing method, current panorama It is empty that video technique is widely used in the networks such as tourist attractions, house property house, car show, leisure club, urban architecture planning In intending showing, these video scenes mainly reach preferably to publicize representing panorama to a kind of experience on the spot in person of people and perfection Purpose.
Summary of the invention
The present invention proposes a kind of 720 degree of panoramic video fast browsing methods, and it achieves level 360 by back projection method Observe video under degree and vertical 360 degree of omnibearing angels, and according to the difference of taking lens length under video difference scene, extract Key frame, forms video summary, reaches the purpose of fast browsing.
A kind of 720 degree of panoramic video fast browsing methods, comprise the following steps:
720 degree of full-view video images are reconstructed by S1 first with Inverse Projection, obtain sphere view space each View sequence corresponding on direction of visual lines.
S2, judges camera lens length by calculating the absolute brightness frame difference of adjacent image frame in video sequence, then extracts Key frame, it is achieved panoramic video fast browsing.
Wherein, S1 comprises the following steps:
S1.1, completes the splicing of 720 degree of panoramic pictures, and sets up centered by the centre of sphere based on sphere view space model Two coordinate systems, represent world coordinate system XYZ and camera coordinate system xyz respectively;Wherein camera coordinate system xyz is by the world Coordinate system XYZ X-axis rotation alpha angle in world coordinate system, obtains further around the Y-axis rotation β angle in world coordinate system.
The method of the splicing completing 720 degree of panoramic pictures based on sphere view space model in S1.1 is: according to photographing unit Straight line vertical with image transverse axis in the image that straight line parallel with y-axis in coordinate system xyz generates according to sphere parameters transform It is still this character of vertical line, multiple real scene images of fish eye lens shooting is carried out rotation transformation correction, obtains every realistic picture As the azimuth information at view space of upper pixel, utilize azimuth information that multiple images are spliced, eliminate real scene image Between duplicate message that may be present, finally project on a sphere, store with the form of Spherical Panorama Image.
S1.2, by unified for the basic measuring unit of pixel under Two coordinate system in S1.1, then to calculate using pixel as base Each pixel is i.e. estimated the viewpoint pixel focal length to view plane under camera coordinate system by the pixel focal length of this unit f。
In S1.2, if image S is the Spherical Panorama Image that a splicing is complete, Q is any one on Spherical Panorama Image S Pixel, image coordinate is;J is the view needing to generate, and some P is that the some Q on Spherical Panorama Image is corresponding on view J Point, its image coordinate is;F represents pixel focal length, and f uses camera lens to estimate according to shooting real scene image.
The method of estimation of the pixel focal distance f of wide-angle lens or standard lens is: sets camera level and rotates a circle captured Real scene image n open, then the horizontal view angle of photographing unit is 360/n, and real scene image width is W, permissible according to trigonometric function relation The pixel focal length estimation formulas drawing common lens is:
F=W/ (2tan (180/n)).
The method of estimation of fish-eye pixel focal distance f is: the width of image after the dark border of fish eye images being removed Degree is designated as W, and the most fish-eye pixel focal length estimation formulas is: f=W/ φ, wherein φ is fish-eye horizontal view.
S1.3, sets up the coordinate of X-Y scheme picture point and turning of the three-dimensional parameter coordinate points corresponding to sphere by pixel focal distance f Change relational expression, further according to world coordinate system XYZ X-axis rotation alpha angle in world coordinate system, further around the Y in world coordinate system During axle rotation β angle, along with the rotation of coordinate axes, pixel expression on each coordinate components also occurs to become accordingly (after coordinate system rotation, the correspondence position of each pixel needs again to represent under new coordinate system, and coordinate components is each leisure in change Respective components on tri-coordinate axess of x, y, z), this change can be closed with trigonometric function and be tied up to represent on respective coordinate components Come, thus obtain the transformation matrix H of corresponding point under Two coordinate system.
S1.4, is set up counter-transformation function by transformation matrix H, finds out on panoramic picture arbitrfary point to each view of spherical space The corresponding relation formula of point, calculates each point coordinates, obtains view corresponding on each direction of visual lines of view space.
In S1.3, set up coordinate and the three-dimensional parameter coordinate points corresponding to sphere of X-Y scheme picture point by pixel focal distance f Transforming relationship formula, as follows:
u = f c o s ( π 2 - y f ) c o s ( x f ) v = f s i n ( π 2 - y f ) w = f c o s ( π 2 - y f ) s i n ( x f ) - - - ( 1 )
Calculating the transformation matrix H of corresponding point under Two coordinate system, expression formula is as follows:
H = 1 0 0 0 c o s α s i n α 0 - s i n α c o s α c o s β 0 - sin β 0 1 0 s i n β 0 c o s β - - - ( 2 )
In S1.4, from the formula (1) in S1.3 and the point under (2), coordinate system XYZ, correspondence under coordinate system xyz Coordinate is.
In known video, the width of captured real scene image is W, and height is H, sets up any point Q on Spherical Panorama Image The functional relation of the some P corresponding on view J with a Q, and utilize formula (3) to calculate the coordinate of each corresponding point, obtain viewpoint View corresponding on each direction of visual lines in space.
x = W 2 - f cos ( π 2 - y ′ f ) cos ( x ′ f + β ) cos ( π 2 - y ′ f ) cos α sin ( x ′ f + β ) - sin ( π 2 - y ′ f ) sin α y = H 2 - f sin ( π 2 - y ′ f ) cos α + f cos ( π 2 - y ′ f ) sin α sin ( x ′ f + β ) cos ( π 2 - y ′ f ) cos α sin ( x ′ f + β ) - sin ( π 2 - y ′ f ) sin α - - - ( 3 )
The S2 of the present invention comprises the following steps:
S2.1, carries out structuring process to panoramic video sequences, and panoramic video is corresponding on each direction of visual lines Being all the video sequence of one group of view composition, video sequence step S1 obtained is according to being projected in regarding on different directions visual angle Figure frame sequence is classified, and obtains the video sequence group that can individually browse on multiple visual angle;
S2.2, splits the video sequence group of each view directions respectively, calculates adjacent image frame in video sequence Absolute brightness frame is poor, it is judged that the switching node of video lens, is multiple camera lens fragments by video signal process;
S2.3, calculates the quantity of motion summation of each camera lens fragment, sets quantity of motion and weighs thresholding, sentences according to shot durations Disconnected current lens is full length shot or short camera lens;
S2.4, extracts key frame respectively to long and short camera lens, extracts a key frame at random for short camera lens, and full length shot is then Multiple image is extracted as key frame by method at equal intervals;
S2.5, recombinates the keyframe sequence of extraction, reverts to generate on different visual angles direction video summary, observes Person by reaching the purpose of video fast browsing to the operation of video summary.
In S2.2, selecting absolute brightness frame difference AIFD as the characteristic quantity of measurement change degree of video content, it defines Formula is as follows:
A I F D ( t ) = Σ x = 1 W Σ y = 1 H a b s ( f ( x , y , t + 1 ) - f ( x , y , t ) ) - - - ( 4 )
In above formula and represent respectively under brightness value and the t of t picture frame pixel at coordinate in video sequence One frame is the brightness value of pixel at coordinate, W and H represents width and the height of frame of video respectively;If regarding in a certain view directions The complete picture frame number play of frequency is N, then the luminance frame difference average of video is:
A I F D ( t ) ‾ = 1 N A I F D ( t ) - - - ( 5 )
By calculating luminance frame difference average as determinating reference, two different coefficient a and b are set (when the value of a and b sets Time the least, easy flase drop;When setting the biggest, easy missing inspection.In experiment, the value of a is 1.2, and b value is 2.3, empirical value) to bright Whether degree frame difference average is weighted obtaining high-low threshold value thresh_low and thresh_high, change as camera lens and with which The decision condition that the mode of kind is changed, wherein
t h r e s h _ l o w = a · A I F D ( t ) ‾ t h r e s h _ h i g h = b · A I F D ( t ) ‾ - - - ( 6 )
In S2.2, the method splitting video sequence group is as follows:
First the video requency frame data of input is initialized, calculate the AIFD eigenvalue of adjacent two frames of t, compare also Judge the eigenvalue of present frame and the size of decision threshold, detect with this and whether there is camera lens between present frame and next frame and turn Change, it is determined that if method for present frame eigenvalue less than thresh_low; there is not Shot change, if more than thresh_low and Then think that less than thresh_high present frame there occurs that gradual shot is changed, if more than thresh_high, thinking that present frame is sent out Give birth to abrupt shot conversion, gradual shot conversion and abrupt shot conversion and all represented that present frame is recorded as there occurs turning of camera lens Changing, present frame is record camera lens switching node.
In S2.3, by the calculating to the quantity of motion summation calculating camera lens, and weigh door with the quantity of motion pre-set Limit compares, it is judged that camera lens belongs to full length shot or short camera lens, wherein represents the relative motion between adjacent two frame of video of t Amount, represents the persistent period of camera lens, is judged to full length shot when camera motion amount summation weighs thresholding more than quantity of motion, is otherwise Short camera lens.
72 degree of panoramic video fast browsing methods of one that the present invention proposes can be quickly generated virtual scene in difference See-through view on direction of visual lines, and on all directions view, analogue camera rotates and zoom motion effectively, improves virtual The surfing of scene, it is possible to well meet this specific application area of virtual reality system.
Accompanying drawing explanation
Fig. 1 panoramic picture back projection coordinate system schematic diagram
Segmentation restructuring block diagram on Fig. 2 panoramic video different views direction
Fig. 3 key-frame extraction block diagram
Fig. 4 is the trigonometric function relation schematic diagram between W, f, θ three
Detailed description of the invention
Below with reference to accompanying drawing, the present invention is described in further details.
In order to enable to carry out 720 degree of panoramic videos effectively browsing at a high speed on all directions view, the first of the present invention Step is to utilize Inverse Projection to be reconstructed panoramic video, obtains corresponding regarding on each direction of visual lines of sphere view space Graphic sequence, and the rotation of analogue camera and zoom motion realize browsing video in different visual angles, specifically comprises the following steps that
S1.1, completes the splicing of 720 degree of panoramic pictures, and sets up centered by the centre of sphere based on sphere view space model Two coordinate systems, represent world coordinate system XYZ and camera coordinate system xyz respectively.
Wherein, the method for the splicing completing 720 degree of panoramic pictures based on sphere view space model is: sit according to photographing unit Straight line vertical with image transverse axis in the image that straight line parallel with y-axis in mark system xyz generates according to sphere parameters transform is still For this character of vertical line, multiple real scene images of fish eye lens shooting are carried out rotation transformation correction, obtains every real scene image The azimuth information at view space of upper pixel, utilizes azimuth information to splice multiple images, eliminate real scene image it Between duplicate message that may be present, finally project on a sphere, store with the form of Spherical Panorama Image.
Wherein camera coordinate system xyz is by the world coordinate system XYZ X-axis anglec of rotation in world coordinate system, further around generation The Y-axis anglec of rotation in boundary's coordinate system and obtain.
Setting image S and splice complete Spherical Panorama Image as one, Q is any one pixel on Spherical Panorama Image S Point, image coordinate is;J is the view (i.e. J is a view on some direction of visual lines to be obtained) needing to generate, As it is shown in figure 1, some P is the point corresponding on view J for some Q on sphere, its image coordinate is;F represents pixel focal length, according to Shooting real scene image uses common lens (general wide-angle lens and standard lens) or fish eyes camera lens to estimate.
S1.2, in order to by unified for the basic unit of pixel under two coordinate systems, carry out the pixel focal distance f of camera lens Estimate.
The pixel focal length method of estimation of common lens (general wide-angle lens and standard lens) is: set camera level rotation The real scene image n circling captured opens, then the horizontal view angle of photographing unit is 360/n, and real scene image width is W, according to triangle Functional relationship can show that the pixel focal length estimation formulas of common lens is: f=W/ (2tan (180/n)).Wherein trigonometric function Relation refers to the relations such as the sine of the interior existence of right angled triangle, cosine, tangent.
With reference to Fig. 4, take a sectional drawing of panorama sketch, by trigonometric function relation, each amount can be indicated.Assume Figure represents with θ the horizontal view angle of photographing unit, then has θ=360/n, as seen from the figure the trigonometric function relation between W, f, θ three Can obtain:
Transformation resolution goes out, f=W/ (2tan (θ/2)), i.e. f=W/ (2tan (180/n))
Fish-eye pixel focal length is estimated to be derived by fish-eye equidistant imaging model, particularly as follows: will After the dark border of fish eye images is removed, the width of image is designated as W, and the most fish-eye pixel focal length estimation formulas is: f= W/ φ, wherein φ is fish-eye horizontal view, can be made a look up by fish eye lens description.
S1.3, generates the inverse operation of process, is under three-dimensional parameter coordinate by two dimensional image coordinate transformation according to panoramic picture Carry out computing, point on image coordinate point Q correspondence sphere, meet following transforming relationship formula:
u = f c o s ( π 2 - y f ) c o s ( x f ) v = f s i n ( π 2 - y f ) w = f cos ( π 2 - y f ) s i n ( x f ) - - - ( 1 )
Calculating the transformation matrix H of corresponding point under Two coordinate system, expression formula is as follows:
H = 1 0 0 0 c o s α s i n α 0 - s i n α c o s α c o s β 0 - sin β 0 1 0 s i n β 0 c o s β - - - ( 2 )
After trying to achieve transformation matrix, from both the above formula, the point under coordinate system XYZ, correspondence under coordinate system xyz Coordinate be.
S1.4, it is known that in video, the width of captured real scene image frame is W, and height is H, sets up Spherical Panorama Image and takes up an official post The functional relation of some P corresponding on view J with it for 1 Q that anticipates, and calculate the coordinate of each corresponding point.
x = W 2 - f cos ( π 2 - y ′ f ) cos ( x ′ f + β ) cos ( π 2 - y ′ f ) cos α sin ( x ′ f + β ) - sin ( π 2 - y ′ f ) sin α y = H 2 - f sin ( π 2 - y ′ f ) cos α + f cos ( π 2 - y ′ f ) sin α sin ( x ′ f + β ) cos ( π 2 - y ′ f ) cos α sin ( x ′ f + β ) - sin ( π 2 - y ′ f ) sin α - - - ( 3 )
So far the first step of the present invention is completed, by the picture frame of 720 degree of panoramic videos through back project, depending on 720 degree of panoramic videos i.e. can be carried out seeing on any sight direction by view corresponding on space of points any sight direction See.Owing to directly 720 degree panoramic videos to be carried out comprehensive browsing one by one, there is the problem that access data volume is excessive that browses, one Aspect easily causes the fatigue of observer, on the other hand affect the efficiency of key message extraction, based on this, present invention additionally comprises the Two steps, second step extracts key frame according to the length shooting video lens under different scenes, is built by the key frame extracted One section of video frequency abstract, observer, by operating the keyframe sequence video of restructuring, reaches quick to 720 degree of panoramic videos The purpose browsed.
Video is made up of multiple different scenes, comprises multiple camera lens, wherein have full length shot, short mirror under each scene Head, each camera lens is formed according to certain played in order by multiframe associated images, and therefore frame of video is to form video base This unit.In order to be able to realize the fast browsing of video, obtaining the key frame in video image becomes the extraction of video effective information Key.It is said that in general, different types of video council according to self theme photographed scene had emphasis and time emphasis point, camera lens Length also can distinguish according to the difference of focus, therefore by detection and the length camera lens that judges in video, for pass The extraction of key frame is more favourable.
S2.1, the first step obtains panoramic video view sequence in different directions, by these video sequences according to projection View sequence on different directions visual angle carries out classifying, and (360 degree of panorama sketch that a width launches are by the image of multiple different visual angles Being spliced, be exactly that panorama sketch reverts to multiple views in different visual angles through back projection, these views are arranged according to sequence Put, classify according to sequence number), obtain the view sequence group that can individually browse on multiple visual angle.A frame panorama sketch in panoramic video What back projection obtained is the view on multiple direction of visual lines, and panoramic video back projection obtains is regarding in multiple view directions Frequently view sequence, each direction has multiple views.
S2.2, carries out dividing processing to the video sequence group on different directions visual angle respectively.
Select absolute brightness frame difference AIFD (absolute intensity frame difference) as weighing video The characteristic quantity of content change degree, its definition is as follows:
A I F D ( t ) = Σ x = 1 W Σ y = 1 H a b s ( f ( x , y , t + 1 ) - f ( x , y , t ) ) - - - ( 4 )
In above formula and represent brightness value and the t+1 moment of t picture frame pixel at coordinate in video sequence respectively Picture frame is the brightness value of pixel at coordinate, W and H represents width and the height of frame of video respectively.If in a certain view directions The picture frame number that video is completely play is N, then the luminance frame difference average of video is:
A I F D ( t ) ‾ = 1 N A I F D ( t ) - - - ( 5 )
Owing to the luminance frame difference change of pixel is little under same camera lens, present the more uniform distribution situation of ratio, therefore Two different coefficient a and b can be set by calculating luminance frame difference average as determinating reference, when the value of a and b sets the least Time, easy flase drop;When setting the biggest, easy missing inspection.(in experiment, the value of a is 1.2, and b value is 2.3, empirical value), to brightness Whether frame difference average is weighted obtaining high-low threshold value thresh_low and thresh_high, change as camera lens and with which kind of The decision condition that mode is changed.
t h r e s h _ l o w = a · A I F D ( t ) ‾ t h r e s h _ h i g h = b · A I F D ( t ) ‾ - - - ( 6 )
The step that implements splitting video sequence is first to initialize the video requency frame data of input, meter Calculate the AIFD eigenvalue of adjacent two frames of t, relatively and judge the eigenvalue of present frame and the size of decision threshold, examine with this Survey and between present frame and next frame, whether there is camera lens conversion.If decision method is present frame eigenvalue is less than thresh_low, There is not Shot change, if more than thresh_low and less than thresh_high, thinking that present frame may have occurred gradual change mirror Head conversion, if more than thresh_high, thinks that present frame may have occurred abrupt shot conversion, and either gradual change or sudden change are all Present frame is recorded as there occurs the conversion of camera lens, i.e. record camera lens switching node.
S2.3, component motion is commonly used to characterize the situation of content change in video, by the quantity of motion calculating camera lens The calculating of summation, and (it is threshold value set in advance that quantity of motion weighs thresholding, it is generally recognized that same to weigh thresholding with the quantity of motion arranged Under one camera lens, the histogram difference of two two field pictures is less, when difference aggregate-value i.e. quantity of motion summation exceedes the quantity of motion measurement of setting During thresholding, it is determined that for full length shot) compare, it is judged that camera lens belongs to full length shot or short camera lens, wherein represents adjacent two videos of t Amount of relative motion between frame, it is simply that the difference between adjacent two frames, weighs by histogram difference rate, even if same mirror Two two field pictures under Tou are the most constant, and simply difference value is less.Represent the persistent period of camera lens, when camera motion amount summation It is judged to full length shot when weighing thresholding more than quantity of motion, is otherwise short camera lens.
S2.4, extracts a key frame for short camera lens according to randomly selecting method, and full length shot is then according to camera lens start frame Start, equally spaced choose the multiple image key frame as full length shot.
S2.5, recombinates to the keyframe sequence extracted, and reverts to generate on different visual angles direction video summary, observes Person by reaching the purpose of video fast browsing to the operation of video summary.

Claims (10)

1. 720 degree of panoramic video fast browsing methods, it is characterised in that comprise the following steps:
720 degree of full-view video images are reconstructed by S1 first with Inverse Projection, obtain each sight line of sphere view space View sequence corresponding on direction;
S2, judges camera lens length by calculating the absolute brightness frame difference of adjacent image frame in video sequence, then extracts key Frame, it is achieved panoramic video fast browsing.
2. according to 720 described in claim 1 degree panoramic video fast browsing method, it is characterised in that the method for S1 is as follows:
S1.1, completes the splicing of 720 degree of panoramic pictures, and sets up two centered by the centre of sphere based on sphere view space model Coordinate system, represents world coordinate system XYZ and camera coordinate system xyz respectively;Wherein camera coordinate system xyz is by world coordinates It is XYZ X-axis rotation alpha angle in world coordinate system, obtains further around the Y-axis rotation β angle in world coordinate system;
S1.2, by unified for the basic measuring unit of pixel under Two coordinate system in S1.1, then to calculate using pixel as fundamental quantity Each pixel is i.e. estimated the viewpoint pixel focal distance f to view plane under camera coordinate system by the pixel focal length of degree unit;
S1.3, sets up the coordinate of X-Y scheme picture point and the conversion pass of the three-dimensional parameter coordinate points corresponding to sphere by pixel focal distance f It is formula, further according to world coordinate system XYZ X-axis rotation alpha angle in world coordinate system, revolves further around the Y-axis in world coordinate system During turning β angle, along with the rotation of coordinate axes, pixel expression on each coordinate components also occurs to change accordingly, This change can be closed with trigonometric function and be tied up to show on respective coordinate components, thus obtains corresponding point under Two coordinate system Transformation matrix H;
S1.4, is set up counter-transformation function by transformation matrix H, find out arbitrfary point on panoramic picture to spherical space each view point Corresponding relation formula, calculates each point coordinates, obtains view corresponding on each direction of visual lines of view space.
3. according to 720 described in claim 1 degree panoramic video fast browsing method, it is characterised in that in S1.2, if image S is a complete Spherical Panorama Image of splicing, and Q is any one pixel on Spherical Panorama Image S, image coordinate be (x ', y′);J is the view needing to generate, and some P is the point corresponding on view J for some Q on Spherical Panorama Image, its image coordinate For (x, y);F represents pixel focal length, and f uses camera lens to estimate according to shooting real scene image;
The method of estimation of the pixel focal distance f of wide-angle lens or standard lens is: sets camera level and rotates a circle captured reality Scape image n opens, then the horizontal view angle of photographing unit is 360/n, and real scene image width is W, can draw according to trigonometric function relation The pixel focal length estimation formulas of common lens is: f=W/ (2tan (180/n));
The method of estimation of fish-eye pixel focal distance f is: the width note of image after the dark border of fish eye images being removed For W, the most fish-eye pixel focal length estimation formulas is: f=W/ φ, wherein φ is fish-eye horizontal view.
4. according to 720 described in claim 3 degree panoramic video fast browsing method, it is characterised in that in S1.3, use pixel Focal distance f sets up the transforming relationship formula of the coordinate of X-Y scheme picture point and the three-dimensional parameter coordinate points corresponding to sphere, as follows:
u = f c o s ( π 2 - y f ) c o s ( x f ) v = f s i n ( π 2 - y f ) w = f c o s ( π 2 - y f ) s i n ( x f ) - - - ( 1 )
Calculating the transformation matrix H of corresponding point under Two coordinate system, expression formula is as follows:
H = 1 0 0 0 c o s α s i n α 0 - s i n α c o s α c o s β 0 - sin β 0 1 0 s i n β 0 c o s β - - - ( 2 ) .
720 degree of panoramic video fast browsing methods the most according to claim 4, it is characterised in that in S1.4, by S1.3 Formula (1) and (2) understand, (u, v, w), coordinate corresponding under coordinate system xyz is (u ', v ', w ') for some Q ' under coordinate system XYZ =H (u, v, w);
In known video, the width of captured real scene image is W, and height is H, set up any point Q on Spherical Panorama Image (x ', Y ') the some P corresponding on view J with some Q (x, functional relation y), and utilize formula (3) to calculate the coordinate of each corresponding point, Obtain view corresponding on each direction of visual lines of view space;
x = W 2 - f cos ( π 2 - y ′ f ) cos ( x ′ f + β ) cos ( π 2 - y ′ f ) cos α sin ( x ′ f + β ) - sin ( π 2 - y ′ f ) sin α y = H 2 - f sin ( π 2 - y ′ f ) cos α + f cos ( π 2 - y ′ f ) sin α sin ( x ′ f + β ) cos ( π 2 - y ′ f ) cos α sin ( x ′ f + β ) - sin ( π 2 - y ′ f ) sin α - - - ( 3 ) .
720 degree of panoramic video fast browsing methods the most according to claim 1, it is characterised in that the method for S2 is as follows:
S2.1, carries out structuring process to panoramic video sequences, and video sequence step S1 obtained is according to being projected in not Tongfang View frames sequence on visual angle is classified, and obtains the video sequence group that can individually browse on multiple visual angle;
S2.2, splits the video sequence group of each view directions respectively, and in calculating video sequence, adjacent image frame is absolute Luminance frame is poor, it is judged that the switching node of video lens, is multiple camera lens fragments by video signal process;
S2.3, calculates the quantity of motion summation of each camera lens fragment, sets quantity of motion and weighs thresholding, judges to work as according to shot durations Front camera lens is full length shot or short camera lens;
S2.4, extracts key frame respectively to long and short camera lens, extracts a key frame at random for short camera lens, and full length shot is then pressed Interval method extracts multiple image as key frame;
S2.5, recombinates the keyframe sequence of extraction, reverts to generate on different visual angles direction video summary, and observer is led to Cross the operation to video summary and reach the purpose of video fast browsing.
720 degree of panoramic video fast browsing methods the most according to claim 6, it is characterised in that in S2.2, select definitely Luminance frame difference AIFD is as the characteristic quantity of measurement change degree of video content, and its definition is as follows:
A I F D ( t ) = Σ x = 1 W Σ y = 1 H a b s ( f ( x , y , t + 1 ) - f ( x , y , t ) ) - - - ( 4 )
In above formula, (with f (x, y, t+1), x, y t) represent that in video sequence, t picture frame is in (x, y) pixel at coordinate to f respectively Point brightness value and t next frame (x, y) brightness value of pixel at coordinate, W and H represent respectively frame of video width and Highly;If the picture frame number that in a certain view directions, video is completely play is N, then the luminance frame difference average of video is:
A I F D ( t ) ‾ = 1 N A I F D ( t ) - - - ( 5 )
By calculating luminance frame difference average as determinating reference, two different coefficient a and b are set, luminance frame difference average is carried out Weighting obtains high-low threshold value thresh_low and thresh_high, and whether that changes as camera lens and change in which way sentences Fixed condition, wherein
t h r e s h _ l o w = a · A I F D ( t ) ‾ t h r e s h _ h i g h = b · A I F D ( t ) ‾ - - - ( 6 ) .
720 degree of panoramic video fast browsing methods the most according to claim 7, it is characterised in that the value of a is 1.2, b Value is 2.3.
720 degree of panoramic video fast browsing methods the most according to claim 7, it is characterised in that in S2.2, to video sequence The method that row group carries out splitting is as follows:
First the video requency frame data of input is initialized, calculate the AIFD eigenvalue of adjacent two frames of t, relatively and judge The eigenvalue of present frame and the size of decision threshold, detect with this and whether there is camera lens conversion between present frame and next frame, If less than thresh_low, there is not Shot change, if being more than thresh_low and being less than for present frame eigenvalue in decision method Thresh_high then thinks that present frame there occurs that gradual shot is changed, if more than thresh_high, thinking that present frame there occurs Abrupt shot is changed, and gradual shot conversion and abrupt shot conversion all represent that present frame is recorded as there occurs the conversion of camera lens, when Front frame is record camera lens switching node.
720 degree of panoramic video fast browsing methods the most according to claim 9, it is characterised in that in S2.3, by right Calculate the quantity of motion summation of camera lensCalculating, and weigh thresholding with the quantity of motion that pre-sets and compare, it is judged that Camera lens belongs to full length shot or short camera lens, and wherein M_f (t) represents the amount of relative motion between adjacent two frame of video of t, S_ Time represents the persistent period of camera lens, is judged to full length shot, otherwise when camera motion amount summation weighs thresholding more than quantity of motion For short camera lens.
CN201610496238.7A 2016-06-29 2016-06-29 720-degree panoramic video fast browsing method Active CN106127680B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610496238.7A CN106127680B (en) 2016-06-29 2016-06-29 720-degree panoramic video fast browsing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610496238.7A CN106127680B (en) 2016-06-29 2016-06-29 720-degree panoramic video fast browsing method

Publications (2)

Publication Number Publication Date
CN106127680A true CN106127680A (en) 2016-11-16
CN106127680B CN106127680B (en) 2019-12-17

Family

ID=57284438

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610496238.7A Active CN106127680B (en) 2016-06-29 2016-06-29 720-degree panoramic video fast browsing method

Country Status (1)

Country Link
CN (1) CN106127680B (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106792151A (en) * 2016-12-29 2017-05-31 上海漂视网络科技有限公司 A kind of virtual reality panoramic video player method
CN107172412A (en) * 2017-06-11 2017-09-15 成都吱吖科技有限公司 A kind of interactive panoramic video storage method and device based on virtual reality
CN107213636A (en) * 2017-05-31 2017-09-29 网易(杭州)网络有限公司 Camera lens moving method, device, storage medium and processor
CN107484004A (en) * 2017-07-24 2017-12-15 北京奇艺世纪科技有限公司 A kind of method for processing video frequency and device
CN108122191A (en) * 2016-11-29 2018-06-05 成都观界创宇科技有限公司 Fish eye images are spliced into the method and device of panoramic picture and panoramic video
WO2018103384A1 (en) * 2016-12-07 2018-06-14 华为技术有限公司 Method, device and system for playing 360 degree panoramic video
CN108269234A (en) * 2016-12-30 2018-07-10 成都观界创宇科技有限公司 A kind of lens of panoramic camera Attitude estimation method and panorama camera
CN108769731A (en) * 2018-05-25 2018-11-06 北京奇艺世纪科技有限公司 The method, apparatus and electronic equipment of target video segment in a kind of detection video
CN111669547A (en) * 2020-05-29 2020-09-15 成都易瞳科技有限公司 Panoramic video structuring method
CN111713106A (en) * 2018-02-14 2020-09-25 高通股份有限公司 Signaling 360 degree video information
CN114342909A (en) * 2022-01-04 2022-04-15 阳光电源股份有限公司 Laser bird repelling method and related device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6560375B1 (en) * 1998-08-26 2003-05-06 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Video image stabilization and registration
CN102833525A (en) * 2012-07-19 2012-12-19 中国人民解放军国防科学技术大学 Browsing operation method of 360-degree panoramic video
CN103338343A (en) * 2013-05-29 2013-10-02 山西绿色光电产业科学技术研究院(有限公司) Multi-image seamless splicing method and apparatus taking panoramic image as reference
CN104219584A (en) * 2014-09-25 2014-12-17 广州市联文信息科技有限公司 Reality augmenting based panoramic video interaction method and system
CN105678693A (en) * 2016-01-25 2016-06-15 成都易瞳科技有限公司 Panorama video browsing-playing method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6560375B1 (en) * 1998-08-26 2003-05-06 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Video image stabilization and registration
CN102833525A (en) * 2012-07-19 2012-12-19 中国人民解放军国防科学技术大学 Browsing operation method of 360-degree panoramic video
CN103338343A (en) * 2013-05-29 2013-10-02 山西绿色光电产业科学技术研究院(有限公司) Multi-image seamless splicing method and apparatus taking panoramic image as reference
CN104219584A (en) * 2014-09-25 2014-12-17 广州市联文信息科技有限公司 Reality augmenting based panoramic video interaction method and system
CN105678693A (en) * 2016-01-25 2016-06-15 成都易瞳科技有限公司 Panorama video browsing-playing method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
夏玉华; 孙建德; 亓靖涛: "《图书馆学术视频快速浏览技术中的关键帧提取》", 《现代图书情报技术》 *
陈立栋,徐玮,包卫东,张茂军,熊志辉: "《一种全景图像浏览器的设计与实现》", 《小型微型计算机系统》 *
陶明明,周源华: "《一种基于亮度帧差的自适应镜头边界检测算法》", 《节目制作与广播》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108122191B (en) * 2016-11-29 2021-07-06 成都美若梦景科技有限公司 Method and device for splicing fisheye images into panoramic image and panoramic video
CN108122191A (en) * 2016-11-29 2018-06-05 成都观界创宇科技有限公司 Fish eye images are spliced into the method and device of panoramic picture and panoramic video
WO2018103384A1 (en) * 2016-12-07 2018-06-14 华为技术有限公司 Method, device and system for playing 360 degree panoramic video
CN108174265A (en) * 2016-12-07 2018-06-15 华为技术有限公司 A kind of playback method, the apparatus and system of 360 degree of panoramic videos
US11184646B2 (en) 2016-12-07 2021-11-23 Huawei Technologies Co., Ltd. 360-degree panoramic video playing method, apparatus, and system
CN106792151A (en) * 2016-12-29 2017-05-31 上海漂视网络科技有限公司 A kind of virtual reality panoramic video player method
CN108269234A (en) * 2016-12-30 2018-07-10 成都观界创宇科技有限公司 A kind of lens of panoramic camera Attitude estimation method and panorama camera
CN107213636A (en) * 2017-05-31 2017-09-29 网易(杭州)网络有限公司 Camera lens moving method, device, storage medium and processor
CN107172412A (en) * 2017-06-11 2017-09-15 成都吱吖科技有限公司 A kind of interactive panoramic video storage method and device based on virtual reality
CN107484004A (en) * 2017-07-24 2017-12-15 北京奇艺世纪科技有限公司 A kind of method for processing video frequency and device
CN107484004B (en) * 2017-07-24 2020-01-03 北京奇艺世纪科技有限公司 Video processing method and device
CN111713106A (en) * 2018-02-14 2020-09-25 高通股份有限公司 Signaling 360 degree video information
CN108769731A (en) * 2018-05-25 2018-11-06 北京奇艺世纪科技有限公司 The method, apparatus and electronic equipment of target video segment in a kind of detection video
CN108769731B (en) * 2018-05-25 2021-09-24 北京奇艺世纪科技有限公司 Method and device for detecting target video clip in video and electronic equipment
CN111669547A (en) * 2020-05-29 2020-09-15 成都易瞳科技有限公司 Panoramic video structuring method
CN111669547B (en) * 2020-05-29 2022-03-11 成都易瞳科技有限公司 Panoramic video structuring method
CN114342909A (en) * 2022-01-04 2022-04-15 阳光电源股份有限公司 Laser bird repelling method and related device

Also Published As

Publication number Publication date
CN106127680B (en) 2019-12-17

Similar Documents

Publication Publication Date Title
CN106127680A (en) A kind of 720 degree of panoramic video fast browsing methods
US11776199B2 (en) Virtual reality environment based manipulation of multi-layered multi-view interactive digital media representations
CN107945282B (en) Rapid multi-view three-dimensional synthesis and display method and device based on countermeasure network
Lee et al. Motion sickness prediction in stereoscopic videos using 3d convolutional neural networks
Zanfir et al. Human appearance transfer
CN103795976B (en) A kind of full-time empty 3 d visualization method
JP2021511729A (en) Extension of the detected area in the image or video data
JP2019534494A (en) Automatic tagging of objects in multi-view interactive digital media representation of dynamic entities
CN101729920B (en) Method for displaying stereoscopic video with free visual angles
CN107240147B (en) Image rendering method and system
Zhang et al. Multiscale-vr: Multiscale gigapixel 3d panoramic videography for virtual reality
CN107862718A (en) 4D holographic video method for catching
Chen et al. Salbinet360: Saliency prediction on 360 images with local-global bifurcated deep network
Ma et al. VommaNet: An End-to-End network for disparity estimation from reflective and texture-less light field images
Shah et al. Video Stitching with Localized 360o Model for Intelligent Car Parking Monitoring and Assistance System
Xu et al. 3D video segmentation using point distance histograms
Khan et al. A review of benchmark datasets and training loss functions in neural depth estimation
Schneider Visual hull
Lu et al. Automatic object extraction and reconstruction in active video
Guo et al. No-reference omnidirectional video quality assessment based on generative adversarial networks
Bagdanov et al. Florence faces: a dataset supporting 2d/3d face recognition
Zhou et al. Time-Series Fusion-Based Multicamera Self-Calibration for Free-View Video Generation in Low-Texture Sports Scene
Patterson 360 Degree photographic imagery for VR: challenges & user experiences
Wang et al. Online photography assistance by exploring geo-referenced photos on MID/UMPC
Zhang et al. Array camera crowd counting method based on Yolov5

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant