CN107846605A - Main broadcaster end stream medium data generation system and method, network direct broadcasting system and method - Google Patents
Main broadcaster end stream medium data generation system and method, network direct broadcasting system and method Download PDFInfo
- Publication number
- CN107846605A CN107846605A CN201710037179.1A CN201710037179A CN107846605A CN 107846605 A CN107846605 A CN 107846605A CN 201710037179 A CN201710037179 A CN 201710037179A CN 107846605 A CN107846605 A CN 107846605A
- Authority
- CN
- China
- Prior art keywords
- video
- module
- audio
- stream medium
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2187—Live feed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/65—Transmission of management data between client and server
- H04N21/658—Transmission by the client directed to the server
- H04N21/6587—Control parameters, e.g. trick play commands, viewpoint selection
Abstract
The invention discloses a kind of main broadcaster end stream medium data generation system and method, network direct broadcasting system and method, wherein main broadcaster end stream medium data generation system includes audio-video collection module, audio/video coding module, Streaming Media package module, scene analysis module and control module, the output end of audio-video collection module is connected by scene analysis module with the input of control module, and the output end of control module is connected with audio/video coding module;The video data that scene analysis module collects to audio-video collection module is analyzed, and judges current scene type NiAnd by NiIt is delivered to control module;Control module is according to NiSelection performs corresponding stream medium data generation scheme in M.The present invention can during network direct broadcasting intelligent distinguishing main broadcaster end state, on the premise of user's viewing experience is ensured, scheme is generated using different stream medium data, reduces the network bandwidth that live broadcast stream media data take, flow is saved, reduces cost.
Description
Technical field
The invention belongs to network direct broadcasting technical field, more particularly to a kind of main broadcaster end stream medium data generation system and side
Method, network direct broadcasting system and method.
Background technology
" network direct broadcasting " is substantially divided to two classes, and the first kind is to provide the viewing of TV signal, such as all kinds of physical culture ratio on the net
Match is live with recreational activities, and this kind of live principle is that TV (simulation) signal is converted into data signal input by collection
Computer, website is uploaded in real time and is watched for people, equivalent to " Web TV ";Second class is then " network direct broadcasting " truly:
Independent signal collecting device collection multi-medium data (including voice data and video data) is set up at the scene, imports main broadcaster end
(instructor in broadcasting's equipment or platform), then uploaded onto the server by network, it is distributed to network address and is watched for people.It is the present invention is directed above-mentioned
The second class situation in network direct broadcasting, especially for personal live, i.e., the most common mode in current live market.Second
Maximum difference of the class network direct broadcasting compared with the first kind is that live independence:Individually controllable audio-video collection, it is entirely different
Watched in single (moreover viewing effect is not so good as the smoothness of television-viewing) of relay television signal.
Multi-medium data during network direct broadcasting generally includes video data and voice data.As shown in figure 1, common
Network direct broadcasting system includes main broadcaster end 1, server 2 and client 3, and wherein main broadcaster end 1 generates stream after collecting multi-medium data
Media data, and stream medium data is uploaded onto the server 2 by network;Server 2 receive client 3 viewing it is live please
After asking, stream medium data is transmitted through the network to client 3.
From accompanying drawing 1 as can be seen that stream medium data needs to be transmitted through journey by uploading and downloading two nets.Because net passes generally
It is required for using the network of operator, therefore how on the premise of Consumer's Experience is ensured, reduces the band taken in network direct broadcasting
Width, flow is saved for user, be always the problem of needing to consider in network direct broadcasting product development.
The stream medium data of net crossing is produced by the stream medium data of main broadcaster end 1 generation system, as shown in Fig. 2 traditional
Main broadcaster end 1 stream medium data generation system include audio-video collection module 4, audio/video coding module 5 and Streaming Media Encapsulation Moulds
Block 6, the output end of audio-video collection module 4 are connected by audio/video coding module 5 with the input of Streaming Media package module 6,
Wherein through the compressed encoding of audio/video coding module 5 after the collection of audio-video collection module 4 audio, video data, then through Streaming Media Encapsulation Moulds
Block 6 is packaged, and finally exports stream medium data.
Generally, netting the stream medium data of biography includes video compression data and audio compression data.Audio compression data is usual
All without too big, the proportion that video compression data accounts in stream medium data is very big, and influence video compression data size because
Element is mainly the complexity (complexity generally includes motion complexity, Texture complication etc.) and encoder of video image, wherein
It is again related to coding parameter after encoder determines.
In network direct broadcasting, the state of main broadcaster is that various, common state is main broadcaster before camera lens and bean vermicelli carries out interaction
Chat, and main broadcaster may sleep while live sometimes, main broadcaster is of short duration sometimes leaves but live continuation, also sometimes may be used
Energy main broadcaster actively shelters from camera lens.The different state in main broadcaster end 1, it is meant that in the different periods, video figure to be encoded
As complexity is that different, current live scheme is all that a set of stream medium data generation scheme tackles all main broadcaster's shapes
State, i.e., the coder parameters of pre-set audio/video coding module 5, the video and voice data that will be generated after compressed encoding,
Stream medium data is packaged into, net spreads out of.The defects of this 1 stream medium data generation method of main broadcaster end is not account for master
The state of broadcasting can change, live picture complexity has significant change (most commonly between static scene and dynamic scene
Change), can have the redundancy of stream medium data for viewing user, waste flow.
All do not go to consider substantially for this problem, in existing direct seeding technique, scheme relatively might have
Following several ways:1. main broadcaster is of short duration when leave, main broadcaster live closing, live interruption in this case, returns manually
Reconnection is needed after coming, Consumer's Experience is influenceed bigger;2. live end setting options, main broadcaster can manually select only transmission sound
Frequency according to, do not propagate video data, this scheme needs main broadcaster's interactive operation, and main broadcaster can only select band with it is straight without video
Broadcast, video data encoding mode is not optimized, solution is relatively rough, and bigger Streaming Media number still be present
According to redundancy.
The content of the invention
Existing main broadcaster end stream medium data generation system tackles all masters using a set of stream medium data generation scheme
State is broadcast, larger stream medium data redundancy be present, increase user watches the cost of network direct broadcasting.It is an object of the present invention to
For above-mentioned the deficiencies in the prior art, there is provided a kind of main broadcaster end stream medium data generation system and method, network direct broadcasting system and
Method, main broadcaster end state can be differentiated during network direct broadcasting, suitable stream medium data generating mode is automatically selected, is protecting
On the premise of demonstrate,proving user's viewing experience, scheme is generated using different stream medium datas, reduces what live broadcast stream media data took
Network bandwidth, and then live flow is saved, reduce the cost that user watches network direct broadcasting.
In order to solve the above technical problems, the technical solution adopted in the present invention is:
A kind of main broadcaster end stream medium data generation system, including audio-video collection module, audio/video coding module and stream matchmaker
Body package module, the output end of audio-video collection module pass through audio/video coding module and the input phase of Streaming Media package module
Even, Streaming Media package module output stream medium data;It, which is structurally characterized in that, also includes scene analysis module and control module, and sound regards
The output end of frequency acquisition module is connected by scene analysis module with the input of control module, the output end and sound of control module
Video encoding module is connected;The video data that wherein scene analysis module is used to collect audio-video collection module divides
Analysis, judges current scene type NiAnd by scene type NiIt is delivered to control module;Control module internal preset has comprising n
Element in scene type set N and stream medium data generation scheme the set M, wherein M and N of element corresponds,Control module is according to NiSelection performs corresponding stream medium data generation scheme in M;The stream medium data generation
Scheme includes setting coding parameter of the audio/video coding module to video data.
The code prosecutor formula of constant code rate generally is used to the coding of video data in existing live application, i.e., in certain time
Interior coding code check can only carry out minor fluctuations on target bit rate.Image subjective quality and figure to be encoded after being compressed due to coding
Size of data after the complexity of picture and compression is relevant, wherein after video compress data size again and current encoder coding
Parameter setting is related.The present invention carries out image procossing and analysis using scene analysis module to video data, judges current scene
Belong to some scene type set in advance, further according to the scene type determined, using different stream medium data generation sides
Case.Because different scene types employs different coding parameters, coding loss size is controlled by adjusting coding parameter, will
The high video image of complexity and the low video image compression of complexity to same subjective quality grade, wherein pass through increase it is complicated
The coding loss of low video image is spent, reaches the purpose of the video data size after reducing compression on the whole.
Further, the output end of the control module is also connected with Streaming Media package module;The stream medium data life
Include control Streaming Media package module into scheme and only encapsulate voice data, or control Streaming Media package module only encapsulates video counts
According to, or control Streaming Media package module encapsulation Voice & Video data.
By said structure, on the premise of user's viewing experience is not influenceed, chosen whether to encapsulate video counts according to scene
According to.Because audio compression data is generally all without too greatly, the proportion that video compression data accounts in stream medium data is very big, at certain
Video data is not packaged under a little scenes, greatly reduces the size of stream medium data.
Based on same inventive concept, present invention also offers a kind of main broadcaster end stream medium data generation method, including step
Suddenly:
Step 1, audio-video collection module collection voice data and video data;
Step 2, the video data that scene analysis module collects to audio-video collection module are analyzed, and are judged current
Scene type NiAnd by scene type NiIt is delivered to control module;
Step 3, control module is according to NiSelection performs corresponding stream medium data generation scheme in M, and according to Streaming Media
Data generation scheme sets coding parameter of the audio/video coding module to video data;Wherein control module (8) internal preset has
A pair of element 1 in scene type set N and stream medium data generation scheme set M, wherein M and N comprising n element
Should,
Step 4, video data of the audio/video coding module in the stream medium data generation scheme selected in step 3
Coding parameter is to video data encoding;Simultaneously to audio data coding.
Further, in addition to step 5, Streaming Media package module generate according to the stream medium data selected in step 3
Scheme only encapsulates voice data, either only encapsulates video data or encapsulation Voice & Video data.
As a kind of preferred embodiment, N={ Ni| i=1,2,3 }, wherein N1=normal scene, N2=picture still scene, N3
=camera lens blocks scene;M={ Mj| j=1,2,3 };Wherein M1=target bit rate of the audio/video coding module to video data is set
For steady state value T1And Streaming Media package module encapsulation Voice & Video data, M2=audio/video coding module is set to video data
Target bit rate be steady state value T2And Streaming Media package module encapsulation Voice & Video data, M3=Streaming Media package module only seals
Fill voice data, wherein T2< T1;N1Corresponding M1, N2Corresponding M2, N3Corresponding M3;
Step 2 Scene analysis module is analyzed video image and judges that the process of current scene classification is as follows:
A. the average brightness value AVG of video image is soughtlumaIf the AVG of continuous S two field pictureslumaLess than preset value Th1, then really
It is N to determine current scene classification3;Otherwise step b is jumped to;
B. noise reduction process is filtered to video image, by frame difference method detect present image relative to previous frame image or
The amount of exercise of former two field pictures, then the image after frame difference method is filtered and handles and count motion pixel number Summov,
If the Sum of continuous S two field picturesmovLess than preset value Th2, it is determined that current scene classification is N2;Otherwise step c is jumped to;
C. determine that current scene classification is N1。
As another preferred embodiment, N={ Ni| i=1~n }, wherein N1=camera lens blocks scene Nm=frame stabilization and
Image motion complexity fluctuates, Nn=picture state is unstable;
M={ Mj| j=1~n }, wherein M1=Streaming Media package module only encapsulates voice data, Mm=set audio frequency and video to compile
Code module is steady state value T to the target bit rate of video datamAnd Streaming Media package module encapsulation Voice & Video data, Mn=set
It is steady state value T to the target bit rate of video data to put audio/video coding module1And Streaming Media package module encapsulation Voice & Video
Data;Wherein m ∈ [2, n-2];TmNumerical value by it is following rule determine:
(if T/ (n-2)) * m >=Th3, then Tm=(T/ (n-2)) * m, otherwise Tm=Th3, wherein T is given object code
Rate;Th3For default target bit rate value;
Step 2 Scene analysis module is analyzed video image and judges that the process of current scene classification is as follows:
A. the average brightness value AVG of video image is soughtlumaIf the AVG of continuous S two field pictureslumaLess than preset value Th1, then really
It is N to determine current scene classification1;Otherwise step b is jumped to;
B. noise reduction process is filtered to video image, by frame difference method detect present image relative to previous frame image or
The amount of exercise of former two field pictures, then the image after frame difference method is filtered and handles and count motion pixel number Summov,
If the Sum of continuous S two field picturesmov∈[Summ,Summ+ Δ Sum), it is determined that current scene classification is Nm;Otherwise step is jumped to
c;Wherein, Summ=(m-1) * (W*H/ (n-2)), Δ Sum=(W*H/ (n-2)), W represent the width of video image, and H represents video
The height of image;
C. determine that current scene classification is Nn。
Based on same inventive concept, present invention also offers a kind of network direct broadcasting system, including described main broadcaster end to flow
Media data generates system.
Based on same inventive concept, present invention also offers a kind of live network broadcast method, including described main broadcaster end to flow
Media data generation method.
Compared with prior art, the present invention can during network direct broadcasting intelligent distinguishing main broadcaster end state, automatically select
Suitable stream medium data generating mode, on the premise of user's viewing experience is ensured, generated using different stream medium datas
Scheme, the network bandwidth that live broadcast stream media data take is reduced, and then save live flow, reduced user and watch network direct broadcasting
Cost.
Brief description of the drawings
Fig. 1 is network direct broadcasting system block diagram.
Fig. 2 is that traditional main broadcaster end stream medium data generates system block diagram.
Fig. 3 is that main broadcaster end of the present invention stream medium data generates system block diagram.
Wherein, 1 is main broadcaster end, and 2 be server, and 3 be client, and 4 be audio-video collection module, and 5 be audio/video coding mould
Block, 6 be Streaming Media package module, and 7 be scene analysis module, and 8 be control module.
Embodiment
As shown in Fig. 2 the main broadcaster end stream medium data generation system in network direct broadcasting system includes audio-video collection module
4th, audio/video coding module 5 and Streaming Media package module 6, the output end of audio-video collection module 4 pass through audio/video coding module 5
It is connected with the input of Streaming Media package module 6, Streaming Media package module 6 exports stream medium data;It is structurally characterized in that and also wrapped
Scene analysis module 7 and control module 8 are included, the output end of audio-video collection module 4 passes through scene analysis module 7 and control module
8 input is connected, and the output end of control module 8 is connected with audio/video coding module 5;Wherein scene analysis module 7 be used for pair
The video data that audio-video collection module 4 collects is analyzed, and judges current scene type NiAnd by scene type NiConveying
To control module 8;The internal preset of control module 8 has the scene type set N comprising n element and stream medium data generation side
Element in case set M, wherein M and N corresponds,Control module 8 is according to NiSelection performs corresponding stream matchmaker in M
Volume data generates scheme;The stream medium data generation scheme includes setting coding of the audio/video coding module 5 to video data
Parameter.
The output end of the control module 8 is also connected with Streaming Media package module 6;The stream medium data generates scheme bag
Include control Streaming Media package module 6 and only encapsulate voice data, or control Streaming Media package module 6 only encapsulates video data, or
Person controls Streaming Media package module 6 to encapsulate Voice & Video data.
The effect of scene analysis module 7 and control module 8 is to carry out algorithm and logic judgment, typically in the form of software
Operate on common processor and (these algorithms and logic judgment can also be realized and solidified by hardware programmable).
The substantially flow of the present invention is, after audio-video collection module 4 gathers voice data and video data, scene analysis module 7 to regarding
Frequency image is analyzed, and intelligently judges that current scene belongs to some scene type set in advance.According to the field determined
Scape classification, scheme is generated using different stream medium datas, these stream medium datas generation scheme includes to be used to Video coding
Different strategy (adjustment coding parameter etc.), the various combination of audio, video data encapsulation, last plug-flow are gone out to carry out net biography.
The implementation process of main broadcaster end stream medium data generation method in inventive network live broadcasting method is as follows:
(1) set-up procedure
(1) some live states often occurred according to main broadcaster end 1 pre-set state scene set N, live state field
Whether for that can be that the scene settings stream medium data generates scheme to reduce stream medium data redundancy, one straight for scape partitioning standards
The state scene of broadcasting is typically the state that main broadcaster can continue within a period of time.
(2) scheme is generated for a set of stream medium data of each state scene category setting, stream medium data generation scheme
Number is consistent with scene state number, and corresponds.Stream medium data generation scheme set is arranged to M, and (M and N set is one by one
It is corresponding), stream medium data generation scheme can use one of following two, or the two is combined:
The first, is adjusted for different state scenes to encoder, including but not limited to encoder to video
The coding parameter of data is adjusted, it is therefore an objective to which finding does not influence the forced coding side of user's subjective effect under the state scene
Case.
Second, the various combination that control Streaming Media package module 6 encapsulates to voice data and video data, including only seal
Fill voice data, only encapsulate video data, encapsulation Voice & Video three kinds of situations of data.
(2) handling process
(1) main broadcaster end 1 carries out adopting for voice data and video data in audio-video collection module 4 by collecting device
Collection;
(2) for the video data after collection, live end state is analyzed using the method for image procossing, and marks
Scene type Si, the foundation of analysis can be color characteristic of the motion complexity of video image, video image etc..Need to indicate
, the specific method of graphical analysis is not limited here, does not also limit the feature chosen in graphical analysis, it is all by figure
As analyzing the scheme sorted out all in the protection domain of this patent;
(3) result analyzed in (2) is sorted out, all scenes is attributed to certain a kind of live scene set in advance
In, i.e. Si∈Ni, wherein (Si∈N).Comprising two kinds of situations, a kind of situation be with correspondingly, i.e., each video image scene
Classification is exactly a live scene, using a kind of Streaming Media generation method;Another situation be with being many-to-one relation, i.e., it is more
The corresponding live scene of individual image scene classification, scheme is generated using same Streaming Media.
(4) it is live scene classification NiStream medium data generation scheme M corresponding to selectioni, wherein Mi∈M。
(5) stream medium data is produced according to the stream medium data of selection in (4) generation scheme, the inside includes audio/video coding
Encapsulated with Streaming Media.
(6) generated stream medium data is transmitted by agreement, so far completes the stream medium data production at main broadcaster end 1
It is raw.
In order to preferably describe whole process, it is assumed that the resolution ratio of live video image is 360 × 640, remembers W × H=360
× 640, normal live target bit rate is arranged to 600kb/s, remembers T1=600.
The live state that the present invention often occurs to main broadcaster end 1 is classified, and the foundation of classification is by substantial amounts of straight
Broadcast video data to be analyzed, and the Streaming Media number that code check is more saved compared to traditional scheme can be made for the state scene
According to generation scheme.Whether corresponded according to graphical analysis scene type and main broadcaster's state, there is following two main broadcaster's state classification
Scheme.
(1) graphical analysis scene type and main broadcaster's state are many-to-one relations, artificially rule of thumb, by common master
The state scene of broadcasting is divided into n (n=3) class:Normal scene, picture still scene, camera lens block scene and (have only arranged three fields here
Scape, scene type can also be increased as needed), it is that a kind of stream medium data of each live state scene setting generates scheme:
Scheme 1:Corresponding normal scene, keeps traditional stream medium data generation method, i.e. stream medium data encapsulation scheme
For voice data+video data, the target bit rate of video data encoding is not changed;
Scheme 2:Corresponding picture still scene (main broadcaster is of short duration to be left, main broadcaster's sleep etc.), video image now to be encoded
Complexity is low, can be by adjusting video coding parameter the methods of, reduces the data volume of video multimedia, and voice data keeps passing
System strategy.Concrete scheme is that stream medium data encapsulation scheme is voice data+video data, changes the object code of Video coding
Rate;
Scheme 3:Corresponding camera lens blocks scene (being usually that main broadcaster is actively blocked), and now whole picture is a piece of black, transmission
Video data is without what meaning, using the scheme for only transmitting voice data.
It can be seen that when graphical analysis scene type and main broadcaster's state are many-to-one relations, N={ Ni| i=1,2,3 }, its
Middle N1=normal scene, N2=picture still scene, N3=camera lens blocks scene;M={ Mj| j=1,2,3 };Wherein M1=set
Audio/video coding module is steady state value T to the target bit rate of video data1And Streaming Media package module encapsulation Voice & Video number
According to M2=setting audio/video coding module is steady state value T to the target bit rate of video data2And Streaming Media package module encapsulation sound
Frequency and video data, M3=Streaming Media package module only encapsulates voice data, wherein T2< T1;N1Corresponding M1, N2Corresponding M2, N3It is right
Answer M3;
Specific implementation steps are as follows:
Step 1, main broadcaster end 1 carry out voice data and video data in audio-video collection module 4 by collecting device
Collection;
Step 2, scene analysis module 7 are directed to the video data that audio-video collection module 4 collects, utilize image procossing
Method live end state is analyzed, judge current scene type NiAnd by scene type NiIt is delivered to control module 8;
The brightness and movement degree of video image are mainly considered during analysis, specific analysis deterministic process is as follows:
A. the average brightness value of video image is sought, counts the luma component values of all pixels in a frame, each pixel
Brightness value is designated as Pi, AVG that average brightness value is designated asluma, wherein,W represents video
The width of image, H represent the height of video image;(if S can elect the integral multiple of GOP length as to continuous S frames, if GOP length is 50
Frame, 2 times of GOP length are then 100 frames) AVG of imagelumaLess than preset value Th1(Th1Be chosen for empirical value, scope be generally (0,
20]), it is determined that current scene classification is N3(camera lens blocks scene);Otherwise step b is jumped to;
B. the pretreatment such as noise reduction is filtered to video image, present image F is detected by frame difference methodcurRelative to the time
Previous frame image F on axlelastThe amount of exercise of (or former two field pictures), then the image after frame difference method is filtered and handles and unites
Meter motion pixel number SummovIf the Sum of continuous S two field picturesmovLess than preset value Th2(Th2It is an empirical, generally
Span is [0, W × H/20]), it is determined that current scene classification is N2(picture still scene);Otherwise step c is jumped to;
C. determine that current scene classification is N1(other states of one of two states of a, b are unsatisfactory for, are all classified as normal
Scene).
Step 3, control module 8 is according to NiSelection performs corresponding stream medium data generation scheme in M, and according to stream matchmaker
Volume data generation scheme sets coding parameter of the audio/video coding module 5 to video data;
Step 4, video counts of the audio/video coding module 5 in the stream medium data generation scheme selected in step 3
According to coding parameter to video data encoding;Simultaneously to audio data coding.
Step 5, Streaming Media package module 6 only encapsulate audio according to the stream medium data generation scheme selected in step 3
Data, either only encapsulate video data or encapsulation Voice & Video data.
The detailed process of step 3 to step 5 is as follows:
Live scene classification of the control module 8 in step 2, the corresponding stream medium data life set in advance of selection
It is as follows into scheme, Scheme Choice:Normal scene selection scheme 1, picture still scene selection scheme 2, camera lens block scene selection
Scheme 3.
Scheme 1:Program implementation process is consistent with traditional live end stream medium data generation method, i.e., compiles audio frequency and video
Code module 5 is arranged to 600kb/s to the target bit rate of video data encoding, and audio data coding still presses traditional scheme, compiles
After code is complete, Streaming Media package module 6 encapsulates into video compression data and audio compression data when encapsulating together;
Scheme 2:The normal scene of ratio that the program sets audio/video coding module 5 to the target bit rate of video data encoding
It is low, 200kb/s is arranged to, after having encoded, by video compression data and audio compression data when Streaming Media package module 6 encapsulates
Encapsulation is entered together;
Scheme 3:In order to not change the flow of voice data and video data encoding, Ke Yizheng in the implementation process of the program
The coding of voice data and video data is often carried out, after simply audio, video data has encoded, when Streaming Media package module 6 encapsulates
Only encapsulation audio compression data, does not encapsulate video compression data.
Generated stream medium data is transmitted by agreement, the stream medium data of main broadcaster end 1 is so far completed and produces.
(2) graphical analysis scene type and main broadcaster's state are man-to-man relations, i.e. main broadcaster's status categories can contemplate mirror
Head blocks, field motion complexity the two factors, it is assumed that main broadcaster's status number is preset as n, and (wherein a kind of classification represents that camera lens hides
Gear, n-2 kind live state classifications are determined that remaining a kind of classification represents that picture state is unstable by different image motion complexities
Fixed live scene), it is corresponding that we should set n kinds stream medium data generation scheme:
Scheme 1:Voice data is only included in stream medium data generation scheme, without video data;
2~scheme of scheme n-2:Stream medium data generation scheme is voice data+video data, is changed according to motion complexity
Become the target bit rate of video data encoding;
Scheme 3:Stream medium data generation scheme is voice data+video data, does not change the target of video data encoding
Code check.
It can be seen that when graphical analysis scene type and main broadcaster's state are many-to-one relations, N={ Ni| i=1~n }, its
Middle N1=camera lens blocks scene, Nm=frame stabilization and the fluctuation of image motion complexity, Nn=picture state is unstable;M={ Mj|
J=1~n }, wherein, M1=Streaming Media package module only encapsulates voice data, Mm=audio/video coding module is set to video counts
According to target bit rate be steady state value TmAnd Streaming Media package module encapsulation Voice & Video data, Mn=audio/video coding mould is set
Block is steady state value T to the target bit rate of video data1And Streaming Media package module encapsulation Voice & Video data;Wherein m ∈ [2,
n-2];
Specific implementation steps are as follows:
Step 1, main broadcaster end 1 carry out voice data and video data in audio-video collection module 4 by collecting device
Collection;
Step 2, scene analysis module 7 are directed to the video data that audio-video collection module 4 collects, utilize image procossing
Method live end state is analyzed, judge current scene type NiAnd by scene type NiIt is delivered to control module 8;
The brightness and movement degree of video image are mainly considered during analysis, specific analysis deterministic process is as follows:
A. the average brightness value of video image is sought, counts the luma component values of all pixels in a frame, each pixel
Brightness value is designated as Pi, AVG that average brightness value is designated asluma, wherein,W represents video
The width of image, H represent the height of video image;(if S can elect the integral multiple of GOP length as to continuous S frames, if GOP length is 50
Frame, 2 times of GOP length are then 100 frames) AVG of imagelumaLess than preset value Th1(Th1Be chosen for empirical value, scope be generally (0,
20]), it is determined that current scene classification is N1(camera lens blocks scene);Otherwise step b is jumped to;
B. the pretreatment such as noise reduction is filtered to video image, present image F is detected by frame difference methodcurRelative to the time
Previous frame image F on axlelastThe amount of exercise of (or former two field pictures), then the image after frame difference method is filtered and handles and unites
Meter motion pixel number Summov.Because the state of a total of n-2 kinds different motion complexity, the judgment mode of each state
For:If Summov∈[Summ,Summ+ Δ Sum), (wherein m ∈ [2, n-2], Summ=(m-1) * (W*H/ (N-2)), Δ Sum
=(W*H/ (N-2)), W represent the width of video image, and H represents the height of video image), and the state continue for the (value selection of S frames
Method be same as above), then current scene kind judging is classification Nm;Otherwise step c is jumped to.What deserves to be explained is, it is based in addition
The classification of motion complexity is not suitable for the meticulous of division.
C. determine that current scene classification is Nn(other states of one of two states of a, b are unsatisfactory for, are all classified as picture
The live scene of state labile).
Step 3, control module 8 is according to NiSelection performs corresponding stream medium data generation scheme in M, and according to stream matchmaker
Volume data generation scheme sets coding parameter of the audio/video coding module 5 to video data;
Step 4, video counts of the audio/video coding module 5 in the stream medium data generation scheme selected in step 3
According to coding parameter to video data encoding;Simultaneously to audio data coding.
Step 5, Streaming Media package module 6 only encapsulate audio according to the stream medium data generation scheme selected in step 3
Data, either only encapsulate video data or encapsulation Voice & Video data.
The detailed process of step 3 to step 5 is as follows:
Live scene classification of the control module 8 in step 2, the corresponding stream medium data life set in advance of selection
It is as follows into scheme, Scheme Choice:Camera lens blocks scene selection scheme 1, frame stabilization and image motion complexity fluctuation selecting party
Case 2, the unstable selection scheme 3 of picture state.
Scheme 1:In order to not change the flow of voice data and video data encoding, Ke Yizheng in the implementation process of the program
The coding of voice data and video data is often carried out, after simply audio, video data has encoded, when Streaming Media package module 6 encapsulates
Only encapsulation audio compression data, does not encapsulate video compression data.
Scheme 2:The program is by mainly adjustment audio/video coding module 5 to the target bit rate of video data encoding.Classification
For NmThe corresponding target bit rate T of scene typemNumerical value by it is following rule determine:
(if T/ (n-2)) * m >=Th3, then Tm=(T/ (n-2)) * m, otherwise Tm=Th3, wherein T takes 600, (in example only
Given this setting scheme of target bit rate, other similar schemes also should be within the scope of the present invention), Th3 choosings
Take with image resolution ratio correlation, the value can be set to 100 under 360 × 640 resolution ratio.After having encoded, Streaming Media Encapsulation Moulds
Block 6 encapsulates into video compression data and audio compression data when encapsulating together.
Scheme 3:Program implementation process is consistent with traditional live end stream medium data generation method, i.e., compiles audio frequency and video
Code module 5 is arranged to 600kb/s to the target bit rate of video data encoding, and audio data coding still presses traditional scheme, compiles
After code is complete, Streaming Media package module 6 encapsulates into video compression data and audio compression data when encapsulating together.
Generated stream medium data is transmitted by agreement, the stream medium data of main broadcaster end 1 is so far completed and produces.
Claims (8)
1. a kind of main broadcaster end stream medium data generate system, including audio-video collection module (4), audio/video coding module (5) and
Streaming Media package module (6), the output end of audio-video collection module (4) are encapsulated by audio/video coding module (5) and Streaming Media
The input of module (6) is connected, Streaming Media package module (6) output stream medium data;Characterized in that, also include scene analysis
Module (7) and control module (8), the output end of audio-video collection module (4) pass through scene analysis module (7) and control module
(8) input is connected, and the output end of control module (8) is connected with audio/video coding module (5);
The video data that wherein scene analysis module (7) is used to collect audio-video collection module (4) is analyzed, and judges to work as
Preceding scene type NiAnd by scene type NiIt is delivered to control module (8);
Control module (8) internal preset has the scene type set N comprising n element and stream medium data generation scheme set
M, wherein M correspond with the element in N,Control module (8) is according to NiSelection performs corresponding Streaming Media number in M
According to generation scheme;The stream medium data generation scheme includes setting audio/video coding module (5) to join the coding of video data
Number.
2. main broadcaster end as claimed in claim 1 stream medium data generates system, it is characterised in that the control module (8)
Output end is also connected with Streaming Media package module (6);The stream medium data generation scheme includes control Streaming Media package module
(6) voice data is only encapsulated, either controls Streaming Media package module (6) only to encapsulate video data or control Streaming Media encapsulation
Module (6) encapsulates Voice & Video data.
3. a kind of main broadcaster end stream medium data generation method, it is characterised in that including step:
Step 1, audio-video collection module (4) collection voice data and video data;
Step 2, the video data that scene analysis module (7) collects to audio-video collection module (4) are analyzed, and judge to work as
Preceding scene type NiAnd by scene type NiIt is delivered to control module (8);
Step 3, control module (8) is according to NiSelection performs corresponding stream medium data generation scheme in M, and according to Streaming Media number
According to generation scheme, coding parameter of the audio/video coding module (5) to video data is set;Wherein control module (8) internal preset has
A pair of element 1 in scene type set N and stream medium data generation scheme set M, wherein M and N comprising n element
Should,
Step 4, video data of the audio/video coding module (5) in the stream medium data generation scheme selected in step 3
Coding parameter is to video data encoding;Simultaneously to audio data coding.
4. main broadcaster end stream medium data generation method as claimed in claim 3, it is characterised in that also include:
Step 5, Streaming Media package module (6) only encapsulate audio number according to the stream medium data generation scheme selected in step 3
According to, either only encapsulate video data or encapsulation Voice & Video data.
5. main broadcaster end stream medium data generation method as claimed in claim 4, it is characterised in that N={ Ni| i=1,2,3 }, its
Middle N1=normal scene, N2=picture still scene, N3=camera lens blocks scene;M={ Mj| j=1,2,3 };Wherein M1=set
Audio/video coding module is steady state value T to the target bit rate of video data1And Streaming Media package module encapsulation Voice & Video number
According to M2=setting audio/video coding module is steady state value T to the target bit rate of video data2And Streaming Media package module encapsulation sound
Frequency and video data, M3=Streaming Media package module only encapsulates voice data, wherein T2< T1;N1Corresponding M1, N2Corresponding M2, N3It is right
Answer M3;
Step 2 Scene analysis module (7) is analyzed video image and judges that the process of current scene classification is as follows:
A. the average brightness value AVG of video image is soughtlumaIf the AVG of continuous S two field pictureslumaLess than preset value Th1, it is determined that when
Preceding scene type is N3;Otherwise step b is jumped to;
B. noise reduction process is filtered to video image, present image is detected relative to previous frame image or former by frame difference method
The amount of exercise of two field picture, then the image after frame difference method is filtered and handles and count motion pixel number SummovIf even
The Sum of continuous S two field picturesmovLess than preset value Th2, it is determined that current scene classification is N2;Otherwise step c is jumped to;
C. determine that current scene classification is N1。
6. main broadcaster end stream medium data generation method as claimed in claim 4, it is characterised in that N={ Ni| i=1~n }, its
Middle N1=camera lens blocks scene, Nm=frame stabilization and the fluctuation of image motion complexity, Nn=picture state is unstable;M={ Mj|
J=1~n }, wherein M1=Streaming Media package module only encapsulates voice data, Mm=audio/video coding module is set to video data
Target bit rate be steady state value TmAnd Streaming Media package module encapsulation Voice & Video data, Mn=audio/video coding module is set
Target bit rate to video data is steady state value T1And Streaming Media package module encapsulation Voice & Video data;Wherein m ∈ [2, n-
2];TmNumerical value by it is following rule determine:
(if T/ (n-2)) * m >=Th3, then Tm=(T/ (n-2)) * m, otherwise Tm=Th3,
Wherein T is given target bit rate;Th3For default target bit rate value;
Step 2 Scene analysis module (7) is analyzed video image and judges that the process of current scene classification is as follows:
A. the average brightness value AVG of video image is soughtlumaIf the AVG of continuous S two field pictureslumaLess than preset value Th1, it is determined that when
Preceding scene type is N1;Otherwise step b is jumped to;
B. noise reduction process is filtered to video image, present image is detected relative to previous frame image or former by frame difference method
The amount of exercise of two field picture, then the image after frame difference method is filtered and handles and count motion pixel number SummovIf even
The Sum of continuous S two field picturesmov∈[Summ,Summ+ Δ Sum), it is determined that current scene classification is Nm;Otherwise step c is jumped to;Its
In, Summ=(m-1) * (W*H/ (n-2)), Δ Sum=(W*H/ (n-2)), W represent the width of video image, and H represents video image
Height;
C. determine that current scene classification is Nn。
7. a kind of network direct broadcasting system, it is characterised in that given birth to including main broadcaster end as claimed in claim 1 or 2 stream medium data
Into system.
8. a kind of live network broadcast method, it is characterised in that including the main broadcaster end Streaming Media as described in any one of claim 3 to 6
Data creation method.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710037179.1A CN107846605B (en) | 2017-01-19 | 2017-01-19 | System and method for generating streaming media data of anchor terminal, and system and method for live network broadcast |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710037179.1A CN107846605B (en) | 2017-01-19 | 2017-01-19 | System and method for generating streaming media data of anchor terminal, and system and method for live network broadcast |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107846605A true CN107846605A (en) | 2018-03-27 |
CN107846605B CN107846605B (en) | 2020-09-04 |
Family
ID=61682781
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710037179.1A Active CN107846605B (en) | 2017-01-19 | 2017-01-19 | System and method for generating streaming media data of anchor terminal, and system and method for live network broadcast |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107846605B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111836060A (en) * | 2019-11-20 | 2020-10-27 | 高群 | Series flow making system based on picture interception |
CN113473148A (en) * | 2020-03-30 | 2021-10-01 | 阿里巴巴集团控股有限公司 | Computing system for video coding and video coding method |
CN115529298A (en) * | 2022-11-30 | 2022-12-27 | 湖南快乐阳光互动娱乐传媒有限公司 | Intensive video and audio transmission system, method and device |
WO2023078204A1 (en) * | 2021-11-02 | 2023-05-11 | 腾讯科技(深圳)有限公司 | Data processing method and apparatus, device, readable storage medium, and program product |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6804295B1 (en) * | 2000-01-07 | 2004-10-12 | International Business Machines Corporation | Conversion of video and audio to a streaming slide show |
CN101330602A (en) * | 2008-04-10 | 2008-12-24 | 王兴忠 | System for monitoring digital video |
CN101453642A (en) * | 2007-11-30 | 2009-06-10 | 华为技术有限公司 | Method, apparatus and system for image encoding/decoding |
CN102577308A (en) * | 2009-09-22 | 2012-07-11 | 高通股份有限公司 | Enhanced block-request streaming using scalable encoding |
CN102625106A (en) * | 2012-03-28 | 2012-08-01 | 上海交通大学 | Scene self-adaptive screen encoding rate control method and system |
CN102780869A (en) * | 2012-06-27 | 2012-11-14 | 宇龙计算机通信科技(深圳)有限公司 | Video recording device and method |
US20130007223A1 (en) * | 2006-06-09 | 2013-01-03 | Qualcomm Incorporated | Enhanced block-request streaming system for handling low-latency streaming |
US20130167187A1 (en) * | 2011-12-21 | 2013-06-27 | Thomson Licensing | Processing cluster and method for processing video content |
CN103617797A (en) * | 2013-12-09 | 2014-03-05 | 腾讯科技(深圳)有限公司 | Voice processing method and device |
CN104243998A (en) * | 2014-09-29 | 2014-12-24 | 广州华多网络科技有限公司 | Data processing method, data processing device and related servers |
-
2017
- 2017-01-19 CN CN201710037179.1A patent/CN107846605B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6804295B1 (en) * | 2000-01-07 | 2004-10-12 | International Business Machines Corporation | Conversion of video and audio to a streaming slide show |
US20130007223A1 (en) * | 2006-06-09 | 2013-01-03 | Qualcomm Incorporated | Enhanced block-request streaming system for handling low-latency streaming |
CN101453642A (en) * | 2007-11-30 | 2009-06-10 | 华为技术有限公司 | Method, apparatus and system for image encoding/decoding |
CN101330602A (en) * | 2008-04-10 | 2008-12-24 | 王兴忠 | System for monitoring digital video |
CN102577308A (en) * | 2009-09-22 | 2012-07-11 | 高通股份有限公司 | Enhanced block-request streaming using scalable encoding |
US20130167187A1 (en) * | 2011-12-21 | 2013-06-27 | Thomson Licensing | Processing cluster and method for processing video content |
CN102625106A (en) * | 2012-03-28 | 2012-08-01 | 上海交通大学 | Scene self-adaptive screen encoding rate control method and system |
CN102780869A (en) * | 2012-06-27 | 2012-11-14 | 宇龙计算机通信科技(深圳)有限公司 | Video recording device and method |
CN103617797A (en) * | 2013-12-09 | 2014-03-05 | 腾讯科技(深圳)有限公司 | Voice processing method and device |
CN104243998A (en) * | 2014-09-29 | 2014-12-24 | 广州华多网络科技有限公司 | Data processing method, data processing device and related servers |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111836060A (en) * | 2019-11-20 | 2020-10-27 | 高群 | Series flow making system based on picture interception |
CN113473148A (en) * | 2020-03-30 | 2021-10-01 | 阿里巴巴集团控股有限公司 | Computing system for video coding and video coding method |
WO2023078204A1 (en) * | 2021-11-02 | 2023-05-11 | 腾讯科技(深圳)有限公司 | Data processing method and apparatus, device, readable storage medium, and program product |
CN115529298A (en) * | 2022-11-30 | 2022-12-27 | 湖南快乐阳光互动娱乐传媒有限公司 | Intensive video and audio transmission system, method and device |
CN115529298B (en) * | 2022-11-30 | 2023-10-13 | 湖南快乐阳光互动娱乐传媒有限公司 | System, method and device for transmitting dense video and audio |
Also Published As
Publication number | Publication date |
---|---|
CN107846605B (en) | 2020-09-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3846477B1 (en) | Preprocessing image data | |
CN107846605A (en) | Main broadcaster end stream medium data generation system and method, network direct broadcasting system and method | |
CN101743753B (en) | A buffer-based rate control exploiting frame complexity, buffer level and position of intra frames in video coding | |
CN108495141A (en) | A kind of synthetic method and system of audio and video | |
CN110139113B (en) | Transmission parameter distribution method and device for video resources | |
CN101466035B (en) | Method for distributing video image set bit based on H.264 | |
US20060050970A1 (en) | Method and apparatus for transmitting a coded video signal | |
CN102625106A (en) | Scene self-adaptive screen encoding rate control method and system | |
CN102137258B (en) | Method for controlling three-dimensional video code rates | |
CN105827633A (en) | Video transmission method and device | |
CN110708570B (en) | Video coding rate determining method, device, equipment and storage medium | |
CN108810530A (en) | A kind of AVC bit rate control methods based on human visual system | |
CN110620924B (en) | Method and device for processing coded data, computer equipment and storage medium | |
CN103051901A (en) | Video data coding device and video data encoding method | |
JP2015513717A (en) | Data, multimedia and video transmission update system | |
CN102780869A (en) | Video recording device and method | |
CN100350801C (en) | Method and equipment for controlling video frequency data quality | |
CN107087192A (en) | Target bit rate method of adjustment and device | |
CN105516721A (en) | Video encoder and bit rate control method thereof | |
CN113573140A (en) | Code rate self-adaptive decision-making method supporting face detection and real-time super-resolution | |
CN105898403A (en) | Online media service code stream self-adaptive method and system | |
CN107222748A (en) | The treating method and apparatus of view data code check | |
WO2023134523A1 (en) | Content adaptive video coding method and apparatus, device and storage medium | |
CN107197266B (en) | HDR video coding method | |
CN110545418B (en) | Self-adaptive video coding method based on scene |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |