CN107613360A - Video data real-time processing method and device, computing device - Google Patents
Video data real-time processing method and device, computing device Download PDFInfo
- Publication number
- CN107613360A CN107613360A CN201710850208.6A CN201710850208A CN107613360A CN 107613360 A CN107613360 A CN 107613360A CN 201710850208 A CN201710850208 A CN 201710850208A CN 107613360 A CN107613360 A CN 107613360A
- Authority
- CN
- China
- Prior art keywords
- processing
- image
- information
- video data
- current frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Image Processing (AREA)
Abstract
The invention discloses a kind of video data real-time processing method and device, computing device, its method includes:Real-time image acquisition collecting device is captured and/or the video recorded in include special object current frame image;Or the current frame image that special object is included in currently played video is obtained in real time;Scene cut processing is carried out to current frame image, obtains the foreground image that current frame image is directed to special object;The input information of external input sources is obtained, at least one information element is extracted from input information;Stylized processing is carried out to background image according at least one information element;Background image after foreground image and stylization processing is subjected to fusion treatment, obtains the image after present frame processing;Image after present frame is handled covers the video data after former current frame image is handled;Show the video data after the processing.The present invention uses deep learning method, realizes completion scene cut and the three dimensional stress processing of the high accuracy of high efficiency.
Description
Technical field
The present invention relates to image processing field, and in particular to a kind of video data real-time processing method and device, calculating are set
It is standby.
Background technology
With the development of science and technology, the technology of image capture device also increasingly improves.Regarded using what image capture device was recorded
Frequency also becomes apparent from, resolution ratio, display effect also greatly improve.But the video of existing recording is only dull recorded material sheet
Body, the increasing individual requirement that user proposes can not be met.Prior art can be manual by user after recorded video
It is further again to video to be handled, to meet the individual requirement of user.But so processing needs user to have higher figure
As treatment technology, and the cost user more time is needed in processing, handle cumbersome, technical sophistication.
Therefore, it is necessary to a kind of video data real-time processing method, to meet the individual requirement of user in real time.
The content of the invention
In view of the above problems, it is proposed that the present invention so as to provide one kind overcome above mentioned problem or at least in part solve on
State the video data real-time processing method and device, computing device of problem.
According to an aspect of the invention, there is provided a kind of video data real-time processing method, it includes:
Real-time image acquisition collecting device is captured and/or the video recorded in include special object present frame figure
Picture;Or the current frame image that special object is included in currently played video is obtained in real time;
Scene cut processing is carried out to current frame image, obtains the foreground image that current frame image is directed to special object;
The input information of external input sources is obtained, at least one information element is extracted from input information;
Stylized processing is carried out to background image according at least one information element;
Background image after foreground image and stylization processing is subjected to fusion treatment, obtains the figure after present frame processing
Picture;
Image after present frame is handled covers the video data after former current frame image is handled;
Show the video data after the processing.
Alternatively, it is music to input information;At least one information element includes:Amplitude, frequency and/or tone color.
Alternatively, stylization processing is carried out to background image according at least one information element to further comprise:
Value according to amplitude, frequency and/or tone color chooses the changing pattern that stylized processing is carried out to background image;Its
In, selected changing pattern is different and different according to the value of amplitude, frequency and/or tone color;
Stylized processing is carried out to background image using changing pattern.
Alternatively, the background image or default background that background image obtains to carry out scene cut to handle to current frame image
Image.
Alternatively, before the background image after foreground image and stylization processing being carried out into fusion treatment, method also includes:
Special object is subjected to three dimensional stress processing.
Alternatively, before the background image after by foreground image and stylization processing carries out fusion treatment, method is also wrapped
Include:
At least one dynamic effect to be loaded is generated according at least one information element.
Alternatively, at least one dynamic effect to be loaded is generated according at least one information element to further comprise:
According at least one information element obtain the colouring information of each dynamic effect to be loaded, positional information and/
Or angle information;
Each dynamic effect is generated according to colouring information, positional information and/or angle information.
Alternatively, colouring information, the position of each dynamic effect to be loaded are obtained according at least one information element
Information and/or angle information further comprise:
Value according to amplitude, frequency and/or tone color obtains the colouring information of each dynamic effect to be loaded, position
Information and/or angle information, wherein, colouring information, positional information and/or angle information are according to amplitude, frequency and/or tone color
Value it is different and different.
Alternatively, the background image after foreground image and stylization processing is subjected to fusion treatment, obtains present frame processing
Image afterwards further comprises:
Background image after foreground image and stylization processing is subjected to fusion treatment and integral color processing, and is loaded onto
A kind of few dynamic effect, obtain the image after present frame processing.
Alternatively, dynamic effect is light-illuminating effect.
Alternatively, the video data after display processing further comprises:By the video data real-time display after processing;
Method also includes:Video data after processing is uploaded to Cloud Server.
Alternatively, the video data after processing is uploaded into Cloud Server to further comprise:
Video data after processing is uploaded to cloud video platform server, so that cloud video platform server is in cloud video
Platform is shown video data.
Alternatively, the video data after processing is uploaded into Cloud Server to further comprise:
Video data after processing is uploaded to cloud direct broadcast server, so that cloud direct broadcast server pushes away video data in real time
Give viewing subscription client.
Alternatively, the video data after processing is uploaded into Cloud Server to further comprise:
Video data after processing is uploaded to cloud public number server, so that cloud public number server pushes away video data
Give public number concern client.
According to another aspect of the present invention, there is provided a kind of video data real-time processing device, it includes:
Acquisition module, suitable for captured by real-time image acquisition collecting device and/or in the video recorded comprising specific right
The current frame image of elephant;Or the current frame image that special object is included in currently played video is obtained in real time;
Split module, suitable for current frame image carry out scene cut processing, obtain current frame image be directed to it is specific right
The foreground image of elephant;
Extraction module, suitable for obtaining the input information of external input sources, extracting at least one information from input information will
Element;
Stylized module, suitable for carrying out stylized processing to background image according at least one information element;
Fusion Module, suitable for the background image after foreground image and stylization processing is carried out into fusion treatment, obtain current
Image after frame processing;
Overlay module, the video data after former two field picture is handled is covered suitable for the image after present frame is handled;
Display module, suitable for the video data after display processing.
Alternatively, it is music to input information;At least one information element includes:Amplitude, frequency and/or tone color.
Alternatively, stylized module is further adapted for:
Value according to amplitude, frequency and/or tone color chooses the changing pattern that stylized processing is carried out to background image;Its
In, selected changing pattern is different and different according to the value of amplitude, frequency and/or tone color;Using changing pattern to the back of the body
Scape image carries out stylized processing.
Alternatively, the background image or default background that background image obtains to carry out scene cut to handle to current frame image
Image.
Alternatively, device also includes:
Three-dimensional process module, suitable for special object is carried out into three dimensional stress processing.
Alternatively, device also includes:
Generation module, suitable for generating at least one dynamic effect to be loaded according at least one information element.
Alternatively, generation module is further adapted for:
According at least one information element obtain the colouring information of each dynamic effect to be loaded, positional information and/
Or angle information;Each dynamic effect is generated according to colouring information, positional information and/or angle information.
Alternatively, generation module is further adapted for:
Value according to amplitude, frequency and/or tone color obtains the colouring information of each dynamic effect to be loaded, position
Information and/or angle information, wherein, colouring information, positional information and/or angle information are according to amplitude, frequency and/or tone color
Value it is different and different.
Alternatively, Fusion Module is further adapted for:
Background image after foreground image and stylization processing is subjected to fusion treatment and integral color processing, and is loaded onto
A kind of few dynamic effect, obtain the image after present frame processing.
Alternatively, dynamic effect is light-illuminating effect.
Alternatively, display module is further adapted for:By the video data real-time display after processing;
Device also includes:
Uploading module, suitable for the video data after processing is uploaded into Cloud Server.
Alternatively, uploading module is further adapted for:
Video data after processing is uploaded to cloud video platform server, so that cloud video platform server is in cloud video
Platform is shown video data.
Alternatively, uploading module is further adapted for:
Video data after processing is uploaded to cloud direct broadcast server, so that cloud direct broadcast server pushes away video data in real time
Give viewing subscription client.
Alternatively, uploading module is further adapted for:
Video data after processing is uploaded to cloud public number server, so that cloud public number server pushes away video data
Give public number concern client.
According to another aspect of the invention, there is provided a kind of computing device, including:Processor, memory, communication interface and
Communication bus, the processor, the memory and the communication interface complete mutual communication by the communication bus;
The memory is used to deposit an at least executable instruction, and the executable instruction makes the computing device above-mentioned
Operated corresponding to video data real-time processing method.
In accordance with a further aspect of the present invention, there is provided a kind of computer-readable storage medium, be stored with the storage medium to
A few executable instruction, the executable instruction make computing device be grasped as corresponding to above-mentioned video data real-time processing method
Make.
According to video data real-time processing method provided by the invention and device, computing device, real-time image acquisition collection
Equipment is captured and/or the video recorded in include special object current frame image;Or obtain in real time currently played
Video in include the current frame image of special object;Scene cut processing is carried out to current frame image, obtains current frame image
It is directed to the foreground image of special object;The input information of external input sources is obtained, at least one letter is extracted from input information
Cease key element;Stylized processing is carried out to background image according at least one information element;After foreground image and stylization processing
Background image carry out fusion treatment, obtain present frame processing after image;Image covering after present frame is handled is former current
Two field picture handled after video data;Show the video data after the processing.The present invention carries out field to current frame image
Scape dividing processing, obtain being directed to the foreground image of special object, will according at least one information in the input information of extraction
Element carries out stylized processing to background image, makes the style of background image and the input information match of external input sources.Again will
Background image after foreground image and stylization processing carries out fusion treatment, the video after processing is integrally presented and outside input
The display effect of the input information match in source.The present invention can directly obtain the video after processing, it is not necessary to which user is to recording
Video carry out extra process, save user time, can with real-time display to user handle after video data, facilitate user
Check display effect.Present invention employs deep learning method, completes scene cut processing with realizing the high accuracy of high efficiency
With three dimensional stress processing.And user's technical merit is not limited simultaneously, facilitate public use.
Described above is only the general introduction of technical solution of the present invention, in order to better understand the technological means of the present invention,
And can be practiced according to the content of specification, and in order to allow above and other objects of the present invention, feature and advantage can
Become apparent, below especially exemplified by the embodiment of the present invention.
Brief description of the drawings
By reading the detailed description of hereafter preferred embodiment, it is various other the advantages of and benefit it is common for this area
Technical staff will be clear understanding.Accompanying drawing is only used for showing the purpose of preferred embodiment, and is not considered as to the present invention
Limitation.And in whole accompanying drawing, identical part is denoted by the same reference numerals.In the accompanying drawings:
Fig. 1 shows the flow chart of video data real-time processing method according to an embodiment of the invention;
Fig. 2 shows the flow chart of video data real-time processing method in accordance with another embodiment of the present invention;
Fig. 3 shows the functional block diagram of video data real-time processing device according to an embodiment of the invention;
Fig. 4 shows the functional block diagram of video data real-time processing device in accordance with another embodiment of the present invention;
Fig. 5 shows a kind of structural representation of computing device according to an embodiment of the invention.
Embodiment
The exemplary embodiment of the disclosure is more fully described below with reference to accompanying drawings.Although the disclosure is shown in accompanying drawing
Exemplary embodiment, it being understood, however, that may be realized in various forms the disclosure without should be by embodiments set forth here
Limited.On the contrary, these embodiments are provided to facilitate a more thoroughly understanding of the present invention, and can be by the scope of the present disclosure
Completely it is communicated to those skilled in the art.
Fig. 1 shows the flow chart of video data real-time processing method according to an embodiment of the invention.Such as Fig. 1 institutes
Show, video data real-time processing method specifically comprises the following steps:
Step S101, real-time image acquisition collecting device is captured and/or the video recorded in comprising special object
Current frame image;Or the current frame image that special object is included in currently played video is obtained in real time.
Image capture device illustrates by taking mobile terminal as an example in the present embodiment.Get mobile terminal camera in real time
Current frame image when current frame image in recorded video or shooting video.Due to the present invention to special object at
Reason, therefore only obtain the current frame image comprising special object during acquisition current frame image.Except real-time image acquisition collecting device
Outside the captured and/or video recorded, it can also obtain in real time current comprising special object in currently played video
Two field picture.
Step S102, scene cut processing is carried out to current frame image, current frame image is obtained and is directed to special object
Foreground image.
Special object, such as human body are contained in current frame image.Scene cut processing is carried out to current frame image, mainly
Special object is split from current frame image, obtains the foreground image that current frame image is directed to special object, before this
Scape image can only include special object.
When carrying out scene cut processing to current frame image, deep learning method can be utilized.Deep learning is machine
It is a kind of based on the method that data are carried out with representative learning in study.Observation (such as piece image) can use various ways
Represent, such as the vector of each pixel intensity value, or be more abstractively expressed as a series of sides, the region etc. of given shape.And make
It is easier with some specific method for expressing from example learning task (for example, recognition of face or human facial expression recognition).Such as profit
Scene cut can be carried out to current frame image, obtain including the foreground image of human body with human body segmentation's method of deep learning.
Step S103, the input information of external input sources is obtained, at least one information element is extracted from input information.
The real-time input information of external input sources is obtained, at least one information element is extracted from real-time input information.
When extracting information element, extracted according to specific external input sources.The information element of extract real-time is worked as according to acquisition
When the input information of external input sources extracted, when the input information for the external input sources that each moment gets is different
When, the specific value of the information element of extraction also can be different.
Step S104, stylized processing is carried out to background image according at least one information element.
, can be according to an information element when carrying out stylization processing to background image according at least one information element
Stylized processing is carried out to background image, or stylized processing is carried out to background image according to multiple information elements.
Background image can be that current frame image is carried out to carry on the back used in the current frame image that scene cut handles to obtain
Scape image, or default background image.
Step S105, the background image after foreground image and stylization processing is subjected to fusion treatment, obtained at present frame
Image after reason.
Background image after foreground image and stylization processing is subjected to fusion treatment, in fusion, to make foreground image
Background image after being handled with stylization can be merged preferably, when carrying out dividing processing to current frame image, to splitting
The edge of the perspective process arrived carries out translucent processing, the edge of special object is obscured, preferably to merge.
Step S106, the image after present frame is handled cover the video data after former current frame image is handled.
Image after being handled using present frame directly overrides former current frame image, the video after directly can be processed
Data.Meanwhile the user of recording can also be immediately seen the image after present frame processing.
In the image after obtaining present frame processing, the image after can present frame be handled directly covers former present frame figure
Picture.Speed during covering, typically completed within 1/24 second.For a user, because the time of covering treatment is relative
Short, human eye is not discovered significantly, i.e., human eye does not perceive the process that the former current frame image in video data is capped.This
During video data of the sample after follow-up display processing, shoot and/or record equivalent to one side and/or during playing video data, one
Side real-time display does not feel as the display effect that two field picture in video data covers for the video data after processing, user
Fruit.
Step S107, the video data after display processing.
After video data after being handled, it can be shown in real time, after user can directly be seen that processing
Video data display effect.
According to video data real-time processing method provided by the invention, real-time image acquisition collecting device it is captured and/or
The current frame image of special object is included in the video recorded;Or obtain in real time in currently played video comprising spy
Determine the current frame image of object;Scene cut processing is carried out to current frame image, current frame image is obtained and is directed to special object
Foreground image;The input information of external input sources is obtained, at least one information element is extracted from input information;According at least
One information element carries out stylized processing to background image;Background image after foreground image and stylization processing is melted
Conjunction is handled, and obtains the image after present frame processing;After the former current frame image of image covering after present frame is handled obtains processing
Video data;Show the video data after the processing.The present invention carries out scene cut processing to current frame image, obtains pin
For the foreground image of special object, sector-style is entered to background image according at least one information element in the input information of extraction
Format processing, make the style of background image and the input information match of external input sources.Again by foreground image and stylization
Background image after reason carries out fusion treatment, makes the video after processing that the input information match with external input sources integrally be presented
Display effect.The present invention can directly obtain the video after processing, it is not necessary to and user carries out extra process to the video of recording,
User time is saved, the video data after being handled with real-time display to user, facilitates user to check display effect.The present invention
Deep learning method is employed, complete scene cut processing with realizing the high accuracy of high efficiency.And simultaneously to user's technology water
It is flat not to be limited, facilitate public use.
Fig. 2 shows the flow chart of video data real-time processing method in accordance with another embodiment of the present invention.Such as Fig. 2 institutes
Show, video data real-time processing method specifically comprises the following steps:
Step S201, real-time image acquisition collecting device is captured and/or the video recorded in comprising special object
Current frame image;Or the current frame image that special object is included in currently played video is obtained in real time.
Step S202, to current frame image carry out scene cut processing, obtain current frame image be directed to it is described specific right
The foreground image of elephant.
Above step will not be repeated here with reference to the description of the step S101-S102 in the embodiment of figure 1.
Step S203, the input information of external input sources is obtained, at least one information element is extracted from input information.
The real-time input information of external input sources is obtained, at least one information element is extracted from real-time input information.
When extracting information element, extracted according to specific external input sources.The input information of external input sources can be outside
Music, sound etc..When such as to input information be music, the information element of extraction includes the information elements such as amplitude, frequency, tone color.
The information element of extract real-time is extracted according to the input information of the external input sources at that time of acquisition, is obtained when each moment
During the input information difference of the external input sources arrived, the specific value of the information element of extraction also can be different.
Step S204, stylized processing is carried out to background image according at least one information element.
Stylized processing is carried out to background image according at least one information element.Specially according to shaking in information element
The value of width, frequency and/or tone color chooses the changing pattern that stylized processing is carried out to background image.Wherein, selected change
Change pattern is different and different according to the value of amplitude, frequency and/or tone color.It can believe when choosing changing pattern according only to one
Breath key element such as amplitude value is chosen, and can also be selected according to multiple information elements such as value of amplitude, frequency and tone color
Take.Stylized processing is carried out to background image using the changing pattern of selection.Changing pattern can be included such as filter, foundation information
Filter corresponding to select factors, filter of such as missing old times or old friends, Blues filter, handsome filter, sets background image according to the filter of selection
Filter style corresponding to being set to.
The current frame image that above-mentioned background image can handle to obtain to carry out current frame image scene cut is made
Background image, or default background image.
Step S205, at least one dynamic effect to be loaded is generated according at least one information element.
One or more dynamic effects to be loaded can be generated according to an information element, or will according to multiple information
Element generates a kind of dynamic effect to be loaded;Different dynamic effects can be generated according to different information elements.
Dynamic effect includes colouring information, positional information, angle information etc..Obtain and treat according at least one information element
Colouring information, positional information and/or the angle information of each dynamic effect of loading.According to colouring information, positional information and/
Or angle information generates each dynamic effect.Specifically, the value according to the amplitude in information element, frequency and/or tone color
Colouring information, positional information and/or the angle information of each dynamic effect to be loaded are obtained, wherein, colouring information, position
Information and/or angle information are according to the different and different of the value of amplitude, frequency and/or tone color.If dynamic effect is light photograph
Effect is penetrated, the value according to the amplitude in information element, frequency and/or tone color can generate the color letter of light-illuminating effect
Breath, positional information, angle information etc..During generation, the colouring information of light-illuminating effect can be generated according to the value of amplitude;Or
Positional information of the person according to the value generation light-illuminating effect of amplitude;Or generate light-illuminating effect according to the value of frequency
Positional information etc..Specifically the light-illuminating effect color information of the generation of amplitude, frequency, the value of tone color and light, position
Information, the corresponding relation of angle information do not limit herein.
Step S206, special object is subjected to three dimensional stress processing.
To make the display effect of the dynamic effect of loading more three-dimensional, special object can be subjected to three dimensional stress processing.It is special
Determine object to illustrate by taking human face as an example, when such as dynamic effect being light-illuminating effect, when light is on the right side of human face
During irradiation, it should be unable to be arrived in real life on the left of human face by light-illuminating., can be with after the processing of human face three dimensional stress
Realize the display effect that will not be arrived on the left of human face by light-illuminating.But if being handled without three dimensional stress, human face is
The image of two dimension, then it can also be arrived on the left of human face by light-illuminating, display effect can be untrue.
When special object is carried out into three dimensional stress processing, three dimensional stress processing can be carried out by deep learning.Specifically, such as make
Human face is subjected to three dimensional stress processing with deep learning, extracts the key message of face.The key message can be specially to close
Key point information, key area information, and/or key lines information etc..Embodiments of the invention are said by taking key point information as an example
It is bright, but the key message of the present invention is not limited to key point information.It can be improved according to key point information using key point information
The processing speed and efficiency of three dimensional stress processing are carried out, three dimensional stress processing directly can be carried out according to key point information, it is not necessary to again
The complex operations such as subsequently calculating, analysis are carried out to key message.Meanwhile key point information is easy to extract, and extract accurate, progress
The effect of three dimensional stress processing is more accurate.When carrying out three dimensional stress processing, the faceform of three-dimensional is first built.It is base to build threedimensional model
Identity and Expression Reformation matrix in 3D face databases, the set for the key point information of a given face can
Identity, Expression Reformation coefficient and rotation scaling translation parameters are tried to achieve by way of coordinate declines (coordinate descent)
Euclidean distance is restrained, and then constructs the three-dimensional structure model of corresponding face.Human face is entered using three-dimensional structure model
The processing of row three dimensional stress, obtains the face of three dimensional stress.It should be noted that the special object after three dimensional stress processing does not have texture special
Reference ceases.Further extract the image texture information of special object in current frame image, image texture information record present frame
The information such as the spatial color distribution of special object and light distribution in image.Can when extracting the image texture information of special object
To be carried using such as the methods of LBP (Local binary patterns) local binary patterns method, gray level co-occurrence matrixes
Take.The special object after three dimensional stress processing is drawn according to the image texture information of the special object extracted, wrapped
The three dimensional stress special object of textural characteristics is contained.
Step S207, the background image after foreground image and stylization processing is carried out at fusion treatment and integral color
Reason, and at least one dynamic effect is loaded, obtain the image after present frame processing.
Background image after foreground image and stylization processing is first subjected to fusion treatment, and carried out at overall tone
Reason, so that the image after fusion is more natural.On this basis, at least one dynamic effect, realization and external input sources are loaded
Input information match present frame processing after image.It is music such as to input information, and dynamic effect is light show irradiation effect
The light-illuminating effect of fruit, background image are the background picture of discotheque style, and one kind is integrally presented in the image after present frame processing
With display effect of the personage of music change in discotheque.
Further, to allow the background image after foreground image and stylization processing preferably to merge, to present frame
When image carries out dividing processing, the edge of the perspective process obtained to segmentation carries out translucent processing, obscures the side of special object
Edge, preferably to merge.
Step S208, the image after present frame is handled cover the video data after former current frame image is handled.
Image after being handled using present frame directly overrides former current frame image, the video after directly can be processed
Data.Meanwhile the user of recording can also be immediately seen the image after present frame processing.
Step S209, the video data after display processing.
After video data after being handled, it can be shown in real time, after user can directly be seen that processing
Video data display effect.
Step S210, the video data after processing is uploaded to Cloud Server.
Video data after processing can be directly uploaded to Cloud Server, specifically, can be by the video counts after processing
According to be uploaded to one or more cloud video platform server, such as iqiyi.com, youku.com, fast video cloud video platform server,
So that cloud video platform server is shown video data in cloud video platform.Or can also be by the video data after processing
Cloud direct broadcast server is uploaded to, can be straight by cloud when the user for having live viewing end is watched into cloud direct broadcast server
Broadcast server and give video data real time propelling movement to viewing subscription client.Or the video data after processing can also be uploaded to
Cloud public number server, when there is user to pay close attention to the public number, video data is pushed to public number by cloud public number server
Pay close attention to client;Further, cloud public number server can also be accustomed to according to the viewing of the user of concern public number, and push meets
The video data of user's custom pays close attention to client to public number.
According to video data real-time processing method provided by the invention, according at least one information element of extraction, generation
At least one dynamic effect to be loaded.Background image after foreground image and stylization processing is subjected to fusion treatment, adjustment
Integral color, and the dynamic effect of load information key element generation, make the video after processing integrally present defeated with external input sources
Enter the display effect of information match.Meanwhile to make the display effect of the dynamic effect of loading more three-dimensional, can will be specific right
As carrying out three dimensional stress processing, so that the display effect of the video after processing is closer to truly.The present invention can directly obtain processing
Video afterwards, the video after processing can also be directly uploaded to Cloud Server, it is not necessary to which user carries out volume to the video of recording
Outer processing, user time is saved, the video data after being handled with real-time display to user, facilitate user to check display effect
Fruit.User's technical merit is not limited simultaneously, facilitates public use.
Fig. 3 shows the functional block diagram of video data real-time processing device according to an embodiment of the invention.Such as Fig. 3 institutes
Show, video data real-time processing device includes following module:
Acquisition module 301, suitable for captured by real-time image acquisition collecting device and/or in the video recorded comprising specific
The current frame image of object;Or the current frame image that special object is included in currently played video is obtained in real time.
Image capture device illustrates by taking mobile terminal as an example in the present embodiment.Acquisition module 301 gets shifting in real time
Move current frame image when current frame image or shooting video of the terminal camera in recorded video.Because the present invention is to spy
Determine object to be handled, therefore the current frame image comprising special object is only obtained during the acquisition current frame image of acquisition module 301.
For acquisition module 301 except captured by real-time image acquisition collecting device and/or in addition to the video recorded, acquisition module 301 can be with
The current frame image that special object is included in currently played video is obtained in real time.Special object can be image in the present invention
In human body, plant, any object such as animal, illustrated in embodiment by taking human body as an example, but be not limited only to human body.
Split module 302, suitable for current frame image carry out scene cut processing, obtain current frame image be directed to it is specific
The foreground image of object.
Current frame image contains special object, such as human body.Split module 302 to carry out at scene cut current frame image
Reason, special object is mainly split, obtain the prospect that current frame image is directed to special object from current frame image
Image, the foreground image can only include special object.
Split module 302 when carrying out scene cut processing to current frame image, deep learning method can be utilized.Depth
Study is a kind of based on the method that data are carried out with representative learning in machine learning.Observation (such as piece image) can use
Various ways represent, a series of such as vector of each pixel intensity value, or be more abstractively expressed as sides, the area of given shape
Domain etc..And some specific method for expressing are used to be easier from example learning task (for example, recognition of face or facial expression are known
Not).Scene cut can be carried out using human body segmentation's method of deep learning to current frame image by such as splitting module 302, be obtained
Foreground image comprising human body.
Extraction module 303, suitable for obtaining the input information of external input sources, at least one information is extracted from input information
Key element.
Extraction module 303 obtains the real-time input information of external input sources, and at least one is extracted from real-time input information
Individual information element.The input information of external input sources can be outside music, sound etc..When such as to input information be music, carry
The information element that modulus block 303 extracts includes the information elements such as amplitude, frequency, tone color.Extraction module 303 is in extraction information element
When, extracted according to specific external input sources.The information element of the extract real-time of extraction module 303 according to acquisition at that time
The input information of external input sources is extracted, and when the input information difference for the external input sources that each moment gets, is carried
The specific value for the information element that modulus block 303 extracts also can be different.
Stylized module 304, suitable for carrying out stylized processing to background image according at least one information element.
Stylized module 304 to background image according at least one information element when carrying out stylization processing, stylization
Module 304 can carry out stylized processing according to an information element to background image, or stylized module 304 is according to multiple
Information element carries out stylized processing to background image.
Specially stylized module 304 is chosen to background according to the value of the amplitude in information element, frequency and/or tone color
Image carries out the changing pattern of stylized processing.Wherein, the changing pattern selected by stylized module 304 is according to amplitude, frequency
And/or the value of tone color is different and different.Can be according only to an information element during the stylized selection of module 304 changing pattern
As amplitude value is chosen, can also be chosen according to multiple information elements such as value of amplitude, frequency and tone color.Style
Change module 304 and stylized processing is carried out to background image using the changing pattern chosen.Changing pattern can be included such as filter, wind
Module of formatting 304 is according to filter, filter of such as missing old times or old friends, Blues filter, handsome filter, stylized mould corresponding to information element selection
Block 304 background image is arranged to according to the filter of selection corresponding to filter style.
Background image can be that current frame image is carried out to carry on the back used in the current frame image that scene cut handles to obtain
Scape image, or default background image.
Fusion Module 305, suitable for the background image after foreground image and stylization processing is carried out into fusion treatment, worked as
Image after previous frame processing.
Background image after foreground image and stylization processing is carried out fusion treatment, Fusion Module 305 by Fusion Module 305
In fusion, to allow the background image after foreground image and stylization processing preferably to merge, segmentation module 302 is to working as
When prior image frame carries out dividing processing, the edge of the perspective process obtained to segmentation carries out translucent processing, obscures special object
Edge, so that Fusion Module 305 preferably merges.
Overlay module 306, the video data after former two field picture is handled is covered suitable for the image after present frame is handled.
Image after overlay module 306 is handled using present frame directly overrides former current frame image, can directly obtain
Video data after processing.Meanwhile the user of recording can also be immediately seen the image after present frame processing.
When Fusion Module 305 obtains the image after present frame processing, overlay module 306 understands the figure after present frame be handled
As directly covering former current frame image.Speed when overlay module 306 covers, was typically completed within 1/24 second.For
For user, because the time of the covering treatment of overlay module 306 is relatively short, human eye is not discovered significantly, i.e., human eye is not examined
Feel the process that the former current frame image in video data is capped.So regarding after the follow-up display processing of display module 307
Frequency according to when, shoot and/or record equivalent to one side and/or during playing video data, the real-time display of one side display module 307
For the video data after processing, user does not feel as the display effect that two field picture in video data covers.
Display module 307, suitable for the video data after display processing.
Display module 307 handled after video data after, it can be shown in real time, user can be direct
See the display effect of the video data after processing.
According to video data real-time processing device provided by the invention, real-time image acquisition collecting device it is captured and/or
The current frame image of special object is included in the video recorded;Or obtain in real time in currently played video comprising spy
Determine the current frame image of object;Scene cut processing is carried out to current frame image, current frame image is obtained and is directed to special object
Foreground image;The input information of external input sources is obtained, at least one information element is extracted from input information;According at least
One information element carries out stylized processing to background image;Background image after foreground image and stylization processing is melted
Conjunction is handled, and obtains the image after present frame processing;After the former current frame image of image covering after present frame is handled obtains processing
Video data;Show the video data after the processing.The present invention carries out scene cut processing to current frame image, obtains pin
For the foreground image of special object, sector-style is entered to background image according at least one information element in the input information of extraction
Format processing, make the style of background image and the input information match of external input sources.Again by foreground image and stylization
Background image after reason carries out fusion treatment, makes the video after processing that the input information match with external input sources integrally be presented
Display effect.The present invention can directly obtain the video after processing, it is not necessary to and user carries out extra process to the video of recording,
User time is saved, the video data after being handled with real-time display to user, facilitates user to check display effect.The present invention
Deep learning method is employed, complete scene cut processing with realizing the high accuracy of high efficiency.And simultaneously to user's technology water
It is flat not to be limited, facilitate public use.
Fig. 4 shows the functional block diagram of video data real-time processing device in accordance with another embodiment of the present invention.Such as Fig. 4
It is shown, it is with Fig. 3 differences, video data real-time processing device also includes:
Generation module 308, suitable for generating at least one dynamic effect to be loaded according at least one information element.
Generation module 308 can generate one or more dynamic effects to be loaded, Huo Zhesheng according to an information element
A kind of dynamic effect to be loaded is generated according to multiple information elements into module 308;Generation module 308 will according to different information
Element can generate different dynamic effects.
Dynamic effect includes colouring information, positional information, angle information etc..Generation module 308 is according at least one letter
Cease colouring information, positional information and/or angle information that key element obtains each dynamic effect to be loaded.Generation module 308
Each dynamic effect is generated according to colouring information, positional information and/or angle information.Specifically, generation module 308 is according to letter
The value for ceasing the amplitude in key element, frequency and/or tone color obtains the colouring information of each dynamic effect to be loaded, position letter
Breath and/or angle information, wherein, colouring information, positional information and/or angle information take according to amplitude, frequency and/or tone color
Value it is different and different.If dynamic effect is light-illuminating effect, generation module 308 is according to the amplitude in information element, frequency
And/or the value of tone color can generate the colouring information of light-illuminating effect, positional information, angle information etc..Generation module 308
During generation, the colouring information of light-illuminating effect can be generated according to the value of amplitude;Or generation module 308 is according to amplitude
Value generates the positional information of light-illuminating effect;Or generation module 308 generates light-illuminating effect according to the value of frequency
Positional information etc..Specifically the light-illuminating effect color information of the generation of amplitude, frequency, the value of tone color and light, position
Information, the corresponding relation of angle information do not limit herein.
Three-dimensional process module 309, suitable for special object is carried out into three dimensional stress processing.
To make the display effect of the dynamic effect of loading more three-dimensional, three-dimensional process module 309 can enter special object
The processing of row three dimensional stress.Special object illustrates by taking human face as an example, when such as dynamic effect being light-illuminating effect, works as light
When being irradiated on the right side of from human face, it should be unable to be arrived in real life on the left of human face by light-illuminating.Three-dimensional process module
309 by after the processing of human face three dimensional stress, it is possible to achieve the display effect that will not be arrived on the left of human face by light-illuminating.But such as
Fruit is handled without three dimensional stress, and human face is the image of two dimension, then can also be arrived on the left of human face by light-illuminating, display effect
Fruit can be untrue.
When special object is carried out three dimensional stress processing by three-dimensional process module 309, three dimensional stress can be carried out by deep learning
Processing.Specifically, human face is carried out three dimensional stress processing by such as three-dimensional process module 309 using deep learning, face is extracted
Key message.The key message can be specially key point information, key area information, and/or key lines information etc..The present invention
Embodiment illustrated by taking key point information as an example, but the present invention key message be not limited to key point information.Use pass
Key point information can improve the processing speed and efficiency that three dimensional stress processing is carried out according to key point information, can be directly according to key
Point information carries out three dimensional stress processing, it is not necessary to carries out the complex operations such as subsequently calculating, analysis to key message again.It is meanwhile crucial
Point information is easy to extract, and extracts accurately, and the effect of progress three dimensional stress processing is more accurate.Three-dimensional process module 309 carries out three-dimensional
When changing processing, the faceform of three-dimensional is first built.It is based on the identity in 3D face databases and expression weight to build threedimensional model
Structure matrix, the set for the key point information of a given face, (coordinate can be declined by coordinate
Descent mode) tries to achieve identity, and Expression Reformation coefficient and rotation scaling translation parameters restrain Euclidean distance, and then build
Go out the three-dimensional structure model of corresponding face.Human face is carried out three dimensional stress by three-dimensional process module 309 using three-dimensional structure model
Processing, obtains the face of three dimensional stress.It should be noted that the special object after three dimensional stress processing does not have texture feature information.
Three-dimensional process module 309 further extracts the image texture information of special object in current frame image, image texture information record
The information such as the spatial color distribution of special object and light distribution in current frame image.Three-dimensional process module 309 is special in extraction
Such as LBP (Local binary patterns) local binary patterns method, ash can be used when determining the image texture information of object
The methods of spending co-occurrence matrix is extracted.Three-dimensional process module 309 is according to the image texture information pair of the special object extracted
Special object after three dimensional stress processing is drawn, and obtains containing the three dimensional stress special object of textural characteristics.
After above-mentioned module is performed, Fusion Module 305 first carries out the background image after foreground image and stylization processing
Fusion treatment, and overall tone processing is carried out, so that the image after fusion is more natural.On this basis, Fusion Module 305
At least one dynamic effect is loaded, is realized and the image after the present frame processing of the input information match of external input sources.Such as
Input information is music, and dynamic effect is the light-illuminating effect of light show radiation response, and background image is the back of the body of discotheque style
A kind of display effect with the personage of music change in discotheque is integrally presented in scape picture, the image after present frame processing.
Uploading module 310, suitable for the video data after processing is uploaded into Cloud Server.
Video data after processing can be directly uploaded to Cloud Server by uploading module 310, specifically, uploading module
310 can be uploaded to the video data after processing the cloud video platform server of one or more, such as iqiyi.com, youku.com, fast
The cloud video platform server such as video, so that cloud video platform server is shown video data in cloud video platform.Or
Video data after processing can also be uploaded to cloud direct broadcast server by uploading module 310, when the user for having live viewing end enters
When entering cloud direct broadcast server and being watched, can by cloud direct broadcast server by video data real time propelling movement to viewing user client
End.Or the video data after processing can also be uploaded to cloud public number server by uploading module 310, it is somebody's turn to do when there is user's concern
During public number, video data is pushed to public number concern client by cloud public number server;Further, cloud public number service
Device can also be accustomed to according to the viewing of the user of concern public number, and the video data that push meets user's custom is paid close attention to public number
Client.
According to video data real-time processing device provided by the invention, according at least one information element of extraction, generation
At least one dynamic effect to be loaded.Background image after foreground image and stylization processing is subjected to fusion treatment, adjustment
Integral color, and the dynamic effect of load information key element generation, make the video after processing integrally present defeated with external input sources
Enter the display effect of information match.Meanwhile to make the display effect of the dynamic effect of loading more three-dimensional, can will be specific right
As carrying out three dimensional stress processing, so that the display effect of the video after processing is closer to truly.The present invention can directly obtain processing
Video afterwards, the video after processing can also be directly uploaded to Cloud Server, it is not necessary to which user carries out volume to the video of recording
Outer processing, user time is saved, the video data after being handled with real-time display to user, facilitate user to check display effect
Fruit.User's technical merit is not limited simultaneously, facilitates public use.
Present invention also provides a kind of nonvolatile computer storage media, the computer-readable storage medium is stored with least
One executable instruction, the computer executable instructions can perform video data in the above-mentioned any means embodiment side of processing in real time
Method.
Fig. 5 shows a kind of structural representation of computing device according to an embodiment of the invention, of the invention specific real
Specific implementation of the example not to computing device is applied to limit.
As shown in figure 5, the computing device can include:Processor (processor) 502, communication interface
(Communications Interface) 504, memory (memory) 506 and communication bus 508.
Wherein:
Processor 502, communication interface 504 and memory 506 complete mutual communication by communication bus 508.
Communication interface 504, for being communicated with the network element of miscellaneous equipment such as client or other servers etc..
Processor 502, for configuration processor 510, it can specifically perform above-mentioned video data real-time processing method embodiment
In correlation step.
Specifically, program 510 can include program code, and the program code includes computer-managed instruction.
Processor 502 is probably central processor CPU, or specific integrated circuit ASIC (Application
Specific Integrated Circuit), or it is arranged to implement the integrated electricity of one or more of the embodiment of the present invention
Road.The one or more processors that computing device includes, can be same type of processor, such as one or more CPU;Also may be used
To be different types of processor, such as one or more CPU and one or more ASIC.
Memory 506, for depositing program 510.Memory 506 may include high-speed RAM memory, it is also possible to also include
Nonvolatile memory (non-volatile memory), for example, at least a magnetic disk storage.
Program 510 specifically can be used for so that processor 502 performs video counts in above-mentioned any means embodiment factually
When processing method.The specific implementation of each step may refer to the phase in the real-time Processing Example of above-mentioned video data in program 510
Corresponding description in step and unit is answered, will not be described here.It is apparent to those skilled in the art that it is description
Convenience and succinct, the equipment of foregoing description and the specific work process of module, may be referred to pair in preceding method embodiment
Process description is answered, will not be repeated here.
Algorithm and display be not inherently related to any certain computer, virtual system or miscellaneous equipment provided herein.
Various general-purpose systems can also be used together with teaching based on this.As described above, required by constructing this kind of system
Structure be obvious.In addition, the present invention is not also directed to any certain programmed language.It should be understood that it can utilize various
Programming language realizes the content of invention described herein, and the description done above to language-specific is to disclose this hair
Bright preferred forms.
In the specification that this place provides, numerous specific details are set forth.It is to be appreciated, however, that the implementation of the present invention
Example can be put into practice in the case of these no details.In some instances, known method, structure is not been shown in detail
And technology, so as not to obscure the understanding of this description.
Similarly, it will be appreciated that in order to simplify the disclosure and help to understand one or more of each inventive aspect,
Above in the description to the exemplary embodiment of the present invention, each feature of the invention is grouped together into single implementation sometimes
In example, figure or descriptions thereof.However, the method for the disclosure should be construed to reflect following intention:I.e. required guarantor
The application claims of shield features more more than the feature being expressly recited in each claim.It is more precisely, such as following
Claims reflect as, inventive aspect is all features less than single embodiment disclosed above.Therefore,
Thus the claims for following embodiment are expressly incorporated in the embodiment, wherein each claim is in itself
Separate embodiments all as the present invention.
Those skilled in the art, which are appreciated that, to be carried out adaptively to the module in the equipment in embodiment
Change and they are arranged in one or more equipment different from the embodiment.Can be the module or list in embodiment
Member or component be combined into a module or unit or component, and can be divided into addition multiple submodule or subelement or
Sub-component.In addition at least some in such feature and/or process or unit exclude each other, it can use any
Combination is disclosed to all features disclosed in this specification (including adjoint claim, summary and accompanying drawing) and so to appoint
Where all processes or unit of method or equipment are combined.Unless expressly stated otherwise, this specification (including adjoint power
Profit requires, summary and accompanying drawing) disclosed in each feature can be by providing the alternative features of identical, equivalent or similar purpose come generation
Replace.
In addition, it will be appreciated by those of skill in the art that although some embodiments described herein include other embodiments
In included some features rather than further feature, but the combination of the feature of different embodiments means in of the invention
Within the scope of and form different embodiments.For example, in the following claims, embodiment claimed is appointed
One of meaning mode can use in any combination.
The all parts embodiment of the present invention can be realized with hardware, or to be run on one or more processor
Software module realize, or realized with combinations thereof.It will be understood by those of skill in the art that it can use in practice
Microprocessor or digital signal processor (DSP) realize device that video data according to embodiments of the present invention is handled in real time
In some or all parts some or all functions.The present invention is also implemented as described herein for performing
The some or all equipment or program of device (for example, computer program and computer program product) of method.So
Realization the present invention program can store on a computer-readable medium, or can have one or more signal shape
Formula.Such signal can be downloaded from internet website and obtained, and either be provided or with any other shape on carrier signal
Formula provides.
It should be noted that the present invention will be described rather than limits the invention for above-described embodiment, and ability
Field technique personnel can design alternative embodiment without departing from the scope of the appended claims.In the claims,
Any reference symbol between bracket should not be configured to limitations on claims.Word "comprising" does not exclude the presence of not
Element or step listed in the claims.Word "a" or "an" before element does not exclude the presence of multiple such
Element.The present invention can be by means of including the hardware of some different elements and being come by means of properly programmed computer real
It is existing.In if the unit claim of equipment for drying is listed, several in these devices can be by same hardware branch
To embody.The use of word first, second, and third does not indicate that any order.These words can be explained and run after fame
Claim.
Claims (10)
1. a kind of video data real-time processing method, it includes:
Real-time image acquisition collecting device is captured and/or the video recorded in include special object current frame image;Or
Person, the current frame image that special object is included in currently played video is obtained in real time;
Scene cut processing is carried out to current frame image, obtains the foreground image that current frame image is directed to the special object;
The input information of external input sources is obtained, at least one information element is extracted from the input information;
Stylized processing is carried out to background image according at least one information element;
Background image after the foreground image and stylization processing is subjected to fusion treatment, obtains the figure after present frame processing
Picture;
Image after present frame is handled covers the video data after former current frame image is handled;
Show the video data after the processing.
2. according to the method for claim 1, wherein, the input information is music;At least one information element bag
Include:Amplitude, frequency and/or tone color.
3. method according to claim 1 or 2, wherein, it is described according at least one information element to background image
Stylization processing is carried out to further comprise:
Value according to the amplitude, frequency and/or tone color chooses the changing pattern that stylized processing is carried out to background image;Its
In, selected changing pattern is different and different according to the value of the amplitude, frequency and/or tone color;
Stylized processing is carried out to background image using the changing pattern.
4. according to the method any one of claim 1-3, wherein, the background image is that the current frame image is entered
Row scene cut handles obtained background image or default background image.
5. according to the method any one of claim 1-4, wherein, by the back of the body after the foreground image and stylization processing
Before scape image carries out fusion treatment, methods described also includes:
The special object is subjected to three dimensional stress processing.
6. according to the method any one of claim 1-5, wherein, the foreground image and stylization are handled described
Before background image afterwards carries out fusion treatment, methods described also includes:
At least one dynamic effect to be loaded is generated according at least one information element.
7. the method according to claim 11, wherein, it is described to be loaded extremely according at least one information element generation
A kind of few dynamic effect further comprises:
Obtained according at least one information element colouring information of each dynamic effect to be loaded, positional information and/
Or angle information;
Each described dynamic effect is generated according to the colouring information, positional information and/or angle information.
8. a kind of video data real-time processing device, it includes:
Acquisition module, suitable for captured by real-time image acquisition collecting device and/or in the video recorded comprising special object
Current frame image;Or the current frame image that special object is included in currently played video is obtained in real time;
Split module, suitable for current frame image carry out scene cut processing, obtain current frame image be directed to it is described specific right
The foreground image of elephant;
Extraction module, suitable for obtaining the input information of external input sources in real time, at least one letter is extracted from the input information
Cease key element;
Stylized module, suitable for carrying out stylized processing to background image according at least one information element;
Fusion Module, suitable for the background image after the foreground image and stylization processing is carried out into fusion treatment, obtain current
Image after frame processing;
Overlay module, the video data after former two field picture is handled is covered suitable for the image after present frame is handled;
Display module, suitable for the video data after the display processing.
9. a kind of computing device, including:Processor, memory, communication interface and communication bus, the processor, the storage
Device and the communication interface complete mutual communication by the communication bus;
The memory is used to deposit an at least executable instruction, and the executable instruction makes the computing device such as right will
Ask and operated corresponding to the video data real-time processing method any one of 1-7.
10. a kind of computer-readable storage medium, an at least executable instruction, the executable instruction are stored with the storage medium
Make operation corresponding to video data real-time processing method of the computing device as any one of claim 1-7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710850208.6A CN107613360A (en) | 2017-09-20 | 2017-09-20 | Video data real-time processing method and device, computing device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710850208.6A CN107613360A (en) | 2017-09-20 | 2017-09-20 | Video data real-time processing method and device, computing device |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107613360A true CN107613360A (en) | 2018-01-19 |
Family
ID=61061452
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710850208.6A Pending CN107613360A (en) | 2017-09-20 | 2017-09-20 | Video data real-time processing method and device, computing device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107613360A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110475150A (en) * | 2019-09-11 | 2019-11-19 | 广州华多网络科技有限公司 | The rendering method and device of virtual present special efficacy, live broadcast system |
CN110493630A (en) * | 2019-09-11 | 2019-11-22 | 广州华多网络科技有限公司 | The treating method and apparatus of virtual present special efficacy, live broadcast system |
CN111869330A (en) * | 2018-02-27 | 2020-10-30 | 昕诺飞控股有限公司 | Rendering dynamic light scenes based on one or more light settings |
CN112752038A (en) * | 2020-12-28 | 2021-05-04 | 广州虎牙科技有限公司 | Background replacing method and device, electronic equipment and computer readable storage medium |
WO2021109876A1 (en) * | 2019-12-02 | 2021-06-10 | Oppo广东移动通信有限公司 | Image processing method, apparatus and device, and storage medium |
CN112995491A (en) * | 2019-12-13 | 2021-06-18 | 阿里巴巴集团控股有限公司 | Video generation method and device, electronic equipment and computer storage medium |
CN115334296A (en) * | 2022-07-01 | 2022-11-11 | 深圳市安博斯技术有限公司 | Stereoscopic image display method and display device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101452582A (en) * | 2008-12-18 | 2009-06-10 | 北京中星微电子有限公司 | Method and device for implementing three-dimensional video specific action |
CN106303555A (en) * | 2016-08-05 | 2017-01-04 | 深圳市豆娱科技有限公司 | A kind of live broadcasting method based on mixed reality, device and system |
CN106803057A (en) * | 2015-11-25 | 2017-06-06 | 腾讯科技(深圳)有限公司 | Image information processing method and device |
CN107071580A (en) * | 2017-03-20 | 2017-08-18 | 北京潘达互娱科技有限公司 | Data processing method and device |
CN107172485A (en) * | 2017-04-25 | 2017-09-15 | 北京百度网讯科技有限公司 | A kind of method and apparatus for being used to generate short-sighted frequency |
-
2017
- 2017-09-20 CN CN201710850208.6A patent/CN107613360A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101452582A (en) * | 2008-12-18 | 2009-06-10 | 北京中星微电子有限公司 | Method and device for implementing three-dimensional video specific action |
CN106803057A (en) * | 2015-11-25 | 2017-06-06 | 腾讯科技(深圳)有限公司 | Image information processing method and device |
CN106303555A (en) * | 2016-08-05 | 2017-01-04 | 深圳市豆娱科技有限公司 | A kind of live broadcasting method based on mixed reality, device and system |
CN107071580A (en) * | 2017-03-20 | 2017-08-18 | 北京潘达互娱科技有限公司 | Data processing method and device |
CN107172485A (en) * | 2017-04-25 | 2017-09-15 | 北京百度网讯科技有限公司 | A kind of method and apparatus for being used to generate short-sighted frequency |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111869330A (en) * | 2018-02-27 | 2020-10-30 | 昕诺飞控股有限公司 | Rendering dynamic light scenes based on one or more light settings |
CN111869330B (en) * | 2018-02-27 | 2024-05-10 | 昕诺飞控股有限公司 | Rendering dynamic light scenes based on one or more light settings |
CN110475150A (en) * | 2019-09-11 | 2019-11-19 | 广州华多网络科技有限公司 | The rendering method and device of virtual present special efficacy, live broadcast system |
CN110493630A (en) * | 2019-09-11 | 2019-11-22 | 广州华多网络科技有限公司 | The treating method and apparatus of virtual present special efficacy, live broadcast system |
CN110493630B (en) * | 2019-09-11 | 2020-12-01 | 广州华多网络科技有限公司 | Processing method and device for special effect of virtual gift and live broadcast system |
WO2021047420A1 (en) * | 2019-09-11 | 2021-03-18 | 广州华多网络科技有限公司 | Virtual gift special effect rendering method and apparatus, and live streaming system |
WO2021109876A1 (en) * | 2019-12-02 | 2021-06-10 | Oppo广东移动通信有限公司 | Image processing method, apparatus and device, and storage medium |
CN112995491A (en) * | 2019-12-13 | 2021-06-18 | 阿里巴巴集团控股有限公司 | Video generation method and device, electronic equipment and computer storage medium |
CN112995491B (en) * | 2019-12-13 | 2022-09-16 | 阿里巴巴集团控股有限公司 | Video generation method and device, electronic equipment and computer storage medium |
CN112752038A (en) * | 2020-12-28 | 2021-05-04 | 广州虎牙科技有限公司 | Background replacing method and device, electronic equipment and computer readable storage medium |
CN112752038B (en) * | 2020-12-28 | 2024-04-19 | 广州虎牙科技有限公司 | Background replacement method, device, electronic equipment and computer readable storage medium |
CN115334296A (en) * | 2022-07-01 | 2022-11-11 | 深圳市安博斯技术有限公司 | Stereoscopic image display method and display device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107613360A (en) | Video data real-time processing method and device, computing device | |
CN107633228A (en) | Video data handling procedure and device, computing device | |
CN107820027A (en) | Video personage dresss up method, apparatus, computing device and computer-readable storage medium | |
CN107483892A (en) | Video data real-time processing method and device, computing device | |
CN107547804A (en) | Realize the video data handling procedure and device, computing device of scene rendering | |
CN107507155A (en) | Video segmentation result edge optimization real-time processing method, device and computing device | |
Gooch et al. | Viewing progress in non-photorealistic rendering through Heinlein's lens | |
Liang et al. | Spatial-separated curve rendering network for efficient and high-resolution image harmonization | |
Zamuda et al. | Vectorized procedural models for animated trees reconstruction using differential evolution | |
CN108109161A (en) | Video data real-time processing method and device based on adaptive threshold fuzziness | |
CN107977927A (en) | Stature method of adjustment and device, computing device based on view data | |
CN108111911A (en) | Video data real-time processing method and device based on the segmentation of adaptive tracing frame | |
CN107945188A (en) | Personage based on scene cut dresss up method and device, computing device | |
CN107665482A (en) | Realize the video data real-time processing method and device, computing device of double exposure | |
CN107743263A (en) | Video data real-time processing method and device, computing device | |
Lopez et al. | Modeling complex unfoliaged trees from a sparse set of images | |
CN107566853A (en) | Realize the video data real-time processing method and device, computing device of scene rendering | |
EP3591618A2 (en) | Method and apparatus for converting 3d scanned objects to avatars | |
CN108124489A (en) | Information processing method and device, cloud processing equipment and computer program product | |
US20160086365A1 (en) | Systems and methods for the conversion of images into personalized animations | |
CN107613161A (en) | Video data handling procedure and device, computing device based on virtual world | |
CN107770606A (en) | Video data distortion processing method, device, computing device and storage medium | |
Delanoy et al. | A Generative Framework for Image‐based Editing of Material Appearance using Perceptual Attributes | |
CN107766803A (en) | Video personage based on scene cut dresss up method, apparatus and computing device | |
CN107633547A (en) | Realize the view data real-time processing method and device, computing device of scene rendering |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180119 |