CN101610421A - Video communication method, Apparatus and system - Google Patents

Video communication method, Apparatus and system Download PDF

Info

Publication number
CN101610421A
CN101610421A CNA2008101270074A CN200810127007A CN101610421A CN 101610421 A CN101610421 A CN 101610421A CN A2008101270074 A CNA2008101270074 A CN A2008101270074A CN 200810127007 A CN200810127007 A CN 200810127007A CN 101610421 A CN101610421 A CN 101610421A
Authority
CN
China
Prior art keywords
content
local
depth value
scene
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CNA2008101270074A
Other languages
Chinese (zh)
Other versions
CN101610421B (en
Inventor
方平
刘琛
刘源
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Device Co Ltd
Huawei Device Shenzhen Co Ltd
Original Assignee
Shenzhen Huawei Communication Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Huawei Communication Technologies Co Ltd filed Critical Shenzhen Huawei Communication Technologies Co Ltd
Priority to CN2008101270074A priority Critical patent/CN101610421B/en
Priority to PL09765408T priority patent/PL2299726T3/en
Priority to ES09765408T priority patent/ES2389401T3/en
Priority to PCT/CN2009/072320 priority patent/WO2009152769A1/en
Priority to EP09765408A priority patent/EP2299726B1/en
Publication of CN101610421A publication Critical patent/CN101610421A/en
Priority to US12/971,392 priority patent/US8446459B2/en
Application granted granted Critical
Publication of CN101610421B publication Critical patent/CN101610421B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

Embodiments of the invention disclose a kind of video communication method, Apparatus and system, relate to the video communication field, have solved now to carry out the problem that special arrangement could strengthen the sense of reality to the scene of communication two party.The embodiment of the invention is at first extracted local object content, secondly local object content and depth value thereof is sent to far-end; Far-end utilizes the relation of depth value, with the synthetic scene of the background of the local object content that receives and far-end, show, make that the local background of distal displayed picture and its is corresponding, in the sense of reality that does not need the scene of communication two party is carried out to strengthen under the situation of special arrangement the user.The embodiment of the invention can be used in the video communication, for example: general Video chat, the visual telephone of office usefulness, video conference etc.

Description

Video communication method, Apparatus and system
Technical field
The present invention relates to the video communication field, particularly a kind of video communication method, and the device and the system that adopt this video communication method.
Background technology
Along with the continuous development of mechanics of communication, the video communication technology has obtained using comparatively widely, and for example: visual telephone, video conference etc. has all used the video communication technology.Present various video communications are used main adopt traditional two dimensional image or video.
Extract for the target of picture material at present, mainly adopt Chroma key (chroma coder) method to cut apart the foreground target that extracts in the video by color.In video communication, thereby with the foreground target and the synthetic sense of reality that increases of other far-end video that extract.For example: the lantern slide lecture notes of people in the video (foreground target) and far-end are synthetic.But there is following defective in Chroma key method:
1, Chroma key method requires blue for adopting, green or other solid color by the background of divided video, could realize cutting apart of foreground target and background like this, and require to occur in the prospect color in the background, because the color of this method background and prospect is strict, use inconvenient.
2, Chroma key method only can be distinguished prospect and background, and scene content can not be divided into more level, can't realize the replacement of prospect middle part partial objectives for, as in conference scenario, may there be desk in the place ahead of people, if the other side's desk is replaced to local desk, then can increase the sense of reality.
3, this technology has only realized the replacement of two-dimensional video content, can't make the user experience the degree of depth of scene, lacks the sense of reality.
Above-mentioned two dimensional image or video can only show the content of scenery, can not reflect distance, the position even depth information of scenery.
Human custom uses two eyes to observe the world, because the existence of binocular parallax makes observed scenery have far and near preferably, location aware, can embody the third dimension of scenery.The three-dimensional video-frequency technology shows slightly variant scene content by the right and left eyes of giving the people, thereby makes the people obtain the depth feelings and the stereovision of scene based on the binocular parallax principle.
In order to increase the sense of reality of video communication, prior art utilization three-dimensional video-frequency technology, and, make the user feel that communication two party is in the Same Scene, to increase the sense of reality to the special decoration of communication two party scene.For example: the indoor environment of communication two party is arranged to the same, in communication process, the user sees the other side in the video like this, just as the other side just be in and own scene in, but this method range of application is subjected to the restriction of both sides' environment layout.
Summary of the invention
Embodiments of the invention provide a kind of video communication method, equipment and system, are not subjected to the restriction of the environment layout of communication two party, increase the sense of reality of communication two party in communication process.
Embodiments of the invention adopt following technical scheme:
A kind of video preprocessor processing method comprises:
Obtain local scene content and depth value thereof;
Depth value according to local scene content is partitioned into local object content from local scene content.
A kind of video preprocessor processing unit comprises:
The information acquisition module is used to obtain local scene content and depth value thereof;
Cut apart module, be used for from local scene content, being partitioned into local object content according to local scene content depth value.
A kind of video receiving method comprises:
Receive object content and depth value thereof that far-end sends;
Obtain the local background content and the depth value of background content;
According to depth value remote target content and local background content are synthesized scene content.
A kind of video receiving apparatus comprises:
The coffret module is used to receive object content and the depth value thereof that far-end sends;
Extraction module is used to obtain local background content and depth value thereof;
Synthesis module is used for according to depth value remote target content and the synthetic scene content of local background content.
A kind of video communications system comprises transmitting terminal and receiving terminal:
Described transmitting terminal, be used to obtain the scene content and the depth value thereof of transmitting terminal, according to the scene content depth value of transmitting terminal, from the scene content of transmitting terminal, be partitioned into the object content of transmitting terminal, and the object content and the depth value thereof of described transmitting terminal sent to receiving terminal;
Described receiving terminal is used for object content and the depth value thereof that receiving end/sending end sends, and obtains the background content and the depth value thereof of receiving terminal, according to the background content synthetic scene content of depth value with the object content and the receiving terminal of transmitting terminal.
By the described embodiment of the invention of technique scheme, in the video communication process, local picture displayed need be synthetic by the object content of local background content and far-end, make the user see that background and own present located scene in the picture are identical, similarly be that the residing environment of communication two party is to increase the sense of reality in the user communication process.And owing to do not need the scene of communication two party is carried out special arrangement in the present embodiment, allow the residing environment difference of communication two party, do not need to change background into solid color yet, so when implementing to be the embodiment of the invention, the communication two party environmental limit can be subjected to, the sense of reality in the communication process can be increased.
Description of drawings
Fig. 1 is the flow chart of first embodiment of the invention video preprocessor processing method;
Fig. 2 is the block diagram of first embodiment of the invention video preprocessor processing unit;
Fig. 3 is the flow chart of first embodiment of the invention video receiving method;
Fig. 4 is the block diagram of first embodiment of the invention video receiving apparatus;
Fig. 5 is the schematic diagram of first embodiment of the invention video communication equipment;
Fig. 6 is the schematic diagram of second embodiment of the invention video communication equipment
Fig. 7 is the schematic diagram of the stereo camera that adopts in the second embodiment of the invention;
Fig. 8 is the schematic diagram of synthetic scene content process in the second embodiment of the invention;
Fig. 9 is a third embodiment of the invention video communication flow chart;
Figure 10 is the structure chart of third embodiment of the invention video communications system.
Embodiment
The embodiment of the invention shows the background content of this locality and synthetic picture of object content of far-end, make that communication two party does not need scene is carried out special arrangement, can allow the scene in the picture identical, increase the sense of reality in the communication process with self scene of living in.Embodiment to video communication method of the present invention, device and equipment is described in detail below in conjunction with accompanying drawing.
Embodiment 1:
Present embodiment provides a kind of video preprocessor processing method, and as shown in Figure 1, this video preprocessor processing method comprises the steps:
101, obtain local scene content and depth value thereof by depth camera or stereo camera.
102, the depth value by local scene content can be divided into many levels with local scene content, so just the described level of local object content can be split, and promptly is partitioned into local object content from local scene content.
103, the local object content that will be partitioned into, and the depth value of local object content correspondence sends to far-end, generally need send to the opposite end of communication.
Mainly finish preliminary treatment to image by step 101 and step 102 in the present embodiment, step 103 is the steps that the preliminary treatment content is sent, and can omit.
Corresponding to above-mentioned video preprocessor processing method, present embodiment also provides a kind of video preprocessor processing unit, and as shown in Figure 2, this video preprocessor processing unit comprises: information acquisition module 21, cut apart module 22 and sending module 23.
Wherein, information acquisition module 21 is used to obtain local scene content and depth value thereof, described information acquisition module can be realized by depth camera or stereo camera, depth camera wherein adopts infrared technique to obtain the degree of depth of image, and three-dimensional video camera adopts dual camera to obtain the degree of depth of image.After obtaining the depth value of local scene content, just local scene content can be divided into many levels, cut apart module 22, be used for from local scene content, being partitioned into local object content according to local scene content depth value.Sending module 23 is used for described local object content and depth value thereof are sent to far-end.
Mainly by information acquisition module 21 with cut apart module 22 and finish video preprocessor and handle, sending module 23 wherein can omit in this video preprocessor processing unit.
In order to finish video communication, present embodiment also provides a kind of and above-mentioned video preprocessor processing method corresponding video receiving method, and as shown in Figure 3, this video receiving method comprises the steps:
301, receive object content and the depth value thereof that far-end sends.
302, obtain the local background content and the depth value of background content.
303, according to the difference of depth value, determine the hiding relation of local background content and remote target content, be generally the little pixel of depth value and block the big pixel of depth value, can remote target content and local background content be synthesized scene content according to the relation of depth value like this.
Corresponding to above-mentioned video receiving method, present embodiment also provides a kind of video receiving apparatus, and as shown in Figure 4, this video receiving apparatus comprises: coffret module 41, extraction module 42 and synthesis module 43.
Wherein, coffret module 41 is used to receive object content and the depth value thereof that far-end sends; Extraction module 42 is used to obtain local background content and depth value thereof; Synthesis module 43 is used for according to the relation of depth value remote target content and local background content being synthesized scene content, is that the little pixel of depth value is blocked the big pixel of depth value generally speaking; Scene content after synthesizing by the demonstration of equipment such as display at last.
As shown in Figure 5, the embodiment of the invention also provides a kind of video communication equipment, specifically comprises: information acquisition module 51, cut apart module 52, coffret module 53, extraction module 54 and synthesis module 55.
Wherein, information acquisition module 51 is used to obtain local scene content and depth value thereof, described information acquisition module 51 can be realized by depth camera or stereo camera, depth camera wherein adopts infrared technique to obtain the degree of depth of image, and three-dimensional video camera adopts dual camera to obtain the degree of depth of image.Cut apart module 52, be used for from local scene content, being partitioned into local object content according to local scene content depth value.Coffret module 53 is used for described local object content and depth value thereof are sent to far-end.
Described coffret module 53 also is used to receive object content and the depth value thereof that far-end sends, and extraction module 54 is used to obtain local background content and depth value thereof; Synthesis module 55 is used for according to the relation of depth value remote target content and local background content being synthesized scene content, is that the little pixel of depth value is blocked the big pixel of depth value generally speaking; Scene content after synthesizing by the display module demonstration at last.
Wherein Ben Di background content can also can be obtained the background content and the depth value thereof on local target opposite by another video camera for cutting apart the residue content after module 54 is partitioned into local object content.
Carry out communication if allow between video preprocessor processing unit in the present embodiment and the video receiving apparatus, for example all be linked into same network, so just enough become a video communications system, the transmitting terminal of this system comprises the video preprocessor processing unit of Fig. 2, and receiving terminal comprises the video receiving apparatus among Fig. 4.
Embodiment 2:
Present embodiment provides a kind of video communication equipment, this equipment is with the local object content in the local scene content, and the depth value of local object content correspondence sends to opposite equip., opposite equip. is after receiving local object content, the object content of described this locality and the background of opposite end are synthesized a width of cloth scene, and be shown to the user of opposite end.The scene that the user saw and self the residing scene that can guarantee the opposite end like this are just the same, relatively have the telepresenc and the sense of reality.Local video communication equipment with remote target content and the synthetic width of cloth scene of local background content, and is shown to local user, to improve telepresenc and the sense of reality of local user in communication process after receiving the object content of far-end.
As shown in Figure 6, this video communication equipment mainly comprises: information acquisition module 61, cut apart module 62, coding module 63, coffret module 64, decoder module 65, synthesis module 66 and display module 67.
Wherein, information acquisition module 61 is used to realize the shooting to local scene content, and the calculating of the corresponding depth value of local scene content, perhaps directly obtains the corresponding depth value of local scene content; Cut apart module 62, be used for being partitioned into local object content from local scene content according to depth value; Coding module 63 is used for the local object content and the corresponding depth value thereof that are partitioned into are encoded; Coffret module 64 is used for sending local object content and depth value thereof, perhaps receives object content and depth value thereof that far-end sends; Decoder module 65, the remote target content received and the decoding of depth value thereof are used to achieve a butt joint; Synthesis module 66, the remote target content and the local background content that are used for decoding is obtained merge, depth value according to correspondence generates three-dimensional view, local background content wherein can be the residue content that is partitioned in the local scene content behind the local object content, also can be the scene content that adopts the local target opposite that another group video camera takes; Display module 67, be used to realize to the demonstration that becomes image, can be stereoscopic display device or ordinary two dimensional display device, if stereoscopic display device, then need the two dimensional image of reconstruct one another viewpoint of the width of cloth.
Respectively each module in the present embodiment video communication equipment is described in detail below.
Information acquisition module 61 can have following two kinds of implementations: one, adopt depth camera to obtain local scene content and depth value thereof simultaneously; Two, adopt multiple cameras to take local scene content, obtain corresponding depth value by the stereo-picture matching process.
Depth camera (Depth Camera) is a kind of novel video camera, and depth camera can be obtained the depth value of each pixel correspondence in the coloured image when taking the RGB coloured image.Present depth camera mainly adopts the infrared mode depth value.
Obtain the method for corresponding depth value by the stereo-picture matching process, two or two above video camera photographed scenes are adopted in requirement when IMAQ, obtain the multiple image of scene different angles, by image is mated, can obtain the parallax of scene on different images, according to the inside and outside parameter of parallax and video camera, can calculate the depth value of each pixel correspondence in the image.To be that example is obtained depth value to the images match mode and described below with two video cameras.
Be illustrated in figure 7 as two parallel vidicon imaging schematic diagrames of horizontal positioned, wherein O1 and O2 are respectively two video camera photocentres, its distance is B, and some A is Z (promptly putting the degree of depth of A) to the distance of the vertical point O of video camera, and A1 and A2 put the imaging point of A at two video cameras respectively.
By triangle A1 O1 O1 ' and similar the getting of triangle A O1 C: A 1 O 1 ′ CO 1 = f Z ;
By triangle A2 O2 O1 ' and similar the getting of triangle A O2 C: A 2 O 2 ′ CO 2 = f Z ;
So release the parallax of two imaging points be: d=A1O1 '-A2O2 '=f* (CO1-CO2)/Z=f*B/Z.
So, can obtain the depth value Z=f*B/d of an A.
Because f is known, B can measure, and d can calculate by the method for images match, so, adopt two video cameras can get access to the corresponding depth value of each point in the scene.
Obtaining depth information by stereo camera comprises: find certain some imaging point corresponding in multiple image in the scene, and then obtain its depth value according to this some coordinate in multiple image.Find in the scene certain a bit in different images process of corresponding imaging point finish by images match.Present image matching technology mainly comprises: based on the coupling of window, based on the coupling of feature and dynamic programming etc.
Wherein, all adopted matching algorithm based on the coupling of window and dynamic programming based on gray scale.Algorithm based on gray scale is that one of them image segmentation is become a plurality of little subregions, in other image, find and its subregion of similar grey value profile as masterplate with its gray value, if two sub regions satisfy the similitude requirement of grey value profile, we can think that the point in the subregion mates, and promptly the imaging point of this two sub regions is the imaging of same point in the scene.In matching process, use correlation function to weigh the similitude in two zones usually.
Directly do not utilize the gray scale of image based on the coupling of feature, but utilize the feature that derives by gradation of image information to mate, compare and utilize simple brightness and grey scale change information more stable.Matching characteristic can be thought potential can describe scene 3D structure key character, as the intersection point (angle point) at edge and edge.Coupling based on feature generally obtains sparse depth information figure earlier, utilizes methods such as interpolate value to obtain the intensive depth information figure of image then.
Cut apart module 62 according to local scene content and corresponding depth value thereof, image is cut apart the local object content that obtains in the local scene.Cutting apart module 62 can be by searching unit 621 and cutting unit 622 is realized, search unit 621, be used for searching local object content in zone that local scene content occurs, cutting unit 622, edge contour extracts to be used for carrying out accurately in local scene content to the zone of local object content, cuts apart to obtain local object content and other local background content.
In general, the zone that local object content occurs in local scene, after can estimating the position of the relative video camera of local target by the local user, set the depth value scope that local object content occurs, in follow-up Video processing, in this depth value scope, search the zone that object content occurs by searching the unit.
The local object content of searching if desired is a figure image, can adopt existing face recognition technology so, from local scene content, automatically identify the position that facial image occurs by face identification unit 623, in the depth value of local scene content, search the depth value of described facial image position correspondence by searching unit 621 then, determine the scope of local object content depth value then according to the depth value that finds, and determine the zone of local object content in scene content according to the scope of described depth value.Thereby determine the depth bounds that personage's target occurs in scene.
Because depth value is fit to and coloured image is corresponding, and is corresponding with the people's object area from coloured image according to people's object area that depth value is partitioned into.To be sent to coding module 63 behind the local object content of the coloured image that obtains and the depth value thereof, coding module 63 sends to far-end to its coding back by coffret module 64.
Because in different size from the local object content that extracts, these local object contents need be adjusted to same size, generally be that these local object contents are adjusted to the size the same with local scene content, thereby each frame is obtained the image to be encoded of identical size, be convenient to coding.This adjustment can not carried out convergent-divergent to local object content itself, has just changed the size of painting canvas that local object content uses.For adjusting the white space that the size back occurs, can adopt 0 value to fill.
63 pairs of coding modules in the present embodiment split local object content and depth value thereof encode.Compare single pass two-dimensional video, three-dimensional video-frequency has much bigger data volume: binocular tri-dimensional video has two data passages.The increase of video data has all brought difficulty to its storage and transmission.Stereo scopic video coding mainly also can be divided into two classes at present: block-based coding and object-based coding.Stereo-picture coding in, except the data redundancy that infra-frame prediction and inter prediction are eliminated on spatial domain and the time domain is outside one's consideration, also must eliminate the spatial domain data redundancy between the multichannel image.Parallax (Parallax) estimation and compensation are key technologies in the stereo scopic video coding, are used to eliminate the spatial domain redundancy between multichannel image.The core of disparity estimation compensation is the correlation that finds between two width of cloth (or several) image.Stereo scopic video coding content herein comprises coloured image and corresponding depth value thereof, can adopt hierarchical coding, is about to the coloured image hybrid coding and puts into basic layer, puts into enhancement layer behind the depth value hybrid coding.
Coffret module 64 in the present embodiment is used for sending coding local object content in back and depth value thereof, and remote target content and depth value thereof behind the coding of reception far-end transmission, delivers to decoder module and carries out decoding processing.Coffret module 64 in the present embodiment can be to realize the various wired or wireless interface that transmits, for example: broadband interface, blue tooth interface, infrared interface or adopt the access technology of the mobile radio communication of mobile phone.The coffret module only need be transmitted wherein local target and depth value thereof in the present embodiment, and for original local scene content, its data volume reduces to some extent, the bandwidth usage in the time of can reducing transfer of data
After the coffret module 64 of present embodiment video communication equipment receives remote target content and depth value thereof, need handle and to show.
Decoder module 65 is used for the remote data that receives is decoded, and obtains the depth value of the object content correspondence of far-end.
Synthesis module 66, be used for remote target content and the local background content that decoding obtains being merged according to depth value, coloured image after remote target content that obtains synthesizing and local background merge, and corresponding depth value, wherein local background content is finished by extraction module 69.To determine hiding relation according to the depth value of remote target content and the depth value of local background content earlier in the building-up process, then according to the synthetic corresponding color image content of hiding relation.When display module 67 is the 3 D stereo display device, need further according to the virtual image of combined color image content with another viewpoint of depth value reconstruct of correspondence, so can also comprise view restructuring module 68 in the present embodiment, be used for the picture material after synthetic is carried out view restructuring, generate a virtual visual point image, this virtual visual point image and combined color image promptly constitute three-dimensional view, send to the 3 D stereo display device and realize stereo display.
As shown in Figure 8, provided the remote target content (personage) that receives, and illustrate the degree of depth of this remote target content, and the local local background content (tree and desk) that adopts the depth camera mode to obtain, and illustrate the degree of depth of this this locality background content, basis depth relationship wherein synthesizes the scene that obtains synthesizing then.Owing to obtained the distance of remote target content, the far-end personage can be inserted between local desk and the tree with the relative video camera of local background content.
In order to allow synthetic image more true to nature, need address the problem:
(1) the convergent-divergent problem of remote target content.Far-end personage and local background content are perfect to be merged in order to make, and may need to adjust by unit for scaling 661 position of the relative video camera of remote target content, at this moment needs simultaneously the remote target content size to be carried out convergent-divergent.When need the remote target content move to nearer apart from the time, when promptly reducing depth value, need amplify the remote target content; When the remote target content arrangement farther apart from the time, when promptly increasing depth value, need dwindle its remote target content.Because the remote target content is a single target, being limited in scope of its change in depth can be reduced to the convergent-divergent of perspective relation the linear scale consistent with its degree of depth when carrying out image zoom.
(2) the mutual occlusion issue between remote target content and the local background content.To remote target content and the fusion of local background content the time, need to consider its mutual occlusion issue by synthesis unit 662.Hiding relation can be determined that when the horizontal and vertical position of pixel overlapped, the pixel that depth value is little blocked the big point (close shot blocks distant view) of depth value by depth value.
(3) empty filling problem.May there be the cavity in the local background content that obtains after having removed local object content, still may have the cavity after itself and remote target content are merged.At this two kinds of settling modes are arranged:
First kind of scene content for using another group camera acquisition to take local target opposite, general is the scene content that the people saw, when synthetic, adopt this scene content directly and the remote target content synthesize, this mode effect is better, and promptly background that the people saw and far-end personage are merged, owing to directly use the scene on opposite, there is not the hole-filling problem, but need increases by one group of video camera at each end of video communication.
Another kind of solution is rejected the local background content of being left behind the local object content for using, and for the cavity that may occur, the method that adopts edge pixel to fill is filled.
When present embodiment video communication equipment adopts the 3 D stereo display device, and display device needs another width of cloth image of reconstruct, thereby realizes stereo display when only supporting that left and right sides image input mode shows.Some automatic stereoscopic display device supports a width of cloth Two-dimensional Color Image and corresponding depth value thereof to carry out the 3 D stereo demonstration, so just do not needed another width of cloth image of reconstruct, but finish the reconstruct of another width of cloth image by automatic stereoscopic display device self, and in restructuring procedure, finish corresponding cavity and fill, as the three-dimensional display of philips.
It is synthetic that view restructuring is also referred to as virtual visual point image, refers generally to from the image at other visual angle of image reconstruction of model or different angles.Present embodiment realizes by view restructuring module 68, when the degree of depth of known image, can calculate parallax between virtual view and the known view according to following formula:
d=A1O1′-A2O2′=f*(CO1-CO2)/Z=f*B/Z。
Wherein, d is the parallax between virtual view view and the known view, and f is a focus of camera, and B is the distance between virtual view and the former camera point, and Z is the degree of depth of image.
When based on the image on its right of composograph and degree of depth reconstruct thereof, certain bar scan line x in the right image rPlace's color of pixel is by corresponding scan line x in the left image (composograph) lPlace's color of pixel is definite, wherein x lCoordinate determine by following formula:
x l = x r + d = x r + fB Z
When determining to synthesize the view content according to above formula, owing to existing occlusion issue to cause some point among the right figure can't in left figure, find corresponding point, promptly have empty problem, adopt the pixel at empty edge that it is filled equally, filling can adopt the bilinear interpolation mode to carry out.
Display module in the present embodiment is used for showing the image to after synthetic.This display module 67 can be that stereoscopic display device comprises auto-stereoscopic display device, and anaglyph spectacles and the demonstration of holographic display device 3 D stereo etc. realize the stereo display of stereo-picture, can let user experiencing the degree of depth of scene, experience stereoeffect.When needs carry out stereo display, generally need finish above-mentioned view restructuring and cavity filling.The present embodiment display module also can be the ordinary two dimensional display device, only shows two-dimentional composograph, when only needing to show two dimensional image, does not then need to carry out view restructuring, directly shows the two dimensional image after synthesizing.
Embodiment 3:
Present embodiment is a communication process example in the video communication system, be specially two users (A and B) and carry out communication by the video communication equipment among the embodiment 2, user A sends video data to user B in its communication process, and user B receives the overall process of the video data of user A, the structure of this video communications system as shown in figure 10, comprise transmitting terminal and receiving terminal, transmitting terminal is connected by network with receiving terminal.
Described transmitting terminal, be used to obtain the scene content and the depth value thereof of transmitting terminal, according to the scene content depth value of transmitting terminal, from the scene content of transmitting terminal, be partitioned into the object content of transmitting terminal, and the object content and the depth value thereof of described transmitting terminal sent to receiving terminal; Described transmitting terminal comprises: information acquisition module 1001, be used to realize shooting to local scene content, and the calculating of the corresponding depth value of local scene content, perhaps directly obtain the corresponding depth value of local scene content; Cut apart module 1002, be used for being partitioned into local object content from local scene content according to depth value; Coding module 1003 is used for the local object content and the corresponding depth value thereof that are partitioned into are encoded; Coffret module 1004 is used for local object content and depth value thereof are sent to receiving terminal.
Described receiving terminal is used for object content and the depth value thereof that receiving end/sending end sends, and obtains the background content and the depth value thereof of receiving terminal, according to the background content synthetic scene content of depth value with the object content and the receiving terminal of transmitting terminal.Described receiving terminal comprises coffret module 1005, is used for receiving object content and the depth value thereof that far-end sends; Decoder module 1006, the remote target content received and the decoding of depth value thereof are used to achieve a butt joint; Synthesis module 1007, the remote target content and the local background content that are used for decoding is obtained merge, depth value according to correspondence generates three-dimensional view, local background content wherein can be the residue content that is partitioned in the local scene content behind the local object content, extracts this residue content by extraction module 1010; Local background content also can be to adopt the scene content on the local target opposite of another group video camera shooting; Display module 1009, be used to realize to the demonstration that becomes image, can be stereoscopic display device or ordinary two dimensional display device, if stereoscopic display device, then need the two dimensional image of reconstruct one another viewpoint of the width of cloth.Another viewpoint two dimensional image of reconstruct can be finished by view restructuring module 1008.
Its communication process specifically comprises the steps: as shown in Figure 9
901, the information acquisition module of the video communication equipment of user A obtains local scene content and depth value thereof; Can obtain the depth value of local scene content and scene content by depth camera (depth camera) or stereo camera.Depth camera can directly be obtained the degree of depth by infrared ray; And three-dimensional video camera generally obtains scene content by two parallel video cameras, calculates the depth value of each pixel in this scene content then, and computing formula is: Z=fB/ Δ x; Wherein f is a focal length, and B is the distance of two video cameras, and Δ x is the position difference of each pixel in two video cameras.
902, the module of cutting apart of the video communication equipment of user A is partitioned into local object content from local scene content, be specially: the local scene content that photographs is carried out the position that recognition of face obtains facial image by the face identification unit of cutting apart in the module, then by described facial image position correspondence is searched in the unit in the depth value of local scene content the depth value of cutting apart in the module of searching, and the scope of personage's depth value in the picture of determining to photograph according to the depth value that finds.So just can determine the zone of local object content in scene content, from local scene content, be partitioned into personage's target by the cutting unit of cutting apart in the module according to the zone of determining at last.
903, cut apart draw local personage's target after, can preserve the residue content after described local scene content is partitioned into local personage's target, and the residue content depth value; Also can obtain the background content and the depth value thereof on personage's target opposite simultaneously, and preserve by another video camera.
904, will be for unified local personage's target size, local personage's target need be expanded to former is the size of gathering picture, perhaps is cut into the picture of other sizes; Can be filled to 0 value owing to cut out the hole region of back generation.
905, respectively resultant local personage's target and depth value thereof in the step 904 are encoded, preferably use hierarchical coding, adopt hierarchical coding to need data quantity transmitted less.
906, described local personage's target and the depth value thereof after will encoding sends to the video communication equipment of user B by the coffret module.
Above step has been finished the transmit operation of user A, and following steps receive data for user B and to the processing procedure of data.
907, the video communication equipment of user B receives personage's target and depth value thereof that user A sends by the coffret module.
908, the video communication equipment of user B obtains personage's target and the depth value thereof of user A by the data decode of decoder module to receiving.The video communication equipment of user B also needs to obtain the depth value of background content and background content simultaneously, generally speaking, and can be with the residue content after the local target of removal in the local scene content as its background content.If obtain the background content and the depth value thereof on user B opposite by another video camera, the picture that user B is seen is truer, and can not produce empty problem when composograph.
909, personage's target and the depth value thereof that user A is sended over by unit for scaling in the synthesis module carries out convergent-divergent, obtain comparatively personage's target of desirable amount, when need the remote target content move to nearer apart from the time, when promptly reducing depth value, need amplify the remote target content; When the remote target content arrangement farther apart from the time, when promptly increasing depth value, need dwindle its remote target content.
Then according to the depth value behind personage's target convergent-divergent of user A and the depth value of background content, determine the hiding relation of far-end personage target and local background content, blocking principle is: when the horizontal and vertical position of pixel overlapped, the pixel that depth value is little blocked the big point (close shot blocks distant view) of depth value.
Synthesis unit in the synthesis module synthesizes a width of cloth scene content according to above-mentioned definite hiding relation with personage's target and background content again.
If background content is the residue content behind the removal object content, needs that then pixel is carried out in the cavity in the synthetic scene content and fill; If background content is directly to obtain the scene on user B opposite, then need not carries out pixel and fill.
910, that described synthetic scene content is carried out virtual visual point image is synthetic for the view restructuring module, is specially according to following formula and calculates parallax between virtual view and the known view:
d=A1O1′-A2O2′=f*(CO1-CO2)/Z=f*B/Z。
Wherein, d is the parallax between virtual view view and the known view, and f is a focus of camera, and B is the distance between virtual view and the former camera point, and Z is the degree of depth of image.
When based on the image on its right of composograph and degree of depth reconstruct thereof, certain bar scan line x in the right image rPlace's color of pixel is by corresponding scan line x in the left image (composograph) lPlace's color of pixel is definite, wherein x lCoordinate determine by following formula:
x l = x r + d = x r + fB Z
After finishing view restructuring, the cavity in the scene content after need synthesizing virtual visual point image is carried out pixel and is filled.
911, show scene content after synthetic by display module, for example: by auto-stereoscopic display device, anaglyph spectacles or the demonstration of holographic display device 3 D stereo etc., realize the stereo display of stereo-picture, perhaps only show two-dimentional composograph by the ordinary two dimensional display device.
In the video communications system of present embodiment, the equipment of user A can also comprise video receiving apparatus, and the equipment of user B can also comprise the video preprocessor processing unit, can send video data to user A to guarantee user B.If user B need send video data to user A, its process is the same with Fig. 9, and just transmit leg and recipient have changed.The embodiment of the invention mainly is used in the video communication, for example: general Video chat, the visual telephone of office usefulness, video conference etc.
The above; only be the specific embodiment of the present invention, but protection scope of the present invention is not limited thereto, anyly is familiar with those skilled in the art in the technical scope that the present invention discloses; can expect easily changing or replacing, all should be encompassed within protection scope of the present invention.Therefore, protection scope of the present invention should be as the criterion with the protection range of claim.

Claims (27)

1, a kind of video preprocessor processing method is characterized in that comprising:
Obtain local scene content and depth value thereof;
Depth value according to local scene content is partitioned into local object content from local scene content.
2, video preprocessor processing method according to claim 1 is characterized in that, also comprises:
Preserve the residue content after described local scene content is partitioned into local object content, and the depth value of residue content, the background content and the depth value thereof on local target opposite perhaps preserved.
3, video preprocessor processing method according to claim 1 is characterized in that, describedly obtains local scene content and depth value comprises:
Obtain the depth value of local scene content and scene content by depth camera; Perhaps
Obtain two width of cloth images of scene content by two parallel video cameras;
Described two width of cloth images are mated, calculate the parallax of each pixel correspondence in two parallel vidicons;
Calculate the depth value of each pixel in the scene content then by following formula: Z=fB/ Δ x, wherein, Z is the depth value of this pixel, and f is a focus of camera, and B is the distance of two video cameras, and Δ x is the parallax of each pixel correspondence in two parallel vidicons.
4, video preprocessor processing method according to claim 1 is characterized in that, described depth value according to local scene content is partitioned into local object content and comprises from local scene content:
Determine the scope of local object content depth value;
Determine the zone of local object content in scene content according to the scope of described depth value;
From local scene content, be partitioned into local object content according to described zone.
5, video preprocessor processing method according to claim 4 is characterized in that, if described local object content is the personage, described method also comprises:
Local scene content is carried out the position that recognition of face obtains facial image;
Search the depth value of described facial image position correspondence at the depth value of local scene content;
The scope of described definite local object content depth value comprises: the scope of determining local object content depth value according to the depth value that finds.
6, video preprocessor processing method according to claim 1 is characterized in that also comprising:
Respectively described local object content and depth value thereof are encoded.
7, a kind of video preprocessor processing unit is characterized in that comprising:
The information acquisition module is used to obtain local scene content and depth value thereof;
Cut apart module, be used for from local scene content, being partitioned into local object content according to local scene content depth value.
8, video preprocessor processing unit according to claim 7 is characterized in that also comprising:
Memory module is used to preserve the residue content after described local scene content is partitioned into local object content, and the depth value of residue content, perhaps preserves the background content and the depth value thereof on local target opposite.
9, video preprocessor processing unit according to claim 7 is characterized in that, the described module of cutting apart comprises:
Search the unit, be used for determining the scope of local object content depth value, and determine the zone of local object content in scene content according to the scope of described depth value;
Cutting unit is used for being partitioned into local object content according to described zone from local scene content.
10, video preprocessor processing unit according to claim 9 is characterized in that, if described local object content is the personage, the described module of cutting apart also comprises:
Face identification unit is used for local scene content is carried out the position that recognition of face obtains facial image;
Describedly search described facial image position correspondence is searched in the unit in the depth value of local scene content depth value, determine the scope of local object content depth value then according to the depth value that finds, and determine the zone of local object content in scene content according to the scope of described depth value.
11, video preprocessor processing unit according to claim 7 is characterized in that also comprising:
Coding module is used for described local object content and depth value thereof are encoded.
12, a kind of video receiving method is characterized in that comprising:
Receive object content and depth value thereof that far-end sends;
Obtain the local background content and the depth value of background content;
According to depth value remote target content and local background content are synthesized scene content.
13, video receiving method according to claim 12 is characterized in that, described background content is: the residue content in the local scene content after the local target of removal, the background content on perhaps local target opposite.
14, video receiving method according to claim 12 is characterized in that, according to depth value remote target content and the synthetic scene content of local background content is comprised:
According to the depth value of remote target content and the depth value of local background content, determine the hiding relation of remote target content and local background content;
According to above-mentioned hiding relation remote target content and local background content are synthesized scene content.
15, video receiving method according to claim 14 is characterized in that, according to depth value remote target content and the synthetic scene content of local background content is also comprised: remote target content and depth value thereof are carried out convergent-divergent.
16, video receiving method according to claim 12 is characterized in that also comprising:
Described synthetic scene content is carried out virtual visual point image synthesizes.
17, video receiving method according to claim 12 is characterized in that also comprising:
Object content and depth value thereof to the far-end that receives are decoded.
18, a kind of video receiving apparatus is characterized in that comprising:
The coffret module is used to receive object content and the depth value thereof that far-end sends;
Extraction module is used to obtain local background content and depth value thereof;
Synthesis module is used for according to depth value remote target content and the synthetic scene content of local background content.
19, video receiving apparatus according to claim 18 is characterized in that, described synthesis module comprises:
Synthesis unit, be used for according to the depth value of remote target content and the depth value of local background content, determine the hiding relation of remote target content and local background content, and remote target content and local background content are synthesized scene content according to above-mentioned hiding relation.
20, video receiving apparatus according to claim 19 is characterized in that also comprising:
Unit for scaling is used for remote target content and depth value thereof are carried out convergent-divergent;
Described synthesis unit after according to the remote target content scaling depth value and the depth value of local background content, determine the hiding relation of remote target content and local background content.
21, video receiving apparatus according to claim 18 is characterized in that also comprising:
The view restructuring module is used for that described synthetic scene content is carried out virtual visual point image and synthesizes.
22, video receiving apparatus according to claim 18 is characterized in that, described background content is: the residue content in the local scene content after the local target of removal, the background content on perhaps local target opposite.
23, video receiving apparatus according to claim 18 is characterized in that also comprising:
The information acquisition module is used to obtain the local scene content and the depth value of scene content;
Cut apart module, be used for being partitioned into local object content from scene content according to the depth value of local scene content;
Described coffret module also is used for local object content and depth value thereof are sent to far-end.
24, want 23 described video receiving apparatus according to right, it is characterized in that, the background content of described this locality is: the residue content in the local scene content after the local target of removal, the background content on perhaps local target opposite.
25, video receiving apparatus according to claim 23 is characterized in that, if described local object content is the personage, the described module of cutting apart comprises:
Face identification unit is used for local scene content is carried out the position that recognition of face obtains facial image;
Search the unit, be used for determining the scope of personage's depth value, and determine the zone of personage in local scene content according to the scope of described depth value according to the position of facial image;
Cutting unit is used for being partitioned into the personage according to described zone from local scene content.
26, video receiving apparatus according to claim 18 is characterized in that also comprising:
Decoder module is used for the object content and the depth value thereof of the far-end that receives are decoded.
27, a kind of video communications system comprises transmitting terminal and receiving terminal, it is characterized in that,
Described transmitting terminal, be used to obtain the scene content and the depth value thereof of transmitting terminal, according to the scene content depth value of transmitting terminal, from the scene content of transmitting terminal, be partitioned into the object content of transmitting terminal, and the object content and the depth value thereof of described transmitting terminal sent to receiving terminal;
Described receiving terminal is used for object content and the depth value thereof that receiving end/sending end sends, and obtains the background content and the depth value thereof of receiving terminal, according to the background content synthetic scene content of depth value with the object content and the receiving terminal of transmitting terminal.
CN2008101270074A 2008-06-17 2008-06-17 Video communication method, video communication device and video communication system Active CN101610421B (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
CN2008101270074A CN101610421B (en) 2008-06-17 2008-06-17 Video communication method, video communication device and video communication system
PL09765408T PL2299726T3 (en) 2008-06-17 2009-06-17 Video communication method, apparatus and system
ES09765408T ES2389401T3 (en) 2008-06-17 2009-06-17 Method, device and communication system through video
PCT/CN2009/072320 WO2009152769A1 (en) 2008-06-17 2009-06-17 Video communication method, apparatus and system
EP09765408A EP2299726B1 (en) 2008-06-17 2009-06-17 Video communication method, apparatus and system
US12/971,392 US8446459B2 (en) 2008-06-17 2010-12-17 Video communication method, device, and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2008101270074A CN101610421B (en) 2008-06-17 2008-06-17 Video communication method, video communication device and video communication system

Publications (2)

Publication Number Publication Date
CN101610421A true CN101610421A (en) 2009-12-23
CN101610421B CN101610421B (en) 2011-12-21

Family

ID=41483960

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2008101270074A Active CN101610421B (en) 2008-06-17 2008-06-17 Video communication method, video communication device and video communication system

Country Status (1)

Country Link
CN (1) CN101610421B (en)

Cited By (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102118625A (en) * 2009-12-30 2011-07-06 晨星软件研发(深圳)有限公司 Image processing device with screen display function and image processing method
CN102223555A (en) * 2010-04-19 2011-10-19 Lg电子株式会社 Image display apparatus and method for controlling the same
CN102368816A (en) * 2011-12-01 2012-03-07 中科芯集成电路股份有限公司 Intelligent front end system of video conference
CN102376077A (en) * 2010-08-20 2012-03-14 英华达(上海)科技有限公司 Palm electronic device capable of being combined with images and image combining method thereof
CN102572464A (en) * 2010-10-15 2012-07-11 卡西欧计算机株式会社 Image composition apparatus and image retrieval method
CN102779359A (en) * 2012-07-13 2012-11-14 南京大学 Automatic ticket checking device for performing passage detection based on depth image
CN103168466A (en) * 2010-09-20 2013-06-19 高通股份有限公司 Virtual video capture device
CN103220540A (en) * 2012-01-21 2013-07-24 瑞昱半导体股份有限公司 Device and method of image processing
CN103279315A (en) * 2013-04-24 2013-09-04 电子科技大学 Real-time desktop remote sharing method
WO2013155984A1 (en) * 2012-04-19 2013-10-24 Huawei Technologies Co., Ltd. Using depth information to assist motion compensation-based video coding
CN103379352A (en) * 2012-04-20 2013-10-30 Nlt科技股份有限公司 Image generation apparatus, image display apparatus and image generation method
CN103795961A (en) * 2012-10-30 2014-05-14 三亚中兴软件有限责任公司 Video conference telepresence system and image processing method thereof
CN103959762A (en) * 2011-11-30 2014-07-30 诺基亚公司 Quality enhancement in multimedia capturing
CN104243966A (en) * 2013-06-18 2014-12-24 西斯维尔科技有限公司 method and device for generating, storing, transmitting, receiving and reproducing depth maps
CN104349111A (en) * 2013-07-24 2015-02-11 华为技术有限公司 Meeting place creating method and system of video conference
CN104836977A (en) * 2014-02-10 2015-08-12 阿里巴巴集团控股有限公司 Method and system for video communication in instant communication process
CN104869346A (en) * 2014-02-26 2015-08-26 中国移动通信集团公司 Method and electronic equipment for processing image in video call
CN104935822A (en) * 2015-06-15 2015-09-23 努比亚技术有限公司 Method and device for processing images
CN104954689A (en) * 2015-06-30 2015-09-30 努比亚技术有限公司 Method and shooting device for acquiring photo through double cameras
CN105187724A (en) * 2015-09-17 2015-12-23 努比亚技术有限公司 Mobile terminal and method for processing images
CN105303543A (en) * 2015-10-23 2016-02-03 努比亚技术有限公司 Image enhancement method and mobile terminal
CN105472309A (en) * 2015-12-31 2016-04-06 杭州华为数字技术有限公司 Data transmission method, device and system
CN105577517A (en) * 2015-12-17 2016-05-11 掌赢信息科技(上海)有限公司 Sending method of short video message and electronic device
CN105611373A (en) * 2015-12-23 2016-05-25 小米科技有限责任公司 Video picture processing method and device
CN105872448A (en) * 2016-05-31 2016-08-17 宇龙计算机通信科技(深圳)有限公司 Display method and device of video images in video calls
CN106303690A (en) * 2015-05-27 2017-01-04 腾讯科技(深圳)有限公司 A kind of method for processing video frequency and device
CN106446883A (en) * 2016-08-30 2017-02-22 西安小光子网络科技有限公司 Scene reconstruction method based on light label
CN106961621A (en) * 2011-12-29 2017-07-18 英特尔公司 Use the communication of incarnation
CN107438161A (en) * 2017-07-31 2017-12-05 广东欧珀移动通信有限公司 Shooting picture processing method, device and terminal
CN107529096A (en) * 2017-09-11 2017-12-29 广东欧珀移动通信有限公司 Image processing method and device
CN107533761A (en) * 2015-04-27 2018-01-02 索尼半导体解决方案公司 Image processing apparatus and image processing system
CN107682656A (en) * 2017-09-11 2018-02-09 广东欧珀移动通信有限公司 Background image processing method, electronic equipment and computer-readable recording medium
CN107734283A (en) * 2017-09-11 2018-02-23 广东欧珀移动通信有限公司 Picture processing method, device and the storage medium of Video chat
CN108076307A (en) * 2018-01-26 2018-05-25 南京华捷艾米软件科技有限公司 Video conferencing system based on AR and the video-meeting method based on AR
WO2018095317A1 (en) * 2016-11-28 2018-05-31 中兴通讯股份有限公司 Data processing method, device, and apparatus
CN108156369A (en) * 2017-12-06 2018-06-12 广东欧珀移动通信有限公司 Image processing method and device
CN108259810A (en) * 2018-03-29 2018-07-06 上海掌门科技有限公司 A kind of method of video calling, equipment and computer storage media
CN108933913A (en) * 2017-05-24 2018-12-04 中兴通讯股份有限公司 A kind of video meeting implementing method, device, system and computer storage medium
CN109120912A (en) * 2018-09-30 2019-01-01 Oppo广东移动通信有限公司 A kind of data processing method, MEC server, terminal device and device
CN109151430A (en) * 2018-09-30 2019-01-04 Oppo广东移动通信有限公司 A kind of data processing method, terminal, server and computer storage medium
CN109246409A (en) * 2018-09-30 2019-01-18 Oppo广东移动通信有限公司 A kind of data processing method, terminal, server and computer storage medium
CN109246408A (en) * 2018-09-30 2019-01-18 Oppo广东移动通信有限公司 A kind of data processing method, terminal, server and computer storage medium
CN109286804A (en) * 2018-09-30 2019-01-29 Oppo广东移动通信有限公司 A kind of data processing method, MEC server, terminal device and device
CN110049378A (en) * 2019-04-17 2019-07-23 珠海格力电器股份有限公司 Interactive approach, control system and terminal under a kind of video mode
CN110430364A (en) * 2019-08-26 2019-11-08 武汉众果科技有限公司 A kind of stingy figure local dynamic station figure synthetic method of taking pictures of photographing
CN110574076A (en) * 2017-12-04 2019-12-13 佳能株式会社 Generation device, generation method, and program
CN111294582A (en) * 2018-12-06 2020-06-16 宏达国际电子股份有限公司 Three-dimensional image processing method, photographing device and non-transitory computer readable storage medium
CN111372062A (en) * 2020-05-02 2020-07-03 北京花兰德科技咨询服务有限公司 Artificial intelligence image communication system and recording method
CN111669662A (en) * 2020-07-03 2020-09-15 海信视像科技股份有限公司 Display device, video call method and server
CN112423142A (en) * 2020-09-30 2021-02-26 北京字跳网络技术有限公司 Image processing method, image processing device, electronic equipment and computer readable medium
WO2022001635A1 (en) * 2020-07-03 2022-01-06 海信视像科技股份有限公司 Display device and display method
CN115486088A (en) * 2021-03-30 2022-12-16 京东方科技集团股份有限公司 Information interaction method, computer readable storage medium and communication terminal
CN117640877A (en) * 2024-01-24 2024-03-01 浙江华创视讯科技有限公司 Picture reconstruction method for online conference and electronic equipment

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10021366B2 (en) * 2014-05-02 2018-07-10 Eys3D Microelectronics, Co. Image process apparatus
US9569830B2 (en) * 2015-07-03 2017-02-14 Mediatek Inc. Image processing method and electronic apparatus with image processing mechanism
CN105141834A (en) * 2015-07-27 2015-12-09 努比亚技术有限公司 Device and method for controlling picture shooting

Cited By (82)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102118625B (en) * 2009-12-30 2013-07-24 晨星软件研发(深圳)有限公司 Image processing device with screen display function and image processing method
CN102118625A (en) * 2009-12-30 2011-07-06 晨星软件研发(深圳)有限公司 Image processing device with screen display function and image processing method
CN102223555A (en) * 2010-04-19 2011-10-19 Lg电子株式会社 Image display apparatus and method for controlling the same
CN102223555B (en) * 2010-04-19 2015-03-18 Lg电子株式会社 Image display apparatus and method for controlling the same
CN102376077A (en) * 2010-08-20 2012-03-14 英华达(上海)科技有限公司 Palm electronic device capable of being combined with images and image combining method thereof
CN103168466A (en) * 2010-09-20 2013-06-19 高通股份有限公司 Virtual video capture device
US9280847B2 (en) 2010-10-15 2016-03-08 Casio Computer Co., Ltd. Image composition apparatus, image retrieval method, and storage medium storing program
CN102572464A (en) * 2010-10-15 2012-07-11 卡西欧计算机株式会社 Image composition apparatus and image retrieval method
CN103959762B (en) * 2011-11-30 2017-10-27 诺基亚技术有限公司 Method and apparatus for the increased quality in multimedia capture
CN103959762A (en) * 2011-11-30 2014-07-30 诺基亚公司 Quality enhancement in multimedia capturing
CN102368816A (en) * 2011-12-01 2012-03-07 中科芯集成电路股份有限公司 Intelligent front end system of video conference
CN106961621A (en) * 2011-12-29 2017-07-18 英特尔公司 Use the communication of incarnation
CN103220540A (en) * 2012-01-21 2013-07-24 瑞昱半导体股份有限公司 Device and method of image processing
CN103220540B (en) * 2012-01-21 2017-04-12 瑞昱半导体股份有限公司 Device and method of image processing
WO2013155984A1 (en) * 2012-04-19 2013-10-24 Huawei Technologies Co., Ltd. Using depth information to assist motion compensation-based video coding
KR20140147123A (en) * 2012-04-19 2014-12-29 후아웨이 테크놀러지 컴퍼니 리미티드 Using depth information to assist motion compensation-based video coding
KR101629746B1 (en) * 2012-04-19 2016-06-13 후아웨이 테크놀러지 컴퍼니 리미티드 Using depth information to assist motion compensation-based video coding
US9584806B2 (en) 2012-04-19 2017-02-28 Futurewei Technologies, Inc. Using depth information to assist motion compensation-based video coding
CN103379352B (en) * 2012-04-20 2016-08-10 Nlt科技股份有限公司 Video generation device, image display device and image generating method
CN103379352A (en) * 2012-04-20 2013-10-30 Nlt科技股份有限公司 Image generation apparatus, image display apparatus and image generation method
CN102779359B (en) * 2012-07-13 2015-07-15 南京大学 Automatic ticket checking device for performing passage detection based on depth image
CN102779359A (en) * 2012-07-13 2012-11-14 南京大学 Automatic ticket checking device for performing passage detection based on depth image
CN103795961A (en) * 2012-10-30 2014-05-14 三亚中兴软件有限责任公司 Video conference telepresence system and image processing method thereof
CN103279315A (en) * 2013-04-24 2013-09-04 电子科技大学 Real-time desktop remote sharing method
CN104243966A (en) * 2013-06-18 2014-12-24 西斯维尔科技有限公司 method and device for generating, storing, transmitting, receiving and reproducing depth maps
CN104243966B (en) * 2013-06-18 2019-10-18 西斯维尔科技有限公司 Method and apparatus for generating, storing, transmit, receive and reproducing depth map
CN104349111A (en) * 2013-07-24 2015-02-11 华为技术有限公司 Meeting place creating method and system of video conference
CN104836977B (en) * 2014-02-10 2018-04-24 阿里巴巴集团控股有限公司 Video communication method and system during instant messaging
CN104836977A (en) * 2014-02-10 2015-08-12 阿里巴巴集团控股有限公司 Method and system for video communication in instant communication process
CN104869346A (en) * 2014-02-26 2015-08-26 中国移动通信集团公司 Method and electronic equipment for processing image in video call
US11089289B2 (en) 2015-04-27 2021-08-10 Sony Semiconductor Solutions Corporation Image processing device, imaging device, image processing method and program
CN107533761A (en) * 2015-04-27 2018-01-02 索尼半导体解决方案公司 Image processing apparatus and image processing system
CN106303690A (en) * 2015-05-27 2017-01-04 腾讯科技(深圳)有限公司 A kind of method for processing video frequency and device
CN104935822A (en) * 2015-06-15 2015-09-23 努比亚技术有限公司 Method and device for processing images
CN104954689A (en) * 2015-06-30 2015-09-30 努比亚技术有限公司 Method and shooting device for acquiring photo through double cameras
CN104954689B (en) * 2015-06-30 2018-06-26 努比亚技术有限公司 A kind of method and filming apparatus that photo is obtained using dual camera
CN105187724A (en) * 2015-09-17 2015-12-23 努比亚技术有限公司 Mobile terminal and method for processing images
CN105187724B (en) * 2015-09-17 2019-07-19 努比亚技术有限公司 A kind of mobile terminal and method handling image
WO2017067526A1 (en) * 2015-10-23 2017-04-27 努比亚技术有限公司 Image enhancement method and mobile terminal
CN105303543A (en) * 2015-10-23 2016-02-03 努比亚技术有限公司 Image enhancement method and mobile terminal
CN105577517A (en) * 2015-12-17 2016-05-11 掌赢信息科技(上海)有限公司 Sending method of short video message and electronic device
CN105611373A (en) * 2015-12-23 2016-05-25 小米科技有限责任公司 Video picture processing method and device
CN105611373B (en) * 2015-12-23 2019-03-08 小米科技有限责任公司 Method of video image processing and device
CN105472309A (en) * 2015-12-31 2016-04-06 杭州华为数字技术有限公司 Data transmission method, device and system
WO2017206456A1 (en) * 2016-05-31 2017-12-07 宇龙计算机通信科技(深圳)有限公司 Method and apparatus for presenting video image in video call
CN105872448A (en) * 2016-05-31 2016-08-17 宇龙计算机通信科技(深圳)有限公司 Display method and device of video images in video calls
CN106446883B (en) * 2016-08-30 2019-06-18 西安小光子网络科技有限公司 Scene reconstruction method based on optical label
CN106446883A (en) * 2016-08-30 2017-02-22 西安小光子网络科技有限公司 Scene reconstruction method based on light label
WO2018095317A1 (en) * 2016-11-28 2018-05-31 中兴通讯股份有限公司 Data processing method, device, and apparatus
CN108377355A (en) * 2016-11-28 2018-08-07 中兴通讯股份有限公司 A kind of video data handling procedure, device and equipment
CN108933913A (en) * 2017-05-24 2018-12-04 中兴通讯股份有限公司 A kind of video meeting implementing method, device, system and computer storage medium
CN107438161A (en) * 2017-07-31 2017-12-05 广东欧珀移动通信有限公司 Shooting picture processing method, device and terminal
CN107682656B (en) * 2017-09-11 2020-07-24 Oppo广东移动通信有限公司 Background image processing method, electronic device, and computer-readable storage medium
CN107529096A (en) * 2017-09-11 2017-12-29 广东欧珀移动通信有限公司 Image processing method and device
CN107682656A (en) * 2017-09-11 2018-02-09 广东欧珀移动通信有限公司 Background image processing method, electronic equipment and computer-readable recording medium
CN107734283A (en) * 2017-09-11 2018-02-23 广东欧珀移动通信有限公司 Picture processing method, device and the storage medium of Video chat
US11012679B2 (en) 2017-12-04 2021-05-18 Canon Kabushiki Kaisha Generating apparatus, generating method, and storage medium
CN110574076B (en) * 2017-12-04 2020-11-06 佳能株式会社 Image generation device, image generation method, and computer-readable storage medium
CN110574076A (en) * 2017-12-04 2019-12-13 佳能株式会社 Generation device, generation method, and program
CN108156369A (en) * 2017-12-06 2018-06-12 广东欧珀移动通信有限公司 Image processing method and device
CN108156369B (en) * 2017-12-06 2020-03-13 Oppo广东移动通信有限公司 Image processing method and device
CN108076307A (en) * 2018-01-26 2018-05-25 南京华捷艾米软件科技有限公司 Video conferencing system based on AR and the video-meeting method based on AR
CN108076307B (en) * 2018-01-26 2021-01-05 南京华捷艾米软件科技有限公司 AR-based video conference system and AR-based video conference method
CN108259810A (en) * 2018-03-29 2018-07-06 上海掌门科技有限公司 A kind of method of video calling, equipment and computer storage media
CN109246409A (en) * 2018-09-30 2019-01-18 Oppo广东移动通信有限公司 A kind of data processing method, terminal, server and computer storage medium
CN109286804A (en) * 2018-09-30 2019-01-29 Oppo广东移动通信有限公司 A kind of data processing method, MEC server, terminal device and device
CN109246408A (en) * 2018-09-30 2019-01-18 Oppo广东移动通信有限公司 A kind of data processing method, terminal, server and computer storage medium
CN109151430A (en) * 2018-09-30 2019-01-04 Oppo广东移动通信有限公司 A kind of data processing method, terminal, server and computer storage medium
CN109120912A (en) * 2018-09-30 2019-01-01 Oppo广东移动通信有限公司 A kind of data processing method, MEC server, terminal device and device
US11089279B2 (en) 2018-12-06 2021-08-10 Htc Corporation 3D image processing method, camera device, and non-transitory computer readable storage medium
CN111294582A (en) * 2018-12-06 2020-06-16 宏达国际电子股份有限公司 Three-dimensional image processing method, photographing device and non-transitory computer readable storage medium
CN111294582B (en) * 2018-12-06 2022-04-15 宏达国际电子股份有限公司 Three-dimensional image processing method, photographing device and non-transitory computer readable storage medium
CN110049378A (en) * 2019-04-17 2019-07-23 珠海格力电器股份有限公司 Interactive approach, control system and terminal under a kind of video mode
CN110430364A (en) * 2019-08-26 2019-11-08 武汉众果科技有限公司 A kind of stingy figure local dynamic station figure synthetic method of taking pictures of photographing
CN111372062A (en) * 2020-05-02 2020-07-03 北京花兰德科技咨询服务有限公司 Artificial intelligence image communication system and recording method
CN111372062B (en) * 2020-05-02 2021-04-20 北京花兰德科技咨询服务有限公司 Artificial intelligence image communication system and recording method
CN111669662A (en) * 2020-07-03 2020-09-15 海信视像科技股份有限公司 Display device, video call method and server
WO2022001635A1 (en) * 2020-07-03 2022-01-06 海信视像科技股份有限公司 Display device and display method
CN112423142A (en) * 2020-09-30 2021-02-26 北京字跳网络技术有限公司 Image processing method, image processing device, electronic equipment and computer readable medium
CN112423142B (en) * 2020-09-30 2024-04-05 北京字跳网络技术有限公司 Image processing method, device, electronic equipment and computer readable medium
CN115486088A (en) * 2021-03-30 2022-12-16 京东方科技集团股份有限公司 Information interaction method, computer readable storage medium and communication terminal
CN117640877A (en) * 2024-01-24 2024-03-01 浙江华创视讯科技有限公司 Picture reconstruction method for online conference and electronic equipment

Also Published As

Publication number Publication date
CN101610421B (en) 2011-12-21

Similar Documents

Publication Publication Date Title
CN101610421B (en) Video communication method, video communication device and video communication system
US8446459B2 (en) Video communication method, device, and system
CN101651841B (en) Method, system and equipment for realizing stereo video communication
CN101416520B (en) Efficient encoding of multiple views
KR101492876B1 (en) 3d video control system to adjust 3d video rendering based on user prefernces
US8736659B2 (en) Method, apparatus, and system for 3D video communication
JP5763184B2 (en) Calculation of parallax for 3D images
CN101472190B (en) Multi-visual angle filming and image processing apparatus and system
US20120139906A1 (en) Hybrid reality for 3d human-machine interface
Po et al. Automatic 2D-to-3D video conversion technique based on depth-from-motion and color segmentation
CN101459857A (en) Communication terminal and information system
You et al. Internet of Things (IoT) for seamless virtual reality space: Challenges and perspectives
KR20170055930A (en) Method and apparatus to display stereoscopic image in 3d display system
Tseng et al. Multiviewpoint video coding with MPEG-2 compatibility
Ziegler et al. Evolution of stereoscopic and three-dimensional video
CN104284193A (en) Encoding device and encoding method, and decoding device and decoding method
CN101527762A (en) Digital photo frame and three-dimensional image display method for same
Chung et al. Bit allocation algorithm with novel view synthesis distortion model for multiview video plus depth coding
CN103748872A (en) Receiver-side adjustment of stereoscopic images
Knorr et al. From 2D-to stereo-to multi-view video
Redert et al. 3-D scene reconstruction with viewpoint adaptation on stereo displays
CN102780894B (en) A kind of decoding method of 3D rendering
Ding et al. Depth map pre-processing algorithm for compression based on 3D-HEVC scheme
US20230115563A1 (en) Method for a telepresence system
Aflaki et al. Unpaired multiview video plus depth compression

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CP01 Change in the name or title of a patent holder
CP01 Change in the name or title of a patent holder

Address after: 518129 Building 2, B District, Bantian HUAWEI base, Longgang District, Shenzhen, Guangdong.

Patentee after: Huawei terminal (Shenzhen) Co.,Ltd.

Address before: 518129 Building 2, B District, Bantian HUAWEI base, Longgang District, Shenzhen, Guangdong.

Patentee before: HUAWEI DEVICE Co.,Ltd.

TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20181217

Address after: 523808 Southern Factory Building (Phase I) Project B2 Production Plant-5, New Town Avenue, Songshan Lake High-tech Industrial Development Zone, Dongguan City, Guangdong Province

Patentee after: HUAWEI DEVICE Co.,Ltd.

Address before: 518129 Building 2, B District, Bantian HUAWEI base, Longgang District, Shenzhen, Guangdong.

Patentee before: Huawei terminal (Shenzhen) Co.,Ltd.