CN111371962A - Video image transmission method and transmission system - Google Patents

Video image transmission method and transmission system Download PDF

Info

Publication number
CN111371962A
CN111371962A CN202010310932.1A CN202010310932A CN111371962A CN 111371962 A CN111371962 A CN 111371962A CN 202010310932 A CN202010310932 A CN 202010310932A CN 111371962 A CN111371962 A CN 111371962A
Authority
CN
China
Prior art keywords
image
video sequence
frame
foreground
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010310932.1A
Other languages
Chinese (zh)
Other versions
CN111371962B (en
Inventor
苑贵全
骞一凡
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou XINGFEIDA Electronic Technology Co.,Ltd.
Original Assignee
Beijing Yingmaiqi Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Yingmaiqi Technology Co ltd filed Critical Beijing Yingmaiqi Technology Co ltd
Priority to CN202010310932.1A priority Critical patent/CN111371962B/en
Publication of CN111371962A publication Critical patent/CN111371962A/en
Application granted granted Critical
Publication of CN111371962B publication Critical patent/CN111371962B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/00095Systems or arrangements for the transmission of the picture signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

The application relates to the technical field of image processing, in particular to a video image transmission method and a video image transmission system.

Description

Video image transmission method and transmission system
Technical Field
The present application relates to the field of image processing technologies, and in particular, to a video image transmission method and a video image transmission system.
Background
In recent years, with the development of photographing apparatuses, the quality of photographed images and videos has been greatly improved, so the data amount of photographed images and videos has also sharply increased. In the prior art, each frame of a video image is completely transmitted in the transmission of the video, so that the transmission of the video with high quality and high data volume in the prior art undoubtedly causes great pressure on the existing network, and the phenomenon of blocking easily occurs in the transmission process.
Therefore, how to transmit video data with high data volume and avoid the phenomenon of stuttering in the transmission process is a technical problem that needs to be solved urgently by those skilled in the art at present.
Disclosure of Invention
The application provides a video image transmission method and a video image transmission system, which are used for transmitting video data with high data volume and avoiding the phenomenon of blocking in the transmission process.
In order to solve the technical problem, the application provides the following technical scheme:
a video image transmission method, comprising the steps of: respectively converting a first frame video sequence to an Nth frame video sequence into a first gray image to an Nth gray image according to the first N frames of video sequences of the collected video images; calculating information entropy of pixel points x in the first gray level image to the Nth gray level image; defining an image corresponding to a region with the information entropy larger than a preset threshold value as a foreground image; according to the area occupied by the foreground image in the first gray image, segmenting the foreground and the background in the first frame video sequence to form a first background model; transmitting a first frame sequence of the video image and the established first background model so as to display the transmitted first frame video sequence as a first frame of the video; foreground segmentation is carried out on an area, changed relative to the first frame video sequence, in the second frame video sequence, and a segmented foreground image is transmitted; and updating the background missing part of the transmitted foreground image on the first background model to update the first background model to the second background model, performing foreground covering on the foreground image on the basis of the second background model, and displaying the image subjected to foreground covering as a second frame of the video.
In the video image transmission method as described above, preferably, the three colors R, G, B of each pixel in each frame of the video sequence are set to have the same value, and a grayscale image is generated by performing a graying process according to the following formula R ═ G ═ wr + wg × G + wb ═ B, where wr, wg, and wb are weighted values R, G, B, respectively.
The video image transmission method as described above, preferably, the first frame video sequence is used as a static background, the gray value of the pixel in the second frame video sequence is subtracted from the gray value of the pixel in the first frame video sequence to obtain a region formed by the pixel with a changed gray value, and the region is used as a foreground image and is segmented from the second frame video sequence.
In the video image transmission method, it is preferable that a part of the segmented foreground images is an area occupied by an object which actually moves or has a changed shape in the second frame video sequence, and another part of the segmented foreground images is an area occupied by the object which actually moves or has a changed shape in the previous frame video sequence.
The video image transmission method as described above, wherein preferably, the area of the object with changed motion or shape in the foreground image in the first frame video sequence is used for updating the missing background part on the basis of the first background model; and the area of the object with the actual motion or the shape change in the foreground image in the second frame video sequence is used for performing foreground coverage on the basis of the second background model.
A video image transmission system comprising: a first terminal and a second terminal, the first terminal comprising: collection module, segmentation module and first transmission module, the second terminal includes: the device comprises a synthesis module, a display module and a second transmission module; the acquisition module acquires each frame of video sequence of a video image, the segmentation module converts the first frame video sequence to the Nth frame video sequence into a first gray image to an Nth gray image respectively according to the first N frames of video sequences of the acquired video image, calculates information entropy for pixel points x in the first gray image to the Nth gray image, defines an image corresponding to a region of which the information entropy is greater than a preset threshold value as a foreground image, and performs foreground and background segmentation in the first frame video sequence according to a region occupied by the foreground image in the first gray image to form a first background model; the first transmission module transmits a first frame sequence of the video image and the established first background model to the second transmission module; the second transmission module receives the first frame video sequence and the established first background model for transmission, and the display module displays the first frame video sequence as a first frame of the video; the segmentation module performs foreground segmentation on a region, which changes relative to the first frame video sequence, in the second frame video sequence, and the first transmission module transmits a segmented foreground image to the second transmission module; the second transmission module receives the foreground image, the synthesis module updates the background missing part of the transmitted foreground image on the first background model so as to update the first background model to the second background model, foreground covering is carried out on the foreground image on the basis of the second background model, and the display module displays the image subjected to foreground covering as a second frame of the video.
In the video image transmission system as described above, preferably, the segmentation module sets the same value for the three colors R, G, B of each pixel in each frame of the video sequence, and performs graying processing according to the following formula R ═ G ═ wr + wg × (G + wb ×) to generate a grayscale image, where wr, wg, and wb are weighted values of R, G, B, respectively.
In the video image transmission system as described above, preferably, the segmentation module takes the first frame video sequence as a static and unchangeable background, subtracts the gray value of the pixel in the second frame video sequence from the gray value of the pixel in the first frame video sequence to obtain an area formed by the pixel with a changed gray value, and segments the area from the second frame video sequence as a foreground image.
In the video image transmission system as described above, it is preferable that a part of the segmented foreground images is an area occupied by an object which actually moves or has a changed shape in the second frame video sequence, and another part of the segmented foreground images is an area occupied by the object which has moved or has a changed shape in the previous frame video sequence.
The video image transmission system as described above, wherein preferably, the synthesis module uses an area occupied by an object with a changed motion or shape in the foreground image in the first frame video sequence to update the missing background part on the basis of the first background model; and the synthesis module is used for covering the foreground of the object with actual motion or changed appearance in the foreground image in the second frame video sequence on the basis of the second background model.
Compared with the background art, the video image transmission method and the video image transmission system provided by the application have the advantages that the transmission of the video sequence after the first frame video sequence only transmits the segmented foreground image, so that the video transmission method can reduce the transmission quantity of the video data with high data quantity, and the phenomenon of blocking during the transmission process is avoided.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments described in the present invention, and other drawings can be obtained by those skilled in the art according to the drawings.
Fig. 1 is a flowchart of a video image transmission method provided in an embodiment of the present application;
FIG. 2 is a flowchart of a first background model established according to a first N-frame video sequence according to an embodiment of the present application;
FIG. 3 is a flow chart of a first background model established according to a first N-frame video sequence according to another embodiment provided by the present application;
FIG. 4 is a schematic diagram of a video sequence of a next frame according to an embodiment of the present application;
fig. 5 is a schematic diagram of a video image transmission system according to an embodiment of the present application.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are illustrative only and should not be construed as limiting the invention.
The video image transmission system comprises at least two terminal devices, and video images are transmitted between the two terminal devices. For example: the video image transmission is carried out between the two mobile phones, between the mobile phones and the computer, between the camera and the computer, and the like.
The video image transmission method provided by the present application is described below by taking the video image transmission between a first terminal and a second terminal as an example, and the terms "first", "second", and the like in the present application are only used for distinguishing different steps, devices, and the like, and do not represent any specific technical meaning nor necessarily indicate a logical order therebetween.
Example one
Referring to fig. 1, fig. 1 is a flowchart of a video image transmission method according to an embodiment of the present application.
The application provides a video image transmission method, which comprises the following steps:
step S110, the first terminal establishes a first background model according to the collected video image, and transmits a first frame sequence of the video image and the established first background model to the second terminal;
the first terminal acquires each frame of the video image, and establishes a first background model according to the previous N frames of the video sequence, referring to fig. 2, which specifically includes the following sub-steps:
step S111, respectively converting the video sequences of the first frame to the Nth frame into a first gray image to an Nth gray image;
the three colors R, G, B of each pixel in each frame of video sequence are set to be the same, so as to perform graying processing on each frame of video sequence, specifically, graying processing is performed through the following formula R ═ G ═ wr ═ R + wg × (G + wb) × (B), where wr, wg, and wb are weights of R, G, B, respectively.
Step S112, calculating the absolute gradient value of each pixel point in the first gray level image to the Nth gray level image in sequence, and taking the pixel point of which the absolute gradient value is greater than a threshold value in each gray level image as an edge point of the gray level image;
the example of finding the edge point of the first gray image corresponding to the first frame of video sequence is described, where the first gray image corresponding to the first frame of video sequence is a two-dimensional function f (x, y), where x and y are coordinate values of a pixel in the first gray image, an absolute gradient value M (x, y) ═ f (x, y) -f (x +1, y +1) | + | f (x +1, y) -f (x, y) +1|, and if the absolute gradient value M (x, y) of the pixel is greater than a preset threshold, the pixel belongs to the edge point of the first gray image corresponding to the first frame of video sequence.
And sequentially solving an absolute gradient value M (x, y) for each pixel point in the second gray scale image to the Nth gray scale image corresponding to the second frame video sequence to the Nth frame video sequence according to the formula, and respectively taking the pixel points of which the absolute gradient values M (x, y) are greater than a preset threshold value as edge points of the corresponding gray scale images.
Step S113, determining whether the image surrounded by the edge points is an image corresponding to an object which is moving or has a changed shape according to the position and shape information of the image surrounded by the edge points in the N gray level images;
the edge points of each gray level image surround to form one or more images, the positions f (x, y) of the same or corresponding edge points of the images are changed in M gray level images, wherein M is more than or equal to 2 and less than or equal to N, the object corresponding to the image is considered to be moving or the shape of the object is changed, and otherwise, the object corresponding to the image is considered to be not moving and the shape of the object is not changed.
Step S114, if the image surrounded by the edge points is the image corresponding to the moving or shape-changing object, defining the image as a foreground image;
in the application, an image corresponding to an object which is moving or has a changed shape (such as a flying bird, a walking person, a running car and a twitch) is defined as a foreground image, and an image corresponding to a relatively static object is defined as a background image to be reused subsequently.
Step S115, segmenting the foreground and the background in the first frame video sequence according to the edge point of the foreground image in the first gray image to form a first background model.
In addition, a first background model is established according to the first N frames of video sequence of the video image acquired by the first terminal, please refer to fig. 3, which may specifically include the following sub-steps:
step S111', respectively converting the video sequences of the first frame to the Nth frame into a first gray image to an Nth gray image;
the three colors R, G, B of each pixel point in each frame of video sequence are set to be the same, so as to perform graying processing on each frame of video sequence, specifically, graying processing is performed through the following formula R ═ G ═ wr ═ R + wg × G + wb ═ B, so as to generate a grayscale image, where wr, wg, and wb are weight values of R, G, B, respectively.
Step S112', calculating information entropy H (x) of pixel points x in the first gray level image to the Nth gray level image;
specifically, let a (x) { a ═ a0,a1,...,aN-1B (x) represents a set of all pixel points in a 3 × 3 neighborhood centered on the point x, and the information entropy of the point x is
Figure BDA0002457816370000071
Wherein p (mu) is the gray probability density of all pixel points in B (x), M is 3 × 3 × N, and represents the total number of all sample point sets in B (x) taking x as the center in the previous N frames of video sequences;
Figure BDA0002457816370000072
m(ai) Is pixel point a in A (x)iCorresponding feature class, i is an integer from 0 to N-1; delta [ m (a) ]i)-μ]Is a Dikela function, which is used to determine whether the color value of the pixel is equal to the color value of the μ -th pixel in the set B (x), which is equal to 1, but not equal to 0.
Step S113', defining an image corresponding to an area with the information entropy H (x) larger than a preset threshold value as a foreground image;
if the information entropy h (x) of a certain area is greater than a preset threshold value, and the difference between pixels of the area is large, the object corresponding to the image of the area is in a dynamic state, and the image corresponding to the object in the dynamic state is defined as a foreground image in the application; otherwise, the information entropy h (x) of the area is smaller than the preset threshold, and the difference between the pixels of the area is small, so that the object corresponding to the image of the area is in a static state, and the image corresponding to the object in the static state is defined as a background image.
Step S114', segmenting foreground and background in the first frame video sequence according to the region occupied by the foreground image in the first gray scale image to form a first background model.
Continuing to refer to fig. 1, in step S120, the second terminal displays the first frame video sequence as a first frame of the video, and uses the received first background model as an initial background model to update the initial background model according to a video image received next;
the second terminal receives a first frame video sequence which is sent by the first terminal and is not divided and a divided first background model which is sent by the first terminal, the first frame video sequence is used as a first frame of a video to be displayed on a display of the second terminal, and the first background model is cached in the second terminal to be used as an initial background model for subsequent repeated utilization.
Step S130, the first terminal carries out foreground segmentation on a region which is changed relative to the first frame video sequence in the second frame video sequence, and transmits a segmented foreground image to the second terminal;
the first terminal takes a first frame video sequence as a static and unchangeable background, wherein pixel points in the first frame video sequence are represented as b (x, y), pixel points in a collected j frame video sequence are represented as g (x, y, j), then pixel points in a second frame video sequence are g (x, y, 2), the gray value of the pixel points in the second frame video sequence is subtracted from the pixel points in the first frame video sequence, namely id (x, y, 2) is g (x, y, 2) -b (x, y), a region formed by the pixel points with the changed gray value is obtained, and the region is taken as a foreground image and is segmented from the second frame video sequence.
In the segmented foreground image, as shown in fig. 4, a is a segmented foreground image, which is an image change condition corresponding to a position a1 of an object (e.g., a sphere) from a previous time to a position a2 of the object at a subsequent time, that is, a part of the segmented foreground image is an area occupied by the object which actually moves or has a changed appearance in the second frame video sequence, that is, a2, and the part is sent to the second terminal for covering the foreground on the basis of the first background model. And the other part of the segmented foreground image is the area occupied by the object with the changed motion or shape in the previous frame of video sequence, namely the area occupied by the object with the changed motion or shape in the first frame of video sequence, namely A1, and the part is sent to the second terminal for updating the missing background part on the basis of the first background model.
Step S140, the second terminal updates the background missing part of the foreground image on the first background model so as to update the first background model to the second background model, performs foreground covering on the basis of the second background model, and displays the image subjected to foreground covering as a second frame of the video;
the second terminal uses the area occupied by the object with changed motion or shape in the received foreground image in the previous frame video sequence, namely the area occupied in the first frame video sequence, to update the background missing part on the basis of the first background model, so as to update the first background model to the second background model, and further update the background model in the following process. And covering the foreground of the area occupied by the object which actually moves or has a changed appearance in the foreground image in the second frame video sequence on the basis of the second background model, and displaying the image with the covered foreground as a second frame of the video.
Next, repeating the steps S130 and S140, the first terminal performing foreground segmentation on the changed region in the video sequence of the next frame relative to the video sequence of the previous frame, and transmitting the segmented foreground image to the second terminal; and the second terminal updates the background missing part of the foreground image on the basis of the previous background model so as to update the foreground image to a next background model, and performs foreground covering so as to display the image after the foreground covering as a next frame of video sequence.
In the application, the transmission of the video sequence after the first frame video sequence only transmits the segmented foreground image, so that the transmission amount of the video data with high data volume can be reduced by the video transmission method, and the phenomenon of blocking during the transmission process is avoided.
Example two
Referring to fig. 5, the present application further provides a video image transmission system, including: a first terminal 510 and a second terminal 520, wherein the first terminal 510 includes: a segmentation module 511, an acquisition module 512 and a first transmission module 513, the second terminal 520 comprising: a display module 521, a synthesis module 522 and a second transmission module 523.
The capture module 512 of the first terminal 510 captures video images, i.e., captures each frame of a video sequence of the video images and transmits each frame of the captured video sequence of the video images to the segmentation module 511, and the capture module 512 transmits the captured first frame of the video sequence to the first transmission module 513; the segmentation module 511 establishes a first background model according to the collected video image, and transmits the established first background model to the first transmission module 513; the first transmission module 513 transmits the first frame sequence of video images and the established first background model to the second transmission module 523 of the second terminal 520.
The capture module 512 of the first terminal 510 captures video images and the segmentation module 511 builds a first background model from the first N frames of the video sequence.
Specifically, the segmentation module 511 converts the video sequences of the first frame to the nth frame into the first gray image to the nth gray image, respectively.
Specifically, the segmentation module 511 sets the same value of the three colors R, G, B of each pixel in each frame of video sequence to perform the graying processing on each frame of video sequence, and specifically performs the graying processing by using the following formula R ═ G ═ wr ═ R + wg ═ G + wb ═ B, where wr, wg, and wb are weights of R, G, B, respectively.
Next, the segmentation module 511 calculates the absolute gradient value of each pixel point in the first gray image to the nth gray image in sequence, and takes the pixel point whose absolute gradient value is greater than the threshold value in each gray image as the edge point of the gray image.
The example of finding the edge point of the first gray image corresponding to the first frame of video sequence is described, where the first gray image corresponding to the first frame of video sequence is a two-dimensional function f (x, y), where x and y are coordinate values of a pixel in the first gray image, an absolute gradient value M (x, y) ═ f (x, y) -f (x +1, y +1) | + | f (x +1, y) -f (x, y) +1|, and if the absolute gradient value M (x, y) of the pixel is greater than a preset threshold, the pixel belongs to the edge point of the first gray image corresponding to the first frame of video sequence.
And sequentially solving an absolute gradient value M (x, y) for each pixel point in the second gray scale image to the Nth gray scale image corresponding to the second frame video sequence to the Nth frame video sequence according to the formula, and respectively taking the pixel points of which the absolute gradient values M (x, y) are greater than a preset threshold value as edge points of the corresponding gray scale images.
Then, the segmentation module 511 determines whether the image surrounded by the edge points is an image corresponding to an object which is moving or has a changed shape according to the position and shape information of the image surrounded by the edge points in the N grayscale images;
the edge points of each gray level image surround to form one or more images, the positions f (x, y) of the same or corresponding edge points of the images are changed in M gray level images, wherein M is more than or equal to 2 and less than or equal to N, the object corresponding to the image is considered to be moving or the shape of the object is changed, and otherwise, the object corresponding to the image is considered to be not moving and the shape of the object is not changed.
If the image surrounded by the edge points is an image corresponding to an object which is moving or has a changed shape, the segmentation module 511 defines the image as a foreground image;
in the application, an image corresponding to an object which is moving or has a changed shape (such as a flying bird, a walking person, a running car and a twitch) is defined as a foreground image, and an image corresponding to a relatively static object is defined as a background image to be reused subsequently.
Finally, the segmentation module 511 performs the segmentation of the foreground and the background in the first frame video sequence according to the edge point of the foreground image in the first gray image to form a first background model.
In addition, the segmentation module 511 establishes a first background model according to the first N frames of video sequence of the video image collected by the first terminal, and may specifically include the following steps:
firstly, the segmentation module 511 converts the video sequences of the first frame to the nth frame into the first gray image to the nth gray image respectively;
next, the segmentation module 511 calculates information entropy h (x) for pixel points x in the first to nth grayscale images;
specifically, let a (x) { a ═ a0,a1,...,aN-1B (x) represents a set of all pixel points in a 3 × 3 neighborhood centered on the point x, and the information entropy of the point x is
Figure BDA0002457816370000111
Wherein p (mu) is the gray probability density of all pixel points in B (x), M is 3 × 3 × N, and represents the total number of all sample point sets in B (x) taking x as the center in the previous N frames of video sequences;
Figure BDA0002457816370000112
m(ai) Is pixel point a in A (x)iCorresponding feature class, i is an integer from 0 to N-1; delta [ m (a) ]i)-μ]Is a Dikela function, which is used to determine whether the color value of the pixel is equal to the color value of the μ -th pixel in the set B (x), which is equal to 1, but not equal to 0.
Next, the segmentation module 511 defines an image corresponding to an area where the information entropy h (x) is greater than a preset threshold as a foreground image;
if the information entropy h (x) of a certain area is greater than a preset threshold value, and the difference between pixels of the area is large, the object corresponding to the image of the area is in a dynamic state, and the image corresponding to the object in the dynamic state is defined as a foreground image in the application; otherwise, the information entropy h (x) of the area is smaller than the preset threshold, and the difference between the pixels of the area is small, so that the object corresponding to the image of the area is in a static state, and the image corresponding to the object in the static state is defined as a background image.
Finally, the segmentation module 511 performs the segmentation of the foreground and the background in the first frame video sequence according to the occupied area of the foreground image in the first gray image to form a first background model.
The second transmission module 523 of the second terminal 520 receives the first frame video sequence which is sent by the first terminal 510 and is not segmented and the first background model sent by the first terminal 510, and the display module 521 of the second terminal 520 displays the first frame video sequence as a first frame of a video on the display module 521 of the second terminal 520.
The segmentation module 511 of the first terminal 510 performs foreground segmentation on a region of the second frame video sequence that changes with respect to the first frame video sequence, and transmits the segmented foreground image to the first transmission module 513 to be transmitted to the second transmission module 523 of the second terminal 520 through the first transmission module 513.
The segmentation module 511 of the first terminal 510 uses the first frame video sequence as a static background, where a pixel in the first frame video sequence is denoted as b (x, y), a pixel in the jth frame video sequence is denoted as g (x, y, j), and a pixel in the second frame video sequence is g (x, y, 2), and subtracts a gray value of a pixel in the second frame video sequence from a gray value of a pixel in the first frame video sequence, that is, i d (x, y, 2) ═ g (x, y, 2) -b (x, y), to obtain an area where the gray value is changed by the pixel, so that the area is segmented from the second frame video sequence as a foreground image.
In the segmented foreground image, as shown in fig. 4, a is a segmented foreground image, which is an image change condition corresponding to a position a1 of an object (e.g., a sphere) from a previous time to a position a2 of the object at a subsequent time, that is, a part of the segmented foreground image is an area occupied by the object which actually moves or has a changed appearance in the second frame video sequence, that is, a2, and the part is sent to the second terminal for covering the foreground on the basis of the first background model. And the other part of the segmented foreground image is the area occupied by the object with the changed motion or shape in the previous frame of video sequence, namely the area occupied by the object with the changed motion or shape in the first frame of video sequence, namely A1, and the part is sent to the second terminal for updating the missing background part on the basis of the first background model.
The second transmission module 523 of the second terminal 520 receives the foreground image transmitted by the first transmission module 513 of the first terminal 510 and transmits the foreground image to the synthesis module 522; the synthesis module 522 updates the foreground image with a background missing part on the first background model to update the first background model to the second background model, and the synthesis module 522 performs foreground coverage on the basis of the second background model and transmits the image subjected to foreground coverage to the display module 521; the display module 521 displays the foreground-overlaid image as a second frame of the video.
The synthesizing module 522 of the second terminal 520 uses the area occupied by the object with changed motion or shape in the foreground image in the previous frame of video sequence, that is, the area occupied in the first frame of video sequence, to update the missing background part on the basis of the first background model, so as to update the first background model to the second background model, and then further update the background model. And the synthesizing module 522 covers the foreground of the area occupied by the actual moving or shape-changing object in the foreground image in the second frame video sequence on the basis of the second background model, and the display module displays the image with the foreground covered as the second frame of the video.
Next, the segmentation module 511 of the first terminal 510 performs foreground segmentation on the changed region of the video sequence of the next frame relative to the video sequence of the previous frame, and the first transmission module 513 transmits the segmented foreground image to the second transmission module 523 of the second terminal 520; the synthesis module 522 of the second terminal 520 updates the foreground image to the background missing part on the basis of the previous background model, so as to update to the next background model, and performs foreground covering, so that the display module 521 displays the image after foreground covering as the next frame of video sequence.
It will be evident to those skilled in the art that the invention is not limited to the details of the foregoing illustrative embodiments, and that the present invention may be embodied in other specific forms without departing from the spirit or essential attributes thereof. The present embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein. Any reference sign in a claim should not be construed as limiting the claim concerned.
Furthermore, it should be understood that although the present description refers to embodiments, not every embodiment may contain only a single embodiment, and such description is for clarity only, and those skilled in the art should integrate the description, and the embodiments may be combined as appropriate to form other embodiments understood by those skilled in the art.

Claims (10)

1. A video image transmission method, comprising the steps of:
respectively converting a first frame video sequence to an Nth frame video sequence into a first gray image to an Nth gray image according to the first N frames of video sequences of the collected video images;
calculating information entropy of pixel points x in the first gray level image to the Nth gray level image;
defining an image corresponding to a region with the information entropy larger than a preset threshold value as a foreground image;
according to the area occupied by the foreground image in the first gray image, segmenting the foreground and the background in the first frame video sequence to form a first background model;
transmitting a first frame sequence of the video image and the established first background model so as to display the transmitted first frame video sequence as a first frame of the video;
foreground segmentation is carried out on an area, changed relative to the first frame video sequence, in the second frame video sequence, and a segmented foreground image is transmitted;
and updating the background missing part of the transmitted foreground image on the first background model to update the first background model to the second background model, performing foreground covering on the foreground image on the basis of the second background model, and displaying the image subjected to foreground covering as a second frame of the video.
2. The video image transmission method according to claim 1, wherein the three colors R, G, B of each pixel in each frame of the video sequence are set to have the same value, and a gray image is generated by performing a graying process according to the following formula R-G-wr-R + wg-G + wb-B, where wr, wg, and wb are weights R, G, B, respectively.
3. The video image transmission method according to claim 1 or 2, wherein the first frame video sequence is used as a static background, the gray value of the pixel in the second frame video sequence is subtracted from the gray value of the pixel in the first frame video sequence to obtain a region formed by the pixel with a changed gray value, and the region is used as a foreground image and is segmented from the second frame video sequence.
4. The video image transmission method according to claim 3, wherein one part of the segmented foreground images is an area occupied by an object with actual motion or a changed appearance in the second frame video sequence, and the other part of the segmented foreground images is an area occupied by the object with the actual motion or the changed appearance in the previous frame video sequence.
5. The video image transmission method according to claim 4, wherein the area of the object with changed motion or shape in the foreground image in the first frame video sequence is used for updating the missing background part based on the first background model;
and the area of the object with the actual motion or the shape change in the foreground image in the second frame video sequence is used for performing foreground coverage on the basis of the second background model.
6. A system for transmitting video images, comprising: a first terminal and a second terminal, the first terminal comprising: collection module, segmentation module and first transmission module, the second terminal includes: the device comprises a synthesis module, a display module and a second transmission module;
the acquisition module acquires each frame of video sequence of a video image, the segmentation module converts the first frame video sequence to the Nth frame video sequence into a first gray image to an Nth gray image respectively according to the first N frames of video sequences of the acquired video image, calculates information entropy for pixel points x in the first gray image to the Nth gray image, defines an image corresponding to a region of which the information entropy is greater than a preset threshold value as a foreground image, and performs foreground and background segmentation in the first frame video sequence according to a region occupied by the foreground image in the first gray image to form a first background model;
the first transmission module transmits a first frame sequence of the video image and the established first background model to the second transmission module;
the second transmission module receives the first frame video sequence and the established first background model for transmission, and the display module displays the first frame video sequence as a first frame of the video;
the segmentation module performs foreground segmentation on a region, which changes relative to the first frame video sequence, in the second frame video sequence, and the first transmission module transmits a segmented foreground image to the second transmission module;
the second transmission module receives the foreground image, the synthesis module updates the background missing part of the transmitted foreground image on the first background model so as to update the first background model to the second background model, foreground covering is carried out on the foreground image on the basis of the second background model, and the display module displays the image subjected to foreground covering as a second frame of the video.
7. The video image transmission system of claim 6, wherein the segmentation module sets the three colors R, G, B of each pixel in each frame of the video sequence to be the same, and generates the gray image by performing the graying process according to the following formula R-G-wr-R + wg-G + wb, where wr, wg, and wb are weights of R, G, B, respectively.
8. The video image transmission system according to claim 6 or 7, wherein the segmentation module takes the first frame video sequence as a static and unchangeable background, subtracts the gray value of the pixel in the second frame video sequence from the gray value of the pixel in the first frame video sequence to obtain a region formed by the pixels with changed gray values, and segments the region from the second frame video sequence as a foreground image.
9. The video image transmission system according to claim 8, wherein one part of the divided foreground images is an area occupied by an object with actual motion or a changed appearance in the second frame video sequence, and the other part of the divided foreground images is an area occupied by the object with the actual motion or the changed appearance in the previous frame video sequence.
10. The video image transmission system according to claim 9, wherein the synthesis module uses an area occupied by an object with a changed motion or shape in the foreground image in the first frame video sequence to update the missing background portion based on the first background model;
and the synthesis module is used for covering the foreground of the object with actual motion or changed appearance in the foreground image in the second frame video sequence on the basis of the second background model.
CN202010310932.1A 2020-04-20 2020-04-20 Video image transmission method and transmission system Active CN111371962B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010310932.1A CN111371962B (en) 2020-04-20 2020-04-20 Video image transmission method and transmission system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010310932.1A CN111371962B (en) 2020-04-20 2020-04-20 Video image transmission method and transmission system

Publications (2)

Publication Number Publication Date
CN111371962A true CN111371962A (en) 2020-07-03
CN111371962B CN111371962B (en) 2021-09-07

Family

ID=71212111

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010310932.1A Active CN111371962B (en) 2020-04-20 2020-04-20 Video image transmission method and transmission system

Country Status (1)

Country Link
CN (1) CN111371962B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1418766A2 (en) * 1998-08-28 2004-05-12 Imax Corporation Method and apparatus for processing images
CN103281539A (en) * 2013-06-07 2013-09-04 华为技术有限公司 Method, device and terminal for image encoding and decoding processing
CN103313061A (en) * 2013-06-28 2013-09-18 汕头大学 Adaptive background updating video compression and transmission method and system
CN106327520A (en) * 2016-08-19 2017-01-11 苏州大学 Moving object detection method and system
CN109509205A (en) * 2017-09-14 2019-03-22 北京君正集成电路股份有限公司 Foreground detection method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1418766A2 (en) * 1998-08-28 2004-05-12 Imax Corporation Method and apparatus for processing images
CN103281539A (en) * 2013-06-07 2013-09-04 华为技术有限公司 Method, device and terminal for image encoding and decoding processing
CN103313061A (en) * 2013-06-28 2013-09-18 汕头大学 Adaptive background updating video compression and transmission method and system
CN106327520A (en) * 2016-08-19 2017-01-11 苏州大学 Moving object detection method and system
CN109509205A (en) * 2017-09-14 2019-03-22 北京君正集成电路股份有限公司 Foreground detection method and device

Also Published As

Publication number Publication date
CN111371962B (en) 2021-09-07

Similar Documents

Publication Publication Date Title
CN111526417B (en) Video image transmission method and transmission system
US10991076B2 (en) Image super-resolution method, image super-resolution device, and computer readable storage medium
CN106162177B (en) Method for video coding and device
CN106875437B (en) RGBD three-dimensional reconstruction-oriented key frame extraction method
US20220222776A1 (en) Multi-Stage Multi-Reference Bootstrapping for Video Super-Resolution
KR20190128686A (en) Method and apparatus, equipment, and storage medium for determining the pose of an object in an image
CN108234882B (en) Image blurring method and mobile terminal
CN110827193A (en) Panoramic video saliency detection method based on multi-channel features
CN102194443A (en) Display method and system for window of video picture in picture and video processing equipment
US11917158B2 (en) Static video recognition
CN110944160A (en) Image processing method and electronic equipment
CN111145192A (en) Image processing method and electronic device
CN108806638B (en) Image display method and device
CN112085031A (en) Target detection method and system
CN110868544B (en) Shooting method and electronic equipment
CN111583357A (en) Object motion image capturing and synthesizing method based on MATLAB system
CN110689498B (en) High-definition video optimization method based on hierarchical blurring of non-focus part
CN112614110A (en) Method and device for evaluating image quality and terminal equipment
CN106846250B (en) Super-resolution reconstruction method based on multi-scale filtering
CN111371962B (en) Video image transmission method and transmission system
CN111797694B (en) License plate detection method and device
CN111291663B (en) Method for quickly segmenting video target object by using space-time information
CN114913095B (en) Depth deblurring method based on domain adaptation
CN112365517A (en) Super-pixel segmentation method based on image color and density characteristics
CN115660994B (en) Image enhancement method based on regional least square estimation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20210819

Address after: 510000 room 307, No. 5, jiantashan Road, Huangpu District, Guangzhou City, Guangdong Province

Applicant after: Guangzhou XINGFEIDA Electronic Technology Co.,Ltd.

Address before: 102400 no.18-d11961, Jianshe Road, Kaixuan street, Liangxiang, Fangshan District, Beijing

Applicant before: Beijing yingmaiqi Technology Co.,Ltd.

GR01 Patent grant
GR01 Patent grant
PE01 Entry into force of the registration of the contract for pledge of patent right
PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: A Video Image Transmission Method and System

Effective date of registration: 20230831

Granted publication date: 20210907

Pledgee: Bank of China Limited by Share Ltd. Guangzhou Tianhe branch

Pledgor: Guangzhou XINGFEIDA Electronic Technology Co.,Ltd.

Registration number: Y2023980054683