CN105657399A - 3D medical video transmission method in wireless network environment - Google Patents

3D medical video transmission method in wireless network environment Download PDF

Info

Publication number
CN105657399A
CN105657399A CN201610004628.8A CN201610004628A CN105657399A CN 105657399 A CN105657399 A CN 105657399A CN 201610004628 A CN201610004628 A CN 201610004628A CN 105657399 A CN105657399 A CN 105657399A
Authority
CN
China
Prior art keywords
centerdot
lsqb
rsqb
theta
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610004628.8A
Other languages
Chinese (zh)
Other versions
CN105657399B (en
Inventor
刘金霞
张增年
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Wanli College
Original Assignee
Zhejiang Wanli College
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Wanli College filed Critical Zhejiang Wanli College
Priority to CN201610004628.8A priority Critical patent/CN105657399B/en
Publication of CN105657399A publication Critical patent/CN105657399A/en
Application granted granted Critical
Publication of CN105657399B publication Critical patent/CN105657399B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/647Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load, bridging between two different networks, e.g. between IP and wireless
    • H04N21/64784Data processing by the network
    • H04N21/64792Controlling the complexity of the content stream, e.g. by dropping packets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/139Format conversion, e.g. of frame-rate or size
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Security & Cryptography (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Medicines Containing Antibodies Or Antigens For Use As Internal Diagnostic Agents (AREA)

Abstract

The invention discloses a 3D medical video transmission method in a wireless network environment. An end to end video distortion in a mode of a current coding QP (quantization parameter), a current authentication Hash quantity and current MCS (modulation and channel coding schemes) is obtained through recursive computation; the end to end video distortions under combination conditions of all coding QPs, authentication Hash quantities and MCS are computed through traversal; the QP, the authentication Hash quantity and the current MCS under a least distortion condition are selected for coding and authenticating an application layer, configuring the physical layer MCS and transmitting current data packets. Through the mode, the strengths of different parameters in authentication and fault-tolerant aspects can be configured and coordinated systematically; and the quality of the 3D medical video can be maximized in the current network state.

Description

3D medical video transmission method in wireless network environment
Technical Field
The invention relates to the field of video coding and wireless transmission, in particular to a 3D medical video transmission method in a wireless network environment.
Background
With the rapid development of electronic information and multimedia communication technologies, modern medical technologies gradually move to a digital distributed application mode. The modern development of the medical health industry is greatly promoted by the electronic health technology generated by combining the information technology and the biotechnology. The core problem of the electronic health technology is to remotely process and monitor the health information of the patient. The combination of video processing and wireless communication technology breaks through the regional limitation of medical level difference, and a remote medical system is produced. Most of the traditional telemedicine systems are based on 2D video communication systems to transmit and interact medical information. In recent years, with the rapid development of technologies such as digital imaging, 3D vision, and 3D display, the digital 3D video technology has become a research hotspot in academia and industry at present. Compared with a 2D video, the 3D video can enable the foreground image and the background image to be distinguished more obviously, and has clearer boundary outline and finer texture patterns, so that better imaging quality and more vivid and natural 3D perception effect can be presented to people, people can see the three-dimensional nature of a natural scene again, and the people can feel visual experience which is completely different from that of a common plane picture.
To provide more accurate third-dimensional information, 3D video may be applied to telemedicine systems. The 3D video system can enable a doctor to obtain depth information data which cannot be captured from the traditional plane display, can read image information in all directions, provides richer and more accurate image data for clinical diagnosis, and greatly reduces missed diagnosis of a focus, thereby improving diagnosis quality.
Currently, 3D medical video communication is mostly based on wireless networks, which can cover remote areas. Since wireless communication technology is more vulnerable to intrusion than wired communication, authentication technology in the medical video communication process is very critical to protecting the privacy of patients. The traditional authentication signature technology facing data flow signs each data packet, and the cost such as complexity and the like is high.
In the data packet encoding process, each encoded slice (slice) is packed as a basic packing unit. And the code rate of each video data packet is basically determined by the quantization parameter of the coding. Therefore, in real-time coding, the Quantization Parameter (QP) of the coded slice can be determined in real time according to the channel bandwidth, and the selection of different QPs will result in different levels of video distortion, i.e. different levels of video quality. Each encoded video data packet is transmitted as a basic application layer unit to be transmitted to other layers of the network.
In order to ensure the security and easy copyright management of video data packets, video transmission usually requires an authentication technique for encryption or authentication processing. In the existing video transmission, a multi-path hash chain authentication mode, that is, a hash authentication processing mode based on a same source multi-chain, may be used to authenticate a group of pictures (GOP). The packets of each frame may be selected by parent nodes (the upper-level indices of the hash chain) that hash the corresponding position packets within the already decoded frame within the GOP. There may be several hash chain path selection schemes (i.e. parent node selection) for each packet, and each scheme estimates the authentication probability of the current authentication scheme of the current packet according to the authentication of the previous node on the path. While calculating the number of bits spent for the authentication.
Assuming that each packet can have a maximum of N hash parents, the number of bits consumed by each hash is b, and the probability of successful authentication isThe probability of channel transmission packet loss is assumed to be p. Each packet may have multiple hash links such that the probability of successful authentication for each packet is different, assuming that N packets as a group have a signature. To reduce authentication and decoding delays, we assume that the current packet can only select as the parent node of the hash the packet on which the current packet depends that must be decoded earlier than the current packet. The ith (i) th node is present, assuming there are M hash parent nodes available<M) probability of successful authentication of data packets is
If it is currently the ith (i)<M) number of hash parent nodes selected by data packets is M (M)<M) and the nth parent node is a distance d from the current packetnThen the probability of successful authentication of the current data packet is revised to
And the number of bits spent for hash authentication is m · b. Assuming that the length of the original data packet is S and the length of the packet after adding the hash authentication data is S + m · b, the number of resource blocks that each data packet can be divided into in the physical layer and the corresponding modulation and channel coding (MCS) may change according to the quality of the channel. Therefore, as the number of hash parent nodes increases, the number of resource blocks corresponding to each packet also increases, and the packet loss probability of each corresponding data packet also increases, so that the hash number of each data packet directly affects the authentication probability and the packet loss probability of data.
In addition to source and channel distortion during video transmission, video packets that fail authentication will also cause video distortion. This is mainly because the current data packet is distorted due to the fact that the current data packet cannot be authenticated because of the packet loss of the data packet on which the hash authentication depends. On the other hand, the authentication bit data occupies bandwidth data, which may cause the loss of the quality of the information source due to the reduction of the information source data, that is, a certain distortion is generated.
The stream-level authentication based on the hash chain (hashchain) signature is simple. In order to improve the fault tolerance of the authentication data packet, a homologous multi-chain authentication mode is generally adopted. But the homologous multiple chain authentication mode greatly increases redundancy of transmitted data. How to effectively select the number of authenticated hash paths directly affects the efficiency of video data transmission and the distortion of a receiving end video. Because the wireless network physical layer supports a certain modulation and channel coding mode, the data stream can be protected to a certain extent, but the selection of the physical layer modulation and channel coding mode also directly influences the packet loss probability and video distortion of the data packet. The traditional authentication method does not select authentication hash from the perspective of the overall optimization of the system, so that the optimization of the wireless communication quality of the 3D medical video cannot be guaranteed under the dynamic channel condition.
Disclosure of Invention
The technical problems to be solved by the invention are as follows: the 3D medical video transmission method under the wireless network environment with good video transmission quality is provided.
In order to solve the technical problems, the invention is realized by the following technical scheme: A3D medical video transmission method under a wireless network environment is characterized in that: the method comprises the following specific steps:
1) according to the selection of the number of the wireless 3D medical video hash capable of fault-tolerant authentication, the selection of an application layer video coding quantization parameter and the selection of an LTE downlink physical layer modulation and coding mode, the following optimal parameters are comprehensively established by minimizing end-to-end 3D video distortion:
( m h a s h ( d 0 , ... d m - 1 ) o p t , QP o p t , MCS o p t ) = arg m i n m &Element; Q , Q P &Element; Q , M C S &Element; Z D ( m , Q P , M C S )
subjecttot<Tmax
wherein m ishash(d0,...dm-1)optIs the optimal Hash father node of the current data packet, H, Q, and Z respectively represent the candidate Hash father node set, QP set and MCS mode set, QP is the quantization parameter of the coding strip, MCS is the physical layer modulation and coding mode, TmaxThe maximum time delay limit of data packet transmission, T is the transmission time delay of the data packet, and can be obtained approximately by dividing the size of the data packet by the transmission rate of a physical layer;
2) setting parameters such as initial QP and MCS and the like: setting QP to minQP, minD to 0, maxMCS to 15, and maxQP to 45;
3) aiming at a 3D medical video acquired by a binocular camera, coding a coding structure with inter-view prediction according to multi-view video coding, and coding by using QP as a quantization parameter;
4) setting the data volume of a Hash father node to be 0, namely, m is 0, obtaining the maximum possible Hash node quantity of the current data packet to be maxH, and obtaining the maximum possible Hash node quantity of the current data packet according to a formulaCalculating the probability of successful authentication of the current ith data packet, wherein the current ith (i) th data packet<maxH) number of hash parent node selected by packet is m (m)<maxH) and the tth parent node is a distance d from the current packettThe probability of successful authentication isThe probability of packet loss of the current data transmission application layer is assumed to be rhoi
5) Setting the coding and modulation mode MCS of the physical layer channel to be 1; for each modulation and coding mode h, calculating the equivalent signal-to-noise ratio &gamma; m i e f f ( h ) = &kappa; ( h ) &lsqb; J - 1 ( 1 N s b &Sigma; j = 1 N s b J ( &gamma; j &kappa; ( h ) ) ) &rsqb; 2 ;
Wherein N issbRepresenting the number of sub-carriers, gammajIs the signal to interference plus noise ratio of the jth sub-carrier, k (h) is the correction factor in modulation and coding mode h, J (x) and J-1(y) is calculated as follows:
J ( x ) &ap; - 0.04210610 x 3 + 0.209252 x 2 - 0.00640081 x , 0 < x < 1.6363 1 - exp ( 0.00181491 x 3 - 0.142675 x 2 - 0.08220540 x + 0.0549608 ) , x &GreaterEqual; 1.6363
J - 1 ( y ) &ap; 1.09542 y 2 + 0.214217 y + 2.33727 y , 0 < y < 0.3646 - 0.706692 l o g ( - 0.386013 ( y - 1 ) ) + 1.75017 y , 0.3646 &le; y &le; 1 ;
6) calculating the physical layer resource block loss rate and the application layer packet loss rate of the current data packet according to the signal-to-noise ratio (SNR):
the block loss probability for a resource block with the selected modulation and coding mode h is calculated as follows:
where erfc (-) is a complementary error function, and b (h) and c (h) can be obtained in advance by means of fitting calculations;
packet loss probability of ith packet in nth frame of application layer video code streamWherein, BnumThe physical layer resource occupied by the current data packet is fast;
7) calculating distortion according to the coding structure, wherein the packet loss rate in the distortion calculation comprises the combination of the authentication probability and the packet loss probability, namely the probability that the current packet is successfully authenticated and does not lose the packet is
8) Calculating end-to-end distortion D in the current data packet transmission process:
D = 1 S &Sigma; i = 1 S E &lsqb; ( f n i - f ~ n i ) 2 = 1 S &Sigma; i = 1 S ( f n i ) 2 - 2 &CenterDot; f n i &CenterDot; E &lsqb; ( f ~ n i ) &rsqb; + E &lsqb; ( f ~ n i ) 2 &rsqb; ;
wherein, S is the number of the image contained in the current data packet in the current nth frame;
if the current frame belongs to the left viewpoint, a traditional distortion calculation method is adopted. The distortion is calculated by using a method in which the pixel is expected to recur. Distortion calculation for intra-frame predictive coding block
E &lsqb; f ~ n , l i &rsqb; = &theta; n , l i &CenterDot; ( f ~ n , l i ) + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; f ~ n - 1 , l k &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; f ~ n - 1 , l i &rsqb; ;
Image number distortion calculation for inter-predicted coding blocks
E &lsqb; f ~ n , l i &rsqb; = &theta; n , l i &CenterDot; ( e ^ n , l i + E &lsqb; f ~ n - 1 , l j &rsqb; ) + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; f ~ n - 1 , l k &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; f ~ n - 1 , l i &rsqb; ;
E &lsqb; ( f ~ n , l i ) 2 &rsqb; = &theta; n , l i &CenterDot; { ( e ^ n , l i ) 2 + 2 e ^ n , l i &CenterDot; E &lsqb; f ~ n - 1 , l j &rsqb; + E &lsqb; ( f ~ n - 1 , l j ) 2 &rsqb; } + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; ( f ~ n - 1 , l j ) 2 &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; ( f ~ n - 1 , l i ) 2 &rsqb; ;
And if the current frame belongs to the right viewpoint, distortion calculation is carried out according to different prediction relations: for theThe calculation of (a) can be classified into the following cases:
(1) the current pixel is located in the coding block by adopting intra-frame prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( f ~ n , r i ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(2) the coding block where the current pixel is located adopts inter prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + E &lsqb; f ~ n - 1 , r j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(3) the coding block where the current pixel is located adopts inter-view prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + E &lsqb; f ~ n , l j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(4) the weighted prediction of the inter prediction and the inter prediction employed by the coding block in which the current pixel is located,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + w 1 &CenterDot; E &lsqb; f ~ n - 1 , r o &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; ;
(1) the current pixel is in the coding block by adopting intra-frame prediction
(2) The coding block where the current pixel is located adopts inter prediction,
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
(3) the current pixel is in the coding block using inter-view prediction
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; E &lsqb; f ~ n , l j &rsqb; + E &lsqb; ( f ~ n , l j ) 2 &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
(4) The coding block where the current pixel is located adopts weighted prediction between views and frames,
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &CenterDot; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb; &ap; &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &rsqb; &CenterDot; E &lsqb; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &CenterDot; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb; &ap; &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &rsqb; &CenterDot; E &lsqb; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
wherein,is the value of the ith pixel in the nth frame,is the ith pixel value, E [. in the nth frame at the decoder end]In order to be a function of the expectation,for the ith frame of the left view at the decoder endThe value of each of the pixels is calculated,for the ith pixel value in the (n-1) th frame of the left view at the decoder side,is the error concealment value of the ith pixel in the (n-1) th frame of the left view at the decoder end,for the reconstructed value of the ith pixel in the nth frame at the encoder end,is a pixelThe probability of the authentication being successful is determined,is a pixelThe probability of the authentication being successful is determined,is a pixelThe coded prediction residual of (2) is,is a pixelThe coded prediction value of (a) is,for the ith pixel value of the nth frame of the right view at the decoder end,for the ith pixel value in the n-1 th frame of the right view at the decoder side,for the error concealment value of the ith pixel of the (n-1) th frame of the right view at the decoder end,is the reconstructed value of the ith pixel in the nth frame of the right view at the encoder end,is a pixelThe probability of the authentication being successful is determined,is a pixelThe probability of the authentication being successful is determined,is a pixelThe residual values of the encoding of (1),is a pixelThe coded prediction value of (a) is,is a pixelA coding prediction value in a left view; the data can be directly obtained in the encoding process;andobtained by step 7;
9) if minD > D, minD is equal to D, and m _ opt is equal to m, MCS _ opt is equal to MCS, QP _ opt is equal to QP, and MCS is equal to MCS + 1;
10) if the MCS is less than or equal to maxMCS, returning to the step 6;
11) if m is less than or equal to maxH, returning to the step 4;
12) QP + 1; if the QP is less than or equal to maxQP, returning to the step 3;
13) and outputting m _ opt, MCS _ opt and QP _ opt, and using the parameters to encode the current 3D video data packet, configure the number of Hash authentication father nodes, select a physical layer MCS and perform wireless transmission.
By adopting the structure, the invention has the advantages that: and selecting one number of the Hash chains according to the number of the candidate Hash chains, and then estimating the corresponding probability of successful authentication. Then, the number of MCS modes of the physical layer candidate is determined, one MCS is selected, an equivalent SNR is determined, and then the block loss rate corresponding to the current MCS is estimated. According to the quantization parameter of the video coding, the alternative video coding code rate of the current data packet is obtained, a Quantization Parameter (QP) is selected, and the current packet loss probability is estimated. And obtaining the current coding QP, the current authentication Hash quantity and the end-to-end video distortion under the current MCS mode through recursive calculation. And performing application layer coding and authentication by traversing and calculating end-to-end distortion under the combination condition of all QP, authentication Hash quantity and MCS, selecting the QP, the authentication Hash quantity and the MCS under the condition of minimum distortion, configuring the MCS of a physical layer, and transmitting the current data packet. The method can configure and coordinate the intensity of different parameters on the aspects of authentication and fault tolerance, and can maximize the quality of the 3D medical video under the current network state.
Drawings
Fig. 1 is an encoding structure of inter-view prediction for a binocular acquired 3D medical video.
Fig. 2 is a hash-based authentication mechanism.
Fig. 3 is a 3D medical video authentication and fault tolerant transmission module framework.
Table 1 shows MCS modes of downlink candidates in a wireless environment.
As shown, I-Intra coded frames, P-uni-directional predictive coded frames, B-bi-directional predictive coded frames.
Detailed Description
The invention is further described below with reference to the accompanying drawings.
In the description of the present invention, it is to be understood that the terms "upper", "lower", "inside", "outside", and the like indicate orientations or positional relationships based on the positional relationships illustrated in the drawings, and are only for convenience in describing the present invention or simplifying the description, but do not indicate that a specific orientation is necessary.
The Intra-coded frames are Intra-prediction coded.
TABLE 1
As shown in fig. 1, fig. 2, fig. 3 and table 1, a 3D medical video transmission method in a wireless network environment is characterized in that: the method comprises the following specific steps:
1) according to the selection of the number of the wireless 3D medical video hash capable of fault-tolerant authentication, the selection of an application layer video coding quantization parameter and the selection of an LTE downlink physical layer modulation and coding mode, the following optimal parameters are comprehensively established by minimizing end-to-end 3D video distortion:
( m h a s h ( d 0 , ... d m - 1 ) o p t , QP o p t , MCS o p t ) = arg m i n m &Element; Q , Q P &Element; Q , M C S &Element; Z D ( m , Q P , M C S )
subjecttot<Tmax
wherein m ishash(d0,...dm-1)optIs the optimal Hash father node of the current data packet, H, Q, and Z respectively represent the candidate Hash father node set, QP set and MCS mode set, QP is the quantization parameter of the coding strip, MCS is the physical layer modulation and coding mode, TmaxThe maximum time delay limit of data packet transmission, T is the transmission time delay of the data packet, and can be obtained approximately by dividing the size of the data packet by the transmission rate of a physical layer;
2) setting parameters such as initial QP and MCS and the like: setting QP to minQP, minD to 0, maxMCS to 15, and maxQP to 45;
3) aiming at a 3D medical video acquired by a binocular camera, coding a coding structure with inter-view prediction according to multi-view video coding, and coding by using QP as a quantization parameter;
4) setting the data volume of a Hash father node to be 0, namely, m is 0, obtaining the maximum possible Hash node quantity of the current data packet to be maxH, and obtaining the maximum possible Hash node quantity of the current data packet according to a formulaCalculating the probability of successful authentication of the current ith data packet, wherein the current ith (i) th data packet<maxH) number of hash parent node selected by packet is m (m)<maxH) and the tth parent node is a distance d from the current packettThe probability of successful authentication isThe probability of packet loss of the current data transmission application layer is assumed to be rhoi
5) Setting the coding and modulation mode MCS of the physical layer channel to be 1; for each modulation and coding mode h, calculating the equivalent signal-to-noise ratio &gamma; m i e f f ( h ) = &kappa; ( h ) &lsqb; J - 1 ( 1 N s b &Sigma; j = 1 N s b J ( &gamma; j &kappa; ( h ) ) ) &rsqb; 2 ;
Wherein N issbRepresentational sonNumber of carriers, gammajIs the signal to interference plus noise ratio of the jth sub-carrier, k (h) is the correction factor in modulation and coding mode h, J (x) and J-1(y) is calculated as follows:
J ( x ) &ap; - 0.04210610 x 3 + 0.209252 x 2 - 0.00640081 x , 0 < x < 1.6363 1 - exp ( 0.00181491 x 3 - 0.142675 x 2 - 0.08220540 x + 0.0549608 ) , x &GreaterEqual; 1.6363
J - 1 ( y ) &ap; 1.09542 y 2 + 0.214217 y + 2.33727 y , 0 < y < 0.3646 - 0.706692 l o g ( - 0.386013 ( y - 1 ) ) + 1.75017 y , 0.3646 &le; y &le; 1 ;
6) calculating the physical layer resource block loss rate and the application layer packet loss rate of the current data packet according to the signal-to-noise ratio (SNR):
the block loss probability for a resource block with the selected modulation and coding mode h is calculated as follows:
where erfc (-) is a complementary error function, and b (h) and c (h) can be obtained in advance by means of fitting calculations;
packet loss probability of ith packet in nth frame of application layer video code streamWherein, BnumThe physical layer resource occupied by the current data packet is fast;
7) calculating distortion according to the coding structure, wherein the packet loss rate in the distortion calculation comprises the combination of the authentication probability and the packet loss probability, namely the probability that the current packet is successfully authenticated and does not lose the packet is
8) Calculating end-to-end distortion D in the current data packet transmission process:
D = 1 S &Sigma; i = 1 S E &lsqb; ( f n i - f ~ n i ) 2 = 1 S &Sigma; i = 1 S ( f n i ) 2 - 2 &CenterDot; f n i &CenterDot; E &lsqb; ( f ~ n i ) &rsqb; + E &lsqb; ( f ~ n i ) 2 &rsqb; ;
wherein, S is the number of the image contained in the current data packet in the current nth frame;
if the current frame belongs to the left viewpoint, a traditional distortion calculation method is adopted. The distortion is calculated by using a method in which the pixel is expected to recur. Distortion calculation for intra-frame predictive coding block
E &lsqb; f ~ n , l i &rsqb; = &theta; n , l i &CenterDot; ( f ~ n , l i ) + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; f ~ n - 1 , l k &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; f ~ n - 1 , l i &rsqb; ;
Image number distortion calculation for inter-predicted coding blocks
E &lsqb; f ~ n , l i &rsqb; = &theta; n , l i &CenterDot; ( e ^ n , l i + E &lsqb; f ~ n - 1 , l j &rsqb; ) + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; f ~ n - 1 , l k &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; f ~ n - 1 , l i &rsqb; ;
E &lsqb; ( f ~ n , l i ) 2 &rsqb; = &theta; n , l i &CenterDot; { ( e ^ n , l i ) 2 + 2 e ^ n , l i &CenterDot; E &lsqb; f ~ n - 1 , l j &rsqb; + E &lsqb; ( f ~ n - 1 , l j ) 2 &rsqb; } + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; ( f ~ n - 1 , l j ) 2 &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; ( f ~ n - 1 , l i ) 2 &rsqb; ;
And if the current frame belongs to the right viewpoint, distortion calculation is carried out according to different prediction relations: for theThe calculation of (a) can be classified into the following cases:
(1) the current pixel is located in the coding block by adopting intra-frame prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( f ~ n , r i ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(2) the coding block where the current pixel is located adopts inter prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + E &lsqb; f ~ n - 1 , r j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(3) the coding block where the current pixel is located adopts inter-view prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + E &lsqb; f ~ n , l j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(4) the weighted prediction of the inter prediction and the inter prediction employed by the coding block in which the current pixel is located,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + w 1 &CenterDot; E &lsqb; f ~ n - 1 , r o &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; ;
(1) the current pixel is in the coding block by adopting intra-frame prediction
(2) The coding block where the current pixel is located adopts inter prediction,
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
(3) the current pixel is in the coding block using inter-view prediction
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; E &lsqb; f ~ n , l j &rsqb; + E &lsqb; ( f ~ n , l j ) 2 &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
(4) The coding block where the current pixel is located adopts weighted prediction between views and frames,
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &CenterDot; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb; &ap; &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &rsqb; &CenterDot; E &lsqb; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &CenterDot; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb; &ap; &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &rsqb; &CenterDot; E &lsqb; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
wherein,is the value of the ith pixel in the nth frame,is the ith pixel value, E [. in the nth frame at the decoder end]In order to be a function of the expectation,for the ith pixel in the nth frame of the left view at the decoder endThe value of the one or more of,for the ith pixel value in the (n-1) th frame of the left view at the decoder side,is the error concealment value of the ith pixel in the (n-1) th frame of the left view at the decoder end,for the reconstructed value of the ith pixel in the nth frame at the encoder end,is a pixelThe probability of the authentication being successful is determined,is a pixelThe probability of the authentication being successful is determined,is a pixelThe coded prediction residual of (2) is,is a pixelThe coded prediction value of (a) is,for the ith pixel value of the nth frame of the right view at the decoder end,for the ith pixel value in the n-1 th frame of the right view at the decoder side,for the error concealment value of the ith pixel of the (n-1) th frame of the right view at the decoder end,is the reconstructed value of the ith pixel in the nth frame of the right view at the encoder end,is a pixelThe probability of the authentication being successful is determined,is a pixelThe probability of the authentication being successful is determined,is a pixelThe residual values of the encoding of (1),is a pixelThe coded prediction value of (a) is,is a pixelA coding prediction value in a left view; the data can be directly obtained in the encoding process;andobtained by step 7;
9) if minD > D, minD is equal to D, and m _ opt is equal to m, MCS _ opt is equal to MCS, QP _ opt is equal to QP, and MCS is equal to MCS + 1;
10) if the MCS is less than or equal to maxMCS, returning to the step 6;
11) if m is less than or equal to maxH, returning to the step 4;
12) QP + 1; if the QP is less than or equal to maxQP, returning to the step 3;
13) and outputting m _ opt, MCS _ opt and QP _ opt, and using the parameters to encode the current 3D video data packet, configure the number of Hash authentication father nodes, select a physical layer MCS and perform wireless transmission.
And selecting one number of the Hash chains according to the number of the candidate Hash chains, and then estimating the corresponding probability of successful authentication. Then, the number of MCS modes of the physical layer candidate is determined, one MCS is selected, an equivalent SNR is determined, and then the block loss rate corresponding to the current MCS is estimated. According to the quantization parameter of the video coding, the alternative video coding code rate of the current data packet is obtained, a Quantization Parameter (QP) is selected, and the current packet loss probability is estimated. And obtaining the current coding QP, the current authentication Hash quantity and the end-to-end video distortion under the current MCS mode through recursive calculation. And performing application layer coding and authentication by traversing and calculating end-to-end distortion under the combination condition of all QP, authentication Hash quantity and MCS, selecting the QP, the authentication Hash quantity and the MCS under the condition of minimum distortion, configuring the MCS of a physical layer, and transmitting the current data packet. The method can configure and coordinate the intensity of different parameters on the aspects of authentication and fault tolerance, and can maximize the quality of the 3D medical video under the current network state.
The present invention and its embodiments have been described above, and the description is not intended to be limiting, and the drawings are only one embodiment of the present invention, and the actual structure is not limited thereto. In summary, those skilled in the art should appreciate that they can readily use the disclosed conception and specific embodiments as a basis for designing or modifying other structures for carrying out the same purposes of the present invention without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (1)

1. A3D medical video transmission method under a wireless network environment is characterized in that: the method comprises the following specific steps:
1) according to the selection of the number of the wireless 3D medical video hash capable of fault-tolerant authentication, the selection of an application layer video coding quantization parameter and the selection of an LTE downlink physical layer modulation and coding mode, the following optimal parameters are comprehensively established by minimizing end-to-end 3D video distortion:
( m h a s h ( d 0 , ... d m - 1 ) o p t , QP o p t , MCS o p t ) = arg min m &Element; H , Q P &Element; Q , M C S &Element; Z D ( m , Q P , M C S )
subjecttot<Tmax
wherein m ishash(d0,...dm-1)optIs the optimal Hash father node of the current data packet, H, Q, and Z respectively represent the candidate Hash father node set, QP set and MCS mode set, QP is the quantization parameter of the coding strip, MCS is the physical layer modulation and coding mode, TmaxThe maximum delay limit of a data packet transmission, T, index the transmission delay of the data packet, can be obtained by dividing the size of the data packet by the physical layer transmissionThe output rate is approximately obtained;
2) setting parameters such as initial QP and MCS and the like: setting QP to minQP, minD to 0, maxMCS to 15, and maxQP to 45;
3) aiming at a 3D medical video acquired by a binocular camera, coding a coding structure with inter-view prediction according to multi-view video coding, and coding by using QP as a quantization parameter;
4) setting the data volume of a Hash father node to be 0, namely, m is 0, obtaining the maximum possible Hash node quantity of the current data packet to be maxH, and obtaining the maximum possible Hash node quantity of the current data packet according to a formulaCalculating the probability of successful authentication of the current ith data packet, wherein the current ith (i) th data packet<maxH) number of hash parent node selected by packet is m (m)<maxH) and the tth parent node is a distance d from the current packettThe probability of successful authentication isThe probability of packet loss of the current data transmission application layer is assumed to be rhoi
5) Setting the coding and modulation mode MCS of the physical layer channel to be 1; for each modulation and coding mode h, calculating the equivalent signal-to-noise ratio &gamma; m i e f f ( h ) = &kappa; ( h ) &lsqb; J - 1 ( 1 N s b &Sigma; j = 1 N s b J ( &gamma; j &kappa; ( h ) ) ) &rsqb; 2 ;
Wherein N issbRepresenting the number of sub-carriers, gammajIs the signal to interference plus noise ratio of the jth sub-carrier, k (h) is the correction factor in modulation and coding mode h, J (x) and J-1(y) is calculated as follows:
J ( x ) &ap; - 0.04210610 x 3 + 0.209252 x 2 - 0.00640081 x , 0 < x < 1.6363 1 - exp ( 0.00181491 x 3 - 0.142675 x 2 - 0.08220540 x + 0.0549608 ) , x &GreaterEqual; 1.6363
J - 1 ( y ) &ap; 1.09542 y 2 + 0.214217 y + 2.33727 y , 0 < y < 0.3646 - 0.706692 l o g ( - 0.386013 ( y - 1 ) ) + 1.75017 y , 0.3646 &le; y &le; 1 ;
6) calculating the physical layer resource block loss rate and the application layer packet loss rate of the current data packet according to the signal-to-noise ratio (SNR):
the block loss probability for a resource block with the selected modulation and coding mode h is calculated as follows:
where erfc (-) is a complementary error function, and b (h) and c (h) can be obtained in advance by means of fitting calculations;
packet loss probability of ith packet in nth frame of application layer video code streamWherein, BnumThe physical layer resource occupied by the current data packet is fast;
7) calculating distortion from coding structure, distortion meterThe packet loss rate in the calculation includes the combination of the authentication probability and the packet loss probability, that is, the probability that the current packet is successfully authenticated and does not lose packet is
8) Calculating end-to-end distortion D in the current data packet transmission process:
D = E &lsqb; ( f n i - f ~ f i ) 2 = ( f n i ) 2 - 2 &CenterDot; f n i &CenterDot; E &lsqb; ( f ~ n i ) + E &lsqb; ( f ~ n i ) 2 &rsqb; ;
wherein, S is the number of the image contained in the current data packet in the current nth frame;
if the current frame belongs to the left viewpoint, a traditional distortion calculation method is adopted. The distortion is calculated by using a method in which the pixel is expected to recur. Distortion calculation for intra-frame predictive coding block
E &lsqb; f ~ n , l i &rsqb; = &theta; n , l i &CenterDot; ( f ~ n , l i ) + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; f ~ n - 1 , l k &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; f ~ n - 1 , l i &rsqb; ;
Image number distortion calculation for inter-predicted coding blocks
E &lsqb; f ~ n , l i &rsqb; = &theta; n , l i &CenterDot; ( e ^ n , l i + E &lsqb; f ~ n - 1 , l j &rsqb; ) + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; f ~ n - 1 , l k &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; f ~ n - 1 , l i &rsqb; ;
E &lsqb; ( f ~ n , l i ) 2 &rsqb; = &theta; n , l i &CenterDot; { ( e ^ n , l i ) 2 + 2 e ^ n , l i &CenterDot; E &lsqb; f ~ n - 1 , l j &rsqb; + E &lsqb; ( f ~ n - 1 , l j ) 2 &rsqb; } + ( 1 - &theta; n , l i ) &CenterDot; &theta; n - 1 , l i &CenterDot; E &lsqb; ( f ~ n - 1 , l k ) 2 &rsqb; + ( 1 - &theta; n , l i ) &CenterDot; ( 1 - &theta; n - 1 , l i ) &CenterDot; E &lsqb; ( f ~ n - 1 , l i ) 2 &rsqb; ;
And if the current frame belongs to the right viewpoint, distortion calculation is carried out according to different prediction relations: for theThe calculation of (a) can be classified into the following cases:
(1) the current pixel is coded using intra prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( f ~ n , r i ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(2) the coding block where the current pixel is located adopts inter prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + E &lsqb; f ~ n - 1 , r j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(3) the coding block where the current pixel is located adopts inter-view prediction,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + E &lsqb; f ~ n , l j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r i &rsqb; ;
(4) the weighted prediction of the inter prediction and the inter prediction employed by the coding block in which the current pixel is located,
E &lsqb; f ~ n , r i &rsqb; = &theta; n , r i &CenterDot; ( e ^ n , r i + w 1 &CenterDot; E &lsqb; f ~ n - 1 , r o &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; f ~ n - 1 , r k &rsqb; ;
(1) the current pixel is in the coding block by adopting intra-frame prediction
(2) The coding block where the current pixel is located adopts inter prediction,
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
(3) the current pixel is in the coding block using inter-view prediction
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; E &lsqb; f ~ n , l j &rsqb; + E &lsqb; ( f ~ n , l j ) 2 &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
(4) The coding block where the current pixel is located adopts weighted prediction between views and frames,
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &CenterDot; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb; &ap; &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &rsqb; &CenterDot; E &lsqb; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
E &lsqb; ( f ~ n , r i ) 2 &rsqb; = &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &CenterDot; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb; &ap; &theta; n , r i &CenterDot; { ( e ^ n , r i ) 2 + 2 e ^ n , r i &CenterDot; ( w 1 &CenterDot; E &lsqb; f ~ n - 1 , r j &rsqb; + w 2 &CenterDot; E &lsqb; f ~ n , l j &rsqb; ) + w 1 2 &CenterDot; E &lsqb; ( f ~ n - 1 , r j ) 2 &rsqb; + w 2 2 E &lsqb; ( f ~ n , l j ) 2 &rsqb; + 2 w 1 w 2 E &lsqb; f ~ n - 1 , r j &rsqb; &CenterDot; E &lsqb; f ~ n , l j &rsqb; } + ( 1 - &theta; n , r i ) &CenterDot; &theta; n - 1 , r i &CenterDot; E &lsqb; ( f ~ n - 1 , r k ) 2 &rsqb; + ( 1 - &theta; n , r i ) &CenterDot; ( 1 - &theta; n - 1 , r i ) &CenterDot; E &lsqb; ( f ~ n - 1 , r i ) 2 &rsqb;
wherein,is the value of the ith pixel in the nth frame,is the ith pixel value, E [. in the nth frame at the decoder end]In order to be a function of the expectation,for the ith pixel value in the nth frame of the left view at the decoder side,for the ith pixel value in the (n-1) th frame of the left view at the decoder side,is the error concealment value of the ith pixel in the (n-1) th frame of the left view at the decoder end,for the reconstructed value of the ith pixel in the nth frame at the encoder end,is a pixelThe probability of the authentication being successful is determined,is a pixelThe probability of the authentication being successful is determined,is a pixelThe coded prediction residual of (2) is,is a pixelThe coded prediction value of (a) is,for the ith pixel value of the nth frame of the right view at the decoder end,for the ith pixel value in the n-1 th frame of the right view at the decoder side,for the error concealment value of the ith pixel of the (n-1) th frame of the right view at the decoder end,is the reconstructed value of the ith pixel in the nth frame of the right view at the encoder end,is a pixelThe probability of the authentication being successful is determined,is a pixelThe probability of the authentication being successful is determined,is a pixelThe residual values of the encoding of (1),is a pixelThe coded prediction value of (a) is,is a pixelA coding prediction value in a left view; the data can be directly obtained in the encoding process;andobtained by step 7;
9) if minD > D, minD is equal to D, and m _ opt is equal to m, MCS _ opt is equal to MCS, QP _ opt is equal to QP, and MCS is equal to MCS + 1;
10) if the MCS is less than or equal to maxMCS, returning to the step 6;
11) if m is less than or equal to maxH, returning to the step 4;
12) QP + 1; if the QP is less than or equal to maxQP, returning to the step 3;
13) and outputting m _ opt, MCS _ opt and QP _ opt, and using the parameters to encode the current 3D video data packet, configure the number of Hash authentication father nodes, select a physical layer MCS and perform wireless transmission.
CN201610004628.8A 2016-01-04 2016-01-04 A kind of 3D medical video transmission methods under wireless network environment Expired - Fee Related CN105657399B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610004628.8A CN105657399B (en) 2016-01-04 2016-01-04 A kind of 3D medical video transmission methods under wireless network environment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610004628.8A CN105657399B (en) 2016-01-04 2016-01-04 A kind of 3D medical video transmission methods under wireless network environment

Publications (2)

Publication Number Publication Date
CN105657399A true CN105657399A (en) 2016-06-08
CN105657399B CN105657399B (en) 2017-11-17

Family

ID=56490488

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610004628.8A Expired - Fee Related CN105657399B (en) 2016-01-04 2016-01-04 A kind of 3D medical video transmission methods under wireless network environment

Country Status (1)

Country Link
CN (1) CN105657399B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112311727A (en) * 2019-07-29 2021-02-02 杭州海康威视数字技术股份有限公司 Control method and system for wireless video transmission

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102355582A (en) * 2011-09-23 2012-02-15 宁波大学 Method for optimizing rate distortion model for three-dimensional video coding
CN103108197A (en) * 2011-11-14 2013-05-15 辉达公司 Priority level compression method and priority level compression system for three-dimensional (3D) video wireless display
CN103856779A (en) * 2014-03-18 2014-06-11 天津大学 Packet switching network oriented multi-view video transmission distortion predication method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102355582A (en) * 2011-09-23 2012-02-15 宁波大学 Method for optimizing rate distortion model for three-dimensional video coding
CN103108197A (en) * 2011-11-14 2013-05-15 辉达公司 Priority level compression method and priority level compression system for three-dimensional (3D) video wireless display
CN103856779A (en) * 2014-03-18 2014-06-11 天津大学 Packet switching network oriented multi-view video transmission distortion predication method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
刘金霞等: "基于跨层码率适配和差错控制的3D视频无线传输方法", 《电子学报》 *
陈莲娜等: "基于SSIM的低复杂度跨层优化无线视频传输方法", 《计算机应用研究》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112311727A (en) * 2019-07-29 2021-02-02 杭州海康威视数字技术股份有限公司 Control method and system for wireless video transmission
CN112311727B (en) * 2019-07-29 2022-05-24 杭州海康威视数字技术股份有限公司 Control method and system for wireless video transmission

Also Published As

Publication number Publication date
CN105657399B (en) 2017-11-17

Similar Documents

Publication Publication Date Title
CN111837396B (en) Error suppression in video coding based on sub-image code stream view correlation
US10397603B1 (en) Method and apparatus for video coding
US11748729B2 (en) Method and apparatus for video coding
US9202289B2 (en) Method for coding and decoding target block partition information using information about neighboring blocks
US11323731B2 (en) Method and apparatus for video coding
WO2018095328A1 (en) System and Method for Coding Intra Prediction Mode using a Second Set of Most Probable Modes
JP5768332B2 (en) Transmitter, receiver and communication system
JP2013523010A5 (en) Method and apparatus for implicit adaptive motion vector predictor selection for video encoding and video decoding
EP3541076B1 (en) Error resilient rate distortion optimization for image and video encoding
US20200021836A1 (en) Method and apparatus for ordering and selection of affine merge candidates in motion compensation
KR102494087B1 (en) Method and apparatus for sub-block motion vector prediction
CN101888566A (en) Estimation method of distortion performance of stereo video encoding rate
US20200280733A1 (en) Method and apparatus for video coding
CN101355707B (en) Error control method for demixing and encoding stereo video
Cen et al. Interview motion compensated joint decoding for compressively sampled multiview video streams
KR102105323B1 (en) A method for adaptive illuminance compensation based on object and an apparatus using it
CN106657961B (en) Hybrid digital-analog encoding of stereoscopic video
KR20230047446A (en) Entropy Coding for Intra Prediction Modes
CN105657399B (en) A kind of 3D medical video transmission methods under wireless network environment
Liu et al. Multiple description coding of free viewpoint video for multi-path network streaming
US20230156185A1 (en) Generalized sample offset
US20230156232A1 (en) Adaptive application of generalized sample offset
Pang et al. Relativity analysis-based error concealment algorithm for entire frame loss of stereo video
Liu et al. Cross-layer optimized authentication and error control for wireless 3D medical video streaming over LTE
US12114009B2 (en) Method and apparatus for adaptive reordering for reference frames

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20171117