CN117313869A - Large model privacy protection reasoning method based on model segmentation - Google Patents

Large model privacy protection reasoning method based on model segmentation Download PDF

Info

Publication number
CN117313869A
CN117313869A CN202311418709.9A CN202311418709A CN117313869A CN 117313869 A CN117313869 A CN 117313869A CN 202311418709 A CN202311418709 A CN 202311418709A CN 117313869 A CN117313869 A CN 117313869A
Authority
CN
China
Prior art keywords
model
client
server
segmentation
reasoning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311418709.9A
Other languages
Chinese (zh)
Other versions
CN117313869B (en
Inventor
乔一帆
邵硕
秦湛
王志波
任奎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202311418709.9A priority Critical patent/CN117313869B/en
Publication of CN117313869A publication Critical patent/CN117313869A/en
Application granted granted Critical
Publication of CN117313869B publication Critical patent/CN117313869B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • G06F21/6245Protecting personal data, e.g. for financial or medical purposes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • G06N3/0455Auto-encoder networks; Encoder-decoder networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses a large model privacy protection reasoning method based on model segmentation, and belongs to the technical field of computer artificial intelligence and large model security. Segmentation by model: deploying the Encoder and the Decode of the original large model at the client, and leaving the middle part of the large model at the local of the server; model compression: the middle layer is compressed through the server side and then sent to the client side, and a small model with the basic function of the original large model is formed at the client side; fine tuning of the model: the client side fine-tunes the model through a loss function; model reasoning: the client sends the trained Encoder to the server according to the protocol, the intermediate result is obtained by reasoning, and the training is completed by combining with the Decoder of the local server. The invention gives consideration to the performance and privacy protection of the model, effectively prevents reconstruction attack, and has no negative influence on the effect of the model; meanwhile, model privacy leakage of a large model and data leakage of a user are prevented, the calculation efficiency is high, and a large amount of calculation resources are not needed at a client.

Description

Large model privacy protection reasoning method based on model segmentation
Technical Field
The invention relates to the technical field of computer artificial intelligence and large model security, in particular to a large model privacy protection reasoning method based on model segmentation.
Background
The large model privacy protection reasoning based on model segmentation has very important application in the artificial intelligence field and the large model security field. Model segmentation techniques refer to the segmentation of a complete neural network model into two or more sub-modules, which are then processed separately to accomplish different tasks. The core idea is to make the model easier to understand, optimize and debug through the modular design. Model segmentation technology originated in the 90 s of the 20 th century, and early work mainly adopted a simple series module connection mode. Into the 21 st century, more complex tree structures and multi-branch connections were proposed. Today, large models of artificial intelligence, which are neural network models of extremely large parameter scale, typically with billions to billions of parameters, are in the fast-evolving phase, and universal language or visual capabilities are obtained by pre-training on massive data. Early large models were GPT series, BERT, etc. language models and Vision Transformer visual models. The parameters of the last two years have been increasing explosively, and models of the billion parameter level such as GPT-3 and Switch Transformer are presented, and the parameters are expected to continue to increase rapidly with further increase in computing power.
Nowadays, with the development of deep learning and large models, model segmentation techniques are widely used in fields such as natural language processing and computer vision. The model segmentation technology reduces training difficulty through modularized design, improves the capability of the model to adapt to new tasks, and is an important technical means for realizing migration learning. With the continuous development of the transfer learning technology, the emerging Offsite-Tuning technology adopts a privacy protection method, so that a data owner does not need to share sensitive data of the data owner to a model owner. Traditional migration learning methods may require the data owners to share their data and pay an expensive fee in order for the model owners to be able to make full fine-Tuning, while Offsite-Tuning reduces the need to share the data by sending lightweight adapters and simulators to the data owners, thereby reducing costs. For large base models, offsite-Tuning is more computationally efficient, and the data owner only needs to fine tune the adapter locally, without accessing the full model weights, thus saving a lot of computation time and resources.
However, the migration learning also has some security problems, such as model leakage by using model extraction for the countermeasure sample attack of the source model or the target model, and privacy leakage risks possibly existing in the sample data of the source domain and the target domain, so the invention provides a large model privacy protection reasoning method based on model segmentation to solve the problems.
Disclosure of Invention
The invention aims to provide a large model privacy protection reasoning method based on model segmentation, which takes model performance and privacy protection into account, effectively prevents reconstruction attack, and has no negative influence on the effect of the model; meanwhile, the model segmentation technology prevents model privacy leakage of a large model and data leakage of a user, so that the calculation efficiency is high, and a large amount of calculation resources are not needed at a client.
In order to achieve the above purpose, the invention adopts the following technical scheme:
a large model privacy protection reasoning method based on model segmentation comprises the following steps:
s1, model segmentation: deploying the front n layers of encoders and the last n layers of encoders of the original large model at the client, and leaving the middle part of the large model at the local of the server;
s2, model compression: the middle layer is compressed through the server side and then sent to the client side, and a small model with the basic function of the original large model is formed at the client side;
s3, fine adjustment of a model: the client side fine-tunes the model through a loss function;
s4, model reasoning: the client sends the trained front n layers of encodings to the server according to the protocol, then performs reasoning to obtain an intermediate result, and then completes training by combining with the last n layers of encodings of the local server to obtain a final result.
Preferably, in step S1, the original data of the client is used to train n model layers without leaving the local area, and the trained Encoder and the Encoder are obtained by training the n layers before the client completes the large model and training the last n layers.
Preferably, in step S2, the remaining intermediate layer in step S1 is compressed to form a simulator module for providing an approximate gradient direction in the adaptation process, the simulator module is sent from the server side to the client side, and the Encoder and the Decoder deployed in the client side and the original large model are combined to form a complete small model.
Preferably, in step S3, the small model obtained in step S2 is trimmed by a loss function, where the loss function L is:
wherein L is 1 Is the loss function of the original large model task, L 2 Is f 2 、f' 2 、f' n-2 Cosine distance f of (f) 2 As an intermediate feature, f' 2 Is f n-2 Intermediate features to the original large model.
Preferably, in step S4, the Encoder performs task fine tuning on the client, then combines with the server, performs joint training with the rest of the large model, coordinates and accords with model parameters between the server and the client, and finally sends a trained intermediate result back to the client, combines with the trained Encoder, refines the model, and obtains final output.
Therefore, the large model privacy protection reasoning method based on model segmentation has the following beneficial effects:
1. according to the invention, the model privacy of the server side and the data privacy of the user side can be simultaneously protected by deploying part of the large model on the local side of the client side and training the part of the large model.
2. The middle part of the large model is compressed into a simulator by locally utilizing a fine tuning method of a model compression technology at the client side, so that the client side is assisted to locally perform fine tuning of the Encoder and the Decoder, meanwhile, the loss functions of model performance and privacy protection are balanced, the performance in fine tuning of the model is not reduced, and reconstruction attacks are prevented.
The technical scheme of the invention is further described in detail through the drawings and the embodiments.
Drawings
FIG. 1 is a flow chart of the present invention.
Detailed Description
The technical scheme of the invention is further described below through the attached drawings and the embodiments.
Unless defined otherwise, technical or scientific terms used herein should be given the ordinary meaning as understood by one of ordinary skill in the art to which this invention belongs. The terms "first," "second," and the like, as used herein, do not denote any order, quantity, or importance, but rather are used to distinguish one element from another. The word "comprising" or "comprises", and the like, means that elements or items preceding the word are included in the element or item listed after the word and equivalents thereof, but does not exclude other elements or items. The terms "disposed," "mounted," "connected," and "connected" are to be construed broadly, and may be, for example, fixedly connected, detachably connected, or integrally connected; can be mechanically or electrically connected; can be directly connected or indirectly connected through an intermediate medium, and can be communication between two elements. "upper", "lower", "left", "right", etc. are used merely to indicate relative positional relationships, which may also be changed when the absolute position of the object to be described is changed.
Examples
As shown in FIG. 1, the invention provides a large model privacy protection reasoning method based on model segmentation, which comprises the following steps:
1. (step S1) model segmentation: the first n layers of encodings and the last n layers of encodings of the original large model are deployed at the client, and the middle part of the large model is left at the local of the server.
The original large model comprises a transducer model, a Seq2Seq model and the like; the first n layers are set to w 1 、w 1 Etc.; the last n layers are set to w n-1 、w n Etc.
The training of the front n layers and the last n layers is completed at the client to obtain the trained Encoder and the trained Decoder, and the strategy ensures that the data of the client are invisible to the server in the training process, and the original data of the client are not separated from the local and are only used for training the n key model layers. Sensitive data can be protected, the sensitive data cannot be exposed in network transmission, and the risk of data leakage is reduced. Only the first n layers and the last n layers need to be trained at the client, compared with the whole large model, the required computing resources are greatly reduced, and even on equipment with limited resources, the training can be easily performed without huge computing clusters, so that the cost and complexity of the training process are reduced by the effective computing resource utilization mode.
2. (step S2) model compression: and compressing the middle layer through the server side, sending the compressed middle layer to the client side, and forming a small model with the basic function of the original large model at the client side.
1) The server compresses the remaining intermediate layer R of step S1 to form a simulator module E for providing an approximate gradient direction during the adaptation process, which contains the main functional information of the original model and is a fixed untrainable part.
2) The building process of the simulator module occurs at the server side, wherein the middle layer is processed by a well-designed compression algorithm to ensure that the size of the simulator is minimized while retaining the key model functional characteristics. This compression process is lossy, and aims to preserve the main information of the model, while reducing the size of the simulator module so that it can be efficiently transmitted over the network.
3) And sending the obtained simulator module to a client from a server, and combining the client with the original large model deployed Encoder and Decoder to form a complete small model. The small model has enough performance and functions to complete specific tasks, and is assisted by a simulator module, and a user adjusts an Encoder and a Decode by using own data with the assistance of the small model.
3. (step S3) fine tuning of the model: the client fine-tunes the model through the loss function.
And (3) fine tuning the small model obtained in the step (S2) by using a loss function, wherein the specific formula is as follows:
wherein L is 1 Is the loss function of the original large model task, L 2 Is f 2 、f' 2 、f' n-2 Cosine distance f of (f) 2 As an intermediate feature, f' 2 Is f n-2 Intermediate features to the original large model.
1) Loss function L of original large model task 1 Generally, depending on the nature of the particular task, may be cross entropy loss, maximum likelihood loss, or other loss function suitable for the task, L 1 The function of (2) is to ensure that the fine-tuned small model keeps high performance on the original task, i.e. the performance of the model is not degraded, and the part of the loss function ensures the effectiveness of the model on the task.
2) Cosine distance L 2 Similarity of models in intermediate feature space is measured by minimizing L 2 Can be protected against reconstruction attacks (i.e. an attacker tries to reconstruct the original data from the intermediate features), and L 2 The introduction of loss enhances the safety of the model, and ensures that the intermediate representation of the model is not easy to leak sensitive information.
4. (step S4) model reasoning: the client sends the trained front n layers of encodings to the server according to the protocol, then performs reasoning to obtain an intermediate result, and then completes training by combining with the last n layers of encodings of the local server to obtain a final result.
The client sends the adjusted and trained Encoder to the server, and the Encoder is adjusted by the task to adapt to the specific application scene or task requirement; the server-side will combine with this Encoder and co-train with the rest of the large model. This process ensures consistent model parameters between the server and client, thus maintaining the performance and functionality of the model to the greatest extent.
The server side sends trained intermediate results back to the client side, wherein the intermediate results comprise further training results of the model by the server side and possible performance improvement; the client combines these intermediate results with the trained encodings of its own, further refining the model, resulting in a final output. Step S4 is the last loop of the overall process, aimed at ensuring that the model reaches an optimal performance level with the cooperation of the server side and the client side.
Therefore, the invention adopts a large model privacy protection reasoning method based on model segmentation, combines model performance and privacy protection, effectively prevents reconstruction attack, and has no negative influence on the effect of the model; meanwhile, the model segmentation technology prevents model privacy leakage of a large model and data leakage of a user, so that the calculation efficiency is high, and a large amount of calculation resources are not needed at a client.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention and not for limiting it, and although the present invention has been described in detail with reference to the preferred embodiments, it will be understood by those skilled in the art that: the technical scheme of the invention can be modified or replaced by the same, and the modified technical scheme cannot deviate from the spirit and scope of the technical scheme of the invention.

Claims (5)

1. The large model privacy protection reasoning method based on model segmentation is characterized by comprising the following steps of:
s1, model segmentation: deploying the front n layers of encoders and the last n layers of encoders of the original large model at the client, and leaving the middle part of the large model at the local of the server;
s2, model compression: the middle layer is compressed through the server side and then sent to the client side, and a small model with the basic function of the original large model is formed at the client side;
s3, fine adjustment of a model: the client side fine-tunes the model through a loss function;
s4, model reasoning: the client sends the trained front n layers of encodings to the server according to the protocol, then performs reasoning to obtain an intermediate result, and then completes training by combining with the last n layers of encodings of the local server to obtain a final result.
2. The large model privacy preserving reasoning method based on model segmentation as set forth in claim 1, wherein: in step S1, the original data of the client is used to train n model layers without leaving the local area, and the trained Encoder and the Encoder are obtained by training the n layers before and the last n layers before the client completes the large model.
3. The large model privacy preserving reasoning method based on model segmentation as set forth in claim 2, wherein: in step S2, the rest intermediate layer in step S1 is compressed to form a simulator module for providing approximate gradient direction in the adapting process, the simulator module is sent to the client side from the server side, and the client side is combined with the Encoder and the Decode deployed by the original large model to form a complete small model.
4. A large model privacy preserving reasoning method based on model segmentation as claimed in claim 3, wherein: in step S3, fine tuning is performed on the small model obtained in step S2 through a loss function, where the loss function L is:
wherein L is 1 Is the loss function of the original large model task, L 2 Is f 2 、f' 2 、f' n-2 Cosine distance f of (f) 2 As an intermediate feature of the device,f' 2 is f n-2 Intermediate features to the original large model.
5. The large model privacy preserving reasoning method based on model segmentation as set forth in claim 4, wherein: in step S4, the Encoder performs task fine tuning on the client, then combines with the server, performs joint training with the rest of the large model, coordinates and accords with model parameters between the server and the client, and finally sends a trained intermediate result back to the client, combines with the trained Encoder, refines the model, and obtains final output.
CN202311418709.9A 2023-10-30 2023-10-30 Large model privacy protection reasoning method based on model segmentation Active CN117313869B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311418709.9A CN117313869B (en) 2023-10-30 2023-10-30 Large model privacy protection reasoning method based on model segmentation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311418709.9A CN117313869B (en) 2023-10-30 2023-10-30 Large model privacy protection reasoning method based on model segmentation

Publications (2)

Publication Number Publication Date
CN117313869A true CN117313869A (en) 2023-12-29
CN117313869B CN117313869B (en) 2024-04-05

Family

ID=89242714

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311418709.9A Active CN117313869B (en) 2023-10-30 2023-10-30 Large model privacy protection reasoning method based on model segmentation

Country Status (1)

Country Link
CN (1) CN117313869B (en)

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110942147A (en) * 2019-11-28 2020-03-31 支付宝(杭州)信息技术有限公司 Neural network model training and predicting method and device based on multi-party safety calculation
CN111832729A (en) * 2020-07-06 2020-10-27 东南数字经济发展研究院 Distributed deep learning reasoning deployment method for protecting data privacy
CN114140478A (en) * 2022-01-30 2022-03-04 电子科技大学 Federal learning method, system, device and medium for medical image segmentation
CN114723057A (en) * 2022-03-31 2022-07-08 北京理工大学 Neural network collaborative reasoning method for multi-access edge computing system
CN114912132A (en) * 2022-05-11 2022-08-16 南京大学 Method for realizing privacy protection convolutional neural network reasoning based on model conversion
CN115775010A (en) * 2022-11-23 2023-03-10 国网江苏省电力有限公司信息通信分公司 Electric power data sharing method based on horizontal federal learning
WO2023050754A1 (en) * 2021-09-30 2023-04-06 清华大学 Model training method and apparatus for private data set
CN116167084A (en) * 2023-02-24 2023-05-26 北京工业大学 Federal learning model training privacy protection method and system based on hybrid strategy
KR20230084407A (en) * 2021-12-03 2023-06-13 연세대학교 산학협력단 An artificial intelligence-based privacy-preserving distribution method for vertically, horizontally and multi-partitioned data and a device thereof
CN116582242A (en) * 2023-04-14 2023-08-11 南京大学 Safe federal learning method of ciphertext and plaintext hybrid learning mode
CN116579418A (en) * 2023-05-18 2023-08-11 杭州电子科技大学 Privacy data protection method for model segmentation optimization under federal edge learning environment
CN116739079A (en) * 2023-05-10 2023-09-12 浙江大学 Self-adaptive privacy protection federal learning method
CN116805082A (en) * 2023-08-23 2023-09-26 南京大学 Splitting learning method for protecting private data of client

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110942147A (en) * 2019-11-28 2020-03-31 支付宝(杭州)信息技术有限公司 Neural network model training and predicting method and device based on multi-party safety calculation
CN111832729A (en) * 2020-07-06 2020-10-27 东南数字经济发展研究院 Distributed deep learning reasoning deployment method for protecting data privacy
WO2023050754A1 (en) * 2021-09-30 2023-04-06 清华大学 Model training method and apparatus for private data set
KR20230084407A (en) * 2021-12-03 2023-06-13 연세대학교 산학협력단 An artificial intelligence-based privacy-preserving distribution method for vertically, horizontally and multi-partitioned data and a device thereof
CN114140478A (en) * 2022-01-30 2022-03-04 电子科技大学 Federal learning method, system, device and medium for medical image segmentation
CN114723057A (en) * 2022-03-31 2022-07-08 北京理工大学 Neural network collaborative reasoning method for multi-access edge computing system
CN114912132A (en) * 2022-05-11 2022-08-16 南京大学 Method for realizing privacy protection convolutional neural network reasoning based on model conversion
CN115775010A (en) * 2022-11-23 2023-03-10 国网江苏省电力有限公司信息通信分公司 Electric power data sharing method based on horizontal federal learning
CN116167084A (en) * 2023-02-24 2023-05-26 北京工业大学 Federal learning model training privacy protection method and system based on hybrid strategy
CN116582242A (en) * 2023-04-14 2023-08-11 南京大学 Safe federal learning method of ciphertext and plaintext hybrid learning mode
CN116739079A (en) * 2023-05-10 2023-09-12 浙江大学 Self-adaptive privacy protection federal learning method
CN116579418A (en) * 2023-05-18 2023-08-11 杭州电子科技大学 Privacy data protection method for model segmentation optimization under federal edge learning environment
CN116805082A (en) * 2023-08-23 2023-09-26 南京大学 Splitting learning method for protecting private data of client

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
JIUYUN XU 等: "IFTS: A Location Privacy Protection Method Based on Initial and Final Trajectory Segments", 《DIGITAL OBJECT IDENTIFIER》, vol. 9, 1 February 2021 (2021-02-01), pages 18112 - 18122, XP011834453, DOI: 10.1109/ACCESS.2021.3052169 *
任奎 等: "人工智能模型数据泄露的攻击与防御研究综述", 《网络与信息安全学报》, vol. 7, no. 1, 28 February 2021 (2021-02-28), pages 1 - 10 *
周俊 等: "联邦学习安全与隐私保护研究综述", 《西华大学学报(自然科学版)》, no. 04, 10 July 2020 (2020-07-10), pages 21 - 29 *
张小玉 等: "基于属性分割的差分隐私异构多属性数据发布", 《计算机系统应用》, vol. 31, no. 10, 7 July 2022 (2022-07-07), pages 225 - 235 *

Also Published As

Publication number Publication date
CN117313869B (en) 2024-04-05

Similar Documents

Publication Publication Date Title
Krishnaraj et al. Deep learning model for real-time image compression in Internet of Underwater Things (IoUT)
Sun et al. Deep pixel‐to‐pixel network for underwater image enhancement and restoration
CN113592718A (en) Mine image super-resolution reconstruction method and system based on multi-scale residual error network
CN113315972B (en) Video semantic communication method and system based on hierarchical knowledge expression
Dong et al. Semantic communication system based on semantic slice models propagation
CN110889416B (en) Salient object detection method based on cascade improved network
CN111681154A (en) Color image steganography distortion function design method based on generation countermeasure network
Zhang et al. Optical image compression and encryption transmission-based ondeep learning and ghost imaging
Wang et al. A survey on digital image steganography
CN116958534A (en) Image processing method, training method of image processing model and related device
CN113076957A (en) RGB-D image saliency target detection method based on cross-modal feature fusion
CN115766159A (en) Private data processing method and device and electronic equipment
CN116258757A (en) Monocular image depth estimation method based on multi-scale cross attention
JP2023001926A (en) Method and apparatus of fusing image, method and apparatus of training image fusion model, electronic device, storage medium and computer program
CN113595993A (en) Vehicle-mounted sensing equipment joint learning method for model structure optimization under edge calculation
CN116580184A (en) YOLOv 7-based lightweight model
CN117313869B (en) Large model privacy protection reasoning method based on model segmentation
Wang et al. Data hiding during image processing using capsule networks
CN115761242B (en) Denoising method and terminal based on convolutional neural network and fuzzy image characteristics
CN116977169A (en) Data processing method, apparatus, device, readable storage medium, and program product
Nakahara et al. Edge Computing-Assisted DNN Image Recognition System With Progressive Image Retransmission
CN114694065A (en) Video processing method, device, computer equipment and storage medium
CN114419517A (en) Video frame processing method and device, computer equipment and storage medium
Zou et al. AF-net: All-scale feature fusion network for road extraction from remote sensing images
Ren et al. Asymmetric Semantic Communication System Based on Diffusion Model in IoT

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant