WO2022001092A1 - 一种数据处理方法、装置及设备 - Google Patents
一种数据处理方法、装置及设备 Download PDFInfo
- Publication number
- WO2022001092A1 WO2022001092A1 PCT/CN2021/073615 CN2021073615W WO2022001092A1 WO 2022001092 A1 WO2022001092 A1 WO 2022001092A1 CN 2021073615 W CN2021073615 W CN 2021073615W WO 2022001092 A1 WO2022001092 A1 WO 2022001092A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- baseline model
- target
- initial
- edge server
- model
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 17
- 238000000034 method Methods 0.000 claims abstract description 57
- 230000004927 fusion Effects 0.000 claims abstract description 38
- 238000012549 training Methods 0.000 claims abstract description 26
- 238000012545 processing Methods 0.000 claims description 33
- 238000006243 chemical reaction Methods 0.000 claims description 30
- 230000008569 process Effects 0.000 claims description 23
- 230000006835 compression Effects 0.000 claims description 12
- 238000007906 compression Methods 0.000 claims description 12
- 230000006837 decompression Effects 0.000 claims description 8
- 238000012546 transfer Methods 0.000 claims description 6
- 230000009471 action Effects 0.000 claims description 2
- 238000013528 artificial neural network Methods 0.000 description 45
- 238000013473 artificial intelligence Methods 0.000 description 21
- 230000006870 function Effects 0.000 description 18
- 238000012360 testing method Methods 0.000 description 17
- 238000010586 diagram Methods 0.000 description 15
- 238000010801 machine learning Methods 0.000 description 8
- 238000004590 computer program Methods 0.000 description 7
- 238000011176 pooling Methods 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 4
- 238000013527 convolutional neural network Methods 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 230000005284 excitation Effects 0.000 description 4
- 230000009466 transformation Effects 0.000 description 4
- 238000013478 data encryption standard Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000012935 Averaging Methods 0.000 description 2
- 230000004913 activation Effects 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 108091028043 Nucleic acid sequence Proteins 0.000 description 1
- 239000006227 byproduct Substances 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000013178 mathematical model Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000012163 sequencing technique Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/60—Protecting data
- G06F21/62—Protecting access to data via a platform, e.g. using keys or access control rules
- G06F21/6218—Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
- G06F21/6245—Protecting personal data, e.g. for financial or medical purposes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/625—License plates
Definitions
- the present application relates to the field of artificial intelligence, and in particular, to a data processing method, apparatus and device.
- Machine learning is a way to realize artificial intelligence. It is a multi-field interdisciplinary subject involving probability theory, statistics, approximation theory, convex analysis, algorithm complexity theory and other disciplines. Machine learning is used to study how computers simulate or implement human learning behaviors to acquire new knowledge or skills, and to reorganize existing knowledge structures to continuously improve their performance. Machine learning pays more attention to algorithm design, enabling computers to automatically learn rules from data and use rules to predict unknown data.
- Machine learning has a wide range of applications, such as deep learning, data mining, computer vision, natural language processing, biometric recognition, search engines, medical diagnosis, detection of credit card fraud, stock market analysis, DNA sequence sequencing, speech and handwriting recognition , strategy games and robot use.
- the application provides a data processing method, the method includes:
- the central server sends the initial baseline model to the edge server
- the edge server trains the initial baseline model through the scene data of the edge server, obtains a target baseline model, and determines whether to deploy the target baseline model;
- the edge server sends the target baseline model to the central server
- the central server generates a fused baseline model based on the target baseline model and the initial baseline model, trains the fused baseline model, determines the trained baseline model as the initial baseline model, and returns to the execution center server to store the initial baseline model. Action sent to the edge server.
- the present application provides a data processing method, which is applied to an edge server, including:
- the present application provides a data processing method, which is applied to a central server, including:
- the fusion baseline model is trained, the trained baseline model is determined as the initial baseline model, and the operation of sending the initial baseline model to the edge server is returned to be executed.
- the present application provides a data processing device applied to an edge server, including:
- the acquisition module is used to acquire the initial baseline model from the central server;
- a training module configured to train the initial baseline model through the scene data of the edge server, obtain a target baseline model, and determine whether to deploy the target baseline model;
- a sending module configured to send the target baseline model to the central server when the target baseline model is not deployed, so that the central server generates a fusion baseline model based on the target baseline model and the initial baseline model , and re-acquire the initial baseline model based on the fusion baseline model.
- the present application provides a data processing device, which is applied to a central server, including:
- a sending module configured to send the initial baseline model to the edge server, so that the edge server trains the initial baseline model through the scene data of the edge server to obtain the target baseline model
- an obtaining module configured to obtain the target baseline model from the edge server
- the generating module is configured to generate a fusion baseline model based on the target baseline model and the initial baseline model; train the fusion baseline model, and determine the trained baseline model as the initial baseline model.
- the present application provides an edge server, including: a processor and a machine-readable storage medium, where the machine-readable storage medium stores machine-executable instructions that can be executed by the processor;
- the processor is configured to execute machine-executable instructions to implement the following steps:
- the present application provides a central server, including: a processor and a machine-readable storage medium, where the machine-readable storage medium stores machine-executable instructions that can be executed by the processor;
- the processor is configured to execute machine-executable instructions to implement the following steps:
- the fusion baseline model is trained, the trained baseline model is determined as the initial baseline model, and the operation of sending the initial baseline model to the edge server is returned to be executed.
- the edge server can obtain the initial baseline model from the central server, train the initial baseline model through the scene data, obtain the target baseline model, and send the target baseline model to the central server, and the central server is based on the target baseline model.
- the model and the initial baseline model generate a fused baseline model, train the fused baseline model, determine the trained baseline model as the initial baseline model, send the initial baseline model to the edge server, and so on. Since the above process can be performed cyclically, the performance of the initial baseline model and the target baseline model is continuously upgraded, and the recognition capabilities of the initial baseline model and the target baseline model are continuously improved, so that the target baseline model can achieve the expected performance and achieve high-precision recognition capabilities.
- the edge server can obtain a target baseline model with better performance, the target baseline model can match the environment where the edge server is located, and the accuracy of the intelligent analysis results is high.
- the edge server sends the target baseline model to the central server, rather than scene data (such as license plate images), so as to protect the privacy of the scene data, and will not send the scene data to the central server, so as to achieve deprivation of privacy. Data protection function to avoid sending private license plate images to the central server. Since the target baseline model is trained based on the scene data, the information of the scene data can be embodied in the initial baseline model and the target baseline model.
- FIG. 1 is a schematic structural diagram of a system in an embodiment of the present application.
- FIG. 2 is a schematic flowchart of a data processing method in an embodiment of the present application.
- FIG. 3 is a schematic structural diagram of a baseline model in an embodiment of the present application.
- FIG. 4 is a schematic flowchart of a data processing method in another embodiment of the present application.
- FIG. 5 is a schematic flowchart of a data processing method in another embodiment of the present application.
- 6A and 6B are schematic structural diagrams of a data processing apparatus in an embodiment of the present application.
- FIG. 7A is a hardware structure diagram of an edge server in an embodiment of the present application.
- FIG. 7B is a hardware structure diagram of a central server in an embodiment of the present application.
- first, second, third, etc. may be used in the embodiments of the present application to describe various information, the information should not be limited to these terms. These terms are only used to distinguish the same type of information from each other.
- the first information may also be referred to as the second information, and similarly, the second information may also be referred to as the first information without departing from the scope of the present application.
- the use of the word "if” can be interpreted as "at the time of" or "when” or “in response to determining”, depending on the context.
- Machine learning is a way to realize artificial intelligence, which is used to study how computers simulate or realize human learning behaviors to acquire new knowledge or skills, and to reorganize existing knowledge structures to continuously improve their performance.
- Deep learning is a subcategory of machine learning and is the process of using mathematical models to model specific problems in the real world in order to solve similar problems in the field.
- Neural network is the implementation of deep learning. For the convenience of description, this paper takes neural network as an example to introduce the structure and function of neural network. For other subclasses of machine learning, the structure and function of neural network are similar.
- Neural networks may include but are not limited to convolutional neural networks (CNN for short), recurrent neural networks (RNN for short), fully connected networks, etc.
- the structural units of neural networks may include but are not limited to convolutional layers (Conv), pooling Pool, excitation layer, fully connected layer (FC), etc., are not limited.
- one or more convolutional layers, one or more pooling layers, one or more excitation layers, and one or more fully connected layers can be combined to construct a neural network according to different requirements.
- the input data features are enhanced by performing convolution operations on the input data features by using a convolution kernel.
- the convolution kernel can be a matrix of size m*n, and the input data features of the convolution layer are By convolution with the convolution kernel, the output data features of the convolution layer can be obtained.
- the convolution operation is actually a filtering process.
- the input data features (such as the output of the convolution layer) are subjected to operations such as taking the maximum value, the minimum value, and the average value, so as to use the principle of local correlation to sub-sample the input data features.
- the pooling layer operation is actually a downsampling process.
- an activation function (such as a nonlinear function) can be used to map the input data features, thereby introducing nonlinear factors, so that the neural network can enhance the expressive ability through nonlinear combinations.
- the activation function may include, but is not limited to, a ReLU (Rectified Linear Units, rectified linear unit) function, where the ReLU function is used to set the features smaller than 0 to 0, while the features larger than 0 remain unchanged.
- all data features input to the fully connected layer are fully connected to obtain a feature vector, and the feature vector may include multiple data features.
- Baseline model of neural network (such as convolutional neural network):
- each neural network parameter in the neural network can be trained with sample data, such as convolutional layer parameters (such as convolution kernel), pooling layer parameters , excitation layer parameters, fully connected layer parameters, etc., which are not limited.
- the neural network can fit the mapping relationship between input and output.
- the trained neural network is the baseline model of the neural network, which is referred to as the baseline model in this paper.
- Artificial intelligence processing can be realized based on the baseline model, such as face detection, human body detection, vehicle detection, license plate recognition, etc. artificial intelligence processing.
- the image including the license plate that is, the license plate image
- the baseline model performs artificial intelligence processing on the image
- the artificial intelligence processing result is the license plate recognition result.
- Sample data and scene data In artificial intelligence scenarios, central servers and edge servers can be deployed.
- the central server obtains data for training the baseline model, trains the baseline model based on these data, and uses the data used by the central server for training the baseline model.
- the data is called sample data.
- the edge server obtains the data for training the baseline model (which will not be sent to the central server), and trains the baseline model based on the data.
- the data used by the edge server for training the baseline model is called scene data.
- the sample data may be image data or other types of data, which are not limited.
- the scene data may be image data or other types of data, which are not limited.
- the above-mentioned baseline model may be a baseline model for realizing license plate recognition.
- the sample data can be the license plate image (that is, the image including the license plate information) obtained by the central server for training the baseline model, and the central server can train the baseline model based on the license plate image.
- the license plate image is only an example of the sample data , there is no restriction on this.
- the scene data can be the license plate image obtained by the edge server for training the baseline model. The license plate image will not be sent to the central server.
- the edge server can train the baseline model based on the license plate image.
- the license plate image is only an example of the scene data. No restrictions.
- the central server since the edge server will not send the scene data to the central server, the central server cannot train the baseline model through the scene data of the edge server, that is, the baseline model cannot match the environment where the edge server is located, and the baseline model is deployed to the edge server. , the performance of the baseline model is lower.
- the central server may use license plate images as sample data to train a baseline model for implementing license plate recognition.
- the edge server since the license plate image belongs to private data, based on the protection requirements for private data, the edge server will not send the license plate image to the central server after obtaining the license plate image, so that the central server cannot pass the license plate image of the edge server.
- the baseline model that is, the baseline model cannot match the environment where the edge server is located.
- the performance of the baseline model is low.
- the edge server can use the scene data of its own environment to train the baseline model to obtain a new baseline model. Since the baseline model is trained by using the scene data of the environment where the edge server is located, the new baseline model can match the environment where the edge server is located, and the performance of the new baseline model is better.
- FIG. 1 is a schematic structural diagram of a system according to an embodiment of the present application.
- the system may include a center server 110 and an edge server 120 , a first relay device 130 connected to the center server 110 , and a first relay device 130 connected to the edge server 120 .
- Two relay devices 140, the first relay device 130 and the second relay device 140 are connected through a network.
- the number of edge servers 120 is at least one, and each edge server 120 is connected to one second transit device 140 , that is, the number of second transit devices 140 is the same as the number of edge servers 120 .
- the central server 110 may constitute an identification center system
- at least one edge server 120 may constitute an identification end-side system
- the first transit device 130 and at least one second transit device 140 may constitute a deprivation system.
- the central server 110 is a service platform for providing a baseline model, and can provide a baseline model for at least one edge server 120 .
- the edge server 120 is a server with baseline model requirements, that is, the baseline model needs to be acquired from the central server 110, and then artificial intelligence processing is implemented according to the baseline model.
- the first relay device 130 is a network device (such as a router, a switch, etc.) connected to the central server 110, and is used to forward the baseline model sent by the central server 110 to the edge server 120, and forward the baseline model sent by the edge server 120 to the central server. 110.
- the second relay device 140 is a network device connected to the edge server 120 , and is used for forwarding the baseline model sent by the center server 110 to the edge server 120 , and forwarding the baseline model sent by the edge server 120 to the center server 110 .
- An embodiment of the present application proposes a data processing method, as shown in FIG. 2 , the method includes steps 201-207.
- Step 201 the central server sends the initial baseline model to the edge server.
- the central server can obtain a large amount of sample data, and there is no restriction on the acquisition method.
- the sample data has label information, such as the actual category and/or target frame, etc., and the label information is not limited. make restrictions.
- the sample data can be a license plate image
- the target frame can be the coordinate information of a rectangular frame in the license plate image (such as the coordinates of the upper left corner of the rectangular frame, the width and height of the rectangular frame, etc.)
- the actual class represents the license plate identification of the rectangular box area.
- the central server trains a neural network (such as a convolutional neural network) according to the sample data and the label information of the sample data to obtain a baseline model.
- the baseline model is called an initial baseline model.
- the neural network can be the initial baseline model.
- output data corresponding to the sample data can be obtained.
- the output data is a feature vector, and whether the neural network has converged can be determined based on the feature vector.
- the loss value of the loss function (which can be configured according to experience) is determined according to the feature vector, and whether the neural network has converged is determined according to the loss value. For example, if the loss value is less than a preset threshold, the neural network has converged, otherwise , the neural network does not converge.
- the neural network has converged, complete the training process and obtain the initial baseline model. If the neural network has not converged, continue to adjust the parameters of each neural network in the neural network to obtain the adjusted neural network. Then, the sample data and The label information corresponding to the sample data is input to the adjusted neural network, and the adjusted neural network is continued to be trained, and so on, until the neural network has converged.
- this determination method is not limited. For example, if the number of iterations reaches a preset number of times threshold, it is determined that the neural network has converged; for another example, if the iteration duration reaches a preset duration threshold, it is determined that the neural network has converged.
- the central server can send the initial baseline model to the edge server.
- the initial baseline model can be sent to all edge servers or some edge servers. The following is combined with an edge server pair. This sending process is explained:
- Mode 1 The central server sends the initial baseline model to the first transit device, the first transit device sends the initial baseline model to the second transit device, and the second transit device can send the initial baseline model to the edge server, so far , the initial baseline model is successfully sent to the edge server.
- Mode 2 The central server sends the initial baseline model to the first relay device, and the first relay device performs a first conversion operation on the initial baseline model to obtain a first low-dimensional baseline model, and sends the first low-dimensional baseline model to the second transit device.
- the second transit device performs a second conversion operation on the first low-dimensional baseline model to obtain the initial baseline model, and sends the initial baseline model to the edge server.
- the first conversion operation may be an encryption operation
- the second conversion operation may be a decryption operation
- the first transformation operation may be a compression operation and the second transformation operation may be a decompression operation
- the first conversion operation may be an encryption operation and a compression operation
- the second conversion operation may be a decryption operation and a decompression operation.
- the above is just an example of the first conversion operation and the second conversion operation, which is not limited.
- the first transit device after receiving the initial baseline model, performs a compression operation on the initial baseline model, so as to convert the initial baseline model into a first low-dimensional baseline model, and convert the first low-dimensional baseline model to the first low-dimensional baseline model. Sent to the second relay device. After receiving the first low-dimensional baseline model, the second transit device decompresses the first low-dimensional baseline model to obtain a decompressed initial baseline model.
- the compressed first low-dimensional baseline model is transmitted in the network instead of the uncompressed initial baseline model, the amount of transmitted data can be reduced and the model transmission time can be saved.
- the first transit device after receiving the initial baseline model, performs an encryption operation on the initial baseline model, so as to convert the initial baseline model into a first low-dimensional baseline model, and convert the first low-dimensional baseline model to the first low-dimensional baseline model. Sent to the second relay device. After receiving the first low-dimensional baseline model, the second transit device performs a decryption operation on the first low-dimensional baseline model to obtain a decrypted initial baseline model.
- the encrypted first low-dimensional baseline model is transmitted in the network instead of the unencrypted initial baseline model, the security of the initial baseline model can be ensured and the attacker can avoid illegally intercepting the initial baseline model.
- the first transit device after receiving the initial baseline model, performs a compression operation on the initial baseline model, and performs an encryption operation on the compressed initial baseline model, so as to convert the initial baseline model into the first low wiki line model, and send the first low-dimensional line model to the second relay device.
- the second relay device After receiving the first low-dimensional baseline model, performs a decryption operation on the first low-dimensional baseline model, and decompresses the decrypted first low-dimensional baseline model to obtain a decompressed initial baseline model.
- the compressed and encrypted first low-dimensional baseline model is transmitted in the network, the amount of transmitted data can be reduced, the model transmission time can be saved, and an attacker can avoid illegally intercepting the initial baseline model.
- the first transit device after receiving the initial baseline model, performs an encryption operation on the initial baseline model, and performs a compression operation on the encrypted initial baseline model, thereby converting the initial baseline model into the first low wiki line model, and send the first low-dimensional line model to the second relay device.
- the second relay device After receiving the first low-dimensional baseline model, performs a decompression operation on the first low-dimensional baseline model, and performs a decryption operation on the decompressed first low-dimensional baseline model to obtain a decrypted initial baseline model.
- the initial baseline model can be understood as a high-dimensional parameter vector because the initial baseline model has been encrypted and/or compressed.
- a line model can be understood as a low-dimensional parameter vector, i.e. a smaller number of parameters.
- a sparse algorithm can be used to compress the initial baseline model, and a sparse algorithm can be used to decompress the first low-dimensional baseline model.
- the sparse algorithm is only an example of a compression algorithm, which is not limited. , any algorithm capable of compressing the initial baseline model can be used.
- a cryptographic algorithm can be used to encrypt the initial baseline model, and the cryptographic algorithm can be used to decrypt the first low-dimensional baseline model.
- the cryptographic algorithm can be any type of cryptographic algorithm, and this cryptographic algorithm is not limited. , as long as the initial baseline model can be encrypted.
- the cryptographic algorithm can be SM1 (Security Manager, security management), SM2, SM3, SM4 and other domestic commercial cryptographic algorithms, or can be DES (Data Encryption Standard, data encryption standard), AES (Advanced Encryption Standard, advanced encryption) Standard), IDEA (International Data Encryption Algorithm, International Data Encryption Algorithm) and other international commercial cryptographic algorithms.
- Step 202 the edge server trains the initial baseline model through the scene data of the edge server to obtain a target baseline model, that is, the trained initial baseline model is the target baseline model.
- the edge server may acquire a large amount of scene data
- the scene data may be scene data of the environment where the edge server is located, and the acquisition method is not limited.
- the scene data has label information, such as an actual category and/or a target frame, and the label information is not limited.
- the scene data can be a license plate image
- the target frame can be the coordinate information of a rectangular frame in the license plate image (such as the coordinates of the upper left corner of the rectangular frame, the width and height of the rectangular frame, etc. )
- the actual category can represent the license plate identification of the rectangular box area.
- the edge server trains the initial baseline model according to the scene data and the label information of the scene data, and obtains the trained baseline model.
- the trained baseline model is called the target baseline model.
- a large amount of scene data and the label information corresponding to the scene data are input into the initial baseline model, so as to use these scene data and label information to train the neural network parameters in the initial baseline model, after the initial baseline model training is completed,
- the initial baseline model that has been trained can be the target baseline model.
- the initial baseline model In the training process of the initial baseline model, after inputting the scene data to the initial baseline model, output data corresponding to the scene data can be obtained, the output data is a feature vector, and whether the initial baseline model has converged is determined based on the feature vector. For example, the loss value of the loss function (which can be configured according to experience) is determined according to the feature vector, and whether the initial baseline model has converged is determined according to the loss value, for example, if the loss value is less than a preset threshold, the initial baseline model has converged, Otherwise, the initial baseline model did not converge.
- the loss value of the loss function which can be configured according to experience
- the training process is completed and the target baseline model is obtained. If the initial baseline model does not converge, continue to adjust the neural network parameters in the initial baseline model to obtain the adjusted initial baseline model, input the scene data and label information into the adjusted initial baseline model, and continue to adjust the adjusted initial baseline model. The initial baseline model is trained, and so on, until the initial baseline model has converged.
- other methods may also be used to determine whether the initial baseline model has converged, which is not limited. For example, if the number of iterations reaches a preset number of thresholds, it is determined that the initial baseline model has converged; for another example, if the iteration duration reaches a preset duration threshold, it is determined that the initial baseline model has converged.
- the initial baseline model can be trained based on the scene data of the edge server to obtain the target baseline model. Since the scene data is the data of the environment where the edge server is located, and the scene data is not sent to the central server, that is, the central server does not use the scene data to train the initial baseline model. Therefore, the edge server uses these scene data to train the initial baseline model.
- the target baseline model has sample data information and new knowledge of scene data to improve model performance.
- Step 203 the edge server determines whether to deploy the target baseline model. Exemplarily, if the target baseline model is deployed, step 204 is performed; if the target baseline model is not deployed, step 205 is performed.
- the target baseline model may be deployed, and if the target baseline model does not meet the performance requirement, the target baseline model may not be deployed.
- a test data set is acquired, where the test data set includes a plurality of test data, and each test data corresponds to an actual category of the test data.
- the test data set For each test data in the test data set, input the test data to the target baseline model, and perform artificial intelligence processing on the test data through the target baseline model to obtain the artificial intelligence processing result, and the artificial intelligence processing result is the result of the test data.
- Predicted category If the predicted category of the test data is consistent with the actual category of the test data, it means that the target baseline model recognizes the test data correctly; if the predicted category of the test data is inconsistent with the actual category of the test data, it means the target baseline model The model identified the test data incorrectly.
- the number of correct identification results (denoted as a1) and the number of incorrect identification results (denoted as a2) can be obtained, and according to the number of correct identification results a1 and the number of wrong identification results a2 , to determine the performance metrics of the target baseline model.
- the performance index can be a1/(a1+a2). Obviously, the larger the performance index, the better the performance of the target baseline model.
- the performance index of the target baseline model is greater than the preset threshold, it means that the target baseline model has met the performance requirements, and the target baseline model can be deployed. If the performance index of the target baseline model is not greater than the preset threshold, it means that the target baseline model does not meet the performance requirements, and the target baseline model may not be deployed.
- the above is just an example of determining whether to deploy the target baseline model, and there is no restriction on the determination method.
- the preset number of times threshold is determined to determine the deployment target baseline model. For another example, when a user instruction is received, whether to deploy the target baseline model is determined based on the user instruction.
- Step 204 the edge server processes the application data through the target baseline model.
- the edge server can deploy the target baseline model on the edge server, and after acquiring the application data, the edge server can input the application data to the target baseline model, so as to perform artificial intelligence processing on the application data through the target baseline model, Get artificial intelligence processing results.
- the edge server can also deploy the target baseline model to the terminal device, that is, the terminal device managed by the edge server, such as an analog camera, IPC (Internet Protocol Camera, network camera), etc.
- the terminal device obtains the application data
- the application data can be input into the target baseline model, so that artificial intelligence processing is performed on the application data through the target baseline model to obtain an artificial intelligence processing result.
- the application data is an image including license plates.
- the target baseline model performs artificial intelligence processing on the application data to obtain the artificial intelligence processing result.
- the artificial intelligence processing result may be a license plate identification.
- Step 205 the edge server sends the target baseline model to the central server.
- the edge server After obtaining the target baseline model, the edge server sends the target baseline model to the central server if it is determined not to deploy the target baseline model. If it is determined to deploy the target baseline model, the target baseline model may be sent to the central server, or the target baseline model may not be sent to the central server.
- the process of sending the target baseline model to the central server from the edge server may include:
- Mode 1 The edge server sends the target baseline model to the second relay device, the second relay device sends the target baseline model to the first relay device, and the first relay device can send the target baseline model to the central server, so far , successfully sending the target baseline model to the central server.
- the edge server sends the target baseline model to the second transit device, and the second transit device performs a first conversion operation on the target baseline model to obtain a second low-dimensional baseline model, and sends the second low-dimensional baseline model to the first transit device.
- the first relay device performs a second conversion operation on the second low-dimensional baseline model to obtain the target baseline model, and sends the target baseline model to the central server.
- the first conversion operation may be an encryption operation
- the second conversion operation may be a decryption operation
- the first transformation operation may be a compression operation and the second transformation operation may be a decompression operation
- the first conversion operation may be an encryption operation and a compression operation
- the second conversion operation may be a decryption operation and a decompression operation.
- the above is just an example of the first conversion operation and the second conversion operation, which is not limited.
- the second relay device performs a compression operation on the target baseline model, thereby converting the target baseline model into a second low-dimensional baseline model, and sends the second low-dimensional baseline model to the first relay device.
- the first transit device decompresses the second low-dimensional baseline model to obtain the target baseline model.
- the second transit device performs an encryption operation on the target baseline model, thereby converting the target baseline model into a second low-dimensional baseline model, and sends the second low-dimensional baseline model to the first transit device.
- the first transit device decrypts the second low-dimensional baseline model to obtain the target baseline model.
- the second transit device performs compression and encryption operations on the target baseline model, thereby converting the target baseline model into a second low-dimensional baseline model, and sends the second low-dimensional baseline model to the first transit device.
- the first transit device performs decryption and decompression operations on the second low-dimensional baseline model to obtain the target baseline model.
- Step 206 the central server generates a fusion baseline model based on the target baseline model and the initial baseline model.
- the central server may perform a fusion operation on the target baseline model and the initial baseline model of at least one edge server to obtain a fusion baseline model;
- the fusion operation may include but is not limited to one of the following operations: weighting operation, averaging operation, Take the maximum value operation, take the minimum value operation.
- the target baseline model and the initial baseline model have the same network structure, for example, both include network layer 1 and network layer 2, network layer 1 includes parameter A and parameter B, and network layer 2 includes parameter C and parameter D.
- the central server obtains the target baseline model 1 and the target baseline model 2.
- the value of parameter A is a11
- the value of parameter B is b11
- the value of parameter C is c11
- the value of parameter D is d11.
- the value of parameter A is a21
- the value of parameter B is b21
- the value of parameter C is c21
- the value of parameter D is d21.
- the value of parameter A is a31
- the value of parameter B is b31
- the value of parameter C is c31
- the value of parameter D is d31.
- the central server can fuse the initial baseline model, target baseline model 1 and target baseline model 2 to obtain a fused baseline model.
- the fused baseline model can include network layer 1 and network layer 2, and network layer 1 includes parameters A and parameter B, network layer 2 includes parameter C and parameter D.
- the value of parameter A is obtained by fusing a11, a21 and a31.
- the value of parameter A is the average value of a11, a21 and a31; or, the value of parameter A is the maximum value of a11, a21 and a31; or, the value of parameter A is the value of a11, a21 and a31 or, the value of parameter A is the weighted value of a11, a21 and a31, if the weight of the initial baseline model (eg 2) is greater than the weight of the target baseline model (eg 1), the value of parameter A is ( a11*2+a21*1+a31*1)/4, when the weight of the initial baseline model (eg 1) is less than the weight of the target baseline model (eg 2), the value of parameter A is (a11*1+a21*2 +a31*2)/5.
- the above are just a few examples, which are not limited.
- the value of parameter B is obtained by fusing b11, b21 and b31.
- the value of parameter B is the average value of b11, b21 and b31; or, the value of parameter B is the maximum value of b11, b21 and b31; or, the value of parameter B is the value of b11, b21 and b31
- the minimum value of ; or, the value of parameter B is the weighted value of b11, b21 and b31, which is not limited.
- the value of parameter C is obtained by fusing c11, c21 and c31.
- the value of parameter C is the average value of c11, c21 and c31; or, the value of parameter C is the maximum value of c11, c21 and c31; or the value of parameter C is the value of c11, c21 and c31
- the minimum value of ; or, the value of parameter C is the weighted value of c11, c21 and c31, which is not limited.
- the value of parameter D is obtained by the fusion operation of d11, d21 and d31.
- the value of parameter D is the average value of d11, d21 and d31; or, the value of parameter D is the maximum value of d11, d21 and d31; or the value of parameter D is the value of d11, d21 and d31
- the minimum value of ; or, the value of parameter D is the weighted value of d11, d21 and d31, which is not limited.
- Step 207 the central server trains the fusion baseline model to obtain a trained baseline model, determines the trained baseline model as the initial baseline model, and returns to step 201 .
- the central server may train the fused baseline model according to the sample data and the label information of the sample data to obtain the trained baseline model.
- the training process refer to step 201, which is not described here. Repeat. Determine the trained baseline model as the initial baseline model, and return to step 201, that is, send the initial baseline model to the edge server.
- the edge server can obtain the initial baseline model from the central server, train the initial baseline model through the scene data, obtain the target baseline model, and send the target baseline model to the central server, and the central server is based on the target baseline model.
- the model and the initial baseline model generate a fused baseline model, train the fused baseline model, determine the trained baseline model as the initial baseline model, send the initial baseline model to the edge server, and so on. Since the above process can be performed cyclically, the performance of the initial baseline model and the target baseline model is continuously upgraded, and the recognition capabilities of the initial baseline model and the target baseline model are continuously improved, so that the target baseline model can achieve the expected performance and achieve high-precision recognition capabilities.
- the edge server can obtain a target baseline model with better performance, the target baseline model can match the environment where the edge server is located, and the accuracy of the intelligent analysis results is high.
- the edge server sends the target baseline model to the central server instead of scene data (such as license plate images), so as to protect the privacy of the scene data, and will not send the scene data to the central server. Since the target baseline model is trained based on the scene data, the information of the scene data can be embodied in the initial baseline model and the target baseline model.
- the scene data of the edge server can be a license plate image
- the sample data of the central server can be a license plate image
- the target baseline model is used to perform license plate recognition processing on the application data, that is, identify the license plate image. License plate identification in license plate images.
- a private license plate recognition method is proposed, which can achieve a good license plate recognition capability without involving user private data, that is, without sending the private license plate image to the central server.
- the network structure of the baseline model is the same, and the initial baseline model trained by the center server has the same network structure as the target baseline model trained by the edge server.
- Figure 3 it is a schematic diagram of the network structure of the baseline model, and the two use the same network structure.
- edge server Since the edge server is not directly connected to the central server, the problem of a single edge server will not directly affect the operation of the central server, and the edge servers can be easily added or deleted.
- the edge server may also be called an end-side server, the edge server is maintained by a legal (meeting local regulations) regional support provider/agent, and the edge server has a model training function.
- the central server is maintained by the identification system service provider, and the central server has the function of model training.
- the edge server can obtain the initial baseline model from the central server, obtain better preliminary recognition ability, and train the initial baseline model according to the local license plate image to obtain the target baseline model, so that the target baseline model can be adapted to the edge server.
- the central server can obtain the target baseline model sent by the edge server, and complete the fusion of the target baseline model and the initial baseline model, so that the initial baseline model is further upgraded and has stronger generalization ability.
- a data processing method is proposed in the embodiment of the present application, and the method includes:
- step S1 the central server obtains an initial baseline model by training according to the sample data.
- Step S2 the central server sends the initial baseline model to the first transit device.
- Step S3 the first transfer device performs a first conversion operation on the initial baseline model to obtain a first low-dimensional baseline model, and sends the first low-dimensional baseline model to the second transfer device.
- the first low-dimensional line model may be sent to each second relay device, and a second relay device is used as an example for description in the following.
- Step S4 the second transit device performs a second conversion operation on the first low-dimensional baseline model to obtain the initial baseline model, and sends the initial baseline model to the edge server.
- the initial baseline model can be sent to the edge server connected to the second transit device, that is, the initial baseline model can be sent to the edge server connected to the second transit device.
- the baseline model is sent to multiple edge servers.
- the processing process of one edge server is used as an example for description.
- Step S5 the edge server trains the initial baseline model through the scene data of the edge server to obtain a target baseline model, that is, the trained initial baseline model is the target baseline model.
- Step S6 the edge server determines whether to deploy the target baseline model. Exemplarily, if the target baseline model is deployed, step S7 is performed, and if the target baseline model is not deployed, step S8 is performed.
- Step S7 the edge server processes the application data through the target baseline model.
- Step S8 the edge server sends the target baseline model to the second transit device.
- Step S9 the second transfer device performs a first conversion operation on the target baseline model to obtain a second low-dimensional baseline model, and sends the second low-dimensional baseline model to the first transfer device.
- Step S10 the first relay device performs a second conversion operation on the second low-dimensional baseline model to obtain the target baseline model, and sends the target baseline model to the central server.
- Step S11 the central server generates a fusion baseline model based on the target baseline model and the initial baseline model.
- Step S12 the central server trains the fusion baseline model to obtain a trained baseline model, determines the trained baseline model as the initial baseline model, and returns to step S2.
- the performance of the initial baseline model of the central server can be continuously upgraded, and the license plate recognition capability can be continuously improved.
- the performance of the initial baseline model and the target baseline model can be continuously improved to achieve the expected performance.
- the method can be applied to an edge server. Referring to FIG. 4 , the method can include steps 401-404.
- Step 401 obtaining an initial baseline model from a central server.
- Step 402 Train the initial baseline model through the scene data of the edge server to obtain a target baseline model, and determine whether to deploy the target baseline model. Exemplarily, if the target baseline model is deployed, step 403 is performed, and if the target baseline model is not deployed, step 404 is performed.
- step 403 the application data is processed through the target baseline model.
- Step 404 sending the target baseline model to the central server, so that the central server generates a fused baseline model based on the target baseline model and the initial baseline model, and re-acquires the initial baseline model based on the fused baseline model, and returns to perform the initial acquisition from the central server.
- step 401 is executed.
- the edge server can retrieve the initial baseline model from the central server.
- the method can be applied to a central server.
- the method can include steps 501-504.
- Step 501 Send the initial baseline model to the edge server, so that the edge server trains the initial baseline model through the scene data of the edge server to obtain the target baseline model.
- Step 502 Obtain the target baseline model from the edge server.
- Step 503 generating a fusion baseline model based on the target baseline model and the initial baseline model.
- Step 504 train the fused baseline model, determine the trained baseline model as the initial baseline model, and return to perform the operation of sending the initial baseline model to the edge server, that is, step 501 is performed.
- FIG. 6A is a schematic structural diagram of the apparatus. Referring to FIG. 6A , the apparatus includes:
- Obtaining module 611 for obtaining the initial baseline model from the central server
- a training module 612 configured to train the initial baseline model through the scene data of the edge server, obtain a target baseline model, and determine whether to deploy the target baseline model;
- Sending module 613 configured to send the target baseline model to the central server when the target baseline model is not deployed, so that the central server generates a fusion baseline based on the target baseline model and the initial baseline model model, and re-acquire the initial baseline model based on the fused baseline model.
- the data processing apparatus may further include: a processing module configured to process application data by using the target baseline model when deploying the target baseline model.
- FIG. 6B is a schematic structural diagram of the apparatus. Referring to FIG. 6B , the apparatus includes:
- the sending module 621 is configured to send the initial baseline model to the edge server, so that the edge server can train the initial baseline model through the scene data of the edge server to obtain the target baseline model;
- an obtaining module 622 configured to obtain the target baseline model from the edge server
- the generating module 623 is configured to generate a fused baseline model based on the target baseline model and the initial baseline model; train the fused baseline model, and determine the trained baseline model as the initial baseline model.
- the generating module 623 When the generating module 623 generates the fusion baseline model based on the target baseline model and the initial baseline model, it is specifically configured to: perform a fusion operation on the target baseline model and the initial baseline model of at least one edge server to obtain the fusion baseline The model; wherein, the fusion operation includes one of the following operations: weighting operation, averaging operation, maximum value operation, and minimum value operation.
- the edge server includes: a processor 711 and a machine-readable storage medium 712, the machine-readable storage medium 712 stores machine-executable instructions that can be executed by the processor 711; the processor 711 is configured to execute the machine-executable instructions to implement the following steps:
- the central server includes: a processor 721 and a machine-readable storage medium 722, the machine-readable storage medium 722 stores machine-executable instructions that can be executed by the processor 721; the processor 721 is configured to execute the machine-executable instructions to implement the following steps:
- the fusion baseline model is trained, the trained baseline model is determined as the initial baseline model, and the operation of sending the initial baseline model to the edge server is returned to be executed.
- an embodiment of the present application further provides a machine-readable storage medium, where several computer instructions are stored on the machine-readable storage medium, and when the computer instructions are executed by a processor, the present invention can be implemented. Apply the data processing method disclosed in the above example.
- the above-mentioned machine-readable storage medium may be any electronic, magnetic, optical or other physical storage device, which may contain or store information, such as executable instructions, data, and the like.
- the machine-readable storage medium may be: RAM (Radom Access Memory, random access memory), volatile memory, non-volatile memory, flash memory, storage drives (such as hard drives), solid-state drives, any type of storage discs (such as optical discs, DVDs, etc.), or similar storage media, or a combination thereof.
- a typical implementing device is a computer, which may be in the form of a personal computer, laptop computer, cellular phone, camera phone, smart phone, personal digital assistant, media player, navigation device, email sending and receiving device, game control desktop, tablet, wearable device, or a combination of any of these devices.
- the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, embodiments of the present application may take the form of a computer program product implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
- computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
- these computer program instructions may also be stored in a computer readable memory capable of directing a computer or other programmable data processing apparatus to operate in a particular manner, such that the instructions stored in the computer readable memory result in an article of manufacture comprising the instruction means,
- the instruction means implements the functions specified in a flow or flows of the flowcharts and/or a block or blocks of the block diagrams.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Medical Informatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioethics (AREA)
- Mathematical Physics (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Signal Processing (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Databases & Information Systems (AREA)
- Computer Hardware Design (AREA)
- Computer Security & Cryptography (AREA)
- Computer And Data Communications (AREA)
- Debugging And Monitoring (AREA)
- Telephonic Communication Services (AREA)
Abstract
Description
Claims (12)
- 一种数据处理方法,所述方法包括:中心服务器将初始基线模型发送给边缘服务器;所述边缘服务器通过所述边缘服务器的场景数据对所述初始基线模型进行训练,得到目标基线模型,并确定是否部署所述目标基线模型;若否,则所述边缘服务器将所述目标基线模型发送给所述中心服务器;所述中心服务器基于所述目标基线模型和所述初始基线模型生成融合基线模型,对所述融合基线模型进行训练,将训练后的基线模型确定为初始基线模型,返回执行中心服务器将初始基线模型发送给边缘服务器的操作。
- 根据权利要求1所述的方法,其中,所述中心服务器将初始基线模型发送给边缘服务器,包括:所述中心服务器将所述初始基线模型发送给第一中转设备;所述第一中转设备对所述初始基线模型进行第一转换操作,得到第一低维基线模型,并将所述第一低维基线模型发送给第二中转设备;所述第二中转设备对所述第一低维基线模型进行第二转换操作,得到所述初始基线模型,并将所述初始基线模型发送给所述边缘服务器。
- 根据权利要求1所述的方法,其中,所述边缘服务器将所述目标基线模型发送给所述中心服务器,包括:所述边缘服务器将所述目标基线模型发送给第二中转设备;所述第二中转设备对所述目标基线模型进行第一转换操作,得到第二低维基线模型,并将所述第二低维基线模型发送给第一中转设备;所述第一中转设备对所述第二低维基线模型进行第二转换操作,得到所述目标基线模型,并将所述目标基线模型发送给所述中心服务器。
- 根据权利要求2或者3所述的方法,其中,所述第一转换操作为加密操作,所述第二转换操作为解密操作;或者,所述第一转换操作为压缩操作,所述第二转换操作为解压缩操作;或者,所述第一转换操作为加密操作和压缩操作,所述第二转换操作为解密操作和解压缩操作。
- 根据权利要求1所述的方法,其中,所述边缘服务器包括至少一个边缘服务器,所述中心服务器基于所述目标基线模型和所述初始基线模型生成融合基线模型,包括:所述中心服务器对所述至少一个边缘服务器的目标基线模型以及所述初始基线模 型进行融合操作,得到所述融合基线模型;其中,所述融合操作包括以下操作中的一种:加权操作,平均操作,取最大值操作,取最小值操作。
- 根据权利要求1所述的方法,其中,所述确定是否部署所述目标基线模型之后,所述方法还包括:若是,则所述边缘服务器通过所述目标基线模型对应用数据进行处理。
- 一种数据处理方法,应用于边缘服务器,所述方法包括:从中心服务器获取初始基线模型;通过所述边缘服务器的场景数据对所述初始基线模型进行训练,得到目标基线模型,并确定是否部署所述目标基线模型;若否,则将所述目标基线模型发送给所述中心服务器,以使所述中心服务器基于所述目标基线模型和所述初始基线模型生成融合基线模型,并基于所述融合基线模型重新获取初始基线模型,返回执行从中心服务器获取初始基线模型的操作。
- 一种数据处理方法,应用于中心服务器,所述方法包括:将初始基线模型发送给边缘服务器,以使所述边缘服务器通过所述边缘服务器的场景数据对所述初始基线模型进行训练,得到目标基线模型;从所述边缘服务器获取所述目标基线模型;基于所述目标基线模型和所述初始基线模型生成融合基线模型;对所述融合基线模型进行训练,将训练后的基线模型确定为初始基线模型,返回执行将初始基线模型发送给边缘服务器的操作。
- 一种数据处理装置,应用于边缘服务器,所述装置包括:获取模块,用于从中心服务器获取初始基线模型;训练模块,用于通过所述边缘服务器的场景数据对所述初始基线模型进行训练,得到目标基线模型,并确定是否部署所述目标基线模型;发送模块,用于在不部署所述目标基线模型时,将所述目标基线模型发送给所述中心服务器,以使所述中心服务器基于所述目标基线模型和所述初始基线模型生成融合基线模型,并基于所述融合基线模型重新获取初始基线模型。
- 一种数据处理装置,应用于中心服务器,所述装置包括:发送模块,用于将初始基线模型发送给边缘服务器,以使边缘服务器通过所述边缘服务器的场景数据对所述初始基线模型进行训练,得到目标基线模型;获取模块,用于从所述边缘服务器获取所述目标基线模型;生成模块,用于基于所述目标基线模型和所述初始基线模型生成融合基线模型;对融合基线模型进行训练,将训练后的基线模型确定为初始基线模型。
- 一种边缘服务器,包括:处理器和机器可读存储介质,所述机器可读存储介质存储有能够被所述处理器执行的机器可执行指令;所述处理器用于执行机器可执行指令,以实现如下的步骤:从中心服务器获取初始基线模型;通过所述边缘服务器的场景数据对所述初始基线模型进行训练,得到目标基线模型,并确定是否部署所述目标基线模型;若否,则将所述目标基线模型发送给所述中心服务器,以使所述中心服务器基于所述目标基线模型和所述初始基线模型生成融合基线模型,并基于所述融合基线模型重新获取初始基线模型,返回执行从中心服务器获取初始基线模型的操作。
- 一种中心服务器,包括:处理器和机器可读存储介质,所述机器可读存储介质存储有能够被所述处理器执行的机器可执行指令;所述处理器用于执行机器可执行指令,以实现如下的步骤:将初始基线模型发送给边缘服务器,以使所述边缘服务器通过所述边缘服务器的场景数据对所述初始基线模型进行训练,得到目标基线模型;从所述边缘服务器获取所述目标基线模型;基于所述目标基线模型和所述初始基线模型生成融合基线模型;对所述融合基线模型进行训练,将训练后的基线模型确定为初始基线模型,返回执行将初始基线模型发送给边缘服务器的操作。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010610046.0A CN111783630B (zh) | 2020-06-29 | 2020-06-29 | 一种数据处理方法、装置及设备 |
CN202010610046.0 | 2020-06-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022001092A1 true WO2022001092A1 (zh) | 2022-01-06 |
Family
ID=72760352
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/073615 WO2022001092A1 (zh) | 2020-06-29 | 2021-01-25 | 一种数据处理方法、装置及设备 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111783630B (zh) |
WO (1) | WO2022001092A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116451276A (zh) * | 2023-06-15 | 2023-07-18 | 杭州海康威视数字技术股份有限公司 | 一种图像处理方法、装置、设备及系统 |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111783630B (zh) * | 2020-06-29 | 2022-07-01 | 上海高德威智能交通系统有限公司 | 一种数据处理方法、装置及设备 |
CN112686300B (zh) * | 2020-12-29 | 2023-09-26 | 杭州海康威视数字技术股份有限公司 | 一种数据处理方法、装置及设备 |
CN113222014A (zh) * | 2021-05-12 | 2021-08-06 | 深圳思谋信息科技有限公司 | 图像分类模型训练方法、装置、计算机设备和存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108510000A (zh) * | 2018-03-30 | 2018-09-07 | 北京工商大学 | 复杂场景下行人细粒度属性的检测与识别方法 |
CN110263936A (zh) * | 2019-06-14 | 2019-09-20 | 深圳前海微众银行股份有限公司 | 横向联邦学习方法、装置、设备及计算机存储介质 |
CN111091199A (zh) * | 2019-12-20 | 2020-05-01 | 哈尔滨工业大学(深圳) | 一种基于差分隐私的联邦学习方法、装置及存储介质 |
CN111222647A (zh) * | 2020-01-09 | 2020-06-02 | 深圳前海微众银行股份有限公司 | 联邦学习系统优化方法、装置、设备及存储介质 |
WO2020115273A1 (en) * | 2018-12-07 | 2020-06-11 | Telefonaktiebolaget Lm Ericsson (Publ) | Predicting network communication performance using federated learning |
CN111783630A (zh) * | 2020-06-29 | 2020-10-16 | 上海高德威智能交通系统有限公司 | 一种数据处理方法、装置及设备 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8194938B2 (en) * | 2009-06-02 | 2012-06-05 | George Mason Intellectual Properties, Inc. | Face authentication using recognition-by-parts, boosting, and transduction |
US10936907B2 (en) * | 2018-08-10 | 2021-03-02 | Buffalo Automation Group Inc. | Training a deep learning system for maritime applications |
CN109711556B (zh) * | 2018-12-24 | 2020-11-03 | 中国南方电网有限责任公司 | 机巡数据处理方法、装置、网级服务器和省级服务器 |
-
2020
- 2020-06-29 CN CN202010610046.0A patent/CN111783630B/zh active Active
-
2021
- 2021-01-25 WO PCT/CN2021/073615 patent/WO2022001092A1/zh active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108510000A (zh) * | 2018-03-30 | 2018-09-07 | 北京工商大学 | 复杂场景下行人细粒度属性的检测与识别方法 |
WO2020115273A1 (en) * | 2018-12-07 | 2020-06-11 | Telefonaktiebolaget Lm Ericsson (Publ) | Predicting network communication performance using federated learning |
CN110263936A (zh) * | 2019-06-14 | 2019-09-20 | 深圳前海微众银行股份有限公司 | 横向联邦学习方法、装置、设备及计算机存储介质 |
CN111091199A (zh) * | 2019-12-20 | 2020-05-01 | 哈尔滨工业大学(深圳) | 一种基于差分隐私的联邦学习方法、装置及存储介质 |
CN111222647A (zh) * | 2020-01-09 | 2020-06-02 | 深圳前海微众银行股份有限公司 | 联邦学习系统优化方法、装置、设备及存储介质 |
CN111783630A (zh) * | 2020-06-29 | 2020-10-16 | 上海高德威智能交通系统有限公司 | 一种数据处理方法、装置及设备 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116451276A (zh) * | 2023-06-15 | 2023-07-18 | 杭州海康威视数字技术股份有限公司 | 一种图像处理方法、装置、设备及系统 |
CN116451276B (zh) * | 2023-06-15 | 2023-09-26 | 杭州海康威视数字技术股份有限公司 | 一种图像处理方法、装置、设备及系统 |
Also Published As
Publication number | Publication date |
---|---|
CN111783630A (zh) | 2020-10-16 |
CN111783630B (zh) | 2022-07-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022001092A1 (zh) | 一种数据处理方法、装置及设备 | |
CN113688855B (zh) | 数据处理方法、联邦学习的训练方法及相关装置、设备 | |
Chen et al. | An edge traffic flow detection scheme based on deep learning in an intelligent transportation system | |
JP6756037B2 (ja) | ユーザ本人確認の方法、装置及びシステム | |
US10726335B2 (en) | Generating compressed representation neural networks having high degree of accuracy | |
CN110941855B (zh) | 一种AIoT场景下的神经网络模型窃取防御方法 | |
CN114144781A (zh) | 身份验证和管理系统 | |
CN102611692B (zh) | 多承租人数据中心中的安全计算方法 | |
CA3063126A1 (en) | System and method for biometric identification | |
CN113505882B (zh) | 基于联邦神经网络模型的数据处理方法、相关设备及介质 | |
CN111680672B (zh) | 人脸活体检测方法、系统、装置、计算机设备和存储介质 | |
CN111428887B (zh) | 一种基于多个计算节点的模型训练控制方法、装置及系统 | |
CN113537633B (zh) | 基于纵向联邦学习的预测方法、装置、设备、介质和系统 | |
CN111191267B (zh) | 一种模型数据的处理方法、装置及设备 | |
CN111612167B (zh) | 机器学习模型的联合训练方法、装置、设备及存储介质 | |
CN110874571A (zh) | 人脸识别模型的训练方法及装置 | |
KR20220076398A (ko) | Ar장치를 위한 객체 인식 처리 장치 및 방법 | |
Ma et al. | Trusted ai in multiagent systems: An overview of privacy and security for distributed learning | |
CN114301850A (zh) | 一种基于生成对抗网络与模型压缩的军用通信加密流量识别方法 | |
KR20220138696A (ko) | 이미지 분류 방법 및 장치 | |
CN112948883A (zh) | 保护隐私数据的多方联合建模的方法、装置和系统 | |
RU2704538C1 (ru) | Сетевая архитектура человекоподобной сети и способ реализации | |
US20230133033A1 (en) | System and method for processing a data subject rights request using biometric data matching | |
CN112291188B (zh) | 注册验证方法及系统、注册验证服务器、云服务器 | |
CN114004265A (zh) | 一种模型训练方法及节点设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21833695 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21833695 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03.07.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21833695 Country of ref document: EP Kind code of ref document: A1 |