WO2023197927A1 - Model fairness evaluation methods and apparatus - Google Patents

Model fairness evaluation methods and apparatus Download PDF

Info

Publication number
WO2023197927A1
WO2023197927A1 PCT/CN2023/086570 CN2023086570W WO2023197927A1 WO 2023197927 A1 WO2023197927 A1 WO 2023197927A1 CN 2023086570 W CN2023086570 W CN 2023086570W WO 2023197927 A1 WO2023197927 A1 WO 2023197927A1
Authority
WO
WIPO (PCT)
Prior art keywords
sample
evaluated
model
evaluation
fairness
Prior art date
Application number
PCT/CN2023/086570
Other languages
French (fr)
Chinese (zh)
Inventor
李进锋
刘翔宇
张�荣
Original Assignee
阿里巴巴(中国)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴(中国)有限公司 filed Critical 阿里巴巴(中国)有限公司
Publication of WO2023197927A1 publication Critical patent/WO2023197927A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures

Definitions

  • the embodiments of this specification relate to the field of computer technology, and in particular to two model fairness evaluation methods.
  • a model fairness assessment method including:
  • a fairness evaluation is performed on the model to be evaluated.
  • a sample processing module configured to, when the credibility detection result satisfies the non-credibility condition, process the sample according to the The credibility test results are used to perform sample processing on the sample to be evaluated to obtain an updated evaluation sample;
  • the first determination module is configured to determine the real data probability distribution of the image and/or text training samples based on the image and/or text training model;
  • the second determination module is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
  • a sample update module configured to perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample when the credibility detection result satisfies the non-credibility condition
  • the result display module is configured to obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
  • a computing device including:
  • the memory is used to store computer-executable instructions
  • the processor is used to execute the computer-executable instructions.
  • the steps of the above model fairness evaluation method are implemented.
  • a computer-readable storage medium which stores computer-executable instructions.
  • the steps of the above model fairness evaluation method are implemented.
  • a computer program is provided, wherein when the computer program is executed in a computer, the computer is caused to perform the steps of the above model fairness evaluation method.
  • the model fairness evaluation method uses a graphic training model to model the real data probability distribution of graphic training samples, and conducts credibility testing on the samples to be evaluated based on the real data probability distribution; and targets non-credible samples. Process and obtain updated evaluation samples to improve the reliability and completeness of the evaluation samples in untrusted environments, thereby ensuring the model's fairness evaluation method and its robustness in both trusted and untrusted environments. and the availability of model evaluation results, thereby ensuring the accuracy of the fairness evaluation of the model to be evaluated through the model fairness evaluation method, so that the subsequent models to be evaluated can be used in actual applications, whether from meeting regulatory compliance or improving user experience. angles, all have good results and can be applied to algorithmic governance.
  • Figure 3 is a process flow chart of a model fairness evaluation method provided by an embodiment of this specification
  • Figure 4 is a schematic structural diagram of a model fairness evaluation device provided by an embodiment of this specification.
  • Figure 5 is a flow chart of another model fairness evaluation method provided by an embodiment of this specification.
  • Figure 7 is a structural block diagram of a computing device provided by an embodiment of this specification.
  • Algorithm fairness The automated decision-making of artificial intelligence algorithms is independent of protected sensitive attributes (natural attributes and social attributes) such as ethnicity, belief, and region. That is, for protected sensitive attributes, artificial intelligence algorithm decisions have no impact on individuals or groups. Prejudice or preference arising from inherent or acquired attributes.
  • embodiments of this specification provide a fairness evaluation system that has the ability to evaluate the fairness of some artificial intelligence algorithm tasks (such as text classification, image classification, etc.), but the fairness evaluation system Only the fairness evaluation of natural inputs in a trusted environment is considered, and the quantification of fairness relies entirely on statistical indicators such as accuracy, recall, and F1-score. In an uncontrolled environment (non-trusted environment), these statistical indicators will be greatly affected by factors such as adversarial perturbation and data selection. Therefore, the fairness evaluation results produced by the above system will not accurately reflect the algorithm ( model) itself, so the validity and usability of the evaluation cannot be guaranteed.
  • Figure 1 shows a schematic diagram of a specific scenario of a model fairness assessment method provided according to an embodiment of this specification, which specifically includes the following steps.
  • Step 102 Based on the large-scale graphic and text training samples collected in the database, use self-supervised learning technology to train a large-scale pre-training model (picture and/or text training model), and initially model the number of graphic and text training samples through the pre-training model. According to the probability distribution; and, based on the pre-training model, deep generation technology is used to construct a generative adversarial network model, and the data probability distribution of graphic training samples is further optimized through the generative adversarial network model.
  • a large-scale pre-training model picture and/or text training model
  • graphic training samples can be understood as picture and text training samples.
  • the fairness evaluation of the model to be evaluated in the embodiments of this specification can be understood as a statistical model performance difference indicator for different groups of the model to be evaluated based on sensitive/protected attributes (such as ethnicity, belief, income, etc.). If false Positive rate, statistical equality, equal opportunity, inconsistent impact and other indicators. Subsequently, the fairness of the model to be tested can be evaluated based on these indicators.
  • Step 106 Return the fairness evaluation result of the model to be evaluated to the user.
  • the model fairness evaluation method provided by the embodiment of this specification proposes a robust graphic algorithm fairness evaluation system, which models the data probability distribution by combining large-scale pre-training technology and deep generation technology, and treats evaluation samples according to the probability distribution. Carry out reliability testing, and perform denoising, adversarial reconstruction and diversity generation respectively for untrusted samples such as adversarial samples or distribution deviation samples, so as to improve the reliability and integrity of evaluation samples in untrusted environments, thereby Ensure the robustness of the fairness evaluation system in an uncontrolled environment and the availability of evaluation results.
  • Figure 2 shows a flow chart of a model fairness assessment method provided by an embodiment of this specification, which specifically includes the following steps.
  • Step 202 Based on the image and/or text training model, determine the real data probability distribution of the image and/or text training sample.
  • pictures include but are not limited to pictures of any type, any size, and contain any content, such as pictures of animals or people; texts include but are not limited to any type, any length, and contain any content, such as academic discussions, literary articles, etc.
  • large-scale image and/or text training samples will first be used to train the image and/or text training model, and then based on The trained image and/or text training model models the real data probability distribution of the image and/or text training sample; then the real data probability distribution is tuned by generating an adversarial network model to obtain the optimized real data. Probability distributions.
  • the specific implementation method is as follows:
  • Determining the real data probability distribution of the picture and/or text training samples based on the picture and/or text training model include:
  • the real data probability distribution of the training sample is adjusted according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample.
  • the image and/or text training model can be understood as a visual Transformer model, etc.; when the training sample is text, the image and/or text training model can be understood as a language model, BERT model, etc.; when the training sample is a graphic training sample, In this case, the image and/or text training model can be understood as a multi-modal fusion model that combines the visual Transformer model and the language model BERT model.
  • the model fairness evaluation method provided by the embodiments of this specification first trains the picture and/or text training model through large-scale picture and/or text training samples, and initially models the picture and/or text based on the picture and/or text training model.
  • the real data probability distribution of the training samples is then optimized based on the generative adversarial network model constructed with deep generation technology to optimize the data probability distribution of the initially modeled image and/or text training samples to determine the accuracy of the image and/or text training samples. performance and availability.
  • the step of adjusting the real data probability distribution of the training sample according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample includes:
  • the real data probability distribution of the training sample is adjusted according to the discrimination module to obtain the adjusted real data probability distribution of the training sample.
  • the construction phase of the generative adversarial network model includes two parts.
  • the first part is the construction of the generative adversarial network model
  • the second part is the training of the generative adversarial network model.
  • the generative adversarial network model consists of two parts: the discriminant module and the generation module. Then when constructing the generative adversarial network model, the image and/or text training sample model obtained by training in the above embodiment can be used as the discriminant module of the generative adversarial network model; and for the generation module, if image data is generated, you can use Multiple upsampling deconvolution networks build a generation module for a generative adversarial network model. If text data is generated, Transformer can be used as the generation module for a generative adversarial network model.
  • the specific implementation method is as follows:
  • the method of constructing a generative adversarial network model based on the picture and/or text training model includes:
  • the generative adversarial network model is constructed according to the discriminating module and the generating module.
  • the image and/or text training sample model obtained by training in the above embodiment is used as the discriminant module of the generative adversarial network model.
  • the discriminant module of the generative adversarial network model is initialized according to the model parameters of the image and/or text training model.
  • module parameters to build the discriminant module of the generative adversarial network model; and the generation module can be constructed based on the type of data to be generated by selecting a deconvolution network or a text generation network; the final generated discriminant module and generation module are constructed and generated Adversarial network model.
  • the generative adversarial network model can be trained; specifically, the generation module and the discriminating module of the generative adversarial network model are alternately trained by constructing a zero-sum game adversarial loss function, so that the generation module generates The data can be closer to the real data distribution, and at the same time, the discriminant module can better distinguish between real data and generated data.
  • the parameters of the pre-training model i.e., the model parameters of the picture and/or text training model
  • the discriminator i.e., the discriminant module.
  • the model fairness evaluation method constructs a generative adversarial network model based on the picture and/or text training model, and trains the generative adversarial network model based on the real data probability distribution of the picture and/or text training samples, Subsequently, according to the generated adversarial network model obtained through training, the real data probability distribution of the picture and/or text training sample can be adjusted to obtain the adjusted real data probability distribution of the picture and/or text training sample to improve the picture and/or text training sample. Or the authenticity of text training samples.
  • Step 204 Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model.
  • determining the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model includes:
  • sample data probability distribution determine the similarity of the sample to be evaluated belonging to the real data probability distribution of the training sample
  • the discriminant module of the generative adversarial network model obtain the sample prediction result of the sample to be evaluated
  • the credibility detection result of the sample to be evaluated is determined.
  • the real data probability distribution in the following embodiments can be understood as the real data probability distribution adjusted by the picture and/or text training sample; and the sample prediction result of the sample to be evaluated can be understood as the sample prediction result of the sample to be evaluated. Samples are generated samples or real samples.
  • the model fairness evaluation method performs credibility detection on the samples to be evaluated based on the target probability distribution of the image and/or text training samples and the generated adversarial network model, thereby determining whether the evaluation samples include adversarial samples or
  • subsequent processing of the samples to be evaluated can be performed to improve the reliability and integrity of the samples to be evaluated in an untrustworthy environment. sex.
  • the credibility test of the sample to be evaluated can be understood as the detection of whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated.
  • the specific implementation method is as follows:
  • Determining the credibility test result of the sample to be evaluated based on the similarity and the sample prediction result includes:
  • the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated.
  • the statistics of the samples to be evaluated belong to the logarithmic likelihood distribution in the real data distribution. If the distribution is more divergent, it means that the distribution diversity of the samples to be evaluated is stronger, and the distribution is more diverse. Concentration can mean that the distribution diversity of the samples to be evaluated is weaker.
  • the detection of the distribution diversity of the sample to be evaluated is the detection of the distribution diversity of the entire sample to be evaluated, rather than the measurement of a single sample to be evaluated, which can be measured by variance, standard deviation, median, central tendency, etc.
  • the indicator of distribution divergence is used to detect the distribution diversity of the samples to be evaluated. The embodiments of this specification do not impose any limitations on this.
  • Step 206 If the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample.
  • the sample to be evaluated when it is determined that the sample to be evaluated is an adversarial sample, the sample can be reconstructed through denoising and adversarial reconstruction; and when it is determined that the distribution diversity of the sample to be evaluated is weak, then it can be Diversity generation is performed on the sample to be evaluated.
  • the specific implementation method is as follows:
  • sample processing is performed on the sample to be evaluated according to the credibility test result to obtain an updated evaluation sample, including:
  • the sample to be evaluated is an adversarial sample
  • sample processing is performed on the samples to be evaluated according to the second preset processing method to obtain a second updated evaluation sample.
  • the credibility detection of the sample to be evaluated it can be detected whether the sample to be evaluated is an adversarial sample and whether the distribution diversity of the sample to be evaluated is weak; Even when the distribution diversity of the evaluation sample is weak, the sample to be evaluated can be considered to be untrustworthy; in this case, the sample to be evaluated needs to be processed.
  • the sample to be evaluated when the credibility test result of the sample to be evaluated is, and the sample to be evaluated is an adversarial sample, it can be determined that the credibility test result of the sample to be evaluated satisfies the non-credibility condition. At this time, then The sample to be evaluated can be processed according to the first preset processing method to obtain the first updated evaluation sample; and when the credibility test result of the sample to be evaluated is that the distribution diversity of the sample to be evaluated is weak, It can be determined that the credibility test result of the sample to be evaluated satisfies the non-credibility condition. At this time, the sample to be evaluated can be processed according to the second preset processing method to obtain a second updated evaluation sample.
  • the sample to be evaluated is processed according to the first preset processing method to obtain the specific details of the first updated evaluation sample.
  • the method is as follows:
  • performing sample processing on the sample to be evaluated according to a first preset processing method to obtain a first updated evaluation sample includes:
  • the sample to be evaluated is processed according to the second preset processing method, and the specific processing method for obtaining the second updated evaluation sample is as follows:
  • sample processing is performed on the samples to be evaluated according to the second preset processing method to obtain the second updated evaluation sample, including:
  • the sample data probability distribution is used to generate new evaluation samples through the generation module of the generative adversarial network model
  • adversarial reconstruction and diversity generation of the sample to be evaluated can be performed, where adversarial reconstruction can be understood as, in the sample to be evaluated,
  • the detected adversarial samples are reconstructed using denoising methods including but not limited to data compression, data randomization, adversarial error correction, etc. to eliminate the interference of adversarial noise; diversity generation can be understood as based on the evaluation to be
  • the sample data probability distribution of the sample uses the generation module of the generative adversarial network model to generate new evaluation samples to expand the diversity of samples to be evaluated.
  • the model fairness evaluation method provided by the embodiments of this specification, after obtaining the log likelihood and sample prediction results that the sample to be evaluated belongs to the real data distribution, can then evaluate the model fairness based on the log likelihood and sample prediction results.
  • the sample is tested for credibility, that is, whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated is tested.
  • the sample to be evaluated is an adversarial sample or the distribution diversity of the sample to be evaluated is weak
  • subsequent data processing can be performed on the untrustworthy sample to be evaluated to improve the reliability and integrity of the sample to be evaluated in an untrustworthy environment, thereby ensuring that the fairness assessment method is used in Robustness and availability of measurement results in untrusted environments.
  • the quality of the new evaluation samples will also be tested based on the discriminant module of the generative adversarial network model to ensure Added accuracy rate of new evaluation samples.
  • the second updated evaluation sample is obtained based on the newly added evaluation sample, including:
  • the new evaluation sample is deleted to obtain a second updated evaluation sample.
  • the model fairness evaluation method in order to ensure the accuracy of the newly added evaluation samples, after generating the new evaluation samples through the generation module of the generative adversarial network model, it will also be based on the generation quality of the new evaluation samples. Perform sample filtering to ensure the accuracy of newly added evaluation samples.
  • the new evaluation samples can also be generated and filtered based on the quality evaluation indicators of the original samples to be evaluated. For example, if the original samples to be evaluated are images, they can be quality filtered based on the Inception Score (initial score) of the images. When the original sample to be evaluated is text, quality filtering can be performed on it based on the fluency of the text.
  • Step 208 Conduct a fairness assessment on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample.
  • the updated evaluation sample includes the first updated evaluation sample and/or the second updated evaluation sample; and the model to be evaluated can also be understood as an image and text recognition model of the same type as the training model.
  • the fairness evaluation of the model to be evaluated based on the sample to be evaluated and the updated evaluation sample includes:
  • the mixed evaluation sample and the model to be evaluated are input into the fairness evaluation module to obtain the fairness evaluation index of the model to be evaluated, including:
  • the predicted value is the output of the model to be evaluated based on the mixed evaluation sample.
  • the mixed evaluation sample and the model to be evaluated are input into the fairness evaluation module.
  • the fairness evaluation module the mixed evaluation sample is input into the model to be evaluated, and the predicted value of the mixed evaluation sample output by the model to be evaluated is obtained; according to The predicted value and the true value of the mixed evaluation sample are used to calculate the prediction accuracy of the model to be evaluated; and after determining the prediction accuracy of the model to be evaluated, the fairness assessment module calculates the prediction accuracy of the model to be evaluated.
  • the model's false positive rate, statistical equality, equal opportunity, inconsistent impact and other indicators; subsequent users or the system can evaluate the fairness of the model to be evaluated based on these indicators.
  • the model fairness evaluation method uses a graphic training model to model the real data probability distribution of graphic training samples, and conducts credibility detection on the samples to be evaluated based on the real data probability distribution; and for non- Trusted samples are processed to obtain updated evaluation samples, thereby improving the reliability and completeness of evaluation samples in non-trusted environments, thereby ensuring the fairness evaluation method of the model, both in trusted environments and non-trusted environments.
  • Robustness and availability of model evaluation results thereby ensuring the accuracy of the fairness evaluation of the model to be evaluated through this model fairness evaluation method, so that subsequent models to be evaluated can be used in practical applications, whether from meeting regulatory compliance or improving From a user experience perspective, both have good results.
  • FIG. 3 shows a process flow chart of a model fairness evaluation method provided by an embodiment of this specification, which specifically includes the following steps.
  • Step 302 Based on the large-scale unsupervised graphic and text training samples collected, use self-supervised learning technology to train a pre-training model, and use the pre-training model to initially model the real data probability distribution of the graphic and text training samples.
  • Step 304 Use deep generation technology to build a generative adversarial network model based on the pre-trained model, and train based on graphics and text Sample training generates an adversarial network model, and optimizes the real data probability distribution of graphic and text training samples based on the generative adversarial network model to obtain the optimized real data probability distribution of graphic and text training samples.
  • Step 306 Based on the optimized real data probability distribution of the graphic training samples and the generated adversarial network model, perform a credibility test on the samples to be evaluated.
  • Step 308 Based on the credibility detection results of the samples to be evaluated, use adversarial defense technology to perform adversarial reconstruction of the samples to be evaluated, or use a generator that generates an adversarial network model to generate diversity for the samples to be evaluated.
  • Step 310 Mix the sample to be evaluated with the adversarially reconstructed sample to be evaluated and/or the diversity-generated sample to be evaluated to obtain a mixed sample, and input the mixed sample and the recommended model into the fairness evaluation module for evaluation to obtain the recommendation. Model fairness evaluation results.
  • the model fairness evaluation method proposes algorithm fairness evaluation technology in an uncontrolled (trusted) environment, based on large-scale and easy-to-obtain unsupervised graphic and text training data, and utilizing large-scale pre-training technology Model the probability distribution of real data by combining it with deep generation technology, and detect untrustworthy samples in the samples to be evaluated based on the probability distribution of the real data (unsupervised graphic training data) obtained through modeling, and use adversarial defense technology to counteract them.
  • noisy reconstruction can effectively eliminate the impact of adversarial noise on fairness evaluation.
  • the model fairness evaluation method in the embodiment of this specification proposes a data distribution based on the sample to be evaluated itself, while combining depth generation technology to achieve diversity Generation can achieve the purpose of improving the reliability and integrity of samples to be evaluated in untrusted environments.
  • the model fairness evaluation method provided by the embodiments of this specification does not require additional evaluation data or manual intervention, which greatly improves the intelligence of the evaluation and also reduces the evaluation cost. .
  • the model fairness evaluation method not only has the ability to evaluate fairness in a trusted environment, but also ensures the robustness of the fairness evaluation and the availability of the evaluation results in an uncontrolled environment. Therefore, it will be applicable to the fairness assessment of algorithms on e-commerce platforms, online social platforms, online social media and other platforms, including but not limited to intelligent customer service, personalized recommendations, intelligent risk control, etc., to eliminate algorithm bias and promote algorithms to meet regulatory compliance. and improve user experience.
  • the probability distribution determination module 402 is configured to determine the real data probability distribution of the picture and/or text training sample according to the picture and/or text training model;
  • the detection result determination module 404 is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
  • the sample processing module 406 is configured to perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample when the credibility detection result satisfies the non-credibility condition;
  • the evaluation module 408 is configured to perform a fairness evaluation on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample.
  • the probability distribution determination module 402 is further configured to:
  • the real data probability distribution of the training sample is adjusted according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample.
  • the probability distribution determination module 402 is further configured to:
  • the probability distribution determination module 402 is further configured to:
  • the generative adversarial network model is constructed according to the discriminating module and the generating module.
  • the detection result determination module 404 is further configured to:
  • sample data probability distribution determine the similarity of the sample to be evaluated belonging to the real data probability distribution of the training sample
  • the discriminant module of the generative adversarial network model obtain the sample prediction result of the sample to be evaluated
  • the credibility detection result of the sample to be evaluated is determined.
  • the detection result determination module 404 is further configured to:
  • the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated.
  • sample processing module 406 is further configured to:
  • the sample to be evaluated is an adversarial sample
  • sample processing module 406 is further configured to:
  • the method can be used to remove errors through data compression, data randomization or adversarial error correction.
  • the noise method is used to reconstruct the sample to be evaluated to obtain the first updated evaluation sample.
  • sample processing module 406 is further configured to:
  • sample processing module 406 is further configured to:
  • the evaluation module 408 is further configured to:
  • the evaluation module 408 is further configured to:
  • the predicted value is the output of the model to be evaluated based on the mixed evaluation sample.
  • model fairness evaluation device The above is a schematic solution of a model fairness evaluation device in this embodiment. It should be noted that the technical solution of the model fairness evaluation device and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. For details that are not described in detail in the technical solution of the model fairness evaluation device, please refer to the above-mentioned model fairness. Description of technical solutions for sexual assessment methods.
  • Step 502 Determine the real data probability distribution of the image and/or text training samples based on the image and/or text training model.
  • Step 504 Receive the sample to be evaluated and the model to be evaluated sent by the user.
  • Step 506 Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model.
  • Step 508 If the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample.
  • Step 512 Obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
  • the model fairness evaluation method provided by the embodiments of this specification is applied to the model fairness evaluation platform.
  • the actual application scenario can be that if a user wants to conduct a fairness evaluation on his or her project model on the model fairness evaluation platform, the user can send the sample to be evaluated and the model to be evaluated to the model fairness evaluation platform.
  • the model fairness evaluation platform After receiving the samples to be evaluated and the models to be evaluated sent by the user, adversarial reconstruction or diversity generation of the samples to be evaluated can be performed according to the above embodiment, thereby ensuring that the evaluation results given by the model fairness evaluation platform are closer to each other. The actual situation of the model to be evaluated.
  • the model fairness evaluation method provided by the embodiments of this specification is first based on large-scale and easy-to-obtain unsupervised graphic training data, and uses a combination of large-scale pre-training technology and deep generation technology to model the training data probability distribution as real data Probability distributions.
  • the reliability of the samples to be evaluated will be tested based on the probability distribution of real data, which effectively alleviates the interference of untrustworthy samples on the evaluation results.
  • this solution denoises and reconstructs the evaluation samples based on adversarial defense technology and generates diversity based on deep generation models, achieving improved performance in untrusted environments.
  • the purpose of ensuring the reliability and integrity of the evaluation samples is to ensure the robustness of the fairness evaluation system in uncontrolled environments and the availability of evaluation results, and effectively make up for the existing system's resistance to disturbance and evaluation in uncontrolled environments.
  • FIG. 6 shows a schematic structural diagram of another model fairness evaluation device provided by an embodiment of this specification. As shown in Figure 6, the device is applied to the model fairness evaluation platform, including:
  • the first determination module 602 is configured to determine the real data probability distribution of the image and/or text training samples according to the image and/or text training model;
  • the data receiving module 604 is configured to receive samples to be evaluated and models to be evaluated sent by the user;
  • the second determination module 606 is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
  • the sample update module 608 is configured to perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample when the credibility detection result satisfies the non-credibility condition;
  • the fairness evaluation module 610 is configured to perform a fairness evaluation on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample;
  • the result display module 612 is configured to obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
  • the model fairness evaluation device provided by the embodiments of this specification is first based on large-scale and easy-to-obtain unsupervised graphic training data, and uses a combination of large-scale pre-training technology and deep generation technology to model the training data probability distribution as real data Probability distributions.
  • the reliability of the samples to be evaluated will be tested based on the probability distribution of real data, which effectively alleviates the interference of untrustworthy samples on the evaluation results.
  • this solution denoises and reconstructs the evaluation samples based on adversarial defense technology and generates diversity based on deep generation models, achieving improved performance in untrusted environments.
  • the purpose of ensuring the reliability and integrity of the evaluation samples is to ensure the robustness of the fairness evaluation system in uncontrolled environments and the availability of evaluation results, and effectively make up for the existing system's resistance to disturbance and evaluation in uncontrolled environments.
  • model fairness evaluation device The above is a schematic solution of a model fairness evaluation device in this embodiment. It should be noted that the technical solution of the model fairness evaluation device and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. For details that are not described in detail in the technical solution of the model fairness evaluation device, please refer to the above-mentioned model fairness. Description of technical solutions for sexual assessment methods.
  • Figure 7 shows a structural block diagram of a computing device 700 provided according to an embodiment of this specification.
  • Components of the computing device 700 include, but are not limited to, memory 710 and processor 720 .
  • the processor 720 and the memory 710 are connected through a bus 730, and the database 750 is used to save data.
  • Computing device 700 also includes an access device 740 that enables computing device 700 to communicate via one or more networks 760 .
  • networks include the Public Switched Telephone Network (PSTN), a local area network (LAN), a wide area network (WAN), a personal area network (PAN), or a combination of communications networks such as the Internet.
  • Access device 740 may include one or more of any type of network interface (eg, a network interface card (NIC)), wired or wireless, such as an IEEE 802.11 Wireless Local Area Network (WLAN) wireless interface, Global Interconnection for Microwave Access ( Wi-MAX) interface, Ethernet interface, Universal Serial Bus (USB) interface, cellular network interface, Bluetooth interface, Near Field Communication (NFC) interface, etc.
  • NIC network interface card
  • the above-mentioned components of the computing device 700 and other components not shown in FIG. 7 may also be connected to each other, such as through a bus. It should be understood that the structural block diagram of the computing device shown in FIG. 7 is for illustrative purposes only and does not limit the scope of this description. Those skilled in the art can add or replace other components as needed.
  • Computing device 700 may be any type of stationary or mobile computing device, including a mobile computer or mobile computing device (e.g., tablet computer, personal digital assistant, laptop computer, notebook computer, netbook, etc.), a mobile telephone (e.g., smartphone ), a wearable computing device (e.g., smart watch, smart glasses, etc.) or other type of mobile device, or a stationary computing device such as a desktop computer or PC.
  • a mobile computer or mobile computing device e.g., tablet computer, personal digital assistant, laptop computer, notebook computer, netbook, etc.
  • a mobile telephone e.g., smartphone
  • a wearable computing device e.g., smart watch, smart glasses, etc.
  • stationary computing device such as a desktop computer or PC.
  • Computing device 700 may also be a mobile or stationary server.
  • the processor 720 is configured to execute the following computer-executable instructions. When the computer-executable instructions are executed by the processor, the steps of the above model fairness evaluation method are implemented.
  • the above is a schematic solution of a computing device in this embodiment. It should be noted that the technical solution of the computing device and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. Details that are not described in detail in the technical solution of the computing device can be found in the technical solution of the above-mentioned model fairness evaluation method. description of.
  • An embodiment of this specification also provides a computer-readable storage medium that stores computer-executable instructions.
  • the computer-executable instructions are executed by a processor, the steps of the above model fairness evaluation method are implemented.
  • An embodiment of this specification also provides a computer program, wherein when the computer program is executed in a computer, the computer is caused to perform the steps of the above model fairness evaluation method.
  • the computer instructions include computer program code, which may be in the form of source code, object code, executable file or some intermediate form.
  • the computer-readable medium may include: any entity or device capable of carrying the computer program code, recording media, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signals, telecommunications signals, and software distribution media, etc.
  • ROM Read-Only Memory
  • RAM Random Access Memory
  • electrical carrier signals telecommunications signals
  • software distribution media etc.
  • the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of legislation and patent practice in the jurisdiction.
  • the computer-readable medium Excludes electrical carrier signals and telecommunications signals.

Abstract

Provided in the embodiments of the present specification are model fairness evaluation methods and apparatus. One model fairness evaluation method comprises: according to a picture and/or text training model, determining real data probability distribution of a picture and/or text training sample; according to the real data probability distribution and a generative adversarial network model, determining a credibility detection result of a sample to be evaluated; when the credibility detection result meets an untrusted condition, performing sample processing on the sample to be evaluated according to the credibility detection result, so as to obtain an updated evaluation sample; and according to the sample to be evaluated and the updated evaluation sample, performing fairness evaluation on a model to be evaluated. The method can be applied to algorithm governance.

Description

模型公平性评估方法及装置Model fairness assessment method and device
本申请要求于2022年04月12日提交中国专利局、申请号为202210379396.X、申请名称为“模型公平性评估方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims priority to the Chinese patent application submitted to the China Patent Office on April 12, 2022, with application number 202210379396. Applying.
技术领域Technical field
本说明书实施例涉及计算机技术领域,特别涉及两种模型公平性评估方法。The embodiments of this specification relate to the field of computer technology, and in particular to two model fairness evaluation methods.
背景技术Background technique
随着人工智能基础理论与技术不断取得新的突破,以人工智能技术为基石的图文算法被广泛地应用于金融、教育、医疗、安防等多个公共领域,催生出了诸如智能安防、智能客服、医疗问诊、个性化推荐等一系列智能应用,不仅极大地丰富和便利了人们的日常生活,同时还推动和促进了社会经济和科技的发展和进步。As the basic theories and technologies of artificial intelligence continue to make new breakthroughs, graphic and text algorithms based on artificial intelligence technology have been widely used in many public fields such as finance, education, medical care, and security, giving rise to concepts such as intelligent security, intelligent A series of intelligent applications such as customer service, medical consultation, and personalized recommendations not only greatly enrich and facilitate people's daily lives, but also promote and promote the development and progress of social economy and technology.
然而,人工智能在自动化决策过程中存在不公平性甚至是歧视的问题饱受社会争议,不仅引发了人们对算法自动化决策的担忧和质疑,也逐渐引起了社会和公众的广泛关注,也有一些地方相继出台算法公平性相关法律法规,明确指出人工智能算法研发与应用必须满足公平性约束。因此,无论是从满足监管合规还是从提升用户体验的角度,开展算法公平性评估、消除算法偏见,在算法全生命周期中均是必不可少的步骤。However, the issue of unfairness and even discrimination in the automated decision-making process of artificial intelligence has been controversial in society. It has not only caused people's concerns and doubts about algorithm automated decision-making, but also gradually attracted widespread attention from society and the public. In some places Laws and regulations related to algorithm fairness have been promulgated one after another, clearly stating that the development and application of artificial intelligence algorithms must meet fairness constraints. Therefore, whether it is from the perspective of meeting regulatory compliance or improving user experience, conducting algorithm fairness assessments and eliminating algorithm bias are essential steps in the entire algorithm life cycle.
发明内容Contents of the invention
有鉴于此,本说明书实施例提供了两种模型公平性评估方法。本说明书一个或者多个实施例同时涉及两种模型公平性评估装置,一种计算设备,一种计算机可读存储介质以及一种计算机程序,以解决现有技术中存在的技术缺陷。In view of this, the embodiments of this specification provide two model fairness evaluation methods. One or more embodiments of this specification relate to two model fairness evaluation devices, a computing device, a computer-readable storage medium, and a computer program at the same time, so as to solve the technical deficiencies existing in the existing technology.
根据本说明书实施例的第一方面,提供了一种模型公平性评估方法,包括:According to the first aspect of the embodiment of this specification, a model fairness assessment method is provided, including:
根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;Based on the image and/or text training model, determine the real data probability distribution of the image and/or text training samples;
根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果;Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;When the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample;
根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估。According to the sample to be evaluated and the updated evaluation sample, a fairness evaluation is performed on the model to be evaluated.
根据本说明书实施例的第二方面,提供了一种模型公平性评估装置,包括:According to the second aspect of the embodiment of this specification, a model fairness evaluation device is provided, including:
概率分布确定模块,被配置为根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;a probability distribution determination module configured to determine the real data probability distribution of the image and/or text training sample based on the image and/or text training model;
检测结果确定模块,被配置为根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果;The detection result determination module is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
样本处理模块,被配置为在所述可信性检测结果满足非可信条件的情况下,根据所述 可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;A sample processing module configured to, when the credibility detection result satisfies the non-credibility condition, process the sample according to the The credibility test results are used to perform sample processing on the sample to be evaluated to obtain an updated evaluation sample;
评估模块,被配置为根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估。The evaluation module is configured to perform a fairness evaluation on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample.
根据本说明书实施例的第三方面,提供了一种模型公平性评估方法,应用于模型公平性评估平台,包括:According to the third aspect of the embodiment of this specification, a model fairness evaluation method is provided, which is applied to the model fairness evaluation platform, including:
根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;Based on the image and/or text training model, determine the real data probability distribution of the image and/or text training samples;
接收用户发送的待测评样本以及待测评模型;Receive samples to be evaluated and models to be evaluated sent by users;
根据所述真实数据概率分布、以及生成对抗网络模型,确定所述待测评样本的可信性检测结果;Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;When the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample;
根据所述待测评样本以及所述更新测评样本,对所述待测评模型进行公平性评估;Conduct a fairness assessment on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample;
获得所述待测评模型的公平性评估结果,并将所述公平性评估结果返回至所述用户。Obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
根据本说明书实施例的第四方面,提供了一种模型公平性评估方法,应用于模型公平性评估平台,包括:According to the fourth aspect of the embodiments of this specification, a model fairness evaluation method is provided, which is applied to a model fairness evaluation platform, including:
第一确定模块,被配置为根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;The first determination module is configured to determine the real data probability distribution of the image and/or text training samples based on the image and/or text training model;
数据接收模块,被配置为接收用户发送的待测评样本以及待测评模型;The data receiving module is configured to receive the samples to be evaluated and the models to be evaluated sent by the user;
第二确定模块,被配置为根据所述真实数据概率分布、以及生成对抗网络模型,确定所述待测评样本的可信性检测结果;The second determination module is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
样本更新模块,被配置为在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;A sample update module configured to perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample when the credibility detection result satisfies the non-credibility condition;
公平性评估模块,被配置为根据所述待测评样本以及所述更新测评样本,对所述待测评模型进行公平性评估;A fairness evaluation module configured to perform a fairness evaluation on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample;
结果展示模块,被配置为获得所述待测评模型的公平性评估结果,并将所述公平性评估结果返回至所述用户。The result display module is configured to obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
根据本说明书实施例的第五方面,提供了一种计算设备,包括:According to a fifth aspect of the embodiments of this specification, a computing device is provided, including:
存储器和处理器;memory and processor;
所述存储器用于存储计算机可执行指令,所述处理器用于执行所述计算机可执行指令,该计算机可执行指令被处理器执行时实现上述模型公平性评估方法的步骤。The memory is used to store computer-executable instructions, and the processor is used to execute the computer-executable instructions. When the computer-executable instructions are executed by the processor, the steps of the above model fairness evaluation method are implemented.
根据本说明书实施例的第六方面,提供了一种计算机可读存储介质,其存储有计算机可执行指令,该指令被处理器执行时实现上述模型公平性评估方法的步骤。According to a sixth aspect of the embodiments of this specification, a computer-readable storage medium is provided, which stores computer-executable instructions. When the instructions are executed by a processor, the steps of the above model fairness evaluation method are implemented.
根据本说明书实施例的第七方面,提供了一种计算机程序,其中,当所述计算机程序在计算机中执行时,令计算机执行上述模型公平性评估方法的步骤。According to a seventh aspect of the embodiments of this specification, a computer program is provided, wherein when the computer program is executed in a computer, the computer is caused to perform the steps of the above model fairness evaluation method.
本说明书一个实施例实现了两种模型公平性评估方法及装置,其中一种模型公平性评估方法包括根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分 布;根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果;在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估。One embodiment of this specification implements two model fairness evaluation methods and devices. One of the model fairness evaluation methods includes training a model based on pictures and/or text, and determining the real data probability score of the picture and/or text training sample. Distribution; determine the credibility detection result of the sample to be evaluated according to the real data probability distribution and the generated adversarial network model; when the credibility detection result satisfies the non-credibility condition, according to the credibility As a result of the detection, the sample to be evaluated is subjected to sample processing to obtain an updated evaluation sample; based on the sample to be evaluated and the updated evaluation sample, a fairness assessment is performed on the model to be evaluated.
具体的,所述模型公平性评估方法,通过图文训练模型,建模图文训练样本的真实数据概率分布,依据该真实数据概率分布对待测评样本进行可信性检测;并针对非可信样本进行处理,获得更新测评样本,以此提高在非可信环境下,测评样本的可靠性和完成性,从而保证该模型公平性评估方法,在可信环境下以及非可信环境下的健壮性和对模型测评结果的可用性,从而保证通过该模型公平性评估方法对待测评模型进行公平性评估的准确性,使得后续待测评模型在实际应用中无论是从满足监管合规还是从提升用户体验的角度,均有较好的效果,可应用于算法治理。Specifically, the model fairness evaluation method uses a graphic training model to model the real data probability distribution of graphic training samples, and conducts credibility testing on the samples to be evaluated based on the real data probability distribution; and targets non-credible samples. Process and obtain updated evaluation samples to improve the reliability and completeness of the evaluation samples in untrusted environments, thereby ensuring the model's fairness evaluation method and its robustness in both trusted and untrusted environments. and the availability of model evaluation results, thereby ensuring the accuracy of the fairness evaluation of the model to be evaluated through the model fairness evaluation method, so that the subsequent models to be evaluated can be used in actual applications, whether from meeting regulatory compliance or improving user experience. angles, all have good results and can be applied to algorithmic governance.
附图说明Description of the drawings
图1是本说明书一个实施例提供的一种模型公平性评估方法的具体场景示意图;Figure 1 is a schematic diagram of a specific scenario of a model fairness evaluation method provided by an embodiment of this specification;
图2是本说明书一个实施例提供的一种模型公平性评估方法的流程图;Figure 2 is a flow chart of a model fairness evaluation method provided by an embodiment of this specification;
图3是本说明书一个实施例提供的一种模型公平性评估方法的处理过程流程图;Figure 3 is a process flow chart of a model fairness evaluation method provided by an embodiment of this specification;
图4是本说明书一个实施例提供的一种模型公平性评估装置的结构示意图;Figure 4 is a schematic structural diagram of a model fairness evaluation device provided by an embodiment of this specification;
图5是本说明书一个实施例提供的另一种模型公平性评估方法的流程图;Figure 5 is a flow chart of another model fairness evaluation method provided by an embodiment of this specification;
图6是本说明书一个实施例提供的另一种模型公平性评估方法的结构示意图;Figure 6 is a schematic structural diagram of another model fairness evaluation method provided by an embodiment of this specification;
图7是本说明书一个实施例提供的一种计算设备的结构框图。Figure 7 is a structural block diagram of a computing device provided by an embodiment of this specification.
具体实施方式Detailed ways
在下面的描述中阐述了很多具体细节以便于充分理解本说明书。但是本说明书能够以很多不同于在此描述的其它方式来实施,本领域技术人员可以在不违背本说明书内涵的情况下做类似推广,因此本说明书不受下面公开的具体实施的限制。In the following description, numerous specific details are set forth to facilitate a thorough understanding of this specification. However, this specification can be implemented in many other ways different from those described here. Those skilled in the art can make similar extensions without violating the connotation of this specification. Therefore, this specification is not limited by the specific implementation disclosed below.
在本说明书一个或多个实施例中使用的术语是仅仅出于描述特定实施例的目的,而非旨在限制本说明书一个或多个实施例。在本说明书一个或多个实施例和所附权利要求书中所使用的单数形式的“一种”、“所述”和“该”也旨在包括多数形式,除非上下文清楚地表示其他含义。还应当理解,本说明书一个或多个实施例中使用的术语“和/或”是指并包含一个或多个相关联的列出项目的任何或所有可能组合。The terminology used in one or more embodiments of this specification is for the purpose of describing particular embodiments only and is not intended to limit the one or more embodiments of this specification. As used in one or more embodiments of this specification and the appended claims, the singular forms "a," "the" and "the" are intended to include the plural forms as well, unless the context clearly dictates otherwise. It will also be understood that the term "and/or" as used in one or more embodiments of this specification refers to and includes any and all possible combinations of one or more of the associated listed items.
应当理解,尽管在本说明书一个或多个实施例中可能采用术语第一、第二等来描述各种信息,但这些信息不应限于这些术语。这些术语仅用来将同一类型的信息彼此区分开。例如,在不脱离本说明书一个或多个实施例范围的情况下,第一也可以被称为第二,类似地,第二也可以被称为第一。取决于语境,如在此所使用的词语“如果”可以被解释成为“在……时”或“当……时”或“响应于确定”。It should be understood that although the terms first, second, etc. may be used to describe various information in one or more embodiments of this specification, the information should not be limited to these terms. These terms are only used to distinguish information of the same type from each other. For example, without departing from the scope of one or more embodiments of this specification, the first may also be called the second, and similarly, the second may also be called the first. Depending on the context, the word "if" as used herein may be interpreted as "when" or "when" or "in response to determining."
首先,对本说明书一个或多个实施例涉及的名词术语进行解释。 First, terminology used in one or more embodiments of this specification will be explained.
图文算法:图文算法指图片算法和文本算法的总称,具体包括图片分类、人脸识别、目标检测、图片检索等以图片为输入的算法,以及文本分类、情感分析、机器翻译、对话生成等以文本为输入的算法。Graphic-text algorithm: Graphic-text algorithm refers to the general term for picture algorithms and text algorithms, specifically including picture classification, face recognition, target detection, picture retrieval and other algorithms that use pictures as input, as well as text classification, sentiment analysis, machine translation, and dialogue generation. and other algorithms that take text as input.
算法公平性:人工智能算法自动化决策独立于民族、信仰、地区等受保护的敏感属性(自然属性和社会属性),也即是,对于受保护敏感属性,人工智能算法决策对个人或群体不存在因其固有或后天属性所引起的偏见或偏爱。Algorithm fairness: The automated decision-making of artificial intelligence algorithms is independent of protected sensitive attributes (natural attributes and social attributes) such as ethnicity, belief, and region. That is, for protected sensitive attributes, artificial intelligence algorithm decisions have no impact on individuals or groups. Prejudice or preference arising from inherent or acquired attributes.
OOD数据:OOD(Out-of-Distribution)数据,也称分布外数据,是指样本数据来自不同于算法模型训练数据分布的数据。如果样本数据分布与算法模型训练数据分布相同,则称该数据为ID数据,即in-distribution数据。OOD data: OOD (Out-of-Distribution) data, also known as out-of-distribution data, means that the sample data comes from data that is different from the distribution of the algorithm model training data. If the sample data distribution is the same as the algorithm model training data distribution, the data is called ID data, that is, in-distribution data.
鲁棒性:鲁棒性也称健壮性或稳健性,是指一个计算机系统在一定参数(结构,大小)改变下,或在执行过程中处理错误以及算法在遭遇输入、运算等异常时继续正常运行且保障其性能稳定的能力。Robustness: Robustness, also known as robustness or robustness, refers to the fact that a computer system continues to operate normally when certain parameters (structure, size) change, or handles errors during execution, and the algorithm encounters abnormalities such as input and operations. The ability to operate and ensure stable performance.
随着人工智能基础理论与技术不断取得新的突破,以人工智能技术为基石的图文算法被广泛地应用于金融、教育、医疗、安防等多个公共领域,催生出了诸如智能安防、智能客服、医疗问诊、个性化推荐等一系列智能应用,不仅极大地丰富和便利了人们的日常生活,同时还推动和促进了社会经济和科技的发展和进步。As the basic theories and technologies of artificial intelligence continue to make new breakthroughs, graphic and text algorithms based on artificial intelligence technology have been widely used in many public fields such as finance, education, medical care, and security, giving rise to concepts such as intelligent security, intelligent A series of intelligent applications such as customer service, medical consultation, and personalized recommendations not only greatly enrich and facilitate people's daily lives, but also promote and promote the development and progress of social economy and technology.
然而,伴随着应用场景的日趋广泛,人工智能算法面临的法律、伦理问题和风险也日渐突出。例如,某种犯罪风险评估算法会对一种民族的人存在系统性歧视等。人工智能在自动化决策过程中存在不公平性甚至是歧视的问题饱受社会争议,不仅引发了人们对算法自动化决策的担忧和质疑,也逐渐引起了社会和公众的广泛关注,一些地方相继出台算法公平性相关法律法规,明确指出人工智能算法研发与应用必须满足公平性约束。因此,无论是从满足监管合规还是从提升用户体验的角度,开展算法公平性评估、消除算法偏见,在算法全生命周期中均是必不可少的步骤。However, as application scenarios become more and more widespread, the legal and ethical issues and risks faced by artificial intelligence algorithms are becoming increasingly prominent. For example, a certain crime risk assessment algorithm may systematically discriminate against people of one ethnic group. The issue of unfairness and even discrimination in the automated decision-making process of artificial intelligence has been controversial in society. It has not only caused people's concerns and doubts about automated decision-making by algorithms, but also gradually attracted widespread attention from society and the public. Some places have successively introduced algorithms Fairness-related laws and regulations clearly state that the development and application of artificial intelligence algorithms must meet fairness constraints. Therefore, whether it is from the perspective of meeting regulatory compliance or improving user experience, conducting algorithm fairness assessments and eliminating algorithm bias are essential steps in the entire algorithm life cycle.
针对上述问题,本说明书实施例提供了一种公平性评估系统,该公平性评估系统具有对部分人工智能算法任务(如文本分类、图像分类等)的公平性测评能力,但该公平性评估系统仅考虑了在可信环境下针对自然输入的公平性评估,并且其公平性的量化完全依赖于准确率、召回率、F1-score等统计意义上的指标。而在非受控环境(非可信环境)下,这些统计指标均会极大地受对抗扰动、数据选择等因素的影响,因此上述系统所产出的公平性评估结果将无法准确地反映算法(模型)本身的真实公平性水平,因此无法保证测评的有效性和可用性。In response to the above problems, embodiments of this specification provide a fairness evaluation system that has the ability to evaluate the fairness of some artificial intelligence algorithm tasks (such as text classification, image classification, etc.), but the fairness evaluation system Only the fairness evaluation of natural inputs in a trusted environment is considered, and the quantification of fairness relies entirely on statistical indicators such as accuracy, recall, and F1-score. In an uncontrolled environment (non-trusted environment), these statistical indicators will be greatly affected by factors such as adversarial perturbation and data selection. Therefore, the fairness evaluation results produced by the above system will not accurately reflect the algorithm ( model) itself, so the validity and usability of the evaluation cannot be guaranteed.
基于此,在本说明书中,提供了两种模型公平性评估方法。本说明书一个或者多个实施例同时涉及两种模型公平性评估装置,一种计算设备,一种计算机可读存储介质以及一种计算机程序,在下面的实施例中逐一进行详细说明。Based on this, in this specification, two model fairness evaluation methods are provided. One or more embodiments of this specification relate to two model fairness evaluation devices, a computing device, a computer-readable storage medium, and a computer program, which will be described in detail one by one in the following embodiments.
参见图1,图1示出了根据本说明书一个实施例提供的一种模型公平性评估方法的具体场景示意图,具体包括以下步骤。Referring to Figure 1, Figure 1 shows a schematic diagram of a specific scenario of a model fairness assessment method provided according to an embodiment of this specification, which specifically includes the following steps.
具体的,本说明书实施例提供的模型公平性评估方法应用于模型公平性评估平台。Specifically, the model fairness evaluation method provided by the embodiment of this specification is applied to the model fairness evaluation platform.
步骤102:根据数据库中收集的大规模的图文训练样本,利用自监督学习技术训练大规模预训练模型(图片和/或文本训练模型),通过预训练模型初始建模图文训练样本的数 据概率分布;并且,基于预训练模型,利用深度生成技术构建生成对抗网络模型,通过该生成对抗网络模型进一步优化图文训练样本的数据概率分布。Step 102: Based on the large-scale graphic and text training samples collected in the database, use self-supervised learning technology to train a large-scale pre-training model (picture and/or text training model), and initially model the number of graphic and text training samples through the pre-training model. According to the probability distribution; and, based on the pre-training model, deep generation technology is used to construct a generative adversarial network model, and the data probability distribution of graphic training samples is further optimized through the generative adversarial network model.
其中,图文训练样本可以理解为图片和文字训练样本。Among them, graphic training samples can be understood as picture and text training samples.
步骤104:接收用户发送的公平性评估请求,该公平性评估请求中携带有待测评样本以及待测评模型;首先,根据图文训练样本的数据概率分布以及生成对抗网络模型,对待测评样本进行可信性检测;其次,基于待测评样本的可信性检测结果,在确定待测评样本中包括对抗样本的情况下,利用对抗防御技术为对抗样本进行去噪和对抗重建,在确定待测评样本的分布多样性检测较弱的情况下,利用生成对抗网络模型对待测评样本进行多样性生成;最后将原始的待测评样本、对抗重建后的测评样本以及多样性生成得到的样本进行混合后,结合待测评模型输入到公平性评估模块进行测评,获得待测评模型的公平性评估结果。Step 104: Receive the fairness evaluation request sent by the user. The fairness evaluation request carries the sample to be evaluated and the model to be evaluated. First, based on the data probability distribution of the graphic training sample and the generated adversarial network model, the sample to be evaluated is trusted. Secondly, based on the credibility test results of the samples to be evaluated, when it is determined that the samples to be evaluated include adversarial samples, adversarial defense technology is used to denoise and adversarially reconstruct the adversarial samples. After determining the distribution of the samples to be evaluated, When the diversity detection is weak, the generative adversarial network model is used to generate diversity for the samples to be evaluated; finally, the original samples to be evaluated, the evaluation samples after adversarial reconstruction and the samples generated by diversity are mixed, and then combined with the samples to be evaluated The model is input to the fairness evaluation module for evaluation, and the fairness evaluation results of the model to be evaluated are obtained.
具体的,本说明书实施例对待测评模型的公平性评估,可以理解为统计待测评模型基于敏感/受保护属性(如民族、信仰、收益等)划分的不同群体上的模型性能差异指标,如假阳率、统计均等、机会均等、不一致影响等指标。后续可以根据这些指标评估待测评模型的公平性。Specifically, the fairness evaluation of the model to be evaluated in the embodiments of this specification can be understood as a statistical model performance difference indicator for different groups of the model to be evaluated based on sensitive/protected attributes (such as ethnicity, belief, income, etc.). If false Positive rate, statistical equality, equal opportunity, inconsistent impact and other indicators. Subsequently, the fairness of the model to be tested can be evaluated based on these indicators.
步骤106:将待测评模型的公平性评估结果返回至用户。Step 106: Return the fairness evaluation result of the model to be evaluated to the user.
本说明书实施例提供的模型公平性评估方法,提出了一种健壮的图文算法公平性评估系统,通过结合大规模预训练技术和深度生成技术建模数据概率分布,并依据概率分布对待测评样本进行可靠性检测,同时针对非可信样本如对抗样本或分布偏差样本,分别进行去噪、对抗重建和多样性生成,以此提高在非可信环境下测评样本的可靠性和完整性,从而保证公平性测评系统在非受控环境下的健壮性和测评结果的可用性。The model fairness evaluation method provided by the embodiment of this specification proposes a robust graphic algorithm fairness evaluation system, which models the data probability distribution by combining large-scale pre-training technology and deep generation technology, and treats evaluation samples according to the probability distribution. Carry out reliability testing, and perform denoising, adversarial reconstruction and diversity generation respectively for untrusted samples such as adversarial samples or distribution deviation samples, so as to improve the reliability and integrity of evaluation samples in untrusted environments, thereby Ensure the robustness of the fairness evaluation system in an uncontrolled environment and the availability of evaluation results.
参见图2,图2示出了本说明书一个实施例提供的一种模型公平性评估方法的流程图,具体包括以下步骤。Referring to Figure 2, Figure 2 shows a flow chart of a model fairness assessment method provided by an embodiment of this specification, which specifically includes the following steps.
步骤202:根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布。Step 202: Based on the image and/or text training model, determine the real data probability distribution of the image and/or text training sample.
其中,图片包括但不限于任意类型,任意大小,里面包含任意内容的图片,例如包含动物或者人的图片;文本包括但不限于任意类型,任意篇幅,里面包含任意内容的文本,例如学术论述、文学文章等。Among them, pictures include but are not limited to pictures of any type, any size, and contain any content, such as pictures of animals or people; texts include but are not limited to any type, any length, and contain any content, such as academic discussions, Literary articles, etc.
而图片和/或文本训练模型可以理解为,图片训练模型、文本训练模型或者图文结合的训练模型等;实际应用中,图片和/或文本训练模型具体为哪种类型的模型,可以根据实际需求进行确定,本说明书实施例对此不作任何限定。The picture and/or text training model can be understood as a picture training model, a text training model or a training model that combines pictures and text, etc.; in actual applications, the specific type of picture and/or text training model can be determined according to the actual situation. The requirements are determined, and the embodiments of this specification do not limit this in any way.
具体实施时,为了保证图片和/或文本训练样本的真实数据概率分布的准确性,会先利用大规模的图片和/或文本训练样本对图片和/或文本训练模型进行模型训练后,先根据训练后的图片和/或文本训练模型,建模该图片和/或文本训练样本的真实数据概率分布;再通过生成对抗网络模型对该真实数据概率分布进行调优,获得调优后的真实数据概率分布。具体实现方式如下所述:During specific implementation, in order to ensure the accuracy of the real data probability distribution of image and/or text training samples, large-scale image and/or text training samples will first be used to train the image and/or text training model, and then based on The trained image and/or text training model models the real data probability distribution of the image and/or text training sample; then the real data probability distribution is tuned by generating an adversarial network model to obtain the optimized real data. Probability distributions. The specific implementation method is as follows:
所述根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布, 包括:Determining the real data probability distribution of the picture and/or text training samples based on the picture and/or text training model, include:
获取图片和/或文本训练样本;Obtain image and/or text training samples;
根据所述训练样本利用自监督学习技术,训练获得图片和/或文本训练模型;Use self-supervised learning technology to train and obtain image and/or text training models based on the training samples;
根据所述图片和/或文本训练模型,获得所述训练样本调整后的真实数据概率分布;According to the picture and/or text training model, obtain the real data probability distribution adjusted by the training sample;
根据生成对抗网络模型对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布。The real data probability distribution of the training sample is adjusted according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample.
其中,为了保证图片和/或文本训练模型的准确率,本说明书实施例中,会获取大规模的图片和/或文本训练样本对其进行模型训练;而当训练样本为图片的情况下,该图片和/或文本训练模型可以理解为视觉Transformer模型等;当训练样本为文本的情况下,该图片和/或文本训练模型可以理解为语言模型BERT模型等;当训练样本为图文训练样本的情况下,该图片和/或文本训练模型则可以理解为视觉Transformer模型与语言模型BERT模型的结合多模态融合模型。Among them, in order to ensure the accuracy of the picture and/or text training model, in the embodiment of this specification, large-scale picture and/or text training samples will be obtained for model training; and when the training samples are pictures, the The image and/or text training model can be understood as a visual Transformer model, etc.; when the training sample is text, the image and/or text training model can be understood as a language model, BERT model, etc.; when the training sample is a graphic training sample, In this case, the image and/or text training model can be understood as a multi-modal fusion model that combines the visual Transformer model and the language model BERT model.
具体的,在获取大规模的图片和/或文本训练样本之后,会根据该大规模的图片和/或文本训练样本利用自监督学习技术,训练获得图片和/或文本训练模型;然后通过该图片和/或文本训练模型初始建模该图片和/或文本训练样本的真实数据概率分布。而为了进一步优化该图片和/或文本训练样本的真实数据概率分布,则可以根据生成对抗网络模型,对该图片和/或文本训练样本的真实数据概率分布进行调优,获得该图片和/或文本训练样本调整后的真实数据概率分布。Specifically, after obtaining a large-scale image and/or text training sample, self-supervised learning technology will be used to train the image and/or text training model based on the large-scale image and/or text training sample; and then use the image to and/or the text training model initially models the real data probability distribution of the image and/or text training sample. In order to further optimize the real data probability distribution of the picture and/or text training sample, the real data probability distribution of the picture and/or text training sample can be tuned according to the generative adversarial network model to obtain the picture and/or Probability distribution of real data adjusted for text training samples.
本说明书实施例提供的模型公平性评估方法,首先通过大规模的图片和/或文本训练样本,训练图片和/或文本训练模型,根据图片和/或文本训练模型初步建模图片和/或文本训练样本的真实数据概率分布,再根据深度生成技术构建的生成对抗网络模型,对初步建模的图片和/或文本训练样本的数据概率分布进行优化,从而确定图片和/或文本训练样本的准确性和可用性。The model fairness evaluation method provided by the embodiments of this specification first trains the picture and/or text training model through large-scale picture and/or text training samples, and initially models the picture and/or text based on the picture and/or text training model. The real data probability distribution of the training samples is then optimized based on the generative adversarial network model constructed with deep generation technology to optimize the data probability distribution of the initially modeled image and/or text training samples to determine the accuracy of the image and/or text training samples. performance and availability.
而在根据生成对抗网络模型对图片和/或文本训练样本的真实数据概率分布进行调优之前,需要根据图片和/或文本训练模型,利用深度生成技术构建生成对抗网络模型,以保证生成对抗网络模型的后续可用性。具体实现方式如下所述:Before tuning the real data probability distribution of image and/or text training samples based on the generative adversarial network model, it is necessary to use deep generation technology to build a generative adversarial network model based on the image and/or text training model to ensure that the generative adversarial network Subsequent availability of the model. The specific implementation method is as follows:
所述根据生成对抗网络模型对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布,包括:The step of adjusting the real data probability distribution of the training sample according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample includes:
根据所述图片和/或文本训练模型,构建生成对抗网络模型;Build a generative adversarial network model based on the image and/or text training model;
根据所述训练样本对所述生成对抗网络模型进行训练,获得训练后的生成对抗网络模型的判别模块和生成模块;Train the generative adversarial network model according to the training samples, and obtain the discriminating module and generating module of the trained generative adversarial network model;
根据所述判别模块对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布。The real data probability distribution of the training sample is adjusted according to the discrimination module to obtain the adjusted real data probability distribution of the training sample.
具体的,先根据图片和/或文本训练模型,构建生成对抗网络模型;再根据图片和/或文本训练样本对生成对抗网络模型进行训练,获得训练后的生成对抗网络模型的判别模块和生成模块;最后根据生成对抗网络模型的判别模块对图片和/或文本训练样本的初始概率分布进行微调,获得图片和/或文本训练样本微调后的真实数据概率分布。 Specifically, first build a generative adversarial network model based on pictures and/or text training models; then train the generative adversarial network model based on picture and/or text training samples to obtain the discriminant module and generation module of the trained generative adversarial network model. ; Finally, the initial probability distribution of the image and/or text training samples is fine-tuned according to the discriminant module of the generative adversarial network model, and the real data probability distribution after fine-tuning of the image and/or text training samples is obtained.
具体实施时,生成对抗网络模型的构建阶段包括两部分,第一部分为生成对抗网络模型的构建,第二部分为生成对抗网络模型的训练;其中,生成对抗网络模型由判别模块和生成模块两部分构建,那么在生成对抗网络模型的构建时,可以利用上述实施例训练获得的图片和/或文本训练样本模型作为生成对抗网络模型的判别模块;而对于生成模块,若生成图像数据,则可以采用多个上采样的反卷积网络构建生成对抗网络模型的生成模块,若生成文本数据,则可以采用Transformer作为生成对抗网络模型的生成模块。具体实现方式如下所述:In specific implementation, the construction phase of the generative adversarial network model includes two parts. The first part is the construction of the generative adversarial network model, and the second part is the training of the generative adversarial network model. Among them, the generative adversarial network model consists of two parts: the discriminant module and the generation module. Then when constructing the generative adversarial network model, the image and/or text training sample model obtained by training in the above embodiment can be used as the discriminant module of the generative adversarial network model; and for the generation module, if image data is generated, you can use Multiple upsampling deconvolution networks build a generation module for a generative adversarial network model. If text data is generated, Transformer can be used as the generation module for a generative adversarial network model. The specific implementation method is as follows:
所述根据所述图片和/或文本训练模型,构建生成对抗网络模型,包括:The method of constructing a generative adversarial network model based on the picture and/or text training model includes:
根据所述图片和/或文本训练模型的模型参数,对生成对抗网络模型的判别模块的模块参数进行初始化,构建所述生成对抗网络模型的判别模块;Initialize the module parameters of the discrimination module of the generative adversarial network model according to the model parameters of the picture and/or text training model, and construct the discrimination module of the generative adversarial network model;
根据反卷积网络和/或文本生成网络,构建所述生成对抗网络模型的生成模块;Construct a generation module of the generative adversarial network model according to the deconvolution network and/or text generation network;
根据所述判别模块和所述生成模块构建所述生成对抗网络模型。The generative adversarial network model is constructed according to the discriminating module and the generating module.
其中,利用上述实施例训练获得的图片和/或文本训练样本模型作为生成对抗网络模型的判别模块,可以理解为,根据图片和/或文本训练模型的模型参数,初始化生成对抗网络模型的判别模块的模块参数,以构建生成对抗网络模型的判别模块;而生成模块,则可以基于待生成的数据的类型,选择反卷积网络或者文本生成网络进行构建;最终生成的判别模块和生成模块构建生成对抗网络模型。Among them, the image and/or text training sample model obtained by training in the above embodiment is used as the discriminant module of the generative adversarial network model. It can be understood that the discriminant module of the generative adversarial network model is initialized according to the model parameters of the image and/or text training model. module parameters to build the discriminant module of the generative adversarial network model; and the generation module can be constructed based on the type of data to be generated by selecting a deconvolution network or a text generation network; the final generated discriminant module and generation module are constructed and generated Adversarial network model.
而在构建生成对抗网络模型之后,即可对该生成对抗网络模型进行训练;具体的,通过构建零和博弈对抗损失函数来交替训练生成对抗网络模型的生成模块和判别模块,使得生成模块生成的数据能更接近真实的数据分布,同时,判别模块又可以更好的区分真实数据和生成数据。After constructing the generative adversarial network model, the generative adversarial network model can be trained; specifically, the generation module and the discriminating module of the generative adversarial network model are alternately trained by constructing a zero-sum game adversarial loss function, so that the generation module generates The data can be closer to the real data distribution, and at the same time, the discriminant module can better distinguish between real data and generated data.
具体的,本说明书实施例中,利用预训练模型预训练的参数(即图片和/或文本训练模型的模型参数)去初始化判别器(即判别模块)的参数,好处在于,预训练模型是基于大规模图文训练样本训练得到的,通过预训练模型初始化判别器可以将预训练模型大规模的训练样本中学到的知识迁移到判别器中,即深度学习中的预训练加微调的技术实现。Specifically, in the embodiment of this specification, the parameters of the pre-training model (i.e., the model parameters of the picture and/or text training model) are used to initialize the parameters of the discriminator (i.e., the discriminant module). The advantage is that the pre-training model is based on It is obtained by training with large-scale image and text training samples. By initializing the discriminator through the pre-training model, the knowledge learned from the large-scale training samples of the pre-training model can be transferred to the discriminator, which is the technical implementation of pre-training and fine-tuning in deep learning.
本说明书实施例提供的模型公平性评估方法,根据图片和/或文本训练模型,构建生成对抗网络模型,并根据图片和/或文本训练样本的真实数据概率分布,对生成对抗网络模型进行训练,后续可以根据训练获得的生成对抗网络模型,对图片和/或文本训练样本的真实数据概率分布进行调优,获得该图片和/或文本训练样本调整后的真实数据概率分布,以提高图片和/或文本训练样本的真实性。The model fairness evaluation method provided by the embodiments of this specification constructs a generative adversarial network model based on the picture and/or text training model, and trains the generative adversarial network model based on the real data probability distribution of the picture and/or text training samples, Subsequently, according to the generated adversarial network model obtained through training, the real data probability distribution of the picture and/or text training sample can be adjusted to obtain the adjusted real data probability distribution of the picture and/or text training sample to improve the picture and/or text training sample. Or the authenticity of text training samples.
步骤204:根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果。Step 204: Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model.
而在获得图片和/或文本训练样本调整后的真实数据概率分布之后,即可结合生成对抗网络模型,对待测评样本的可信性进行检测。After obtaining the adjusted real data probability distribution of the image and/or text training samples, it can be combined with the generated adversarial network model to detect the credibility of the samples to be evaluated.
具体的,所述根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果,包括:Specifically, determining the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model includes:
根据所述图片和/或文本训练模型,获得所述待测评样本的样本数据概率分布; According to the picture and/or text training model, obtain the sample data probability distribution of the sample to be evaluated;
根据所述样本数据概率分布,确定所述待测评样本属于所述训练样本的真实数据概率分布的相似度;According to the sample data probability distribution, determine the similarity of the sample to be evaluated belonging to the real data probability distribution of the training sample;
根据所述生成对抗网络模型的判别模块,获得待测评样本的样本预测结果;According to the discriminant module of the generative adversarial network model, obtain the sample prediction result of the sample to be evaluated;
根据所述相似度以及所述样本预测结果,确定所述待测评样本的可信性检测结果。Based on the similarity and the sample prediction result, the credibility detection result of the sample to be evaluated is determined.
其中,为了便于理解,以下实施例中的真实数据概率分布,均可以理解为该图片和/或文本训练样本调整后的真实数据概率分布;而待测评样本的样本预测结果,可以理解为待测评样本为生成样本或者真实样本。Among them, for ease of understanding, the real data probability distribution in the following embodiments can be understood as the real data probability distribution adjusted by the picture and/or text training sample; and the sample prediction result of the sample to be evaluated can be understood as the sample prediction result of the sample to be evaluated. Samples are generated samples or real samples.
具体实施时,首先,根据图片和/或文本训练模型,获得待测评样本的样本数据概率分布,根据该待测评样本的样本数据概率分布,计算该待测评样本属于真实数据(图片和/或文本训练样本)分布中的相似度(即对数似然度);同时,根据生成对抗网络模型的判别模块,获得待测评样本的样本预测结果;然后根据该相似度以及样本预测结果,确定待测评样本的可信性检测结果。In specific implementation, first, train the model based on pictures and/or text to obtain the sample data probability distribution of the sample to be evaluated, and calculate whether the sample to be evaluated belongs to real data (pictures and/or text) based on the sample data probability distribution of the sample to be evaluated. training samples) distribution; at the same time, according to the discriminant module of the generative adversarial network model, the sample prediction results of the samples to be evaluated are obtained; and then based on the similarity and the sample prediction results, the sample to be evaluated is determined The credibility test results of the sample.
本说明书实施例提供的模型公平性评估方法,根据图片和/或文本训练样本的目标概率分布以及生成对抗网络模型,对待测评样本进行可信性检测,以此判断测评样本中是否包括对抗样本或者待测评样本的多样性较弱等情况,在确定待测评样本的可信性出现问题的情况下,可以对待测评样本进行后续处理,从而提高在非可信环境下待测评样本的可靠性和完整性。The model fairness evaluation method provided by the embodiments of this specification performs credibility detection on the samples to be evaluated based on the target probability distribution of the image and/or text training samples and the generated adversarial network model, thereby determining whether the evaluation samples include adversarial samples or When the diversity of the samples to be evaluated is weak and there is a problem with the credibility of the samples to be evaluated, subsequent processing of the samples to be evaluated can be performed to improve the reliability and integrity of the samples to be evaluated in an untrustworthy environment. sex.
实际应用中,对待测评样本的可信性检测,即可以理解为对待测评样本是否为对抗样本以及待测评样本的分布多样性的检测。具体实现方式如下所述:In practical applications, the credibility test of the sample to be evaluated can be understood as the detection of whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated. The specific implementation method is as follows:
所述根据所述相似度以及所述样本预测结果,确定所述待测评样本的可信性检测结果,包括:Determining the credibility test result of the sample to be evaluated based on the similarity and the sample prediction result includes:
根据所述相似度以及所述样本预测结果,确定所述待测评样本是否为对抗样本以及所述待测评样本的分布多样性。According to the similarity and the sample prediction result, it is determined whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated.
实际应用中,在获得对数似然度以及样本预测结果之后,则可以根据该对数似然度以及样本预测结果,判断待测评样本是否为OOD样本,如对抗样本,理论上对数似然度越小(实际应用中,可以根据样本集和模型任务为其设置一个对数似然度阈值,小于该阈值,则可以认为该对数似然度较小),该待测评样本为对抗样本的概率越大。In practical applications, after obtaining the log-likelihood and sample prediction results, you can judge whether the sample to be evaluated is an OOD sample based on the log-likelihood and sample prediction results. Such as adversarial samples, theoretical log-likelihood The smaller the degree (in practical applications, a log-likelihood threshold can be set based on the sample set and model task. If it is less than this threshold, the log-likelihood can be considered smaller), the sample to be evaluated is an adversarial sample. The greater the probability.
而对于待测评样本的分布多样性检测,统计待测评样本属于真实数据分布中的对数似然度的分布,如果该分布越发散则说明待测评样本的分布多样性越强,而该分布越集中则可以说明待测评样本的分布多样性越弱。实际应用中,待测评样本的分布多样性检测是对整个待测评样本的分布多样性的检测,而不是针对单个待测评样本的度量,可以通过方差、标准差、中位数、集中趋势等衡量分布散度的指标,对待测评样本的分布多样性进行检测,本说明书实施例对此不做任何限定。For the distribution diversity detection of the samples to be evaluated, the statistics of the samples to be evaluated belong to the logarithmic likelihood distribution in the real data distribution. If the distribution is more divergent, it means that the distribution diversity of the samples to be evaluated is stronger, and the distribution is more diverse. Concentration can mean that the distribution diversity of the samples to be evaluated is weaker. In practical applications, the detection of the distribution diversity of the sample to be evaluated is the detection of the distribution diversity of the entire sample to be evaluated, rather than the measurement of a single sample to be evaluated, which can be measured by variance, standard deviation, median, central tendency, etc. The indicator of distribution divergence is used to detect the distribution diversity of the samples to be evaluated. The embodiments of this specification do not impose any limitations on this.
本说明书实施例提供的模型公平性评估方法,在获得待测评样本属于真实数据分布中的对数似然度以及样本预测结果之后,即可根据该对数似然度以及样本预测结果,对待测评样本进行可信性检测,即对待测评样本是否为对抗样本、待测评样本的分布多样性进行检测,当确定待测评样本为对抗样本或者待测评样本的分布多样性较弱的情况下,则可以 确定该待测评样本为不可信,后续则可以对不可信的待测评样本进行数据处理,以提高非可信环境下,待测评样本的可靠性和完整性。The model fairness evaluation method provided by the embodiments of this specification, after obtaining the log likelihood and sample prediction results that the sample to be evaluated belongs to the real data distribution, can then evaluate the model fairness based on the log likelihood and sample prediction results. The sample is tested for credibility, that is, whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated is tested. When it is determined that the sample to be evaluated is an adversarial sample or the distribution diversity of the sample to be evaluated is weak, then After determining that the sample to be evaluated is untrustworthy, the untrustworthy sample to be evaluated can be subsequently processed to improve the reliability and integrity of the sample to be evaluated in an untrustworthy environment.
步骤206:在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本。Step 206: If the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample.
具体的,在确定待测评样本为对抗样本的情况下,则可以通过去噪、对抗重建的方式,对其进行样本重建;而在确定待测评样本的分布多样性较弱的情况下,则可以对该待测评样本进行多样性生成。具体实现方式如下所述:Specifically, when it is determined that the sample to be evaluated is an adversarial sample, the sample can be reconstructed through denoising and adversarial reconstruction; and when it is determined that the distribution diversity of the sample to be evaluated is weak, then it can be Diversity generation is performed on the sample to be evaluated. The specific implementation method is as follows:
所述在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本,包括:In the case where the credibility test result meets the non-credibility condition, sample processing is performed on the sample to be evaluated according to the credibility test result to obtain an updated evaluation sample, including:
在所述待测评样本为对抗样本的情况下,根据第一预设处理方式对所述待测评样本进行样本处理,获得第一更新测评样本;和/或When the sample to be evaluated is an adversarial sample, perform sample processing on the sample to be evaluated according to a first preset processing method to obtain a first updated evaluation sample; and/or
在所述待测评样本的分布多样性满足预设分布条件的情况下,根据第二预设处理方式对所述待测评样本进行样本处理,获得第二更新测评样本。When the distribution diversity of the samples to be evaluated satisfies the preset distribution conditions, sample processing is performed on the samples to be evaluated according to the second preset processing method to obtain a second updated evaluation sample.
其中,第一预设处理方式和第二预设处理方式可以根据实际应用进行设置,本说明书实施例对此不作任何限定。The first preset processing method and the second preset processing method can be set according to actual applications, and the embodiments of this specification do not limit this in any way.
实际应用中,对于待测评样本的可信性检测,可以检测出待测评样本是否为对抗样本,以及该待测评样本的分布多样性是否较弱;而在待测评样本为对抗样本,或者该待测评样本的分布多样性较弱的情况下,均可以认为该待测评样本为不可信;此时,则需要对该待测评样本进行处理。In practical applications, for the credibility detection of the sample to be evaluated, it can be detected whether the sample to be evaluated is an adversarial sample and whether the distribution diversity of the sample to be evaluated is weak; Even when the distribution diversity of the evaluation sample is weak, the sample to be evaluated can be considered to be untrustworthy; in this case, the sample to be evaluated needs to be processed.
具体实施时,在待测评样本的可信性检测结果为,该待测评样本为对抗样本的情况下,可以确定该待测评样本的可信性检测结果满足非可信性条件,此时,则可以根据第一预设处理方式对待测评样本进行样本处理,获得第一更新测评样本;而在待测评样本的可信性检测结果为,该待测评样本的分布多样性较弱的情况下,也可以确定该待测评样本的可信性检测结果满足非可信性条件,此时,则可以根据第二预设处理方式对待测评样本进行样本处理,获得第二更新测评样本。During specific implementation, when the credibility test result of the sample to be evaluated is, and the sample to be evaluated is an adversarial sample, it can be determined that the credibility test result of the sample to be evaluated satisfies the non-credibility condition. At this time, then The sample to be evaluated can be processed according to the first preset processing method to obtain the first updated evaluation sample; and when the credibility test result of the sample to be evaluated is that the distribution diversity of the sample to be evaluated is weak, It can be determined that the credibility test result of the sample to be evaluated satisfies the non-credibility condition. At this time, the sample to be evaluated can be processed according to the second preset processing method to obtain a second updated evaluation sample.
那么,在第一预设处理方式包括数据压缩、数据随机化或者对抗纠错的去噪方法的情况下,根据第一预设处理方式对待测评样本进行样本处理,获得第一更新测评样本的具体方式如下所述:Then, when the first preset processing method includes data compression, data randomization or a denoising method against error correction, the sample to be evaluated is processed according to the first preset processing method to obtain the specific details of the first updated evaluation sample. The method is as follows:
所述在所述待测评样本为对抗样本的情况下,根据第一预设处理方式对所述待测评样本进行样本处理,获得第一更新测评样本,包括:When the sample to be evaluated is an adversarial sample, performing sample processing on the sample to be evaluated according to a first preset processing method to obtain a first updated evaluation sample includes:
在所述待测评样本为对抗样本的情况下,通过数据压缩、数据随机化或对抗纠错的去噪方法对所述待测评样本进行重建,获得第一更新测评样本。When the sample to be evaluated is an adversarial sample, the sample to be evaluated is reconstructed through a denoising method such as data compression, data randomization or adversarial error correction to obtain a first updated evaluation sample.
而在第二预设处理方法为生成新增测评样本的情况下,根据第二预设处理方式对待测评样本进行样本处理,获得第二更新测评样本的具体处理方式如下所述:When the second preset processing method is to generate a new evaluation sample, the sample to be evaluated is processed according to the second preset processing method, and the specific processing method for obtaining the second updated evaluation sample is as follows:
所述在所述待测评样本的分布多样性满足预设分布条件的情况下,根据第二预设处理方式对所述待测评样本进行样本处理,获得第二更新测评样本,包括:When the distribution diversity of the samples to be evaluated meets the preset distribution conditions, sample processing is performed on the samples to be evaluated according to the second preset processing method to obtain the second updated evaluation sample, including:
在所述待测评样本的分布多样性满足预设分布条件的情况下,根据所述待测评样本的 样本数据概率分布,通过所述生成对抗网络模型的生成模块,生成新增测评样本;When the distribution diversity of the samples to be evaluated satisfies the preset distribution conditions, according to the distribution diversity of the samples to be evaluated, The sample data probability distribution is used to generate new evaluation samples through the generation module of the generative adversarial network model;
根据所述新增测评样本,获得第二更新测评样本。According to the newly added evaluation sample, a second updated evaluation sample is obtained.
具体的,在待测评样本的可信性由于对抗样本以及分布多样性较弱产生的情况下,则可以对待测评样本进行对抗重建以及多样性生成,其中,对抗重建可以理解为,对待测评样本中检测到的对抗样本,采用包括但不限于数据压缩、数据随机化、对抗纠错等去噪方法对其进行重建,以消除对抗噪声的干扰;对多样性生成,则可以理解为,基于待测评样本的样本数据概率分布,利用生成对抗网络模型的生成模块,生成新增测评样本,以扩充待测评样本的多样性。Specifically, when the credibility of the sample to be evaluated is due to weak adversarial samples and distribution diversity, adversarial reconstruction and diversity generation of the sample to be evaluated can be performed, where adversarial reconstruction can be understood as, in the sample to be evaluated, The detected adversarial samples are reconstructed using denoising methods including but not limited to data compression, data randomization, adversarial error correction, etc. to eliminate the interference of adversarial noise; diversity generation can be understood as based on the evaluation to be The sample data probability distribution of the sample uses the generation module of the generative adversarial network model to generate new evaluation samples to expand the diversity of samples to be evaluated.
本说明书实施例提供的模型公平性评估方法,在获得待测评样本属于真实数据分布中的对数似然度以及样本预测结果之后,即可根据该对数似然度以及样本预测结果,对待测评样本进行可信性检测,即对待测评样本是否为对抗样本、待测评样本的分布多样性进行检测,当确定待测评样本为对抗样本或者待测评样本的分布多样性较弱的情况下,则可以确定该待测评样本为不可信,后续则可以对不可信的待测评样本进行数据处理,以提高非可信环境下,待测评样本的可靠性和完整性,进而保证了该公平性评估方法在非可信环境下的健壮性和测评结果的可用性。The model fairness evaluation method provided by the embodiments of this specification, after obtaining the log likelihood and sample prediction results that the sample to be evaluated belongs to the real data distribution, can then evaluate the model fairness based on the log likelihood and sample prediction results. The sample is tested for credibility, that is, whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated is tested. When it is determined that the sample to be evaluated is an adversarial sample or the distribution diversity of the sample to be evaluated is weak, then After determining that the sample to be evaluated is untrustworthy, subsequent data processing can be performed on the untrustworthy sample to be evaluated to improve the reliability and integrity of the sample to be evaluated in an untrustworthy environment, thereby ensuring that the fairness assessment method is used in Robustness and availability of measurement results in untrusted environments.
此外,为了保证新增测评样本的准确性,在通过生成对抗网络模型的生成模块生成新增测评样本之后,还会根据生成对抗网络模型的判别模块对新增测评样本的质量进行检测,以保证新增测评样本的准确率。具体实现方式如下所述:In addition, in order to ensure the accuracy of the new evaluation samples, after the new evaluation samples are generated through the generation module of the generative adversarial network model, the quality of the new evaluation samples will also be tested based on the discriminant module of the generative adversarial network model to ensure Added accuracy rate of new evaluation samples. The specific implementation method is as follows:
所述根据所述新增测评样本,获得第二更新测评样本,包括:The second updated evaluation sample is obtained based on the newly added evaluation sample, including:
将所述新增测评样本输入所述生成对抗网络模型的判别模块,获得所述新增测评样本的预测结果;Input the newly added evaluation sample into the discrimination module of the generative adversarial network model to obtain the prediction results of the newly added evaluation sample;
根据所述新增测评样本的预测结果,对所述新增测评样本进行删减,获得第二更新测评样本。According to the prediction results of the new evaluation sample, the new evaluation sample is deleted to obtain a second updated evaluation sample.
本说明书实施例提供的模型公平性评估方法,为了保证新增测评样本的准确性,在通过生成对抗网络模型的生成模块生成新增测评样本之后,还会基于新增测评样本的生成质量对其进行样本过滤,以保证新增测评样本的准确率。此外,也可以根据原始待测评样本的质量评估指标对新增测评样本进行生成质量过滤,比如原始待测评样本为图像的情况下,可以根据图像的Inception Score(初始分数)对其进行质量过滤,当原始待测评样本为文本的情况下,可以根据文本的流畅度对其进行质量过滤等。The model fairness evaluation method provided by the embodiments of this specification, in order to ensure the accuracy of the newly added evaluation samples, after generating the new evaluation samples through the generation module of the generative adversarial network model, it will also be based on the generation quality of the new evaluation samples. Perform sample filtering to ensure the accuracy of newly added evaluation samples. In addition, the new evaluation samples can also be generated and filtered based on the quality evaluation indicators of the original samples to be evaluated. For example, if the original samples to be evaluated are images, they can be quality filtered based on the Inception Score (initial score) of the images. When the original sample to be evaluated is text, quality filtering can be performed on it based on the fluency of the text.
步骤208:根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估。Step 208: Conduct a fairness assessment on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample.
其中,更新测评样本包括第一更新测评样本和/或第二更新测评样本;而待测评模型也可以理解为与训练模型同类型的图文识别模型。Among them, the updated evaluation sample includes the first updated evaluation sample and/or the second updated evaluation sample; and the model to be evaluated can also be understood as an image and text recognition model of the same type as the training model.
在更新测评样本包括第一更新测评样本的情况下,则将待测评样本以及第一更新测评样本进行混合,生成混合样本后,对待测评模型进行公平性评估;在更新测评样本包括第二更新测评样本的情况下,则将待测评样本以及第二更新测评样本进行混合,生成混合样本后,对待测评模型进行公平性评估;在更新测评样本包括第一更新测评样本以及第二更 新测评样本的情况下,则将待测评样本、第一更新测评样本以及第二更新测评样本进行混合,生成混合样本后,对待测评模型进行公平性评估。具体实现方式如下所述:When the updated evaluation sample includes the first updated evaluation sample, the sample to be evaluated and the first updated evaluation sample are mixed. After the mixed sample is generated, the fairness of the model to be evaluated is evaluated; when the updated evaluation sample includes the second updated evaluation sample In the case of samples, the samples to be evaluated and the second updated evaluation samples are mixed. After the mixed samples are generated, the fairness of the model to be evaluated is evaluated; the updated evaluation samples include the first updated evaluation samples and the second updated evaluation samples. In the case of new evaluation samples, the samples to be evaluated, the first updated evaluation samples, and the second updated evaluation samples are mixed. After the mixed samples are generated, the fairness of the model to be evaluated is evaluated. The specific implementation method is as follows:
所述根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估,包括:The fairness evaluation of the model to be evaluated based on the sample to be evaluated and the updated evaluation sample includes:
将所述待测评样本和所述更新测评样本进行混合,获得混合测评样本;Mix the sample to be evaluated and the updated evaluation sample to obtain a mixed evaluation sample;
将所述混合测评样本以及待测评模型输入公平性评估模块,获得所述待测评模型的公平性评估指标;Input the mixed evaluation sample and the model to be evaluated into the fairness evaluation module to obtain the fairness evaluation index of the model to be evaluated;
根据所述待测评模型的公平性评估指标,对所述待测评模型进行公平性评估。Conduct a fairness evaluation on the model to be evaluated according to the fairness evaluation index of the model to be evaluated.
其中,公平性评估指标包括但不限于假阳率、统计均等、机会均等、不一致影响等指标。Among them, fairness evaluation indicators include but are not limited to false positive rate, statistical equality, equal opportunity, inconsistent impact and other indicators.
具体实施时,所述将所述混合测评样本以及待测评模型输入公平性评估模块,获得所述待测评模型的公平性评估指标,包括:During specific implementation, the mixed evaluation sample and the model to be evaluated are input into the fairness evaluation module to obtain the fairness evaluation index of the model to be evaluated, including:
将所述混合测评样本以及所述待测评模型输入公平性评估模块;Input the mixed evaluation sample and the model to be evaluated into the fairness evaluation module;
接收所述公平性评估模块输出的、根据所述混合测评样本的真实值以及预测值的比对结果,确定的所述待测评模型的公平性评估指标,receiving the fairness evaluation index of the model to be evaluated determined based on the comparison result of the real value and the predicted value of the mixed evaluation sample output by the fairness evaluation module,
其中,所述预测值为所述待测评模型根据所述混合测评样本输出。Wherein, the predicted value is the output of the model to be evaluated based on the mixed evaluation sample.
具体的,混合测评样本的真实值以及预测值的比对结果,可以理解为待测评模型的预测准确性。Specifically, the comparison result between the true value and the predicted value of the mixed evaluation sample can be understood as the prediction accuracy of the model to be evaluated.
实际应用中,将混合测评样本以及待测评模型输入公平性评估模块,在公平性评估模块中,将混合测评样本输入待测评模型,获得该待测评模型输出的该混合测评样本的预测值;根据该预测值以及该混合测评样本的真实值,计算该待测评模型的预测准确性;而该公平性评估模块在确定该待测评模型的预测准确性之后,根据该预测准确性计算出该待测评模型的假阳率、统计均等、机会均等、不一致影响等指标;后续用户或者系统可以根据这些指标评估该待测评模型的公平性。In practical applications, the mixed evaluation sample and the model to be evaluated are input into the fairness evaluation module. In the fairness evaluation module, the mixed evaluation sample is input into the model to be evaluated, and the predicted value of the mixed evaluation sample output by the model to be evaluated is obtained; according to The predicted value and the true value of the mixed evaluation sample are used to calculate the prediction accuracy of the model to be evaluated; and after determining the prediction accuracy of the model to be evaluated, the fairness assessment module calculates the prediction accuracy of the model to be evaluated. The model's false positive rate, statistical equality, equal opportunity, inconsistent impact and other indicators; subsequent users or the system can evaluate the fairness of the model to be evaluated based on these indicators.
本说明书实施例提供的所述模型公平性评估方法,通过图文训练模型,建模图文训练样本的真实数据概率分布,依据该真实数据概率分布对待测评样本进行可信性检测;并针对非可信样本进行处理,获得更新测评样本,以此提高在非可信环境下,测评样本的可靠性和完成性,从而保证该模型公平性评估方法,在可信环境下以及非可信环境下的健壮性和对模型测评结果的可用性,从而保证通过该模型公平性评估方法对待测评模型进行公平性评估的准确性,使得后续待测评模型在实际应用中无论是从满足监管合规还是从提升用户体验的角度,均有较好的效果。The model fairness evaluation method provided by the embodiment of this specification uses a graphic training model to model the real data probability distribution of graphic training samples, and conducts credibility detection on the samples to be evaluated based on the real data probability distribution; and for non- Trusted samples are processed to obtain updated evaluation samples, thereby improving the reliability and completeness of evaluation samples in non-trusted environments, thereby ensuring the fairness evaluation method of the model, both in trusted environments and non-trusted environments. Robustness and availability of model evaluation results, thereby ensuring the accuracy of the fairness evaluation of the model to be evaluated through this model fairness evaluation method, so that subsequent models to be evaluated can be used in practical applications, whether from meeting regulatory compliance or improving From a user experience perspective, both have good results.
下述结合附图3,以本说明书提供的模型公平性评估方法在对推荐模型进行公平性评估的应用为例,对所述模型公平性评估方法进行进一步说明。其中,图3示出了本说明书一个实施例提供的一种模型公平性评估方法的处理过程流程图,具体包括以下步骤。The following describes the model fairness evaluation method further in conjunction with Figure 3, taking the application of the model fairness evaluation method provided in this specification in fairness evaluation of the recommended model as an example. Among them, FIG. 3 shows a process flow chart of a model fairness evaluation method provided by an embodiment of this specification, which specifically includes the following steps.
步骤302:根据收集的大规模的无监督图文训练样本,利用自监督学习技术训练预训练模型,并通过该预训练模型初始建模图文训练样本的真实数据概率分布。Step 302: Based on the large-scale unsupervised graphic and text training samples collected, use self-supervised learning technology to train a pre-training model, and use the pre-training model to initially model the real data probability distribution of the graphic and text training samples.
步骤304:根据预训练模型利用深度生成技术构建生成对抗网络模型,根据图文训练 样本训练生成对抗网络模型,并根据生成对抗网络模型优化图文训练样本的真实数据概率分布,获得图文训练样本优化后的真实数据概率分布。Step 304: Use deep generation technology to build a generative adversarial network model based on the pre-trained model, and train based on graphics and text Sample training generates an adversarial network model, and optimizes the real data probability distribution of graphic and text training samples based on the generative adversarial network model to obtain the optimized real data probability distribution of graphic and text training samples.
步骤306:根据图文训练样本优化后的真实数据概率分布以及生成对抗网络模型,对待测评样本进行可信性检测。Step 306: Based on the optimized real data probability distribution of the graphic training samples and the generated adversarial network model, perform a credibility test on the samples to be evaluated.
其中,根据图文训练样本优化后的真实数据概率分布以及生成对抗网络模型,对待测评样本进行可信性检测,的具体实现方式可以参见上述实施例的详细介绍,在此不再赘述。Among them, based on the optimized real data probability distribution of the graphic training samples and the generated adversarial network model, the credibility of the samples to be evaluated is tested. The specific implementation method can be found in the detailed introduction of the above embodiments, and will not be described again here.
步骤308:根据待测评样本的可信性检测结果,利用对抗防御技术对待测评样本进行对抗重建、或者利用生成对抗网络模型的生成器对待测评样本进行多样性生成。Step 308: Based on the credibility detection results of the samples to be evaluated, use adversarial defense technology to perform adversarial reconstruction of the samples to be evaluated, or use a generator that generates an adversarial network model to generate diversity for the samples to be evaluated.
其中,待测评样本的对抗重建以及多样性生成的具体实现方式可以参见上述实施例的详细介绍,在此不再赘述。For specific implementation methods of adversarial reconstruction and diversity generation of samples to be evaluated, please refer to the detailed introduction of the above embodiments and will not be described again here.
步骤310:将待测评样本与、对抗重建的待测评样本和/或多样性生成的待测评样本进行混合,得到混合样本,并将混合样本以及推荐模型输入公平性评估模块进行测评,获得该推荐模型的公平性评估结果。Step 310: Mix the sample to be evaluated with the adversarially reconstructed sample to be evaluated and/or the diversity-generated sample to be evaluated to obtain a mixed sample, and input the mixed sample and the recommended model into the fairness evaluation module for evaluation to obtain the recommendation. Model fairness evaluation results.
本说明书实施例提供的模型公平性评估方法,提出了针对非受控(可信)环境下的算法公平性评估技术,基于大规模易获取的无监督图文训练数据,利用大规模预训练技术和深度生成技术相结合的方式建模真实数据概率分布,并基于建模得到的真实数据(无监督图文训练数据)概率分布检测待测评样本中的不可信样本同时通过对抗防御技术进行对抗去噪重建,可以有效地消除对抗噪声对公平性评估的影响。The model fairness evaluation method provided by the embodiments of this specification proposes algorithm fairness evaluation technology in an uncontrolled (trusted) environment, based on large-scale and easy-to-obtain unsupervised graphic and text training data, and utilizing large-scale pre-training technology Model the probability distribution of real data by combining it with deep generation technology, and detect untrustworthy samples in the samples to be evaluated based on the probability distribution of the real data (unsupervised graphic training data) obtained through modeling, and use adversarial defense technology to counteract them. Noisy reconstruction can effectively eliminate the impact of adversarial noise on fairness evaluation.
对于待测评数据分布偏差(样本分布单一,无法覆盖整个数据分布)的问题,本说明书实施例的模型公平性评估方法,提出了基于待测评样本本身的数据分布,同时结合深度生成技术进行多样性生成,可以达到提高在非可信环境下待测评样本的可靠性和完整性的目的。此外,在整个公平性评估流程中,本说明书实施例提供的模型公平性评估方法,不需要额外的测评数据,也无需人工介入,极大地提高了测评的智能化程度,同时也降低了测评成本。综上,本说明书实施例提供的模型公平性评估方法,不仅具备在可信环境下的公平性测评能力,同时还能保障在非受控环境下公平性测评的健壮性和测评结果的可用性,因此将适用于电商平台、在线社交平台、在线社交媒体等平台上包括但不限于智能客服、个性化推荐、智能风控等算法的公平性评估,达到消除算法偏见并促使算法满足监管合规和提升用户体验的目的。Regarding the problem of distribution deviation of the data to be evaluated (the sample distribution is single and cannot cover the entire data distribution), the model fairness evaluation method in the embodiment of this specification proposes a data distribution based on the sample to be evaluated itself, while combining depth generation technology to achieve diversity Generation can achieve the purpose of improving the reliability and integrity of samples to be evaluated in untrusted environments. In addition, in the entire fairness evaluation process, the model fairness evaluation method provided by the embodiments of this specification does not require additional evaluation data or manual intervention, which greatly improves the intelligence of the evaluation and also reduces the evaluation cost. . In summary, the model fairness evaluation method provided by the embodiments of this specification not only has the ability to evaluate fairness in a trusted environment, but also ensures the robustness of the fairness evaluation and the availability of the evaluation results in an uncontrolled environment. Therefore, it will be applicable to the fairness assessment of algorithms on e-commerce platforms, online social platforms, online social media and other platforms, including but not limited to intelligent customer service, personalized recommendations, intelligent risk control, etc., to eliminate algorithm bias and promote algorithms to meet regulatory compliance. and improve user experience.
与上述方法实施例相对应,本说明书还提供了模型公平性评估装置实施例,图4示出了本说明书一个实施例提供的一种模型公平性评估装置的结构示意图。如图4所示,该装置包括:Corresponding to the above method embodiments, this specification also provides an embodiment of a model fairness evaluation device. Figure 4 shows a schematic structural diagram of a model fairness evaluation device provided by an embodiment of this specification. As shown in Figure 4, the device includes:
概率分布确定模块402,被配置为根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;The probability distribution determination module 402 is configured to determine the real data probability distribution of the picture and/or text training sample according to the picture and/or text training model;
检测结果确定模块404,被配置为根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果;The detection result determination module 404 is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
样本处理模块406,被配置为在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本; The sample processing module 406 is configured to perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample when the credibility detection result satisfies the non-credibility condition;
评估模块408,被配置为根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估。The evaluation module 408 is configured to perform a fairness evaluation on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample.
可选地,所述概率分布确定模块402,进一步被配置为:Optionally, the probability distribution determination module 402 is further configured to:
获取图片和/或文本训练样本;Obtain image and/or text training samples;
根据所述训练样本利用自监督学习技术,训练获得图片和/或文本训练模型;Use self-supervised learning technology to train and obtain image and/or text training models based on the training samples;
根据所述图片和/或文本训练模型,获得所述训练样本调整后的真实数据概率分布;According to the picture and/or text training model, obtain the real data probability distribution adjusted by the training sample;
根据生成对抗网络模型对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布。The real data probability distribution of the training sample is adjusted according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample.
可选地,所述概率分布确定模块402,进一步被配置为:Optionally, the probability distribution determination module 402 is further configured to:
根据所述图片和/或文本训练模型,构建生成对抗网络模型;Build a generative adversarial network model based on the image and/or text training model;
根据所述训练样本对所述生成对抗网络模型进行训练,获得训练后的生成对抗网络模型的判别模块和生成模块;Train the generative adversarial network model according to the training samples, and obtain the discriminating module and generating module of the trained generative adversarial network model;
根据所述判别模块对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布。The real data probability distribution of the training sample is adjusted according to the discrimination module to obtain the adjusted real data probability distribution of the training sample.
可选地,所述概率分布确定模块402,进一步被配置为:Optionally, the probability distribution determination module 402 is further configured to:
根据所述图片和/或文本训练模型的模型参数,对生成对抗网络模型的判别模块的模块参数进行初始化,构建所述生成对抗网络模型的判别模块;Initialize the module parameters of the discrimination module of the generative adversarial network model according to the model parameters of the picture and/or text training model, and construct the discrimination module of the generative adversarial network model;
根据反卷积网络和/或文本生成网络,构建所述生成对抗网络模型的生成模块;Construct a generation module of the generative adversarial network model according to the deconvolution network and/or text generation network;
根据所述判别模块和所述生成模块构建所述生成对抗网络模型。The generative adversarial network model is constructed according to the discriminating module and the generating module.
可选地,所述检测结果确定模块404,进一步被配置为:Optionally, the detection result determination module 404 is further configured to:
根据所述图片和/或文本训练模型,获得所述待测评样本的样本数据概率分布;According to the picture and/or text training model, obtain the sample data probability distribution of the sample to be evaluated;
根据所述样本数据概率分布,确定所述待测评样本属于所述训练样本的真实数据概率分布的相似度;According to the sample data probability distribution, determine the similarity of the sample to be evaluated belonging to the real data probability distribution of the training sample;
根据所述生成对抗网络模型的判别模块,获得待测评样本的样本预测结果;According to the discriminant module of the generative adversarial network model, obtain the sample prediction result of the sample to be evaluated;
根据所述相似度以及所述样本预测结果,确定所述待测评样本的可信性检测结果。Based on the similarity and the sample prediction result, the credibility detection result of the sample to be evaluated is determined.
可选地,所述检测结果确定模块404,进一步被配置为:Optionally, the detection result determination module 404 is further configured to:
根据所述相似度以及所述样本预测结果,确定所述待测评样本是否为对抗样本以及所述待测评样本的分布多样性。According to the similarity and the sample prediction result, it is determined whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated.
可选地,所述样本处理模块406,进一步被配置为:Optionally, the sample processing module 406 is further configured to:
在所述待测评样本为对抗样本的情况下,根据第一预设处理方式对所述待测评样本进行样本处理,获得第一更新测评样本;和/或When the sample to be evaluated is an adversarial sample, perform sample processing on the sample to be evaluated according to a first preset processing method to obtain a first updated evaluation sample; and/or
在所述待测评样本的分布多样性满足预设分布条件的情况下,根据第二预设处理方式对所述待测评样本进行样本处理,获得第二更新测评样本。When the distribution diversity of the samples to be evaluated satisfies the preset distribution conditions, sample processing is performed on the samples to be evaluated according to the second preset processing method to obtain a second updated evaluation sample.
可选地,所述样本处理模块406,进一步被配置为:Optionally, the sample processing module 406 is further configured to:
在所述待测评样本为对抗样本的情况下,通过数据压缩、数据随机化或对抗纠错的去 噪方法对所述待测评样本进行重建,获得第一更新测评样本。When the sample to be evaluated is an adversarial sample, the method can be used to remove errors through data compression, data randomization or adversarial error correction. The noise method is used to reconstruct the sample to be evaluated to obtain the first updated evaluation sample.
可选地,所述样本处理模块406,进一步被配置为:Optionally, the sample processing module 406 is further configured to:
在所述待测评样本的分布多样性满足预设分布条件的情况下,根据所述待测评样本的样本数据概率分布,通过所述生成对抗网络模型的生成模块,生成新增测评样本;When the distribution diversity of the samples to be evaluated meets the preset distribution conditions, a new evaluation sample is generated through the generation module of the generative adversarial network model according to the sample data probability distribution of the samples to be evaluated;
根据所述新增测评样本,获得第二更新测评样本。According to the newly added evaluation sample, a second updated evaluation sample is obtained.
可选地,所述样本处理模块406,进一步被配置为:Optionally, the sample processing module 406 is further configured to:
将所述新增测评样本输入所述生成对抗网络模型的判别模块,获得所述新增测评样本的预测结果;Input the newly added evaluation sample into the discrimination module of the generative adversarial network model to obtain the prediction results of the newly added evaluation sample;
根据所述新增测评样本的预测结果,对所述新增测评样本进行删减,获得第二更新测评样本。According to the prediction results of the new evaluation sample, the new evaluation sample is deleted to obtain a second updated evaluation sample.
可选地,所述评估模块408,进一步被配置为:Optionally, the evaluation module 408 is further configured to:
将所述待测评样本和所述更新测评样本进行混合,获得混合测评样本;Mix the sample to be evaluated and the updated evaluation sample to obtain a mixed evaluation sample;
将所述混合测评样本以及待测评模型输入公平性评估模块,获得所述待测评模型的公平性评估指标;Input the mixed evaluation sample and the model to be evaluated into the fairness evaluation module to obtain the fairness evaluation index of the model to be evaluated;
根据所述待测评模型的公平性评估指标,对所述待测评模型进行公平性评估。Conduct a fairness evaluation on the model to be evaluated according to the fairness evaluation index of the model to be evaluated.
可选地,所述评估模块408,进一步被配置为:Optionally, the evaluation module 408 is further configured to:
将所述混合测评样本以及所述待测评模型输入公平性评估模块;Input the mixed evaluation sample and the model to be evaluated into the fairness evaluation module;
接收所述公平性评估模块输出的、根据所述混合测评样本的真实值以及预测值的比对结果,确定的所述待测评模型的公平性评估指标,receiving the fairness evaluation index of the model to be evaluated determined based on the comparison result of the real value and the predicted value of the mixed evaluation sample output by the fairness evaluation module,
其中,所述预测值为所述待测评模型根据所述混合测评样本输出。Wherein, the predicted value is the output of the model to be evaluated based on the mixed evaluation sample.
本说明书实施例提供的所述模型公平性评估装置,通过图文训练模型,建模图文训练样本的真实数据概率分布,依据该真实数据概率分布对待测评样本进行可信性检测;并针对非可信样本进行处理,获得更新测评样本,以此提高在非可信环境下,测评样本的可靠性和完成性,从而保证该模型公平性评估方法,在可信环境下以及非可信环境下的健壮性和对模型测评结果的可用性,从而保证通过该模型公平性评估方法对待测评模型进行公平性评估的准确性,使得后续待测评模型在实际应用中无论是从满足监管合规还是从提升用户体验的角度,均有较好的效果。The model fairness evaluation device provided by the embodiment of this specification uses a graphic training model to model the real data probability distribution of graphic training samples, and conducts credibility detection on the samples to be evaluated based on the real data probability distribution; and for non- Trusted samples are processed to obtain updated evaluation samples, thereby improving the reliability and completeness of evaluation samples in non-trusted environments, thereby ensuring the fairness evaluation method of the model, both in trusted environments and non-trusted environments. Robustness and availability of model evaluation results, thereby ensuring the accuracy of the fairness evaluation of the model to be evaluated through this model fairness evaluation method, so that subsequent models to be evaluated can be used in practical applications, whether from meeting regulatory compliance or improving From a user experience perspective, both have good results.
上述为本实施例的一种模型公平性评估装置的示意性方案。需要说明的是,该模型公平性评估装置的技术方案与上述的模型公平性评估方法的技术方案属于同一构思,模型公平性评估装置的技术方案未详细描述的细节内容,均可以参见上述模型公平性评估方法的技术方案的描述。The above is a schematic solution of a model fairness evaluation device in this embodiment. It should be noted that the technical solution of the model fairness evaluation device and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. For details that are not described in detail in the technical solution of the model fairness evaluation device, please refer to the above-mentioned model fairness. Description of technical solutions for sexual assessment methods.
参见图5,图5示出了本说明书一个实施例提供的另一种模型公平性评估方法的流程图,具体包括以下步骤。Referring to Figure 5, Figure 5 shows a flow chart of another model fairness evaluation method provided by an embodiment of this specification, which specifically includes the following steps.
步骤502:根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布。Step 502: Determine the real data probability distribution of the image and/or text training samples based on the image and/or text training model.
步骤504:接收用户发送的待测评样本以及待测评模型。 Step 504: Receive the sample to be evaluated and the model to be evaluated sent by the user.
步骤506:根据所述真实数据概率分布、以及生成对抗网络模型,确定所述待测评样本的可信性检测结果。Step 506: Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model.
步骤508:在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本。Step 508: If the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample.
步骤510:根据所述待测评样本以及所述更新测评样本,对所述待测评模型进行公平性评估。Step 510: Conduct a fairness assessment on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample.
步骤512:获得所述待测评模型的公平性评估结果,并将所述公平性评估结果返回至所述用户。Step 512: Obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
本说明书实施例提供的模型公平性评估方法,应用于模型公平性评估平台。The model fairness evaluation method provided by the embodiments of this specification is applied to the model fairness evaluation platform.
实际应用场景可以为,用户想要在模型公平性评估平台对其项目模型进行公平性评估,那么用户则可以将其待测评样本以及待测评模型发送至模型公平性评估平台,模型公平性评估平台在接收用户发送的待测评样本以及待测评模型之后,即可根据上述实施例的方式对其待测评样本进行对抗重建或者是多样性生成,从而保证模型公平性评估平台给出的测评结果更接近待测评模型的真实情况。The actual application scenario can be that if a user wants to conduct a fairness evaluation on his or her project model on the model fairness evaluation platform, the user can send the sample to be evaluated and the model to be evaluated to the model fairness evaluation platform. The model fairness evaluation platform After receiving the samples to be evaluated and the models to be evaluated sent by the user, adversarial reconstruction or diversity generation of the samples to be evaluated can be performed according to the above embodiment, thereby ensuring that the evaluation results given by the model fairness evaluation platform are closer to each other. The actual situation of the model to be evaluated.
本说明书实施例提供的模型公平性评估方法,首先基于大规模易获取的无监督图文训练数据,利用大规模预训练技术和深度生成技术相结合的方式建模训练数据概率分布以作为真实数据概率分布。对于待测评样本,在进入公平性评估模块之前,将依据真实数据概率分布对待测评样本进行可靠性检测,有效地缓解了不可信样本对测评结果的干扰。对于检测到的非可信样本如对抗样本或分布偏差样本,本方案分别基于对抗防御技术对测评样本进行进行去噪、重建和基于深度生成模型对多样性生成,达到了提高在非可信环境下测评样本的可靠性和完整性的目的,保障了公平性测评系统在非受控环境下的健壮性和测评结果的可用性,有效弥补了现有系统对非受控环境下的对抗扰动和测评样本分布偏差敏感的缺点。The model fairness evaluation method provided by the embodiments of this specification is first based on large-scale and easy-to-obtain unsupervised graphic training data, and uses a combination of large-scale pre-training technology and deep generation technology to model the training data probability distribution as real data Probability distributions. For the samples to be evaluated, before entering the fairness assessment module, the reliability of the samples to be evaluated will be tested based on the probability distribution of real data, which effectively alleviates the interference of untrustworthy samples on the evaluation results. For detected untrusted samples such as adversarial samples or distribution deviation samples, this solution denoises and reconstructs the evaluation samples based on adversarial defense technology and generates diversity based on deep generation models, achieving improved performance in untrusted environments. The purpose of ensuring the reliability and integrity of the evaluation samples is to ensure the robustness of the fairness evaluation system in uncontrolled environments and the availability of evaluation results, and effectively make up for the existing system's resistance to disturbance and evaluation in uncontrolled environments. The disadvantage of being sensitive to sample distribution deviation.
与上述方法实施例相对应,本说明书还提供了另一种模型公平性评估装置实施例,图6示出了本说明书一个实施例提供的另一种模型公平性评估装置的结构示意图。如图6所示,该装置应用于模型公平性评估平台,包括:Corresponding to the above method embodiment, this specification also provides another embodiment of a model fairness evaluation device. Figure 6 shows a schematic structural diagram of another model fairness evaluation device provided by an embodiment of this specification. As shown in Figure 6, the device is applied to the model fairness evaluation platform, including:
第一确定模块602,被配置为根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;The first determination module 602 is configured to determine the real data probability distribution of the image and/or text training samples according to the image and/or text training model;
数据接收模块604,被配置为接收用户发送的待测评样本以及待测评模型;The data receiving module 604 is configured to receive samples to be evaluated and models to be evaluated sent by the user;
第二确定模块606,被配置为根据所述真实数据概率分布、以及生成对抗网络模型,确定所述待测评样本的可信性检测结果;The second determination module 606 is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
样本更新模块608,被配置为在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;The sample update module 608 is configured to perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample when the credibility detection result satisfies the non-credibility condition;
公平性评估模块610,被配置为根据所述待测评样本以及所述更新测评样本,对所述待测评模型进行公平性评估;The fairness evaluation module 610 is configured to perform a fairness evaluation on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample;
结果展示模块612,被配置为获得所述待测评模型的公平性评估结果,并将所述公平性评估结果返回至所述用户。 The result display module 612 is configured to obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
本说明书实施例提供的模型公平性评估装置,首先基于大规模易获取的无监督图文训练数据,利用大规模预训练技术和深度生成技术相结合的方式建模训练数据概率分布以作为真实数据概率分布。对于待测评样本,在进入公平性评估模块之前,将依据真实数据概率分布对待测评样本进行可靠性检测,有效地缓解了不可信样本对测评结果的干扰。对于检测到的非可信样本如对抗样本或分布偏差样本,本方案分别基于对抗防御技术对测评样本进行进行去噪、重建和基于深度生成模型对多样性生成,达到了提高在非可信环境下测评样本的可靠性和完整性的目的,保障了公平性测评系统在非受控环境下的健壮性和测评结果的可用性,有效弥补了现有系统对非受控环境下的对抗扰动和测评样本分布偏差敏感的缺点。The model fairness evaluation device provided by the embodiments of this specification is first based on large-scale and easy-to-obtain unsupervised graphic training data, and uses a combination of large-scale pre-training technology and deep generation technology to model the training data probability distribution as real data Probability distributions. For the samples to be evaluated, before entering the fairness assessment module, the reliability of the samples to be evaluated will be tested based on the probability distribution of real data, which effectively alleviates the interference of untrustworthy samples on the evaluation results. For detected untrusted samples such as adversarial samples or distribution deviation samples, this solution denoises and reconstructs the evaluation samples based on adversarial defense technology and generates diversity based on deep generation models, achieving improved performance in untrusted environments. The purpose of ensuring the reliability and integrity of the evaluation samples is to ensure the robustness of the fairness evaluation system in uncontrolled environments and the availability of evaluation results, and effectively make up for the existing system's resistance to disturbance and evaluation in uncontrolled environments. The disadvantage of being sensitive to sample distribution deviation.
上述为本实施例的一种模型公平性评估装置的示意性方案。需要说明的是,该模型公平性评估装置的技术方案与上述的模型公平性评估方法的技术方案属于同一构思,模型公平性评估装置的技术方案未详细描述的细节内容,均可以参见上述模型公平性评估方法的技术方案的描述。The above is a schematic solution of a model fairness evaluation device in this embodiment. It should be noted that the technical solution of the model fairness evaluation device and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. For details that are not described in detail in the technical solution of the model fairness evaluation device, please refer to the above-mentioned model fairness. Description of technical solutions for sexual assessment methods.
图7示出了根据本说明书一个实施例提供的一种计算设备700的结构框图。该计算设备700的部件包括但不限于存储器710和处理器720。处理器720与存储器710通过总线730相连接,数据库750用于保存数据。Figure 7 shows a structural block diagram of a computing device 700 provided according to an embodiment of this specification. Components of the computing device 700 include, but are not limited to, memory 710 and processor 720 . The processor 720 and the memory 710 are connected through a bus 730, and the database 750 is used to save data.
计算设备700还包括接入设备740,接入设备740使得计算设备700能够经由一个或多个网络760通信。这些网络的示例包括公用交换电话网(PSTN)、局域网(LAN)、广域网(WAN)、个域网(PAN)或诸如因特网的通信网络的组合。接入设备740可以包括有线或无线的任何类型的网络接口(例如,网络接口卡(NIC))中的一个或多个,诸如IEEE802.11无线局域网(WLAN)无线接口、全球微波互联接入(Wi-MAX)接口、以太网接口、通用串行总线(USB)接口、蜂窝网络接口、蓝牙接口、近场通信(NFC)接口,等等。Computing device 700 also includes an access device 740 that enables computing device 700 to communicate via one or more networks 760 . Examples of these networks include the Public Switched Telephone Network (PSTN), a local area network (LAN), a wide area network (WAN), a personal area network (PAN), or a combination of communications networks such as the Internet. Access device 740 may include one or more of any type of network interface (eg, a network interface card (NIC)), wired or wireless, such as an IEEE 802.11 Wireless Local Area Network (WLAN) wireless interface, Global Interconnection for Microwave Access ( Wi-MAX) interface, Ethernet interface, Universal Serial Bus (USB) interface, cellular network interface, Bluetooth interface, Near Field Communication (NFC) interface, etc.
在本说明书的一个实施例中,计算设备700的上述部件以及图7中未示出的其他部件也可以彼此相连接,例如通过总线。应当理解,图7所示的计算设备结构框图仅仅是出于示例的目的,而不是对本说明书范围的限制。本领域技术人员可以根据需要,增添或替换其他部件。In one embodiment of this specification, the above-mentioned components of the computing device 700 and other components not shown in FIG. 7 may also be connected to each other, such as through a bus. It should be understood that the structural block diagram of the computing device shown in FIG. 7 is for illustrative purposes only and does not limit the scope of this description. Those skilled in the art can add or replace other components as needed.
计算设备700可以是任何类型的静止或移动计算设备,包括移动计算机或移动计算设备(例如,平板计算机、个人数字助理、膝上型计算机、笔记本计算机、上网本等)、移动电话(例如,智能手机)、可佩戴的计算设备(例如,智能手表、智能眼镜等)或其他类型的移动设备,或者诸如台式计算机或PC的静止计算设备。计算设备700还可以是移动式或静止式的服务器。Computing device 700 may be any type of stationary or mobile computing device, including a mobile computer or mobile computing device (e.g., tablet computer, personal digital assistant, laptop computer, notebook computer, netbook, etc.), a mobile telephone (e.g., smartphone ), a wearable computing device (e.g., smart watch, smart glasses, etc.) or other type of mobile device, or a stationary computing device such as a desktop computer or PC. Computing device 700 may also be a mobile or stationary server.
其中,处理器720用于执行如下计算机可执行指令,该计算机可执行指令被处理器执行时实现上述模型公平性评估方法的步骤。The processor 720 is configured to execute the following computer-executable instructions. When the computer-executable instructions are executed by the processor, the steps of the above model fairness evaluation method are implemented.
上述为本实施例的一种计算设备的示意性方案。需要说明的是,该计算设备的技术方案与上述的模型公平性评估方法的技术方案属于同一构思,计算设备的技术方案未详细描述的细节内容,均可以参见上述模型公平性评估方法的技术方案的描述。The above is a schematic solution of a computing device in this embodiment. It should be noted that the technical solution of the computing device and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. Details that are not described in detail in the technical solution of the computing device can be found in the technical solution of the above-mentioned model fairness evaluation method. description of.
本说明书一实施例还提供一种计算机可读存储介质,其存储有计算机可执行指令,该计算机可执行指令被处理器执行时实现上述模型公平性评估方法的步骤。 An embodiment of this specification also provides a computer-readable storage medium that stores computer-executable instructions. When the computer-executable instructions are executed by a processor, the steps of the above model fairness evaluation method are implemented.
上述为本实施例的一种计算机可读存储介质的示意性方案。需要说明的是,该存储介质的技术方案与上述的模型公平性评估方法的技术方案属于同一构思,存储介质的技术方案未详细描述的细节内容,均可以参见上述模型公平性评估方法的技术方案的描述。The above is a schematic solution of a computer-readable storage medium in this embodiment. It should be noted that the technical solution of the storage medium and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. For details that are not described in detail in the technical solution of the storage medium, please refer to the technical solution of the above-mentioned model fairness evaluation method. description of.
本说明书一实施例还提供一种计算机程序,其中,当所述计算机程序在计算机中执行时,令计算机执行上述模型公平性评估方法的步骤。An embodiment of this specification also provides a computer program, wherein when the computer program is executed in a computer, the computer is caused to perform the steps of the above model fairness evaluation method.
上述为本实施例的一种计算机程序的示意性方案。需要说明的是,该计算机程序的技术方案与上述的模型公平性评估方法的技术方案属于同一构思,计算机程序的技术方案未详细描述的细节内容,均可以参见上述模型公平性评估方法的技术方案的描述。The above is a schematic solution of a computer program in this embodiment. It should be noted that the technical solution of this computer program and the technical solution of the above-mentioned model fairness evaluation method belong to the same concept. For details that are not described in detail in the technical solution of the computer program, please refer to the technical solution of the above-mentioned model fairness evaluation method. description of.
上述对本说明书特定实施例进行了描述。其它实施例在所附权利要求书的范围内。在一些情况下,在权利要求书中记载的动作或步骤可以按照不同于实施例中的顺序来执行并且仍然可以实现期望的结果。另外,在附图中描绘的过程不一定要求示出的特定顺序或者连续顺序才能实现期望的结果。在某些实施方式中,多任务处理和并行处理也是可以的或者可能是有利的。The foregoing describes specific embodiments of this specification. Other embodiments are within the scope of the appended claims. In some cases, the actions or steps recited in the claims can be performed in a different order than in the embodiments and still achieve desired results. Additionally, the processes depicted in the figures do not necessarily require the specific order shown, or sequential order, to achieve desirable results. Multitasking and parallel processing are also possible or may be advantageous in certain implementations.
所述计算机指令包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、电载波信号、电信信号以及软件分发介质等。需要说明的是,所述计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括电载波信号和电信信号。The computer instructions include computer program code, which may be in the form of source code, object code, executable file or some intermediate form. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording media, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signals, telecommunications signals, and software distribution media, etc. It should be noted that the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of legislation and patent practice in the jurisdiction. For example, in some jurisdictions, according to legislation and patent practice, the computer-readable medium Excludes electrical carrier signals and telecommunications signals.
需要说明的是,对于前述的各方法实施例,为了简便描述,故将其都表述为一系列的动作组合,但是本领域技术人员应该知悉,本说明书实施例并不受所描述的动作顺序的限制,因为依据本说明书实施例,某些步骤可以采用其它顺序或者同时进行。其次,本领域技术人员也应该知悉,说明书中所描述的实施例均属于优选实施例,所涉及的动作和模块并不一定都是本说明书实施例所必须的。It should be noted that for the convenience of description, each of the foregoing method embodiments is expressed as a series of action combinations. However, those skilled in the art should know that the embodiments of this specification are not limited by the described action sequence. limitation, because according to the embodiments of this specification, certain steps may be performed in other orders or at the same time. Secondly, those skilled in the art should also know that the embodiments described in the specification are preferred embodiments, and the actions and modules involved are not necessarily necessary for the embodiments of this specification.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述的部分,可以参见其它实施例的相关描述。In the above embodiments, each embodiment is described with its own emphasis. For parts that are not described in detail in a certain embodiment, please refer to the relevant descriptions of other embodiments.
以上公开的本说明书优选实施例只是用于帮助阐述本说明书。可选实施例并没有详尽叙述所有的细节,也不限制该发明仅为所述的具体实施方式。显然,根据本说明书实施例的内容,可作很多的修改和变化。本说明书选取并具体描述这些实施例,是为了更好地解释本说明书实施例的原理和实际应用,从而使所属技术领域技术人员能很好地理解和利用本说明书。本说明书仅受权利要求书及其全部范围和等效物的限制。 The preferred embodiments of this specification disclosed above are only used to help explain this specification. Alternative embodiments are not described in all details, nor are the inventions limited to the specific embodiments described. Obviously, many modifications and changes can be made based on the contents of the embodiments of this specification. These embodiments are selected and described in detail in this specification to better explain the principles and practical applications of the embodiments in this specification, so that those skilled in the art can better understand and utilize this specification. This specification is limited only by the claims and their full scope and equivalents.

Claims (14)

  1. 一种模型公平性评估方法,包括:A model fairness assessment method, including:
    根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;Based on the image and/or text training model, determine the real data probability distribution of the image and/or text training samples;
    根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果;Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
    在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;When the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample;
    根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估。According to the sample to be evaluated and the updated evaluation sample, a fairness evaluation is performed on the model to be evaluated.
  2. 根据权利要求1所述的模型公平性评估方法,所述根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布,包括:The model fairness evaluation method according to claim 1, wherein the training model based on pictures and/or text and determining the real data probability distribution of picture and/or text training samples includes:
    获取图片和/或文本训练样本;Obtain image and/or text training samples;
    根据所述训练样本利用自监督学习技术,训练获得图片和/或文本训练模型;Use self-supervised learning technology to train and obtain image and/or text training models based on the training samples;
    根据所述图片和/或文本训练模型,获得所述训练样本调整后的真实数据概率分布;According to the picture and/or text training model, obtain the real data probability distribution adjusted by the training sample;
    根据生成对抗网络模型对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布。The real data probability distribution of the training sample is adjusted according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample.
  3. 根据权利要求2所述的模型公平性评估方法,所述根据生成对抗网络模型对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布,包括:The model fairness evaluation method according to claim 2, adjusting the real data probability distribution of the training sample according to the generative adversarial network model to obtain the adjusted real data probability distribution of the training sample, including:
    根据所述图片和/或文本训练模型,构建生成对抗网络模型;Build a generative adversarial network model based on the image and/or text training model;
    根据所述训练样本对所述生成对抗网络模型进行训练,获得训练后的生成对抗网络模型的判别模块和生成模块;Train the generative adversarial network model according to the training samples, and obtain the discriminating module and generating module of the trained generative adversarial network model;
    根据所述判别模块对所述训练样本的真实数据概率分布进行调整,获得所述训练样本调整后的真实数据概率分布。The real data probability distribution of the training sample is adjusted according to the discrimination module to obtain the adjusted real data probability distribution of the training sample.
  4. 根据权利要求3所述的模型公平性评估方法,所述根据所述图片和/或文本训练模型,构建生成对抗网络模型,包括:The model fairness evaluation method according to claim 3, wherein the training model based on the pictures and/or text and constructing a generative adversarial network model includes:
    根据所述图片和/或文本训练模型的模型参数,对生成对抗网络模型的判别模块的模块参数进行初始化,构建所述生成对抗网络模型的判别模块;Initialize the module parameters of the discrimination module of the generative adversarial network model according to the model parameters of the picture and/or text training model, and construct the discrimination module of the generative adversarial network model;
    根据反卷积网络和/或文本生成网络,构建所述生成对抗网络模型的生成模块;Construct a generation module of the generative adversarial network model according to the deconvolution network and/or text generation network;
    根据所述判别模块和所述生成模块构建所述生成对抗网络模型。The generative adversarial network model is constructed according to the discriminating module and the generating module.
  5. 根据权利要求1所述的模型公平性评估方法,所述根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果,包括:The model fairness evaluation method according to claim 1, wherein determining the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model includes:
    根据所述图片和/或文本训练模型,获得所述待测评样本的样本数据概率分布; According to the picture and/or text training model, obtain the sample data probability distribution of the sample to be evaluated;
    根据所述样本数据概率分布,确定所述待测评样本属于所述训练样本的真实数据概率分布的相似度;According to the sample data probability distribution, determine the similarity of the sample to be evaluated belonging to the real data probability distribution of the training sample;
    根据所述生成对抗网络模型的判别模块,获得待测评样本的样本预测结果;According to the discriminant module of the generative adversarial network model, obtain the sample prediction result of the sample to be evaluated;
    根据所述相似度以及所述样本预测结果,确定所述待测评样本的可信性检测结果。Based on the similarity and the sample prediction result, the credibility detection result of the sample to be evaluated is determined.
  6. 根据权利要求5所述的模型公平性评估方法,所述根据所述相似度以及所述样本预测结果,确定所述待测评样本的可信性检测结果,包括:The model fairness evaluation method according to claim 5, wherein determining the credibility detection result of the sample to be evaluated based on the similarity and the sample prediction result includes:
    根据所述相似度以及所述样本预测结果,确定所述待测评样本是否为对抗样本以及所述待测评样本的分布多样性。According to the similarity and the sample prediction result, it is determined whether the sample to be evaluated is an adversarial sample and the distribution diversity of the sample to be evaluated.
  7. 根据权利要求6所述的模型公平性评估方法,所述在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本,包括:The model fairness evaluation method according to claim 6, wherein when the credibility detection result satisfies the non-credibility condition, sample processing is performed on the sample to be evaluated according to the credibility detection result, Get updated test samples, including:
    在所述待测评样本为对抗样本的情况下,根据第一预设处理方式对所述待测评样本进行样本处理,获得第一更新测评样本;和/或When the sample to be evaluated is an adversarial sample, perform sample processing on the sample to be evaluated according to a first preset processing method to obtain a first updated evaluation sample; and/or
    在所述待测评样本的分布多样性满足预设分布条件的情况下,根据第二预设处理方式对所述待测评样本进行样本处理,获得第二更新测评样本。When the distribution diversity of the samples to be evaluated satisfies the preset distribution conditions, sample processing is performed on the samples to be evaluated according to the second preset processing method to obtain a second updated evaluation sample.
  8. 根据权利要求7所述的模型公平性评估方法,所述在所述待测评样本为对抗样本的情况下,根据第一预设处理方式对所述待测评样本进行样本处理,获得第一更新测评样本,包括:The model fairness evaluation method according to claim 7, wherein when the sample to be evaluated is an adversarial sample, sample processing is performed on the sample to be evaluated according to a first preset processing method to obtain a first updated evaluation. Samples include:
    在所述待测评样本为对抗样本的情况下,通过数据压缩、数据随机化或对抗纠错的去噪方法对所述待测评样本进行重建,获得第一更新测评样本。When the sample to be evaluated is an adversarial sample, the sample to be evaluated is reconstructed through a denoising method such as data compression, data randomization or adversarial error correction to obtain a first updated evaluation sample.
  9. 根据权利要求7所述的模型公平性评估方法,所述在所述待测评样本的分布多样性满足预设分布条件的情况下,根据第二预设处理方式对所述待测评样本进行样本处理,获得第二更新测评样本,包括:The model fairness evaluation method according to claim 7, wherein when the distribution diversity of the samples to be evaluated satisfies preset distribution conditions, sample processing is performed on the samples to be evaluated according to a second preset processing method. , get the second updated evaluation sample, including:
    在所述待测评样本的分布多样性满足预设分布条件的情况下,根据所述待测评样本的样本数据概率分布,通过所述生成对抗网络模型的生成模块,生成新增测评样本;When the distribution diversity of the samples to be evaluated meets the preset distribution conditions, a new evaluation sample is generated through the generation module of the generative adversarial network model according to the sample data probability distribution of the samples to be evaluated;
    根据所述新增测评样本,获得第二更新测评样本。According to the newly added evaluation sample, a second updated evaluation sample is obtained.
  10. 根据权利要求9所述的模型公平性评估方法,所述根据所述新增测评样本,获得第二更新测评样本,包括:The model fairness evaluation method according to claim 9, wherein obtaining a second updated evaluation sample based on the newly added evaluation sample includes:
    将所述新增测评样本输入所述生成对抗网络模型的判别模块,获得所述新增测评样本的预测结果;Input the newly added evaluation sample into the discrimination module of the generative adversarial network model to obtain the prediction results of the newly added evaluation sample;
    根据所述新增测评样本的预测结果,对所述新增测评样本进行删减,获得第二更新测评样本。According to the prediction results of the new evaluation sample, the new evaluation sample is deleted to obtain a second updated evaluation sample.
  11. 根据权利要求1所述的模型公平性评估方法,所述根据所述待测评样本以及所述 更新测评样本,对待测评模型进行公平性评估,包括:The model fairness evaluation method according to claim 1, wherein the sample to be evaluated and the Update the evaluation samples and conduct a fairness assessment on the model to be evaluated, including:
    将所述待测评样本和所述更新测评样本进行混合,获得混合测评样本;Mix the sample to be evaluated and the updated evaluation sample to obtain a mixed evaluation sample;
    将所述混合测评样本以及待测评模型输入公平性评估模块,获得所述待测评模型的公平性评估指标;Input the mixed evaluation sample and the model to be evaluated into the fairness evaluation module to obtain the fairness evaluation index of the model to be evaluated;
    根据所述待测评模型的公平性评估指标,对所述待测评模型进行公平性评估。Conduct a fairness evaluation on the model to be evaluated according to the fairness evaluation index of the model to be evaluated.
  12. 根据权利要求11所述的模型公平性评估方法,所述将所述混合测评样本以及待测评模型输入公平性评估模块,获得所述待测评模型的公平性评估指标,包括:The model fairness evaluation method according to claim 11, said inputting the mixed evaluation sample and the model to be evaluated into a fairness evaluation module to obtain the fairness evaluation index of the model to be evaluated, including:
    将所述混合测评样本以及所述待测评模型输入公平性评估模块;Input the mixed evaluation sample and the model to be evaluated into the fairness evaluation module;
    接收所述公平性评估模块输出的、根据所述混合测评样本的真实值以及预测值的比对结果,确定的所述待测评模型的公平性评估指标,receiving the fairness evaluation index of the model to be evaluated determined based on the comparison result of the real value and the predicted value of the mixed evaluation sample output by the fairness evaluation module,
    其中,所述预测值为所述待测评模型根据所述混合测评样本输出。Wherein, the predicted value is the output of the model to be evaluated based on the mixed evaluation sample.
  13. 一种模型公平性评估装置,包括:A model fairness assessment device, including:
    概率分布确定模块,被配置为根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;a probability distribution determination module configured to determine the real data probability distribution of the image and/or text training sample based on the image and/or text training model;
    检测结果确定模块,被配置为根据所述真实数据概率分布、以及生成对抗网络模型,确定待测评样本的可信性检测结果;The detection result determination module is configured to determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
    样本处理模块,被配置为在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;A sample processing module configured to perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample when the credibility detection result satisfies the non-credibility condition;
    评估模块,被配置为根据所述待测评样本以及所述更新测评样本,对待测评模型进行公平性评估。The evaluation module is configured to perform a fairness evaluation on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample.
  14. 一种模型公平性评估方法,应用于模型公平性评估平台,包括:A model fairness assessment method, applied to the model fairness assessment platform, including:
    根据图片和/或文本训练模型,确定图片和/或文本训练样本的真实数据概率分布;Based on the image and/or text training model, determine the real data probability distribution of the image and/or text training samples;
    接收用户发送的待测评样本以及待测评模型;Receive samples to be evaluated and models to be evaluated sent by users;
    根据所述真实数据概率分布、以及生成对抗网络模型,确定所述待测评样本的可信性检测结果;Determine the credibility detection result of the sample to be evaluated based on the real data probability distribution and the generated adversarial network model;
    在所述可信性检测结果满足非可信条件的情况下,根据所述可信性检测结果对所述待测评样本进行样本处理,获得更新测评样本;When the credibility detection result satisfies the non-credibility condition, perform sample processing on the sample to be evaluated according to the credibility detection result to obtain an updated evaluation sample;
    根据所述待测评样本以及所述更新测评样本,对所述待测评模型进行公平性评估;Conduct a fairness assessment on the model to be evaluated based on the sample to be evaluated and the updated evaluation sample;
    获得所述待测评模型的公平性评估结果,并将所述公平性评估结果返回至所述用户。 Obtain the fairness evaluation result of the model to be evaluated, and return the fairness evaluation result to the user.
PCT/CN2023/086570 2022-04-12 2023-04-06 Model fairness evaluation methods and apparatus WO2023197927A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210379396.XA CN114970670A (en) 2022-04-12 2022-04-12 Model fairness assessment method and device
CN202210379396.X 2022-04-12

Publications (1)

Publication Number Publication Date
WO2023197927A1 true WO2023197927A1 (en) 2023-10-19

Family

ID=82977853

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/086570 WO2023197927A1 (en) 2022-04-12 2023-04-06 Model fairness evaluation methods and apparatus

Country Status (2)

Country Link
CN (1) CN114970670A (en)
WO (1) WO2023197927A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114970670A (en) * 2022-04-12 2022-08-30 阿里巴巴(中国)有限公司 Model fairness assessment method and device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200143231A1 (en) * 2018-11-02 2020-05-07 Microsoft Technology Licensing, Llc Probabilistic neural network architecture generation
CN111753918A (en) * 2020-06-30 2020-10-09 浙江工业大学 Image recognition model for eliminating sex bias based on counterstudy and application
CN112700408A (en) * 2020-12-28 2021-04-23 中国银联股份有限公司 Model training method, image quality evaluation method and device
CN113220553A (en) * 2021-05-13 2021-08-06 支付宝(杭州)信息技术有限公司 Method and device for evaluating performance of text prediction model
CN114139601A (en) * 2021-11-01 2022-03-04 国家电网有限公司大数据中心 Evaluation method and system for artificial intelligence algorithm model of power inspection scene
CN114970670A (en) * 2022-04-12 2022-08-30 阿里巴巴(中国)有限公司 Model fairness assessment method and device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200143231A1 (en) * 2018-11-02 2020-05-07 Microsoft Technology Licensing, Llc Probabilistic neural network architecture generation
CN111753918A (en) * 2020-06-30 2020-10-09 浙江工业大学 Image recognition model for eliminating sex bias based on counterstudy and application
CN112700408A (en) * 2020-12-28 2021-04-23 中国银联股份有限公司 Model training method, image quality evaluation method and device
CN113220553A (en) * 2021-05-13 2021-08-06 支付宝(杭州)信息技术有限公司 Method and device for evaluating performance of text prediction model
CN114139601A (en) * 2021-11-01 2022-03-04 国家电网有限公司大数据中心 Evaluation method and system for artificial intelligence algorithm model of power inspection scene
CN114970670A (en) * 2022-04-12 2022-08-30 阿里巴巴(中国)有限公司 Model fairness assessment method and device

Also Published As

Publication number Publication date
CN114970670A (en) 2022-08-30

Similar Documents

Publication Publication Date Title
US10678997B2 (en) Machine learned models for contextual editing of social networking profiles
WO2022041979A1 (en) Information recommendation model training method and related device
US20220414464A1 (en) Method and server for federated machine learning
Wang et al. Methods for correcting inference based on outcomes predicted by machine learning
CN109447156B (en) Method and apparatus for generating a model
Häggström Data‐driven confounder selection via Markov and Bayesian networks
WO2015130928A1 (en) Real estate evaluating platform methods, apparatuses, and media
Pauly et al. Permutation‐based inference for the AUC: A unified approach for continuous and discontinuous data
WO2023197927A1 (en) Model fairness evaluation methods and apparatus
CN110929799A (en) Method, electronic device, and computer-readable medium for detecting abnormal user
CN112231570A (en) Recommendation system trust attack detection method, device, equipment and storage medium
KR20230078785A (en) Analysis of augmented reality content item usage data
US11934926B2 (en) Sensitivity in supervised machine learning with experience data
EP3561735A1 (en) Integrating deep learning into generalized additive mixed-effect (game) frameworks
US10572835B2 (en) Machine-learning algorithm for talent peer determinations
CN113330462A (en) Neural network training using soft nearest neighbor loss
CN113485993A (en) Data identification method and device
CN110955840B (en) Joint optimization of notifications and pushes
WO2020093817A1 (en) Identity verification method and device
WO2022105117A1 (en) Method and device for image quality assessment, computer device, and storage medium
US20220004937A1 (en) Determining application path for execution by bot
CN114238968A (en) Application program detection method and device, storage medium and electronic equipment
Zhang et al. Usable region estimate for assessing practical usability of medical image segmentation models
US20230185953A1 (en) Selecting differential privacy parameters in neural networks
CN113010784B (en) Method, apparatus, electronic device and medium for generating prediction information

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23787572

Country of ref document: EP

Kind code of ref document: A1