US20230325726A1 - Techniques for deriving and/or leveraging application-centric model metric - Google Patents

Techniques for deriving and/or leveraging application-centric model metric Download PDF

Info

Publication number
US20230325726A1
US20230325726A1 US18/203,967 US202318203967A US2023325726A1 US 20230325726 A1 US20230325726 A1 US 20230325726A1 US 202318203967 A US202318203967 A US 202318203967A US 2023325726 A1 US2023325726 A1 US 2023325726A1
Authority
US
United States
Prior art keywords
model
prototype
exemplars
features
performance
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/203,967
Inventor
Arnold BOEDIHARDJO
Adam Estrada
Andrew Jenkins
Nathan CLEMENT
Alan Schoen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Radiant Mission Solutions Inc
Maxar Mission Solutions Inc
Original Assignee
Maxar Mission Solutions Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Maxar Mission Solutions Inc filed Critical Maxar Mission Solutions Inc
Priority to US18/203,967 priority Critical patent/US20230325726A1/en
Assigned to RADIANT ANALYTIC SOLUTIONS INC. reassignment RADIANT ANALYTIC SOLUTIONS INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BOEDIHARDJO, Arnold, CLEMENT, Nathan, ESTRADA, ADAM, JENKINS, ANDREW, SCHOEN, Alan
Assigned to MAXAR MISSION SOLUTIONS INC. reassignment MAXAR MISSION SOLUTIONS INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: RADIANT MISSION SOLUTIONS INC.
Assigned to RADIANT MISSION SOLUTIONS INC. reassignment RADIANT MISSION SOLUTIONS INC. MERGER (SEE DOCUMENT FOR DETAILS). Assignors: RADIANT ANALYTIC SOLUTIONS INC., RADIANT GEOSPATIAL SOLUTIONS INC., THE HUMAN GEO GROUP LLC
Assigned to SIXTH STREET LENDING PARTNERS, AS ADMINISTRATIVE AGENT reassignment SIXTH STREET LENDING PARTNERS, AS ADMINISTRATIVE AGENT INTELLECTUAL PROPERTY SECURITY AGREEMENT Assignors: Maxar Intelligence Inc., MAXAR MISSION SOLUTIONS INC., MAXAR SPACE LLC, MAXAR TECHNOLOGIES HOLDINGS INC.
Publication of US20230325726A1 publication Critical patent/US20230325726A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/76Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries based on eigen-space representations, e.g. from pose or different illumination conditions; Shape manifolds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/7715Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Definitions

  • Certain example embodiments described herein relate to techniques usable in connection with models supporting machine learning (ML) and/or artificial intelligence (AI) applications. More particularly, certain example embodiments described herein relate to techniques for deriving and/or leveraging an application-centric model metric in connection with a model that supports ML and/or AI functionality.
  • ML machine learning
  • AI artificial intelligence
  • AI Artificial intelligence
  • Machine learning may be thought of as being a subset of AI, as ML in general may be thought of as the scientific study of algorithms and statistical models that computer systems use to help perform specific tasks effectively without using explicit instructions, instead relying on, for example, patterns and inferences.
  • AI problems oftentimes are addressed using search and optimization tools, logic programming and automated reasoning, probabilistic approaches to uncertain reasoning (using, for example, Bayesian networks, Hidden Markov Models, etc.), classifiers and statistical learning methods, neural networks, etc.
  • Models oftentimes underlie these and/or other AI-related techniques. For instance, ML algorithms quite often build a mathematical model based on training data. Once trained, a model can be used to make predictions or decisions without being explicitly programmed to perform the task.
  • AI and ML techniques are being actively researched and are already implemented in many facets of everyday life, affecting millions of people around the world. Indeed, AI and ML technologies already have practical applications in computer visioning, ecommerce and other recommendation engines, financial market analysis, fraud detection, game playing, image recognition, medical diagnosis, natural language processing, search engines, speech recognition, weather prediction, and many other areas.
  • a presumed “neutral” model may in essence be trained to have a “bias,” making it function better or worse for a given problem in a given space based on, for example, assumptions made during training, the training methodology or data set used, and/or the like.
  • Certain example embodiments address these and/or other concerns.
  • one aspect of certain example embodiments relates to quantifying the accuracy of a given model for a general input space.
  • one aspect of certain example embodiments relates to techniques that provide the ability to make quantifiable, reliable, and reproducible statements about the quality of a given model.
  • Another aspect of certain example embodiments relates to the ability to quantify the accuracy of models so that they can be rated and/or compared to one another using a common and understandable metric.
  • Another aspect of certain example embodiments relates to the ability to perform the above with respect to specific input spaces or areas of interest. For example, certain example embodiments are able to suggest a particular model or training data set of interest for a specific input space or area of interest.
  • a method of quantifying accuracy of a prediction model that has been trained on a data set parameterized by a plurality of features is provided.
  • the model performs in accordance with a theoretical performance manifold over an intractable input space in connection with the features.
  • a determination is made as to which of the plurality of features are strongly correlated with performance of the model.
  • a plurality of parameterized sub-models are created such that, in aggregate, they approximate the intractable input space.
  • Prototype exemplars are generated for each of the created sub-models, with the prototype exemplars for each created sub-model being objects to which the model can be applied to result in a match with the respective sub-model.
  • the accuracy of the model is quantified using the generated prototype exemplars.
  • a method of recommending a prediction model from among a plurality of different prediction models is provided.
  • Each one of the different prediction models has been trained based on a respective training data set, and each one of the different prediction models performs in accordance with a respective theoretical performance manifold.
  • An indication of a region definable in relation to the theoretical performance manifolds of the different prediction models is received as input.
  • the indication of the region is linked to features parameterizing the respective performance manifold.
  • One or more portions of the respective performance manifold is/are identified based on the features determined by the linking, with the one or more portions having a volume and a shape that collectively denote an expected performance of the respective model for the input.
  • the expected performance of the different prediction models are compared for the input. Based on the comparison, one or more of the different prediction models is/are recommended.
  • FIG. 1 is an example performance manifold defined on the set of all valid image collections to which a hypothetical model M can be applied;
  • FIG. 2 shows a subspace of the performance manifold from FIG. 1 , with the training set S and the test set S′;
  • FIG. 3 shows how the performance manifold shown in FIG. 1 can be tessellated so that the aggregate of the tessellation approximates the performance manifold from FIG. 1 ;
  • FIG. 4 shows the performance manifold of FIG. 1 , with images of interest confined to a relatively tight area remote from the training set S and the test set S′;
  • FIG. 5 is a block diagram showing, at a notional level, example components for quantifying accuracy of a prediction model, in accordance with certain example embodiments;
  • FIG. 6 is a flowchart showing an approach for quantifying accuracy of a prediction model, in accordance with the approach shown in and described in connection with FIG. 5 ;
  • FIG. 7 is a block diagram showing example computer system components that may be used to implement the approach shown in and described in connection with FIGS. 5 - 6 ;
  • FIG. 8 is a flowchart showing an approach for recommending a prediction model, in accordance with certain example embodiments.
  • Certain example embodiments described herein relate to techniques for deriving and/or leveraging an application-centric model metric in connection with a model that supports machine learning (ML) and/or artificial intelligence (AI) functionality. That is, certain example embodiments develop and implement an index rating system for quantifying the accuracy of detection and/or other models.
  • Such models can be used in a variety of applications including, for example, identifying objects in satellite imageries, facial recognition, etc.
  • a model in this sense is a learning algorithm that has been trained on a data set parameterized by a plurality of features. In the satellite imagery context, for example, the data set may be parameterized by spatial extent, National Imagery Interpretability Rating Scale (NIIRS), off-nadir angle, cloud coverage, and other attributes.
  • NIRS National Imagery Interpretability Rating Scale
  • features may include spatial extent, NIIRS, off-nadir angle, cloud coverage, and other attributes.
  • the rating index of certain example embodiments helps measure the expected performance of that model on new and/or unseen imageries, i.e., imageries on which the model has not been trained. As such, the rating index of certain example embodiments considers the large and potentially intractable class of all possible unseen imagery inputs.
  • certain example embodiments generate a compact and accurate approximate representation of the input space, produce prototypical exemplars from the approximate representation, and efficiently evaluate the accuracy of the model for the set of exemplars to estimate a given model’s expected performance.
  • the expected performance in this sense is a statistical measure of the model’s accuracy in the “general case.”
  • a recommender system based implementation may be provided for use in connection with this index rating. That is, although the expected performance metric may represent the model’s accuracy across the entire spectrum of available inputs (the “general case” alluded to above), some users may be concerned with only limited input spaces (a subset of the general case).
  • a limited input space might involve, for example, locating objects in a desert where coloration and topography might be relatively uniform, in a cityscape where off-nadir images may predominate, etc.
  • coloration and topography might be relatively uniform within desert and snow images respectively, but coloration and topography may different between the two.
  • a detection model M can be thought of as being trained on object class type T and an image collection S.
  • a set of features F [ ⁇ 1 .. ⁇ k ] parameterizes the training regime of model M.
  • feature f may represent cloud coverage, NIIRS, off-nadir angle, etc., as noted in the above-mentioned satellite imagery example.
  • the performance of model M is defined by its model architecture and F, i.e., M(F).
  • G 1..n (M,P) is the set of performance metrics for model M and an arbitrary image collection P.
  • the determination of the rating metric in certain example embodiments involves estimating the expected performance of model M for detecting object T from a novel (unseen) image collection S′ or E[G i (M,S′)], and the variation in the performance metrics of the model on the novel (unseen) image collection S′ or VAR[G i (M,S′)].
  • This technique may be used on a number of different model types. This includes, for example, convolutional neural networks (CNNs), na ⁇ ve Bayesian networks (NBNs), decision trees, etc.
  • Typical performance metrics may include mean average precision, receiver operating characteristic (ROC) related curves (such as precision-recall ROC or P-ROC), an F-beta score (e.g., F1), single scatter, latent feature measures from item-response theory (IRT), and/or the like.
  • ROC receiver operating characteristic
  • F1 F-beta score
  • single scatter single scatter
  • latent feature measures from item-response theory (IRT) and/or the like.
  • Constraints are introduced, including on the novel (unseen) image collection S′, in order to enable computations related to E[G i (M,S′)] and VAR[G i (M,S′)].
  • training inputs are modeled so as to be constrained to argmin S′ (DISCREPANCY (S′,U)) where U may represent training data S, or the set of all valid image collections to which the model M can be applied, ⁇ .
  • argmin S′ refers to the ML model assumption that the actual data is similar to a given collection of data (e.g., training data S), with some small discrepancy, and DISCREPANCY can be measured in different ways to in essence query how well S′ matches the input data set.
  • DISCREPANCY can be symmetric KL-divergence on probability distributions P(S), P(S′), P( ⁇ ), where a is the set of all valid image collections to which the model M can be applied. It will be appreciated that inverse SIMILARITY can be used in place of DISCREPANCY in certain example embodiments. These and/or other constraints may be introduced in different example instances, e.g., based on domain knowledge, specific applications or use cases, etc.
  • FIG. 1 is an example performance manifold defined on the set of all valid image collections to which the model M can be applied.
  • the model M is parameterized by features.
  • features f1 and f2 are shown, and additional features are implied by the color gradient.
  • the surface area of subspace projections (coverage) of the performance manifold is normalized against the volume (G) of S′ in the performance manifold.
  • FIG. 2 shows a subspace of the performance manifold from FIG. 1 , with the training set S and the test set S′. The subspace thus implies an additional constraint on the model training inputs.
  • Valid inputs are modeled so as to be constrained to argmin S′ (DISCREPANCY (S′, ⁇ )).
  • DISCREPANCY S′, ⁇
  • inverse SIMILARITY can be used in place of DISCREPANCY, etc.
  • certain example embodiments generate a compact and accurate approximate representation of the input space, produce prototypical exemplars from the approximate representation, and efficiently evaluate the accuracy of the model for the set of exemplars to estimate the model’s expected performance.
  • certain example embodiments create a tractable model of ⁇ , generate S′ from proxy of ⁇ , and estimate E[G i (M,S′)] and VAR[G i (M,S′)].
  • the creation of the tractable model of ⁇ involves developing a proxy for ⁇ to which computation of G i () can be applied. This is possible because the proxy for ⁇ is tractable (e.g., as enabled by pruning).
  • the generation of S′ from the proxy of ⁇ involves creating and/or deriving prototype exemplars of S′ from the proxy of ⁇ . And the estimation of E[G i (M,S′)] and VAR[G i (M,S′)] involves computing G i () based on the prototypes of S′ and estimating the probabilities of the exemplars. These operations will be discussed, in turn, below.
  • F When it comes to creating a tractable model of ⁇ , the data image features F that are strongly correlated to the model M’s performance are determined.
  • F can be defined a priori to include features known or suspected to be of interest.
  • cloud cover, NIIRS, and/or other factors may be specified by a user a priori.
  • domain experience may be taken into account in identifying features.
  • F can be defined using embedding projections such as feature extractors. Feature extractors based on a residual network algorithm (e.g., ResNet50) may be used in this regard.
  • ResNet50 residual network algorithm
  • a piecewise aggregate model is created.
  • This piecewise aggregate model in essence partitions ⁇ into “simplified” parameterized subsets or sub-models.
  • FIG. 3 shows how the performance manifold shown in FIG. 1 can be tessellated so that the aggregate of the tessellation approximates the performance manifold from FIG. 1 .
  • Different approaches may be used to create the sub-models in different example embodiments.
  • Example approaches may include a priori creation of sub-models, data-driven sub-model creation, hybrid approaches that use a priori creation of sub-models and data-driven sub-model creation, etc. Further details regarding some example approaches are provided below.
  • a data-driven sub-model creation approach may be used in certain example embodiments.
  • a data-driven sub-model creation approach may involve deriving sub-models through clustering or the like. For instance, a similar metric on F and G() can be generated and used as input to a clustering algorithm. This may be performed on any images that may be represented by the manifold including, for example, training and test samples, in certain example embodiments.
  • a priori creation of sub-model may be facilitated by developing or leveraging an existing taxonomy of image classes in certain example embodiments.
  • a taxonomy may specify images with cloud cover of 0-10%, 11-20%, etc.
  • Such taxonomies may be “flat” or multi-level hierarchies in different example embodiments.
  • Data analytics can be used to develop clusters of images that should be grouped together in certain example embodiments, a priori, similar to the approach set forth in the preceding paragraph.
  • each sub-model or partition may be desirable to have each sub-model or partition converge to some uniform distribution of values.
  • the value that each sub-model or partition converges to may be one or more performance metrics G(), for example. This may help to determine areas where the model performs uniformly. It will be appreciated that this strict assumption may be loosened in certain example embodiments (e.g., in response to a determination that there are too many partitions, the partitions created are too small, clusters are too close together, etc.).
  • certain example embodiments may involve determining what features matter to a given part, and/or the whole, of the model (a form of sensitivity testing with respect to, or perturbation of, individual features) in generating synthetic images.
  • the exemplar is not limited to a particular instance of an image or collection of images (real or otherwise) but instead can be objects parameterized explicitly on F. This may allow for estimates of G() based on F, explicitly.
  • the proxy exemplars are to be estimated.
  • a regression is created for G on the set of exemplars q.
  • FIG. 4 shows the performance manifold of FIG. 1 , with images of interest confined to a relatively tight area remote from the training set S and the test set S′.
  • a recommender system may be employed.
  • the performance metric is computed based on the user’s area of interest, U c S′, where U is small (and in some cases potentially extremely small) subset of S′. This will allow differentiated scoring of models based on user’s unique needs in certain example embodiments.
  • a user may specify features of interest directly, the system may infer features of interest (e.g., by analyzing example images and deriving those features, etc.), and/or the like.
  • the recommender system will output scores based on the desert region. However, if that user is interested in those models as they apply to a snow region, then the system will (re)compute scores based on snow terrain. Relative performance of X and Y may be different for desert and snow, e.g., X may outperform Y in the desert but not in snow. Thus, the recommender system may help the user determine which models are best for a given interest or set of interests.
  • FIG. 5 is a block diagram showing, at a notional level, example components for quantifying accuracy of a prediction model, in accordance with certain example embodiments
  • FIG. 6 is a flowchart showing an approach for quantifying accuracy of a prediction model, in accordance with the approach shown in and described in connection with FIG. 5 .
  • the model 500 is parameterized by features 502 .
  • the model 500 performs in accordance with a theoretical performance manifold over an intractable input space (e.g., all valid data sets to which the model can be applied) in connection with those features 502 .
  • the model 500 may, for example, be trained to identify objects in images (e.g., with the images being satellite images or the like), and the features 502 may include, for example, spatial extent, National Imagery Interpretability Rating Scale (NIIRS), off-nadir angle, signal-to-noise ratio (SNR), cloud coverage, and/or other values.
  • NIRS National Imagery Interpretability Rating Scale
  • SNR signal-to-noise ratio
  • cloud coverage and/or other values.
  • the objects themselves may be images and/or image collections, with the objects potentially being parameterized explicitly on the features.
  • the intractable input space may be limited to a continuous region defined by the data set on which the prediction model is trained, a plurality of discontinuous regions, and/or the like.
  • the feature deriver module 504 determines which of the features 502 are strongly correlated with performance of the model 500 . This may be aided based on user input 506 and/or embedding projections 508 (step 602 in FIG. 6 ).
  • at least one of the features 510 determined to be strongly correlated with part of the model 500 may include a non-linear mapping based on a feature from the data set on which the prediction model is trained. For instance, the non-linear mapping may be kernel principal component analysis.
  • the feature deriver module 504 outputs those features 510 , and based on the features 510 determined to be strongly correlated with performance of the model 500 , the sub-model creation engine 512 creates parameterized sub-models that, in an aggregate tessellation 514 , approximate the intractable input space (step 604 in FIG. 6 ).
  • the exemplar generator 516 generates prototype exemplars for each of the created sub-models in the aggregate tessellation 514 , with the prototype exemplars for each created sub-model being objects to which the model 500 can be applied to result in a match with the respective sub-model (step 606 in FIG. 6 ).
  • the accuracy of the model 500 is quantified using the generated prototype exemplars in one form or another. The quantified accuracy may, for example, reflect the expected performance of the model identifying an object of a given type from new and/or unseen images.
  • Those output metrics may be output for display (e.g., as being a rating on a common rating scale representing accuracy of the model 500 ), provided for comparison purposes (e.g., with other models), used to power suggestions (e.g., as to whether this model is good or bad for a particular task), etc. (see step 608 in FIG. 6 ).
  • the data set on which the prediction model is trained may be for a first application
  • the accuracy of the model may be quantified for a data set for a second application
  • the first and second applications may be different from one another.
  • the data set on which the prediction model is trained may be for a first geospatial and/or geotemporal image type
  • the accuracy of the model may be quantified for a data set for a second geospatial and/or geotemporal image type
  • the first and second geospatial and/or geotemporal image types may be different from one another.
  • modules, engines, generators, etc., shown in and described in connection with FIG. 5 may be implemented in connection with processing resources including at least one processor and a memory.
  • program logic corresponding to these elements may be executed by the processing resources.
  • the output may be displayed on a display device, transmitted to a remote computing device, etc.
  • FIG. 7 is a block diagram showing example computer system components that may be used to implement the approach shown in and described in connection with FIGS. 5 - 6 .
  • FIG. 7 shows a computing system 702 including processing resources such as, for example, at least one processor 704 operably coupled to a memory 706 .
  • the computing system further includes network interface(s) 710 and a display interface(s) 712 .
  • Each of these interfaces 710 , 712 may include hardware input/output devices (e.g., a network card in the case of the network interface(s) 710 ), software Application Programming Interfaces (APIs) callable to make use of the hardware elements, and/or the like.
  • the network interface(s) 710 may, for example, provide incoming and/or outgoing connections from/to a computer network such as, for example the Internet.
  • a representation of the model 500 (with the model features 502 ) at least initially may be external to the computing system 702 .
  • it may be stored to an external storage device and uploaded or otherwise made accessible to the computing system 702 .
  • An electronic interface making use of the network interface(s) 710 (such as, for example, a network upload, API and/or web service calls, etc.) may enable the computing system 702 to access or otherwise interact with the model 500 and the model features 502 .
  • the memory 706 in the computing system 702 includes volatile and/or nonvolatile storage. As shown in the FIG. 7 example, the memory 706 stores an operating system 708 , along with software modules suitable for implementing the functionality described above in connection with FIG. 5 . For instance, in certain example embodiments, software modules are stored in the memory and enable the processor(s) 704 to perform functionality needed for the execution of the feature deriver 504 , sub-model creation engine 512 , exemplar generator 516 , etc.
  • Working storage may be used to buffer or otherwise store user input 506 and/or projections 508 usable by the feature deriver 504 ; output from the modules executed under control of the processor(s) 704 such as, for example, the features 510 determined to be strongly correlated with performance of the model 500 , the aggregate tessellation 514 , prototype exemplars for each created sub-model, output metrics 518 , etc.; and/or other data useful for operation of the overall system.
  • the output generator module 716 provides the display, comparison, suggestions, etc. 520 , described above.
  • the network interface(s) 710 may make use of the network interface(s) 710 to generate results to be displayed on a remote display (e.g., if the computing system is a server and is accessed by a remote client computing device), the display interface(s) 712 to output results to a local display device, local storage where results may be cached, etc.
  • the processing resources (including the processor(s) 704 and the memory 706 ) of the computing system 702 may be configured to control the computing system 702 to perform functionality such as that described above in connection with FIGS. 5 - 6 .
  • the computing system 702 itself may including one or more computers (e.g., a server or series of servers) or it may be a standalone machine.
  • the model 500 may be collocated with the computing system 702 initially, or the model 500 may be provided or otherwise made accessible to the computing system 702 . Results may be output locally and/or provided to a remote user accessing a program implemented over the Internet, for example.
  • FIG. 8 is a flowchart showing an approach for recommending a prediction model, in accordance with certain example embodiments. It will be appreciated that the same or similar architecture as that shown in and described above in connection with FIG. 7 may be used to execute the functionality of that described in connection with FIG. 8 .
  • the model is recommended from among a plurality of different prediction models, with each one of the different prediction models having been trained based on a respective training data set and performing in accordance with a respective theoretical performance manifold. In some cases, each of the different prediction models may be trained based on a different training data set.
  • the models may be received over an electronic interface (including, for example, a hardware network interface, using API calls, web service calls, etc.) if they are remote from the computing system, etc., as described above.
  • an indication of a region definable in relation to the theoretical performance manifolds of the different prediction models is received as input.
  • the indication of the region may be defined as a set of one or more attributes that parameterize at least one of the different models, using a plurality of images, etc.
  • User input may be provided locally or from a remote source (e.g., via a web application, over the internet, using a standalone application operating under control of the processing resources, etc.).
  • the indication of the region is linked to features parameterizing the respective performance manifold, and one or more portions of the respective performance manifold is identified based on the features determined by the linking.
  • the one or more portions have a volume and a shape that collectively denote an expected performance of the respective model for the input.
  • a representation of the respective performance manifold may be generated, and the linking and the identifying may be practiced in connection with the representations of the performance manifolds of the respective prediction models.
  • the generating of the representations of the performance manifolds may comprise, for each of the different prediction models: determining which of the plurality of features parameterizing the respective model are strongly correlated with performance of the respective model (see example techniques disclosed above); based on the features determined to be strongly correlated with performance of the respective model, creating a plurality of parameterized sub-models that, in aggregate, approximate the performance manifold; and defining the representation of the respective model as the aggregation of the sub-models.
  • generating of the representations may be performed prior to reception of the input.
  • Certain example embodiments may further include, for each of the different prediction models, generating prototype exemplars for each of the created sub-models, with the prototype exemplars for each created sub-model being objects to which the respective model can be applied to result in a match with the respective sub-model, and with the prototype exemplars characterizing the volume and/or shape for an estimated portion of the performance manifold.
  • the techniques described above may be used in these regards. For instance, the software modules shown in and described in connection with FIG. 7 may be used in certain example embodiments.
  • the expected performance of the different prediction models are compared for the input, in step 806 . Based on the comparison, one or more of the different prediction models is recommended in step 808 .
  • the expected performance of the models may in some instances reflect accuracy of identifying an object of a given type from new and/or unseen images.
  • the objects may be images and/or image collections, with those objects potentially being parameterized explicitly on the features.
  • the training data sets and/or the input may include geospatial and/or geotemporal data such as, for example, satellite imagery.
  • the features parameterizing the performance manifolds include spatial extent, National Imagery Interpretability Rating Scale (NIIRS), off-nadir angle, signal-to-noise ratio (SNR), and/or cloud coverage values.
  • NIRS National Imagery Interpretability Rating Scale
  • SNR signal-to-noise ratio
  • certain example embodiments can be used to quantify the accuracy of a given model for a general or specific input space, e.g., such that statements can be made about the quality of a given model. It thus becomes possible to quantify the accuracy of models so that they can be rated and/or compared to one another using a common and understandable metric.
  • the same or similar techniques can also be used to suggest a particular model or training data set of interest for a specific input space or area of interest. Points of failure in a model can also be identified using the technology disclosed herein, and the technology disclosed herein also can be used to help combat adversarial ML.
  • certain example embodiments are technically advantageous in that they provide technology-based solutions to technology-based problems.
  • certain example embodiments provide technical solutions to technical problems associated with the general inability to objectively and reproducibly compare model performance across different models, a single model with an intractable input space, and/or the like.
  • certain example embodiments provide technical solutions to technical problems associated with the difficulties in locating where innate “biases” in ML models and/or AI systems might lie, e.g., because of poorly trained models, intentionally manipulated training data sets, and/or the like.
  • certain example embodiments provide technical improvements to current ML and AI related modeling tools.
  • certain example embodiments improve current ML and AI related tools by enabling objective quantification of model accuracy for models being or already produced, yielding expected performance data for a particular data set even when the input space potentially is unknown, facilitating model-to-model comparisons, etc.
  • certain example embodiments improve current ML and AI related tools by enabling specific already-produced models to be suggested, and/or by suggesting how to produce models expected to perform well (e.g., by suggesting what models to use, what training data should be used, etc.).
  • certain example embodiments provide technical improvements to tools that incorporate models because it becomes possible to know which models to include when performing analyses or making suggestions, which data sets to train on, etc. This may lead to better performance (e.g., increased accuracy, reduced processing time, etc.) of the tools that incorporate such models because they can be expected to perform better based on their quantified performance metrics.
  • system, subsystem, service, engine, module, programmed logic circuitry, and the like may be implemented as any suitable combination of software, hardware, firmware, and/or the like.
  • storage locations, stores, and repositories discussed herein may be any suitable combination of disk drive devices, memory locations, solid state drives, CD-ROMs, DVDs, tape backups, storage area network (SAN) systems, and/or any other appropriate tangible non-transitory computer readable storage medium.
  • Cloud and/or distributed storage e.g., using file sharing means, for instance, also may be used in certain example embodiments.
  • the techniques described herein may be accomplished by having at least one processor execute instructions that may be tangibly stored on a non-transitory computer readable storage medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Medical Informatics (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computational Linguistics (AREA)
  • Image Analysis (AREA)

Abstract

Techniques for quantifying accuracy of a prediction model that has been trained on a data set parameterized by multiple features are provided. The model performs in accordance with a theoretical performance manifold over an intractable input space in connection with the features. A determination is made as to which of the features are strongly correlated with performance of the model. Based on the features determined to be strongly correlated with performance of the model, parameterized sub-models are created such that, in aggregate, they approximate the intractable input space. Prototype exemplars are generated for each of the created sub-models, with the prototype exemplars for each created sub-model being objects to which the model can be applied to result in a match with the respective sub-model. The accuracy of the model is quantified using the generated prototype exemplars. A recommendation engine is provided for when there are particular areas of interest.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application is a continuation of U.S. Application No. 16/887,670 filed May 29, 2020, which claims priority to U.S. Application No. 62/855,138 filed on May 31, 2019, the entire contents of which are hereby incorporated herein by reference.
  • TECHNICAL FIELD
  • Certain example embodiments described herein relate to techniques usable in connection with models supporting machine learning (ML) and/or artificial intelligence (AI) applications. More particularly, certain example embodiments described herein relate to techniques for deriving and/or leveraging an application-centric model metric in connection with a model that supports ML and/or AI functionality.
  • BACKGROUND AND SUMMARY
  • Artificial intelligence (AI) in general refers to intelligence demonstrated by machines rather than humans and in this sense describes machines that mimic “cognitive” functions (such as, for example such as “learning” and “problem solving”) that humans typically associate with other human minds. Areas of AI research typically include reasoning, knowledge representation, planning, learning, natural language processing, perception, and the ability to move and manipulate objects, etc. In this sense, machine learning (ML) may be thought of as being a subset of AI, as ML in general may be thought of as the scientific study of algorithms and statistical models that computer systems use to help perform specific tasks effectively without using explicit instructions, instead relying on, for example, patterns and inferences.
  • AI problems oftentimes are addressed using search and optimization tools, logic programming and automated reasoning, probabilistic approaches to uncertain reasoning (using, for example, Bayesian networks, Hidden Markov Models, etc.), classifiers and statistical learning methods, neural networks, etc. Models oftentimes underlie these and/or other AI-related techniques. For instance, ML algorithms quite often build a mathematical model based on training data. Once trained, a model can be used to make predictions or decisions without being explicitly programmed to perform the task.
  • Far from being relegated to topics of science fiction, AI and ML techniques are being actively researched and are already implemented in many facets of everyday life, affecting millions of people around the world. Indeed, AI and ML technologies already have practical applications in computer visioning, ecommerce and other recommendation engines, financial market analysis, fraud detection, game playing, image recognition, medical diagnosis, natural language processing, search engines, speech recognition, weather prediction, and many other areas.
  • There is a perception that models trained for a particular application will be good at dealing with issues associated with that application. For instance, an AI model trained to play chess will be good at playing chess. Yet most people would not realistically expect that AI model to be good at poker (even though chess and poker are both games). Nor would most people realistically expect the AI model trained to play chess to be good at detecting credit card fraud or predicting whether a storm will affect a particular geographic region.
  • Although these are somewhat apparent limits to trained models, other limitations may be imposed by the training data used. These limitations may be intentionally or unintentionally introduced, and they may or may not be apparent. For example, there is evidence that Microsoft’s experiment with “Tay,” an AI chat bot on Twitter, was intentionally manipulated by a group of people to create a racist AI “personality.” As another example, an MIT Media Lab researcher found that facial recognition algorithms created by Microsoft, IBM, and Face++ could identify a white man with near-perfect precision, but had significant problems with a darker-skinned woman.
  • These issues might seem extreme to some, but they underscore problems that might arise in connection with models that are presumed to be quite accurate ― and these problems might be not even be known or appreciated by people who rely on output from those models. In other words, a presumed “neutral” model may in essence be trained to have a “bias,” making it function better or worse for a given problem in a given space based on, for example, assumptions made during training, the training methodology or data set used, and/or the like.
  • Certain example embodiments address these and/or other concerns.
  • For instance, one aspect of certain example embodiments relates to quantifying the accuracy of a given model for a general input space. In other words, one aspect of certain example embodiments relates to techniques that provide the ability to make quantifiable, reliable, and reproducible statements about the quality of a given model.
  • Another aspect of certain example embodiments relates to the ability to quantify the accuracy of models so that they can be rated and/or compared to one another using a common and understandable metric.
  • Another aspect of certain example embodiments relates to the ability to perform the above with respect to specific input spaces or areas of interest. For example, certain example embodiments are able to suggest a particular model or training data set of interest for a specific input space or area of interest.
  • In certain example embodiments, a method of quantifying accuracy of a prediction model that has been trained on a data set parameterized by a plurality of features is provided. The model performs in accordance with a theoretical performance manifold over an intractable input space in connection with the features. A determination is made as to which of the plurality of features are strongly correlated with performance of the model. Based on the features determined to be strongly correlated with performance of the model, a plurality of parameterized sub-models are created such that, in aggregate, they approximate the intractable input space. Prototype exemplars are generated for each of the created sub-models, with the prototype exemplars for each created sub-model being objects to which the model can be applied to result in a match with the respective sub-model. The accuracy of the model is quantified using the generated prototype exemplars.
  • In certain example embodiments, a method of recommending a prediction model from among a plurality of different prediction models is provided. Each one of the different prediction models has been trained based on a respective training data set, and each one of the different prediction models performs in accordance with a respective theoretical performance manifold. An indication of a region definable in relation to the theoretical performance manifolds of the different prediction models is received as input. For each of the different prediction models, the indication of the region is linked to features parameterizing the respective performance manifold. One or more portions of the respective performance manifold is/are identified based on the features determined by the linking, with the one or more portions having a volume and a shape that collectively denote an expected performance of the respective model for the input. The expected performance of the different prediction models are compared for the input. Based on the comparison, one or more of the different prediction models is/are recommended.
  • In addition to the features described in the previous paragraphs and in the detailed description below, counterpart systems, computing platforms, non-transitory computer readable storage media tangibly storing instructions for performing such methods, executable computer programs, and the like, are contemplated herein, as well.
  • These features, aspects, advantages, and example embodiments may be used separately and/or applied in various combinations to achieve yet further embodiments of this invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • These and other features and advantages may be better and more completely understood by reference to the following detailed description of exemplary illustrative embodiments in conjunction with the drawings, of which:
  • FIG. 1 is an example performance manifold defined on the set of all valid image collections to which a hypothetical model M can be applied;
  • FIG. 2 shows a subspace of the performance manifold from FIG. 1 , with the training set S and the test set S′;
  • FIG. 3 shows how the performance manifold shown in FIG. 1 can be tessellated so that the aggregate of the tessellation approximates the performance manifold from FIG. 1 ;
  • FIG. 4 shows the performance manifold of FIG. 1 , with images of interest confined to a relatively tight area remote from the training set S and the test set S′;
  • FIG. 5 is a block diagram showing, at a notional level, example components for quantifying accuracy of a prediction model, in accordance with certain example embodiments;
  • FIG. 6 is a flowchart showing an approach for quantifying accuracy of a prediction model, in accordance with the approach shown in and described in connection with FIG. 5 ;
  • FIG. 7 is a block diagram showing example computer system components that may be used to implement the approach shown in and described in connection with FIGS. 5-6 ; and
  • FIG. 8 is a flowchart showing an approach for recommending a prediction model, in accordance with certain example embodiments.
  • DETAILED DESCRIPTION
  • Certain example embodiments described herein relate to techniques for deriving and/or leveraging an application-centric model metric in connection with a model that supports machine learning (ML) and/or artificial intelligence (AI) functionality. That is, certain example embodiments develop and implement an index rating system for quantifying the accuracy of detection and/or other models. Such models can be used in a variety of applications including, for example, identifying objects in satellite imageries, facial recognition, etc. A model in this sense is a learning algorithm that has been trained on a data set parameterized by a plurality of features. In the satellite imagery context, for example, the data set may be parameterized by spatial extent, National Imagery Interpretability Rating Scale (NIIRS), off-nadir angle, cloud coverage, and other attributes. Thus, in the context of this satellite imagery example, features may include spatial extent, NIIRS, off-nadir angle, cloud coverage, and other attributes. The rating index of certain example embodiments helps measure the expected performance of that model on new and/or unseen imageries, i.e., imageries on which the model has not been trained. As such, the rating index of certain example embodiments considers the large and potentially intractable class of all possible unseen imagery inputs.
  • As will be appreciated from the more detailed description below, certain example embodiments generate a compact and accurate approximate representation of the input space, produce prototypical exemplars from the approximate representation, and efficiently evaluate the accuracy of the model for the set of exemplars to estimate a given model’s expected performance. The expected performance in this sense is a statistical measure of the model’s accuracy in the “general case.” However, because a user’s particular interest may markedly differ from the general case, a recommender system based implementation may be provided for use in connection with this index rating. That is, although the expected performance metric may represent the model’s accuracy across the entire spectrum of available inputs (the “general case” alluded to above), some users may be concerned with only limited input spaces (a subset of the general case). In the geospatial context, a limited input space (a subset of the general case) might involve, for example, locating objects in a desert where coloration and topography might be relatively uniform, in a cityscape where off-nadir images may predominate, etc. Of course, in this context, coloration and topography might be relatively uniform within desert and snow images respectively, but coloration and topography may different between the two.
  • From a data science perspective, the inventors have observed that a detection model M can be thought of as being trained on object class type T and an image collection S. A set of features F=[ƒ1..ƒk] parameterizes the training regime of model M. For example, feature f may represent cloud coverage, NIIRS, off-nadir angle, etc., as noted in the above-mentioned satellite imagery example. Thus, the performance of model M is defined by its model architecture and F, i.e., M(F). G1..n(M,P) is the set of performance metrics for model M and an arbitrary image collection P. For a given performance metric Gi, the determination of the rating metric in certain example embodiments involves estimating the expected performance of model M for detecting object T from a novel (unseen) image collection S′ or E[Gi(M,S′)], and the variation in the performance metrics of the model on the novel (unseen) image collection S′ or VAR[Gi(M,S′)].
  • This technique may be used on a number of different model types. This includes, for example, convolutional neural networks (CNNs), naïve Bayesian networks (NBNs), decision trees, etc. Typical performance metrics may include mean average precision, receiver operating characteristic (ROC) related curves (such as precision-recall ROC or P-ROC), an F-beta score (e.g., F1), single scatter, latent feature measures from item-response theory (IRT), and/or the like. It therefore will be appreciated that the techniques described herein may be used in connection with a variety of different possible performance metrics for a plurality of different possible model types, with the same or different performance metric(s) being applicable to the same or different models in different instances.
  • Constraints are introduced, including on the novel (unseen) image collection S′, in order to enable computations related to E[Gi(M,S′)] and VAR[Gi(M,S′)]. For example, training inputs are modeled so as to be constrained to argminS′ (DISCREPANCY (S′,U)) where U may represent training data S, or the set of all valid image collections to which the model M can be applied, α. Here, argminS′ refers to the ML model assumption that the actual data is similar to a given collection of data (e.g., training data S), with some small discrepancy, and DISCREPANCY can be measured in different ways to in essence query how well S′ matches the input data set. For instance, DISCREPANCY can be symmetric KL-divergence on probability distributions P(S), P(S′), P(α), where a is the set of all valid image collections to which the model M can be applied. It will be appreciated that inverse SIMILARITY can be used in place of DISCREPANCY in certain example embodiments. These and/or other constraints may be introduced in different example instances, e.g., based on domain knowledge, specific applications or use cases, etc.
  • Models perform in accordance with a theoretical performance manifold. FIG. 1 is an example performance manifold defined on the set of all valid image collections to which the model M can be applied. The model M is parameterized by features. In FIG. 1 , features f1 and f2 are shown, and additional features are implied by the color gradient. The surface area of subspace projections (coverage) of the performance manifold is normalized against the volume (G) of S′ in the performance manifold. In this regard, FIG. 2 shows a subspace of the performance manifold from FIG. 1 , with the training set S and the test set S′. The subspace thus implies an additional constraint on the model training inputs.
  • Valid inputs are modeled so as to be constrained to argminS′ (DISCREPANCY (S′,α)). As above, the discrepancy can be measured in different ways, inverse SIMILARITY can be used in place of DISCREPANCY, etc.
  • Each of these constraint scenarios involves the determination of S′ against a complex and potentially infinite input set. Other challenges relate to the fact that a is in general intractable and/or extremely large, and that there are high computational requirements for calculating the performance metrics Gi(M,.), i.e., metrics of the model M on S′ and/or different conditions placed on S′ (such as, for example, α).
  • As noted above, certain example embodiments generate a compact and accurate approximate representation of the input space, produce prototypical exemplars from the approximate representation, and efficiently evaluate the accuracy of the model for the set of exemplars to estimate the model’s expected performance. Using the definitions and “problem setup” discussed above, then, certain example embodiments create a tractable model of α, generate S′ from proxy of α, and estimate E[Gi(M,S′)] and VAR[Gi(M,S′)]. The creation of the tractable model of α involves developing a proxy for α to which computation of Gi() can be applied. This is possible because the proxy for α is tractable (e.g., as enabled by pruning). The generation of S′ from the proxy of α involves creating and/or deriving prototype exemplars of S′ from the proxy of α. And the estimation of E[Gi(M,S′)] and VAR[Gi(M,S′)] involves computing Gi() based on the prototypes of S′ and estimating the probabilities of the exemplars. These operations will be discussed, in turn, below.
  • When it comes to creating a tractable model of α, the data image features F that are strongly correlated to the model M’s performance are determined. There are a number of different ways that F can be determined. For example, F can be defined a priori to include features known or suspected to be of interest. In the geospatial image context, for example, cloud cover, NIIRS, and/or other factors may be specified by a user a priori. In general, domain experience may be taken into account in identifying features. As another example, F can be defined using embedding projections such as feature extractors. Feature extractors based on a residual network algorithm (e.g., ResNet50) may be used in this regard.
  • Based on F, a piecewise aggregate model is created. This piecewise aggregate model in essence partitions α into “simplified” parameterized subsets or sub-models. FIG. 3 , for example, shows how the performance manifold shown in FIG. 1 can be tessellated so that the aggregate of the tessellation approximates the performance manifold from FIG. 1 . Different approaches may be used to create the sub-models in different example embodiments. Example approaches may include a priori creation of sub-models, data-driven sub-model creation, hybrid approaches that use a priori creation of sub-models and data-driven sub-model creation, etc. Further details regarding some example approaches are provided below.
  • First, as indicated above, a data-driven sub-model creation approach may be used in certain example embodiments. In this regard, a data-driven sub-model creation approach may involve deriving sub-models through clustering or the like. For instance, a similar metric on F and G() can be generated and used as input to a clustering algorithm. This may be performed on any images that may be represented by the manifold including, for example, training and test samples, in certain example embodiments.
  • Second, a priori creation of sub-model may be facilitated by developing or leveraging an existing taxonomy of image classes in certain example embodiments. In the geospatial context, for example, a taxonomy may specify images with cloud cover of 0-10%, 11-20%, etc. Such taxonomies may be “flat” or multi-level hierarchies in different example embodiments. Data analytics can be used to develop clusters of images that should be grouped together in certain example embodiments, a priori, similar to the approach set forth in the preceding paragraph.
  • In general, it may be desirable to have each sub-model or partition converge to some uniform distribution of values. The value that each sub-model or partition converges to may be one or more performance metrics G(), for example. This may help to determine areas where the model performs uniformly. It will be appreciated that this strict assumption may be loosened in certain example embodiments (e.g., in response to a determination that there are too many partitions, the partitions created are too small, clusters are too close together, etc.).
  • With respect to generating S′ from the proxy of α, prototype exemplars are generated for each sub-model from above. This in essence helps to create a sufficient sample size for each partition so that a check can be performed as to how well the partitions are formed. Synthetics may be used to generate foreground or background images, for instance. This approach differs from typical uses of synthetics. That is, synthetics typically are used to determine training data in a manner that is careful to avoid creating artifacts that could hamper performance of the model (e.g., creating too many shadows in an image). Here, however, synthetics can be used for testing the model (or at least parts thereof), and the creation of artifacts might not be problematic. To the contrary, such artifacts might or might not actually be relevant to the tested part of the model. Thus, certain example embodiments may involve determining what features matter to a given part, and/or the whole, of the model (a form of sensitivity testing with respect to, or perturbation of, individual features) in generating synthetic images. It will be appreciated that the exemplar is not limited to a particular instance of an image or collection of images (real or otherwise) but instead can be objects parameterized explicitly on F. This may allow for estimates of G() based on F, explicitly.
  • If there exists a sufficient set of (prototype) proxy exemplars of α within a predefined neighborhood of S′ (suggesting that the prototype exemplars are representative of the intractable input space), then it follows that estimates on the expected performance of the exemplars from S′ will approximate the expected performance of the model M. In this regard, it can be assumed that:
    • E[Gi(M,S′)] ≈ ∑Gi(M,q)*P(q), where q is a proxy exemplar and P(q) is the probability of sub-model represented by the exemplar
    • VAR[Gi(M,S′)] ≈ ∑Gi(M,q)^2*P(q)] - [∑Gi(M,q)*P(q)]^2
    • Etc.
  • On the other hand, if no sufficient proxy exemplars exist within the predefined neighborhood of S′ (suggesting that the prototype exemplars are not representative of the intractable input space and/or that the prototype exemplars are completely parameterized by the features), the proxy exemplars are to be estimated. In this case, a regression is created for G on the set of exemplars q. The formulae above will apply here as well, except that G=RM(q). Regression may be advantageous in that it becomes possible to interpolate and/or extrapolate performance once sufficient data is obtained. This in turn enables the performance manifold to be calculated directly.
  • In some scenarios, it may be possible to attain a tight lower bound on E[Gi(M,S′)] and VAR[Gi(M,S′)]. Deriving new and accurate exemplars within the small hyperball may be a much less challenging task than deriving exemplars at an arbitrary point within the proxy of α. That is, for an arbitrary image i, it typically will be easier to generate synthetics or obtain real images that are extremely similar to i. A small hyperball may be superimposed for each q. The performance metrics G() may be computed using exemplars inside the ball, and VAR() can be computed as well. The MAX of VAR() from all hyperballs can be considered and may in some instances serve as an empirical lower bound on the variance component of the metric, a mean component, etc.
  • As noted above, a user might have a limited area of interest. FIG. 4 , for example, shows the performance manifold of FIG. 1 , with images of interest confined to a relatively tight area remote from the training set S and the test set S′. Because the model M’s performance for a user’s area of interest can deviate significantly from E[Gi(M,S′)], a recommender system may be employed. In certain example embodiments, instead of computing the performance metric for S′, the performance metric is computed based on the user’s area of interest, U c S′, where U is small (and in some cases potentially extremely small) subset of S′. This will allow differentiated scoring of models based on user’s unique needs in certain example embodiments. In certain example embodiments, a user may specify features of interest directly, the system may infer features of interest (e.g., by analyzing example images and deriving those features, etc.), and/or the like.
  • For example, if a user is interested in models X and Y as they apply to a desert region, then the recommender system will output scores based on the desert region. However, if that user is interested in those models as they apply to a snow region, then the system will (re)compute scores based on snow terrain. Relative performance of X and Y may be different for desert and snow, e.g., X may outperform Y in the desert but not in snow. Thus, the recommender system may help the user determine which models are best for a given interest or set of interests.
  • FIG. 5 is a block diagram showing, at a notional level, example components for quantifying accuracy of a prediction model, in accordance with certain example embodiments, and FIG. 6 is a flowchart showing an approach for quantifying accuracy of a prediction model, in accordance with the approach shown in and described in connection with FIG. 5 . As shown in FIG. 5 , the model 500 is parameterized by features 502. The model 500 performs in accordance with a theoretical performance manifold over an intractable input space (e.g., all valid data sets to which the model can be applied) in connection with those features 502. The model 500 may, for example, be trained to identify objects in images (e.g., with the images being satellite images or the like), and the features 502 may include, for example, spatial extent, National Imagery Interpretability Rating Scale (NIIRS), off-nadir angle, signal-to-noise ratio (SNR), cloud coverage, and/or other values. The objects themselves may be images and/or image collections, with the objects potentially being parameterized explicitly on the features. In some instances, the intractable input space may be limited to a continuous region defined by the data set on which the prediction model is trained, a plurality of discontinuous regions, and/or the like.
  • The feature deriver module 504 determines which of the features 502 are strongly correlated with performance of the model 500. This may be aided based on user input 506 and/or embedding projections 508 (step 602 in FIG. 6 ). In some cases, at least one of the features 510 determined to be strongly correlated with part of the model 500 may include a non-linear mapping based on a feature from the data set on which the prediction model is trained. For instance, the non-linear mapping may be kernel principal component analysis.
  • The feature deriver module 504 outputs those features 510, and based on the features 510 determined to be strongly correlated with performance of the model 500, the sub-model creation engine 512 creates parameterized sub-models that, in an aggregate tessellation 514, approximate the intractable input space (step 604 in FIG. 6 ). The exemplar generator 516 generates prototype exemplars for each of the created sub-models in the aggregate tessellation 514, with the prototype exemplars for each created sub-model being objects to which the model 500 can be applied to result in a match with the respective sub-model (step 606 in FIG. 6 ).
  • A determination is made as to whether the exemplars are representative of the intractable input space. If so, then the performance metrics for the exemplars are used in connection with output metrics 518. If not, then a regression or the like is run and the results thereof are used in connection with output metrics 518. In other words, the accuracy of the model 500 is quantified using the generated prototype exemplars in one form or another. The quantified accuracy may, for example, reflect the expected performance of the model identifying an object of a given type from new and/or unseen images. Those output metrics may be output for display (e.g., as being a rating on a common rating scale representing accuracy of the model 500), provided for comparison purposes (e.g., with other models), used to power suggestions (e.g., as to whether this model is good or bad for a particular task), etc. (see step 608 in FIG. 6 ).
  • In certain example embodiments, the data set on which the prediction model is trained may be for a first application, the accuracy of the model may be quantified for a data set for a second application, and the first and second applications may be different from one another. In some instances, the data set on which the prediction model is trained may be for a first geospatial and/or geotemporal image type, the accuracy of the model may be quantified for a data set for a second geospatial and/or geotemporal image type, and the first and second geospatial and/or geotemporal image types may be different from one another.
  • It will be appreciated that the modules, engines, generators, etc., shown in and described in connection with FIG. 5 may be implemented in connection with processing resources including at least one processor and a memory. For example, program logic corresponding to these elements may be executed by the processing resources. The output may be displayed on a display device, transmitted to a remote computing device, etc.
  • FIG. 7 is a block diagram showing example computer system components that may be used to implement the approach shown in and described in connection with FIGS. 5-6 . FIG. 7 shows a computing system 702 including processing resources such as, for example, at least one processor 704 operably coupled to a memory 706. The computing system further includes network interface(s) 710 and a display interface(s) 712. Each of these interfaces 710, 712 may include hardware input/output devices (e.g., a network card in the case of the network interface(s) 710), software Application Programming Interfaces (APIs) callable to make use of the hardware elements, and/or the like. The network interface(s) 710 may, for example, provide incoming and/or outgoing connections from/to a computer network such as, for example the Internet.
  • In certain example embodiments, a representation of the model 500 (with the model features 502) at least initially may be external to the computing system 702. For example, it may be stored to an external storage device and uploaded or otherwise made accessible to the computing system 702. An electronic interface making use of the network interface(s) 710 (such as, for example, a network upload, API and/or web service calls, etc.) may enable the computing system 702 to access or otherwise interact with the model 500 and the model features 502.
  • The memory 706 in the computing system 702 includes volatile and/or nonvolatile storage. As shown in the FIG. 7 example, the memory 706 stores an operating system 708, along with software modules suitable for implementing the functionality described above in connection with FIG. 5 . For instance, in certain example embodiments, software modules are stored in the memory and enable the processor(s) 704 to perform functionality needed for the execution of the feature deriver 504, sub-model creation engine 512, exemplar generator 516, etc. Working storage may be used to buffer or otherwise store user input 506 and/or projections 508 usable by the feature deriver 504; output from the modules executed under control of the processor(s) 704 such as, for example, the features 510 determined to be strongly correlated with performance of the model 500, the aggregate tessellation 514, prototype exemplars for each created sub-model, output metrics 518, etc.; and/or other data useful for operation of the overall system. The output generator module 716 provides the display, comparison, suggestions, etc. 520, described above. It may make use of the network interface(s) 710 to generate results to be displayed on a remote display (e.g., if the computing system is a server and is accessed by a remote client computing device), the display interface(s) 712 to output results to a local display device, local storage where results may be cached, etc.
  • In view of the foregoing description, it will be appreciated that the processing resources (including the processor(s) 704 and the memory 706) of the computing system 702 may be configured to control the computing system 702 to perform functionality such as that described above in connection with FIGS. 5-6 . The computing system 702 itself may including one or more computers (e.g., a server or series of servers) or it may be a standalone machine. In certain example embodiments, the model 500 may be collocated with the computing system 702 initially, or the model 500 may be provided or otherwise made accessible to the computing system 702. Results may be output locally and/or provided to a remote user accessing a program implemented over the Internet, for example.
  • FIG. 8 is a flowchart showing an approach for recommending a prediction model, in accordance with certain example embodiments. It will be appreciated that the same or similar architecture as that shown in and described above in connection with FIG. 7 may be used to execute the functionality of that described in connection with FIG. 8 . The model is recommended from among a plurality of different prediction models, with each one of the different prediction models having been trained based on a respective training data set and performing in accordance with a respective theoretical performance manifold. In some cases, each of the different prediction models may be trained based on a different training data set. The models may be received over an electronic interface (including, for example, a hardware network interface, using API calls, web service calls, etc.) if they are remote from the computing system, etc., as described above.
  • In step 802, an indication of a region definable in relation to the theoretical performance manifolds of the different prediction models is received as input. The indication of the region may be defined as a set of one or more attributes that parameterize at least one of the different models, using a plurality of images, etc. User input may be provided locally or from a remote source (e.g., via a web application, over the internet, using a standalone application operating under control of the processing resources, etc.). In step 804, for each of the different prediction models, the indication of the region is linked to features parameterizing the respective performance manifold, and one or more portions of the respective performance manifold is identified based on the features determined by the linking. The one or more portions have a volume and a shape that collectively denote an expected performance of the respective model for the input.
  • In certain example embodiments, for each of the different prediction models, a representation of the respective performance manifold may be generated, and the linking and the identifying may be practiced in connection with the representations of the performance manifolds of the respective prediction models. The generating of the representations of the performance manifolds may comprise, for each of the different prediction models: determining which of the plurality of features parameterizing the respective model are strongly correlated with performance of the respective model (see example techniques disclosed above); based on the features determined to be strongly correlated with performance of the respective model, creating a plurality of parameterized sub-models that, in aggregate, approximate the performance manifold; and defining the representation of the respective model as the aggregation of the sub-models. In some cases, generating of the representations may be performed prior to reception of the input. Certain example embodiments may further include, for each of the different prediction models, generating prototype exemplars for each of the created sub-models, with the prototype exemplars for each created sub-model being objects to which the respective model can be applied to result in a match with the respective sub-model, and with the prototype exemplars characterizing the volume and/or shape for an estimated portion of the performance manifold. The techniques described above may be used in these regards. For instance, the software modules shown in and described in connection with FIG. 7 may be used in certain example embodiments.
  • The expected performance of the different prediction models are compared for the input, in step 806. Based on the comparison, one or more of the different prediction models is recommended in step 808. The expected performance of the models may in some instances reflect accuracy of identifying an object of a given type from new and/or unseen images.
  • In certain example embodiments, similar to as noted above, the objects may be images and/or image collections, with those objects potentially being parameterized explicitly on the features. In some instances, the training data sets and/or the input may include geospatial and/or geotemporal data such as, for example, satellite imagery. For example, the features parameterizing the performance manifolds include spatial extent, National Imagery Interpretability Rating Scale (NIIRS), off-nadir angle, signal-to-noise ratio (SNR), and/or cloud coverage values.
  • It will be appreciated that the techniques used herein may be used in connection with a variety of different models of a variety of different model types trained in connection with a variety of different training sets. Location of objects in geospatial and/or geotemporal images is just one example in this regard.
  • As noted above, certain example embodiments can be used to quantify the accuracy of a given model for a general or specific input space, e.g., such that statements can be made about the quality of a given model. It thus becomes possible to quantify the accuracy of models so that they can be rated and/or compared to one another using a common and understandable metric. The same or similar techniques can also be used to suggest a particular model or training data set of interest for a specific input space or area of interest. Points of failure in a model can also be identified using the technology disclosed herein, and the technology disclosed herein also can be used to help combat adversarial ML.
  • Based on the description provided above, it will be appreciated that certain example embodiments are technically advantageous in that they provide technology-based solutions to technology-based problems. For example, certain example embodiments provide technical solutions to technical problems associated with the general inability to objectively and reproducibly compare model performance across different models, a single model with an intractable input space, and/or the like. As another example, certain example embodiments provide technical solutions to technical problems associated with the difficulties in locating where innate “biases” in ML models and/or AI systems might lie, e.g., because of poorly trained models, intentionally manipulated training data sets, and/or the like.
  • In a similar vein, certain example embodiments provide technical improvements to current ML and AI related modeling tools. In that regard, certain example embodiments improve current ML and AI related tools by enabling objective quantification of model accuracy for models being or already produced, yielding expected performance data for a particular data set even when the input space potentially is unknown, facilitating model-to-model comparisons, etc. As an example, certain example embodiments improve current ML and AI related tools by enabling specific already-produced models to be suggested, and/or by suggesting how to produce models expected to perform well (e.g., by suggesting what models to use, what training data should be used, etc.). As still another example, certain example embodiments provide technical improvements to tools that incorporate models because it becomes possible to know which models to include when performing analyses or making suggestions, which data sets to train on, etc. This may lead to better performance (e.g., increased accuracy, reduced processing time, etc.) of the tools that incorporate such models because they can be expected to perform better based on their quantified performance metrics.
  • It will be appreciated that as used herein, the terms system, subsystem, service, engine, module, programmed logic circuitry, and the like may be implemented as any suitable combination of software, hardware, firmware, and/or the like. It also will be appreciated that the storage locations, stores, and repositories discussed herein may be any suitable combination of disk drive devices, memory locations, solid state drives, CD-ROMs, DVDs, tape backups, storage area network (SAN) systems, and/or any other appropriate tangible non-transitory computer readable storage medium. Cloud and/or distributed storage (e.g., using file sharing means), for instance, also may be used in certain example embodiments. It also will be appreciated that the techniques described herein may be accomplished by having at least one processor execute instructions that may be tangibly stored on a non-transitory computer readable storage medium.
  • While the invention has been described in connection with what is presently considered to be the most practical and preferred embodiment, it is to be understood that the invention is not to be limited to the disclosed embodiment, but on the contrary, is intended to cover various modifications and equivalent arrangements included within the spirit and scope of the appended claims.

Claims (20)

What is claimed is:
1. A method of quantifying accuracy of a prediction model that has been trained on a data set parameterized by a plurality of features, the model operating over an input space in connection with the features, the method comprising:
determining which of the plurality of features are strongly correlated with performance of the model;
based on the features determined to be strongly correlated with performance of the model, creating a plurality of parameterized sub-models that, in aggregate, approximate the input space;
generating prototype exemplars for each of the created sub-models, the prototype exemplars for each created sub-model being objects to which the model can be applied to result in a match with the respective sub-model; and
quantifying the accuracy of the model using the generated prototype exemplars,
wherein the quantifying of the accuracy of the model further comprises, provided that the prototype exemplars are representative of the input space, approximating the variance of the model on a new data set as:
(a) the sum of a set of one or more performance metrics for the model on each of the prototype exemplars squared multiplied by the probability of the respective prototype exemplar matching its respective sub-model, subtracting (b) the square of the sum of the set of the one or more performance metrics for the model on each of the prototype exemplars, multiplied by the probability of the respective prototype exemplar matching its respective sub-model.
2. The method of claim 1, wherein the model is trained to identify objects in images.
3. The method of claim 2, wherein the images are satellite images, and
wherein the features include spatial extent, National Imagery Interpretability Rating Scale (NIIRS), off-nadir angle, signal-to-noise ratio (SNR), and/or cloud coverage values.
4. The method of claim 2, wherein the quantified accuracy reflects the expected performance of the model identifying an object of a given type from new and/or unseen images.
5. The method of claim 1, wherein the quantifying of the accuracy of the model further comprises, provided that the prototype exemplars are completely parameterized by the features:
defining the set of one or more performance metrics for the model as a regression on the model for the prototype exemplars; and
(i) approximating the expected performance of the model on the new data set as the sum of the regression on each prototype exemplar multiplied by the probability of the respective prototype exemplar matching its respective sub-model; and/or (ii) approximating the variance of the model on the new data set as (a) the sum of the regression on each prototype exemplar squared multiplied by the probability of the respective prototype exemplar matching its respective sub-model, subtracting (b) the square of the sum of the regression on each prototype exemplar multiplied by the probability of the respective prototype exemplar matching its respective sub-model.
6. The method of claim 1, wherein the input space represents all valid data sets to which the model can be applied.
7. The method of claim 1, wherein the data set on which the prediction model is trained is for a first application, the accuracy of the model is quantified for a data set for a second application, and the first and second applications are different from one another.
8. The method of claim 1, wherein the data set on which the prediction model is trained is for a first geospatial and/or geotemporal image type, the accuracy of the model is quantified for a data set for a second geospatial and/or geotemporal image type, and the first and second geospatial and/or geotemporal image types are different from one another.
9. A non-transitory computer readable storage medium tangibly storing instructions that, when executed by at least one hardware processor of a computing system, quantify accuracy of a prediction model that has been trained on a data set parameterized by a plurality of features and that operates over an input space in connection with the features, by performing functionality comprising:
determining which of the plurality of features are strongly correlated with performance of the model;
based on the features determined to be strongly correlated with performance of the model, creating a plurality of parameterized sub-models that, in aggregate, approximate the input space;
generating prototype exemplars for each of the created sub-models, the prototype exemplars for each created sub-model being objects to which the model can be applied to result in a match with the respective sub-model; and
quantifying the accuracy of the model using the generated prototype exemplars,
wherein the quantifying of the accuracy of the model further comprises, provided that the prototype exemplars are representative of the input space, approximating the variance of the model on a new data set as:
(a) the sum of a set of one or more performance metrics for the model on each of the prototype exemplars squared multiplied by the probability of the respective prototype exemplar matching its respective sub-model, subtracting (b) the square of the sum of the set of the one or more performance metrics for the model on each of the prototype exemplars, multiplied by the probability of the respective prototype exemplar matching its respective sub-model.
10. The non-transitory computer readable storage medium of claim 9, wherein the model is trained to identify objects in images.
11. The non-transitory computer readable storage medium of claim 10, wherein the quantified accuracy reflects the expected performance of the model identifying an object of a given type from new and/or unseen images.
12. The non-transitory computer readable storage medium of claim 9, wherein the quantifying of the accuracy of the model further comprises, provided that the prototype exemplars are completely parameterized by the features:
defining the set of one or more performance metrics for the model as a regression on the model for the prototype exemplars; and
(i) approximating the expected performance of the model on the new data set as the sum of the regression on each prototype exemplar multiplied by the probability of the respective prototype exemplar matching its respective sub-model; and/or (ii) approximating the variance of the model on the new data set as (a) the sum of the regression on each prototype exemplar squared multiplied by the probability of the respective prototype exemplar matching its respective sub-model, subtracting (b) the square of the sum of the regression on each prototype exemplar multiplied by the probability of the respective prototype exemplar matching its respective sub-model.
13. The non-transitory computer readable storage medium of claim 9, wherein the data set on which the prediction model is trained is for a first geospatial and/or geotemporal image type, the accuracy of the model is quantified for a data set for a second geospatial and/or geotemporal image type, and the first and second geospatial and/or geotemporal image types are different from one another.
14. A system for quantifying accuracy of a prediction model that has been trained on a data set parameterized by a plurality of features, the model operating over an input space in connection with the features, the system comprising:
an electronic interface over which the model is received; and
processing resources including at least one processor and a memory coupled thereto, the processing resources being configured to at least:
determine which of the plurality of features are strongly correlated with performance of the model;
based on the features determined to be strongly correlated with performance of the model, create a plurality of parameterized sub-models that, in aggregate, approximate the input space;
generate prototype exemplars for each of the created sub-models, the prototype exemplars for each created sub-model being objects to which the model can be applied to result in a match with the respective sub-model; and
quantify the accuracy of the model using the generated prototype exemplars,
wherein the quantifying of the accuracy of the model further comprises, provided that the prototype exemplars are representative of the input space, approximating the variance of the model on a new data set as:
(a) the sum of a set of one or more performance metrics for the model on each of the prototype exemplars squared multiplied by the probability of the respective prototype exemplar matching its respective sub-model, subtracting (b) the square of the sum of the set of the one or more performance metrics for the model on each of the prototype exemplars, multiplied by the probability of the respective prototype exemplar matching its respective sub-model.
15. The system of claim 14, wherein the model is trained to identify objects in images.
16. The system of claim 15, wherein the quantified accuracy reflects the expected performance of the model identifying an object of a given type from new and/or unseen images.
17. The system of claim 14, wherein the processing resources are further configured to at least determine which features are strongly correlated with performance of the model by receiving a user-specified list of one or more features and/or by running a residual network feature extractor.
18. The system of claim 14, wherein the prototype exemplars are generated using synthetics.
19. The system of claim 14, wherein the objects are images and/or image collections, the objects being parameterized explicitly on the features.
20. The system of claim 14, wherein at least one of the features determined to be strongly correlated with part of the model includes a non-linear mapping based on a feature from the data set on which the prediction model is trained.
US18/203,967 2019-05-31 2023-05-31 Techniques for deriving and/or leveraging application-centric model metric Pending US20230325726A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/203,967 US20230325726A1 (en) 2019-05-31 2023-05-31 Techniques for deriving and/or leveraging application-centric model metric

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201962855138P 2019-05-31 2019-05-31
US16/887,670 US11699108B2 (en) 2019-05-31 2020-05-29 Techniques for deriving and/or leveraging application-centric model metric
US18/203,967 US20230325726A1 (en) 2019-05-31 2023-05-31 Techniques for deriving and/or leveraging application-centric model metric

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/887,670 Continuation US11699108B2 (en) 2019-05-31 2020-05-29 Techniques for deriving and/or leveraging application-centric model metric

Publications (1)

Publication Number Publication Date
US20230325726A1 true US20230325726A1 (en) 2023-10-12

Family

ID=73549719

Family Applications (4)

Application Number Title Priority Date Filing Date
US16/887,670 Active 2041-03-21 US11699108B2 (en) 2019-05-31 2020-05-29 Techniques for deriving and/or leveraging application-centric model metric
US16/887,705 Active 2041-05-25 US11657334B2 (en) 2019-05-31 2020-05-29 Techniques for deriving and/or leveraging application-centric model metric
US18/134,370 Pending US20230252362A1 (en) 2019-05-31 2023-04-13 Techniques for deriving and/or leveraging application-centric model metric
US18/203,967 Pending US20230325726A1 (en) 2019-05-31 2023-05-31 Techniques for deriving and/or leveraging application-centric model metric

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US16/887,670 Active 2041-03-21 US11699108B2 (en) 2019-05-31 2020-05-29 Techniques for deriving and/or leveraging application-centric model metric
US16/887,705 Active 2041-05-25 US11657334B2 (en) 2019-05-31 2020-05-29 Techniques for deriving and/or leveraging application-centric model metric
US18/134,370 Pending US20230252362A1 (en) 2019-05-31 2023-04-13 Techniques for deriving and/or leveraging application-centric model metric

Country Status (1)

Country Link
US (4) US11699108B2 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11297083B1 (en) * 2019-08-15 2022-04-05 Ca, Inc. Identifying and protecting against an attack against an anomaly detector machine learning classifier
US11455555B1 (en) 2019-12-31 2022-09-27 Meta Platforms, Inc. Methods, mediums, and systems for training a model
US11971796B2 (en) * 2021-05-18 2024-04-30 International Business Machines Corporation Goal seek analysis based on status models

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104346372B (en) * 2013-07-31 2018-03-27 国际商业机器公司 Method and apparatus for assessment prediction model
WO2018209057A1 (en) * 2017-05-11 2018-11-15 The Research Foundation For The State University Of New York System and method associated with predicting segmentation quality of objects in analysis of copious image data
US11544494B2 (en) * 2017-09-28 2023-01-03 Oracle International Corporation Algorithm-specific neural network architectures for automatic machine learning model selection
US11423259B1 (en) * 2017-12-12 2022-08-23 Amazon Technologies, Inc. Trained model approximation
US20190325342A1 (en) * 2018-04-20 2019-10-24 Sri International Embedding multimodal content in a common non-euclidean geometric space
JP6992883B2 (en) * 2018-05-07 2022-01-13 日本電気株式会社 Model delivery system, method and program
US10410090B1 (en) * 2018-08-23 2019-09-10 Palantir Technologies Inc. Machine learning assistant for image analysis
US10936768B2 (en) * 2018-10-15 2021-03-02 Aible, Inc. Interface for visualizing and improving model performance

Also Published As

Publication number Publication date
US20230252362A1 (en) 2023-08-10
US20200380307A1 (en) 2020-12-03
US11657334B2 (en) 2023-05-23
US11699108B2 (en) 2023-07-11
US20200380308A1 (en) 2020-12-03

Similar Documents

Publication Publication Date Title
US20230325726A1 (en) Techniques for deriving and/or leveraging application-centric model metric
US10891524B2 (en) Method and an apparatus for evaluating generative machine learning model
CN111783902B (en) Data augmentation, service processing method, device, computer equipment and storage medium
US11403554B2 (en) Method and apparatus for providing efficient testing of systems by using artificial intelligence tools
CN111724083A (en) Training method and device for financial risk recognition model, computer equipment and medium
CN111369299B (en) Identification method, device, equipment and computer readable storage medium
US20210166150A1 (en) Integrated bottom-up segmentation for semi-supervised image segmentation
US20200320381A1 (en) Method to explain factors influencing ai predictions with deep neural networks
US11663486B2 (en) Intelligent learning system with noisy label data
US11681914B2 (en) Determining multivariate time series data dependencies
US11580420B2 (en) Determining feature impact within machine learning models using prototypes across analytical spaces
KR20190029083A (en) Apparatus and Method for learning a neural network
Domeniconi et al. Composite kernels for semi-supervised clustering
KR20200047006A (en) Method and system for constructing meta model based on machine learning
US20220036232A1 (en) Technology for optimizing artificial intelligence pipelines
Rehman et al. Federated self-supervised learning for video understanding
CN109685104B (en) Determination method and device for recognition model
CN112817563A (en) Target attribute configuration information determination method, computer device, and storage medium
CN114298299A (en) Model training method, device, equipment and storage medium based on course learning
CN111161238A (en) Image quality evaluation method and device, electronic device, and storage medium
US20210149793A1 (en) Weighted code coverage
KR102413588B1 (en) Object recognition model recommendation method, system and computer program according to training data
US20220171985A1 (en) Item recommendation with application to automated artificial intelligence
CN114255381A (en) Training method of image recognition model, image recognition method, device and medium
US20240152440A1 (en) Game performance prediction across a device ecosystem

Legal Events

Date Code Title Description
AS Assignment

Owner name: RADIANT ANALYTIC SOLUTIONS INC., VIRGINIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BOEDIHARDJO, ARNOLD;ESTRADA, ADAM;JENKINS, ANDREW;AND OTHERS;SIGNING DATES FROM 20190913 TO 20190917;REEL/FRAME:063839/0523

Owner name: MAXAR MISSION SOLUTIONS INC., COLORADO

Free format text: CHANGE OF NAME;ASSIGNOR:RADIANT MISSION SOLUTIONS INC.;REEL/FRAME:063839/0515

Effective date: 20210101

Owner name: RADIANT MISSION SOLUTIONS INC., COLORADO

Free format text: MERGER;ASSIGNORS:RADIANT ANALYTIC SOLUTIONS INC.;RADIANT GEOSPATIAL SOLUTIONS INC.;THE HUMAN GEO GROUP LLC;REEL/FRAME:063839/0502

Effective date: 20210101

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: SIXTH STREET LENDING PARTNERS, AS ADMINISTRATIVE AGENT, TEXAS

Free format text: INTELLECTUAL PROPERTY SECURITY AGREEMENT;ASSIGNORS:MAXAR INTELLIGENCE INC.;MAXAR MISSION SOLUTIONS INC.;MAXAR SPACE LLC;AND OTHERS;REEL/FRAME:064907/0036

Effective date: 20230912