US20230004536A1 - Systems and methods for a data search engine based on data profiles - Google Patents

Systems and methods for a data search engine based on data profiles Download PDF

Info

Publication number
US20230004536A1
US20230004536A1 US17/930,926 US202217930926A US2023004536A1 US 20230004536 A1 US20230004536 A1 US 20230004536A1 US 202217930926 A US202217930926 A US 202217930926A US 2023004536 A1 US2023004536 A1 US 2023004536A1
Authority
US
United States
Prior art keywords
data
model
dataset
datasets
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/930,926
Inventor
Austin Walters
Vincent Pham
Galen Rafferty
Anh Truong
Mark Watson
Jeremy Goodsitt
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Capital One Services LLC
Original Assignee
Capital One Services LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Capital One Services LLC filed Critical Capital One Services LLC
Priority to US17/930,926 priority Critical patent/US20230004536A1/en
Assigned to CAPITAL ONE SERVICES, LLC reassignment CAPITAL ONE SERVICES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GOODSITT, JEREMY, PHAM, VINCENT, RAFFERTY, GALEN, TRUONG, ANH, WALTERS, AUSTIN, WATSON, MARK
Publication of US20230004536A1 publication Critical patent/US20230004536A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/242Query formulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/21Design, administration or maintenance of databases
    • G06F16/211Schema design and management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2458Special types of queries, e.g. statistical queries, fuzzy queries or distributed queries
    • G06F16/2468Fuzzy queries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/285Clustering or classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30003Arrangements for executing specific machine instructions
    • G06F9/30007Arrangements for executing specific machine instructions to perform operations on data operands
    • G06F9/30036Instructions to perform operations on packed data, e.g. vector, tile or matrix operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/01Dynamic search techniques; Heuristics; Dynamic trees; Branch-and-bound
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks

Definitions

  • search engines may search for user-generated search terms in an index of website data. Results may be based on the frequency of key words and/or clustering of key words.
  • search-term based approaches may be inefficient, produce misleading or irrelevant results, and have limited scope (i.e., the search may be limited to a small number of drives, databases, or online resources).
  • search-term based methods may not account for the profile of a dataset (e.g., a data schema and/or statistical metric).
  • an agent e.g., a user, a system, a model, or other agent
  • the objective may be to identify regional voting trends and find data related to that objective.
  • the agent may possess a sample dataset that includes addresses, phone numbers, names, dates, and voting records.
  • the sample dataset may have a data schema (i.e., a data structure or organization).
  • a search-term based search e.g., a search for a name
  • a search-term based search may be slow or miss relevant results because a search term is highly specific. Further, conventional systems may not allow the user to control how the search is performed (i.e., to set search parameters such as filtering results that contain a term, etc.). For example, some systems may determine which indexes to use on a search without an understanding of which index will perform better,
  • non-transitory computer readable storage media may store program instructions, which are executed by at least one processor device and perform any of the methods described herein.
  • FIG. 1 depicts an exemplary system for searching data, consistent with disclosed embodiments.
  • FIG. 5 depicts an exemplary process for profiling data, consistent with disclosed embodiments.
  • embodiments may include submitting a sample dataset (e.g., a CSV file or other dataset).
  • a sample dataset e.g., a CSV file or other dataset.
  • disclosed embodiments may include searching one or more databases, computing environments, and/or data indexes based on the sample datasets.
  • Embodiments may include receiving one or more search results.
  • Search results may include a dataset, a data vector, and/or a data source that comprises the dataset.
  • the dataset may be the same as or similar to the sample dataset according to a similarity metric representing a measure of similarity between data in a dataset and data in the sample dataset.
  • FIG. 1 is a diagram of an exemplary system 100 for searching for data, consistent with disclosed embodiments.
  • system 100 may include a model optimizer 102 , a model storage 104 , a data-searching system 106 , an aggregation database 108 , a data-migration system 110 , an interface 112 , a client device 114 , one or more computing environments 116 a, 116 b, 116 n, and one or more databases 118 a , 118 b, 118 n, Components of system 100 may be connected to each other through a network 120 .
  • Model optimizer 102 may be configured to generate models based on instructions received from a user or another system (e.g., via interface 112 ). For example, model optimizer 102 may be configured to receive a visual (graphical) depiction of a machine learning model and parse that graphical depiction into instructions for creating and training a corresponding neural network. Model optimizer 102 may be configured to select model training hyperparameters. This selection may be based on model performance feedback produced by model optimizer 102 during model training and/or received from another component of system 100 . Model optimizer 102 may be configured to provide trained models and descriptive information concerning the trained models to model storage 104 .
  • Model optimizer 102 may be configured to train machine learning models by optimizing model parameters and/or hyperparameters (hyperparameter tuning) using an optimization technique, consistent with disclosed embodiments.
  • the optimization technique may include a grid search, a random search, a gaussian process, a Bayesian process, a Covariance Matrix Adaptation Evolution Strategy (CMA-ES), a derivative-based search, a stochastic hill-climb, a neighborhood search, an adaptive random search, or the like.
  • Tuning a hyperparameter may include iteratively selecting a hyperparameter and training model parameters using the hyperparameter until a training criterion is satisfied, consistent with disclosed embodiments.
  • Model storage 104 may be configured to store models, including machine learning models, consistent with disclosed embodiments. In some embodiments, some or all components of model storage 104 may be hosted on one or more servers, one or more clusters of servers, or one or more cloud services. Model storage 104 may be connected to network 120 and may additionally or alternatively be directly connected to model optimizer 102 (not shown). In some embodiments, model storage 104 may be a component of model optimizer 102 , data-searching system 106 , or client device 114 (not shown).
  • Model storage 104 may include one or more databases configured to store data models (e.g., machine learning models or statistical models) and descriptive information of the data models. Model storage 104 may be configured to provide information regarding available data models to a user or another system.
  • the databases may include cloud-based databases, cloud-based buckets, or on-premises databases.
  • the information may include model information, such as the type and/or purpose of the model and any measures of classification error.
  • Model storage 104 may include one or more databases configured to store indexed and clustered models for use by system 100 .
  • model storage 104 may store models associated with generalized representations of those models (e.g., neural network architectures stored in TENSORFLOW or other standardized formats).
  • the databases may include cloud-based databases (e.g., AMAZON WEB SERVICES S3 buckets) or on-premises databases.
  • Data-searching system 106 may include a computing device, a computer, a server, a server duster, a plurality of dusters, and/or a cloud service, consistent with disclosed embodiments.
  • Data-searching system 106 may include one or more memory units and one or more processors configured to perform operations consistent with disclosed embodiments.
  • Data-searching system 106 may include computing systems configured to manage training of models for system 100 , profile datasets, index datasets, and perform other methods consistent with disclosed embodiments.
  • Data-searching system 106 may be configured to receive and/or retrieve models, and/or datasets from other components of system 100 or from computing components outside system 100 (e.g., via interface 112 ). In some embodiments not shown, data-searching system 106 may be a component of data-migration system 110 . Data-searching system 106 is disclosed in greater detail below (in reference to FIG. 2 ).
  • Aggregation database 108 may include one or more databases configured to store data for use by system 100 , consistent with disclosed embodiments.
  • the databases may include cloud-based databases (e.g., AMAZON WEB SERVICES S3 buckets) or on-premises databases.
  • aggregation database 108 may store one or more datasets, data vectors, and data indexes.
  • aggregation database 108 may be configured to store one or more data vectors comprising statistical metrics of a dataset and information based on a corresponding data schema of the dataset.
  • Aggregation database 108 may include data received from other components of system 100 , including, for example, data-migration system 110 and/or data-searching system 106 .
  • Data-migration system 110 may include a computing device, a computer, a server, a server cluster, a plurality of clusters, and/or a cloud service, consistent with disclosed embodiments.
  • Data-migration system 110 may include one or more memory units and one or more processors configured to perform operations consistent with disclosed embodiments.
  • Data-migration system 110 may include computing systems configured to manage movement (i.e., transfer) of data between components of system 100 , to manage movement of data between components of system 100 and computing components outside system 100 , and to perform other methods consistent with disclosed embodiments.
  • Data-migration system 110 may be configured to receive and/or retrieve models, and/or datasets from other components of system 100 or from computing components outside system 100 (e.g., via interface 112 ).
  • data-migration system 110 may be a component of data-searching system 106 .
  • Data-migration system 110 is disclosed in greater detail below (in reference to FIG. 3 ).
  • Interface 112 may be configured to manage interactions between system 100 and other systems using network 120 , consistent with disclosed embodiments.
  • interface 112 may be configured to publish data received from other components of system 100 . This data may be published in a publication and subscription framework (e.g., using APACHE KAFKA), through a network socket, in response to queries from other systems, or using other known methods. The data may be synthetic data, as described herein.
  • interface 112 may be configured to provide information received from model storage 104 regarding available datasets.
  • interface 112 may be configured to provide data or instructions received from other systems to components of system 100 .
  • interface 112 may be configured to receive instructions for generating data models (e.g., type of data model, data model parameters, training data indicators, training hyperparameters, or the like) from another system and provide this information to model optimizer 102 .
  • interface 112 may be configured to receive data including sensitive portions from another system (e.g., in a file, a message in a publication and subscription framework, a network socket, or the like) and provide that components of system 100 .
  • Client device 114 may include one or more memory units and one or more processors configured to perform operations consistent with disclosed embodiments.
  • client device 114 may include hardware, software, and/or firmware modules.
  • Client device 114 may be a terminal, a kiosk, a mobile device, a tablet, a personal computer, a server, a server cluster, a cloud service, a storage device, a specialized device configured to perform methods according to disclosed embodiments, or the like.
  • Databases 118 a, 118 b, 118 n may include one or more databases configured to store data for use by system 100 , consistent with disclosed embodiments.
  • the databases may include cloud-based databases (e.g., AMAZON WEB SERVICES 53 buckets) or on-premises databases,
  • one or more databases 118 a, 118 b, 118 n may store one or more datasets, data vectors, and data indexes.
  • at least one of databases 118 a, 118 b, 118 n may be configured to store one or more data vectors comprising statistical metrics of a dataset and information based on a corresponding data schema of the dataset.
  • databases 118 a, 118 b, 118 n are exemplary only and system 100 may include additional or fewer databases.
  • model optimizer 102 may connect to network 120 .
  • model storage 104 may connect to network 120 .
  • data-searching system 106 may connect to network 120 .
  • aggregation database 108 may connect to network 120 .
  • client device 114 may connect to network 120 .
  • Network 120 may be a public network or private network and may include, for example, a wired or wireless network, including, without limitation, a Local Area Network, a Wide Area Network, a Metropolitan Area Network, an IEEE 1002.11 wireless network (e.g., Fr), a network of networks (e.g., the Internet), a land-line telephone network, or the like.
  • Network 120 may be connected to other networks (not depicted in FIG. 1 ) to connect the various system components to each other and/or to external systems or devices.
  • network 120 may be a secure network and require a password to access the network.
  • FIG. 2 depicts exemplary data-searching system 106 , consistent with disclosed embodiments.
  • Data-searching system 106 may include a computing device, a computer, a server, a server cluster, a plurality of clusters, and/or a cloud service, consistent with disclosed embodiments.
  • data-searching system 106 may include one or more processors 210 , one or more I/O devices 220 , and one or more memory units 230 .
  • some or all components of data-searching system 106 may be hosted on a device, a computer, a server, a cluster of servers, or a cloud service.
  • data-searching system 106 may be a scalable system configured to efficiently manage resources and enhance security by provisioning computing resources in response to triggering events and terminating resources after completing a task (e.g., a scalable cloud service that spins up and terminates container instances).
  • FIG. 2 depicts an exemplary configuration of data-searching system 106 .
  • data-searching system 106 may include a larger or smaller number of processors, I/O devices, or memory units.
  • data-searching system 106 may further include other components or devices not depicted that perform or assist in the performance of one or more processes consistent with the disclosed embodiments.
  • the components and arrangements shown in FIG. 2 are not intended to limit the disclosed embodiments, as the components used to implement the disclosed processes and features may vary.
  • Processor 210 may comprise known computing processors, including a microprocessor.
  • Processor 210 may constitute a single-core or multiple-core processor that executes parallel processes simultaneously.
  • processor 210 may be a single-core processor configured with virtual processing technologies.
  • processor 210 may use logical processors to simultaneously execute and control multiple processes.
  • Processor 210 may implement virtual machine technologies, or other known technologies to provide the ability to execute, control, run, manipulate, store, etc., multiple software processes, applications, programs, etc.
  • processor 210 may include a multiple-core processor arrangement (e.g., dual core, quad core, etc.) configured to provide parallel processing functionalities to allow execution of multiple processes simultaneously.
  • Programs 235 may include one or more programs (e.g., modules, code, scripts, or functions) used to perform methods consistent with disclosed embodiments.
  • Programs may include operating systems (not shown) that perform known operating system functions when executed by one or more processors.
  • Disclosed embodiments may operate and function with computer systems running any type of operating system.
  • Programs 235 may be written in one or more programming or scripting languages, One or more of such software sections or modules of memory 230 may be integrated into a computer system, non-transitory computer-readable media, or existing communications software.
  • Programs 235 may also be implemented or replicated as firmware or circuit logic.
  • Programs 235 may include a model-training module 236 , data-profiling module 237 , aggregator 238 , searching module 239 and/or other modules not depicted to perform methods of the disclosed embodiments.
  • modules of programs 235 may be configured to generate (“spin up”) one or more ephemeral container instances to perform a task and/or to assign a task to a running (warm) container instance, consistent with disclosed embodiments.
  • Modules of programs 235 may be configured to receive, retrieve, and/or generate models, consistent with disclosed embodiments.
  • Modules of programs 235 may be configured to receive, retrieve, and/or generate datasets (e.g., to generate synthetic datasets, data samples, or other datasets), consistent with disclosed embodiments.
  • Modules of programs 235 may be configured to perform operations in coordination with one another.
  • Model-training module 236 may be configured to train one or more models and/or perform hyperparameter tuning of one or more models, including machine learning models.
  • model-training module 236 may be configured to receive input of one or more thresholds, one or more loss functions, and/or one or more limits on a number of interactions and apply the input for optimizing a received model and/or corresponding generalized representation such as a neural network.
  • training of a model terminates when a training criterion is satisfied.
  • model-training module 236 is configured to adjust model parameters during training.
  • the model parameters may include weights, coefficients, offsets, or the like. Training may be supervised or unsupervised.
  • Model-training module 236 may be configured to select or generate model parameters (e.g., number of layers for a neural network, kernel function for a kernel density estimator, or the like), update training hyperparameters, and evaluate model characteristics.
  • model characteristics may include a model type (e.g., an INN, a convolutional neural network (CNN), a random forest, or another model type), a model parameter, a model hyperparameter (including training a hyperparameter and/or an architectural hyperparameter), a desired outcome, belongingness to a model cluster, and/or belonginess of a model training dataset to a dataset duster, the similarity of synthetic data generated by a model to actual data, or other characteristics,
  • model parameters e.g., number of layers for a neural network, kernel function for a kernel density estimator, or the like
  • model characteristics may include a model type (e.g., an INN, a convolutional neural network (CNN), a random forest, or another model type), a model parameter,
  • model-training module 236 may be configured to perform a search of a hyperparameter space and select new hyperparameters. This search may or may not depend on the values of a performance metric obtained for other trained models. In some aspects, model-training module 236 may be configured to perform a grid search or a random search.
  • the hyperparameters may include training hyperparameters, which may affect how training of the model occurs, or architectural hyperparameters, which may affect the structure of the model.
  • model-training module 236 may include programs to apply one or more templates to a data model (e.g., a model retrieved from model storage 104 ) and apply the templates to generate a generalized representation of the retrieved model (e.g., a neural network).
  • Model-training module 236 may include programs to provide trained generalized representations to model storage 104 for storing in association with corresponding models.
  • Data-profiling module 237 may include, may be configured generate, and/or may be configured to implement one or more data-profiling models.
  • a data-profiling model may include machine-learning models and statistical models to determine the data schema and/or a statistical profile of a dataset (i.e., to profile a dataset), consistent with disclosed embodiments.
  • the data-profiling model may include an RNN model, a CNN model, a variational autoencoder (VAE) model, an autoencoder model, or another machine-learning model.
  • VAE variational autoencoder
  • the data-profiling model may include algorithms to determine a data type, key-value pairs, row-column data structure, statistical distributions of information such as keys or values, or other property of a data schema.
  • Searching module 239 may include programs (scripts, functions, algorithms, routines, or the like) to conduct a search, consistent with disclosed embodiments.
  • searching module 239 may include programs to generate and maintain searchable data indexes as data trees based on the data vectors (e.g., a B-Tree).
  • searching module 239 may include programs based on POSTGRESQL or other object-relational database management system.
  • searching module 239 may be configured to generate a data index (e.g., a RUM-index and/or GIN-index and/or Locality-Sensitive Hashing).
  • the data index may be configured to be searched (e.g., as a data tree).
  • the data index may be based on the data vectors and/or data profiles and organized for efficient searching with or without using k-nearest neighbors clustering-based methods.
  • Searching module 239 may be configured to execute a search of the data index.
  • searching module 239 may be configured to receive search parameters such as the types of data to search or other information related to data types and schemas to guide the search of the data index.
  • the search parameters may indicate that the search will weigh the data schema of a dataset more than the particular data values or statistical metrics of the dataset.
  • Search parameters may be received from another component of system 100 or a computing component outside system 100 (e.g., via interface 112 ).
  • the search parameters may include instructions to search a data index by first comparing data schemas then by comparing statistical metrics.
  • Programs 335 may include one or more programs (e.g., modules, code, scripts, or functions) used to perform methods consistent with disclosed embodiments.
  • programs 335 may include a model-training module 336 , data-profiling module 337 , aggregator 338 , migration module 339 and/or other modules not depicted to perform methods of the disclosed embodiments.
  • Model-training module 336 may include, may be configured generate, and/or may be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to model-training module 236 , above.
  • Data-profiling module 337 may include, may be configured generate, and/or may be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to data-profiling module 237 , above.
  • Aggregator 338 may include, may be configured generate, and/or may be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to model-training module 236 , above,
  • FIG. 4 depicts an exemplary configuration of computing environment 116 .
  • computing environment 116 may include a larger or smaller number of processors 410 , I/O devices 420 , or memory units 430 .
  • computing environment 116 may further include other components or devices not depicted that perform or assist in the performance of one or more processes consistent with the disclosed embodiments.
  • the components and arrangements shown in FIG. 4 are not intended to limit the disclosed embodiments, as the components used to implement the disclosed processes and features may vary.
  • steps of process 500 may be repeated.
  • data-migration system 110 may receive an additional dataset and/or an additional data vector and repeat steps of process 500 to update a data index
  • the data-profiling model may include algorithms to determine a data type, key-value pairs, row-column data structure, statistical distributions of information such as keys or values, or other property of a data schema.
  • the data-profiling model may be configured to implement univariate and multivariate statistical methods.
  • the data-profiling model may include a regression model, a Bayesian model, a statistical model, a linear discriminant analysis model, or other classification model configured to determine one or more descriptive metrics of a dataset.
  • the search parameters may first specify to conduct an initial search based on comparisons of data schemas and conduct a search based on comparisons of statistical metrics.
  • the search parameters may include instructions to search the data index by searching data profiles (data schema and statistical metrics) of the datasets and searching data profiles of variables of the datasets.
  • a similarity metric may include a likelihood that a reference dataset derives from sample dataset. In some embodiments, a similarity metric may include a likelihood that the sample dataset derives from a reference dataset. For example, a dataset may derive from another dataset when the data schema and/or data values match according to a threshold or frequency. As an example, a small dataset may be a subset of a large dataset, i.e. it derives from the larger dataset. As another example, one dataset may contain some overlapping values with another dataset and both may contain values that are unique from the other dataset. In that example, either dataset may be said to derive from the other dataset.
  • data-searching system 106 may provide the one or more similarity metrics, consistent with disclosed embodiments.
  • Providing the similarity metrics at step 614 may include displaying the similarity metrics at an interface (e.g., interface 112 , a display of I/O devices 220 , and/or a display of I/O devices 320 ).
  • Providing the similarity metrics at step 614 may include transmitting the similarity metrics to another component of system 100 (e.g., to client device 114 ) and/or to a computing component outside system 100 (e.g., via interface 112 ).
  • data-searching system 106 may select one or more data vectors, consistent with disclosed embodiments.
  • selecting data vectors may be based on the similarity metrics and one or more predetermined selection criteria.
  • the predetermined selection criteria may be received from a user (e.g., at step 602 or as a user-input received via an interface at step 616 ).
  • the selection criteria may include a command to select any data vectors with similarity criteria falling within a range (e.g., any data vectors between 50% to 90% match).
  • Selecting data vectors may include receiving metadata (e.g., one or more tags or labels) associated with the data vectors.
  • data-searching system 106 may retrieve one or more datasets associated with the selected data vectors, consistent with disclosed embodiments, In some embodiments, data-searching system 106 retrieves the datasets from one of data 231 , aggregation database 108 , data-migration system 110 , computing environments 116 a, 116 b, 116 n, databases 118 a, 118 b, 118 n, and/or a computing outside system 100 (e.g., via the interface 112 ).
  • Retrieving a dataset may be based on the metadata of the selected data vectors, Retrieving a dataset may be include identifying a location (e.g., a database, an internet address) or an identifier of the dataset based on a log and the selected data vector.
  • the log may be stored in data aggregation database 108 .
  • data-searching system 106 may provide the datasets associated with the selected data vectors, consistent with disclosed embodiments.
  • Providing the datasets at step 620 may include displaying the datasets at an interface (e.g., interface 112 , a display of I/O devices 220 , and/or a display of I/O devices 320 ).
  • Providing the datasets at step 620 may include transmitting the datasets to another component of system 100 (e.g., to client device 114 ) and/or to a computing component outside system 100 (e.g., via interface 112 ).
  • Providing the datasets may include storing the datasets in a data storage.
  • Providing the datasets may include providing a link to the datasets.
  • Providing the datasets may include providing a compressed file comprising one or more datasets.
  • Providing the datasets may include sending a request to data-migration system 110 to transfer a dataset to a computing component, consistent with disclosed embodiments.
  • Example Implentation Conducting a search based on a sample dataset in .CSV-format comprising public health indicators.
  • a researcher may be conducting a meta-analysis of a plurality of health studies to identify health indicators that predict a health outcome, such as risk factors that contribute to heart attacks.
  • the researcher may use data-searching system 106 to expand the study by identifying additional data sources to include in the study.
  • the researcher may possess an initial sample dataset.
  • the disclosed system and method provide appropriate datasets to address the researcher's goal of expanding the study, because the system provides datasets that are similar to the sample dataset based on the structure and content of sample dataset.
  • conventional search methods may identify datasets that merely happen to contain keywords or other sequences of characters which overlap with the sample datasets.
  • a conventional search may return redundant or dissimilar datasets that cannot be included in the meta-analysis.
  • data-searching system 100 may receive a sample dataset from the researcher's device (e.g., client device 114 ), consistent with disclosed embodiments.
  • the sample dataset may be a .CSV file.
  • Column headers may include an identifier (e.g., an anonymous patient ID, a social security number, etc.), a data of birth, a date associated with the measurement of other parameters, and one or more other parameters.
  • data-searching system 106 identifies a data schema of the sample dataset (step 604 ), consistent with disclosed embodiments.
  • the data schema may include column headers.
  • the columns may be unlabeled and a data-profiling model may identify a data type and provide a label (e.g., provide a column header).
  • Identifying the data schema may include identifying complex data types using a trained model, consistent with disclosed embodiments.
  • data-searching system 106 generates a sample data vector of the sample dataset (step 606 ).
  • generating the sample data vector may include translating the residential address into Global Positioning System coordinates and developing statistical metrics of the GPS coordinates (e.g., the centroid of the GPS coordinates, parameters of a fit to a Poisson distribution or other statistical distribution of the GPS coordinates).
  • the sample data vector may then include the statistical metrics of the GPS coordinates.
  • generating the sample data vector may include developing a mean, median, and variance of the age of individuals in the dataset based on the date of birth and a data associated with the measurement of parameters.
  • the sample data vector may include the mean, median, and variance of the age.
  • data vectors may be generated using machine learning models (e.g., a data-profiling model) and have no human-understandable interpretation.
  • a machine-learning model comprising an autoencoder model, a VAE model, and/or other model may perform feature embedding and generate vectors based on the identified features, consistent with disclosed embodiments.
  • Data-searching system 106 may set the search parameters (step 608 ).
  • the system may receive a set of instructions from client device 114 , the instructions be based on user inputs.
  • the instructions might specify that the search should be performed over a data index stored in aggregation database 108 .
  • the instructions may include a command to assign a weight of 0.8 to a search parameter associated with the data schema (here, column labels).
  • the instructions may further include a command to assign a weight of 0.05 to search parameters corresponding to each of four statistical metrics (e.g., the GPS centroid, the mean age, the median age, and the variance of the age).
  • the search parameters may be determined using a machine-learning model.
  • Data-searching system 106 may then identify a data index and perform a search of the data index (step 610 ), consistent with disclosed embodiments.
  • the data index may be stored in aggregation database 108 and include stored data vectors associated with a plurality of public health datasets from a plurality of sources (i.e., systems).
  • the systems may include systems associated with a university, a research institution, an agency (e.g., Center for Disease Control), a health care provider (e.g., a hospital), an insurance provider, the internet (e.g., a dataset retrieved from the internet by data-migration system 110 ), and/or other systems.
  • Data-searching system 106 may generate similarity metrics based on the search (step 612 ), consistent with disclosed embodiments.
  • the similarity metric may indicate a “percent match” between one or more stored data vectors and the sample data vector.
  • the stored and sample data vectors may include statistical metrics based on complex data types.
  • data-searching system 106 may identify two datasets that share the same data schema as the sample data vector. Data-searching system 106 may give a first one of these datasets a 100% match because it has the same GPS centroid, mean age, median age, and variance of the age as the sample dataset, i.e., data-searching system 106 found the same dataset as the sample dataset.
  • data-searching system 106 may give the second one of these datasets an 80% match because it has a different GPS centroid, mean age, median age, and variance of the age.
  • data-searching system 106 may identify a third dataset has a 75% match because it shares some properties of the data schema and has different statistical metrics.
  • data-searching system 106 may provide the similarity metrics to the researcher's device (step 614 ).
  • the device e.g., client device 114
  • the researcher may display the metrics via an interface.
  • the researcher may then be able to provide user inputs to reject the 100% match data vector and select the 80%- and 75% match data vector (step 618 ).
  • the researcher may be able to click an icon on the interface to select a data vector.
  • data-searching system 106 may then retrieve datasets associated with the selected data vector and provide the retrieved datasets to the user (steps 618 and 620 ).
  • data-searching system 106 may determine that a vector corresponds to a dataset stored in computing environment 116 a and retrieve or send a request to receive the corresponding dataset from computing environment 116 a.
  • the interface may display an option to download the retrieved dataset, and may transmit the received dataset to the a researcher's local device when the researcher selects the option.
  • Computer programs based on the written description and methods of this specification are within the skill of a software developer.
  • the various functions, scripts, programs, or modules can be created using a variety of programming techniques.
  • programs, scripts, functions, program sections or program modules can be designed in or by means of languages, including JAVASCRIPT, C, C++, JAVA, PHP, PYTHON, RUBY, PERL, BASH, or other programming or scripting languages.
  • One or more of such software sections or modules can be integrated into a computer system, non-transitory computer-readable media, or existing communications software.
  • the programs, modules, or code can also be implemented or replicated as firmware or circuit logic.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Probability & Statistics with Applications (AREA)
  • Fuzzy Systems (AREA)
  • Molecular Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Automation & Control Theory (AREA)
  • Algebra (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

Systems and methods for searching data are disclosed. For example, the system may include one or more memory units storing instructions and one or more processors configured to execute the instructions to perform operations. The operations may include receiving a sample dataset and identifying a data schema of the sample dataset. The operations may include generating a sample data vector that includes statistical metrics of the sample dataset and information based on the data schema of the sample dataset. The operations may include searching a data index comprising a plurality of stored data vectors corresponding to a plurality of reference datasets. The stored data vectors may include statistical metrics of the reference datasets and information based on corresponding data schema. The operations may include generating, based on the search and the sample data vector, one or more similarity metrics of the sample dataset to individual ones of the reference datasets.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application No. 62/694,968, filed Jul. 6, 2018, the content of which is expressly incorporated herein by reference in its entirety.
  • BACKGROUND
  • Conventional systems and methods of searching data typically involve searching data trees for a search term (e.g., a keyword, a sequence of characters, and/or a number). For example, a conventional search engine may search for user-generated search terms in an index of website data. Results may be based on the frequency of key words and/or clustering of key words.
  • Data analysis and data modeling efforts often need to identify data sources (e.g., to search for datasets). Problems arise, however, when applying conventional approaches to a search for datasets. Search-term based approaches may be inefficient, produce misleading or irrelevant results, and have limited scope (i.e., the search may be limited to a small number of drives, databases, or online resources). For example, search-term based methods may not account for the profile of a dataset (e.g., a data schema and/or statistical metric).
  • For example, an agent (e.g., a user, a system, a model, or other agent) may set an objective (i.e., a desired outcome). The objective may be to identify regional voting trends and find data related to that objective. As an illustrative example, the agent may possess a sample dataset that includes addresses, phone numbers, names, dates, and voting records. The sample dataset may have a data schema (i.e., a data structure or organization). However, a search-term based search (e.g., a search for a name) may yield large numbers of irrelevant datasets because it may return results unrelated to the desired objective (e.g., it may return phone records but not voting records). A search-term based search may be slow or miss relevant results because a search term is highly specific. Further, conventional systems may not allow the user to control how the search is performed (i.e., to set search parameters such as filtering results that contain a term, etc.). For example, some systems may determine which indexes to use on a search without an understanding of which index will perform better,
  • Therefore, in view of the shortcomings and problems with conventional approaches to searching for datasets, there is a need for unconventional approaches to data searches that are efficient, effective, customizable, and broad in scope. In particular, there is a need for unconventional systems to search data based on sample data.
  • SUMMARY
  • The disclosed embodiments provide unconventional methods and systems for searching datasets based on one or more data profiles. The disclosed systems and methods may be implemented using a combination of conventional hardware and software as well as specialized hardware and software, such as a machine constructed and/or programmed specifically for performing functions associated with the disclosed method steps.
  • Consistent with the present embodiments, a system for searching datasets is disclosed. The system may include one or more memory units storing instructions and one or more processors configured to execute the instructions to perform operations. The operations may include receiving a sample dataset and identifying a data schema of the sample dataset. The operations may include generating a sample data vector that includes statistical metrics of the sample dataset and information based on the data schema of the sample dataset. The operations may include searching a data index comprising a plurality of stored data vectors corresponding to a plurality of reference datasets. The stored data vectors may include statistical metrics of the reference datasets and information based on corresponding data schema of the reference datasets. The operations may include generating, based on the search and the sample data vector, one or more similarity metrics of the sample dataset to individual ones of the reference datasets.
  • Consistent with the present embodiments, a method for searching datasets is disclosed. The method may include receiving a sample dataset and identifying a data schema of the sample dataset. The method may include generating a sample data vector that includes statistical metrics of the sample dataset and information based on the data schema of the sample dataset. The method may include searching a data index comprising a plurality of stored data vectors corresponding to a plurality of reference datasets. The stored data vectors may include statistical metrics of the reference datasets and information based on corresponding data schema of the reference datasets. The method may include generating, based on the search and the sample data vector, one or more similarity metrics of the sample dataset to individual ones of the reference datasets.
  • Consistent with other disclosed embodiments, non-transitory computer readable storage media may store program instructions, which are executed by at least one processor device and perform any of the methods described herein.
  • The foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the claims.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate several embodiments and, together with the description, serve to explain the disclosed principles. In the drawings:
  • FIG. 1 depicts an exemplary system for searching data, consistent with disclosed embodiments.
  • FIG. 2 depicts an exemplary data-searching system, consistent with disclosed embodiments.
  • FIG. 3 depicts an exemplary data-migration system, consistent with disclosed embodiments.
  • FIG. 4 depicts an exemplary computing environment, consistent with disclosed embodiments.
  • FIG. 5 depicts an exemplary process for profiling data, consistent with disclosed embodiments.
  • FIG. 6 depicts an exemplary process for searching for data, consistent with disclosed embodiments.
  • DESCRIPTION OF THE EMBODIMENTS
  • Consistent with disclosed embodiments, systems and methods to search data sources based on a sample dataset are disclosed. For example, embodiments may include submitting a sample dataset (e.g., a CSV file or other dataset). Disclosed embodiments may include searching one or more databases, computing environments, and/or data indexes based on the sample datasets. Embodiments may include receiving one or more search results. Search results may include a dataset, a data vector, and/or a data source that comprises the dataset. The dataset may be the same as or similar to the sample dataset according to a similarity metric representing a measure of similarity between data in a dataset and data in the sample dataset.
  • The disclosed embodiments may involve profiling data and searching for data based on a data profile that may include a data schema and a statistical description of the data. Disclosed embodiments improve methods of searching remote data resources by generating data vectors based on data profiles and data indices. To achieve these improvements, disclosed embodiments may include one or more memory units for storing instructions and one or more processors configured to execute the instructions to perform operations. The embodiments provide unconventional systems and methods for generating data vectors based on data profiles, storing the data vectors in a data index, and searching the data index.
  • Systems and methods of disclosed embodiments may involve datasets comprising actual data reflecting real-world conditions, events, and/or measurement. However, in some embodiments, disclosed systems and methods may fully or partially involve synthetic data (e.g., anonymized actual data or fake data). Datasets may involve numeric data, text data, and/or image data, For example, datasets may include transaction data, financial data, demographic data, public data, government data, environmental data, traffic data, network data, transcripts of video data, genomic data, proteomic data, and/or other data. Datasets of the embodiments may be in a variety of data formats including, but not limited to, PARQUET, AVRO, SQLITE, POSTGRESQL, MYSQL, ORACLE, HADOOP, CSV, JSON, and/or other data formats.
  • Datasets of disclosed embodiments may have a respective data schema (i.e., structure), including a data type, key-value pair, label, metadata, field, relationship, view, index, package, procedure, function, trigger, sequence, synonym, link, directory, queue, or the like. Datasets of the embodiments may contain foreign keys, i.e., data elements that appear in multiple datasets and may be used to cross-reference data and determine relationships between datasets. Foreign keys may be unique (e.g., a personal identifier) or shared (e.g., a postal code). Datasets of the embodiments may be “clustered,” i.e., a group of datasets may share common features, such as overlapping data, shared statistical properties, etc. Clustered datasets may share hierarchical relationships (i.e., data lineage).
  • Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings and disclosed herein. Wherever convenient, the same reference numbers will be used throughout the drawings to refer to the same or like parts. The disclosed embodiments are described in sufficient detail to enable those skilled in the art to practice the disclosed embodiments. It is to be understood that other embodiments may be utilized and that changes may be made without departing from the scope of the disclosed embodiments. Thus, the materials, methods, and examples are illustrative only and are not intended to be necessarily limiting.
  • FIG. 1 is a diagram of an exemplary system 100 for searching for data, consistent with disclosed embodiments. As shown, system 100 may include a model optimizer 102, a model storage 104, a data-searching system 106, an aggregation database 108, a data-migration system 110, an interface 112, a client device 114, one or more computing environments 116 a, 116 b, 116 n, and one or more databases 118 a, 118 b, 118 n, Components of system 100 may be connected to each other through a network 120.
  • In some embodiments, aspects of system 100 may be implemented on one or more cloud services designed to generate (“spin-up”) one or more ephemeral container instances in response to event triggers, assign one or more tasks to a container instance, and terminate (“spin-down”) a container instance upon completion of a task. By implementing methods using cloud services, disclosed systems may efficiently provision resources based on demand and provide security advantages because the ephemeral container instances may be dosed and destroyed upon completion of a task, That is, the container instances do not permit access from outside using terminals or remote shell tools like SSH, RTP, FTP, or CURL, for example. Further, terminating container instances may include destroying data, thereby protecting sensitive data. Destroying data can provide security advantages because it may involve permanently deleting data (e.g., overwriting data) and associated file pointers.
  • As will be appreciated by one skilled in the art, the components of system 100 can be arranged in various ways and implemented with any suitable combination of hardware, firmware, and/or software, as applicable. For example, as compared to the depiction in FIG. 1 , system 100 may include a larger or smaller number of model optimizers, model storages, data-searching systems, aggregation databases, data-migration systems, interfaces, client devices, computing environments, and/or databases. In addition, system 100 may further include other components or devices not depicted that perform or assist in the performance of one or more processes, consistent with the disclosed embodiments. The exemplary components and arrangements shown in FIG. 1 are not intended to limit the disclosed embodiments.
  • Model optimizer 102 may include one or more computing systems configured to manage training of models for system 100, consistent with disclosed embodiments. Model optimizer 102 may be configured to receive models and/or datasets from other components of system 100 or other components not shown (e.g., via interface 112). Model optimizer 102 may be configured to train models. Model optimizer 102 may be configured to generate models and/or corresponding generalized representations. In some embodiments, model optimizer 102 may be configured to export models to other components of system 100 and/or to external systems or devices.
  • Model optimizer 102 may be configured to generate models based on instructions received from a user or another system (e.g., via interface 112). For example, model optimizer 102 may be configured to receive a visual (graphical) depiction of a machine learning model and parse that graphical depiction into instructions for creating and training a corresponding neural network. Model optimizer 102 may be configured to select model training hyperparameters. This selection may be based on model performance feedback produced by model optimizer 102 during model training and/or received from another component of system 100. Model optimizer 102 may be configured to provide trained models and descriptive information concerning the trained models to model storage 104.
  • Model optimizer 102 may be configured to train machine learning models by optimizing model parameters and/or hyperparameters (hyperparameter tuning) using an optimization technique, consistent with disclosed embodiments. The optimization technique may include a grid search, a random search, a gaussian process, a Bayesian process, a Covariance Matrix Adaptation Evolution Strategy (CMA-ES), a derivative-based search, a stochastic hill-climb, a neighborhood search, an adaptive random search, or the like. Tuning a hyperparameter may include iteratively selecting a hyperparameter and training model parameters using the hyperparameter until a training criterion is satisfied, consistent with disclosed embodiments.
  • Model storage 104 may be configured to store models, including machine learning models, consistent with disclosed embodiments. In some embodiments, some or all components of model storage 104 may be hosted on one or more servers, one or more clusters of servers, or one or more cloud services. Model storage 104 may be connected to network 120 and may additionally or alternatively be directly connected to model optimizer 102 (not shown). In some embodiments, model storage 104 may be a component of model optimizer 102, data-searching system 106, or client device 114 (not shown).
  • Model storage 104 may include one or more databases configured to store data models (e.g., machine learning models or statistical models) and descriptive information of the data models. Model storage 104 may be configured to provide information regarding available data models to a user or another system. The databases may include cloud-based databases, cloud-based buckets, or on-premises databases. The information may include model information, such as the type and/or purpose of the model and any measures of classification error. Model storage 104 may include one or more databases configured to store indexed and clustered models for use by system 100. For example, model storage 104 may store models associated with generalized representations of those models (e.g., neural network architectures stored in TENSORFLOW or other standardized formats). The databases may include cloud-based databases (e.g., AMAZON WEB SERVICES S3 buckets) or on-premises databases.
  • Data-searching system 106 may include a computing device, a computer, a server, a server duster, a plurality of dusters, and/or a cloud service, consistent with disclosed embodiments. Data-searching system 106 may include one or more memory units and one or more processors configured to perform operations consistent with disclosed embodiments. Data-searching system 106 may include computing systems configured to manage training of models for system 100, profile datasets, index datasets, and perform other methods consistent with disclosed embodiments. Data-searching system 106 may be configured to receive and/or retrieve models, and/or datasets from other components of system 100 or from computing components outside system 100 (e.g., via interface 112). In some embodiments not shown, data-searching system 106 may be a component of data-migration system 110. Data-searching system 106 is disclosed in greater detail below (in reference to FIG. 2 ).
  • Aggregation database 108 may include one or more databases configured to store data for use by system 100, consistent with disclosed embodiments. The databases may include cloud-based databases (e.g., AMAZON WEB SERVICES S3 buckets) or on-premises databases. In some embodiments, aggregation database 108 may store one or more datasets, data vectors, and data indexes. For example, aggregation database 108 may be configured to store one or more data vectors comprising statistical metrics of a dataset and information based on a corresponding data schema of the dataset. Aggregation database 108 may include data received from other components of system 100, including, for example, data-migration system 110 and/or data-searching system 106.
  • Data-migration system 110 may include a computing device, a computer, a server, a server cluster, a plurality of clusters, and/or a cloud service, consistent with disclosed embodiments. Data-migration system 110 may include one or more memory units and one or more processors configured to perform operations consistent with disclosed embodiments. Data-migration system 110 may include computing systems configured to manage movement (i.e., transfer) of data between components of system 100, to manage movement of data between components of system 100 and computing components outside system 100, and to perform other methods consistent with disclosed embodiments. Data-migration system 110 may be configured to receive and/or retrieve models, and/or datasets from other components of system 100 or from computing components outside system 100 (e.g., via interface 112). In some embodiments not shown, data-migration system 110 may be a component of data-searching system 106. Data-migration system 110 is disclosed in greater detail below (in reference to FIG. 3 ).
  • Interface 112 may be configured to manage interactions between system 100 and other systems using network 120, consistent with disclosed embodiments. In some aspects, interface 112 may be configured to publish data received from other components of system 100. This data may be published in a publication and subscription framework (e.g., using APACHE KAFKA), through a network socket, in response to queries from other systems, or using other known methods. The data may be synthetic data, as described herein. As an additional example, interface 112 may be configured to provide information received from model storage 104 regarding available datasets. In various aspects, interface 112 may be configured to provide data or instructions received from other systems to components of system 100. For example, interface 112 may be configured to receive instructions for generating data models (e.g., type of data model, data model parameters, training data indicators, training hyperparameters, or the like) from another system and provide this information to model optimizer 102. As an additional example, interface 112 may be configured to receive data including sensitive portions from another system (e.g., in a file, a message in a publication and subscription framework, a network socket, or the like) and provide that components of system 100.
  • Client device 114 may include one or more memory units and one or more processors configured to perform operations consistent with disclosed embodiments. In some embodiments, client device 114 may include hardware, software, and/or firmware modules. Client device 114 may be a terminal, a kiosk, a mobile device, a tablet, a personal computer, a server, a server cluster, a cloud service, a storage device, a specialized device configured to perform methods according to disclosed embodiments, or the like.
  • Computing environments 116 a, 116 b, 116 n may include one or more memory units and one or more processors configured to perform operations consistent with disclosed embodiments. In some embodiments, computing environments 116 a, 116 b, 116 n may be or include a virtual private cloud, a local computing environment (e.g., a personal computer, terminal, kiosk, or workstation), a server, a server cluster, a device (e.g., a mobile device, a tablet, a laptop, etc.), and/or another computing environment. In some embodiments, at least one computing environment may connect to system 100 via interface 112 (not shown in FIG. 1 ). One or more computing environments 116 a, 116 b, 116 n may be unconnected to other computing environments except via network 120 (i.e., they may be disparate computing environments). In some embodiments, one or more computing environments may be directly connected to each other (not shown in FIG. 1 ). As noted above, computing environments 116 a, 116 b, 116 n are exemplary only and system 100 may include additional or fewer computing environments. Computing environments 116 a, 116 b, 116 n are disclosed in greater detail below (in reference to FIG. 4 ).
  • Databases 118 a, 118 b, 118 n may include one or more databases configured to store data for use by system 100, consistent with disclosed embodiments. The databases may include cloud-based databases (e.g., AMAZON WEB SERVICES 53 buckets) or on-premises databases, In some embodiments, one or more databases 118 a, 118 b, 118 n may store one or more datasets, data vectors, and data indexes. For example, at least one of databases 118 a, 118 b, 118 n may be configured to store one or more data vectors comprising statistical metrics of a dataset and information based on a corresponding data schema of the dataset. As noted above, databases 118 a, 118 b, 118 n are exemplary only and system 100 may include additional or fewer databases.
  • As shown in FIG. 1 , at least one of model optimizer 102, model storage 104, data-searching system 106, aggregation database 108, data-migration system 110, interface 112, client device 114, computing environments 116 a, 116 b, 116 n, and/or databases 118 a, 118 b, 118 n may connect to network 120. Network 120 may be a public network or private network and may include, for example, a wired or wireless network, including, without limitation, a Local Area Network, a Wide Area Network, a Metropolitan Area Network, an IEEE 1002.11 wireless network (e.g., Fr), a network of networks (e.g., the Internet), a land-line telephone network, or the like. Network 120 may be connected to other networks (not depicted in FIG. 1 ) to connect the various system components to each other and/or to external systems or devices. In some embodiments, network 120 may be a secure network and require a password to access the network.
  • FIG. 2 depicts exemplary data-searching system 106, consistent with disclosed embodiments. Data-searching system 106 may include a computing device, a computer, a server, a server cluster, a plurality of clusters, and/or a cloud service, consistent with disclosed embodiments. As shown, data-searching system 106 may include one or more processors 210, one or more I/O devices 220, and one or more memory units 230. In some embodiments, some or all components of data-searching system 106 may be hosted on a device, a computer, a server, a cluster of servers, or a cloud service. In some embodiments, data-searching system 106 may be a scalable system configured to efficiently manage resources and enhance security by provisioning computing resources in response to triggering events and terminating resources after completing a task (e.g., a scalable cloud service that spins up and terminates container instances).
  • FIG. 2 depicts an exemplary configuration of data-searching system 106. As will be appreciated by one skilled in the art, the components and arrangement of components included in data-searching system 106 may vary, For example, as compared to the depiction in FIG. 2 , data-searching system 106 may include a larger or smaller number of processors, I/O devices, or memory units. In addition, data-searching system 106 may further include other components or devices not depicted that perform or assist in the performance of one or more processes consistent with the disclosed embodiments. The components and arrangements shown in FIG. 2 are not intended to limit the disclosed embodiments, as the components used to implement the disclosed processes and features may vary.
  • Processor 210 may comprise known computing processors, including a microprocessor. Processor 210 may constitute a single-core or multiple-core processor that executes parallel processes simultaneously. For example, processor 210 may be a single-core processor configured with virtual processing technologies. In some embodiments, processor 210 may use logical processors to simultaneously execute and control multiple processes. Processor 210 may implement virtual machine technologies, or other known technologies to provide the ability to execute, control, run, manipulate, store, etc., multiple software processes, applications, programs, etc. In another embodiment, processor 210 may include a multiple-core processor arrangement (e.g., dual core, quad core, etc.) configured to provide parallel processing functionalities to allow execution of multiple processes simultaneously. One of ordinary skill in the art would understand that other types of processor arrangements could be implemented that provide for the capabilities disclosed herein. The disclosed embodiments are not limited to any type of processor(s) 210. Processor 210 may execute various instructions stored in memory 230 to perform various functions of the disclosed embodiments described in greater detail below. Processor 210 is configured to execute functions written in one or more known programming languages. I/O devices 220 may include at least one of a display, an LED, a router, a touchscreen, a keyboard, a microphone, a speaker, a haptic device, a camera, a button, a dial, a switch, a knob, a transceiver, an input device, an output device, or another I/O device to perform methods of the disclosed embodiments. I/O devices 220 may be components of an interface of data-searching system 106 (e.g., an interface such as interface 112).
  • Memory 230 may be a volatile or non-volatile, magnetic, semiconductor, optical, removable, non-removable, or other type of storage device or tangible (i.e., non-transitory) computer-readable medium, consistent with disclosed embodiments. As shown, memory 230 may include data 231, including of at least one of encrypted data or unencrypted data. Data 231 may include one or more model indexes, model parameters, model hyperparameters, model codes, dataset indexes, data vectors, and/or datasets, consistent with disclosed embodiments.
  • Programs 235 may include one or more programs (e.g., modules, code, scripts, or functions) used to perform methods consistent with disclosed embodiments. Programs may include operating systems (not shown) that perform known operating system functions when executed by one or more processors. Disclosed embodiments may operate and function with computer systems running any type of operating system. Programs 235 may be written in one or more programming or scripting languages, One or more of such software sections or modules of memory 230 may be integrated into a computer system, non-transitory computer-readable media, or existing communications software. Programs 235 may also be implemented or replicated as firmware or circuit logic.
  • Programs 235 may include a model-training module 236, data-profiling module 237, aggregator 238, searching module 239 and/or other modules not depicted to perform methods of the disclosed embodiments. In some embodiments, modules of programs 235 may be configured to generate (“spin up”) one or more ephemeral container instances to perform a task and/or to assign a task to a running (warm) container instance, consistent with disclosed embodiments. Modules of programs 235 may be configured to receive, retrieve, and/or generate models, consistent with disclosed embodiments. Modules of programs 235 may be configured to receive, retrieve, and/or generate datasets (e.g., to generate synthetic datasets, data samples, or other datasets), consistent with disclosed embodiments. Modules of programs 235 may be configured to perform operations in coordination with one another.
  • Model-training module 236 may be configured to train one or more models and/or perform hyperparameter tuning of one or more models, including machine learning models. For example, model-training module 236 may be configured to receive input of one or more thresholds, one or more loss functions, and/or one or more limits on a number of interactions and apply the input for optimizing a received model and/or corresponding generalized representation such as a neural network. In some embodiments, training of a model terminates when a training criterion is satisfied. In some embodiments, model-training module 236 is configured to adjust model parameters during training. The model parameters may include weights, coefficients, offsets, or the like. Training may be supervised or unsupervised.
  • Model-training module 236 may be configured to select or generate model parameters (e.g., number of layers for a neural network, kernel function for a kernel density estimator, or the like), update training hyperparameters, and evaluate model characteristics. For example, model characteristics may include a model type (e.g., an INN, a convolutional neural network (CNN), a random forest, or another model type), a model parameter, a model hyperparameter (including training a hyperparameter and/or an architectural hyperparameter), a desired outcome, belongingness to a model cluster, and/or belonginess of a model training dataset to a dataset duster, the similarity of synthetic data generated by a model to actual data, or other characteristics,
  • In some embodiments, model-training module 236 may be configured to perform a search of a hyperparameter space and select new hyperparameters. This search may or may not depend on the values of a performance metric obtained for other trained models. In some aspects, model-training module 236 may be configured to perform a grid search or a random search. The hyperparameters may include training hyperparameters, which may affect how training of the model occurs, or architectural hyperparameters, which may affect the structure of the model.
  • Consistent with disclosed embodiments, hyperparameters may include training hyperparameters such as learning rate, batch size, number of training batches, number of epochs, chunk size, time window, input noise dimension, or the like. Hyperparameters may include architectural parameters such as number of layers in a neural network, the choice of activation function for a neural network node, the layers in a CNN or the like. For example, when the stored model comprises a generative adversarial network (GAN), training hyperparameters for the model may include a weight for a loss function penalty term that penalizes the generation of training data according to a similarity metric. As a further example, when the stored model comprises a neural network, the training hyperparameters may include a learning rate for the neural network. As an additional example, when the model is a CNN, architectural hyperparameters may include the number and type of layers in the convolutional neural network.
  • In some embodiments, model-training module 236 may include programs to apply one or more templates to a data model (e.g., a model retrieved from model storage 104) and apply the templates to generate a generalized representation of the retrieved model (e.g., a neural network). Model-training module 236 may include programs to provide trained generalized representations to model storage 104 for storing in association with corresponding models.
  • Data-profiling module 237 may be configured to profile a dataset, consistent with disclosed embodiments. A data profile may include information related to a data schema (i.e., the organization or structure of a dataset), including a data type, key-value pair, a label, metadata, a field, a relationship, a view, an index, a package, a procedure, a function, a trigger, a sequence, a synonym, a link, a directory, a queue, or the like. A data profile may include a statistical profile of a dataset that includes a plurality of descriptive (statistical) metrics of the dataset. For example, the statistical metrics may include an average, a mean, a standard deviation, a range, a moment, a variance, a covariance, a covariance matrix or any other statistical metric of the selected dataset.
  • In some embodiments a statistical metric may be a similarity metric representing a measure of similarity between data in a dataset. The similarity metric may be based on a covariance matrix, a variance, a frequency of overlapping values, or other measure of statistical similarity. In some embodiments, data-profiling module 237 implements a machine learning model (e.g., a data-profiling model) to generate the similarity metric
  • Data-profiling module 237 may include, may be configured generate, and/or may be configured to implement one or more data-profiling models. A data-profiling model may include machine-learning models and statistical models to determine the data schema and/or a statistical profile of a dataset (i.e., to profile a dataset), consistent with disclosed embodiments. The data-profiling model may include an RNN model, a CNN model, a variational autoencoder (VAE) model, an autoencoder model, or another machine-learning model. The data-profiling model may include algorithms to determine a data type, key-value pairs, row-column data structure, statistical distributions of information such as keys or values, or other property of a data schema. The data-profiling model may be configured to implement univariate and multivariate statistical methods. The data-profiling model may include a regression model, a Bayesian model, a statistical model, a linear discriminant analysis model, or other classification model configured to determine one or more descriptive metrics of a dataset. For example, the data-profiling model may include algorithms to determine an average, a mean, a standard deviation, a quantile, a quartile, a probability distribution function, a range, a moment, a variance, a covariance, a covariance matrix, a dimension and/or dimensional relationship (e.g., as produced by dimensional analysis such as length, time, mass, etc.) or any other descriptive metric of a dataset.
  • In some embodiments, the data-profiling model may be trained to identify complex data types, i.e. a datatype whose structure may be based on an underlying rule (e.g., a social security number, a longitude, a zip code, an account number, an international telephone number, an address, a license plate number, a driver's license number, a government identification number, a passport number, or the like). In some embodiments, the data-profiling model may be trained to identify a complex variable as an address and may be trained to further identify complex variables as including United States addresses and/or Japanese addresses. The data-profiling model may be configured to tag identified variables. In some embodiments, the data-profiling model may be configured to perform embedding (e.g., feature embedding).
  • Data-profiling module 237 may be configured to generate a data vector based on a data profile, consistent with disclosed embodiments. A data vector may include one or more statistical metrics of a corresponding reference datasets or of a variable (e.g., a data column) of the reference dataset. A data vector may include information based on a corresponding data schema of the reference dataset. For example, the data vector may include data relating to a data schema including a data type, key-value pair, a label, metadata, a field, a relationship, a view, an index, a package, a procedure, a function, a trigger, a sequence, a synonym, a link, a directory, a queue, or the like. In some embodiments, data-profile module 237 may be configured to generate a plurality of data vectors based on corresponding data profiles. Data-profiling module 237 may be configured to generate metadata associated with a data vector. For example, data-profiling module 237 may generate a tag, a label, a unique identifier, a title, a data class name (e.g., “health data”, “financial data”, “environmental data”, etc.), a data format, or other metadata. Data-profiling module 237 may be configured to perform calculations on the dataset (e.g., converting measurement units, normalizing data values, classifying data values into a category, or other calculations).
  • Data-profiling module 237 may be configured to aggregate and/or store the plurality of data vectors in a data storage, consistent with disclosed embodiments. For example, the plurality of data vectors may be stored as a data index, a matrix, in a relational database, or in another format. In some embodiments, the data vector may include statistical metrics of a plurality of data vectors based on a plurality of corresponding data profiles, consistent with disclosed embodiments. In some embodiments, the data vector may be a subset of a data profile. In some embodiments, the data vector may be the data profile. In some embodiments, the data vector may include a statistical metric of the data profile (e.g., a mean of a plurality of means).
  • Aggregator 238 may include programs (scripts, functions, algorithms) to aggregate datasets and/or data indexes, consistent with disclosed embodiments. In some embodiments, aggregator 238 may be configured to retrieve and/or receive data (e.g., a dataset, a data vector, a data index, a model, and/or a model index), transmit data, and/or store data in a data storage. In some embodiments, aggregator 238 may be configured to transmit data to, store data in, receive data from, and/or retrieve data from other components of system 100 (e.g., data 231; computing environments 116 a, 116 b, 116 n; databases 118 a, 118 b, 118 n; aggregation database 108; model storage 104; data-migration system 110; or other components). In some embodiments, aggregator 238 may be configured to transmit data to, store data in, receive data from, and/or retrieve data from a computing component outside system 100 (e.g., via interface 112). In some embodiments, aggregator 238 may receive data from or transmit data to other modules of data-searching system 106. In some embodiments, aggregator 238 may be configured to transmit, store, receive, and/or retrieve data according to at least one of a predetermined schedule or a triggering event. A triggering event may include receiving a notification, receiving a dataset, receiving a request, and/or another event.
  • Searching module 239 may include programs (scripts, functions, algorithms, routines, or the like) to conduct a search, consistent with disclosed embodiments. In some embodiments, searching module 239 may include programs to generate and maintain searchable data indexes as data trees based on the data vectors (e.g., a B-Tree). In some embodiments, searching module 239 may include programs based on POSTGRESQL or other object-relational database management system. For example, searching module 239 may be configured to generate a data index (e.g., a RUM-index and/or GIN-index and/or Locality-Sensitive Hashing). The data index may be configured to be searched (e.g., as a data tree). The data index may be based on the data vectors and/or data profiles and organized for efficient searching with or without using k-nearest neighbors clustering-based methods.
  • Searching module 239 may be configured to execute a search of the data index. In some embodiments, searching module 239 may be configured to receive search parameters such as the types of data to search or other information related to data types and schemas to guide the search of the data index. For example, the search parameters may indicate that the search will weigh the data schema of a dataset more than the particular data values or statistical metrics of the dataset. Search parameters may be received from another component of system 100 or a computing component outside system 100 (e.g., via interface 112). In some embodiments, the search parameters may include instructions to search a data index by first comparing data schemas then by comparing statistical metrics. The search parameters may include instructions to search the data profile (i.e., schema and statistical metrics) of a dataset then to search variables (e.g., data columns) of a dataset. In some embodiments, searching module 239 may be configured to generate search parameters using a machine learning model trained to use learned features of data to improve search accuracy.
  • FIG. 3 depicts an exemplary data-migration system 110, consistent with disclosed embodiments. Data-migration system 110 may include computing systems configured to manage movement of data between components of system 100 and/or to manage movement of data between components of system 100 and computing components outside system 100. Data-migration system 110 may include a computing device, a computer, a server, a server duster, a plurality of dusters, and/or a cloud service, consistent with disclosed embodiments. As shown, data-migration system 110 includes one or more processors 310, one or more I/O devices 320, and one or more memory units 330. In some embodiments, some or all components of data-migration system 110 may be hosted on a device, a computer, a server, a duster of servers, or a cloud service. In some embodiments, data-migration system 110 is a scalable system configured to efficiently manage resources and enhance security by provisioning computing resources in response to triggering events and terminating resources after completing a task (e.g., a scalable cloud service that spins up and terminates container instances).
  • FIG. 3 depicts an exemplary configuration of data-migration system 110. As will be appreciated by one skilled in the art, the components and arrangement of components included in data-migration system 110 may vary. For example, as compared to the depiction in FIG. 3 , data-migration system 110 may include a larger or smaller number of processors 310, I/O devices 320, or memory units 330. In additon, data-migration system 110 may further include other components or devices not depicted that perform or assist in the performance of one or more processes consistent with the disclosed embodiments. The components and arrangements shown in FIG. 3 are not intended to limit the disclosed embodiments, as the components used to impler ent the disclosed processes and features may vary.
  • Processor 310 may comprise known computing processors, including a microprocessor. Processor 310 may constitute a single-core or multiple-core processor that executes parallel processes simultaneously. For example, processor 310 may be a single-core processor configured with virtual processing technologies. In some embodiments, processor 310 may use logical processors to simultaneously execute and control multiple processes. Processor 310 may implement virtual machine technologies, or other known technologies to provide the ability to execute, control, run, manipulate, store, etc., multiple software processes, applications, programs, etc. In another embodiment, processor 310 may include a multiple-core processor arrangement (e.g., dual core, quad core, etc.) configured to provide parallel processing functionalities to allow execution of multiple processes simultaneously. One of ordinary skill in the art would understand that other types of processor arrangements could be implemented that provide for the capabilities disclosed herein. The disclosed embodiments are not limited to any type of processor(s) 310. Processor 310 may execute various instructions stored in memory 330 to perform various functions of the disclosed embodiments described in greater detail below. Processor 310 is configured to execute functions written in one or more known programming languages.
  • I/O devices 320 may include at least one of a display, an LED, a router, a touchscreen, a keyboard, a microphone, a speaker, a haptic device, a camera, a button, a dial, a switch, a knob, a transceiver, an input device, an output device, or another I/O device to perform methods of the disclosed embodiments. I/O devices 320 may be components of an interface of data-migration system 110.
  • Memory 330 may be a volatile or non-volatile, magnetic, semiconductor, optical, removable, non-removable, or other type of storage device or tangible (i.e., non-transitory) computer-readable medium, consistent with disclosed embodiments. As shown, memory 330 may include data 331, including of at least one of encrypted data or unencrypted data, Data 331 may include one or more model indexes, model parameters, model hyperparameters, model codes, data indexes, data vectors, and/or datasets, consistent with disclosed embodiments.
  • Programs 335 may include one or more programs (e.g., modules, code, scripts, or functions) used to perform methods consistent with disclosed embodiments.
  • Programs may include operating systems (not shown) that perform known operating system functions when executed by one or more processors. Disclosed embodiments may operate and function with computer systems running any type of operating system. Programs 335 may be written in one or more programming or scripting languages. One or more of such software sections or modules of memory 330 may be integrated into a computer system, non-transitory computer-readable media, or existing communications software. Programs 335 may also be implemented or replicated as firmware or circuit logic.
  • In some embodiments, modules of programs 335 may be configured to generate (“spin up”) one or more ephemeral container instances to perform a task and/or to assign a task to a running (warm) container instance, consistent with disclosed embodiments. Modules of programs 335 may be configured to receive, retrieve, and/or generate models, consistent with disclosed embodiments. Modules of programs 335 may be configured to receive, retrieve, and/or generate datasets (e.g., to generate synthetic datasets, data samples, or other datasets), consistent with disclosed embodiments. Modules of programs 335 may be configured to perform operations in coordination with one another.
  • As shown, programs 335 may include a model-training module 336, data-profiling module 337, aggregator 338, migration module 339 and/or other modules not depicted to perform methods of the disclosed embodiments. Model-training module 336 may include, may be configured generate, and/or may be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to model-training module 236, above. Data-profiling module 337 may include, may be configured generate, and/or may be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to data-profiling module 237, above. Aggregator 338 may include, may be configured generate, and/or may be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to model-training module 236, above,
  • Migration module 339 may be configured to receive and/or retrieve data from one computing component and transmit the data to another component. The computing components may be part of system 100 or outside system 100, Migration module 339 may be configured to generate and maintain logs or other records related to data transfers. In some embodiments, migration module 339 may be configured to generate and/or maintain logs or other records of data transfers that other components of system 100 perform. For example, migration module 339 may receive a notification that includes a statement of a data transfer to or from a component of system 100 and generate or update a log based on the notification. Migration module 339 may aggregate logs or other records of data transfer that are generated by other components of system 100. For example, migration module 339 may receive a log from a component of system 100 and update an aggregated log file. Updating a log file or aggregated log file may include identifying and removing duplicate records.
  • Data-migration system 110 may manage some or all dataset transfers to and from components of system 100. For example, in some embodiments, some or all datasets that enter system 100 may be received at data-migration system 110 prior to being transferred to another component of system 100. In some embodiments, some or all datasets that exits system 100 may be received at data-migration system 110 prior to being transferred to a computing component outside system 100. In some embodiments, some or all datasets transferred from a component of system 100 may be received at data-migration system 110 before being transferred to another component of system 100. In some embodiments, data-migration system 110 may maintain a log or record of some or all dataset transfers between components of system 100. In some embodiments, data-migration system 110 may maintain a log or record of some or all dataset transfers between components of system 100 and computing components outside system 100.
  • In some embodiments, data-migration system 110 may be configured to receive a request to transfer data. Data-migration system 110 may be configured to transfer data based on the request. The request may include a statement indicating a data source and/or a desired data destination.
  • FIG. 4 depicts an exemplary computing environment 116, consistent with disclosed embodiments. Computing environment 116 may be one of, for example, computing environments 116 a, 116 b, 116 n. Computing environment 116 may be or include a virtual private cloud, a local computing environment (e.g., a personal computer, terminal, kiosk, or workstation), a server, a server cluster, a device (e.g., a mobile device, a tablet, a laptop, etc.), and/or another computing environment.
  • As shown, computing environment 116 includes one or more processors 410, one or more I/O devices 420, and one or more memory units 430. In some embodiments, some or all components of computing environment 116 may be hosted on a device, a computer, a server, a duster of servers, or a cloud service. In some embodiments, computing environment 116 is a scalable system configured to efficiently manage resources and enhance security by provisioning computing resources in response to triggering events and terminating resources after completing a task (e.g., a scalable cloud service that spins up and terminates container instances).
  • FIG. 4 depicts an exemplary configuration of computing environment 116. As will be appreciated by one skilled in the art, the components and arrangement of components included in computing environment 116 may vary. For example, as compared to the depiction in FIG. 4 , computing environment 116 may include a larger or smaller number of processors 410, I/O devices 420, or memory units 430. In addition, computing environment 116 may further include other components or devices not depicted that perform or assist in the performance of one or more processes consistent with the disclosed embodiments. The components and arrangements shown in FIG. 4 are not intended to limit the disclosed embodiments, as the components used to implement the disclosed processes and features may vary.
  • Processor 410 may comprise known computing processors, including a microprocessor. Processor 410 may constitute a single-core or multiple-core processor that executes parallel processes simultaneously. For example, processor 410 may be a single-core processor configured with virtual processing technologies. In some embodiments, processor 410 may use logical processors to simultaneously execute and control multiple processes. Processor 410 may implement virtual machine technologies, or other known technologies to provide the ability to execute, control, run, manipulate, store, etc., multiple software processes, applications, programs, etc. In another embodiment, processor 410 may include a multiple-core processor arrangement (e.g., dual core, quad core, etc.) configured to provide parallel processing functionalities to allow execution of multiple processes simultaneously. One of ordinary skill in the art would understand that other types of processor arrangements could be implemented that provide for the capabilities disclosed herein. The disclosed embodiments are not limited to any type of processor(s) 410. Processor 410 may execute various instructions stored in memory 330 to perform various functions of the disclosed embodiments described in greater detail below. Processor 410 is configured to execute functions written in one or more known programming languages.
  • I/O devices 420 may include at least one of a display, an LED, a router, a touchscreen, a keyboard, a microphone, a speaker, a haptic device, a camera, a button, a dial, a switch, a knob, a transceiver, an input device, an output device, or another UO device to perform methods of the disclosed embodiments. I/O devices 420 may be components of an interface of computing environment 116.
  • Memory 430 may be a volatile or non-volatile, magnetic, semiconductor, optical, removable, non-removable, or other type of storage device or tangible (i.e., non-transitory) computer-readable medium, consistent with disclosed embodiments. As shown, memory 430 may include data 431, including of at least one of encrypted data or unencrypted data, Data 431 may include one or more model indexes, model parameters, model hyperparameters, model codes, data indexes, data vectors, and/or datasets, consistent with disclosed embodiments.
  • Programs 435 may include one or more programs (e.g., modules, code, scripts, or functions) used to perform methods consistent with disclosed embodiments. Programs may include operating systems (not shown) that perform known operating system functions when executed by one or more processors. Disclosed embodiments may operate and function with computer systems running any type of operating system. Programs 435 may be written in one or more programming or scripting languages. One or more of such software sections or modules of memory 430 may be integrated into a computer system, non-transitory computer-readable media, or existing communications software. Programs 435 may also be implemented or replicated as firmware or circuit logic.
  • In some embodiments, modules of programs 435 may be configured to generate (“spin up”) one or more ephemeral container instances to perform a task and/or to assign a task to a running (warm) container instance, consistent with disclosed embodiments. Modules of programs 435 may be configured to receive, retrieve, and/or generate models, consistent with disclosed embodiments. Modules of programs 435 may be configured to receive, retrieve, and/or generate datasets (e.g., to generate synthetic datasets, data samples, or other datasets), consistent with disclosed embodiments. Modules of programs 435 may be configured to perform operations in coordination with one another.
  • As shown, programs 435 may include a model-training module 436, data-profiling module 437, and/or other modules not depicted to perform methods of the disclosed embodiments. Model-training module 436 may include, may be configured generate, and/or ay be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to model-training module 236, above. Data-profiling module 437 may include, may be configured generate, and/or may be configured to implement any of the models, algorithms, functions, programs, scripts and/or other computing code as disclosed in reference to data-profiling module 237, above.
  • FIG. 5 depicts an exemplary process for profiling data 500, consistent with disclosed embodiments. In some embodiments, data-migration system 110 may perform process 500 using programs 335. One or more of model-training module 336, data-profiling module 337, aggregator 338, migration module 339, or other module of programs 335 may perform operations of process 500, consistent with disclosed embodiments. In some embodiments, data-searching system 106 may perform steps of process 500. One or more of model-training module 236, data-profiling module 237, aggregator 238, searching module 239, or other module of programs 235 may perform operations of process 500, consistent with disclosed embodiments. It should be noted that other components of system 100, including, for example, model optimizer 102, client device 114, and/or one or more computing environments (e.g., computing environment 116 a, 116 b, 116 n) may perform operations of one or more steps of process 500.
  • Consistent with disclosed embodiments, steps of process 500 may be performed on one or more cloud services using one or more ephemeral container instances. For example, at any of the steps of process 500, data-migration system 110 or data-searching system 106 may generate (spin up) an ephemeral container instance to execute a task, assign a task to an already-running ephemeral container instance (warm container instance), or terminate a container instance upon completion of a task. As one of skill in the art will appreciate, steps of process 500 may be performed as part of an application interface (API) call.
  • At step 502, data-migration system 110 may receive one or more reference datasets, consistent with disclosed embodiments. In some embodiments, receiving a reference dataset includes retrieving the reference dataset from another computing component. In some embodiments, the reference datasets may be received from client device 114, computing environments 116 a, 116 b, 116 n, databases 118 a, 118 b, 118 n, and/or an external computing component (e.g., via interface 112). In some embodiments, the reference datasets may be received based on a schedule.
  • In some embodiments, receiving a reference dataset at step 502 includes a receiving a request. In some embodiments, the request includes a command to retrieve one or more reference datasets from another component of system 100. In some embodiments, the request includes a command to retrieve a data vector and/or a data index. In some embodiments, a data vector and/or data index is received at step 502, In some embodiments, data-migration system 110 generates or updates a log file based on the received or retrieved datasets, consistent with disclosed embodiments.
  • In some embodiments, at step 502, data-migration system transmits a dataset request to a computing component of system 100 (e.g., a computing environment 116 a, 116 b, 116 n or outside system 100) and receives a dataset in response to the request, consistent with disclosed embodiments. The dataset request may be transmitted according to a schedule. The dataset request may be transmitted based on a triggering event, such as receiving a user request via an interface
  • At step 504, data-migration system 110 may identify one or more data schemas corresponding to individual ones of the one or more reference datasets, consistent with disclosed embodiments. Step 504 may include using one or more data-profiling model to identify the data schemas. The data-profiling model may be stored in, for example, model storage 104, data 331, or another data storage.
  • At step 506, data-migration system 110 may generate data vectors based on the one or more data schemas, consistent with disclosed embodiments. As previously described, the data vectors may include statistical metrics of the sample dataset and information based on the data schema of the sample dataset. Step 506 may include using one or more data-profiling models to generate data vectors. The data-profiling model may be stored in, for example, model storage 104, data 331, or another data storage. Step 506 may include generating metadata (e.g., one or more tags or labels) associated with the data vectors. Step 506 may include performing calculations on the dataset (e.g., converting measurement units, normalizing data values, classifying data values into a category, or other calculations) as before generating the data vectors.
  • At step 508, data-migration system 110 may generate a data index based on the data vectors, consistent with disclosed embodiments. The data index may include the data vectors and metadata associated with the data vectors. The data index may be a searchable data index, consistent with disclosed embodiments. For example, the data index may be stored as a tree.
  • At step 510, data-migration system 110 may store the data index, consistent with disclosed embodiments. Storing the data index may include storing the data index in aggregation database 108 and/or one or more data storages (e.g., data 231, data 331, or data 431). Storing the data index may include storing the data vectors. Storing the data index may include generating and/or updating a log based on the data index and storing the log.
  • It should be noted that steps of process 500 may be repeated. For example, data-migration system 110 may receive an additional dataset and/or an additional data vector and repeat steps of process 500 to update a data index,
  • FIG, 6 depicts an exemplary process for searching for data 600, consistent with disclosed embodiments. In some embodiments, data-searching system 106 performs process 600. One or more of model-training module 236, data-profiling module 237, aggregator 238, searching module 239, or other module of programs 235 may perform operations of process 600, consistent with disclosed embodiments. It should be noted that other components of system 100, including, for example, model optimizer 102, client device 114, data-migration system 110, and/or one or more computing environments (e.g., computing environment 116 a, 116 b, 116 n) may perform operations of one or more steps of process 600.
  • Consistent with disclosed embodiments, steps of process 600 may be performed on one or more cloud services using one or more ephemeral container instances. For example, at any of the steps of process 600, data-migration system 110 or data-searching system 106 may generate (spin up) an ephemeral container instance to execute a task, assign a task to an already-running ephemeral container instance (warm container instance), or terminate a container instance upon completion of a task. As one of skill in the art will appreciate, steps of process 600 may be performed as part of an application interface (API) call.
  • At step 602, data-searching system 106 may receive a sample dataset, consistent with disclosed embodiments. In some embodiments, receiving a sample dataset includes retrieving the reference sample from another computing component. In some embodiments, the sample dataset is received from data-migration system 110, client device 114, computing environments 116 a, 116 b, 116 n, databases 118 a, 118 b, 118 n, and/or an external computing component (e.g., via interface 112). In some embodiments, receiving a sample dataset at step 602 includes a receiving a request. In some embodiments, the request includes a command to retrieve the sample dataset from another component of system 100. In some embodiments, the request includes a command to retrieve one or more reference datasets from another component of system 100. In some embodiments, the request includes a command to retrieve a data vector and/or a data index. In some embodiments, a data vector and/or data index is received at step 602.
  • At step 604, data-searching system 106 may identify a data schema of the sample dataset, consistent with disclosed embodiments. Step 604 may include using one or more data-profiling model to identify the data schema. The data-profiling models may be stored in, for example, model storage 104, data 231, or another data storage. As previously described, data-profiling model may include machine-learning models and statistical models to determine the data schema and/or a statistical profile of a dataset (i.e., to profile a dataset). The data-profiling model may include an RNN model, a CNN model, or other machine-learning model. The data-profiling model may include algorithms to determine a data type, key-value pairs, row-column data structure, statistical distributions of information such as keys or values, or other property of a data schema. The data-profiling model may be configured to implement univariate and multivariate statistical methods. The data-profiling model may include a regression model, a Bayesian model, a statistical model, a linear discriminant analysis model, or other classification model configured to determine one or more descriptive metrics of a dataset. For example, the data-profiling model may include algorithms to determine an average, a mean, a standard deviation, a quantile, a quartile, a probability distribution function, a range, a moment, a variance, a covariance, a covariance matrix, a dimension and/or dimensional relationship (e.g., as produced by dimensional analysis such as length, time, mass, etc.) or any other descriptive metric of a dataset. In some embodiments, the data-profiling model may be trained to identify complex data types, i.e. a datatype whose structure may be based on an underlying rule (e.g., a social security number, a longitude, a zip code, an account number, an international telephone number, an address, a license plate number, a driver's license number, a government identification number, a passport number, or the like)
  • At step 606, data-searching system 106 may generate a sample data vector based on the data schema, consistent with disclosed embodiments. As previously described, the sample data vectors may include statistical metrics of the sample dataset and information based on the data schema of the sample dataset. Step 606 may include using one or more data-profiling models to generate the sample data vector. The data-profiling model may be stored in, for example, model storage 104, data 231, or another data storage. Step 606 may include generating metadata (e.g., one or more tags or labels) associated with the sample data vector. Step 606 may include performing calculations on the sample dataset (e.g., converting measurement units, normalizing data values, classifying data values into a category, or other calculations) as before generating the sample data vector.
  • At step 608, data-searching system 106 may set search parameters, consistent with disclosed embodiments. In some embodiments, the search parameters may be received at step 602. In some embodiments, step 608 may include sending a request for search parameter to a component of system 100 or a computing component outside system 100 and receiving the search parameters in response to the request. For example, the search parameters may include instructions to search the data index based on a comparison of data vector components representing data schemas of the datasets or data schemas of variables of the dataset (e.g., column data). The search parameters may include instructions to search the data index based on a comparison of data vector components representing statistical metrics of the dataset and/or statistical metrics of variables of the dataset. As an example, the search parameters may first specify to conduct an initial search based on comparisons of data schemas and conduct a search based on comparisons of statistical metrics. In some embodiments, the search parameters may include instructions to search the data index by searching data profiles (data schema and statistical metrics) of the datasets and searching data profiles of variables of the datasets.
  • At step 610, data-searching system 106 may search a data index comprising stored data vectors, consistent with disclosed embodiments. The search may be based on search parameters, consistent with disclosed embodiments. In some embodiments, the search may be conducted without k-nearest neighbor clustering. A stored vector may correspond to a reference dataset or to a variable of a reference dataset (e.g., a data column),
  • Step 610 may include identifying the data index. The data index may be stored in, for example, aggregation database 108 or another component of system 100, and data-searching system 106 may identify the data index based on the sample dataset or sample vector (e.g., based on metadata). In some embodiments, the data index may be received at step 602. In some embodiments, identifying a data index may be performed based on user inputs received at step 610 (e.g., receiving an instruction via interface 112). In some embodiments, the data index may be identified based on metadata of associated with the sample data vector.
  • At step 612, data-searching system 106 may generate one or more similarity metrics between the stored data vectors and the sample data vector, consistent with disclosed embodiments. As previously described, the similarity metric may represent a measure of similarity between data. For example, the similarity metric may be based on a covariance matrix, a variance, a frequency of overlapping values, or other measure of statistical similarity. The similarity metric may include a percent match and/or a confidence interval. In some embodiments, data-searching system 106 may generate the similarity metrics using one or more data-profiling models, consistent with disclosed embodiments.
  • In some embodiments, a similarity metric may include a likelihood that a reference dataset derives from sample dataset. In some embodiments, a similarity metric may include a likelihood that the sample dataset derives from a reference dataset. For example, a dataset may derive from another dataset when the data schema and/or data values match according to a threshold or frequency. As an example, a small dataset may be a subset of a large dataset, i.e. it derives from the larger dataset. As another example, one dataset may contain some overlapping values with another dataset and both may contain values that are unique from the other dataset. In that example, either dataset may be said to derive from the other dataset.
  • At step 614, data-searching system 106 may provide the one or more similarity metrics, consistent with disclosed embodiments. Providing the similarity metrics at step 614 may include displaying the similarity metrics at an interface (e.g., interface 112, a display of I/O devices 220, and/or a display of I/O devices 320). Providing the similarity metrics at step 614 may include transmitting the similarity metrics to another component of system 100 (e.g., to client device 114) and/or to a computing component outside system 100 (e.g., via interface 112).
  • At step 616, data-searching system 106 may select one or more data vectors, consistent with disclosed embodiments. In some embodiments, selecting data vectors may be based on the similarity metrics and one or more predetermined selection criteria. The predetermined selection criteria may be received from a user (e.g., at step 602 or as a user-input received via an interface at step 616). The selection criteria may include a command to select any data vectors with similarity criteria falling within a range (e.g., any data vectors between 50% to 90% match). Selecting data vectors may include receiving metadata (e.g., one or more tags or labels) associated with the data vectors.
  • At step 618, data-searching system 106 may retrieve one or more datasets associated with the selected data vectors, consistent with disclosed embodiments, In some embodiments, data-searching system 106 retrieves the datasets from one of data 231, aggregation database 108, data-migration system 110, computing environments 116 a, 116 b, 116 n, databases 118 a, 118 b, 118 n, and/or a computing outside system 100 (e.g., via the interface 112). Retrieving a dataset may be based on the metadata of the selected data vectors, Retrieving a dataset may be include identifying a location (e.g., a database, an internet address) or an identifier of the dataset based on a log and the selected data vector. In some embodiments, the log may be stored in data aggregation database 108.
  • At step 620, data-searching system 106 may provide the datasets associated with the selected data vectors, consistent with disclosed embodiments. Providing the datasets at step 620 may include displaying the datasets at an interface (e.g., interface 112, a display of I/O devices 220, and/or a display of I/O devices 320). Providing the datasets at step 620 may include transmitting the datasets to another component of system 100 (e.g., to client device 114) and/or to a computing component outside system 100 (e.g., via interface 112). Providing the datasets may include storing the datasets in a data storage. Providing the datasets may include providing a link to the datasets. Providing the datasets may include providing a compressed file comprising one or more datasets. Providing the datasets may include sending a request to data-migration system 110 to transfer a dataset to a computing component, consistent with disclosed embodiments.
  • Example Implentation: Conducting a search based on a sample dataset in .CSV-format comprising public health indicators.
  • As an illustrative example, a researcher may be conducting a meta-analysis of a plurality of health studies to identify health indicators that predict a health outcome, such as risk factors that contribute to heart attacks. The researcher may use data-searching system 106 to expand the study by identifying additional data sources to include in the study. The researcher may possess an initial sample dataset, Advantageously, the disclosed system and method provide appropriate datasets to address the researcher's goal of expanding the study, because the system provides datasets that are similar to the sample dataset based on the structure and content of sample dataset. In contrast, conventional search methods may identify datasets that merely happen to contain keywords or other sequences of characters which overlap with the sample datasets. Hence, a conventional search may return redundant or dissimilar datasets that cannot be included in the meta-analysis.
  • In the example, following steps of process 600 beginning with step 602, data-searching system 100 may receive a sample dataset from the researcher's device (e.g., client device 114), consistent with disclosed embodiments. The sample dataset may be a .CSV file. Column headers may include an identifier (e.g., an anonymous patient ID, a social security number, etc.), a data of birth, a date associated with the measurement of other parameters, and one or more other parameters. The parameters may include a current residential address, a sex, a wealth indicator (e.g., home equity, income, debt), one or more health indicators (e.g., a cholesterol measurement, a resting heart rate, a percent body fat, a weight, a height, a smoking status, or the like), and a health event (e.g., a heart attack). Row data may include records associated individual persons.
  • In the example, data-searching system 106 identifies a data schema of the sample dataset (step 604), consistent with disclosed embodiments. Here, the data schema may include column headers. In some cases, the columns may be unlabeled and a data-profiling model may identify a data type and provide a label (e.g., provide a column header). Identifying the data schema may include identifying complex data types using a trained model, consistent with disclosed embodiments.
  • Next, data-searching system 106 generates a sample data vector of the sample dataset (step 606). As an example, generating the sample data vector may include translating the residential address into Global Positioning System coordinates and developing statistical metrics of the GPS coordinates (e.g., the centroid of the GPS coordinates, parameters of a fit to a Poisson distribution or other statistical distribution of the GPS coordinates). The sample data vector may then include the statistical metrics of the GPS coordinates. In addition, generating the sample data vector may include developing a mean, median, and variance of the age of individuals in the dataset based on the date of birth and a data associated with the measurement of parameters. The sample data vector may include the mean, median, and variance of the age.
  • Consistent with disclosed embodiments, in the example, data vectors may be generated using machine learning models (e.g., a data-profiling model) and have no human-understandable interpretation. For example, a machine-learning model comprising an autoencoder model, a VAE model, and/or other model may perform feature embedding and generate vectors based on the identified features, consistent with disclosed embodiments.
  • Data-searching system 106 may set the search parameters (step 608). For example, the system may receive a set of instructions from client device 114, the instructions be based on user inputs. The instructions might specify that the search should be performed over a data index stored in aggregation database 108. The instructions may include a command to assign a weight of 0.8 to a search parameter associated with the data schema (here, column labels). The instructions may further include a command to assign a weight of 0.05 to search parameters corresponding to each of four statistical metrics (e.g., the GPS centroid, the mean age, the median age, and the variance of the age). Alternatively, in the example, the search parameters may be determined using a machine-learning model.
  • Data-searching system 106 may then identify a data index and perform a search of the data index (step 610), consistent with disclosed embodiments. For example, the data index may be stored in aggregation database 108 and include stored data vectors associated with a plurality of public health datasets from a plurality of sources (i.e., systems). The systems may include systems associated with a university, a research institution, an agency (e.g., Center for Disease Control), a health care provider (e.g., a hospital), an insurance provider, the internet (e.g., a dataset retrieved from the internet by data-migration system 110), and/or other systems.
  • Data-searching system 106 may generate similarity metrics based on the search (step 612), consistent with disclosed embodiments. For example, the similarity metric may indicate a “percent match” between one or more stored data vectors and the sample data vector. The stored and sample data vectors may include statistical metrics based on complex data types. In a search, data-searching system 106 may identify two datasets that share the same data schema as the sample data vector. Data-searching system 106 may give a first one of these datasets a 100% match because it has the same GPS centroid, mean age, median age, and variance of the age as the sample dataset, i.e., data-searching system 106 found the same dataset as the sample dataset. However, data-searching system 106 may give the second one of these datasets an 80% match because it has a different GPS centroid, mean age, median age, and variance of the age. In addition, data-searching system 106 may identify a third dataset has a 75% match because it shares some properties of the data schema and has different statistical metrics.
  • In the example, data-searching system 106 may provide the similarity metrics to the researcher's device (step 614). The device (e.g., client device 114) may display the metrics via an interface. The researcher may then be able to provide user inputs to reject the 100% match data vector and select the 80%- and 75% match data vector (step 618). For example, the researcher may be able to click an icon on the interface to select a data vector. Based on the selection, data-searching system 106 may then retrieve datasets associated with the selected data vector and provide the retrieved datasets to the user (steps 618 and 620). For example, data-searching system 106 may determine that a vector corresponds to a dataset stored in computing environment 116 a and retrieve or send a request to receive the corresponding dataset from computing environment 116 a. The interface may display an option to download the retrieved dataset, and may transmit the received dataset to the a researcher's local device when the researcher selects the option.
  • Systems and methods disclosed herein involve unconventional improvements over conventional data processing approaches. Descriptions of the disclosed embodiments are not exhaustive and are not limited to the precise forms or embodiments disclosed. Modifications and adaptations of the embodiments will be apparent from consideration of the specification and practice of the disclosed embodiments. Additionally, the disclosed embodiments are not limited to the examples discussed herein.
  • The foregoing description has been presented for purposes of illustration. It is not exhaustive and is not limited to the precise forms or embodiments disclosed. Modifications and adaptations of the embodiments will be apparent from consideration of the specification and practice of the disclosed embodiments. For example, the described implementations include hardware and software, but systems and methods consistent with the present disclosure can be implemented as hardware alone.
  • Computer programs based on the written description and methods of this specification are within the skill of a software developer. The various functions, scripts, programs, or modules can be created using a variety of programming techniques. For example, programs, scripts, functions, program sections or program modules can be designed in or by means of languages, including JAVASCRIPT, C, C++, JAVA, PHP, PYTHON, RUBY, PERL, BASH, or other programming or scripting languages. One or more of such software sections or modules can be integrated into a computer system, non-transitory computer-readable media, or existing communications software. The programs, modules, or code can also be implemented or replicated as firmware or circuit logic.
  • Moreover, while illustrative embodiments have been described herein, the scope includes any and all embodiments having equivalent elements, modifications, omissions, combinations (e.g., of aspects across various embodiments), adaptations or alterations based on the present disclosure. The elements in the claims are to be interpreted broadly based on the language employed in the claims and not limited to examples described in the present specification or during the prosecution of the application, which examples are to be construed as non-exclusive. Further, the steps of the disclosed methods can be modified in any manner, including by reordering steps or inserting or deleting steps. It is intended, therefore, that the specification and examples be considered as exemplary only, with a true scope and spirit being indicated by the following claims and their full scope of equivalents.

Claims (21)

1-20. (canceled)
21. A system for searching data, comprising:
at least one processor; and
a memory comprising instructions that, when executed by the at least one processor, cause the at least one processor to perform operations comprising:
receiving a search request comprising a sample dataset and a vector similarity threshold of similarity between vectors;
in response to the received search request, performing:
identifying, using a data profiling model comprising a machine learning model, a data schema of the sample dataset;
generating a sample data vector;
searching a data index comprising a plurality of stored data vectors corresponding to a plurality of reference datasets, the stored data vectors comprising information describing corresponding data schemas of the reference datasets, wherein searching the data index comprises performing data schema comparisons between the data schema associated with the sample data vector and the data schemas of the reference datasets;
generating one or more similarity metrics of the sample dataset to individual ones of the reference datasets;
determining, based on the one or more similarity metrics, at least a portion of the reference datasets having at least one data vector satisfying the vector similarity threshold; and
returning, as a result of the received search request, the at least a portion of the reference datasets.
22. The system of claim 21, the operations further comprising:
receiving a request for at least one of a sample dataset, data vector, or data index;
retrieving the sample dataset, data vector or data index.
23. The system of claim 21, wherein the data profiling model includes at least one of an RNN model, a CNN model, or other machine learning model.
24. The system of claim 21, wherein the data profiling model is trained to identify complex data types.
25. The system of claim 21, wherein the data profiling model is stored with a plurality of data profiling models in a model storage.
26. The system of claim 21, wherein the operations further comprise performing calculations on the sample dataset.
27. The system of claim 21, wherein the operations further comprise receiving, by an aggregator, search parameters;
28. The system of claim 27, wherein the search parameters may include instructions to search the data index based on a comparison of data vector components of statistical metrics of the dataset and statistical metrics of variables of the dataset.
29. The system of claim 21, wherein the operations further comprise identifying the data index.
30. The system of claim 21, wherein the operations further comprise returning the at least a portion of the reference datasets by performing at least one of: storing the reference datasets in a data storage, providing a link to the reference datasets, or providing a compressed file of the reference datasets.
31. A method for searching data, comprising:
receiving a search request comprising a sample dataset and a vector similarity threshold of similarity between vectors;
in response to the received search request, performing:
identifying, using a data profiling model comprising a machine learning model, a data schema of the sample dataset;
searching a data index comprising a plurality of stored data vectors corresponding to a plurality of reference datasets, the stored data vectors comprising information describing corresponding data schemas of the reference datasets, wherein searching the data index comprises performing data schema comparisons between the data schema associated with the sample data vector and data schemas of the reference datasets;
generating one or more similarity metrics of the sample dataset to individual ones of the reference datasets;
determining, based on the one or more similarity metrics, at least a portion of the reference datasets having at least one data vector satisfying the vector similarity threshold; and
returning, as a result of the received search request, the at least a portion of the reference datasets.
32. The method of claim 21, the operations further comprising:
receiving a request for at least one of a sample dataset, data vector, or data index;
retrieving the sample dataset, data vector or data index.
33. The method of claim 21, wherein the data profiling model includes at least one of an RNN model, a CNN model, or other machine learning model.
34. The method of claim 21, wherein the data profiling model is trained to identify complex data types.
35. The method of claim 21, wherein the data profiling model is stored with a plurality of data profiling models in a model storage.
36. The method of claim 21, wherein the operations further comprise performing calculations on the sample dataset.
37. The method of claim 21, wherein the operations further comprise receiving, by an aggregator, search parameters;
38. The method of claim 37, wherein the search parameters may include instructions to search the data index based on a comparison of data vector components of statistical metrics of the dataset and statistical metrics of variables of the dataset.
39. The method of claim 21, wherein the operations further comprise identifying the data index.
40. The method of claim 21, wherein the operations further comprise returning the at least a portion of the reference datasets by at least of: storing the reference datasets in a data storage, providing a link to the reference datasets, or providing a compressed file of the reference datasets.
US17/930,926 2018-07-06 2022-09-09 Systems and methods for a data search engine based on data profiles Pending US20230004536A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/930,926 US20230004536A1 (en) 2018-07-06 2022-09-09 Systems and methods for a data search engine based on data profiles

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201862694968P 2018-07-06 2018-07-06
US16/405,956 US11474978B2 (en) 2018-07-06 2019-05-07 Systems and methods for a data search engine based on data profiles
US17/930,926 US20230004536A1 (en) 2018-07-06 2022-09-09 Systems and methods for a data search engine based on data profiles

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/405,956 Continuation US11474978B2 (en) 2018-07-06 2019-05-07 Systems and methods for a data search engine based on data profiles

Publications (1)

Publication Number Publication Date
US20230004536A1 true US20230004536A1 (en) 2023-01-05

Family

ID=67211544

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/405,956 Active US11474978B2 (en) 2018-07-06 2019-05-07 Systems and methods for a data search engine based on data profiles
US17/930,926 Pending US20230004536A1 (en) 2018-07-06 2022-09-09 Systems and methods for a data search engine based on data profiles

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/405,956 Active US11474978B2 (en) 2018-07-06 2019-05-07 Systems and methods for a data search engine based on data profiles

Country Status (2)

Country Link
US (2) US11474978B2 (en)
EP (1) EP3591585A1 (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200320616A1 (en) * 2019-04-04 2020-10-08 Capital One Services, Llc Systems and methods for proactively recognizing reasons for account engagement and providing notifications
US11704494B2 (en) * 2019-05-31 2023-07-18 Ab Initio Technology Llc Discovering a semantic meaning of data fields from profile data of the data fields
US11522881B2 (en) * 2019-08-28 2022-12-06 Nec Corporation Structural graph neural networks for suspicious event detection
US11327665B2 (en) * 2019-09-20 2022-05-10 International Business Machines Corporation Managing data on volumes
CN111309984B (en) * 2020-03-10 2023-09-05 支付宝(杭州)信息技术有限公司 Method and device for retrieving node vector from database by index
CN111626472B (en) * 2020-04-17 2023-10-27 国网甘肃省电力公司信息通信公司 Scene trend judgment index computing system and method based on depth hybrid cloud model
US11500886B2 (en) 2020-12-11 2022-11-15 International Business Machines Corporation Finding locations of tabular data across systems
US11741058B2 (en) * 2021-08-06 2023-08-29 Capital One Services, Llc Systems and methods for architecture embeddings for efficient dynamic synthetic data generation

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130246322A1 (en) * 2012-03-15 2013-09-19 Cept Systems Gmbh Methods, Apparatus and Products for Semantic Processing of Text
US20160232160A1 (en) * 2014-11-26 2016-08-11 Vobis, Inc. Systems and methods to determine and utilize conceptual relatedness between natural language sources
CA3040145A1 (en) * 2016-10-11 2018-04-19 Genomsys Sa Method and systems for the indexing of bioinformatics data
US20180181641A1 (en) * 2015-06-23 2018-06-28 Entit Software Llc Recommending analytic tasks based on similarity of datasets
US10915524B1 (en) * 2017-06-27 2021-02-09 Amazon Technologies, Inc. Scalable distributed data processing and indexing
JP2022114050A (en) * 2021-01-26 2022-08-05 株式会社リコー Information processing apparatus, information processing method, and program

Family Cites Families (135)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5911139A (en) * 1996-03-29 1999-06-08 Virage, Inc. Visual image database search engine which allows for different schema
US5915250A (en) * 1996-03-29 1999-06-22 Virage, Inc. Threshold-based comparison
DE19703965C1 (en) 1997-02-03 1999-05-12 Siemens Ag Process for transforming a fuzzy logic used to simulate a technical process into a neural network
US5974549A (en) 1997-03-27 1999-10-26 Soliton Ltd. Security monitor
US6269351B1 (en) 1999-03-31 2001-07-31 Dryken Technologies, Inc. Method and system for training an artificial neural network
US6137912A (en) 1998-08-19 2000-10-24 Physical Optics Corporation Method of multichannel data compression
US6922699B2 (en) 1999-01-26 2005-07-26 Xerox Corporation System and method for quantitatively representing data objects in vector space
US7630986B1 (en) * 1999-10-27 2009-12-08 Pinpoint, Incorporated Secure data interchange
US20020103793A1 (en) 2000-08-02 2002-08-01 Daphne Koller Method and apparatus for learning probabilistic relational models having attribute and link uncertainty and for performing selectivity estimation using probabilistic relational models
US20030003861A1 (en) 2000-12-07 2003-01-02 Hideki Kagemoto Data broadcast-program production system, data broadcast-program method, data broadcast- program production computer-program, and computer-readable recorded medium
US6804669B2 (en) * 2001-08-14 2004-10-12 International Business Machines Corporation Methods and apparatus for user-centered class supervision
US20090055477A1 (en) 2001-11-13 2009-02-26 Flesher Kevin E System for enabling collaboration and protecting sensitive data
WO2004061702A1 (en) 2002-12-26 2004-07-22 The Trustees Of Columbia University In The City Of New York Ordered data compression system and methods
US20080288889A1 (en) 2004-02-20 2008-11-20 Herbert Dennis Hunt Data visualization application
US7249135B2 (en) * 2004-05-14 2007-07-24 Microsoft Corporation Method and system for schema matching of web databases
US20050278139A1 (en) * 2004-05-28 2005-12-15 Glaenzer Helmut K Automatic match tuning
US20060031622A1 (en) 2004-06-07 2006-02-09 Jardine Robert L Software transparent expansion of the number of fabrics coupling multiple processsing nodes of a computer system
US7599952B2 (en) * 2004-09-09 2009-10-06 Microsoft Corporation System and method for parsing unstructured data into structured data
GB0428191D0 (en) 2004-12-23 2005-01-26 Cambridge Display Tech Ltd Digital signal processing methods and apparatus
US7596560B2 (en) * 2004-12-23 2009-09-29 Raytheon Company System and method for adaptive query identification and acceleration
US20080133573A1 (en) * 2004-12-24 2008-06-05 Michael Haft Relational Compressed Database Images (for Accelerated Querying of Databases)
US7353226B2 (en) * 2005-04-22 2008-04-01 The Boeing Company Systems and methods for performing schema matching with data dictionaries
US20070169017A1 (en) 2006-01-13 2007-07-19 Coward Daniel R Method and apparatus for translating an application programming interface (API) call
US7774288B2 (en) * 2006-05-16 2010-08-10 Sony Corporation Clustering and classification of multimedia data
US7849030B2 (en) * 2006-05-31 2010-12-07 Hartford Fire Insurance Company Method and system for classifying documents
US20080168339A1 (en) 2006-12-21 2008-07-10 Aquatic Informatics (139811) System and method for automatic environmental data validation
US20080270363A1 (en) 2007-01-26 2008-10-30 Herbert Dennis Hunt Cluster processing of a core information matrix
WO2008092147A2 (en) 2007-01-26 2008-07-31 Information Resources, Inc. Analytic platform
US8595559B2 (en) 2007-08-23 2013-11-26 International Business Machines Corporation Method and apparatus for model-based testing of a graphical user interface
JP2009111691A (en) 2007-10-30 2009-05-21 Hitachi Ltd Image-encoding device and encoding method, and image-decoding device and decoding method
US8375014B1 (en) * 2008-06-19 2013-02-12 BioFortis, Inc. Database query builder
US20100036884A1 (en) 2008-08-08 2010-02-11 Brown Robert G Correlation engine for generating anonymous correlations between publication-restricted data and personal attribute data
CA3014839C (en) * 2008-10-23 2019-01-08 Arlen Anderson Fuzzy data operations
US8683554B2 (en) 2009-03-27 2014-03-25 Wavemarket, Inc. System and method for managing third party application program access to user information via a native application program interface (API)
JP5222205B2 (en) 2009-04-03 2013-06-26 Kddi株式会社 Image processing apparatus, method, and program
US8375032B2 (en) 2009-06-25 2013-02-12 University Of Tennessee Research Foundation Method and apparatus for predicting object properties and events using similarity-based information retrieval and modeling
US8990236B2 (en) * 2009-10-09 2015-03-24 Qrative Ltd. Method, computer product program and system for analysis of data
US8438122B1 (en) 2010-05-14 2013-05-07 Google Inc. Predictive analytic modeling platform
US8666998B2 (en) * 2010-09-14 2014-03-04 International Business Machines Corporation Handling data sets
US8832836B2 (en) 2010-12-30 2014-09-09 Verisign, Inc. Systems and methods for malware detection and scanning
US8548951B2 (en) 2011-03-10 2013-10-01 Textwise Llc Method and system for unified information representation and applications thereof
US8533224B2 (en) 2011-05-04 2013-09-10 Google Inc. Assessing accuracy of trained predictive models
US11074495B2 (en) 2013-02-28 2021-07-27 Z Advanced Computing, Inc. (Zac) System and method for extremely efficient image and pattern recognition and artificial intelligence platform
US9916538B2 (en) 2012-09-15 2018-03-13 Z Advanced Computing, Inc. Method and system for feature detection
US8914859B2 (en) 2011-11-07 2014-12-16 International Business Machines Corporation Managing the progressive legible obfuscation and de-obfuscation of public and quasi-public broadcast messages
WO2013089695A1 (en) 2011-12-14 2013-06-20 Intel Corporation Method and system for monitoring calls to an application program interface (api) function
WO2013101723A1 (en) 2011-12-27 2013-07-04 Wellpoint, Inc. Method and system for data pattern matching, masking and removal of sensitive data
US11282287B2 (en) 2012-02-24 2022-03-22 Matterport, Inc. Employing three-dimensional (3D) data predicted from two-dimensional (2D) images using neural networks for 3D modeling applications and other applications
US8782744B1 (en) 2012-06-15 2014-07-15 Amazon Technologies, Inc. Managing API authorization
US9311283B2 (en) 2012-08-16 2016-04-12 Realnetworks, Inc. System for clipping webpages by traversing a dom, and highlighting a minimum number of words
EP2701020A1 (en) 2012-08-22 2014-02-26 Siemens Aktiengesellschaft Monitoring of the initial equipment of a technical system for fabricating a product
US20180173730A1 (en) 2012-09-28 2018-06-21 Clinigence, LLC Generating a Database with Mapped Data
WO2014063124A1 (en) 2012-10-19 2014-04-24 Mcafee, Inc. Mobile application management
WO2014110167A2 (en) 2013-01-08 2014-07-17 Purepredictive, Inc. Integrated machine learning for a data management product
US9274935B1 (en) 2013-01-15 2016-03-01 Google Inc. Application testing system with application programming interface
CN105188516B (en) 2013-03-11 2017-12-22 奇跃公司 For strengthening the System and method for virtual reality
US9104867B1 (en) 2013-03-13 2015-08-11 Fireeye, Inc. Malicious content analysis using simulated user interaction without user involvement
US20140278339A1 (en) 2013-03-15 2014-09-18 Konstantinos (Constantin) F. Aliferis Computer System and Method That Determines Sample Size and Power Required For Complex Predictive and Causal Data Analysis
US9191411B2 (en) 2013-03-15 2015-11-17 Zerofox, Inc. Protecting against suspect social entities
US20140324760A1 (en) 2013-04-30 2014-10-30 Hewlett-Packard Development Company, L.P. Synthetic time series data generation
US9658899B2 (en) 2013-06-10 2017-05-23 Amazon Technologies, Inc. Distributed lock management in a cloud computing environment
US9716842B1 (en) 2013-06-19 2017-07-25 Amazon Technologies, Inc. Augmented reality presentation
US10120838B2 (en) 2013-07-25 2018-11-06 Facebook, Inc. Systems and methods for weighted sampling
US9740662B2 (en) 2013-08-26 2017-08-22 Wright State University Fractional scaling digital filters and the generation of standardized noise and synthetic data series
US20150100537A1 (en) 2013-10-03 2015-04-09 Microsoft Corporation Emoji for Text Predictions
US9531609B2 (en) 2014-03-23 2016-12-27 Ca, Inc. Virtual service automation
US20150309987A1 (en) 2014-04-29 2015-10-29 Google Inc. Classification of Offensive Words
US11069009B2 (en) 2014-05-16 2021-07-20 Accenture Global Services Limited System, method and apparatuses for identifying load volatility of a power customer and a tangible computer readable medium
US9785719B2 (en) 2014-07-15 2017-10-10 Adobe Systems Incorporated Generating synthetic data
US9729506B2 (en) 2014-08-22 2017-08-08 Shape Security, Inc. Application programming interface wall
US20170249432A1 (en) 2014-09-23 2017-08-31 Surgical Safety Technologies Inc. Operating room black-box device, system, method and computer readable medium
US9954893B1 (en) 2014-09-23 2018-04-24 Shape Security, Inc. Techniques for combating man-in-the-browser attacks
US10915233B2 (en) 2014-09-26 2021-02-09 Oracle International Corporation Automated entity correlation and classification across heterogeneous datasets
US10210246B2 (en) * 2014-09-26 2019-02-19 Oracle International Corporation Techniques for similarity analysis and data enrichment using knowledge sources
US20160110657A1 (en) 2014-10-14 2016-04-21 Skytree, Inc. Configurable Machine Learning Method Selection and Parameter Optimization System and Method
US9560075B2 (en) 2014-10-22 2017-01-31 International Business Machines Corporation Cognitive honeypot
US9886247B2 (en) 2014-10-30 2018-02-06 International Business Machines Corporation Using an application programming interface (API) data structure in recommending an API composite
US20160132787A1 (en) * 2014-11-11 2016-05-12 Massachusetts Institute Of Technology Distributed, multi-model, self-learning platform for machine learning
US9460288B2 (en) 2014-12-08 2016-10-04 Shape Security, Inc. Secure app update server and secure application programming interface (“API”) server
US9678999B1 (en) * 2014-12-31 2017-06-13 Teradata Us, Inc. Histogram generation on multiple dimensions
US10643138B2 (en) 2015-01-30 2020-05-05 Micro Focus Llc Performance testing based on variable length segmentation and clustering of time series data
US9614670B1 (en) 2015-02-05 2017-04-04 Ionic Security Inc. Systems and methods for encryption and provision of information security using platform services
WO2016168368A1 (en) 2015-04-13 2016-10-20 Secful, Inc. System and method for identifying and preventing malicious api attacks
US9462013B1 (en) 2015-04-29 2016-10-04 International Business Machines Corporation Managing security breaches in a networked computing environment
US10163061B2 (en) 2015-06-18 2018-12-25 International Business Machines Corporation Quality-directed adaptive analytic retraining
US11567962B2 (en) 2015-07-11 2023-01-31 Taascom Inc. Computer network controlled data orchestration system and method for data aggregation, normalization, for presentation, analysis and action/decision making
US10750161B2 (en) 2015-07-15 2020-08-18 Fyusion, Inc. Multi-view interactive digital media representation lock screen
GB2541231A (en) * 2015-08-13 2017-02-15 Fujitsu Ltd Hybrid data storage system and method and program for storing hybrid data
US10055430B2 (en) * 2015-10-14 2018-08-21 International Business Machines Corporation Method for classifying an unmanaged dataset
US20180253894A1 (en) 2015-11-04 2018-09-06 Intel Corporation Hybrid foreground-background technique for 3d model reconstruction of dynamic scenes
US9934397B2 (en) 2015-12-15 2018-04-03 International Business Machines Corporation Controlling privacy in a face recognition application
US10031745B2 (en) 2016-02-02 2018-07-24 International Business Machines Corporation System and method for automatic API candidate generation
US11769193B2 (en) 2016-02-11 2023-09-26 Ebay Inc. System and method for detecting visually similar items
CN107292326A (en) 2016-03-31 2017-10-24 阿里巴巴集团控股有限公司 The training method and device of a kind of model
US10403043B2 (en) 2016-04-14 2019-09-03 The Research Foundation For The State University Of New York System and method for generating a progressive representation associated with surjectively mapped virtual and physical reality image data
EP3239863A1 (en) 2016-04-29 2017-11-01 QlikTech International AB System and method for interactive discovery of inter-data set relationships
US10462181B2 (en) 2016-05-10 2019-10-29 Quadrant Information Security Method, system, and apparatus to identify and study advanced threat tactics, techniques and procedures
US10009363B2 (en) * 2016-06-09 2018-06-26 Adobe Systems Incorporated Selecting representative metrics datasets for efficient detection of anomalous data
AU2017300259A1 (en) 2016-07-18 2019-02-14 Nant Holdings Ip, Llc Distributed machine learning systems, apparatus, and methods
US10503829B2 (en) 2016-10-13 2019-12-10 Booxby Inc. Book analysis and recommendation
JP2018067115A (en) 2016-10-19 2018-04-26 セイコーエプソン株式会社 Program, tracking method and tracking device
JP7100422B2 (en) * 2016-10-21 2022-07-13 富士通株式会社 Devices, programs, and methods for recognizing data properties
US10609284B2 (en) 2016-10-22 2020-03-31 Microsoft Technology Licensing, Llc Controlling generation of hyperlapse from wide-angled, panoramic videos
US10621210B2 (en) * 2016-11-27 2020-04-14 Amazon Technologies, Inc. Recognizing unknown data objects
US9754190B1 (en) 2016-11-29 2017-09-05 Seematics Systems Ltd System and method for image classification based on Tsallis entropy
US11068949B2 (en) 2016-12-09 2021-07-20 365 Retail Markets, Llc Distributed and automated transaction systems
US10713384B2 (en) 2016-12-09 2020-07-14 Massachusetts Institute Of Technology Methods and apparatus for transforming and statistically modeling relational databases to synthesize privacy-protected anonymized data
US10430661B2 (en) 2016-12-20 2019-10-01 Adobe Inc. Generating a compact video feature representation in a digital medium environment
US10163003B2 (en) 2016-12-28 2018-12-25 Adobe Systems Incorporated Recognizing combinations of body shape, pose, and clothing in three-dimensional input images
US11496286B2 (en) 2017-01-08 2022-11-08 Apple Inc. Differential privacy with cloud data
US10212428B2 (en) 2017-01-11 2019-02-19 Microsoft Technology Licensing, Llc Reprojecting holographic video to enhance streaming bandwidth/quality
US10448054B2 (en) 2017-01-11 2019-10-15 Groq, Inc. Multi-pass compression of uncompressed data
US10360517B2 (en) 2017-02-22 2019-07-23 Sas Institute Inc. Distributed hyperparameter tuning system for machine learning
US10554607B2 (en) 2017-02-24 2020-02-04 Telefonaktiebolaget Lm Ericsson (Publ) Heterogeneous cloud controller
US11068453B2 (en) * 2017-03-09 2021-07-20 data.world, Inc Determining a degree of similarity of a subset of tabular data arrangements to subsets of graph data arrangements at ingestion into a data-driven collaborative dataset platform
US10891545B2 (en) 2017-03-10 2021-01-12 International Business Machines Corporation Multi-dimensional time series event prediction via convolutional neural network(s)
US20180260474A1 (en) 2017-03-13 2018-09-13 Arizona Board Of Regents On Behalf Of The University Of Arizona Methods for extracting and assessing information from literature documents
US11574164B2 (en) 2017-03-20 2023-02-07 International Business Machines Corporation Neural network cooperation
US11734584B2 (en) 2017-04-19 2023-08-22 International Business Machines Corporation Multi-modal construction of deep learning networks
US20170228697A1 (en) * 2017-04-25 2017-08-10 Stella.Ai, Inc. System and method for matching candidates to a job using job similarity and candidate similarity
US20170228696A1 (en) * 2017-04-25 2017-08-10 Stella.Ai, Inc. System and method for determining job similarity using a collation of career streams to match candidates to a job
US20180336463A1 (en) 2017-05-18 2018-11-22 General Electric Company Systems and methods for domain-specific obscured data transport
US10348658B2 (en) 2017-06-15 2019-07-09 Google Llc Suggested items for use with embedded applications in chat conversations
CN109214238B (en) 2017-06-30 2022-06-28 阿波罗智能技术(北京)有限公司 Multi-target tracking method, device, equipment and storage medium
US10902322B2 (en) * 2017-07-26 2021-01-26 Adobe Inc. Classification training techniques to map datasets to a standardized data model
US10467039B2 (en) 2017-08-07 2019-11-05 Open Data Group Inc. Deployment and management platform for model execution engine containers
US10929987B2 (en) 2017-08-16 2021-02-23 Nvidia Corporation Learning rigidity of dynamic scenes for three-dimensional scene flow estimation
US10942914B2 (en) * 2017-10-19 2021-03-09 Adobe Inc. Latency optimization for digital asset compression
US10990901B2 (en) 2017-11-13 2021-04-27 Accenture Global Solutions Limited Training, validating, and monitoring artificial intelligence and machine learning models
US10122969B1 (en) 2017-12-07 2018-11-06 Microsoft Technology Licensing, Llc Video capture systems and methods
WO2019126625A1 (en) 2017-12-22 2019-06-27 Butterfly Network, Inc. Methods and apparatuses for identifying gestures based on ultrasound data
US10453220B1 (en) 2017-12-29 2019-10-22 Perceive Corporation Machine-trained network for misalignment-insensitive depth perception
US10706267B2 (en) 2018-01-12 2020-07-07 Qualcomm Incorporated Compact models for object recognition
US11023473B2 (en) * 2018-02-09 2021-06-01 Microsoft Technology Licensing, Llc Recurrent binary embedding for information retrieval
US10831752B2 (en) * 2018-04-25 2020-11-10 International Business Machines Corporation Semantic relational database operations

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130246322A1 (en) * 2012-03-15 2013-09-19 Cept Systems Gmbh Methods, Apparatus and Products for Semantic Processing of Text
US20160232160A1 (en) * 2014-11-26 2016-08-11 Vobis, Inc. Systems and methods to determine and utilize conceptual relatedness between natural language sources
US20180181641A1 (en) * 2015-06-23 2018-06-28 Entit Software Llc Recommending analytic tasks based on similarity of datasets
CA3040145A1 (en) * 2016-10-11 2018-04-19 Genomsys Sa Method and systems for the indexing of bioinformatics data
US10915524B1 (en) * 2017-06-27 2021-02-09 Amazon Technologies, Inc. Scalable distributed data processing and indexing
JP2022114050A (en) * 2021-01-26 2022-08-05 株式会社リコー Information processing apparatus, information processing method, and program

Also Published As

Publication number Publication date
US20200012626A1 (en) 2020-01-09
US11474978B2 (en) 2022-10-18
EP3591585A1 (en) 2020-01-08

Similar Documents

Publication Publication Date Title
US20210365305A1 (en) Systems and methods for quickly searching datasets by indexing synthetic data generating models
US20230004536A1 (en) Systems and methods for a data search engine based on data profiles
US9646262B2 (en) Data intelligence using machine learning
EP3591586A1 (en) Data model generation using generative adversarial networks and fully automated machine learning system which generates and optimizes solutions given a dataset and a desired outcome
US20230139783A1 (en) Schema-adaptable data enrichment and retrieval
US11810000B2 (en) Systems and methods for expanding data classification using synthetic data generation in machine learning models
US20210279249A1 (en) Systems and methods for formatting data using a recurrent neural network
US11409772B2 (en) Active learning for data matching
US20220215931A1 (en) Generating multi-dimensional recommendation data objects based on decentralized crowd sourcing
AU2020103522A4 (en) DAMA- Dataset Discovery: Dataset Discovery in Data Analytics and Machine Learning Algorithm
AU2020104034A4 (en) IML-Cloud Data Performance: Cloud Data Performance Improved using Machine Learning.

Legal Events

Date Code Title Description
AS Assignment

Owner name: CAPITAL ONE SERVICES, LLC, VIRGINIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WALTERS, AUSTIN;PHAM, VINCENT;RAFFERTY, GALEN;AND OTHERS;REEL/FRAME:061049/0703

Effective date: 20190506

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED