JPWO2020159568A5 - - Google Patents

Download PDF

Info

Publication number
JPWO2020159568A5
JPWO2020159568A5 JP2021517369A JP2021517369A JPWO2020159568A5 JP WO2020159568 A5 JPWO2020159568 A5 JP WO2020159568A5 JP 2021517369 A JP2021517369 A JP 2021517369A JP 2021517369 A JP2021517369 A JP 2021517369A JP WO2020159568 A5 JPWO2020159568 A5 JP WO2020159568A5
Authority
JP
Japan
Prior art keywords
data
training
iteration
machine learning
predictions
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2021517369A
Other languages
Japanese (ja)
Other versions
JP7308262B2 (en
JP2022518646A (en
Publication date
Priority claimed from US16/458,924 external-priority patent/US20200242511A1/en
Application filed filed Critical
Publication of JP2022518646A publication Critical patent/JP2022518646A/en
Publication of JPWO2020159568A5 publication Critical patent/JPWO2020159568A5/ja
Application granted granted Critical
Publication of JP7308262B2 publication Critical patent/JP7308262B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Claims (14)

動的なデータ選択を用いる機械学習予測モデルを実装するための方法であって、
訓練済みの機械学習モデルが生成する複数のデータ予測にアクセスするステップを含み、前記データ予測は、対応する観測データから構成され、前記方法は、さらに、
前記アクセスしたデータ予測の数および前記対応する観測データに基づいて前記機械学習モデルの正解率を算出するステップと、
可変数のデータ予測を用いて前記アクセスするステップおよび前記算出するステップを繰り返すステップとを含み、
前記可変数のデータ予測は、前回のイテレーション(繰り返し)中に実行された操作に基づいて調整され、
前記算出した正解率が所与のイテレーション中に正解率基準を満たさない場合、前記機械学習モデルに対する訓練がトリガされる、方法。
A method for implementing machine learning predictive models that use dynamic data selection.
The data prediction comprises the steps of accessing multiple data predictions generated by the trained machine learning model, the data prediction being composed of corresponding observational data, the method further comprising:
A step of calculating the accuracy rate of the machine learning model based on the number of data predictions accessed and the corresponding observation data, and
Includes the step of accessing and repeating the step of calculating using a variable number of data predictions.
The variable number of data predictions are adjusted based on the operations performed during the last iteration.
A method in which training for the machine learning model is triggered if the calculated accuracy rate does not meet the accuracy criteria during a given iteration.
現在のイテレーションの前記可変数のデータ予測は、前記前回のイテレーション中に訓練がトリガされたかどうかに基づいて調整される、請求項1に記載の方法。 The method of claim 1, wherein the variable number of data predictions for the current iteration are adjusted based on whether training was triggered during the previous iteration. 前記前回のイテレーションは、前記現在のイテレーションの直前のイテレーションから構成される、請求項2に記載の方法。 The method of claim 2, wherein the previous iteration comprises the iteration immediately preceding the current iteration. 前記トリガされた訓練は、前記訓練済み機械学習モデルに対する再訓練または更新済みの訓練から構成される、請求項2または3に記載の方法。 The method of claim 2 or 3 , wherein the triggered training comprises retraining or updated training on the trained machine learning model. 訓練がトリガされた場合、前記アクセスするステップおよび前記算出するステップの次のイテレーションは、前記トリガされた訓練が生成した前記機械学習モデルによって生成されたデータ予測を用いる、請求項2~4のいずれか1項に記載の方法。 When the training is triggered, the next iteration of the accessing step and the calculating step uses any of claims 2-4 using the data prediction generated by the machine learning model generated by the triggered training. Or the method described in item 1 . 前記繰り返すステップは、所定時間に従って実行される、請求項2~5のいずれか1項に記載の方法。 The method according to any one of claims 2 to 5 , wherein the repeating step is executed according to a predetermined time. 前記所定時間は、所定期間、または対応する観測データを有する所定量のデータ予測である、請求項6に記載の方法。 The method of claim 6, wherein the predetermined time is a predetermined period or a predetermined amount of data prediction having corresponding observation data. 前記前回のイテレーション中に訓練がトリガされた場合、前記現在のイテレーションの前記データ予測の数は増やされ、前記前回のイテレーション中に訓練がトリガされなかった場合、前記データ予測の数は減らされる、請求項6または7に記載の方法。 If the training was triggered during the previous iteration, the number of said data predictions for the current iteration is increased, and if the training was not triggered during the previous iteration, the number of said data predictions is decreased. The method according to claim 6 or 7 . 前記前回のイテレーション中に訓練がトリガされた場合、前記現在のイテレーションの前記データ予測の数は係数で乗算され、前記前回のイテレーション中に訓練がトリガされなかった場合、前記データ予測の数は係数で除算される、請求項8に記載の方法。 If the training was triggered during the previous iteration, the number of said data predictions in the current iteration is multiplied by a factor, and if the training was not triggered during the previous iteration, the number of said data predictions is a factor. The method of claim 8, which is divided by. 前記係数は、1以上の値から構成される、請求項9に記載の方法。 The method according to claim 9, wherein the coefficient is composed of one or more values. 前記係数の前記値は、複数のイテレーションにわたって減らされる、請求項10に記載の方法。 10. The method of claim 10, wherein the value of the coefficient is reduced over a plurality of iterations. 設定されたデータ予測の数が前記繰り返すステップに基づいて決定されるよう、複数の前記イテレーションを実行した後に前記アクセスするステップおよび前記算出するステップを前記可変数のデータ予測を用いて終了するステップをさらに含み、前記繰り返すステップの後、前記設定されたデータ予測の数は、前記機械学習モデルの正解率を算出するために用いられ、前記正解率は、前記機械学習モデルの訓練をトリガするために用いられる、請求項11に記載の方法。 A step of ending the accessing step and the calculating step using the variable number of data predictions after performing the plurality of iterations so that the number of set data predictions is determined based on the repeating steps. Further including, after the repeating step, the set number of data predictions is used to calculate the accuracy rate of the machine learning model, the accuracy rate to trigger training of the machine learning model. The method of claim 11 used. コンピュータに請求項1~12のいずれか1項に記載の方法を実行させる、プログラム。A program that causes a computer to perform the method according to any one of claims 1 to 12. 請求項13に記載のプログラムを格納したメモリと
前記プログラムを実行するプロセッサとを備え、システム。
The memory in which the program according to claim 13 is stored and
A system comprising a processor that executes the program .
JP2021517369A 2019-01-30 2019-07-05 Dynamic data selection for machine learning models Active JP7308262B2 (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
IN201941003803 2019-01-30
IN201941003803 2019-01-30
US16/458,924 US20200242511A1 (en) 2019-01-30 2019-07-01 Dynamic Data Selection for a Machine Learning Model
US16/458,924 2019-07-01
PCT/US2019/040693 WO2020159568A1 (en) 2019-01-30 2019-07-05 Dynamic data selection for a machine learning model

Publications (3)

Publication Number Publication Date
JP2022518646A JP2022518646A (en) 2022-03-16
JPWO2020159568A5 true JPWO2020159568A5 (en) 2022-04-13
JP7308262B2 JP7308262B2 (en) 2023-07-13

Family

ID=71731357

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2021517369A Active JP7308262B2 (en) 2019-01-30 2019-07-05 Dynamic data selection for machine learning models

Country Status (5)

Country Link
US (1) US20200242511A1 (en)
EP (1) EP3918541A1 (en)
JP (1) JP7308262B2 (en)
CN (1) CN112789633A (en)
WO (1) WO2020159568A1 (en)

Families Citing this family (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10678244B2 (en) 2017-03-23 2020-06-09 Tesla, Inc. Data synthesis for autonomous control systems
US11409692B2 (en) 2017-07-24 2022-08-09 Tesla, Inc. Vector computational unit
US11157441B2 (en) 2017-07-24 2021-10-26 Tesla, Inc. Computational array microprocessor system using non-consecutive data formatting
US11893393B2 (en) 2017-07-24 2024-02-06 Tesla, Inc. Computational array microprocessor system with hardware arbiter managing memory requests
US10671349B2 (en) 2017-07-24 2020-06-02 Tesla, Inc. Accelerated mathematical engine
US11561791B2 (en) 2018-02-01 2023-01-24 Tesla, Inc. Vector computational unit receiving data elements in parallel from a last row of a computational array
US11215999B2 (en) 2018-06-20 2022-01-04 Tesla, Inc. Data pipeline and deep learning system for autonomous driving
US11361457B2 (en) 2018-07-20 2022-06-14 Tesla, Inc. Annotation cross-labeling for autonomous control systems
US11636333B2 (en) 2018-07-26 2023-04-25 Tesla, Inc. Optimizing neural network structures for embedded systems
US11562231B2 (en) 2018-09-03 2023-01-24 Tesla, Inc. Neural networks for embedded devices
CN115512173A (en) 2018-10-11 2022-12-23 特斯拉公司 System and method for training machine models using augmented data
US11196678B2 (en) 2018-10-25 2021-12-07 Tesla, Inc. QOS manager for system on a chip communications
US11816585B2 (en) 2018-12-03 2023-11-14 Tesla, Inc. Machine learning models operating at different frequencies for autonomous vehicles
US11537811B2 (en) 2018-12-04 2022-12-27 Tesla, Inc. Enhanced object detection for autonomous vehicles based on field view
US11610117B2 (en) 2018-12-27 2023-03-21 Tesla, Inc. System and method for adapting a neural network model on a hardware platform
US10997461B2 (en) 2019-02-01 2021-05-04 Tesla, Inc. Generating ground truth for machine learning from time series elements
US11567514B2 (en) 2019-02-11 2023-01-31 Tesla, Inc. Autonomous and user controlled vehicle summon to a target
US10956755B2 (en) 2019-02-19 2021-03-23 Tesla, Inc. Estimating object properties using visual image data
US11966818B2 (en) 2019-02-21 2024-04-23 Hewlett Packard Enterprise Development Lp System and method for self-healing in decentralized model building for machine learning using blockchain
US11966840B2 (en) * 2019-08-15 2024-04-23 Noodle Analytics, Inc. Deep probabilistic decision machines
US11556117B2 (en) * 2019-10-21 2023-01-17 Applied Materials, Inc. Real-time anomaly detection and classification during semiconductor processing
US11190425B2 (en) 2019-12-30 2021-11-30 Viavi Solutions Inc. Anomaly detection in a network based on a key performance indicator prediction model
US11615347B2 (en) * 2019-12-31 2023-03-28 Paypal, Inc. Optimizing data processing and feature selection for model training
US20210241183A1 (en) * 2020-01-31 2021-08-05 Hewlett Packard Enterprise Development Lp Adaptively synchronizing learning of multiple learning models
KR20220038907A (en) * 2020-09-21 2022-03-29 삼성에스디에스 주식회사 Data prediction method based on generative adversarial network and apparatus implementing the same method
US11769318B2 (en) 2020-11-23 2023-09-26 Argo AI, LLC Systems and methods for intelligent selection of data for building a machine learning model
CN114764967A (en) * 2021-01-14 2022-07-19 新智数字科技有限公司 Equipment fault alarm method under combined learning framework
US11657591B2 (en) * 2021-01-15 2023-05-23 Argo AI, LLC Autonomous vehicle system for intelligent on-board selection of data for building a remote machine learning model
CN113345538B (en) * 2021-06-23 2022-09-30 北京理工大学重庆创新中心 Material performance prediction method based on dynamic selection training set
US11507922B1 (en) 2021-06-24 2022-11-22 Coupang Corp. Computer-implemented systems and methods for artificial intelligence (AI)-based inbound plan generation using fungibility logic
WO2022269331A1 (en) * 2021-06-24 2022-12-29 Coupang Corp. Computer-implemented systems and methods for artificial intelligence (ai)-based inbound plan generation

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9405389B2 (en) * 2014-08-29 2016-08-02 Microsoft Technology Licensing, Llc Noise reduction through democratic alpha smoothing
US10163061B2 (en) 2015-06-18 2018-12-25 International Business Machines Corporation Quality-directed adaptive analytic retraining
US10755196B2 (en) * 2016-05-06 2020-08-25 Accenture Global Solutions Limited Determining retraining of predictive models
CN106127248A (en) 2016-06-24 2016-11-16 平安科技(深圳)有限公司 Car plate sorting technique based on degree of depth study and system
US20180197087A1 (en) * 2017-01-06 2018-07-12 Accenture Global Solutions Limited Systems and methods for retraining a classification model
CN107330522B (en) * 2017-07-04 2021-06-08 北京百度网讯科技有限公司 Method, device and system for updating deep learning model
RU2672394C1 (en) * 2017-07-26 2018-11-14 Общество С Ограниченной Ответственностью "Яндекс" Methods and systems for evaluation of training objects through a machine training algorithm
US20190130303A1 (en) * 2017-10-26 2019-05-02 International Business Machines Corporation Smart default threshold values in continuous learning
US20200167669A1 (en) * 2018-11-27 2020-05-28 Oracle International Corporation Extensible Software Tool with Customizable Machine Prediction

Similar Documents

Publication Publication Date Title
JPWO2020159568A5 (en)
JP6817431B2 (en) Neural architecture search
US20190361808A1 (en) Cache configuration performance estimation
KR20180091842A (en) Training of neural networks with prioritized experience memory
JP2005010140A5 (en)
CN107644630A (en) Melody generation method and device based on neutral net
JPWO2019221985A5 (en)
JP2021500658A5 (en)
US11907821B2 (en) Population-based training of machine learning models
WO2020164271A1 (en) Pooling method and device for convolutional neural network, storage medium and computer device
CN107457780B (en) Method and device for controlling mechanical arm movement, storage medium and terminal equipment
US20210224692A1 (en) Hyperparameter tuning method, device, and program
JP2020091843A5 (en)
CN112508190A (en) Method, device and equipment for processing structured sparse parameters and storage medium
JP2023052555A5 (en)
JP7318383B2 (en) Information processing program, information processing method, and information processing apparatus
JPWO2021113044A5 (en)
KR20220032861A (en) Neural architecture search method and attaratus considering performance in hardware
JP7150651B2 (en) Neural network model reducer
CN110866403B (en) End-to-end conversation state tracking method and system based on convolution cycle entity network
JP7301801B2 (en) Hyperparameter tuning method, device and program
KR102559605B1 (en) Method and apparatus for function optimization
JP7452648B2 (en) Learning methods, learning devices and programs
CN113419424A (en) Modeling reinforcement learning robot control method and system capable of reducing over-estimation
WO2020054402A1 (en) Neural network processing device, computer program, neural network manufacturing method, neural network data manufacturing method, neural network use device, and neural network downscaling method