WO2024174136A1 - Method for improving prediction accuracy of ai model, electronic device and medium - Google Patents
Method for improving prediction accuracy of ai model, electronic device and medium Download PDFInfo
- Publication number
- WO2024174136A1 WO2024174136A1 PCT/CN2023/077710 CN2023077710W WO2024174136A1 WO 2024174136 A1 WO2024174136 A1 WO 2024174136A1 CN 2023077710 W CN2023077710 W CN 2023077710W WO 2024174136 A1 WO2024174136 A1 WO 2024174136A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- model
- accuracy
- rendering
- current
- rendering condition
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 36
- 238000009877 rendering Methods 0.000 claims abstract description 124
- 238000012549 training Methods 0.000 claims abstract description 11
- 238000004590 computer program Methods 0.000 claims description 6
- 238000012360 testing method Methods 0.000 claims description 6
- 238000010200 validation analysis Methods 0.000 claims description 4
- 238000012544 monitoring process Methods 0.000 claims 4
- 238000013473 artificial intelligence Methods 0.000 description 53
- 241000282326 Felis catus Species 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 2
- 238000000429 assembly Methods 0.000 description 1
- 230000000712 assembly Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000013145 classification model Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
Definitions
- Embodiments of the present application mainly relate to the field of artificial intelligence (AI) models, and in particular to a method for improving the prediction accuracy of an AI model, an electronic device and a medium.
- AI artificial intelligence
- Embodiments of the present application provide a method for improving the prediction accuracy of an AI model, an electronic device and a medium. By means of the embodiments of the present application, relevant conditions for improving the prediction accuracy of the AI model can be accurately and rapidly obtained.
- a method for improving the prediction accuracy of an AI model includes: acquiring a 3D model and a first group of rendering condition set, where the first group of rendering condition set includes a plurality of default rendering conditions and at least one default rendering condition in the plurality of default rendering conditions contains a first numerical interval; acquiring label information according to the 3D model; determining the first group of rendering condition set as a current group of rendering condition set; determining a first AI model as a current AI model; randomly determining combinations of a plurality of groups of rendering conditions from the current group of rendering condition set, and rendering same respectively to obtain corresponding 2D images, so as to form a first image data set, where each 2D image in the first image data set contains corresponding label information; training the current AI model by means of the first image data set; evaluating a prediction performance of the trained current AI model to obtain a first accuracy; judging whether the first accuracy meets a preset condition; when the first accuracy does not meet the preset condition, changing at least one rendering condition in the current group
- an electronic device which includes: at least one memory, configured to store computer readable code; and at least one processor, configured to call the computer readable code to perform all steps in the method provided in the first aspect.
- a computer readable medium is provided; computer readable instructions are stored on the computer readable medium, and when executed, the computer readable instructions cause the processor to perform all the steps in the method provided in the first aspect.
- a computer program product which includes a computer program; and when executed by the processor, the computer program performs all the steps in the method provided in the first aspect.
- FIG. 1 is a flowchart of a method for improving the prediction accuracy of an AI model according to an embodiment of the present application.
- FIG. 2 is a schematic diagram of an electronic apparatus according to an embodiment of the present application.
- the term “include” and its variations represent an open term, and mean “include but not limited to” .
- the term “based on” represents “based at least in part on” .
- the terms “one embodiment” and “an embodiment” represent “at least one embodiment” .
- the term “another embodiment” represents “at least one other embodiment” .
- the terms “first” , “second” and so on may refer to different or identical object. Other definitions, either clear or implicit, may also be included below. Unless otherwise indicated obviously from the context, the definition of one term is consistent throughout the specification.
- FIG. 1 is a flowchart of a method for improving the prediction accuracy of an AI model according to an embodiment of the present application. As shown in FIG. 1, the method 100 for improving the prediction accuracy of an AI model includes the following steps:
- Step 110 A 3D model and a first group of rendering condition set are acquired.
- the first group of rendering condition set includes a plurality of default rendering conditions and at least one default rendering condition in the plurality of default rendering conditions contains a first numerical interval.
- At least one 3D model and the first group of rendering condition set may be acquired from a renderer.
- the first group of rendering condition set includes a type of a rendering engine, setting requirements of a camera, setting requirements of light, and setting requirements of a texture.
- the first group of rendering condition set may also include: employing/or not employing image enhancement to obtain more image data, or specific means of the image enhancement, for example, rotation, scaling, random cropping, or adding noises.
- Step 120 Label information is acquired according to the 3D model.
- acquiring label information is specifically to acquire the label information of the at least one 3D model and/or at least one component in the at least one 3D model.
- the acquiring label information in step 120 is to acquire the label information of the car wheels in a relevant rendering scene.
- the label information includes a bounding box of car wheel and a label name corresponding to the car wheel.
- the acquiring label information is to acquire the label information of a scene itself containing the 3D model in the renderer.
- the acquiring label information in step 120 is to acquire the label information of a relevant rendering scene itself.
- the label information is a label name corresponding to the cat or the dog.
- Step 130 The first group of rendering condition set is determined as a current group of rendering condition set.
- a first AI model is determined as a current AI model.
- Step 140 Combinations of a plurality of groups of rendering conditions are randomly determined from the current group of rendering condition set, and same is rendered respectively to obtain corresponding 2D images, so as to form a first image data set.
- Each 2D image in the first image data set contains corresponding label information.
- N numerical values are randomly selected from a numerical interval of at least one rendering condition.
- the N numerical values are combined with other rendering conditions in the current group of rendering condition set, respectively, to determine the combinations of the plurality of groups of rendering conditions.
- the current group of rendering condition set includes three rendering conditions A, B, and C, where the rendering condition A: [0°, 10°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] .
- the randomly determining combinations of a plurality of groups of rendering conditions from the current group of rendering condition set refers to randomly selecting corresponding numerical values from the rendering conditions with the numerical intervals in the current group of rendering condition set, and combining same with the rendering conditions without the numerical intervals respectively, so as to determine the combinations of the plurality of groups of rendering conditions.
- one of the combinations of the plurality of groups of rendering conditions may be: the rendering condition A: [0°] ; the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] ;
- one of the combinations of the plurality of groups of rendering conditions may be: the rendering condition A: [1°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] , ...
- one of the combinations of the plurality of groups of rendering conditions may be: the rendering condition A: [10°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] .
- Step 150 The current AI model is trained by means of the first image data set.
- the first image data set is divided into a first training set, a first validation set, and a first testing set.
- the current AI model may be trained by means of the first training set and the first validation set.
- Step 160 A prediction performance of the trained current AI model is evaluated to obtain a first accuracy.
- the prediction performance of the trained current AI model is evaluated by means of the first testing set or a real image data set with the label information to obtain the first accuracy.
- the prediction performance of the trained current AI model may be evaluated by means of a cloud AI evaluation module or a local server with strong computing performance.
- the prediction performance of the trained current AI model is evaluated by means of the first testing set or the real image data set with the label information to obtain an accuracy at a first stage.
- the trained current AI model is deployed to an edge device.
- the prediction performance of the trained current AI model is monitored on the edge device to obtain an accuracy at a second stage. Both the accuracy at the first stage and the accuracy at the second stage are taken as the first accuracy, that is, the first accuracy may include data at two stages.
- the trained current AI model is deployed to the edge device.
- the prediction performance of the trained current AI model is monitored to obtain the first accuracy.
- the prediction performance of the trained current AI model is monitored by means of a machine vision system or a sensor or artificially.
- a sensor such as a radar or a distance measurement sensor may assist in determining whether there is an object in a relevant position, which is predicted by means of the current AI model, so as to evaluate whether the prediction of the current AI model is accurate.
- the current group of rendering condition set corresponding to the first accuracy and corresponding trained AI model information are inputted into a knowledge graph.
- cases with relatively low accuracy therein may also be inputted into the knowledge graph, with the purpose of being ready for further data analysis and capable of retraining the AI model.
- Step 170 Whether the first accuracy meets a preset condition is judged.
- step 171 is performed to change at least one rendering condition in the current group of rendering condition set to obtain an updated current group of rendering condition set. The method returns back to perform step 140 to step 170 until the first accuracy meets the preset condition.
- the at least one rendering condition in the current group of rendering condition set is changed within a preset range corresponding to each rendering condition in the current group of rendering condition set.
- it may be randomly changed or may be manually changed by a user.
- each rendering condition in the current group of rendering condition set includes A, B, and C
- the preset range of the rendering condition A is [0°, 100°]
- the preset range of the rendering condition B is [texture 1, texture 2, texture 3]
- the preset range of the rendering condition C is [light source 1, light source 2] .
- the first group of rendering condition set includes: the rendering condition A: [0°, 10°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1]
- the updated current group of rendering condition set for example: the rendering condition A: [20°, 30°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1]
- the rendering condition A [20°, 30°]
- the rendering condition B [texture 2]
- the rendering condition C [light source 1]
- Step 172 A rendering condition set corresponding to the first accuracy meeting the preset condition, and corresponding trained AI model information are outputted.
- the until the first accuracy meets the preset condition refers to until the first accuracy meets a highest accuracy within a preset number of cycles.
- a second AI model is determined as the current AI model.
- the method returns back to perform step 140 to step 170 until the second accuracy meets the highest accuracy within the preset number of cycles.
- K AI models are preset, the highest accuracy corresponding to each AI model is obtained, respectively.
- the rendering condition set corresponding to the highest accuracy and the corresponding trained AI model information are selected from all the accuracies corresponding to the K AI models for outputting.
- the AI model information may include an AI model name, hyper-parameters, and the like.
- the embodiments of the present application may continuously improve the quality of a synthetic image data set through result orientation, specifically, continuously evaluate the prediction performance of the AI model after the synthetic image data set generated under different rendering conditions is trained on the AI model, and obtain the trained model information of the relevant AI model, so as to accurately and quickly obtain corresponding conditions for improving the prediction accuracy of the AI model.
- FIG. 2 is a schematic diagram of a device control platform 200 according to an embodiment of the present application.
- the device control platform 200 includes a processor 201 and a memory 202. Instructions are stored in the memory 202. When executed by the processor 201, the instructions implement the method 100 as described above.
- the at least one processor 201 may include a microprocessor, an application specific integrated circuit (ASIC) , a digital signal processor (DSP) , a central processing unit (CPU) , a graphics processing unit (GPU) , a state machine, and the like.
- Embodiments of the computer readable medium include, but are not limited to, a floppy disk, a CD-ROM, a magnetic disk, a memory chip, a ROM, an RAM, an ASIC, a configured processor, an all-optical medium, all magnetic tapes or other magnetic media, or any other media capable of reading instructions from a computer processor.
- various other forms of computer readable media may send instructions to a computer or carry the instructions, and include a router, a private or public network, or other wired and wireless transmission devices or channels.
- the instructions may include codes of any computer programming language, including C, C++, C language, Visual Basic, Java and JavaScript.
- the embodiments of the present application further provide a computer readable medium.
- Computer readable instructions are stored on the computer readable medium, and when executed, the computer readable instructions cause the processor to perform the foregoing method for improving the prediction accuracy of an AI model.
- the embodiments of the computer readable medium include a floppy disk, a hard disk, a magneto-optical disk, an optical disk (such as a CD-ROM, a CD-R, a CD-RW, a DVD-ROM, a DVD-RAM, a DVD-RW, a DVD+RW) , a magnetic tape, a non-volatile storage card, and a ROM.
- the computer readable instructions may be downloaded from a server computer or a cloud via a communication network.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
Acquiring a 3D model and a first group of rendering condition set; acquiring label information according to the 3D model; determining the first group of rendering condition set as a current group of rendering condition set; randomly determining combinations of a plurality of groups of rendering conditions from the current group of rendering condition set, and rendering same respectively to obtain corresponding 2D images, so as to form a first image data set; training the current AI model by means of the first image data set; evaluating a prediction performance of the trained current AI model to obtain a first accuracy; judging whether the first accuracy meets a preset condition; if not, changing at least one rendering condition in the current group of rendering condition set; returning back to the step of rendering to the step of judging until the first accuracy meets the preset condition.
Description
Embodiments of the present application mainly relate to the field of artificial intelligence (AI) models, and in particular to a method for improving the prediction accuracy of an AI model, an electronic device and a medium.
In recent years, training artificial intelligence models by generating synthetic image data has attracted more and more attention during machine learning studies. This is because the preparation of real image data is time-consuming and error-prone, and corresponding objects such as a car, a person, or a table need to be manually marked. However, there is still a certain gap between the synthetic image data and real image data, thereby making it impossible to ensure the performance of trained AI models.
SUMMARY
Embodiments of the present application provide a method for improving the prediction accuracy of an AI model, an electronic device and a medium. By means of the embodiments of the present application, relevant conditions for improving the prediction accuracy of the AI model can be accurately and rapidly obtained.
In a first aspect, a method for improving the prediction accuracy of an AI model is provided, which includes: acquiring a 3D model and a first group of rendering condition set, where the first group of rendering condition set includes a plurality of default rendering conditions and at least one default rendering condition in the plurality of default rendering conditions contains a first numerical interval; acquiring label information according to the 3D model; determining the first group of rendering condition set as a current group of rendering condition set; determining a first AI model as a current AI model; randomly determining combinations of a plurality of groups of rendering conditions from the current group of rendering condition set, and rendering same respectively to obtain corresponding 2D images, so as to form a first image data set, where each 2D image in the first image data set contains corresponding label information; training the current AI model by means of the first image
data set; evaluating a prediction performance of the trained current AI model to obtain a first accuracy; judging whether the first accuracy meets a preset condition; when the first accuracy does not meet the preset condition, changing at least one rendering condition in the current group of rendering condition set to obtain an updated current group of rendering condition set; returning back to the step of rendering same respectively to obtain corresponding 2D images, the step of training, the step of evaluating, and the step of judging until the first accuracy meets the preset condition; and outputting a rendering condition set corresponding to the first accuracy meeting the preset condition, and corresponding trained AI model information.
In a second aspect, an electronic device is provided, which includes: at least one memory, configured to store computer readable code; and at least one processor, configured to call the computer readable code to perform all steps in the method provided in the first aspect.
In a third aspect, a computer readable medium is provided; computer readable instructions are stored on the computer readable medium, and when executed, the computer readable instructions cause the processor to perform all the steps in the method provided in the first aspect.
In a fourth aspect, a computer program product is provided, which includes a computer program; and when executed by the processor, the computer program performs all the steps in the method provided in the first aspect.
The following accompanying drawings are merely intended to illustratively explain and interpret the embodiments of the present application, but are not intended to limit the scope of the present application, where:
FIG. 1 is a flowchart of a method for improving the prediction accuracy of an AI model according to an embodiment of the present application; and
FIG. 2 is a schematic diagram of an electronic apparatus according to an embodiment of the present application.
Reference Numerals
100: method for improving the prediction accuracy of an AI model
110-172: method steps
200: electronic device 201: memory 202: processor
The subject matter described herein will now be discussed with reference to exemplary implementations. It should be understood that these implementations are discussed only to enable those skilled in the art to better understand and implement the subject matter described herein, but do not limit the scope of protection set forth in the claims, applicability, or examples. The functions and arrangement of discussed elements can be changed, without departing from the scope of protection of the content of the embodiments of the present application. For each example, various processes or assemblies can be omitted, replaced, or added as needed. For example, the method described may be performed according to a sequence different from the sequence described, and each step may be added, omitted, or combined. Additionally, features described with respect to some examples may also be combined in other examples.
As used herein, the term “include” and its variations represent an open term, and mean “include but not limited to” . The term “based on” represents “based at least in part on” . The terms “one embodiment” and “an embodiment” represent “at least one embodiment” . The term “another embodiment” represents “at least one other embodiment” . The terms “first” , “second” and so on may refer to different or identical object. Other definitions, either clear or implicit, may also be included below. Unless otherwise indicated obviously from the context, the definition of one term is consistent throughout the specification.
The embodiments of the present application are further described below in detail below with reference to the accompanying drawings.
FIG. 1 is a flowchart of a method for improving the prediction accuracy of an AI model according to an embodiment of the present application. As shown in FIG. 1, the method 100 for improving the prediction accuracy of an AI model includes the following steps:
Step 110: A 3D model and a first group of rendering condition set are acquired. The first group of rendering condition set includes a plurality of default rendering conditions and at least one default rendering condition in the plurality of default rendering conditions contains a first numerical interval.
Optionally, at least one 3D model and the first group of rendering condition set may be acquired from a renderer. The first group of rendering condition set includes a type of a rendering engine, setting requirements of a camera, setting requirements of light, and setting requirements of a texture. Optionally, the first group of rendering condition set may also include: employing/or not employing image enhancement to obtain more image data, or specific means of the image enhancement, for example, rotation, scaling, random cropping, or adding noises.
Step 120: Label information is acquired according to the 3D model.
In an embodiment, in a case that the AI model used in step 130 is an object detection model, then acquiring label information is specifically to acquire the label information of the at least one 3D model and/or at least one component in the at least one 3D model. For example, there is a scene where a car runs on a viaduct, in a case that the AI model is required to detect car wheels therein, the acquiring label information in step 120 is to acquire the label information of the car wheels in a relevant rendering scene. The label information includes a bounding box of car wheel and a label name corresponding to the car wheel.
In an embodiment, in a case that the AI model used in step 130 is a classification model, the acquiring label information is to acquire the label information of a scene itself containing the 3D model in the renderer. For example, there is a scene with a cat and/or a dog, in a case that the cat or the dog needs to be identified by using the AI model, the acquiring label information in step 120 is to acquire the label information of a relevant rendering scene itself. In this example, the label information is a label name corresponding to the cat or the dog.
Step 130: The first group of rendering condition set is determined as a current group of rendering condition set. A first AI model is determined as a current AI model.
Step 140: Combinations of a plurality of groups of rendering conditions are randomly determined from the current group of rendering condition set, and same is rendered respectively to obtain corresponding 2D images, so as to form a first image data set. Each 2D image in the first image data set contains corresponding label information.
Optionally, N numerical values are randomly selected from a numerical interval of at least one rendering condition. The N numerical values are combined with other rendering conditions in the current group of rendering condition set, respectively, to determine the combinations of the plurality of groups of rendering conditions. Optionally, it is assumed that the current group of rendering condition set includes three rendering conditions A, B, and C, where the rendering condition A: [0°, 10°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] . The randomly determining combinations of a plurality of groups of rendering conditions from the current group of rendering condition set refers to randomly selecting corresponding numerical values from the rendering conditions with the numerical intervals in the current group of rendering condition set, and combining same with the rendering conditions without the numerical intervals respectively, so as to determine the combinations of the plurality of groups of rendering conditions. For example: one of the combinations of the plurality of groups of rendering conditions may be: the rendering condition A: [0°] ; the rendering condition B: [texture 2] , and the rendering
condition C: [light source 1] ; one of the combinations of the plurality of groups of rendering conditions may be: the rendering condition A: [1°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] , ..., and one of the combinations of the plurality of groups of rendering conditions may be: the rendering condition A: [10°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] .
Step 150: The current AI model is trained by means of the first image data set.
Optionally, the first image data set is divided into a first training set, a first validation set, and a first testing set. The current AI model may be trained by means of the first training set and the first validation set.
Step 160: A prediction performance of the trained current AI model is evaluated to obtain a first accuracy.
Optionally, the prediction performance of the trained current AI model is evaluated by means of the first testing set or a real image data set with the label information to obtain the first accuracy. Optionally, the prediction performance of the trained current AI model may be evaluated by means of a cloud AI evaluation module or a local server with strong computing performance.
Optionally, the prediction performance of the trained current AI model is evaluated by means of the first testing set or the real image data set with the label information to obtain an accuracy at a first stage. The trained current AI model is deployed to an edge device. The prediction performance of the trained current AI model is monitored on the edge device to obtain an accuracy at a second stage. Both the accuracy at the first stage and the accuracy at the second stage are taken as the first accuracy, that is, the first accuracy may include data at two stages.
Optionally, the trained current AI model is deployed to the edge device. The prediction performance of the trained current AI model is monitored to obtain the first accuracy. Optionally, the prediction performance of the trained current AI model is monitored by means of a machine vision system or a sensor or artificially. For example, it is assumed that the current AI model is to predict whether there is an object in a certain position, a sensor such as a radar or a distance measurement sensor may assist in determining whether there is an object in a relevant position, which is predicted by means of the current AI model, so as to evaluate whether the prediction of the current AI model is accurate.
In an embodiment, after the first accuracy is obtained, the current group of rendering condition set corresponding to the first accuracy and corresponding trained AI model information are inputted into a knowledge graph. Optionally, cases with relatively low
accuracy therein may also be inputted into the knowledge graph, with the purpose of being ready for further data analysis and capable of retraining the AI model.
Step 170: Whether the first accuracy meets a preset condition is judged.
In a case that the first accuracy does not meet the preset condition, step 171 is performed to change at least one rendering condition in the current group of rendering condition set to obtain an updated current group of rendering condition set. The method returns back to perform step 140 to step 170 until the first accuracy meets the preset condition.
Specifically, the at least one rendering condition in the current group of rendering condition set is changed within a preset range corresponding to each rendering condition in the current group of rendering condition set. Optionally, it may be randomly changed or may be manually changed by a user.
In an embodiment, it is assumed that each rendering condition in the current group of rendering condition set includes A, B, and C, the preset range of the rendering condition A is [0°, 100°] , the preset range of the rendering condition B is [texture 1, texture 2, texture 3] , and the preset range of the rendering condition C is [light source 1, light source 2] . It is assumed that the first group of rendering condition set includes: the rendering condition A: [0°, 10°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] , the updated current group of rendering condition set, for example: the rendering condition A: [20°, 30°] , the rendering condition B: [texture 2] , and the rendering condition C: [light source 1] , may be obtained by changing at least one rendering condition on the basis of the first group of rendering condition set within the preset range of each rendering condition.
Step 172: A rendering condition set corresponding to the first accuracy meeting the preset condition, and corresponding trained AI model information are outputted.
In an embodiment, the until the first accuracy meets the preset condition refers to until the first accuracy meets a highest accuracy within a preset number of cycles.
In an embodiment, after step 172, a second AI model is determined as the current AI model. The method returns back to perform step 140 to step 170 until the second accuracy meets the highest accuracy within the preset number of cycles. When K AI models are preset, the highest accuracy corresponding to each AI model is obtained, respectively. The rendering condition set corresponding to the highest accuracy and the corresponding trained AI model information are selected from all the accuracies corresponding to the K AI models for outputting.
In an embodiment, the AI model information may include an AI model name, hyper-parameters, and the like.
The embodiments of the present application may continuously improve the quality of a synthetic image data set through result orientation, specifically, continuously evaluate the prediction performance of the AI model after the synthetic image data set generated under different rendering conditions is trained on the AI model, and obtain the trained model information of the relevant AI model, so as to accurately and quickly obtain corresponding conditions for improving the prediction accuracy of the AI model.
The embodiments of the present application further provide an electronic device 200. FIG. 2 is a schematic diagram of a device control platform 200 according to an embodiment of the present application. As shown in FIG. 2, the device control platform 200 includes a processor 201 and a memory 202. Instructions are stored in the memory 202. When executed by the processor 201, the instructions implement the method 100 as described above.
The at least one processor 201 may include a microprocessor, an application specific integrated circuit (ASIC) , a digital signal processor (DSP) , a central processing unit (CPU) , a graphics processing unit (GPU) , a state machine, and the like. Embodiments of the computer readable medium include, but are not limited to, a floppy disk, a CD-ROM, a magnetic disk, a memory chip, a ROM, an RAM, an ASIC, a configured processor, an all-optical medium, all magnetic tapes or other magnetic media, or any other media capable of reading instructions from a computer processor. In addition, various other forms of computer readable media may send instructions to a computer or carry the instructions, and include a router, a private or public network, or other wired and wireless transmission devices or channels. The instructions may include codes of any computer programming language, including C, C++, C language, Visual Basic, Java and JavaScript.
In addition, the embodiments of the present application further provide a computer readable medium. Computer readable instructions are stored on the computer readable medium, and when executed, the computer readable instructions cause the processor to perform the foregoing method for improving the prediction accuracy of an AI model. The embodiments of the computer readable medium include a floppy disk, a hard disk, a magneto-optical disk, an optical disk (such as a CD-ROM, a CD-R, a CD-RW, a DVD-ROM, a DVD-RAM, a DVD-RW, a DVD+RW) , a magnetic tape, a non-volatile storage card, and a ROM. Optionally, the computer readable instructions may be downloaded from a server computer or a cloud via a communication network.
It should be noted that not all steps and modules in all flows and all system structure diagrams described above are necessary, and certain steps or modules may be neglected according to actual requirements. The sequence of performing each step is not permanent,
and may be adjusted according to requirements. System structures described in all the embodiments described above may be physical structures and may also be logical structures. That is, some modules may be implemented by the same physical entity. Alternatively, some modules may be implemented by a plurality of physical entities. Alternatively, some modules may be commonly implemented by some components in a plurality of independent devices.
Claims (15)
- A method for improving the prediction accuracy of an AI model, comprising:- acquiring (110) a 3D model and a first group of rendering condition set, wherein the first group of rendering condition set comprises a plurality of default rendering conditions, and at least one default rendering condition in the plurality of default rendering conditions contains a first numerical interval;- acquiring (120) label information according to the 3D model;- determining (130) the first group of rendering condition set as a current group of rendering condition set, and determining a first AI model as a current AI model;- randomly determining combinations of a plurality of groups of rendering conditions from the current group of rendering condition set, and rendering (140) same respectively to obtain corresponding 2D images, so as to form a first image data set, wherein each 2D image in the first image data set contains corresponding label information;- training (150) the current AI model by means of the first image data set;- evaluating (160) a prediction performance of the trained current AI model to obtain a first accuracy;- judging (170) whether the first accuracy meets a preset condition;- in a case that the first accuracy does not meet the preset condition, changing (171) at least one rendering condition in the current group of rendering condition set to obtain an updated current group of rendering condition set;- returning back to perform step 140 to step 170 until the first accuracy meets the preset condition; and- outputting (172) a rendering condition set corresponding to the first accuracy meeting the preset condition, and corresponding trained AI model information.
- The method according to claim 1, wherein the training (150) the current AI model by means of the first image data set comprises:- dividing the first image data set into a first training set, a first validation set, and a first testing set; and- training the current AI model by means of the first training set and the first validation set.
- The method according to claim 2, wherein the evaluating (160) a prediction performance of the trained current AI model to obtain a first accuracy comprises:- evaluating the prediction performance of the trained current AI model by means of the first testing set or a real image data set with the label information to obtain the first accuracy.
- The method according to claim 2, wherein the evaluating (160) a prediction performance of the trained current AI model to obtain a first accuracy comprises:- evaluating the prediction performance of the trained current AI model by means of the first testing set or the real image data set with the label information to obtain an accuracy at a first stage;- deploying the trained current AI model to an edge device;- monitoring the prediction performance of the trained current AI model on the edge device to obtain an accuracy at a second stage; and- taking both the accuracy at the first stage and the accuracy at the second stage as the first accuracy.
- The method according to claim 1, wherein the evaluating (160) a prediction performance of the trained current AI model to obtain a first accuracy comprises:- deploying the trained current AI model to an edge device; and- monitoring the prediction performance of the trained current AI model to obtain the first accuracy.
- The method according to claim 5, wherein the monitoring the prediction performance of the trained current AI model comprises:- monitoring the prediction performance of the trained current AI model by means of a machine vision system or a sensor or artificially.
- The method according to claim 1, wherein after the first accuracy is obtained, the method further comprises:- inputting the current group of rendering condition set corresponding to the first accuracy, and corresponding trained AI model information into a knowledge graph.
- The method according to claim 1, wherein the randomly determining combinations of a plurality of groups of rendering conditions from the current group of rendering condition set comprises:- randomly selecting N numerical values from a numerical interval of at least one rendering condition; and- combining the N numerical values with other rendering conditions in the current group of rendering condition set, respectively, to determine the combinations of the plurality of groups of rendering conditions.
- The method according to claim 1 or 7, wherein the first group of rendering condition set comprises:- a type of a rendering engine, setting requirements of a camera, setting requirements of light, and setting requirements of a texture.
- The method according to claim 1, wherein the changing (171) at least one rendering condition in the current group of rendering condition set comprises:- changing at least one rendering condition in the current group of rendering condition set within a preset range corresponding to each rendering condition in the current group of rendering condition set.
- The method according to claim 1, wherein- the until the first accuracy meets the preset condition comprises:- until the first accuracy meets a highest accuracy within a preset number of cycles;- after the outputting (172) a rendering condition set corresponding to the first accuracy meeting the preset condition, and corresponding trained AI model information, the method further comprises:- determining a second AI model as a current AI model; and- returning back to perform step 140 to step 170 until the second accuracy meets the highest accuracy within the preset number of cycles;- when K AI models are preset, obtaining the highest accuracy corresponding to each of the AI models, respectively; and- selecting a rendering condition set corresponding to the highest accuracy and the corresponding trained AI model information from all the accuracies corresponding to the K AI models for outputting.
- The method according to claim 1 or 11, wherein the AI model information comprises: an AI model name and hyper-parameters.
- An electronic device, comprising:a processor (201) ;a memory (202) , configured to store executable instructions of the processor (201) ;wherein the processor (201) is configured to read the executable instructions from the memory (202) and execute the executable instructions to implement the method for improving the prediction accuracy of an AI model according to any one of claims 1 to 12.
- A computer readable storage medium on which computer instructions are stored, wherein when executed by the processor, the computer instructions implement the method for improving the prediction accuracy of an AI model according to any one of claims 1 to 12.
- A computer program product, comprising a computer program, wherein when executed by the processor, the computer program implements the method for improving the prediction accuracy of an AI model according to any one of claims 1 to 12.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2023/077710 WO2024174136A1 (en) | 2023-02-22 | 2023-02-22 | Method for improving prediction accuracy of ai model, electronic device and medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2023/077710 WO2024174136A1 (en) | 2023-02-22 | 2023-02-22 | Method for improving prediction accuracy of ai model, electronic device and medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2024174136A1 true WO2024174136A1 (en) | 2024-08-29 |
Family
ID=92500104
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2023/077710 WO2024174136A1 (en) | 2023-02-22 | 2023-02-22 | Method for improving prediction accuracy of ai model, electronic device and medium |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2024174136A1 (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115004701A (en) * | 2019-12-31 | 2022-09-02 | 宝缇嘉工作室有限公司 | System and method for dynamic image virtualization |
US20220398872A1 (en) * | 2021-06-15 | 2022-12-15 | Microsoft Technology Licensing, Llc | Generation and management of notifications providing data associated with activity determinations pertaining to a vehicle |
WO2023014369A1 (en) * | 2021-08-06 | 2023-02-09 | Siemens Corporation | Synthetic dataset creation for object detection and classification with deep learning |
-
2023
- 2023-02-22 WO PCT/CN2023/077710 patent/WO2024174136A1/en unknown
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115004701A (en) * | 2019-12-31 | 2022-09-02 | 宝缇嘉工作室有限公司 | System and method for dynamic image virtualization |
US20220398872A1 (en) * | 2021-06-15 | 2022-12-15 | Microsoft Technology Licensing, Llc | Generation and management of notifications providing data associated with activity determinations pertaining to a vehicle |
WO2023014369A1 (en) * | 2021-08-06 | 2023-02-09 | Siemens Corporation | Synthetic dataset creation for object detection and classification with deep learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110751678A (en) | Moving object detection method and device and electronic equipment | |
JP2015062121A (en) | System and method of alerting driver that visual perception of pedestrian may be difficult | |
US20210397907A1 (en) | Methods and Systems for Object Detection | |
JP6653361B2 (en) | Road marking image processing apparatus, road marking image processing method, and road marking image processing program | |
CN113030990B (en) | Fusion ranging method, device, ranging equipment and medium for vehicle | |
JP6698191B1 (en) | Road marking failure detection device, road marking failure detection method, and road marking failure detection program | |
CN111753729B (en) | False face detection method and device, electronic equipment and storage medium | |
CN114120071A (en) | Detection method of image with object labeling frame | |
WO2024174136A1 (en) | Method for improving prediction accuracy of ai model, electronic device and medium | |
Philipp et al. | Functional decomposition of automated driving systems for the classification and evaluation of perceptual threats | |
Ngo | A methodology for validation of a radar simulation for virtual testing of autonomous driving | |
CN115713750A (en) | Lane line detection method and device, electronic equipment and storage medium | |
US20230252658A1 (en) | Depth map completion in visual content using semantic and three-dimensional information | |
Wagener et al. | Automated scenario generation for testing advanced driver assistance systems based on post-processed reference laser scanner data | |
CN115482352A (en) | Method and apparatus for training machine learning algorithms | |
CN115471725A (en) | Image processing method and device for anomaly detection and neural network construction method | |
CN112560853A (en) | Image processing method, device and storage medium | |
CN113989774A (en) | Traffic light detection method and device, vehicle and readable storage medium | |
CN113139072A (en) | Data labeling method and device and electronic equipment | |
CN112837326A (en) | Remnant detection method, device and equipment | |
US10896333B2 (en) | Method and device for aiding the navigation of a vehicle | |
US20230410532A1 (en) | Object detection device, monitoring device, training device, and model generation method | |
US20240095945A1 (en) | Method for Uncertainty Estimation in Object Detection Models | |
US20230351741A1 (en) | Method and device for training a neural network | |
US20220402504A1 (en) | Methods and Systems for Generating Ground Truth Data |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 23923317 Country of ref document: EP Kind code of ref document: A1 |