CN113762394B - Blasting block prediction method - Google Patents

Blasting block prediction method Download PDF

Info

Publication number
CN113762394B
CN113762394B CN202111054833.2A CN202111054833A CN113762394B CN 113762394 B CN113762394 B CN 113762394B CN 202111054833 A CN202111054833 A CN 202111054833A CN 113762394 B CN113762394 B CN 113762394B
Authority
CN
China
Prior art keywords
random forest
model
training
layer
blasting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111054833.2A
Other languages
Chinese (zh)
Other versions
CN113762394A (en
Inventor
李祥龙
赵品喆
姚永鑫
方程
武永博
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Kunming University of Science and Technology
Original Assignee
Kunming University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Kunming University of Science and Technology filed Critical Kunming University of Science and Technology
Priority to CN202111054833.2A priority Critical patent/CN113762394B/en
Publication of CN113762394A publication Critical patent/CN113762394A/en
Application granted granted Critical
Publication of CN113762394B publication Critical patent/CN113762394B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/02Agriculture; Fishing; Forestry; Mining

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Strategic Management (AREA)
  • Human Resources & Organizations (AREA)
  • Economics (AREA)
  • Evolutionary Computation (AREA)
  • General Business, Economics & Management (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Tourism & Hospitality (AREA)
  • Artificial Intelligence (AREA)
  • Marketing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Agronomy & Crop Science (AREA)
  • Health & Medical Sciences (AREA)
  • Mining & Mineral Resources (AREA)
  • Primary Health Care (AREA)
  • Marine Sciences & Fisheries (AREA)
  • Animal Husbandry (AREA)
  • General Health & Medical Sciences (AREA)
  • Development Economics (AREA)
  • Game Theory and Decision Science (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to a blasting block size prediction method, and belongs to the technical field of blasting block size prediction. Firstly, constructing a CART decision tree model; then resampling the training sample set; calculating the relation between different decision tree trees (Ntree) and model Mean Square Error (MSE) by utilizing MATLAB script language writing codes; then adopting GINI coefficients as attribute splitting criteria for the decision tree, and selecting nodes based on the criteria; based on the training set of the nodes, different decision trees are obtained, and a first layer of random forest is finally generated; generating a second layer random forest according to the training residual error of the first layer random forest as a training sample set of the second layer random forest; and sequentially superposing the outputs of the first layer random forest model and the second layer random forest model to obtain the final output of the double-layer random forest prediction model, and predicting the blasting block according to the final output. Compared with the traditional prediction method, the method has higher reliability and stability.

Description

Blasting block prediction method
Technical Field
The invention relates to a blasting block size prediction method, and belongs to the technical field of blasting block size prediction.
Background
The drilling and blasting method is one of the rock breaking methods frequently adopted in mining and tunnel construction, and the blasting effect brings about a plurality of negative effects such as blasting vibration, air shock waves, flying stones, noise, dust and the like while breaking the rock, wherein the blasting vibration hazard is particularly remarkable. The explosion regulations in China adopt the speed to measure the vibration intensity, accurately predict the particle vibration speed caused by explosion and effectively control the harm of explosion vibration.
The accuracy and reliability of the blasting block prediction method in the prior art are low.
Disclosure of Invention
The invention aims to solve the technical problems that: the blasting block degree prediction method is provided, the problem of inaccurate and fuzzy information processing which needs to consider a plurality of factors and conditions simultaneously is solved, and the accuracy and reliability of blasting block degree prediction are improved.
The technical scheme adopted by the invention is as follows: a blasting block prediction method, comprising the steps of:
1) Forming a training data set;
2) Selecting a split node;
3) Generating a first layer of random forest;
4) Generating a second layer of random forest;
5) And sequentially superposing the outputs of the first layer of random forest and the second layer of random forest to obtain the final output of the double-layer random forest prediction model.
Specifically, the step 1) includes the steps of:
Step1.1: taking the blasting average block size value actually measured in a blasting test as an output variable (label value) of a model, acquiring a resistance line distance (B), a drilling row distance (S), a step height (H), a drilling diameter (d), a blocking length (L), a rock mass elastic modulus (E), an in-situ rock block size (x) and a blasting unit consumption (q) index as input variables (attribute values) of the model by corresponding blasting test occasions, and forming a training data set by the output variables and the input variables
Step1.2: the data set obtained by using Step1.1 is subjected to resampling of training sample set by adopting Bootstrap method, m data samples are extracted, and n training data sets are randomly generatedD= { x i1,xi2,…,xin,yi } (i e1, m).
Based on a Bootstrap data set, a Bagging method is adopted to randomly select training data with a put back, then a classifier is constructed, and finally the overall acquisition effect is increased by combining the learned models.
Step1.3: using a decision tree algorithm in each subset, selecting an optimal mode to split nodes according to a 'minimum coefficient criterion', and setting the prediction result of a single decision tree predictor f (x, theta k) as f i (x) without pruning in the splitting process, wherein the final prediction result of a random forest regression model is expressed as that a parameter set RFP of a modeling process of the random forest regression algorithm is defined in a formula (1); equation (2) is used to make a prediction of blasting block based on random forest regression.
RFP={Ntree,Mtry} (1)
Where x represents the input vector, θ k is the vector representing the generation of each tree growth path, and F (x) represents the predicted blast block. Ntree is the number of decision trees in the model, and Ntree values affect the training degree and accuracy of the random forest model. In order to obtain the best evaluation result, codes are written by using MATLAB script language, and the relation between different decision tree trees (Ntree) and the model Mean Square Error (MSE) is simulated and calculated. Mtry is the number of features randomly extracted from the features. The Mtry value controls the disturbance degree of the random forest model attribute and is an important parameter in the model. The Mtry value directly affects the accuracy of the model. The Mtry values were calculated according to the following empirical formula:
Mtry&=[log2M] (3)
Mtry&=[M/3 (4)
wherein: m is the number of model input parameters; [] Representing a rounding down operation.
Specifically, the step 2) includes the steps of:
step2.1: a binary tree form decision tree is adopted, and a binary recursion is utilized to divide the data space into different subsets. In classification, assuming that there are K classes, the base index of the probability distribution is:
where p represents the probability of the sample point and p k represents the probability that the sample point belongs to the K-th class.
Step2.2: the decision tree uses GINI coefficients as criteria for attribute splitting. The feature with the lowest coefficient of the base is selected as the root node. And so on, selecting the leaf node with the smallest characteristic coefficient.
Step2.3: from the training dataset, starting from the root node, step 3 is recursively performed for each node, constructing a binary decision tree.
Specifically, the step 3) includes the steps of:
step3.1: the subset D is divided into two parts D 1 and D 2 according to whether the feature a takes a certain possible value a, D 1={(x,y)∈D|A(x)=a},D2=D-D1, and under the condition of the feature a, the base index of the set D is defined as:
where D 1 represents the training set of nodes, and the Basil index of the existing feature to the dataset is calculated.
Step3.2: for each feature a, for each possible value a, the base index at a=a is calculated by dividing D into two parts D 1 and D 2, depending on whether the test of the sample point pair a=a is yes or no. Among all possible features a and all possible cut points a, the cut point with the smallest keni index is selected as the cut point. Generating two child nodes from the current segmentation point according to the optimal characteristics and the optimal segmentation point, and distributing the training data set to two leaf nodes according to the characteristics.
Step3.3: step2.3 to Step3.2 are invoked recursively for both leaf nodes.
Step3.4: repeating the steps 3.1 to 3.3 to obtain different decision trees, recursively branching and growing each decision tree from top to bottom, stopping growing the regression tree after the segmentation termination condition is met, and finally combining all the regression trees together to form a random forest.
Step3.5: the input vector X in the prediction set X is input into a first layer random forest, and arithmetic average is carried out on the predicted value y i of the single regression tree T i by using a simple average method
Specifically, the step 4) includes the steps of:
step4.1: calculating the output of a layer 1 random forest on a training sample set Subtracting the output value/>, from the actual value y i of the training sampleTraining residual error/>, can be obtainedTraining residual/>Substituting the original training sample to construct a new data set as a training sample of a layer 2 random forest, so that the input is x i, and the expected output is/>
Step4.2: because the training residual value obtained by Step4.1 is too small, the training residual needs to be normalized:
Wherein: l in is the original input data; l min is the minimum of the same kind of data in the original input data; l max is the maximum value of the same kind of data in the original input data; l is the input data after normalization processing. And after normalization treatment, a training data set of the layer 2 random forest is formed.
Step4.3: and (3) establishing a layer 2 random forest model according to Step1.1 to Step4.1 to obtain a prediction result of the layer 2 random forest model.
The beneficial effects of the invention are as follows: the invention is particularly suitable for dealing with imprecise and ambiguous information processing problems requiring simultaneous consideration of many factors and conditions, and has higher reliability and stability than conventional prediction methods.
Drawings
FIG. 1 is a flowchart of a blast block predictive value calculation based on random forest regression;
FIG. 2 is a schematic diagram of a step blasting parameter;
FIG. 3 is a plot of a matrix of raw data set distribution points;
FIG. 4 is a comparison of measured and predicted blocking values for a burst test;
FIG. 5 shows the relative error between measured and predicted block sizes for a burst test.
Detailed Description
The invention will be further described with reference to the drawings and the specific embodiments.
Example 1: the random forest is used as an advanced algorithm tool, is particularly suitable for processing the imprecise and fuzzy information processing problem which needs to consider a plurality of factors and conditions at the same time, introduces a random forest regression method into the blasting vibration speed prediction, and compares the result of the random forest model prediction with the result of the traditional method prediction so as to provide an optimal prediction method for blasting block monitoring.
As shown in fig. 1, a blasting block prediction method includes the following steps:
1) Forming a training data set;
2) Selecting a split node;
3) Generating a first layer of random forest;
4) Generating a second layer of random forest;
5) And sequentially superposing the outputs of the first layer of random forest and the second layer of random forest to obtain the final output of the double-layer random forest prediction model.
The specific implementation process is as follows: in order to solve the problems of low model prediction precision, poor model generalization capability and the like in the research on the prediction of the block degree of blasting materials at the present stage, the block degree of the pile is difficult to accurately control, and the block degree meets the dam-loading conditions of the pile for blasting mining. Aiming at the defects of the current blasting prediction model, the blasting block size of the rock-fill dam material is effectively controlled, and the blasting block size prediction model is established based on a random forest regression method.
Further, the step 1) includes the steps of:
Step1.1: taking the blasting average block size value actually measured in a blasting test as an output variable (label value) of a model, acquiring a resistance line distance (B), a drilling row distance (S), a step height (H), a drilling diameter (d), a blocking length (L), a rock mass elastic modulus (E), an in-situ rock block size (x) and a blasting unit consumption (q) index as input variables (attribute values) of the model by corresponding blasting test occasions, and forming a training data set by the output variables and the input variables The step blasting parameter is schematically shown in fig. 2.
Step1.2: the data set obtained by using Step1.1 is subjected to resampling of training sample set by adopting Bootstrap method, m data samples are extracted, and n training data sets are randomly generatedD= { x i1,xi2,…,xin,yi } (i e1, m).
Based on a Bootstrap data set, a Bagging method is adopted to randomly select training data with a put back, then a classifier is constructed, and finally the overall acquisition effect is increased by combining the learned models.
Step1.3: using a decision tree algorithm in each subset, selecting an optimal mode to split nodes according to a 'minimum coefficient criterion', and setting the prediction result of a single decision tree predictor f (x, theta k) as f i (x) without pruning in the splitting process, wherein the final prediction result of a random forest regression model is expressed as that a parameter set RFP of a modeling process of the random forest regression algorithm is defined in a formula (1); equation (2) is used to make a prediction of blasting block based on random forest regression.
RFP={Ntree,Mtry} (1)
Where x represents the input vector, θ k is the vector representing the generation of each tree growth path, and F (x) represents the predicted blast block. Ntree is the number of decision trees in the model, and Ntree values affect the training degree and accuracy of the random forest model. In order to obtain the best evaluation result, codes are written by using MATLAB script language, and the relation between different decision tree trees (Ntree) and the model Mean Square Error (MSE) is simulated and calculated. Mtry is the number of features randomly extracted from the features. The Mtry value controls the disturbance degree of the random forest model attribute and is an important parameter in the model. The Mtry value directly affects the accuracy of the model. The Mtry values were calculated according to the following empirical formula:
Mtry&=[log2M] (3)
Mtry&=[M/3 (4)
wherein: m is the number of model input parameters; [] Representing a rounding down operation.
Further, the step 2) includes the steps of:
step2.1: a binary tree form decision tree is adopted, and a binary recursion is utilized to divide the data space into different subsets. In classification, assuming that there are K classes, the base index of the probability distribution is:
where p represents the probability of the sample point and p k represents the probability that the sample point belongs to the K-th class.
Step2.2: the decision tree uses GINI coefficients as criteria for attribute splitting. The feature with the lowest coefficient of the base is selected as the root node. And so on, selecting the leaf node with the smallest characteristic coefficient.
Step2.3: from the training dataset, starting from the root node, step 3 is recursively performed for each node, constructing a binary decision tree.
Further, the step 3) includes the steps of:
step3.1: the subset D is divided into two parts D 1 and D 2 according to whether the feature a takes a certain possible value a, D 1={(x,y)∈D|A(x)=a},D2=D-D1, and under the condition of the feature a, the base index of the set D is defined as:
where D 1 represents the training set of nodes, and the Basil index of the existing feature to the dataset is calculated.
Step3.2: for each feature a, for each possible value a, the base index at a=a is calculated by dividing D into two parts D 1 and D 2, depending on whether the test of the sample point pair a=a is yes or no. Among all possible features a and all possible cut points a, the cut point with the smallest keni index is selected as the cut point. Generating two child nodes from the current segmentation point according to the optimal characteristics and the optimal segmentation point, and distributing the training data set to two leaf nodes according to the characteristics.
Step3.3: step2.3 to Step3.2 are invoked recursively for both leaf nodes.
Step3.4: repeating the steps 3.1 to 3.3 to obtain different decision trees, recursively branching and growing each decision tree from top to bottom, stopping growing the regression tree after the segmentation termination condition is met, and finally combining all the regression trees together to form a random forest.
Step3.5: the input vector X in the prediction set X is input into a first layer random forest, and arithmetic average is carried out on the predicted value y i of the single regression tree by using a simple average method
Further, the step 4) includes the steps of:
step4.1: calculating the output of a layer 1 random forest on a training sample set Subtracting the output value/>, from the actual value y i of the training sampleTraining residual error/>, can be obtainedTraining residual/>Substituting the original training sample to construct a new data set as a training sample of a layer 2 random forest, so that the input is x i, and the expected output is/>
Step4.2: because the training residual value obtained by Step4.1 is too small, the training residual needs to be normalized:
Wherein: l in is the original input data; l min is the minimum of the same kind of data in the original input data; l max is the maximum value of the same kind of data in the original input data; l is the input data after normalization processing. And after normalization treatment, a training data set of the layer 2 random forest is formed.
Step4.3: and (3) establishing a layer 2 random forest model according to Step1.1 to Step4.1 to obtain a prediction result of the layer 2 random forest model.
Further, according to the statistical analysis result of the training sample set, the Range (Range), average value (Mean) and standard deviation (std. Dev) indexes of each modeling parameter are respectively given in table 1. As can be seen from fig. 2, the obvious correlation between the parameters is not shown in the figure, which indicates that the selected parameters have independence.
TABLE 1 random forest regression model training sample input and output parameter descriptive statistics
Case:
In order to further illustrate the accuracy and reliability of the method of the present invention, a step blasting experiment is taken as an example. The median block sizes of each field are respectively 0.200m,0.135m,0.105m,0.115m and 0.122m through a screening test after the field blasting test and the actual measurement of the blasting block size grading curve of each field; and calculating according to the explosion design parameters, the rock information parameters and the explosive information parameters adopted in the field explosion test to obtain the input parameters of the field explosion test model, wherein the input parameters are shown in table 2.
Table 2 in-situ burst test model input parameters
According to the established blasting block prediction system based on the RF-R method, 10 simulation prediction calculations are respectively carried out on the 5 groups of blasting tests of the experiment, and finally the prediction block average value of each test field is obtained, and the result is shown in Table 3.
TABLE 3 simulation experiment prediction blockiness results
Through calculation, the relative errors of the predicted block value and the actually measured block value of each blasting test field are respectively as follows: 0.60%, 1.04%, 12.76%, 2.63%, 2.80%, wherein the maximum error is 12.76%, the minimum error is 0.60%, and the average error is 3.96%.
The relative error between the predicted and measured block values for each burst test run is shown in fig. 5, for example, in fig. 4.
In conclusion, the explosion block prediction precision based on the RF-R model is higher than that of other prediction models, and the maximum relative error aiming at the experiment is not more than 15%, so that the method meets the engineering requirements. Therefore, by means of the quantitative relation between the blasting parameters and the blasting block sizes, the blasting parameters can be properly adjusted according to the blasting block sizes predicted by the RF-R model and the system and the actual conditions so as to meet the dam feeding block size requirements.
While the present invention has been described in detail with reference to the drawings, the present invention is not limited to the above embodiments, and various changes can be made without departing from the spirit of the present invention within the knowledge of those skilled in the art.

Claims (1)

1. A blasting block prediction method is characterized in that: the method comprises the following steps:
1) Forming a training data set;
2) Selecting a split node;
3) Generating a first layer of random forest;
4) Generating a second layer of random forest;
5) Sequentially superposing the output of the first layer random forest and the output of the second layer random forest to obtain the final output of the double-layer random forest prediction model;
the step 1) comprises the following steps:
Step1.1: taking the blasting average block value actually measured by a blasting test as an output variable of a model, acquiring a resistance line distance, a drilling row distance, a step height, a drilling diameter, a blocking length, a rock mass elastic modulus, an in-situ rock block and a blasting unit consumption index of a corresponding blasting test field as input variables of the model, and forming a training data set by the output variables and the input variables;
step1.2: the data set obtained by using Step1.1 is subjected to resampling of training sample set by adopting Bootstrap method, m data samples are extracted, and n training data sets are randomly generated D= { x b1,xb2,…,xbn,yb } (b e1, m ]);
based on a Bootstrap data set, a Bagging method is adopted to randomly select training data with a put back, then a classifier is constructed, and finally the overall acquisition effect is increased by combining the learned models;
Step1.3: and (3) using a decision tree algorithm in each subset, selecting an optimal mode to split nodes according to a minimum coefficient criterion, and if the prediction result of a single decision tree predictor f (x, theta k) is f i (x) in the splitting process, the final prediction result of a random forest regression model is expressed as: equation (1) defines a random forest regression algorithm modeling process parameter set RFP; the formula (2) is used for carrying out blasting block prediction based on random forest regression;
RFP={Ntree,Mtry} (1)
wherein x represents an input vector, θ k represents a vector for generating a growth path of each tree, F (x) represents predicted blasting block, ntree represents a decision tree number in a model, a MATLAB script language is used for writing codes and carrying out relation simulation calculation between different decision tree numbers Ntree and model mean square errors MSEs, mtry represents feature numbers randomly extracted from features, mtry values control disturbance degrees of random forest model attributes, and Mtry values are calculated according to the following empirical formula:
Mtry&=[log2M] (3)
Mtry&=[M/3] (4)
wherein: m is the number of model input parameters; [] Representing a downward rounding operation;
said step 2) comprises the steps of:
Step2.1: adopting a binary tree form decision tree, continuously dividing a data space into different subsets by utilizing binary recursion, and assuming K classes in classification, the radix index of probability distribution is:
wherein p represents the probability of the sample point, and p j represents the probability that the sample point belongs to the K-th class;
Step2.2: the decision tree adopts GINI coefficients as the attribute splitting standard, selects the feature with the lowest coefficient of the key as the root node, and so on, and selects the leaf node with the smallest coefficient of the rest features;
step2.3: recursively performing step 3) on each node from the root node according to the training data set, and constructing a binary decision tree;
said step 3) comprises the steps of:
Step3.1: the subset D is divided into two parts according to whether the feature A takes a certain value a;
Step3.2: for each feature A, taking a value a of each feature A, dividing D into two parts of D 1 and D 2 according to whether the test of the sample point pair A=a is yes or no, and calculating a base index when A=a; selecting a segmentation point with the minimum base index from all the features A and all the segmentation points a as a segmentation point, generating two child nodes from the current segmentation point according to the optimal features and the optimal segmentation point, and distributing a training data set into two leaf nodes according to the features;
step3.3: recursively invoking step2.3 to step3.2 for both leaf nodes;
Step3.4: repeating the steps 3.1 to 3.3 to obtain different decision trees, recursively branching and growing each decision tree from top to bottom, stopping growing the regression tree after the segmentation termination condition is met, and finally combining all the regression trees together to form a random forest;
step3.5: inputting an input vector X in a prediction set X into a first layer random forest, and calculating an arithmetic average value of predicted values of a single regression tree T i by using a simple average method;
The step 4) comprises the following steps:
Step4.1: calculating the output of a first layer of random forest on a training sample set Subtracting the output value/>, from the actual value y i of the training sampleTraining residual error/>, can be obtainedTraining residual/>Substituting the original training sample to construct a new data set as a training sample of a second layer random forest, so that the input is x i, and the expected output is/>
Step4.2: because the training residual value obtained by Step4.1 is too small, the training residual needs to be normalized:
Wherein: l in is the original input data; l min is the minimum of the same kind of data in the original input data; l max is the maximum value of the same kind of data in the original input data; l is input data after normalization processing, and a training data set of a second layer random forest is formed after normalization processing;
Step4.3: and establishing a second layer random forest model according to Step1.1 to Step4.1 to obtain a prediction result of the second layer random forest model.
CN202111054833.2A 2021-09-09 2021-09-09 Blasting block prediction method Active CN113762394B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111054833.2A CN113762394B (en) 2021-09-09 2021-09-09 Blasting block prediction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111054833.2A CN113762394B (en) 2021-09-09 2021-09-09 Blasting block prediction method

Publications (2)

Publication Number Publication Date
CN113762394A CN113762394A (en) 2021-12-07
CN113762394B true CN113762394B (en) 2024-04-26

Family

ID=78794212

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111054833.2A Active CN113762394B (en) 2021-09-09 2021-09-09 Blasting block prediction method

Country Status (1)

Country Link
CN (1) CN113762394B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115659263B (en) * 2022-10-14 2023-08-08 长江三峡通航管理局 Ship control behavior risk assessment system and method based on big data

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105574544A (en) * 2015-12-16 2016-05-11 平安科技(深圳)有限公司 Data processing method and device
CN109146195A (en) * 2018-09-06 2019-01-04 北方爆破科技有限公司 A kind of blast fragmentation size prediction technique based on cart tree regression algorithm

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10885259B2 (en) * 2019-08-30 2021-01-05 Intel Corporation Random forest model for prediction of chip layout attributes

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105574544A (en) * 2015-12-16 2016-05-11 平安科技(深圳)有限公司 Data processing method and device
CN109146195A (en) * 2018-09-06 2019-01-04 北方爆破科技有限公司 A kind of blast fragmentation size prediction technique based on cart tree regression algorithm

Also Published As

Publication number Publication date
CN113762394A (en) 2021-12-07

Similar Documents

Publication Publication Date Title
WO2020125668A1 (en) Method and system for automatically identifying surrounding rock level by applying while-drilling parameters
CN110442666B (en) Mineral resource prediction method and system based on neural network model
CN112529341B (en) Drilling well leakage probability prediction method based on naive Bayesian algorithm
Arbin et al. Comparative analysis between k-means and k-medoids for statistical clustering
CN110135167B (en) Edge computing terminal security level evaluation method for random forest
CN112989708B (en) Well logging lithology identification method and system based on LSTM neural network
CN109409647A (en) A kind of analysis method of the salary level influence factor based on random forests algorithm
CN103617147A (en) Method for identifying mine water-inrush source
CN108280289B (en) Rock burst danger level prediction method based on local weighted C4.5 algorithm
CN110344824B (en) Acoustic curve generation method based on random forest regression
CN113705110B (en) Blasting vibration speed prediction method based on double random forest regression method
CN113221960B (en) Construction method and collection method of high-quality vulnerability data collection model
CN113762394B (en) Blasting block prediction method
CN114662405A (en) Rock burst prediction method based on few-sample measurement and ensemble learning
CN104915679A (en) Large-scale high-dimensional data classification method based on random forest weighted distance
Chu et al. Co-training based on semi-supervised ensemble classification approach for multi-label data stream
CN105740521A (en) Small grid removing method and apparatus in oil reservoir numerical simulation system solution process
CN114091333A (en) Shale gas content artificial intelligence prediction method based on machine learning
CN102621953B (en) Automatic online quality monitoring and prediction model updating method for rubber hardness
CN117541095A (en) Agricultural land soil environment quality classification method
CN115809601A (en) Sedimentary rock structure background distinguishing method
CN118133104A (en) Rapid identification method for lithofacies of deep sea-phase shale gas well
CN111985782A (en) Automatic tramcar driving risk assessment method based on environment perception
CN114372640A (en) Wind power prediction method based on fluctuation sequence classification correction
CN111125082B (en) Cement process parameter data analysis method based on association rule mining

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant