CN115049026A - Regression analysis method of space non-stationarity relation based on GSNNR - Google Patents

Regression analysis method of space non-stationarity relation based on GSNNR Download PDF

Info

Publication number
CN115049026A
CN115049026A CN202210984054.0A CN202210984054A CN115049026A CN 115049026 A CN115049026 A CN 115049026A CN 202210984054 A CN202210984054 A CN 202210984054A CN 115049026 A CN115049026 A CN 115049026A
Authority
CN
China
Prior art keywords
attribute
distance
space
model
spatial
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210984054.0A
Other languages
Chinese (zh)
Inventor
倪巳涵
王中一
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ocean University of China
Original Assignee
Ocean University of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ocean University of China filed Critical Ocean University of China
Priority to CN202210984054.0A priority Critical patent/CN115049026A/en
Publication of CN115049026A publication Critical patent/CN115049026A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a regression analysis method of spatial non-stationarity relation based on GSNNR, belonging to the technical field of combination of deep learning and spatial analysis. The method comprises the following steps: collecting spatial information data; inputting the spatial features and the attribute spatial features into a full-space adjacent nonlinear fusion neural network model, wherein the SAPDNN neural network model takes GNNWR as a basic model, and the attribute spatial features are added into an input layer; obtaining a full-space adjacent expression matrix through the SAPDNN neural network model; and inputting the full-space adjacent expression matrix into an SWNN module for processing, and outputting a weight matrix. The invention introduces the attribute space as an important characteristic for analyzing the non-stationarity process, provides the full-space expression of the fusion space and the attribute space, fuses the geographic space and the attribute space by using the deep neural network, and further improves the accuracy of the measurement of the non-stationarity.

Description

Regression analysis method of space non-stationarity relation based on GSNNR
Technical Field
The invention belongs to the technical field of combination of deep learning and spatial analysis, and particularly relates to a regression analysis method based on a spatial non-stationarity relation of improved GSNNR.
Background
In the field of spatial analysis, analysis of non-stationarity is very critical, and generally, a mathematical model is used for analyzing a non-stationarity relation of a corresponding space in analysis and prediction. The accuracy of the measure for the non-stationarity has become a core evaluation method of the geospatial analysis model.
Geoneural network weighted regression (GNNWR) is a more advanced model structure in the field of geospatial non-stationarity analysis. The model adopts the deep neural network to replace a kernel function for nonlinear fitting in a classical GWR model, and overcomes the defect that the kernel function cannot fit complex nonlinear mapping. In the GNNWR model, the strong nonlinear fitting capability of an advanced deep neural network is utilized, and a nonlinear mapping process from original geographic space data to high-dimensional space hidden feature data is fitted by constructing a Spatial Weighted Neural Network (SWNN). Firstly, calculating the geographical spatial position distances between a plurality of sample points and a point to be estimated to obtain a spatial distance matrix between an unknown point to be estimated and a plurality of known sample points. And then inputting the spatial distance matrix into the SWNN, carrying out nonlinear mapping on the original data in a high-dimensional space by the deep neural network, and obtaining a corresponding spatial weight matrix through learning the data. And finally, taking the space weight matrix as the input of a linear regression model to obtain a final fitting value.
The mathematical modeling of the GNNWR for the non-stationary process is limited to a single geospatial domain, and only the distance characteristic between the sample point and the estimation point is considered. And the non-stationarity of the space in reality is also influenced by the attribute, and the GNNWR is not fully considered on the representation of the non-stationarity data feature of the space, so that the accuracy is unstable.
Disclosure of Invention
The invention aims to provide a regression analysis method of spatial non-stationarity relation based on GSNNR, so as to make up for the defects of the prior art.
In order to achieve the purpose, the invention adopts the following specific technical scheme:
a regression analysis method of spatial non-stationarity relation based on total space neural network regression (GSNNR), comprising the following steps:
s1: collecting spatial information data, dividing the spatial information data into a training set and a test set, and preprocessing the data to obtain characteristic information including spatial characteristics and attribute spatial characteristics;
s2: inputting the spatial features and attribute spatial features obtained in the step S1 into a full-space adjacent nonlinear fusion neural network model (SAPDNN), wherein the SAPDNN neural network model takes GNNWR as a basic model, and attribute spatial features are added into an input layer; obtaining a full-space adjacent expression matrix through the SAPDNN neural network model;
s3: inputting the full-space adjacent expression matrix into an SWNN module for processing, and outputting a weight matrix W;
Figure 156900DEST_PATH_IMAGE001
outputting a final prediction result y ^ by the weight matrix W after being input into a linear regression model OLR;
s4: and the GNNWR and the SWNN form a GSNNR model, the GSNNR model is trained by using the training set to obtain the trained GSNNR model, test data are input into the trained GSNNR model, and a result is output.
Further, in S1: the spatial characteristics refer to position information in a geographic space, such as information characteristics of longitude and latitude, altitude, position coordinates and the like; the attribute spatial features refer to self attributes owned by the geographic entities, such as information features of temperature, wind direction, vegetation types, tree diameters and the like.
Further, in S1: and measuring the spatial features by adopting Euclidean distance:
Figure 710372DEST_PATH_IMAGE002
for the measurement of the attribute space feature, the absolute difference distance of a designated attribute value or the weighted difference distance of a plurality of attribute values of the geographic attribute in the vector space is referred to; the mathematical expression of the Attribute Distance (Attribute Distance) is as follows:
Figure 817000DEST_PATH_IMAGE003
wherein the content of the first and second substances,
Figure 264162DEST_PATH_IMAGE004
indicating the attribute distance between the ith and j sample points, superscriptAIs the mark of the attribute characteristic, n is the number of the attribute categories of the sample points participating in the calculation,
Figure 544182DEST_PATH_IMAGE005
is a weighting coefficient of the k-th attribute value and satisfies
Figure 506321DEST_PATH_IMAGE006
In order to eliminate the difference of the position distance and the attribute distance in the measurement scale in the vector space, a scale weight parameter is introduced, and the position distance is measured
Figure 885481DEST_PATH_IMAGE007
Distance from attribute
Figure 593674DEST_PATH_IMAGE008
Fusing to construct 'position-attribute' unified distance expression
Figure 16565DEST_PATH_IMAGE009
Expressed as follows:
Figure 419996DEST_PATH_IMAGE010
wherein λ and φ are a position distance scale weight parameter and an attribute distance scale weight parameter, respectively.
Further, in S2: for in space𝑖And𝑗two sample points, assuming a uniform distance representation that takes into account location distance and attribute distance
Figure 117694DEST_PATH_IMAGE009
The non-linear fusion function of (a), which is mathematically expressed as follows:
Figure 821339DEST_PATH_IMAGE011
fitting a "location-attribute" uniform distance representation using a neural network
Figure 759339DEST_PATH_IMAGE009
Constructing a 'position-attribute' fusion Neural Network (SAPNN) between two sample points by using the nonlinear fusion function of the system; by distance of position
Figure 696071DEST_PATH_IMAGE007
Distance from attribute
Figure 151454DEST_PATH_IMAGE008
As input, through several fully connected layers, obtain𝑖And with𝑗Uniform distance between two sample points characterisation: by distance of position
Figure 224452DEST_PATH_IMAGE007
Distance from attribute
Figure 5458DEST_PATH_IMAGE008
As input, through several fully connected layers, obtain𝑖And𝑗the uniform distance between two sample points is characterized by the following formula:
Figure 242535DEST_PATH_IMAGE012
the SAPNN is used for fusing the spatial features and the attribute features of the two sample points; considering the interaction that the uniform distance relation of the space-attribute exists between any two sample points in the point set, a space-attribute fusion Deep Neural Network (SAPDNN) is constructed;
for any sample point i, the position distance characterization vector of the point and other points in the point set in the sample space can be obtained
Figure 282035DEST_PATH_IMAGE013
Characterization vector of distance from attribute
Figure 163535DEST_PATH_IMAGE014
Wherein n is the total number of sample points; for simplicity, the above two distance characterization vectors are simplified as
Figure 364709DEST_PATH_IMAGE015
And
Figure 761186DEST_PATH_IMAGE016
(ii) a By the distance between the position of the sample point i and all other sample points
Figure 683006DEST_PATH_IMAGE015
Distance from attribute
Figure 933859DEST_PATH_IMAGE016
As input, to the sample points
Figure 56666DEST_PATH_IMAGE017
The distance from the position between two points of each sample point
Figure 455287DEST_PATH_IMAGE007
Distance from attribute
Figure 649639DEST_PATH_IMAGE008
The sample point can be obtained by adopting the SAPNN network to carry out the unified distance fusion calculation of the position-attribute
Figure 568047DEST_PATH_IMAGE017
Unified distance characterization vector for fused position distance and attribute distance with all sample points
Figure 111024DEST_PATH_IMAGE018
And then carrying out nonlinear fusion on a plurality of full connection layers to obtain a sample point in a representable space
Figure 419777DEST_PATH_IMAGE017
And all other sample points are subjected to a 'position-attribute' unified distance measurement
Figure 604771DEST_PATH_IMAGE019
The formula is as follows:
Figure 428284DEST_PATH_IMAGE020
furthermore, the SAPDNN neural network model adopts a neural network architecture with three layers, namely an input layer, a hidden layer and an output layer, and technologies such as He parameter initialization, a PReLU activation function, batch normalization and a variable learning rate are used in training to improve the generalization of the model.
Further, the He parameter initialization, the PReLU activation function, the batch normalization and the learning variation rate are specifically as follows:
initializing the He parameter: exponential amplification or reduction of signals during antecedent propagation and backward propagation in the network is avoided, and therefore gradient disappearance or explosion is avoided;
the PReLU activation function, ai, is a learnable parameter,
Figure 158474DEST_PATH_IMAGE021
the PReLU activation function improves the fitting performance of the model under the condition of almost not increasing parameters, and reduces the risk of overfitting;
the batch normalization is as follows: the output of each layer of the model is normalized before passing through the activation function, so that the numerical value is ensured to be stable when being transmitted in the middle of the network, the network is easier to converge, and the overfitting risk is reduced;
the learning rate is as follows: in model training, it is generally desired that the initial learning rate is slightly higher and the late learning rate is slightly lower. The learning rate can be adapted to the degree of model training by using the variable learning rate, and the learning rate is smaller and smaller when the model is more and more accurate.
Further, the SWNN module is a neural network architecture of an input layer, two hidden layers (which may be more than two layers), and an output layer; calculating the weight of the input group full-space adjacent expression matrix; the same training optimization technique as SAPDNN is used in training.
Compared with the prior art, the invention has the advantages and beneficial effects that:
(1) the method introduces an attribute space as an important characteristic for analyzing a non-stationarity process, and the attribute space is brought into the input of a spatial non-stationarity detection model; attribute Space (Attribute Space) refers to the attributes that are owned within a geospatial scope. The spatial difference of the geographic attributes in combination with the geographic spatio-temporal distribution is of great significance for revealing complex geographic phenomena.
(2) The invention provides the full-space expression of the fusion space and the attribute space, the deep neural network is used for fusing the geographic space and the attribute space, and the fused composite feature can more accurately represent the actual space non-stationarity process compared with the single geographic space feature, thereby further improving the accuracy of the measurement of the non-stationarity.
(3) The invention also provides a full-space adjacent nonlinear fusion neural network (SAPDNN), and the neural network model takes GNNWR as a basic model, adds attribute space characteristics and improves the accuracy of prediction. The network is used for fusing the geographic spatial characteristics and the geographic attribute characteristics to obtain a full-space expression of the geographic characteristics.
Drawings
Fig. 1 is a basic framework diagram of the SAPDNN neural network model.
Fig. 2 is a process diagram of the SAPNN neural network model.
Fig. 3 is a process diagram of the SAPDNN neural network model.
Fig. 4 is a basic framework diagram of the SWNN module.
FIG. 5 is a flow chart of the SWNN module output weight matrix.
Fig. 6 is a diagram of the input and output structure of the GSNNWR model.
FIG. 7 is a cross-training validation flow diagram of the present invention.
Detailed Description
The technical solution of the present invention will be further described and illustrated with reference to the following examples.
Example 1:
a regression analysis method of spatial non-stationarity relation based on whole space neural network regression (GSNNR) comprises the following steps:
s1: collecting spatial information data, dividing the spatial information data into a training set and a test set, and preprocessing the data to obtain characteristic information including spatial characteristics and attribute spatial characteristics; the spatial characteristics refer to position information in a geographic space, such as information characteristics of longitude and latitude, altitude, position coordinates and the like; the attribute spatial features refer to self attributes owned by the geographic entities, such as information features of temperature, wind direction, vegetation types, tree diameters and the like.
And measuring the spatial features by adopting Euclidean distance:
Figure 469369DEST_PATH_IMAGE002
for the measurement of the attribute space feature, the absolute difference distance of a designated attribute value or the weighted difference distance of a plurality of attribute values of the geographic attribute in the vector space is referred to; the mathematical expression of the Attribute Distance (Attribute Distance) is as follows:
Figure 943207DEST_PATH_IMAGE003
wherein the content of the first and second substances,
Figure 898525DEST_PATH_IMAGE004
indicating the attribute distance between the ith and j sample points, superscriptAIs the mark of the attribute characteristic, n is the number of the attribute categories of the sample points participating in the calculation,
Figure 48883DEST_PATH_IMAGE005
is a weighting coefficient of the kth attribute value and satisfies
Figure 332228DEST_PATH_IMAGE006
For eliminating location distance and attribute distanceThe difference of distance on the measurement scale in the vector space is introduced into the scale weight parameter, and the position distance is measured
Figure 859025DEST_PATH_IMAGE007
Distance from attribute
Figure 403269DEST_PATH_IMAGE008
Fusing to construct 'position-attribute' unified distance expression
Figure 475262DEST_PATH_IMAGE009
Expressed as follows:
Figure 760750DEST_PATH_IMAGE010
wherein λ and φ are a position distance scale weight parameter and an attribute distance scale weight parameter, respectively.
S2: inputting the spatial features and attribute spatial features obtained in S1 into a full-space proximity nonlinear fusion neural network model (SAPDNN), which takes GNNWR as a basic model and adds attribute spatial features to an input layer, as shown in fig. 1; obtaining a full-space adjacent expression matrix through the SAPDNN neural network model;
for in space𝑖And𝑗two sample points, assuming a uniform distance representation that takes into account location distance and attribute distance
Figure 310811DEST_PATH_IMAGE009
The non-linear fusion function of (a), which is mathematically expressed as follows:
Figure 162092DEST_PATH_IMAGE011
fitting a "location-attribute" uniform distance representation using a neural network
Figure 732882DEST_PATH_IMAGE009
To construct a 'position-attribute' fusion between two sample pointsSynthetic Neural networks (SAPNN), as shown in fig. 2; by distance of position
Figure 787557DEST_PATH_IMAGE007
Distance from attribute
Figure 593839DEST_PATH_IMAGE008
As input, through several fully connected layers, obtain𝑖And𝑗uniform distance between two sample points characterisation: by distance of position
Figure 50359DEST_PATH_IMAGE007
Distance from attribute
Figure 713421DEST_PATH_IMAGE008
As input, through several fully connected layers, obtain𝑖And𝑗the uniform distance between two sample points is characterized by the following formula:
Figure 255392DEST_PATH_IMAGE012
the SAPNN is used for fusing the spatial features and the attribute features of the two sample points; considering the interaction of the uniform distance relationship of the space-attribute between any two sample points in the point set, a Spatial-attribute fused Deep Neural Network (SAPDNN) is constructed, as shown in fig. 3;
for any one sample point
Figure 396524DEST_PATH_IMAGE017
All the characteristic vectors of the position distance between the point and other points in the point set in the sample space can be obtained
Figure 35447DEST_PATH_IMAGE013
Characterization vector of distance from attribute
Figure 620143DEST_PATH_IMAGE014
Wherein
Figure 898677DEST_PATH_IMAGE022
Is the total number of sample points; for simplicity, the above two distance characterization vectors are simplified as
Figure 797494DEST_PATH_IMAGE015
And
Figure 477874DEST_PATH_IMAGE016
(ii) a By the distance between the position of the sample point i and all other sample points
Figure 561368DEST_PATH_IMAGE015
Distance from attribute
Figure 89650DEST_PATH_IMAGE016
As input, to the sample points
Figure 572584DEST_PATH_IMAGE017
The distance from the position between two points of each sample point
Figure 61465DEST_PATH_IMAGE007
Distance from attribute
Figure 237232DEST_PATH_IMAGE008
The sample point can be obtained by adopting the SAPNN network to carry out the unified distance fusion calculation of the position-attribute
Figure 709932DEST_PATH_IMAGE017
Unified distance characterization vector for fused position distance and attribute distance with all sample points
Figure 262136DEST_PATH_IMAGE018
And then nonlinear fusion is carried out on the plurality of full connection layers to obtain the position-attribute unified distance measurement between the sample point and all other sample points in the representation space
Figure 995737DEST_PATH_IMAGE019
The formula is as follows:
Figure 296400DEST_PATH_IMAGE020
the SAPDNN neural network model adopts a neural network architecture with three layers of an input layer, a hidden layer and an output layer, and the generalization of the model is improved by using the technologies of He parameter initialization, a PReLU activation function, batch normalization, learning rate variation and the like in training.
The He parameter initialization, the PReLU activation function, the batch normalization and the learning variation rate are specifically as follows:
initializing the He parameter: exponential amplification or reduction of signals during antecedent propagation and backward propagation in the network is avoided, and therefore gradient disappearance or explosion is avoided;
the PReLU activation function, ai, is a learnable parameter,
Figure 115451DEST_PATH_IMAGE021
the PReLU activation function improves the fitting performance of the model under the condition of almost not increasing parameters, and reduces the risk of overfitting;
the batch normalization is as follows: the output of each layer of the model is normalized before passing through the activation function, so that the numerical value is ensured to be stable when being transmitted in the middle of the network, the network is easier to converge, and the overfitting risk is reduced;
the learning rate is as follows: in model training, it is generally desired that the initial learning rate is slightly higher and the late learning rate is slightly lower. The learning rate can be adapted to the degree of model training by using the variable learning rate, and the learning rate is smaller and smaller when the model is more and more accurate.
S3: inputting the full-space adjacent expression matrix into an SWNN module for processing, and outputting a weight matrix W as shown in FIG. 5;
Figure 205767DEST_PATH_IMAGE023
outputting a final prediction result y ^ by the weight matrix W after being input into a linear regression model OLR;
as shown in fig. 4, the SWNN module is a neural network architecture with four layers, an input layer, two hidden layers (which may be more than two layers), and an output layer; calculating the weight of the input group full-space adjacent expression matrix; the same training optimization technique as SAPDNN is used in training.
S4: the GNNWR and SWNN constitute a GSNNR model, the GSNNR model is trained by using the training set to obtain a trained GSNNR model, and then the test data is input into the trained GSNNR model to output a result, as shown in fig. 6.
The technical features of this embodiment include the following:
(1) the "space-attribute" feature fusion. And inputting the spatial characteristics and the geographic attribute characteristics corresponding to each sample point into SAPDNN, and obtaining a space-attribute full-proximity characteristic expression matrix through operation. The outputs of multiple sample points are combined into one large matrix as input to the next module.
(2) And calculating a space-attribute characteristic weight matrix. For the fusion characteristic matrix output by the former module, a deep neural network is adopted to extract characteristics, the neural network adopts a multi-layer perceptron structure, and optimization technologies such as Dropout, He parameter initialization, a PReLU activation function and the like are adopted in training to enhance the generalization capability of the model.
(3) And (5) calculating a prediction result. And multiplying the non-stationary weight value by the least square coefficient to obtain the non-stationary coefficient. The final output fit value y ̂ _ i of the model is the result of the multiplicative sum of all the non-stationary coefficients and their corresponding arguments. The least squares coefficients are derived from the training set.
(4) And (5) verifying and testing. In the future, the effectiveness of algorithm design is verified, a data set is divided into a training set and a test set according to the proportion of 3:1, 10-fold cross validation is performed in the training set according to the proportion of 9:1, and the process of cross validation is shown in fig. 7.
Example 2
In this embodiment, based on embodiment 1, the spatial non-stationarity relationship of the PM2.5 concentration in the atmosphere is used as a research object, and the actual PM2.5 concentration value is predicted by using the algorithm model.
In order to ensure the representativeness of data, monitoring point data in 2018 nationwide is selected as research data, and the influence of full-space proximity expression fused with space-attribute characteristics on the calculation precision of the non-stationarity relation is emphasized and contrasted. For data processing, Wind Direction (WD) is selected as a geographical attribute feature related to PM2.5 concentration, PM2.5 concentration is taken as a prediction object, and input features of the model further include Data of Elevation (DEM), relative humidity (r), 10m Wind Speed (WS), Aerosol (AOD), precipitation (TP) and 2m Temperature (TEMP).
In the training set and the test set, all data samples were set according to 3:1 into a cross-validation set and a test set, and a ratio of 9: the 10-fold cross-validation at the ratio of 1 ensures the generalization capability of the model. The used data is obtained by randomly sampling monitoring points in the whole country, and the data is randomly distributed in the geographical space range in the whole country, so that the conclusion of the case is general and representative.
Compared with GWR and GNNWR models, the improvement of the invention mainly lies in the following two points:
firstly, introducing a geographic attribute space as one of characteristics input by an algorithm, and performing fusion processing on the space-attribute characteristics by using SAPDNN to obtain a full-space expression. Compared with the original scheme only considering the geographical space position relation, the new scheme provided by the invention has more representativeness on the basic data level, and the actual geographical space non-stationarity relation can be represented by a characteristic processing mode combining space and attributes.
And secondly, the resolving precision is improved after a new characteristic representation mode is introduced, and compared with a GNNWR model without considering the geographic attribute spatial characteristics and a GWR model adopting different kernel functions, the resolving precision is improved by about 10% on average.
Introduction of geographic attribute space
The existing model scheme only excavates the spatial non-stationarity relation between samples from the two-dimensional distance on the geographic spatial position relation, however, in practice, the relation between the samples is influenced by various factors, and after the geographic attribute characteristics are introduced, the expression of the samples on the data is closer to the actual situation, and more semantic information is included.
Two features are fused by adopting a deep neural network
After new feature expression is introduced, the two different features are processed in a deep neural network mode for fusion, and a full-space expression matrix of space-attribute is obtained to be used as input data of subsequent calculation. As the data contains more semantic information, the resolving precision of the model is improved.
On the basis of the above embodiments, the present invention continues to describe the technical features and functions of the technical features in the present invention in detail to help those skilled in the art fully understand the technical solutions of the present invention and reproduce them.
Finally, although the present description refers to embodiments, not every embodiment contains only a single technical solution, and such description of the present description is for clarity reasons only, and those skilled in the art should make the description as a whole, and the technical solutions in the embodiments can be appropriately combined to form other embodiments that can be understood by those skilled in the art.

Claims (7)

1. A regression analysis method of spatial non-stationarity relation based on GSNNR is characterized by comprising the following steps:
s1: collecting spatial information data, dividing the spatial information data into a training set and a test set, and preprocessing the data to obtain characteristic information including spatial characteristics and attribute spatial characteristics;
s2: inputting the spatial features and the attribute spatial features obtained in the step S1 into a full-space adjacent nonlinear fusion neural network model SAPDNN, wherein the SAPDNN takes GNNWR as a basic model, and attribute spatial features are added into an input layer; obtaining a full-space adjacent expression matrix through the SAPDNN;
s3: inputting the full-space adjacent expression matrix into an SWNN module for processing, and outputting a weight matrix W;
Figure 233593DEST_PATH_IMAGE001
outputting a final prediction result y ^ by the weight matrix W after being input into a linear regression model OLR;
s4: and the GNNWR and the SWNN form a GSNNR model, the GSNNR model is trained by using the training set to obtain the trained GSNNR model, test data are input into the trained GSNNR model, and a result is output.
2. The regression analysis method according to claim 1, wherein in S1: the spatial characteristics refer to position information in a geographic space, and comprise longitude and latitude, altitude and position coordinates; the attribute spatial characteristics refer to the attributes of the geographic entity, including temperature, wind direction, vegetation type and tree diameter.
3. The regression analysis method according to claim 1, wherein in S1: the spatial features are measured using Euclidean distances:
Figure 822837DEST_PATH_IMAGE002
the measurement of the attribute space feature is the absolute difference distance of a designated attribute value or the weighted difference distance of a plurality of attribute values of the geographic attribute in the vector space; the mathematical expression of the attribute distance is as follows:
Figure 523946DEST_PATH_IMAGE003
wherein the content of the first and second substances,
Figure 341992DEST_PATH_IMAGE004
represents the attribute distance between the ith and j sample pointsSign boardAIs the mark of the attribute characteristic, n is the number of the attribute categories of the sample points participating in the calculation,
Figure 562889DEST_PATH_IMAGE005
is a weighting coefficient of the k-th attribute value and satisfies
Figure 41143DEST_PATH_IMAGE006
Introducing a scale weight parameter and calculating the position distance
Figure 731013DEST_PATH_IMAGE007
Distance from attribute
Figure 336438DEST_PATH_IMAGE008
Fusing to construct 'position-attribute' unified distance expression
Figure 926688DEST_PATH_IMAGE009
Expressed as follows:
Figure 342888DEST_PATH_IMAGE010
wherein λ and φ are a position distance scale weight parameter and an attribute distance scale weight parameter, respectively.
4. The regression analysis method according to claim 1, wherein in S2: for in space𝑖And𝑗two sample points, assuming a uniform distance representation that takes into account location distance and attribute distance
Figure 769321DEST_PATH_IMAGE009
The non-linear fusion function of (a), which is mathematically expressed as follows:
Figure 427705DEST_PATH_IMAGE011
fitting a "location-attribute" uniform distance representation using a neural network
Figure 373926DEST_PATH_IMAGE009
The nonlinear fusion function of (1) and constructing a 'position-attribute' fusion neural network SAPNN between two sample points; by distance of position
Figure 210295DEST_PATH_IMAGE007
Distance from attribute
Figure 638871DEST_PATH_IMAGE008
As input, through several fully connected layers, obtain𝑖And𝑗uniform distance between two sample points characterisation: by distance of position
Figure 586099DEST_PATH_IMAGE007
Distance from attribute
Figure 906266DEST_PATH_IMAGE008
As input, through several fully connected layers, obtain𝑖And𝑗the uniform distance between two sample points is characterized by the following formula:
Figure 179116DEST_PATH_IMAGE012
the SAPNN is used for fusing the spatial features and the attribute features of the two sample points; considering the interaction of the unified distance relation of 'space-attribute' between any two sample points in the point set, constructing a 'space-attribute' fusion deep neural network SAPDNN;
for any sample point i, the position distance characterization vector of the point and other points in the point set in the sample space can be obtained
Figure 94988DEST_PATH_IMAGE013
Characterization vector of distance from attribute
Figure 111486DEST_PATH_IMAGE014
Wherein n is the total number of sample points; for simplicity, the above two distance characterization vectors are simplified as
Figure 297879DEST_PATH_IMAGE015
And
Figure 272788DEST_PATH_IMAGE016
(ii) a By the distance between the position of the sample point i and all other sample points
Figure 410377DEST_PATH_IMAGE015
Distance from attribute
Figure 964986DEST_PATH_IMAGE016
As input, to the sample points
Figure 271465DEST_PATH_IMAGE017
The distance from the position between two points of each sample point
Figure 886117DEST_PATH_IMAGE007
Distance from attribute
Figure 511002DEST_PATH_IMAGE008
The sample point can be obtained by adopting the SAPNN network to carry out the unified distance fusion calculation of the position-attribute
Figure 134882DEST_PATH_IMAGE017
Unified distance characterization vector for fused position distance and attribute distance with all sample points
Figure 295867DEST_PATH_IMAGE018
And then nonlinear fusion is carried out on the sample points through a plurality of full connection layers to obtain sample points in the representational space
Figure 347000DEST_PATH_IMAGE017
And all other sample points are subjected to a 'position-attribute' unified distance measurement
Figure 459181DEST_PATH_IMAGE019
The formula is as follows:
Figure 886751DEST_PATH_IMAGE020
5. the regression analysis method of claim 1, wherein in S2, said SAPDNN adopts a neural network architecture with three layers of input layer, hidden layer and output layer, and He parameter initialization, prilu activation function, batch normalization, and learning rate variation are used in training to improve the generalization of the model.
6. The regression analysis method of claim 5, wherein the He parameter initialization, the PReLU activation function, the batch normalization, and the learning rate variation are as follows:
initializing the He parameter: exponential amplification or reduction of signals during antecedent propagation and backward propagation in the network is avoided, and therefore gradient disappearance or explosion is avoided;
the PReLU activation function, ai, is a learnable parameter,
Figure 371084DEST_PATH_IMAGE021
the PReLU activation function improves the fitting performance of the model under the condition of almost not increasing parameters;
the batch normalization is as follows: normalizing the output of each layer of the model before the output passes through the activation function;
the learning rate is as follows: the learning rate can be adapted to the degree of model training by using the variable learning rate, and the learning rate is smaller and smaller when the model is more and more accurate.
7. The regression analysis method of claim 1, wherein said SWNN module is a neural network architecture of an input layer, two or more hidden layers, and an output layer; calculating the weight of the input group full-space adjacent expression matrix; the same training optimization technique as the SAPDNN is used in training.
CN202210984054.0A 2022-08-17 2022-08-17 Regression analysis method of space non-stationarity relation based on GSNNR Pending CN115049026A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210984054.0A CN115049026A (en) 2022-08-17 2022-08-17 Regression analysis method of space non-stationarity relation based on GSNNR

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210984054.0A CN115049026A (en) 2022-08-17 2022-08-17 Regression analysis method of space non-stationarity relation based on GSNNR

Publications (1)

Publication Number Publication Date
CN115049026A true CN115049026A (en) 2022-09-13

Family

ID=83168429

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210984054.0A Pending CN115049026A (en) 2022-08-17 2022-08-17 Regression analysis method of space non-stationarity relation based on GSNNR

Country Status (1)

Country Link
CN (1) CN115049026A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115759291A (en) * 2022-11-21 2023-03-07 武汉大学 Space nonlinear regression method and system based on ensemble learning
CN117932456A (en) * 2024-03-22 2024-04-26 中国科学院地理科学与资源研究所 Integrated spatial prediction method considering spatial heterogeneity
CN117932456B (en) * 2024-03-22 2024-06-07 中国科学院地理科学与资源研究所 Integrated spatial prediction method considering spatial heterogeneity

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200311848A1 (en) * 2019-04-01 2020-10-01 International Business Machines Corporation Parking continuity with unused duration between automated vehicles
CN113064220A (en) * 2021-06-03 2021-07-02 四川九通智路科技有限公司 Visibility measuring system and measuring method based on nonlinear autoregressive neural network
CN113591685A (en) * 2021-07-29 2021-11-02 武汉理工大学 Geographic object spatial relationship identification method and system based on multi-scale pooling

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200311848A1 (en) * 2019-04-01 2020-10-01 International Business Machines Corporation Parking continuity with unused duration between automated vehicles
CN113064220A (en) * 2021-06-03 2021-07-02 四川九通智路科技有限公司 Visibility measuring system and measuring method based on nonlinear autoregressive neural network
CN113591685A (en) * 2021-07-29 2021-11-02 武汉理工大学 Geographic object spatial relationship identification method and system based on multi-scale pooling

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
孟耀伟等: "多级空间数据插值及其应用方法", 《许昌学院学报》 *
王中一: "顾及全域时空复杂邻近性的神经网络加权回归方法", 《中国优秀博士学位论文全文数据库》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115759291A (en) * 2022-11-21 2023-03-07 武汉大学 Space nonlinear regression method and system based on ensemble learning
CN117932456A (en) * 2024-03-22 2024-04-26 中国科学院地理科学与资源研究所 Integrated spatial prediction method considering spatial heterogeneity
CN117932456B (en) * 2024-03-22 2024-06-07 中国科学院地理科学与资源研究所 Integrated spatial prediction method considering spatial heterogeneity

Similar Documents

Publication Publication Date Title
CN110533631B (en) SAR image change detection method based on pyramid pooling twin network
CN111126575B (en) Gas sensor array mixed gas detection method and device based on machine learning
CN108399248A (en) A kind of time series data prediction technique, device and equipment
Li et al. Comparative analysis of BPNN, SVR, LSTM, Random Forest, and LSTM-SVR for conditional simulation of non-Gaussian measured fluctuating wind pressures
CN111340132B (en) Machine olfaction mode identification method based on DA-SVM
CN113049500B (en) Water quality detection model training and water quality detection method, electronic equipment and storage medium
CN113901384A (en) Ground PM2.5 concentration modeling method considering global spatial autocorrelation and local heterogeneity
CN108846200B (en) Quasi-static bridge influence line identification method based on iteration method
Coulibaly et al. Rule-based machine learning for knowledge discovering in weather data
CN116610998A (en) Switch cabinet fault diagnosis method and system based on multi-mode data fusion
CN113378473A (en) Underground water arsenic risk prediction method based on machine learning model
Yao et al. Variable selection for nonlinear soft sensor development with enhanced binary differential evolution algorithm
CN115049026A (en) Regression analysis method of space non-stationarity relation based on GSNNR
CN114626304A (en) Soft measurement modeling method for online prediction of copper grade in ore pulp
CN116702005A (en) Neural network-based data anomaly diagnosis method and electronic equipment
Cornford et al. Modelling frontal discontinuities in wind fields
CN115510763A (en) Air pollutant concentration prediction method and system based on data-driven exploration
Roigé et al. Self-organizing maps for analysing pest profiles: Sensitivity analysis of weights and ranks
Zhu et al. Rapid freshness prediction of crab based on a portable electronic nose system
CN115062551A (en) Wet physical process parameterization method based on time sequence neural network
Wang et al. A Research of Neural Network Optimization Technology for Apple Freshness Recognition Based on Gas Sensor Array
CN113553708A (en) Method and device for tracing key influence factors of simulation model
CN113108949A (en) Model fusion-based sonde temperature sensor error prediction method
He et al. A predictive model for the sensory aroma characteristics of flue-cured tobacco based on a back-propagation neural network
CN111382147A (en) Meteorological data missing interpolation method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination