CN114511058B - Load element construction method and device for electric power user portrait - Google Patents

Load element construction method and device for electric power user portrait Download PDF

Info

Publication number
CN114511058B
CN114511058B CN202210101090.8A CN202210101090A CN114511058B CN 114511058 B CN114511058 B CN 114511058B CN 202210101090 A CN202210101090 A CN 202210101090A CN 114511058 B CN114511058 B CN 114511058B
Authority
CN
China
Prior art keywords
user
active power
feature vector
load element
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210101090.8A
Other languages
Chinese (zh)
Other versions
CN114511058A (en
Inventor
王进
华淼杰
倪格格
翁蓓蓓
徐兴春
鞠玲
刘振扬
郑俊杰
赵剑锋
董坤
毛妍纯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Taizhou Power Supply Co of State Grid Jiangsu Electric Power Co Ltd
Original Assignee
Taizhou Power Supply Co of State Grid Jiangsu Electric Power Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Taizhou Power Supply Co of State Grid Jiangsu Electric Power Co Ltd filed Critical Taizhou Power Supply Co of State Grid Jiangsu Electric Power Co Ltd
Priority to CN202210101090.8A priority Critical patent/CN114511058B/en
Publication of CN114511058A publication Critical patent/CN114511058A/en
Application granted granted Critical
Publication of CN114511058B publication Critical patent/CN114511058B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/31Indexing; Data structures therefor; Storage structures
    • G06F16/316Indexing structures
    • G06F16/325Hash tables
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biophysics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Databases & Information Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Biology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a load element construction method and a device for an electric power user portrait, wherein the method comprises the following steps: collecting total power of a household user and active power of a single load element as a training set; carrying out thermal single code coding on the total power of the home subscriber and the active power of a single load element; carrying out hash processing on the user characteristic vector and the active power characteristic vector; respectively inputting the low-dimensional user feature vector and the low-dimensional active power feature vector into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and respectively outputting each user embedded feature vector and each load element embedded feature vector; calculating the relevance of each user embedded feature vector and the load element embedded feature vector based on cosine similarity, and converting a relevance calculation result into posterior probability by utilizing a logistic regression activation function to obtain a load element with highest relevance to the user; the method can improve the efficiency of establishing the user portrait model and has strong reliability.

Description

Load element construction method and device for electric power user portrait
Technical Field
The invention relates to the technical field of power, in particular to a method and a device for constructing a load element for a power consumer portrait.
Background
With the continuous development of electronic technology and big data technology, the formation of household electricity information is gradually converted into a data form, so that willingness information can be intuitively and conveniently obtained, the time consumption is further reduced, and the efficiency is improved. In order to find the target data more quickly and accurately in the network, various attribute information of household user electricity consumption needs to be divided by using a related classification technology. Supply and demand information of the home user can be further known through portrait construction, data such as behavior habits and the like are accurately positioned, and the overall view of the user information is conveniently known.
In the technical field of portrayal, embedded features are generated by utilizing artificially constructed features through a deep learning method and are input into a load element which is predicted to be opened by a user in a neural network, and finally, the habit of electricity utilization behavior of a household user is displayed. Load element opening prediction is the most important consideration of model effect, and a user is crucial to the historical opening behavior of the load element. However, most current power consumer representation models utilize some artificial features as raw features to input into the multi-layer neural network, which are derived from the consumer representation, the consumer's historical behavior, the load element properties, etc.
For example, patent document CN112417308A discloses a user portrait tag generating method based on electric power big data, which generates a user characteristic tag by using big data processing technology, wherein a basic database configuration of the big data processing technology is constructed around customer appeal, opinion and consultation data streams of channels such as electric power official telephones, electric power intranet and extranet, mobile phone APP, weChat public number lines, business hall opinion books are imported into the basic database as tag data original sources, and clients are marked in a tag form through data analysis, so that a user portrait is established.
The method can integrate various source data, build a multi-dimensional and three-dimensional customer portrait by means of big data analysis technology, and describe deep behavior characteristics of users through labels. However, the method is biased to adopt consultation opinions provided by users to describe the characteristics of the users, lacks acquisition of direct electricity information, does not well integrate structural features and node features of complex recommended network scenes, is poor in model reliability, is low in operation speed, and is low in user portrait model establishment efficiency.
Disclosure of Invention
The invention provides a method and a device for constructing a load element for an electric power user portrait, which are used for carrying out model training based on the total power of a household user and the active power of a single load element, and are used for integrating the structural characteristics and the node characteristics of a complex recommended network scene, so that the efficiency of establishing a user portrait model is improved, and the reliability is high.
A load element construction method for an electric power user portrait includes:
collecting total power of a household user and active power of a single load element as a training set;
performing thermal single code coding on the total power of the home user and the active power of a single load element to respectively obtain a user characteristic vector and an active power characteristic vector;
respectively carrying out hash processing on the user characteristic vector and the active power characteristic vector to respectively obtain a low-dimensional user characteristic vector and a low-dimensional active power characteristic vector;
the low-dimensional user feature vector and the low-dimensional active power feature vector are respectively input into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and each user embedded feature vector and each load element embedded feature vector are respectively output;
and calculating the correlation between each user embedded feature vector and the load element embedded feature vector based on the cosine similarity, and converting the correlation calculation result into posterior probability by utilizing a logistic regression activation function to obtain the load element with the highest degree of correlation with the user.
Further, performing hot unique code encoding on the total power of the home user, including:
dividing the same family user total power into user similar characteristics;
the similar characteristics of each user are expressed by binary system, and the user characteristic vector is obtained;
performing thermal single code encoding on the active power of the single load element to obtain an active power characteristic vector of the single load element, including:
dividing the active power of the same single load element into active power similar characteristics;
and (5) representing each active power similar characteristic by using a binary system to obtain an active power characteristic vector.
Further, hash processing is performed on the user feature vector to obtain a low-dimensional user feature vector, including:
respectively adding a start mark and an end mark before and after the user feature vector;
setting step length and sliding window based on N-GRAM algorithm, recording user characteristic vector slices of each scratch, recording all user characteristic vector slices as vector representations composed of a plurality of numbers, and obtaining low-dimension user characteristic vectors;
performing hash processing on the active power feature vector to obtain a low-dimension active power feature vector, including:
respectively adding a start mark and an end mark before and after the active power characteristic vector;
based on an N-GRAM algorithm, setting a step length and a sliding window, recording active power characteristic vector slices which are marked each time, recording all the active power characteristic vector slices as vector representations consisting of a plurality of numbers, and obtaining a low-dimensional active power characteristic vector.
Further, the low-dimensional user feature vector and the low-dimensional active power feature vector are respectively input into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and parameters of the first multi-layer perceptron network and the second multi-layer perceptron network are respectively obtained according to a minimized loss function.
Further, the parameters of the first multi-layer perceptron network and the second multi-layer perceptron network comprise weight matrixes and bias items of all hidden layers.
Further, the load element embedding feature vector is calculated by the following formula:
l i =f(W i l i-1 +b i )i=2,…,N-1;
y L =f(W N l N-1 +b N );
wherein ,yL Embedding feature vectors for load elements, l i Hidden layer of the ith layer of the first multi-layer perceptron network, W i B is a weight matrix of an ith hidden layer of the first multi-layer perceptron network i Bias item representing ith hidden layer of first multi-layer perceptron network, N is hidden layer number of first multi-layer perceptron network, b N Bias items for the output layer of the first multi-layer perceptron network;
the user-embedded feature vector is calculated by the following formula:
l m =f(W m l m-1 +b m )m=2,…,M-1;
y U =f(W M l M-1 +b M );
wherein ,yU Embedding feature vectors for users, l m Hidden layer of the m-th layer of the second multi-layer perceptron network, W m B is a weight matrix of an mth hidden layer of the second multi-layer perceptron network m Bias term representing mth hidden layer of second multi-layer perceptron network, M is hidden layer number of second multi-layer perceptron network, b M And the bias item of the output layer of the second multi-layer perceptron network.
Further, the correlation of the user embedded feature vector and the load element embedded feature vector is calculated by the following formula:
Figure BDA0003492475560000041
wherein ,yU Embedding feature vectors for users, y L The feature vector is embedded for the load element,
Figure BDA0003492475560000043
representing a transpose of the user embedded feature vector, R (U, L) representing a correlation of the user embedded feature vector with the load element embedded feature vector.
Further, the posterior probability is calculated by the following formula:
Figure BDA0003492475560000042
wherein ,P(Lj U) is a posterior probability, γ is a smoothing factor of a logistic regression activation function, L represents a set of load elements, U represents a set of users, L' represents a sum of all load elements in the set of load elements, j is "+" or "-", L + Representing a positive sample of the load cell, L - Representing a load cell negative sample.
Further, the minimized loss function is expressed by the following formula:
Figure BDA0003492475560000044
wherein Λ is a parameter of the multi-layer perceptron network, P (L + U) represents the posterior probability, U represents the set of users, L + Representing a positive sample of the load cell.
A load cell construction apparatus for a representation of an electric power consumer, comprising:
the acquisition module is used for acquiring the total power of the household users and the active power of the single load element as a training set;
the coding module is used for carrying out hot unique code coding on the total power of the home user and the active power of the single load element to respectively obtain a user characteristic vector and an active power characteristic vector;
the processing module is used for respectively carrying out hash processing on the user characteristic vector and the active power characteristic vector to respectively obtain a low-dimensional user characteristic vector and a low-dimensional active power characteristic vector;
the training module is used for respectively inputting the low-dimensional user characteristic vector and the low-dimensional active power characteristic vector into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and respectively outputting each user embedded characteristic vector and each load element embedded characteristic vector;
and the calculating module is used for calculating the correlation between each user embedded feature vector and the load element embedded feature vector based on the cosine similarity, converting the correlation calculation result into posterior probability by utilizing a logistic regression activation function, and obtaining the load element with the highest degree of correlation with the user.
The invention provides a load element construction method and a load element construction device for an electric power user portrait, which at least comprise the following beneficial effects:
the method comprises the steps of carrying out model training based on total power of household users and active power of load elements, carrying out hash processing on characteristics obtained through thermal single code coding, carrying out multi-layer perceptron network training, generating similarity between users and loads by means of a double-tower model, and finally obtaining the load element with highest degree of association with the users, wherein structural characteristics and node characteristics of complex recommended network scenes are fused, manual priori deviation is avoided to a certain extent, operation speed during on-line prediction is effectively improved, user portrait establishment efficiency is high, and reliability is high.
Drawings
FIG. 1 is a flow chart of one embodiment of a method for constructing a load cell for an electrical consumer representation provided by the present invention.
FIG. 2 is a schematic diagram of an embodiment of a load cell construction apparatus for consumer representation of electrical power provided by the present invention.
Description of the drawings: the system comprises a 1-acquisition module, a 2-coding module, a 3-processing module, a 4-training module and a 5-calculating module.
Detailed Description
In order to better understand the above technical solutions, the following detailed description will be given with reference to the accompanying drawings and specific embodiments.
Referring to FIG. 1, in some embodiments, a load cell construction method for a representation of an electrical consumer is provided, comprising:
s1, collecting total power of a household user and active power of a single load element as a training set;
s2, performing hot unique code coding on the total power of the home user and the active power of a single load element to respectively obtain a user characteristic vector and an active power characteristic vector;
s3, carrying out hash processing on the user characteristic vector and the active power characteristic vector respectively to obtain a low-dimensional user characteristic vector and a low-dimensional active power characteristic vector respectively;
s4, respectively inputting the low-dimensional user feature vector and the low-dimensional active power feature vector into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and respectively outputting each user embedded feature vector and each load element embedded feature vector;
s5, calculating the correlation between each user embedded feature vector and the load element embedded feature vector based on cosine similarity, and converting a correlation calculation result into posterior probability by using a logistic regression activation function to obtain the load element with the highest degree of correlation with the user.
Specifically, in step S2, the hot-unique code encoding is performed on the total power of the home subscriber, including:
s211, dividing the total power of the same household user into user similar characteristics;
s212, representing the similar characteristics of each user by using binary system to obtain the user characteristic vector;
performing thermal single code encoding on the active power of the single load element to obtain an active power characteristic vector of the single load element, including:
s221, dividing the active power of the same single load element into active power similar characteristics;
s222, representing each active power similar characteristic by binary system to obtain an active power characteristic vector.
In a specific embodiment, the total power of the home subscriber and the active power of each load element are processed separately, the first multi-layer perceptron network is used for processing the total power of the home subscriber, the second multi-layer perceptron network is used for processing the active power of each load element, and the first multi-layer perceptron network and the second multi-layer perceptron network form a double-tower model.
Specifically, the present active power data is first used as a dictionary, the number of word banks is the dimension of the thermal single code coding, the dictionary is a zero vector of the word bank number x 1 dimension, and the zero vector is 1 only where the current active power value is represented, for example, the power dictionary of a refrigerator is [0,1,2,3,4 … … ], and if the current power is 0, the thermal single code coding result is [1,0 … ].
When the active power of the same single load element is divided into the same kind of characteristics, only one bit of each characteristic in each sample needs to be enabled, and the rest bits are disabled.
In step S3, hash processing is performed on the user feature vector to obtain a low-dimensional user feature vector, including:
s311, respectively adding a start mark and an end mark before and after the user feature vector;
s312, setting step length and sliding window based on an N-GRAM algorithm, recording user feature vector slices scratched each time, recording all user feature vector slices as vector representations consisting of a plurality of numbers, and obtaining a low-dimensional user feature vector;
performing hash processing on the active power feature vector to obtain a low-dimension active power feature vector, including:
s321, respectively adding a start mark and an end mark before and after the active power characteristic vector;
s322, setting step length and sliding window based on an N-GRAM algorithm, recording active power characteristic vector slices of each pass, recording all active power characteristic vector slices as vector representations composed of a plurality of numbers, and obtaining a low-dimensional active power characteristic vector. The algorithm may be represented by the following formula:
l 1 =W 1 x
wherein x represents an input feature vector of the thermal single code coding, l 1 Represents a first hidden layer, W 1 Representing the weight matrix of the first layer.
In a specific embodiment, taking a refrigerator as an example of a load element, when the active power of the refrigerator at a certain moment is 258W, a sign #258# is added to the beginning and the end of a user feature vector, then the active power is divided into n numbers, and when n=3 is taken, the numbers are #25, 258 and 58#, so that the dictionary obtained by the hot unique code coding can be represented by a vector formed by a plurality of 3 numbers, and the dimension of the dictionary is greatly reduced. For example, when the active power is 259W and the hot-unique code dictionary is [ … #25, … 258, 259, 260, … 59# … ], the hash-processed code is [ … 1, … 0,1,0 … 1 … ].
In step S4, the low-dimensional user feature vector and the low-dimensional active power feature vector are respectively input into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and parameters of the first multi-layer perceptron network and the second multi-layer perceptron network are respectively obtained according to a minimized loss function.
The parameters of the first multi-layer perceptron network and the second multi-layer perceptron network comprise weight matrixes and bias items of all hidden layers.
The load element embedding feature vector is calculated by the following formula:
l i =f(W i l i-1 +b i )i=2,…,N-1;
y L =f(W N l N-1 +b N );
wherein ,yL Embedding feature vectors for load elements, l i Hidden layer of the ith layer of the first multi-layer perceptron network, W i B is a weight matrix of an ith hidden layer of the first multi-layer perceptron network i Bias item representing ith hidden layer of first multi-layer perceptron network, N is hidden layer number of first multi-layer perceptron network, b N Bias items for the output layer of the first multi-layer perceptron network;
the user-embedded feature vector is calculated by the following formula:
l m =f(W m l m-1 +b m )m=2,…,M-1;
y U =f(W M l M-1 +b M );
wherein ,yU Embedding feature vectors for users, l m Hidden layer of the m-th layer of the second multi-layer perceptron network, W m B is a weight matrix of an mth hidden layer of the second multi-layer perceptron network m Bias term representing mth hidden layer of second multi-layer perceptron network, M is hidden layer number of second multi-layer perceptron network, b M And the bias item of the output layer of the second multi-layer perceptron network.
Wherein, the activation functions of the hidden layer and the output layer adopt a tanh function, which is expressed by the following formula:
Figure BDA0003492475560000081
in a specific embodiment, the first multi-layer perceptron network and the second multi-layer perceptron network each comprise an input layer, a hidden layer and an output layer, after the low-dimensional user feature vector and the low-dimensional active power feature vector are input to the input layer, activating and embedding feature vector calculation is performed in the hidden layer and the output layer, and then the finally obtained 128-dimensional vector is output as an embedding feature vector.
In step S5, the correlation between the user embedded feature vector and the load element embedded feature vector is calculated by the following formula:
Figure BDA0003492475560000091
wherein ,yU For embedding feature vectors for users, y L The feature vector is embedded for the load element,
Figure BDA0003492475560000092
representing a transpose of the user embedded feature vector, R (U, L) representing a correlation of the user embedded feature vector with the load element embedded feature vector.
In order to obtain parameters of the multi-layer perceptron network and the embedded feature vector output network, the parameters comprise weight matrixes and bias items, and the parameters of the multi-layer perceptron network and the embedded feature vector output network are obtained based on a maximum likelihood estimation method according to the condition probability that all loads are started on the premise of maximizing the total power data of power consumption of a given household user.
The posterior probability is calculated by the following formula:
Figure BDA0003492475560000093
wherein ,P(Lj U) is a posterior probability, γ is a smoothing factor of a logistic regression activation function, L represents a set of load elements, U represents a set of users, L' represents a sum of all load elements in the set of load elements, j is "+" or "-", L + Representing a positive sample of the load cell, L - Representing a load cell negative sample.
In one embodiment, the activation function may employ a softmax function to convert cosine similarity of the user and the load element to posterior probability, with gamma values typically set based on human experience.
Ideally, L should contain all load elements that may participate in the matching. For a scenario where one query of home users and turn on load element behavior occurs, use L + Representing the load that was turned on, i.e., positive samples, using negative sampling to randomly select 4 different loads as negative samples for which no turn-on behavior occurred, specifically expressed as
Figure BDA0003492475560000094
In training, model parameter estimation aims at maximizing the probability of load element start-up behavior to occur given the entire training of querying home users. Therefore, the minimization loss function employs maximum likelihood estimation, which is expressed by the following formula:
Figure BDA0003492475560000095
wherein Λ is a parameter of the multi-layer perceptron network, P (L + U) represents the posterior probability, U represents the set of users, L + Representing a positive sample of the load cell.
The posterior probability accuracy obtained according to the initial parameters of the double-tower model is low, so that the parameters of the model can be optimized through a random gradient descent method, the posterior probability is maximized, and the parameters of each layer of the double-tower model are updated.
Finally, the load element with the highest degree of association with the user is obtained, the obtained embedded feature vector can be used for comparing the similarity between the user and the load element, so that the load possibly started at the next moment can be found, the similarity between different family users can be compared, the family users with the same type can be found, and the information features of the family users can be collected.
Referring to FIG. 2, in some embodiments, there is provided a load cell construction apparatus for a representation of an electric utility, comprising:
the acquisition module 1 is used for acquiring the total power of the household users and the active power of a single load element as a training set;
the encoding module 2 is used for performing hot unique code encoding on the total power of the home user and the active power of the single load element to respectively obtain a user characteristic vector and an active power characteristic vector;
the processing module 3 is used for respectively carrying out hash processing on the user characteristic vector and the active power characteristic vector to respectively obtain a low-dimensional user characteristic vector and a low-dimensional active power characteristic vector;
the training module 4 is used for respectively inputting the low-dimensional user feature vector and the low-dimensional active power feature vector into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and respectively outputting each user embedded feature vector and each load element embedded feature vector;
and the calculating module 5 is used for calculating the correlation between each user embedded feature vector and the load element embedded feature vector based on cosine similarity, and converting the correlation calculation result into posterior probability by utilizing a logistic regression activation function to obtain the load element with the highest degree of correlation with the user.
Specifically, the encoding module 2 is further configured to perform hot-independent encoding on the total power of the home subscriber, and includes:
dividing the same family user total power into user similar characteristics;
the similar characteristics of each user are expressed by binary system, and the user characteristic vector is obtained;
performing thermal single code encoding on the active power of the single load element to obtain an active power characteristic vector of the single load element, including:
dividing the active power of the same single load element into active power similar characteristics;
and (5) representing each active power similar characteristic by using a binary system to obtain an active power characteristic vector.
The processing module 3 is further configured to hash the user feature vector to obtain a low-dimensional user feature vector, and includes:
respectively adding a start mark and an end mark before and after the user feature vector;
setting step length and sliding window based on N-GRAM algorithm, recording user characteristic vector slices of each scratch, recording all user characteristic vector slices as vector representations composed of a plurality of numbers, and obtaining low-dimension user characteristic vectors;
performing hash processing on the active power feature vector to obtain a low-dimension active power feature vector, including:
respectively adding a start mark and an end mark before and after the active power characteristic vector;
based on an N-GRAM algorithm, setting a step length and a sliding window, recording active power characteristic vector slices which are marked each time, recording all the active power characteristic vector slices as vector representations consisting of a plurality of numbers, and obtaining a low-dimensional active power characteristic vector.
The training module 4 is further configured to input the low-dimensional user feature vector and the low-dimensional active power feature vector to the first multi-layer perceptron network and the second multi-layer perceptron network, respectively, to perform training, and obtain parameters of the first multi-layer perceptron network and the second multi-layer perceptron network, respectively, according to a minimized loss function.
The parameters of the first multi-layer perceptron network and the second multi-layer perceptron network comprise weight matrixes and bias items of all hidden layers.
The load element embedding feature vector is calculated by the following formula:
l i =f(W i l i-1 +b i )i=2,…,N-1;
y L =f(W N l N-1 +b N );
wherein ,yL Embedding feature vectors for load elements, l i Hidden layer of the ith layer of the first multi-layer perceptron network, W i B is a weight matrix of an ith hidden layer of the first multi-layer perceptron network i The bias item representing the ith hidden layer of the first multi-layer perceptron network, wherein N is the number of hidden layers of the first multi-layer perceptron network;
the user-embedded feature vector is calculated by the following formula:
l m =f(W m l m-1 +b m )m=2,…,M-1;
y U =f(W M l M-1 +b M );
wherein ,yU Embedding feature vectors for users, l m Hidden layer of the m-th layer of the second multi-layer perceptron network, W m B is a weight matrix of an mth hidden layer of the second multi-layer perceptron network m And the bias item representing the M-th hidden layer of the second multi-layer perceptron network, wherein M is the number of hidden layers of the second multi-layer perceptron network.
The calculating module 5 is further configured to calculate a correlation between the user embedded feature vector and the load element embedded feature vector, where the calculation is performed by the following formula:
Figure BDA0003492475560000121
wherein ,yU for embedding feature vectors for users, y L The feature vector is embedded for the load element,
Figure BDA0003492475560000122
representing a transpose of the user embedded feature vector, R (U, L) representing a correlation of the user embedded feature vector with the load element embedded feature vector.
The posterior probability is calculated by the following formula:
Figure BDA0003492475560000123
wherein ,P(Lj U) is a posterior probability, γ is a smoothing factor of a logistic regression activation function, L represents a set of load elements, U represents a set of users, L' represents a sum of all load elements in the set of load elements, j is "+" or "-", L + Representing a positive sample of the load cell, L - Representing a load cell negative sample.
By setting L', traversing the load elements within the set L can be achieved.
The minimized loss function is represented by the following formula:
Figure BDA0003492475560000124
wherein Λ is a parameter of the multi-layer perceptron network, P (L + U) represents the posterior probability, U represents the set of users, L + Representing a positive sample of the load cell.
According to the load element construction method and device for the electric power user portrait, model training is carried out based on total power of a household user and active power of the load element, hash processing is carried out on the characteristics obtained through thermal unique code coding, multi-layer perceptron network training is carried out, similarity between the user and each load is generated by means of a double-tower model, and finally the load element with the highest degree of association with the user is obtained, wherein structural characteristics and node characteristics of a complex recommended network scene are fused, manual priori deviation is avoided to a certain extent, operation speed during online prediction is effectively improved, user portrait model construction efficiency is high, and reliability is strong.
While preferred embodiments of the present invention have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. It is therefore intended that the following claims be interpreted as including the preferred embodiments and all such alterations and modifications as fall within the scope of the invention. It will be apparent to those skilled in the art that various modifications and variations can be made to the present invention without departing from the spirit or scope of the invention. Thus, it is intended that the present invention also include such modifications and alterations insofar as they come within the scope of the appended claims or the equivalents thereof.

Claims (9)

1. A load cell construction method for an electricity consumer representation, comprising:
collecting total power of a household user and active power of a single load element as a training set;
performing thermal single code coding on the total power of the home user and the active power of a single load element to respectively obtain a user characteristic vector and an active power characteristic vector;
respectively carrying out hash processing on the user characteristic vector and the active power characteristic vector to respectively obtain a low-dimensional user characteristic vector and a low-dimensional active power characteristic vector;
the low-dimensional user feature vector and the low-dimensional active power feature vector are respectively input into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and each user embedded feature vector and each load element embedded feature vector are respectively output;
calculating the relevance of each user embedded feature vector and the load element embedded feature vector based on cosine similarity, and converting a relevance calculation result into posterior probability by utilizing a logistic regression activation function to obtain a load element with highest relevance to the user;
performing hot unique code encoding on the total power of the home user, including:
dividing the same family user total power into user similar characteristics;
the similar characteristics of each user are expressed by binary system, and the user characteristic vector is obtained;
performing thermal single code encoding on the active power of the single load element to obtain an active power characteristic vector of the single load element, including:
dividing the active power of the same single load element into active power similar characteristics;
and (5) representing each active power similar characteristic by using a binary system to obtain an active power characteristic vector.
2. The method of claim 1, wherein hashing the user feature vector to obtain a low-dimensional user feature vector comprises:
respectively adding a start mark and an end mark before and after the user feature vector;
setting step length and sliding window based on N-GRAM algorithm, recording user characteristic vector slices of each scratch, recording all user characteristic vector slices as vector representations composed of a plurality of numbers, and obtaining low-dimension user characteristic vectors;
performing hash processing on the active power feature vector to obtain a low-dimension active power feature vector, including:
respectively adding a start mark and an end mark before and after the active power characteristic vector;
based on an N-GRAM algorithm, setting a step length and a sliding window, recording active power characteristic vector slices which are marked each time, recording all the active power characteristic vector slices as vector representations consisting of a plurality of numbers, and obtaining a low-dimensional active power characteristic vector.
3. The method of claim 1, wherein the low-dimensional user feature vector and the low-dimensional active power feature vector are input to a first multi-layer perceptron network and a second multi-layer perceptron network, respectively, for training, and parameters of the first multi-layer perceptron network and the second multi-layer perceptron network are obtained, respectively, according to a minimization loss function.
4. The method of claim 3, wherein the parameters of the first and second multi-layer perceptron networks comprise a weight matrix and bias terms for each hidden layer.
5. The method of claim 4, wherein the load element embedded feature vector is calculated by the formula:
Figure QLYQS_1
Figure QLYQS_2
;/>
Figure QLYQS_3
wherein ,
Figure QLYQS_6
embedding feature vectors for load elements, < >>
Figure QLYQS_8
Conceal layer for the i-th layer of the first multi-layer perceptron network->
Figure QLYQS_9
Weight matrix for the i-th hidden layer of the first multi-layer perceptron network,/for the i-th hidden layer>
Figure QLYQS_5
Bias item representing ith hidden layer of first multi-layer perceptron network, number of hidden layers of first multi-layer perceptron network is N-2, < ->
Figure QLYQS_7
Bias items for the output layer of the first multi-layer perceptron network; />
Figure QLYQS_10
Input feature vector representing a thermal unique code, < ->
Figure QLYQS_11
Representing the input layer->
Figure QLYQS_4
Weight matrix representing input layer, W N A weight matrix of the output layer of the first multi-layer perceptron network;
the user-embedded feature vector is calculated by the following formula:
Figure QLYQS_12
Figure QLYQS_13
wherein ,
Figure QLYQS_14
embedding feature vectors for a user->
Figure QLYQS_15
Conceal layer for the m-th layer of the second multi-layer perceptron network->
Figure QLYQS_16
Weight matrix for the m-th hidden layer of the second multi-layer perceptron network,/for the m-th hidden layer>
Figure QLYQS_17
Bias item representing mth hidden layer of second multi-layer perceptron network, number of hidden layers of second multi-layer perceptron network is M-2,/or->
Figure QLYQS_18
Bias item W for output layer of second multi-layer perceptron network M And the weight matrix is the weight matrix of the second multi-layer perceptron network output layer.
6. The method of claim 5, wherein the correlation of the user embedded feature vector with the load element embedded feature vector is calculated by the following formula:
Figure QLYQS_19
wherein ,
Figure QLYQS_20
embedding feature vectors for a user->
Figure QLYQS_21
Embedding feature vectors for load elements, < >>
Figure QLYQS_22
Representing the transpose of the user's embedded feature vector, +.>
Figure QLYQS_23
The correlation of the user embedded feature vector with the load element embedded feature vector is represented.
7. The method of claim 6, wherein the posterior probability is calculated by the following formula:
Figure QLYQS_24
wherein ,
Figure QLYQS_25
for posterior probability>
Figure QLYQS_26
For the smoothing factor of the logistic regression activation function, L represents the set of load elements, U represents the set of users, L' represents the sum of all load elements in the set of load elements, j is "+" or "-", "and>
Figure QLYQS_28
representing a positive sample of the load cell, < >>
Figure QLYQS_29
Representing a load cell negative sample.
8. The method of claim 7, wherein the minimization of the loss function is represented by the following formula:
Figure QLYQS_30
wherein ,
Figure QLYQS_31
is a parameter of a multi-layer perceptron network, +.>
Figure QLYQS_32
Representing posterior probability, U representing the set of users, +.>
Figure QLYQS_33
Representing a positive sample of the load cell.
9. A load cell construction device for an electric power consumer figure, comprising:
the acquisition module is used for acquiring the total power of the household users and the active power of the single load element as a training set;
the coding module is used for carrying out hot unique code coding on the total power of the home user and the active power of the single load element to respectively obtain a user characteristic vector and an active power characteristic vector;
the processing module is used for respectively carrying out hash processing on the user characteristic vector and the active power characteristic vector to respectively obtain a low-dimensional user characteristic vector and a low-dimensional active power characteristic vector;
the training module is used for respectively inputting the low-dimensional user characteristic vector and the low-dimensional active power characteristic vector into a first multi-layer perceptron network and a second multi-layer perceptron network for training, and respectively outputting each user embedded characteristic vector and each load element embedded characteristic vector;
the computing module is used for computing the relevance of each user embedded feature vector and the load element embedded feature vector based on cosine similarity, converting a relevance computing result into posterior probability by utilizing a logistic regression activation function, and obtaining the load element with the highest relevance to the user;
the encoding module is also used for carrying out hot unique code encoding on the total power of the home user, and comprises the following steps:
dividing the same family user total power into user similar characteristics;
the similar characteristics of each user are expressed by binary system, and the user characteristic vector is obtained;
performing thermal single code encoding on the active power of the single load element to obtain an active power characteristic vector of the single load element, including:
dividing the active power of the same single load element into active power similar characteristics;
and (5) representing each active power similar characteristic by using a binary system to obtain an active power characteristic vector.
CN202210101090.8A 2022-01-27 2022-01-27 Load element construction method and device for electric power user portrait Active CN114511058B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210101090.8A CN114511058B (en) 2022-01-27 2022-01-27 Load element construction method and device for electric power user portrait

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210101090.8A CN114511058B (en) 2022-01-27 2022-01-27 Load element construction method and device for electric power user portrait

Publications (2)

Publication Number Publication Date
CN114511058A CN114511058A (en) 2022-05-17
CN114511058B true CN114511058B (en) 2023-06-02

Family

ID=81549435

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210101090.8A Active CN114511058B (en) 2022-01-27 2022-01-27 Load element construction method and device for electric power user portrait

Country Status (1)

Country Link
CN (1) CN114511058B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115017333A (en) * 2022-06-13 2022-09-06 四川大学 Method for converting modeless data of material genetic engineering into knowledge map
CN114881538B (en) * 2022-06-24 2022-12-13 东南大学溧阳研究院 Demand response user selection method based on perceptron

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109598446A (en) * 2018-12-09 2019-04-09 国网江苏省电力有限公司扬州供电分公司 A kind of tariff recovery Warning System based on machine learning algorithm
WO2019141040A1 (en) * 2018-01-22 2019-07-25 佛山科学技术学院 Short term electrical load predication method
WO2019238096A1 (en) * 2018-06-13 2019-12-19 国网江苏省电力有限公司 Method and apparatus for estimating weather-sensitive load power
CN110928990A (en) * 2019-10-31 2020-03-27 南方电网调峰调频发电有限公司 Method special for recommending standing book data of power equipment based on user portrait
CN111428355A (en) * 2020-03-18 2020-07-17 东南大学 Modeling method for power load digital statistics intelligent synthesis
CN111651668A (en) * 2020-05-06 2020-09-11 上海晶赞融宣科技有限公司 User portrait label generation method and device, storage medium and terminal
CN112598303A (en) * 2020-12-28 2021-04-02 宁波迦南智能电气股份有限公司 Non-invasive load decomposition method based on combination of 1D convolutional neural network and LSTM
CN113554241A (en) * 2021-09-02 2021-10-26 国网山东省电力公司泰安供电公司 User layering method and prediction method based on user electricity complaint behaviors

Family Cites Families (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101297318B (en) * 2005-08-23 2013-01-23 株式会社理光 Data organization and access for mixed media document system
JP5690296B2 (en) * 2012-03-01 2015-03-25 日本電信電話株式会社 Load balancing program and load balancing apparatus
CN104777383B (en) * 2015-04-16 2018-01-05 武汉阿帕科技有限公司 A kind of non-intrusion type electrical load monitoring and load decomposition device
CN106936129B (en) * 2017-03-23 2020-04-24 东北大学 Power load identification method and system based on multi-feature fusion
JP6832783B2 (en) * 2017-04-20 2021-02-24 株式会社日立製作所 Data analyzers, data analysis methods, and data analysis programs
CN107121210B (en) * 2017-05-19 2019-12-03 四川成瑞科技有限公司 Temp monitoring alarm, method and system
CN107578288B (en) * 2017-09-08 2020-09-18 东南大学 Non-invasive load decomposition method considering user power consumption mode difference
US20190325293A1 (en) * 2018-04-19 2019-10-24 National University Of Singapore Tree enhanced embedding model predictive analysis methods and systems
CN109101620B (en) * 2018-08-08 2022-07-05 阿里巴巴(中国)有限公司 Similarity calculation method, clustering method, device, storage medium and electronic equipment
CN109523057A (en) * 2018-10-18 2019-03-26 国网山东省电力公司经济技术研究院 A kind of regional power grid Methods of electric load forecasting considering economic transition background
CN110212542A (en) * 2019-05-28 2019-09-06 国网江苏省电力有限公司 A kind of accurate cutting load method and device
CN110472041B (en) * 2019-07-01 2021-08-03 浙江工业大学 Text classification method for customer service online quality inspection
JP6691280B1 (en) * 2019-07-30 2020-04-28 特許庁長官 Management system and management method
CN110635833B (en) * 2019-09-25 2020-12-15 北京邮电大学 Power distribution method and device based on deep learning
CN110866471A (en) * 2019-10-31 2020-03-06 Oppo广东移动通信有限公司 Face image quality evaluation method and device, computer readable medium and communication terminal
CN110991263B (en) * 2019-11-12 2022-03-18 华中科技大学 Non-invasive load identification method and system for resisting background load interference
CN113128494A (en) * 2019-12-30 2021-07-16 华为技术有限公司 Method, device and system for recognizing text in image
CN111489036B (en) * 2020-04-14 2023-06-09 天津相和电气科技有限公司 Resident load prediction method and device based on electrical load characteristics and deep learning
CN111428816B (en) * 2020-04-17 2023-01-20 贵州电网有限责任公司 Non-invasive load decomposition method
CN111612319A (en) * 2020-05-11 2020-09-01 上海电力大学 Load curve depth embedding clustering method based on one-dimensional convolution self-encoder
CN111799782A (en) * 2020-06-29 2020-10-20 中国电力科学研究院有限公司 Power equipment power failure window period correction method and system based on machine learning
CN111860977B (en) * 2020-06-30 2022-07-01 清华大学 Probability prediction method and probability prediction device for short-term load
CN111753058B (en) * 2020-06-30 2023-06-02 北京信息科技大学 Text viewpoint mining method and system
CN111949707B (en) * 2020-08-06 2022-03-22 杭州电子科技大学 Shadow field-based hidden Markov model non-invasive load decomposition method
CN112379159B (en) * 2020-11-09 2021-10-26 北华航天工业学院 Non-invasive household load decomposition method based on electric appliance operation mode
CN112364203B (en) * 2020-11-30 2023-04-28 未来电视有限公司 Television video recommendation method, device, server and storage medium
AU2020104000A4 (en) * 2020-12-10 2021-02-18 Guangxi University Short-term Load Forecasting Method Based on TCN and IPSO-LSSVM Combined Model
CN113094198A (en) * 2021-04-13 2021-07-09 中国工商银行股份有限公司 Service fault positioning method and device based on machine learning and text classification
CN113761250A (en) * 2021-04-25 2021-12-07 腾讯科技(深圳)有限公司 Model training method, merchant classification method and device
CN113687182A (en) * 2021-08-12 2021-11-23 湖南工学院 Household load identification method, program and system based on noise reduction automatic encoder
CN113935401A (en) * 2021-09-18 2022-01-14 北京三快在线科技有限公司 Article information processing method, article information processing device, article information processing server and storage medium
CN113887833A (en) * 2021-10-28 2022-01-04 西安热工研究院有限公司 Distributed energy user side time-by-time load prediction method and system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019141040A1 (en) * 2018-01-22 2019-07-25 佛山科学技术学院 Short term electrical load predication method
WO2019238096A1 (en) * 2018-06-13 2019-12-19 国网江苏省电力有限公司 Method and apparatus for estimating weather-sensitive load power
CN109598446A (en) * 2018-12-09 2019-04-09 国网江苏省电力有限公司扬州供电分公司 A kind of tariff recovery Warning System based on machine learning algorithm
CN110928990A (en) * 2019-10-31 2020-03-27 南方电网调峰调频发电有限公司 Method special for recommending standing book data of power equipment based on user portrait
CN111428355A (en) * 2020-03-18 2020-07-17 东南大学 Modeling method for power load digital statistics intelligent synthesis
CN111651668A (en) * 2020-05-06 2020-09-11 上海晶赞融宣科技有限公司 User portrait label generation method and device, storage medium and terminal
CN112598303A (en) * 2020-12-28 2021-04-02 宁波迦南智能电气股份有限公司 Non-invasive load decomposition method based on combination of 1D convolutional neural network and LSTM
CN113554241A (en) * 2021-09-02 2021-10-26 国网山东省电力公司泰安供电公司 User layering method and prediction method based on user electricity complaint behaviors

Also Published As

Publication number Publication date
CN114511058A (en) 2022-05-17

Similar Documents

Publication Publication Date Title
Zhang et al. Improved deep hashing with soft pairwise similarity for multi-label image retrieval
CN112214685B (en) Knowledge graph-based personalized recommendation method
CN111291836B (en) Method for generating student network model
CN114511058B (en) Load element construction method and device for electric power user portrait
CN111414461B (en) Intelligent question-answering method and system fusing knowledge base and user modeling
CN111897913A (en) Semantic tree enhancement based cross-modal retrieval method for searching video from complex text
CN101346718A (en) Method for providing user of chosen content item
CN109033107A (en) Image search method and device, computer equipment and storage medium
CN113177141B (en) Multi-label video hash retrieval method and device based on semantic embedded soft similarity
CN113343125B (en) Academic accurate recommendation-oriented heterogeneous scientific research information integration method and system
CN114693397B (en) Attention neural network-based multi-view multi-mode commodity recommendation method
CN109272332B (en) Client loss prediction method based on recurrent neural network
CN111737578A (en) Recommendation method and system
CN111159485A (en) Tail entity linking method, device, server and storage medium
CN112800344B (en) Deep neural network-based movie recommendation method
CN112860930B (en) Text-to-commodity image retrieval method based on hierarchical similarity learning
CN113569001A (en) Text processing method and device, computer equipment and computer readable storage medium
CN114462420A (en) False news detection method based on feature fusion model
CN115408603A (en) Online question-answer community expert recommendation method based on multi-head self-attention mechanism
CN114781503A (en) Click rate estimation method based on depth feature fusion
CN114004220A (en) Text emotion reason identification method based on CPC-ANN
CN116244484B (en) Federal cross-modal retrieval method and system for unbalanced data
CN116108836B (en) Text emotion recognition method and device, computer equipment and readable storage medium
CN111859979A (en) Ironic text collaborative recognition method, ironic text collaborative recognition device, ironic text collaborative recognition equipment and computer readable medium
CN115952360A (en) Domain-adaptive cross-domain recommendation method and system based on user and article commonality modeling

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant