WO2022079280A1 - Dispositifs et processus de sélection d'échantillons de données pour gestion de consommation d'énergie - Google Patents

Dispositifs et processus de sélection d'échantillons de données pour gestion de consommation d'énergie Download PDF

Info

Publication number
WO2022079280A1
WO2022079280A1 PCT/EP2021/078681 EP2021078681W WO2022079280A1 WO 2022079280 A1 WO2022079280 A1 WO 2022079280A1 EP 2021078681 W EP2021078681 W EP 2021078681W WO 2022079280 A1 WO2022079280 A1 WO 2022079280A1
Authority
WO
WIPO (PCT)
Prior art keywords
samples
past
share
power consumption
data samples
Prior art date
Application number
PCT/EP2021/078681
Other languages
English (en)
Inventor
Gaëtan MILLERAND
Matthieu BOUSSARD
François Pellissier
Chahram BECHARAT
Original Assignee
Craft.Ai
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from EP20306227.8A external-priority patent/EP3985582A1/fr
Priority claimed from US17/072,330 external-priority patent/US11539598B2/en
Application filed by Craft.Ai filed Critical Craft.Ai
Publication of WO2022079280A1 publication Critical patent/WO2022079280A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management

Definitions

  • the invention relates to the domain of power consumption management, and regards more particularly the upstream selection of proper data samples.
  • Another aspect of power management is the ability to identify failures, so as to avoid or promptly overcome potential damages, ranging from useless expenses or comfort inconvenience to serious accidents.
  • ML machine learning
  • a particular aspect of those ML constraints amount to an appropriate tradeoff between remembering and forgetting. Indeed, due to restraining bounds in computing resources and to resulting approximation models, remembering too much past information, i.e. usually too old information, may disturb the efficacy of the system. This could e.g. take the form of outdated data interfering with presently relevant ones, or of behavioral interpretation distorted by past model adjustments. On the other hand, while forgetting techniques exist for disposing of old data, they may naturally be prejudicial in a number of situations, in particular where rare events need consideration. This difficulty is well known to the person skilled in the art as the sensitivity- stability dilemma or stability-plasticity dilemma.
  • time forgetting a strategy known as time forgetting consists in sliding a time window applied to an incoming datastream and deleting oldest samples as and when new samples are added.
  • Such a method is described by X. H. Dang et al. in “Incremental and adaptive clustering stream data over sliding window”, International Conference on Database and Expert Systems Applications, DEXA ’09, pages 660-674, 2009. This may prove appropriate for relatively stable behaviors between successive time window periods, and where time windows are not too large.
  • document US-2019/0122132 Al (to GRID4C) describes a method for identifying abnormalities in personal energy consumption.
  • a personal dynamic forecast model of energy usage patterns is generated in sub-hourly resolution for defined time periods, and an adaptive gradient boost iterative learning algorithm is dynamically applied.
  • the latter uses predefined periodical features and determining abnormalities of actual energy usage by comparing predictions (relying on regression-tree based historical usage data) and actual usage.
  • SARIMA Seasonal Autoregressive Integrated Moving Average
  • a drawback of this model is that tuning the associated parameters makes it too specific and may not leave enough flexibility to learn new behaviors that can occur after a first training time.
  • RNNs Recurrent Neural Networks
  • connections between neurons form a directed graph along a temporal sequence, which provides a temporal dynamic behavior. This ensures that concept drifts are taken into account.
  • a traditional RNN is exposed to a gradient vanishing issue while backpropagating errors through the network, which makes it unsuited to learn long term dependency. A consequence is that it usually misbehaves on anomaly detection.
  • An RNN with a gated state or gated memory (in which the internal state is under a direct control by the neural network) enables to overcome that difficulty, by enabling to store long-term dependencies. It can in particular take the form of a Long Short-Term Memory (LSTM), comprising a cell and three gates regulating the time flow of information into and out of the cell, or of a Gate Recurrent Unit (GRU), comprising a cell and two such gates.
  • LSTM Long Short-Term Memory
  • GRU Gate Recurrent Unit
  • RNNs are exploited in a dual way: a first network set aids in pattern recognition of an input sequence or classification of a typical event or set of behavior, while a second network set branched to outputs of the first network set is dedicated to each of various target appliances or to various time schedules, so as to help dynamically validate real-time outputs from the first network set on key properties, to test data or to recognize a separate set of characteristics.
  • the second network set may further be arranged to re-enforce or train the first network set.
  • Such implementations may help to alleviate to some extent the previously mentioned issues, but however subject to specific arrangements and preprocessing configurations, thereby potentially requiring demanding developments when dealing with various situations.
  • Such models may be particularly suited to the processing of data streams, because increasing weights in the various leaves may provide a balance over time between learned behaviors. This may further ensure a proper mitigated forgetting, because oldest samples corresponding to a common behavior, i.e. falling in leaves that have many samples, are forgotten, while rare samples are kept over several learning periods.
  • a purpose of the present disclosure is to provide a flexible tool for power consumption management, which may be potentially efficient notably when dealing with datastreams.
  • the present disclosure is also directed to judiciously selecting items among available data samples, which may be then exploited for power consumption management, notably via machine learning operations.
  • the disclosure sets forth solutions that may potentially provide a flexible and efficient tradeoff between remembering and forgetting, thereby enabling to avoid excessive or prohibitive computation costs without sacrificing the relevance of the selected data samples to effective and reliable power consumption management.
  • the present disclosure may further apply to various fields of power consumption management, including notably centralized power supply operations, failure detection, autonomous local power management, collective interactive energy sharing, energy cost control, distributed management of energy networked systems, and controlled integration of renewables in energy mix.
  • power consumption management including notably centralized power supply operations, failure detection, autonomous local power management, collective interactive energy sharing, energy cost control, distributed management of energy networked systems, and controlled integration of renewables in energy mix.
  • the term “power consumption management” broadly designates dealing with or controlling operations pertaining to using up energy sources, whatever their nature (e.g. electric, solar, wind, hydraulic, nuclear, hydrokinetic, geothermal, fossil), the management stage (e.g. supply, storage, exploitation, distribution, sharing), the exploitation scale (e.g. individual, collective) and the exploitation frame (e.g. household, commercial, industrial).
  • up energy sources whatever their nature (e.g. electric, solar, wind, hydraulic, nuclear, hydrokinetic, geothermal, fossil)
  • the management stage e.g. supply, storage, exploitation, distribution, sharing
  • the exploitation scale e.g. individual, collective
  • the exploitation frame e.g. household, commercial, industrial.
  • Data samples related to power consumption refers to any kind of information items potentially relevant to power consumption management, and may include valued, nominal or ordinal data, while a single data sample may be unidimensional or multidimensional.
  • Such data samples may e.g. include any combination of levels of power consumption, supply or demand, outside weather conditions such as temperature, humidity or wind, inside environment conditions such as temperature or humidity, occupancy patterns, schedule information such as time of day, day of week, month or season, activity patterns such as work or leisure, energy-related financial expenses, power storage levels, type of energy source, localization such as household room, building office, or P2P energy trading unit.
  • decisional guidance relevant to power consumption management information, instructions or analysis results leading to some actions regarding power consumption management.
  • This decisional guidance may be automatically executed in machine processes or by actuators. It may provide clear and complete instructions or indications to one or more users or operators as to how to proceed. It may give decisional information pertaining e.g. to a failure or anomaly, so that users or operators can properly react. It may provide instructions or indications pertaining to parts of complex systems, so that proper decisions should be taken on complementary grounds, whether automatically or by appropriate operators, taking account of other factors pertaining to those complex systems.
  • the terms “adapted” and “configured” are used in the present disclosure as broadly encompassing initial configuration, later adaptation or complementation of the present device, or any combination thereof alike, whether effected through material or software means (including firmware).
  • processor should not be construed to be restricted to hardware capable of executing software, and refers in a general way to a processing device, which can for example include a computer, a microprocessor, an integrated circuit, or a programmable logic device (PLD).
  • the processor may also encompass one or more Graphics Processing Units (GPU), whether exploited for computer graphics and image processing or other functions.
  • GPU Graphics Processing Unit
  • the instructions and/or data enabling to perform associated and/or resulting functionalities may be stored on any processor-readable medium such as, e.g., an integrated circuit, a hard disk, a CD (Compact Disc), an optical disc such as a DVD (Digital Versatile Disc), a RAM (Random- Access Memory) or a ROM (Read-Only Memory). Instructions may be notably stored in hardware, software, firmware or in any combination thereof.
  • processor-readable medium such as, e.g., an integrated circuit, a hard disk, a CD (Compact Disc), an optical disc such as a DVD (Digital Versatile Disc), a RAM (Random- Access Memory) or a ROM (Read-Only Memory).
  • Instructions may be notably stored in hardware, software, firmware or in any combination thereof.
  • a similarity consists in the identification of equivalence relationships between objects based on a similarity measure (i.e. a real-valued function that quantifies the similarity between two objects) applied to couples of objects, two objects being equivalent when they fall in a same equivalence class or cluster.
  • the objects may be unidimensional or multi-dimensional, and may pertain to valued data (metric processing) as well as to nominal or ordinal data (non-metric processing).
  • the considered equivalence may be structural, automorphic or regular (as known to a person skilled in the art).
  • Machine learning designates in a traditional way computer algorithms improving automatically through experience, on the ground of training data enabling to adjust parameters of computer models through gap reductions between expected outputs extracted from the training data and evaluated outputs computed by the computer models.
  • a white box is a subsystem whose internals can be viewed, while a black box can only be viewed in terms of its inputs and outputs or of its transfer characteristics.
  • a hyper-parameter presently means a parameter used to carry out an upstream control of a model construction, such as a remembering-forgetting balance in sample selection or a width of a time window, by contrast with a parameter of a model itself, which depends on specific situations.
  • hyper-parameters are used to control the learning process.
  • Datasets are collections of data used to build an ML mathematical model, so as to make data-driven predictions or decisions.
  • supervised learning i.e. inferring functions from known input-output examples in the form of labelled training data
  • three types of ML datasets are typically dedicated to three respective kinds of tasks: training, i.e. fitting the parameters, validation, i.e. tuning ML hyperparameters (which are parameters used to control the learning process), and testing, i.e. checking independently of a training dataset exploited for building a mathematical model that the latter model provides satisfying results.
  • a Decision Tree is a decision support tool using a tree-like model of decisions, having internal nodes, branches and leaf nodes (or leaves), each internal node amounting to testing an attribute, each branch representing an outcome of such testing, and each leaf representing a class label associated with a decision taken from attribute computations.
  • Each leaf may notably be viewed as an equivalence class or cluster, thereby providing a similarity model with structural equivalence.
  • Gradient boosting is an ML method that combines iteratively weak ML prediction models (such as e.g. Decision Trees) into a single stronger ML prediction model, by optimizing a cost function over a function space by an iterative gradient descent (i.e. by iteratively choosing a function pointing in the negative gradient direction).
  • iteratively weak ML prediction models such as e.g. Decision Trees
  • iterative gradient descent i.e. by iteratively choosing a function pointing in the negative gradient direction.
  • Reservoir sampling is a randomized algorithm for maintaining a reservoir having a preset size from a population of items having an unknown size and revealed over time, in a single pass over the items, by initializing the reservoir with the first items of the population and then iteratively and randomly selecting in the population an item to be selected and an associated replacement position in the reservoir.
  • a neural network or artificial neural network designates a category of ML comprising nodes (called neurons), and connections between neurons modeled by weights. For each neuron, an output is given in function of an input or a set of inputs by an activation function. Neurons are generally organized into multiple layers, so that neurons of one layer connect only to neurons of the immediately preceding and immediately following layers.
  • An ensemble model is an ML model that uses multiple constituent ML algorithms to obtain better predictive performance than the constituent ML algorithms taken alone.
  • a random forest is a particular ensemble model relying on constructing a multitude of decision trees at training time, and on outputting results based on the joint consideration of the individual trees, i.e. the mode of their classes in classification, and the mean prediction in regression.
  • An object of the present disclosure is notably a device for selecting data samples related to power consumption so as to proceed with power consumption management.
  • the device comprises:
  • At least one input adapted to receive available data samples related to power consumption and associated with respective sample times, those available data samples being distributed into a sliding time window including at least one new batch of data samples, the available data samples being then called current samples, and into a past period preceding said time window, the available data samples being then called past samples;
  • - at least one processor configured for determining in the past period, selected past samples to be kept for the power consumption management, by eliminating part of the past samples in function of at least some of the current samples and of elimination conditions depending on at least one similarity criterion applied to at least the past samples; - at least one output adapted to provide the selected past samples as a complement to the current samples for the power consumption management.
  • the processor(s) is/are configured for determining the selected past samples by more precisely keeping among the past samples:
  • a second share of the past samples through eliminating among the past samples deprived from the first share, called a complementary share of the past samples, part of the past samples in function of the at least some of the current samples, and of the elimination conditions depending on the similarity criterion/criteria applied to at least the first share and the complementary share of the past samples.
  • the data samples related to power consumption may be repeatedly received, including possibly via streaming. They may be derived from automatic measurements, e.g. from meters, sensors or monitoring units.
  • the received available data samples may be derived from a preselection among a broader set of received data samples, e.g. through random sampling executed on the flight on incoming samples. This may notably prevent an excessive amount of entered data while ensuring a representativeness of the gathered available data samples within the sliding time window.
  • a reservoir sampling strategy may be exploited. In this way, all incoming samples have the same probability to be in the reservoir, the updated probability of the reservoir making it possible to do it on the flight.
  • the sliding time window may have a constant length, e.g. corresponding to a cyclic behavior such as a week, a month or a year. It may be set by a user.
  • the samples are said “distributed” into the sliding time window and the past period for sake of conciseness, for meaning that they are associated with respective times located in the relevant time periods.
  • the targeted downstream power consumption management may be automatized, e.g. by machine-controlled supply, storage or consumption power adjustment, or by system fixing of detected anomalies. It may be repeated over time, e.g. iteratively, based on progressively incoming information.
  • the first share of the past samples is used together with the complementary share of the past samples in applying the similarity criterion/criteria on which the elimination conditions depend, while that first share is immune from deletion contrary to the complementary share.
  • the current samples at least some of them are determining in the deletion of some of the past samples, through the elimination conditions. They may be kept together with the selected past samples for the power consumption management.
  • the device for selecting data samples may be coupled to a device for providing power consumption management information, instructions or guidance (included in the form of a joint device grouping related functionalities), which may implement machine learning. Both devices may rely on same or like similarity criteria.
  • the device for selecting data samples as disclosed may be particularly efficient for avoiding excessive computation costs while potentially enabling reliable power consumption management. This is due to its dual nature, combining the exploitation of the first share and of the second share.
  • keeping the first share of the past samples amounts to some kind of time forgetting (or random forgetting), in which deleted past samples are associated with times lying before the joining of the time window and of a preceding time interval.
  • keeping the second share of the past samples amounts to some kind of generalization of leaf forgetting, presently called “similarity forgetting”, in which clusters may be obtained from other models than decision trees, or similarity criteria may rely on similarity measures without involving clusters, while part of the past samples (i.e. the first share) is immune from deletion.
  • time forgetting offers a potentially powerful tool for efficiently facing the sensitivity-stability dilemma, able to quickly follow concept drifts, while keeping track of older behaviors helpful in e.g. detecting anomalies.
  • similarity forgetting (the second share) may allow to keep only track of rare behaviors throughout ages thanks to exploiting similarity criteria in sample deletion. For example, samples falling in common clusters may be more prone to be replaced by new ones, because more new samples fall in those clusters. This may give the possibility to remind behaviors on a long term scale while updating quickly common behaviors and following global concept drifts.
  • time forgetting (the first share) may allow to turn down excessively old behaviors while preserving most recent past information, the deletion actions being focused on the past samples preceding the first share.
  • a remarkable property of the present device for selecting data samples is that in some embodiments, it may enable a controlled balance between time forgetting and similarity forgetting, which may be stricken in function of the targeted kind of power consumption management. This may be reflected in the respective sizes (i.e. the numbers of the past samples) of the first share and of the second share.
  • an appropriate tuning suited to detecting anomalies related to seasonality on an electric network may emphasize similarity forgetting rather than time forgetting so as to keep odd behavior in memory, while setting a time window large enough to take into account a past year.
  • a proper tuning may on the contrary substantially emphasize time forgetting rather than similarity forgetting, so as to forget old behavior as fast as possible.
  • the balance between time forgetting and similarity forgetting may be modified along time, included possibly along successive positions of the sliding time window.
  • the device for selecting data samples may further make possible a basically white box solution, in particular when using decision trees, potentially providing fully and readily explainable results, by contrast with ANN solutions such as e.g. LSTM or GRU.
  • ANN solutions such as e.g. LSTM or GRU.
  • the input(s) is/are adapted to repeatedly receive over time updated sets of the available data samples, derived from keeping in previous sets of the available data samples at least part of the current samples and of the selected past samples;
  • the processor(s) is/are configured for repeatedly determining the selected past samples among the updated sets of the available data samples;
  • the output(s) is/are adapted to repeatedly provide the selected past samples of the updated sets, for dynamically proceeding with the power consumption management over time.
  • the processor(s) is/are configured for eliminating oldest ones of the past samples in respective clusters of the available data samples obtained from the similarity criterion/criteria applied to at least the first share and the complementary share of the past samples, in function of the elimination conditions depending on assignments to the clusters of the at least some of the current samples.
  • the oldest past samples eliminated in the respective clusters are then considered at cluster levels, rather than at global level. Namely, some past samples may be deleted while older ones may be kept, insofar as the latter belong to other clusters than the former.
  • the processor(s) is/are configured for eliminating in the respective clusters the oldest ones of the past samples through taking account of the at least some of the current samples in a chronological order of the current samples.
  • the time arrival of the current samples is then decisive for the identification and elimination of the oldest past samples in the respective clusters, i.e. for properly focusing on the appropriate clusters and on their respective oldest past samples.
  • the current samples may drive away the concerned past samples as and when considered, which may be particularly suited notably to a streaming system.
  • the elimination conditions for at least one of the clusters include a cumulated amount of the available data samples in that cluster being above a preset threshold.
  • This may take the form of weights assigned to the samples in the cluster.
  • the clusters are built from at least one decision tree model, the clusters corresponding to leaf nodes.
  • one or more decision tree models may be produced, and the similarity criteria may correspond to belonging to same leaves of the built trees.
  • the same may apply to tree-based models such as notably random forests, or the gradient boosting framework known as XGBoost (open-source software library) and described e.g. by T. Chen and C. Guestrin in “XGBoost: a scalable tree boosting system”, Proceedings of the 22 nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 785-794, 2016.
  • XGBoost open-source software library
  • the similarity criterion/criteria used for eliminating the part of the complementary share is given by lowest prediction errors associated with the past samples in relation with a machine learning model for the power consumption management, that machine learning model providing predictions based on at least the first share and the complementary share of the past samples.
  • the processor(s) is/are configured for eliminating the part of the past samples based on the similarity criterion/criteria being applied to the past samples irrespective of the current samples.
  • the current samples may be ignored in the formation of the clusters, while at least some of them are determining in the deletion of part of the past samples, through the elimination conditions applied to the clusters.
  • the current samples may be ignored in the predictions of the ML model providing the prediction errors, while at least some of those current samples are determining in the deletion of part of the past samples, through e.g. proceeding with replacement operations based on the total number of the current samples.
  • the processor(s) is/are configured for determining the first share and the second share of the past samples so that the first share and the second share have relative proportions given by a hyper-parameter.
  • the input(s) is/are adapted to receive evaluations of relevance of the selected past samples, those evaluations of relevance being respectively associated with successive time positions of the time window and determined with respect to machine learning processing for the power consumption management corresponding to next time positions of the time window;
  • the processor(s) is/are configured for determining the hyper-parameter by selecting among at least two candidate values of the hyper-parameter a best candidate value providing smallest averaged errors associated with the evaluations of relevance over the successive time positions of the time window.
  • This hyper-parameter determination may thus be provided by a rolling validation, proper to time-series data, ensuring that the evaluation is based on not yet exploited data rather than already seen data.
  • a model may notably be built for every time position of the time window and evaluated with the data of the next time position of the time window, obtained errors on each validation fold being averaged over all the considered successive positions of the time window.
  • This strategy may be able to detect and take account of concept drifts in data, which often correspond to an important drop of predictive performances of a model.
  • the number of tested candidate values may e.g. be comprised between 5 and 12 (included), and more precisely between 8 and 10 (included).
  • the width of the sliding time window may be automatically adjusted likewise.
  • the power consumption management includes power failure detection and/or power consumption dynamic adaptation.
  • the similarity criterion/criteria includes at least one criterion used in machine learning processing for obtaining decisional guidance relevant to the power consumption management.
  • This may establish a strong connection between the selection of the kept data samples and the downstream processing of those data samples, thereby potentially providing particularly efficient synergies.
  • the at least one similarity criterion may be fully applied at the data sample selection stage and then exploited in obtaining the decisional guidance.
  • the disclosure also relates to a system for power consumption management. That system comprises a device for selecting data samples according to the present disclosure, and a device for providing decisional guidance relevant to the power consumption management based on the selected past samples and the current samples through machine learning processing.
  • Another object of the disclosure is a method for selecting data samples related to power consumption so as to proceed with power consumption management.
  • the method comprises:
  • the method comprises determining with the processor(s) the selected past samples by more precisely keeping among the past samples:
  • a second share of the past samples through eliminating among the past samples deprived from the first share, called a complementary share of the past samples, part of the past samples in function of the at least some of the current samples, and of the elimination conditions depending on the similarity criterion/criteria applied to at least the first share and the complementary share of the past samples.
  • the method for selecting data samples is advantageously executed by a device for selecting data samples according to any of the embodiments of the disclosure.
  • a further object of the disclosure is a method for power consumption management comprising a method for selecting data samples according to the disclosure, and a method for providing decisional guidance relevant to the power consumption management based on the selected past samples and the current samples through machine learning processing.
  • the disclosure relates to a computer program comprising software code adapted to perform a method for selecting data samples and/or a method for power consumption management according to the disclosure, when it is executed by a processor.
  • the present disclosure further pertains to a non-transitory program storage device, readable by a computer, tangibly embodying a program of instructions executable by the computer to perform a method for selecting data samples and/or a method for power consumption management, compliant with the present disclosure.
  • a non-transitory program storage device can be, without limitation, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor device, or any suitable combination of the foregoing.
  • Figure 1 is a block diagram representing schematically a particular mode of a system for power consumption management including a device for selecting data samples, compliant with the present disclosure
  • Figure 2A, 2B, 2C, 2D schematically illustrates various applications of the system for power consumption management of figure 1 in individual, collective and interacting energy use situations;
  • Figure 3 represents a timeline with successive positions of a sliding time window, as exploited in the device for selecting data samples of figure 1;
  • Figure 4 focuses on a particular part of the timeline of figure 3, pertaining to data samples exploited in the device for selecting data samples of figure 1 while preceding the sliding time window (past samples);
  • Figure 5 focuses on the processing of the past samples by the device for selecting data samples of figure 1;
  • Figure 6 schematically illustrates a decision tree relied on in a first processing mode of the past samples of figure 5;
  • Figure 7 shows the mechanism exploited with the decision tree of figure 6 in one of its leaves
  • Figure 8 schematically illustrates a distribution of data samples in function of machine learning prediction errors, as relied on in a second processing mode of the past samples of figure 5;
  • Figure 9 represents the exploitation of a similarity ratio in tuning a balance of a first share and a second share of the past samples in the processing of figure 5;
  • Figure 10 illustrates a model for assessing the relevance of an instance of the similarity ratio of figure 9, in relation with machine learning processing
  • Figure 11 show the exploitation of the model of figure 10 in determining a proper value of the similarity ratio of figure 9;
  • Figures 12A is a flow chart showing successive steps executed with the system for power consumption management of figure 1 ;
  • Figures 12B details steps of data sample selection in the flow chart of figure 12A
  • Figures 13 is a flow chart showing successive steps executed with the device for selecting data sample of figure 1 in determining a proper value of the similarity ratio dealt with in figures 9 to 11;
  • Figure 14 diagrammatically shows an apparatus integrating the functions of the system for power consumption management of figure 1.
  • the functions of the various elements shown in the figures may be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software.
  • the functions may be provided by a single dedicated processor, a single shared processor, or a plurality of individual processors, some of which may be shared.
  • Each of the devices 10 and 15 is advantageously an apparatus, or a physical part of an apparatus, designed, configured and/or adapted for performing the mentioned functions and produce the mentioned effects or results.
  • any of the device 10 and the device 15 is embodied as a set of apparatus or physical parts of apparatus, whether grouped in a same machine or in different, possibly remote, machines.
  • the device 10 and/or the device 15 may e.g. have functions distributed over a cloud infrastructure and be available to users as a cloud-based service, or have remote functions accessible through an API.
  • the device 10 for selecting data samples and the device 15 for providing decisional guidance may be integrated in a same apparatus or set of apparatus corresponding to the system 1 for power consumption management, and intended to same users.
  • the structure of device 10 may be completely independent of the structure of device 15, and may be provided for other users.
  • the device 10 may be exploited by a dedicated operator proposing proper upstream online selection of data samples to power management entities provided with ML capabilities embodied in the device 15, as subscribeware services (a.k.a. SaaS, for Software as a Service).
  • the system 1 for power consumption management may have various kinds of applications, and be e.g. directed to private or professional energy consumers, individuals or collectivities, power sharing networked customers (such as with P2P energy trading), power suppliers, energy storage players, power network maintenance teams. It may be suited to proper balancing of non-renewable and renewable energy sources.
  • the system 1 may be suited to one or more target goals integrated in the internal processes of the system 1.
  • Those goals may comprise any of e.g. decreasing energy losses, reducing power bills, mitigating incident risks, raising the share of renewables in an energy mix, supporting disabled people by facilitated energy management, timely adjusting power supply to industrial, customer or household needs, reducing levels of inside or outside pollution.
  • the system 1 may rely on ML operations, and automatically produce, set or adjust ML models, based on the processed data samples and the target goals as known to a skilled person.
  • the system 1 may implement reinforcement learning (in which interactions with a dynamic environment for performing a goal are exploited for maximizing feedbacks amounting to rewards), associated with any of the above-mentioned goals.
  • unsupervised learning in which algorithms find themselves appropriate structures in raw inputs, i.e. without learning stage with labelled inputs
  • feature learning in which representations needed for feature detection or classification are automatically discovered
  • the system 1 may exploit supervised learning, for example for initializing appropriate mappings between input information and targeted goals.
  • the system 1 may rely on one or more expert systems, fed with the data samples selected by the device 10.
  • the system 1 may be fully automatized, the decisional guidance 35 being directly exploited for controlling, directing or adjusting proper actuators, or being transferred to one or more processors for intermediate operations before effective energy management circuit control.
  • the system 1 may provide the decisional guidance 35 as user-directed information, so as to drive users to undertake appropriate actions or to prompt proper reactions.
  • the system 1 may alert on risks linked to automatically detected anomalies, or to potentially damaging or dangerous failures.
  • the system 1 may indicate behavior, structure or networking changes required in view of one or more target goals.
  • the system 1 is configured for repeated consideration over time of incoming data samples. This is all the more attractive in ML operations, in which the data samples may contribute to evolving the ML model.
  • the system 1 may be more particularly suited to datastreams, being thereby able to regularly or continuously process new batches of incoming information.
  • the system 1 is adapted to deal with a single set of data samples previously obtained over time.
  • the system 1 is exploited for punctual anomaly detection, analysis or diagnostic, e.g. in preventing power failures.
  • the system 1 is coupled with a software engine dealing with an incoming stream of data samples and exploited for power consumption management in a simpler way, less demanding in computing resources than the system 1 (e.g. from an expert system while the system 1 relies on ML operations, or based on simplified supervised ML while the system 1 relies on reinforcement learning), so as to complement and strengthen on demand the relevance of decisional guidance.
  • the data samples input to the system 1 are associated with respective sample times, e.g. via time stamps, and may comprise any kind of information pertaining to power consumption management, including targeted goals. They may notably comprise any selection among or combination of: power consumption or supply load or demand, monitored information from an outside environment, such as e.g. temperature, humidity, wind or other weather condition, or from an inside environment, such as e.g. temperature, humidity or atmospheric composition (such as CO2 content), temporal information (irrespective of and complementary to the sample times) such as time of day, day of week, month or season, activity patterns such as working or leisure, occupancy patterns such as an absence of any people or a number of present people, locations in concerned buildings (e.g. floor, particular or open-plan office, reception desk), houses (e.g. rooms, stairs), apartments, factories, industrial plants or facilities.
  • monitored information from an outside environment, such as e.g. temperature, humidity, wind or other weather condition, or from an inside environment, such as e.g. temperature, humidity or atmospheric composition (such as CO
  • the data samples may be provided to the system 1 via any appropriate communication network involving wired (e.g. Ethernet), wireless (e.g. WiFi, WiMax - standing for Worldwide interoperability for Microwave Access, or Bluetooth) or cellular (e.g. UMTS - standing for Universal Mobile Telecommunications System, LTE - standing for Long-Term Evolution, or 5G) transmissions, as well known to a person skilled in the field.
  • wired e.g. Ethernet
  • wireless e.g. WiFi, WiMax - standing for Worldwide interoperability for Microwave Access, or Bluetooth
  • cellular e.g. UMTS - standing for Universal Mobile Telecommunications System, LTE - standing for Long-Term Evolution, or 5G
  • They may be automatically obtained e.g. by sensors, meters, detectors or monitors. Alternatively or in combination, at least part of them may be user-entered.
  • FIG. 2A power consumption in a house 41 is managed by a standalone computer 410, which may be e.g. a desktop or a laptop, so as to control the temperature in each room in function of user heating habits.
  • the computer 410 embodies the system 1 with ML capabilities.
  • Available input data together with respective associated times include the outdoor temperature (thermometer 411), a user presence (infrared detectors 412 arranged in the rooms) and whether the user is working or not on the concerned day (user-entered information).
  • Output data include the temperatures in the respective rooms, which may take a limited number of values (for example low, medium and high).
  • Additional inputs directed to learning are given by the user’s reactions to the selected room temperatures, whether by keeping or by changing them, which provides a measure of the user’s satisfaction with the effected choices.
  • Those inputs include the recording of whether the user reacted (binary data for each room) and the amended temperature levels (monitored by switches 413). It deserves noting that both the above inputs and outputs are parts of the data samples entered into the system 1. Also, those data samples are updated regularly over time, e.g. every 10 minutes.
  • An efficiency feedback to minimize can consist in a number of user actions in modifying the automatically selected temperature rooms.
  • Input data associated with respective times include occupancy patterns in respective rooms 421 (monitored by sensors 422), reference working hours (pre-entered), and effective energy expenses in the rooms 421 (as assessed from local switches 424, e.g. light, air conditioning or heating, computers on or off).
  • Output data (with respective times) include overall power consumption (provided by overall meters 423) and energy invoices amounts. Minimizing the energy losses and expenses may be based on automatic default adjustment of air conditioning, heating and lighting, under constraints of efficient working conditions, e.g. appropriate ranges of room temperatures and lighting levels. This provides additional inputs and outputs, all of them being parts of the data samples feeding the system 1. For example, the situation is upgraded, and the input and output data entered accordingly, every half hour.
  • a third example ( Figure 2C) relates to an interconnected energy sharing network 43, in which user dwellings 431, 432 are equipped with solar panels in addition to electricity supply, and able to receive as well as to provide energy depending on their needs and accumulated power loads. Power is thus regularly and flexibly exchanged over the energy network 43 between users.
  • Some of the dwellings 431 own computing systems 430 embodying the system 1 for optimizing their energy balance, i.e. minimizing their energy expenses while maximizing their energy sales, while others 432 are not.
  • Appropriate examples of inputs and outputs will not be developed here, but at least some of them should be apparent to a skilled person.
  • the computing systems 430 of the dwellings 431 are not functioning as standalone agents, but are communicating together (via any proper wired or wireless telecommunication means) for sake of global power management optimization, in the frame of a distributed implementation of the system 1. This may contribute to significantly enhancing the performance of the network 43 as a whole compared with separate actions.
  • system 1 is further adapted to identify anomalies in the network 43 and to automatically trigger energy exchange remedies and/or to alert involved users or operators.
  • a fourth example regards a power supply network 44 involving a power supply facility 441 and energy customers 442.
  • the power supply facility 441 is provided with a mainframe computer 440 embodying the system 1, for optimizing the load supply distribution over the network 44 so as to reconcile over time the energy resources available to the facility 441 (storage, production, reception) and the customer demands.
  • Appropriate examples of inputs and outputs will not be developed here, but at least some of them should be apparent to a skilled person.
  • the system is adapted to a dynamic process, so that one or more batches 21 of data samples ranging over a time interval 210 are completing previously obtained and kept data samples 20.
  • each batch 21 of data may correspond to newly collected input and output data at a given time slot or at successive time slots (e.g. at 5 or 10 successive time slots).
  • the time interval 210 is considered to start at the end of the previous data recording and to finish with the lastly collected data sample.
  • a sliding time window 230 having a window width 30 is taken into account in the system 1.
  • the window width 30 may be set by a user so as to cover a roughly recurring cycle of events and situations. For example, the width may correspond to one day, one week, one month or one year.
  • the window width 30 may be set to cover multiple batches of data samples, so that the width 30 is larger, and possibly substantially larger, than the time interval 210.
  • the window width 30 may be automatically determined by the system 1, or may have been previously determined before being exploited by the system 1. Notably, it may be derived from periodicity detection in time series. This may be achieved by Fourier transform analysis through a periodogram (giving an estimate of the spectral density of a signal) in the time domain and/or by autocorrelation of the time signal, as known to a skilled person.
  • periodicity detection in time series This may be achieved by Fourier transform analysis through a periodogram (giving an estimate of the spectral density of a signal) in the time domain and/or by autocorrelation of the time signal, as known to a skilled person.
  • An example of an enhanced hybrid solution of the latter methods is presented by T. Puech et al. in “A fully automated periodicity detection in time series”, International Workshop on Advanced Analysis and Learning on Temporal Data, pages 43-54, AALTD 2019.
  • Another solution for potentially determining the window width in a dynamic way will be described below.
  • time window 230 Since the processing by the system 1 is relying on the data samples of the sliding time window 230, called the current samples 23, having the time window 230 correspond to a time cycle makes possible an integration of time variations over that cycle in power consumption management.
  • sliding the time window 230 so as to include the time interval 210 corresponding to the latest batches 21 of data may enable to take finely account of time variations extending beyond the time cycle.
  • the time window 230 may thus have an end time position ti, t2 adapted to catch latest batches 21 of data samples.
  • the system 1 is thus relying on the data samples of the new batches 21 and on previously kept data samples 20, distributed over the part of the time window 230 preceding the time interval 210.
  • the previously kept samples 20 extend however below the time window 230, and further covers past samples 25 older than the start of the time window 230 and distributed over a time interval 250.
  • a processing time range 220 comprising the time window 230 and the contiguously preceding time interval 250 is taken into account by the system 1.
  • the system 1 is interacting with a user interface 18, via which information can be entered and retrieved by a user.
  • the user interface 18 includes any means appropriate for entering or retrieving data, information or instructions, notably visual, tactile and/or audio capacities that can encompass any or several of the following means as well known by a person skilled in the art: a screen, a keyboard, a trackball, a touchpad, a touchscreen, a loudspeaker, a voice recognition system.
  • modules are to be understood as functional entities rather than material, physically distinct, components. They can consequently be embodied either as grouped together in a same tangible and concrete component, or distributed into several such components. Also, each of those modules is possibly itself shared between at least two physical components. In addition, the modules are implemented in hardware, software, firmware, or any mixed form thereof as well. They are preferably embodied within at least one processor of the device 10 or of the device 15.
  • the device 15 for obtaining decisional guidance it is provided for receiving selected data samples from the device 10 (including inputs and outputs) and for inducing from those samples the decisional guidance 35, on the ground of a module 16 for ML processing.
  • the latter may be built from usual techniques directed to achieving the objectives detailed in the present disclosure, including notably Decision Trees, ANN, ensemble models or gradient boosting, as well known to a person skilled in the art. It will therefore not be developed here.
  • the module 16 is configured for producing an ML prediction model 36 from the received samples, suited to produce the decisional guidance 35 based on those samples and possibly on further received samples.
  • expert systems are used in the device 15 instead of ML processing. This may be relevant in some specific cases as simpler and faster, but usually at the cost of sensitive preliminary adjustment of the related parameters, and of less flexibility and lower robustness to variations in the concerned situations.
  • the device 10 for selecting data samples comprises an input module 11 adapted to receive the new batches 21 of data samples (e.g. via a received datastream) and the previously kept data samples 20, for example from one or more local or remote database(s) 17.
  • the latter can take the form of storage resources available from any kind of appropriate storage means, which can be notably a RAM or an EEPROM (Electrically-Erasable Programmable Read-Only Memory) such as a Flash memory, possibly within an SSD (Solid-State Disk).
  • the previously kept data samples 20 may be streamed to the device 10.
  • the input module 11 is also adapted to receive the time window width 30 (e.g. from the database 17), when already available rather than being computed by the device 10.
  • Other parameters can be received by the input module 11 (e.g. from the database 17), including similarity criteria 31, elimination conditions 32 and optionally a similarity ratio 33, as developed below (in following examples, the similarity ratio 33 is instead computed by the device 10).
  • the device 10 optionally comprises a preselection module 12, adapted to preselect part of the newly received data samples, e.g. in the new batches 21 when a preselection has already been effected on the previously kept data samples 20, or more broadly on the current samples 23 of the time window 230.
  • the preselection module 12 may be particularly useful when receiving a datastream, since it enables to reduce in real time the size of the sample load to a desired dimension.
  • a sampling round executed on the flight while gathering data ensures that the selected data are representative of what happens within the time window 230.
  • the sampling may use a reservoir sampling strategy, so that all received samples have a same probability to be in the reservoir, the updated probability of the reservoir making it possible to do it on the flight.
  • a set of available data samples 22 is ready to be processed, which consists in the union of the past samples 25 and of the preselected samples of the time window 230 (i.e. the current samples 23).
  • the preselection module 12 is absent from the device 10 (whether the preselection has been executed upstream or all received samples are kept), so that the set of available data samples 22 to be processed is directly received as such by the device 10 instead of being mixed up with other data samples.
  • the device 10 also comprises a data sample selection module 13, configured for effecting an extraction of data samples among the received past samples 25, those extracted samples being called selected past samples 24. Such a selection introduces a forgetting mechanism in the oldest data, so as to avoid excessive computational costs while potentially getting a good sensitivity- stability tradeoff in a flexible way.
  • the past samples 25 consist of a first share 26 and a complementary share 27, respectively distributed in the sub-intervals 260 and 270.
  • a first aspect of the selection carried out by the module 13 is that all data samples of the first share 26 are kept, i.e. that share is immune from deletion and falls within the selected past samples 24.
  • a second aspect of the selection is that part of the data samples belonging to the complementary share 27 are deleted, which leads to a reduced set forming a second share 29 of the selected past samples 24.
  • the module 13 determines the deleted past samples 28 by the elimination conditions 32, which depend on both the similarity criteria 31 applied to the past samples 25 and on the current samples 23 associated with the time window 230.
  • a decision tree 50 is built on the ground of the past samples 25, and comprises (in compliance with usual practice) a root 51, internal nodes 52, and leaves 53.
  • the decision tree 50 may be obtained based on the automatic identification of feature similarities in the past samples 25, as known to a skilled person.
  • the decision tree 50 may be balanced in an enhanced way so as to improve accuracy, by using a gain function as described by A. Leroux et al. in “Information gain ratio correction: Improving prediction with more balanced decision tree splits”, arXiv:18Ol.O831O, 2018. At least part of the current samples 23 is then mapped to the decision tree 50, where each of the leaves 53 may encompass one or more of the past samples 25. In each leaf 530 ( Figure 7), the arrival of one of the current samples 231 may induce the deletion of one of the past samples 281 belonging to the complementary share 27. For example, the assignment of any of the current samples 23 to one of the leaves 53 causes the incrementation of a leaf counter, until a deletion threshold is reached. The oldest sample 281 of the complementary share 27 in the leaf 530 is then removed.
  • each of the samples belonging to the complementary share 27 in the leaf 530 is pre-weighted further to the decision tree construction, as an increasing function of the number of the past samples 25 in the leaf 530.
  • the related weights are then increased whenever one of the current samples 321 arrives. Once a preset deletion threshold is reached for any of the weighted samples 281, the latter is removed.
  • the current samples 23 are mapped to the decision tree 50 in a chronological order of those current samples 23 for the deletion process.
  • the current samples 23 are mapped to the decision tree 50 while the deletion process is executed in a chronological order of the past samples 25 of the leaves 53.
  • the total number of deleted samples may be compliant with a preset number of kept samples in the complementary share 27, i.e. the target size of the second share 29.
  • the total number of the selected past samples 24, i.e. in the first share 26 and in the second share 29, may be determined from an overall desired number of data samples for downstream processing by the device 15 (this overall number regarding the union of the current samples 23 and of the selected past samples 24) and by a desired balance between the first share 25 and the second share 29.
  • the latter balance may be given by the similarity ratio 33 (as a hyper-parameter), noted q, giving the number of samples in the second share 29 divided by the total number of the selected past samples 24.
  • a decision tree is to be built by the device 15 based on inputs from the device 10, and a maximum number of related operations is set. If e.g.
  • the maximum number is 50,000 and the current samples 23 associated with the time window 230 are 25,000, 25,000 samples can be kept in the time interval 250 associated with the past samples 25.
  • (1 - q) x 25,000 samples can be taken for the first share 26 (which is close to a time forgetting approach), and q x 25,000 samples for the second share 29 (which is close to a similarity forgetting approach).
  • (1 - ⁇ ?) x 25,000 samples of the first share 26 provides the time length of the preserved sub-interval 260, while the q x 25,000 samples of the second share 29 may be possibly scattered within the sub-interval 290.
  • the device 15 is thus fed with the appropriate number 50,000 of samples for building its decision tree.
  • the current samples 23 or part of them are used together with the past samples 25 to build the decision tree 50, which is then exploited as described above.
  • multiples decision trees are built and exploited as above, e.g. with random forest or XGBoost models.
  • the module 13 is adapted to assign the samples to the leaves of each of the trees, and to compute a similarity score related to the number of samples in each of the decision trees. That similarity score is exploited in a similar way to the above for past sample deletions.
  • the first share 26 and the second share 29 of the past samples 25 are used in an ML model providing predictions for power consumption management, and which may be carried out by the device 15.
  • the past samples 25 are then respectively associated with resulting prediction errors, and lowest values of the latter are privileged for sample deletion in relation with the arrival of the current samples 23. Indeed, the prediction error provides a fair estimation of a similarity of a particular point.
  • the current samples 23 or part of them are used together with the past samples 25 to build the ML model providing predictions for power consumption management, which is then exploited as described above.
  • the similarity ratio 33 is provided to the device 10, and properly set in function of the situations. For example, a high value is chosen for detecting anomalies on an electric network (the high level of similarity remembering enabling to keep odd behavior in memory), while a low value is chosen for reflecting a change of consumer behavior (the low level of similarity remembering enabling to forget old behavior as fast as possible).
  • the similarity ratio 33 is automatically tuned by the module 13 of the device 10, taking account of the ML processing of the device 15. This may enable to enhance the performances in a flexible and dynamic way.
  • a rolling validation is for example carried out on the ground of time-series data, as explained below in relation with Figures 10 and 11.
  • the ML prediction model 36 is built by the ML processing module 16 from the available data samples 22, through obtaining the current samples 23 and the selected past samples 24 with the module 13 for data sample selection.
  • the available data samples 22, the current samples 23 and selected past samples 24, and the ML prediction model 36 usually differ at each time position step of the time window 230 (while being presently noted in the same way for sake of convenience).
  • the selected past samples 24 at a given step are then submitted to the ML prediction model 36 of a next step, presently n+1, from which an overall evaluation of relevance 37 can be induced.
  • the evaluation 37 thus reflects the relevance of the samples selected at a given step to the data of the next step.
  • the evaluations 37 are obtained for successive time positions of the time window 230 and averaged, leading to an error 38 ( Figure 11).
  • the same process is repeated for various candidate values qi, qi... q m of the similarity ratio 33, m being for example equal to 10.
  • the value selected for the similarity ratio 33 is then the one entailing the smallest error 38.
  • This automatic selection of the similarity ratio 33 ensures that the algorithm is not based on already seen data, while being able to detect concept drift.
  • testing around 10 values of the similarity ratio 33 may provide a satisfying selection of a proper value q while avoiding additional computation costs.
  • time window width 30 (which is another hyper-parameter), instead of the previously described related implementations.
  • the device 10 further includes an output module 14, adapted to receive the current samples 23 and the selected past samples 24 from the data sample selection module 13, and to provide them to the device 15.
  • system 1 may for example execute the following process, in relation with Figures 12A and 12B:
  • step 62 preselect the available data samples 22 from the new batches 21 and from the previously kept samples 20
  • step 62 proceed with data sample selection among the available data samples 22, by using the time width 30, the similarity criteria 31 and the elimination conditions 32, so as to produce the current samples 23 and the selected past samples 24 (step 63)
  • step 66 apply an ML processing to the selected data samples and induce therefrom the decisional guidance 35 (step 66).
  • the data sample selection step 63 may hinge on the following operations:
  • system 1 may for example execute the following process for automatically and dynamically determining the similarity ratio 33, as shown on Figure 13:
  • step 672 switch to a next available time window position associated with a current value 2305 of the time window position and with current available data samples 225 (step 672), and proceed with the data selection step 63 and the ML processing step 66 as above based on the updated data, - if a previous time window position is available, apply the ML prediction to the past samples 245 selected at the previous step so as to obtain the related evaluation of relevance 37 (step 673),
  • step 672 if a next time window position is available, switch to that next position having a current value 2305 and associated with current available data samples 225 (step 672), and proceed with the data selection step 63, the ML processing step 66 and the ML prediction step 673 as above based on the updated data,
  • a particular apparatus 7, visible on Figure 14, is embodying the system 1 as described above. It corresponds for example to a mainframe computer, a workstation, a laptop, a tablet, a smartphone, or a head-mounted display (HMD).
  • HMD head-mounted display
  • That apparatus 7 is suited to selecting data samples and providing decisional guidance. It comprises the following elements, connected to each other by a bus 75 of addresses and data that also transports a clock signal:
  • microprocessor 71 or CPU
  • a graphics card 72 comprising several Graphical Processing Units (or GPUs) 720 and a Graphical Random Access Memory (GRAM) 721; the GPUs are quite suited to repeated computations on the data samples, due to their highly parallel structure;
  • GPUs Graphical Processing Units
  • GRAM Graphical Random Access Memory
  • I/O devices 74 such as for example a keyboard, a mouse, a trackball, a webcam; other modes for introduction of commands such as for example vocal recognition are also possible;
  • radiofrequency unit 79 a radiofrequency unit 79.
  • the power supply 78 is external to the apparatus 7.
  • the apparatus 7 also comprises a display device 73 of display screen type directly connected to the graphics card 72 to display synthesized images calculated and composed in the graphics card.
  • the use of a dedicated bus 730 to connect the display device 73 to the graphics card 72 offers the advantage of having much greater data transmission bitrates and thus reducing the latency time for the displaying of images composed by the graphics card, e.g. for ML representations.
  • a display device is external to apparatus 7 and is connected thereto by a cable or wirelessly for transmitting the display signals.
  • the apparatus 7, for example through the graphics card 72 comprises an interface for transmission or connection adapted to transmit a display signal to an external display means such as for example an LCD or plasma screen or a video-projector.
  • the RF unit 79 can be used for wireless transmissions.
  • register used hereinafter in the description of memories 77 and 721 can designate in each of the memories mentioned, a memory zone of low capacity (some binary data) as well as a memory zone of large capacity (enabling a whole program to be stored or all or part of the data representative of data calculated or to be displayed). Also, the registers represented for the RAM 77 and the GRAM 721 can be arranged and constituted in any manner, and each of them does not necessarily correspond to adjacent memory locations and can be distributed otherwise (which covers notably the situation in which one register includes several smaller registers).
  • the microprocessor 71 When switched-on, the microprocessor 71 loads and executes the instructions of the program contained in the RAM 77.
  • the random access memory 77 comprises notably:
  • Algorithms implementing the steps of the method specific to the present disclosure and described above are stored in the memory GRAM 721.
  • the graphic processors 720 of graphics card 72 load appropriate information and parameters into the GRAM 721 and execute the instructions of algorithms in the form of microprograms.
  • the random access memory GRAM 721 comprises notably:
  • the presence of the graphics card 72 is not mandatory, and can be replaced with entire CPU processing and/or simpler visualization implementations .
  • the apparatus 7 may include only the functionalities of the device 10 for selecting data samples, or only the functionalities of the device 15 for providing decisional guidance.
  • the device 10 and/or the device 15 may be implemented differently than a standalone software, and an apparatus or set of apparatus comprising only parts of the apparatus 7 may be exploited through an API call or via a cloud interface.

Landscapes

  • Business, Economics & Management (AREA)
  • Human Resources & Organizations (AREA)
  • Engineering & Computer Science (AREA)
  • Strategic Management (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Economics (AREA)
  • Development Economics (AREA)
  • Educational Administration (AREA)
  • Operations Research (AREA)
  • Marketing (AREA)
  • Game Theory and Decision Science (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Supply And Distribution Of Alternating Current (AREA)

Abstract

Selon l'invention, des échantillons de données sont sélectionnés en vue d'une gestion de la consommation d'énergie. Ceci comporte la réception d'échantillons associés à des instants respectifs, répartis dans une fenêtre temporelle glissante en tant qu'échantillons courants (23) et dans une période passée en tant qu'échantillons passés (25). Des échantillons passés sélectionnés sont déterminés en conservant une première part (26) des échantillons passés, constituée des plus récents, et une seconde part (29) en éliminant parmi les échantillons passés privés de la première part, dite part complémentaire (27), une partie des échantillons passés en fonction d'au moins certains des échantillons courants et de conditions (32) d'élimination dépendant de critères (31) de similarité appliqués au moins à la première part et à la part complémentaire. Les échantillons passés sélectionnés sont fournis avec les échantillons courants pour la gestion de la consommation d'énergie. L'invention possède des applications à la détection de pannes d'alimentation et à l'adaptation dynamique de la consommation d'énergie.
PCT/EP2021/078681 2020-10-16 2021-10-15 Dispositifs et processus de sélection d'échantillons de données pour gestion de consommation d'énergie WO2022079280A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US17/072,330 2020-10-16
EP20306227.8A EP3985582A1 (fr) 2020-10-16 2020-10-16 Dispositifs et procédés de sélection d'échantillon de données pour la gestion de la consommation d'énergie
EP20306227.8 2020-10-16
US17/072,330 US11539598B2 (en) 2020-10-16 2020-10-16 Devices and processes for data sample selection for power consumption management

Publications (1)

Publication Number Publication Date
WO2022079280A1 true WO2022079280A1 (fr) 2022-04-21

Family

ID=78212132

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2021/078681 WO2022079280A1 (fr) 2020-10-16 2021-10-15 Dispositifs et processus de sélection d'échantillons de données pour gestion de consommation d'énergie

Country Status (1)

Country Link
WO (1) WO2022079280A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023242926A1 (fr) * 2022-06-13 2023-12-21 日本電信電話株式会社 Dispositif de gestion d'apprentissage, procédé de gestion d'apprentissage, et programme

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2543281A (en) * 2015-10-13 2017-04-19 British Gas Trading Ltd System for energy consumption prediction
US20180299944A1 (en) * 2017-04-14 2018-10-18 National Tsing Hua University Production Management Method and System Using Power Consumption Features
US20190122132A1 (en) 2016-04-19 2019-04-25 Grid4C Method and system for energy consumption prediction
WO2019243524A1 (fr) 2018-06-22 2019-12-26 Moixa Energy Holdings Limited Systèmes d'apprentissage automatique, d'optimisation et de gestion de flexibilité locale multi-actif de ressources distribuées de stockage d'énergie

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2543281A (en) * 2015-10-13 2017-04-19 British Gas Trading Ltd System for energy consumption prediction
US20190122132A1 (en) 2016-04-19 2019-04-25 Grid4C Method and system for energy consumption prediction
US20180299944A1 (en) * 2017-04-14 2018-10-18 National Tsing Hua University Production Management Method and System Using Power Consumption Features
WO2019243524A1 (fr) 2018-06-22 2019-12-26 Moixa Energy Holdings Limited Systèmes d'apprentissage automatique, d'optimisation et de gestion de flexibilité locale multi-actif de ressources distribuées de stockage d'énergie

Non-Patent Citations (9)

* Cited by examiner, † Cited by third party
Title
A. D'AMATOM. BOUSSARD: "Forgetting methods for white box learning", INTERNATIONAL CONFERENCE ON PRACTICAL APPLICATIONS FOR AGENTS AND MULTI-AGENT SYSTEMS, IN TRENDS IN PRACTICAL APPLICATIONS OF SCALABLE MULTI-AGENT SYSTEMS, 2016
A. LEROUX ET AL.: "Information gain ratio correction: Improving prediction with more balanced decision tree splits", ARXIV:1801.08310, 2018
A. ROBINS: "Catastrophic Forgetting, rehearsal and pseudorehearsal", CONNECTION SCIENCE, vol. 7, 1995, pages 123 - 146, XP055728980, DOI: 10.1080/09540099550039318
KIRKPATRICK ET AL.: "Overcoming catastrophic forgetting in neural networks", PNAS, vol. 114, no. 13, 2017, pages 3521 - 3526, XP055707922, DOI: 10.1073/pnas.1611835114
R. M. FRENCH: "Pseudo-recurrent connectionist networks: an approach to the 'sensitivity-stability' dilemma", CONNECTION SCIENCE, vol. 9, no. 4, 1997, pages 353 - 379
S. M. LUNDBERGS.-I. LEE: "A unified approach to interpreting model predictions", NIPS CONFERENCE, 2017, ARXIV: 1705.07874
T. CHENC. GUESTRIN: "XGBoost: a scalable tree boosting system", PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, pages 785 - 794
T. PUECH ET AL.: "A fully automated periodicity detection in time series", INTERNATIONAL WORKSHOP ON ADVANCED ANALYSIS AND LEARNING ON TEMPORAL DATA, pages 43 - 54
X. H. DANG ET AL.: "Incremental and adaptive clustering stream data over sliding window", INTERNATIONAL CONFERENCE ON DATABASE AND EXPERT SYSTEMS APPLICATIONS, DEXA '09, 2009, pages 660 - 674, XP047390053, DOI: 10.1007/978-3-642-03573-9_55

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023242926A1 (fr) * 2022-06-13 2023-12-21 日本電信電話株式会社 Dispositif de gestion d'apprentissage, procédé de gestion d'apprentissage, et programme

Similar Documents

Publication Publication Date Title
Zhang et al. A review of machine learning in building load prediction
Gopinath et al. Energy management using non-intrusive load monitoring techniques–State-of-the-art and future research directions
Sekhar et al. Robust framework based on hybrid deep learning approach for short term load forecasting of building electricity demand
Laouafi et al. Online electricity demand forecasting based on an effective forecast combination methodology
Xenochristou et al. An ensemble stacked model with bias correction for improved water demand forecasting
Ntakolia et al. Machine learning applied on the district heating and cooling sector: A review
Dey et al. Forecasting energy consumption from smart home sensor network by deep learning
Chadoulos et al. Mobile apps meet the smart energy grid: A survey on consumer engagement and machine learning applications
Aouad et al. A CNN-Sequence-to-Sequence network with attention for residential short-term load forecasting
Lazzari et al. User behaviour models to forecast electricity consumption of residential customers based on smart metering data
Qiao et al. Feature selection strategy for machine learning methods in building energy consumption prediction
Natgunanathan et al. Deakin microgrid digital twin and analysis of AI models for power generation prediction
US11539598B2 (en) Devices and processes for data sample selection for power consumption management
EP3985582A1 (fr) Dispositifs et procédés de sélection d'échantillon de données pour la gestion de la consommation d'énergie
WO2022079280A1 (fr) Dispositifs et processus de sélection d'échantillons de données pour gestion de consommation d'énergie
Tsalikidis et al. Energy load forecasting: One-step ahead hybrid model utilizing ensembling
Batbaatar et al. DeepEnergy: Prediction of appliances energy with long-short term memory recurrent neural network
Cheng et al. PowerNet: a smart energy forecasting architecture based on neural networks
Irankhah et al. A parallel CNN-BiGRU network for short-term load forecasting in demand-side management
Meyyappan et al. Wavelet neural network-based wind-power forecasting in economic dispatch: A differential evolution, bacterial foraging technology, and primal dual-interior point approach
Candelieri et al. Layered machine learning for short-term water demand forecasting.
Vargas-Solar et al. Greenhome: a household energy consumption and CO2 footprint metering environment
Motlagh et al. A neural approach to electricity demand forecasting
Kemper et al. Forecasting of residential unit’s heat demands: a comparison of machine learning techniques in a real-world case study
Moon et al. User Behavior Analytics with Machine Learning for Household Electricity Demand Forecasting

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21793928

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21793928

Country of ref document: EP

Kind code of ref document: A1