This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Machine Learning Infused Distributed Optimization for Coordinating Virtual Power Plant Assets

Meiyi Li, , Javad Mohammadi, 
Abstract

Amid the increasing interest in the deployment of Distributed Energy Resources (DERs), the Virtual Power Plant (VPP) has emerged as a pivotal tool for aggregating diverse DERs and facilitating their participation in wholesale energy markets. These VPP deployments have been fueled by the Federal Energy Regulatory Commission’s Order 2222, which makes DERs and VPPs competitive across market segments. However, the diversity and decentralized nature of DERs present significant challenges to the scalable coordination of VPP assets. To address efficiency and speed bottlenecks, this paper presents a novel machine learning-assisted distributed optimization to coordinate VPP assets. Our method, named as 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} (Learning to Optimize the Optimization Process for Multi-agent Coordination), adopts a multi-agent coordination perspective where each VPP agent manages multiple DERs and utilizes neural network approximators to expedite the solution search. The 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} method employs a gauge map to guarantee strict compliance with local constraints, effectively reducing the need for additional post-processing steps. Our results highlight the advantages of 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}, showcasing accelerated solution times per iteration and significantly reduced convergence times. The 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} method outperforms conventional centralized and distributed optimization methods in optimization tasks that require repetitive and sequential execution.

Index Terms:
Virtual Power Plants (VPPs), Alternating Direction Method of Multipliers (ADMM), Distributed Optimization, Distributed Energy Resources (DERs), Learning to Optimize the Optimization Process (LOOP), Collaborative Problem-solving

I Introduction

I-A Motivation

As global energy sectors transition towards sustainability, the role of Distributed Energy Resources (DERs) has become increasingly significant. However, the participation of DERs in competitive electricity markets remains a challenge [1]. While many DERs are capable of providing wholesale market services, they often individually fall short of the minimum size thresholds established by Independent System Operators (ISOs) and may not meet performance requirements [2]. As a solution to these challenges, Virtual Power Plants (VPPs) have emerged to aggregate diverse DERs, creating a unified operating profile for participation in wholesale markets and providing services to system operators [3]. Further promoting the aggregation of DERs, the Federal Energy Regulatory Commission’s (FERC’s) Order 2222, issued in September 2020, allowed DERs to compete on equal terms with other resources in ISO energy, capacity, and ancillary service markets [2]. The FERC regulatory advancement strengthens the position of DERs and VPPs in the market.

Despite their promising potential, the massive, decentralized, diverse, heterogeneous, and small-scale nature of DERs poses significant challenges to traditional centralized approaches, especially in terms of computational efficiency and speed. Centralized controls for VPPs require global information from all DERs, making them susceptible to catastrophic failures if centralized nodes fail and potentially compromising the privacy of DER owners’ information. To address these issues, there is a growing demand for efficient, scalable, distributed and decentralized optimization techniques. Our study aims to tackle these challenges and develop a solution that can efficiently harness the benefits of DERs, thereby unlocking the full potential of VPPs.

I-B Related Work

I-B1 VPP Functionalities and Objectives

VPPs act as aggregators for a variety of DERs, playing a pivotal role in mitigating integration barriers between DERs and grid operations [4]. In what folows, we will highlight recent insights gained from extensive research conducted on strategies for coordinating DERs within VPPs. For instance, optimization schemes for coordinating DERs within VPPs can be customized to achieve various objectives including:

  • VPP’s self financial and operational objectives:

    • Maximizing revenue from energy trading across different markets [5].

    • Decreasing operational and maintenance costs of operating VPPs [5, 6, 7, 8, 9].

    • Optimizing load curtailment [10] or energy exportation [11].

    • Reducing end-user discomfort from joining demand response efforts [12].

    • Narrowing the discrepancy between actual power consumption and predetermined set points and schedules [13, 14].

    • Mitigating financial burden of operational risks [15, 16, 17, 18].

  • Contributing to system-level initiatives:

    • Curtailing greenhouse gas emissions [19].

    • Advancing the reliability and resilience of the overall energy system [17, 18, 20].

I-B2 Shortcomings of Centralized Coordination Methods

Today’s centralized optimization methods are not designed to cope with decentralized, diverse, heterogeneous, and small-scale nature of DERs. Recent studies have shown that integrating DERs at scale may adversely impact today’s tools operation’s efficiency and performance speed [21].

Major challenges of centralized management strategies include:

  • Scalability issues become more pronounced with the addition of more DERs to the network, resulting in increased computational demands due to the management of a growing set of variables

  • Security and privacy risks as centralized decision-making models requires comprehensive data from all DERs [22].

  • Severe system disruptions resulting from dependence on a single centralized node, as a failure in that node may pose a significant operational risk.

  • Significant delays in the decision-making process due to the strain on the communication infrastructure, a situation worsened by continuous data communication and the intermittent nature of DERs.

  • Adaptability challenges as the centralized systems struggle to provide timely responses to network changes. This limitation stems from their requirement for a comprehensive understanding of the entire system to make informed decisions [23].

  • Logistical and political challenges given the diverse and intricate nature of DERs within a comprehensive centralized optimization strategy that spans across various regions and utilities [24].

In response to these challenges, there is a growing demand and interest in the development and implementation of efficient, scalable, and decentralized optimization approaches.

I-B3 State-of-the-art in Distributed Coordination

Distributed coordination methods organize DERs into clusters, with each one treated as an independent agent with capabilities for communication, computation, data storage, and operation, as demonstrated in previous work [25]. A distributed configuration enables DERs to function efficiently without dependence on a central controller. Distributed coordination paradigms, which leverage the autonomy of individual agents, have played a crucial role in the decentralized dispatch of DERs, as highlighted in recent surveys [22].

Among the numerous distributed optimization methods proposed in power systems, the Alternating Direction Method of Multipliers (ADMM) has gained popularity for its versatility across different optimization scenarios. Recent examples include a distributed model to minimize the dispatch cost of DERs in VPPs, while accounting for network constraints [18]. Another noteworthy contribution is a fully distributed methodology that, combines ADMM and consensus optimization protocols to address transmission line limits in VPPs [21]. Li et al. [26] introduced a decentralized algorithm to enable demand response optimization for electric vehicles within a VPP. Contributing to the robustness of VPPs, another decentralized algorithm based on message queuing has been proposed to enhance system resilience, particularly in cases of coordinator disruptions [27].

I-B4 Challenges of Existing Distributed Coordination Methods

Despite their many advantages, most distributed optimization techniques, even those with convergence guarantees, require significant parameter tuning to ensure numerical stability and practical convergence. Real-time energy markets impose operational constraints that require frequent updates, sometimes as frequently as every five minutes throughout the day, as indicated by [28]. The frequent update demands that the optimization of DERs dispatch within VPPs is resolved frequently and in a timely manner. Nevertheless, the iterative nature of these optimization techniques can significantly increase computation time, restricting their utility in time-sensitive scenarios. Moreover, the optimization performance may not necessarily improve, even when encountering identical or analogous dispatching problems frequently, leading to computational inefficiency.

To address these limitations, Machine Learning (ML) has been deployed to enhance the efficiency of optimization procedures, as discussed in [29]. The utilization of neural networks can expedite the search process and reduce the number of iterations needed to identify optimal solutions. Furthermore, neural approximators can continually enhance their performance as they encounter increasingly complex optimization challenges, as demonstrated in [30].

ML-assisted distributed optimizers can be broadly categorized into three distinct models: supervised learning, unsupervised learning, and reinforcement learning. In the realm of supervised learning, a data-driven method to expedite the convergence of ADMM in solving distributed DC optimal power flow (DC-OPF) is presented in [31], where authors employ penalty-based techniques to achieve local feasibility. Also, we have proposed an ML-based ADMM method to solve the DC-OPF problem which provides a rapid solution for primal and dual sub-problems in each iteration [32]. Additional applications of supervised learning are demonstrated in [33] and [34], where ML algorithms are used to provide warm-start points for ADMM. On the other hand, unsupervised learning is exemplified in [35], where a learning-assisted asynchronous ADMM method is proposed, leveraging k-means for anomaly detection. Reinforcement learning has been applied to train neural network controllers for achieving DER voltage control [36], frequency control [37], and optimal transactions [38].

Although these studies showcase the potential of ML for adaptive, real-time DER optimization in decentralized VPP models, they do not fully develop ML-infused distributed optimization methods to improve computation speed while ensuring solution feasibility.

I-C Contributions

In this paper, we propose an ML-assisted method to replace the building blocks of the ADMM-based distributed optimization technique with neural approximators. Our method is referred to as 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} (Learning to Optimize the Optimization Process for Multi-agent Coordination). We will employ our 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method to find a multi-agent solution for the power dispatch problem in DER coordination within a VPP. In the muti-agent VPP configuration, each agent may control multiple DERs. The proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method enables each agent to predict local power profiles by communicating with its neighbors. All agents collaborate to achieve a near-optimal solution for power dispatch while adhering to both local and system-level constraints.

The utilization of neural networks expedites the search process and reduces the number of iterations required to identify optimal solutions. Additionally, unlike restoration-based methods, the 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  approach doesn’t necessitate post-processing steps to enhance feasibility because local constraints are inherently enforced through a gauge mapping method [39], and coupled constraints are penalized through ADMM iterations. This paper advances our recent work in [32] that is focused on speeding up the ADMM-based DC-OPF calculations through efficient approximation of primal and dual sub-problems. While [32] tackled the DC-OPF problem, the present paper extends our previous model to incorporate individual VPP assets, addressing the DER coordination problem. In terms of methodology, [32] employs ML to facilitate both primal and dual updates of the ADMM method. This requires neighboring agents to share updated global variables, local copies of global variables, and Lagrangian multipliers. This work, however, replaces the two ADMM update procedures with a single data infusion step that reduces agents’ communication and computation burden.

II Problem formulation

II-A Compact Formulation

II-A1 The compact formulation for original optimization problem

The centralized optimization function is:

min𝐮f(𝐮,𝐱)s.t.𝐮𝒮(𝐱)\displaystyle\min_{\mathbf{u}}f(\mathbf{u},\mathbf{x})~{}~{}\text{s.t.}~{}~{}\mathbf{u}\in\mathcal{S}(\mathbf{x}) (1)

where 𝐮=i𝐮i\mathbf{u}=\bigoplus_{i}\mathbf{u}^{i} represents the collection of optimization variables across all agents. Note, \bigoplus denotes vector concatenation, and 𝐮i\mathbf{u}^{i} indicates the optimization variable vector of agent ii. Similarly, 𝐱=i𝐱i\mathbf{x}=\bigoplus_{i}\mathbf{x}^{i} encompasses all input parameters across agents, with 𝐱i\mathbf{x}^{i} indicating agent ii’s input parameter vector. The overall objective function is captured by f=ifi(𝐱i,𝐮i)f=\sum_{i}f^{i}(\mathbf{x}^{i},\mathbf{u}^{i}) where fif^{i} stands for agent ii’s objective. Lastly, 𝒮\mathcal{S} is the collection of all agent’s constraint sets.

II-A2 The compact formulation at the multi-agent-level

Here, we introduce the agent-based method to distribute computation responsibilities among agents. Let the variable vector of each agent, 𝐮i\mathbf{u}^{i}, consist of both local and global variables, which can be partitioned as 𝐮i=[𝐮li,𝐮gi]\mathbf{u}^{i}=[\mathbf{u}^{i}_{\texttt{l}},\mathbf{u}^{i}_{\texttt{g}}]. Here, 𝐮li\mathbf{u}^{i}_{\texttt{l}} captures the local variables of agent ii, while 𝐮gi\mathbf{u}^{i}_{\texttt{g}} encapsulates the global variables shared among neighboring agents. To enable distributed computations, each agent ii maintains a local copy vector of other agents’ variables, 𝐮g,Copyi\mathbf{u}^{i}_{\texttt{g,Copy}}, from which this vector mimics the global variables owned by neighboring agents.

Agent-level computations

Solving (1) in a distributed fashion requires agent ii to solve (2) before communication.

min𝐮i,𝐮g,Copyifi([𝐮li,𝐮gi],𝐮g,Copyi,𝐱i)\displaystyle\min_{\mathbf{u}^{i},\mathbf{u}^{i}_{\texttt{g,Copy}}}f^{i}([\mathbf{u}^{i}_{\texttt{l}},\mathbf{u}^{i}_{\texttt{g}}],\mathbf{u}^{i}_{\texttt{g,Copy}},\mathbf{x}^{i}) (2a)
s.t. local constraints: [𝐮i𝐮g,Copyi]𝒮Locali(𝐱i)\displaystyle\textup{local constraints: }\begin{bmatrix}\mathbf{u}^{i}\\ \mathbf{u}^{i}_{\texttt{g,Copy}}\end{bmatrix}\in\mathcal{S}_{\texttt{Local}}^{i}(\mathbf{x}^{i}) (2b)
consensus constraints: 𝐮g,Copyi=𝐈ci[ji𝐮gj]\displaystyle\textup{consensus constraints: }\mathbf{u}^{i}_{\texttt{g,Copy}}=\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}^{j}_{\texttt{g}}] (2c)

where 𝒮Locali\mathcal{S}_{\texttt{Local}}^{i} denotes the agent ii’s local constraint set. Here, 𝐈ci[ji𝐮gj]\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}^{j}_{\texttt{g}}] denotes the global variables owned by neighboring agents, and 𝐈ci\mathbf{I}_{\texttt{c}}^{i} is an element selector matrix. The distributed optimization process and intra-agent information exchange will ensure agreement among local copies of shared global variables.

Intra-agent Information Exchange
dual update:𝝀i[k]=𝝀i[k1]+ρ(𝐮g,Copyi[k1]𝐈ci[ji𝐮gj[k1]])\displaystyle\textup{dual update:}\boldsymbol{\lambda}^{i^{[k]}}\!\!\!\!\!=\boldsymbol{\lambda}^{i^{[k-1]}}\!\!\!\!+\rho(\mathbf{u}^{i^{[k-1]}}_{\texttt{g,Copy}}-\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j^{[k-1]}}]) (3)
primal optimization: [𝐮i[k]𝐮g,Copyi[k]]=hi(𝝀i[k])\displaystyle\textup{primal optimization: }\begin{bmatrix}\mathbf{u}^{i^{[k]}}\\ \mathbf{u}^{i^{[k]}}_{\texttt{g,Copy}}\end{bmatrix}=h^{i}(\boldsymbol{\lambda}^{i^{[k]}}) (4)

The dual update procedure (3) adjusts the Lagrangian multipliers 𝝀i\boldsymbol{\lambda}^{i}, which enforces consensus between agent ii and its neighbors. Here, 𝝀i\boldsymbol{\lambda}^{i} represents the differences between agent ii’s local copies and the global variables from neighboring agents, and ρ>0\rho>0 is a penalty parameter.

In (4), hih^{i} captures the compact form of an optimization problem that reduces the gap between local copies of global variables while respecting the constraints of individual agents.

II-B VPP Model

The considered VPP consists of a number of NAN_{\texttt{A}} agents, each denoted by index i,i𝒩Ai,i\in\mathcal{N}_{\texttt{A}}. Every agent is responsible for aggregating a diverse set of DERs, which encompasses flexible loads (FLs), energy storage systems (ESSs), heating, ventilation, and air conditioning (HVAC) systems, plug-in electric vehicles (PEVs), and photovoltaic (PV) arrays, as shown in Fig.1. These agents might be connected to networks of different utilities. The primary objective of the VPP is to optimize the aggregate behavior of all agents while accounting for agents’ utility functions.

Refer to caption
Figure 1: Examples of agents controlled by a VPP.

In this paper, we propose that the VPP operates within a two-settlement energy market, composed of a day-ahead and a real-time market. Upon the clearing of the day-ahead market, the VPP decides on hourly production schedules. The real-time market, also known as the imbalance market, is designed to settle potential day-ahead commitment violations. The real-time market productions are set in 5-minute increments. The production schedules every 5 minutes are denoted as 𝐏Sch\mathbf{P}_{\texttt{Sch}}.

The 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method is designed for the real-time market, where a VPP solves a dispatch optimization across its assets (agents) to honor its commitment over a given time scale, [ts,te][t_{\texttt{s}},t_{\texttt{e}}], where tst_{\texttt{s}} and tet_{\texttt{e}} represent the starting and ending times, respectively. Put differently, the VPP needs to fulfill the production schedule 𝐏Sch=[PSchtt=ts,,te]\mathbf{P}_{\texttt{Sch}}=\left[{P}_{\texttt{Sch}}^{t}\mid t=t_{\texttt{s}},\ldots,t_{\texttt{e}}\right] while minimizing the overall cost of agents. Generally, the VPP implements 5-minute binding intervals (Δt=5/60\Delta t=5/60 h) for the real-time market, and adopts look-ahead horizon (tets)(t_{\texttt{e}}-t_{\texttt{s}}), ranging from 5 minutes up to 2 hours [40], for the real-time dispatch optimization. The detailed dispatch optimization problem is presented next.

II-C Centralized Formulation of the VPP Coordination Problem

This subsection presents the centralized form of the power dispatch problem solved by a VPP over various assets for every time step t[ts,te]t\in[t_{\texttt{s}},t_{\texttt{e}}]. The asset constraints are:

II-C1 Constraints Pertaining to Flexible Loads

The power of a flexible load should be within a pre-defined operation range [PFLmini,t,PFLmaxi,t][P_{\texttt{FLmin}}^{i,t},P_{\texttt{FLmax}}^{i,t}], t[ts,te]\forall t\in[t_{\texttt{s}},t_{\texttt{e}}], i𝒩A\forall i\in\mathcal{N}_{\texttt{A}}:

PFLmini,tPFLi,tPFLmaxi,t\displaystyle P_{\texttt{FLmin}}^{i,t}\leq P_{\texttt{FL}}^{i,t}\leq P_{\texttt{FLmax}}^{i,t} (5)

II-C2 Constraints Pertaining to Energy Storage Systems

i\forall i and t[ts,te]\forall t\in[t_{\texttt{s}},t_{\texttt{e}}], the charging PESSCi,tP_{\texttt{ESSC}}^{i,t} (or discharging PESSDi,tP_{\texttt{ESSD}}^{i,t}) power of the energy storage system must not exceed PESSmaxiP_{\texttt{ESSmax}}^{i}, as indicated in (6). Also, (7) and (8) define RSoCi,tR_{\texttt{SoC}}^{i,t} as the state of charge (SoC) and bound its limits. Here ηESSCi\eta_{\texttt{ESSC}}^{i} and ηESSDi\eta_{\texttt{ESSD}}^{i} denote the charging and discharging efficiency. Finally EESSNiE_{\texttt{ESSN}}^{i} refers to the capacity.

0PESSCi,t\displaystyle 0\leq P_{\texttt{ESSC}}^{i,t} PESSmaxi,0PESSDi,tPESSmaxi\displaystyle\leq P_{\texttt{ESSmax}}^{i},~{}~{}~{}0\leq P_{\texttt{ESSD}}^{i,t}\leq P_{\texttt{ESSmax}}^{i} (6)
RSoCi,t+1=\displaystyle R_{\texttt{SoC}}^{i,t+1}= RSoCi,t+(PESSCi,tηESSCiPESSDi,tηESSDi)ΔtEESSNi\displaystyle R_{\texttt{SoC}}^{i,t}+\frac{(P_{\texttt{ESSC}}^{i,t}\eta_{\texttt{ESSC}}^{i}-\frac{P_{\texttt{ESSD}}^{i,t}}{\eta_{\texttt{ESSD}}^{i}})\Delta t}{E_{\texttt{ESSN}}^{i}} (7)
RSoCminiRSoCi,t+1RSoCmaxi\displaystyle R_{\texttt{SoCmin}}^{i}\!\!\leq R_{\texttt{SoC}}^{i,t+1}\leq\!\!R_{\texttt{SoCmax}}^{i} (8)

II-C3 Constraints Pertaining to Heating, Ventilation, and Air Conditioning Systems

The inverter-based heating, ventilation, and air conditioning model [41] is presented below with consumption power denoted as PHVACi,tP_{\texttt{HVAC}}^{i,t}.

THVACi,t+1=εHVACiTHVACi,t+(1εHVACi)(Touti,tηHVACiAHVACiPHVACi,t)\displaystyle T_{\texttt{HVAC}}^{i,t+1}\!\!=\!\!\varepsilon^{i}_{\texttt{HVAC}}T_{\texttt{HVAC}}^{i,t}\!\!+\!\!(1\!\!-\!\!\varepsilon^{i}_{\texttt{HVAC}})\left(T_{\texttt{out}}^{i,t}\!\!-\!\!\frac{\eta_{\texttt{HVAC}}^{i}}{A_{\texttt{HVAC}}^{i}}P_{\texttt{HVAC}}^{i,t}\right) (9)

Where THVACi,tT_{\texttt{HVAC}}^{i,t} is the indoor temperature at time tt, Touti,tT_{\texttt{out}}^{i,t} is the forecasted outdoor temperature, εHVACi\varepsilon^{i}_{\texttt{HVAC}} is the factor of inertia, ηHVACi\eta_{\texttt{HVAC}}^{i} is the coefficient of performance, AHVACiA_{\texttt{HVAC}}^{i} is thermal conductivity. Equation (10) introduces the concept of adaptive comfort model [Tmini,Tmaxi][T^{i}_{\texttt{min}},T^{i}_{\texttt{max}}]. Equation (11) enforces the control range within the size of air-conditioning PHVACmaxiP^{i}_{\texttt{HVACmax}}.

TminiTHVACi,t+1Tmaxi\displaystyle T^{i}_{\texttt{min}}\leq T_{\texttt{HVAC}}^{i,t+1}\leq T^{i}_{\texttt{max}} (10)
0PHVACi,tPHVACmaxi\displaystyle 0\leq P_{\texttt{HVAC}}^{i,t}\leq P^{i}_{\texttt{HVACmax}} (11)

II-C4 Constraints Pertaining to Plug-in Electric Vehicles (PEV)

i\forall i and t[ts,te]\forall t\in[t_{\texttt{s}},t_{\texttt{e}}], the PEV charging power PPEVi,tP_{\texttt{PEV}}^{i,t} must adhere to the range [PPEVmini,PPEVmaxi][P_{\texttt{PEVmin}}^{i},P_{\texttt{PEVmax}}^{i}] as described in (12). Further, (13) mandates that agent ii’s cumulative charging power meet the necessary energy EPEViE_{\texttt{PEV}}^{i} for daily commute [42].

PPEVminiPPEVi,tPPEVmaxi\displaystyle P_{\texttt{PEVmin}}^{i}\leq P_{\texttt{PEV}}^{i,t}\leq P_{\texttt{PEVmax}}^{i} (12)
t=tstePPEVi,tEPEVi\displaystyle\sum_{t=t_{\texttt{s}}}^{t_{\texttt{e}}}P_{\texttt{PEV}}^{i,t}\geq E_{\texttt{PEV}}^{i} (13)

II-C5 Constraints Pertaining to Photovoltaic Arrays

The photovoltaic power generation, given by (14) and is determined by the solar irradiance-power conversion function. Here, RPVtR_{\texttt{PV}}^{t}, represents the solar radiation intensity, APVA_{\texttt{PV}} denotes the surface area, and ηPV\eta_{\texttt{PV}} is the transformation efficiency.

PPVi,t=RPVtAPVηPV\displaystyle P_{\texttt{PV}}^{i,t}=R_{\texttt{PV}}^{t}A_{\texttt{PV}}\eta_{\texttt{PV}} (14)

II-C6 Constraints of Network Sharing

The net power of agent ii, POi,tP_{\texttt{O}}^{i,t}, is given below. Note, PILi,tP_{\texttt{IL}}^{i,t} indicates the inflexible loads.

POi,t=PPVi,t+\displaystyle P_{\texttt{O}}^{i,t}\!\!=\!\!P_{\texttt{PV}}^{i,t}+ PESSDi,tPESSCi,tPILi,tPFLi,tPHVACi,tPPEVi,t\displaystyle P_{\texttt{ESSD}}^{i,t}\!\!-\!\!P_{\texttt{ESSC}}^{i,t}\!\!-\!\!P_{\texttt{IL}}^{i,t}\!\!-\!\!P_{\texttt{FL}}^{i,t}\!\!-\!\!P_{\texttt{HVAC}}^{i,t}\!\!-\!\!P_{\texttt{PEV}}^{i,t} (15)

Local distribution utility constraints are enforced by (16), while (17) guarantees that VPP’s output honors the production schedule of both energy markets.

POminiPOi,tPOmaxi\displaystyle P_{\texttt{Omin}}^{i}\leq P_{\texttt{O}}^{i,t}\leq P_{\texttt{Omax}}^{i} (16)
i𝒩APOi,t=PScht\displaystyle\sum_{i\in\mathcal{N}_{\texttt{A}}}P_{\texttt{O}}^{i,t}=P_{\texttt{Sch}}^{t} (17)

II-C7 Objective Function

The objective function for the power dispatch problem, i.e., (18), includes:

Minimizing maintenance & operation costs of energy storage systems

αESSi\alpha_{\texttt{ESS}}^{i} represents the unit maintenance cost.

Balancing the differences between actual and preset consumption profiles for flexible loads

αFLi,t\alpha_{\texttt{FL}}^{i,t} is the inconvenience coefficient. Here, PFLrefi,tP_{\texttt{FLref}}^{i,t} specifies the preferred consumption level [43].

Mitigating thermal discomfort costs for HVAC systems

αHVACi,t\alpha_{\texttt{HVAC}}^{i,t} is the cost coefficient, TRefi,tT_{\texttt{Ref}}^{i,t} indicates the optimal comfort level, and binary variable βHVACi,t\beta_{\texttt{HVAC}}^{i,t} denotes occupancy state, where 1 means occupied and 0 indicates vacancy.

f=t=tstei𝒩A(αESSi(PESSCi,t+PESSDi,t)+αFLi(PFLi,t\displaystyle f=\sum_{t=t_{\texttt{s}}}^{t_{\texttt{e}}}\sum_{i\in\mathcal{N}_{\texttt{A}}}\left(\alpha_{\texttt{ESS}}^{i}(P_{\texttt{ESSC}}^{i,t}+P_{\texttt{ESSD}}^{i,t})+\alpha_{\texttt{FL}}^{i}(P_{\texttt{FL}}^{i,t}\right.
PFLrefi,t)2+βHVACi,tαHVACi(THVACi,tTRefi,t)2)\displaystyle\left.-P_{\texttt{FLref}}^{i,t})^{2}+\beta_{\texttt{HVAC}}^{i,t}\alpha_{\texttt{HVAC}}^{i}{(T_{\texttt{HVAC}}^{i,t}-T_{\texttt{Ref}}^{{i,t}})}^{2}\right) (18)

II-C8 Centralized Optimization Problem

Combining the constraints (5)-(17) and the objective function (18), we formulate the power dispatch problem. Note the formulated dispatch problem requires frequent resolution at each time instance tst_{\texttt{s}} in the real-time market. For a given agent ii, the optimization variables over the time interval [ts,te][t_{\texttt{s}},t_{\texttt{e}}] are denoted by 𝐮i(t)\mathbf{u}^{i}(t), while its inputs over the same interval are represented as 𝐱i\mathbf{x}^{i};

𝐮i=[PFLi,t,PESSCi,t,PESSDi,t,RSoCi,t+1,PHVACi,t,THVACi,t+1,PPEVi,t,PPVi,t,\displaystyle\mathbf{u}^{i}={\left[P_{\texttt{FL}}^{i,t},P_{\texttt{ESSC}}^{i,t},P_{\texttt{ESSD}}^{i,t},R_{\texttt{SoC}}^{i,t+1},P_{\texttt{HVAC}}^{i,t},T_{\texttt{HVAC}}^{i,t+1},P_{\texttt{PEV}}^{i,t},P_{\texttt{PV}}^{i,t},\right.}
POi,tt=ts,,te]\displaystyle{\left.P_{\texttt{O}}^{i,t}\mid t=t_{\texttt{s}},\ldots,t_{\texttt{e}}\right]} (19)
𝐱i=[PFLmini,t,PFLmaxi,t,RSoCi,ts,THVACi,ts,Touti,t,EPEVi,RPVt,PILi,t,\displaystyle\mathbf{x}^{i}={\left[P_{\texttt{FLmin}}^{i,t},P_{\texttt{FLmax}}^{i,t},R_{\texttt{SoC}}^{i,t_{\texttt{s}}},T_{\texttt{HVAC}}^{i,t_{\texttt{s}}},T_{\texttt{out}}^{i,t},E_{\texttt{PEV}}^{i},R_{\texttt{PV}}^{t},P_{\texttt{IL}}^{i,t},\right.}
PScht,PFLrefi,t,βHVACi,t,TRefi,tt=ts,,te]\displaystyle{\left.P_{\texttt{Sch}}^{t},P_{\texttt{FLref}}^{i,t},\beta_{\texttt{HVAC}}^{i,t},T_{\texttt{Ref}}^{i,t}\mid t=t_{\texttt{s}},\ldots,t_{\texttt{e}}\right]} (20)

Let 𝐮=i𝐮i\mathbf{u}=\bigoplus_{i}\mathbf{u}^{i} and 𝐱=i𝐱i\mathbf{x}=\bigoplus_{i}\mathbf{x}^{i}. The DER coordination problem can be formulated as (21) or as follows,

minf(𝐮,𝐱)\displaystyle\min f(\mathbf{u},\mathbf{x}) (21a)
s.t. 𝐀eq𝐮+𝐁eq𝐱+𝐛eq=𝟎\displaystyle\mathbf{A}_{\texttt{eq}}\mathbf{u}+\mathbf{B}_{\texttt{eq}}\mathbf{x}+\mathbf{b}_{\texttt{eq}}=\mathbf{0} (21b)
𝐀ineq𝐮+𝐁ineq𝐱+𝐛ineq𝟎\displaystyle\mathbf{A}_{\texttt{ineq}}\mathbf{u}+\mathbf{B}_{\texttt{ineq}}\mathbf{x}+\mathbf{b}_{\texttt{ineq}}\leq\mathbf{0} (21c)

where 𝐀eq\mathbf{A}_{\texttt{eq}}, 𝐁eq\mathbf{B}_{\texttt{eq}} and 𝐛eq\mathbf{b}_{\texttt{eq}} represent the compact form of parameters in equations (7), (9), (14), (15), and (17) we have formed before. And 𝐀ineq\mathbf{A}_{\texttt{ineq}}, 𝐁ineq\mathbf{B}_{\texttt{ineq}} and 𝐛ineq\mathbf{b}_{\texttt{ineq}} captures parameters in equations (5), (6), (8), (10)-(13), (16).

II-D Agent-based Model for the VPP Coordination Problem

Agent-based problem-solving lends itself well to addressing the computational needs of the VPP coordination problem. In this subsection, we focus on finding a distributed solution for (21) (or (1)). While each sub-problem optimizes the operation of individual agents, communication enables individual agents to collectively find the system-level optimal solution.

In the context of distributed problem-solving, it is important to point out the unique challenges posed by coupling constraints such as (17). These constraints introduce intricate relationships among several agents where some variables of agent ii are tied with those of agent jj. These coupled constraints prevent separating (21) into disjointed sub-problems.

As discussed in Section IIA, we define the variables present among multiple agents’ constraints as global variables, 𝐮gi\mathbf{u}^{i}_{\texttt{g}},

𝐮gi=[POi,tt=ts,,te]T\displaystyle\mathbf{u}^{i}_{\texttt{g}}={\left[P_{\texttt{O}}^{i,t}\mid t=t_{\texttt{s}},\ldots,t_{\texttt{e}}\right]}^{\texttt{T}} (22)

In contrast, the variables solely managed by non-overlapping constraints are referred to as local variables. That is, 𝐮i=[𝐮li,𝐮gi]\mathbf{u}^{i}=[\mathbf{u}^{i}_{\texttt{l}},\mathbf{u}^{i}_{\texttt{g}}]. We refer to agents whose variables are intertwined in a constraint as neighboring agents.

The ADMM method finds a decentralized solution for (21) by creating local copies of neighboring agents’ global variables and adjusting local copies iteratively to satisfy both local and consensus constraints. The adjustment continues until alignment with original global variables is achieved, at which point the global minimum has been found in a decentralized manner.

In the power dispatch problem, we introduce POCopyi,j,tP_{\texttt{OCopy}}^{i,j,t}, which is owned by agent ii, and represents a copy of POj,tP_{\texttt{O}}^{j,t}. Then, coupled constraint (17) become a local constraint (23) and a consensus constraint (24):

POi,t+jiPOCopyi,j,t=PScht\displaystyle P_{\texttt{O}}^{i,t}+\sum_{j\neq i}P_{\texttt{OCopy}}^{i,j,t}=P_{\texttt{Sch}}^{t} (23)
POCopyi,j,t=POj,t,ji\displaystyle P_{\texttt{OCopy}}^{i,j,t}=P_{\texttt{O}}^{j,t},\forall j\neq i (24)

Let 𝐮g,Copyi=[POCopyi,j,tt=ts,,te]\mathbf{u}^{i}_{\texttt{g,Copy}}={[P_{\texttt{OCopy}}^{i,j,t}\mid t=t_{\texttt{s}},\ldots,t_{\texttt{e}}]} denote all local copies owned by agent ii imitating other neighboring agents’ global variables. Then, one could reformulate the problem (21) in accordance to 𝐮i\mathbf{u}^{i} and 𝐮g,Copyi\mathbf{u}^{i}_{\texttt{g,Copy}} as,

minifi(𝐮i,𝐮g,Copyi,𝐱i)\displaystyle\min\sum_{i}f^{i}(\mathbf{u}^{i},\mathbf{u}^{i}_{\texttt{g,Copy}},\mathbf{x}^{i}) (25a)
s.t.𝐀eqi[𝐮i𝐮g,Copyi]+𝐁eqi𝐱i+𝐛eqi=𝟎,i\displaystyle\texttt{s.t.}~{}~{}~{}\mathbf{A}^{i}_{\texttt{eq}}\begin{bmatrix}\mathbf{u}^{i}\\ \mathbf{u}^{i}_{\texttt{g,Copy}}\end{bmatrix}+\mathbf{B}^{i}_{\texttt{eq}}\mathbf{x}^{i}+\mathbf{b}^{i}_{\texttt{eq}}=\mathbf{0},\forall i (25b)
𝐀ineqi[𝐮i𝐮g,Copyi]+𝐁ineqi𝐱i+𝐛ineqi𝟎,i\displaystyle\mathbf{A}^{i}_{\texttt{ineq}}\begin{bmatrix}\mathbf{u}^{i}\\ \mathbf{u}^{i}_{\texttt{g,Copy}}\end{bmatrix}+\mathbf{B}^{i}_{\texttt{ineq}}\mathbf{x}^{i}+\mathbf{b}^{i}_{\texttt{ineq}}\leq\mathbf{0},\forall i (25c)
𝐮g,Copyi=𝐈ci[ji𝐮gj],i\displaystyle\mathbf{u}^{i}_{\texttt{g,Copy}}=\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j}],\forall i (25d)

where, 𝐀eqi\mathbf{A}^{i}_{\texttt{eq}}, 𝐁eqi\mathbf{B}^{i}_{\texttt{eq}}, 𝐛eqi\mathbf{b}^{i}_{\texttt{eq}}, 𝐀ineqi\mathbf{A}^{i}_{\texttt{ineq}}, 𝐁ineqi\mathbf{B}^{i}_{\texttt{ineq}}, and 𝐛ineqi\mathbf{b}^{i}_{\texttt{ineq}} in (25b) and (25c) capture the compact form of constraints (5)-(16), (23). And (25d) is the compact form of constraints (24). Here 𝐈ci\mathbf{I}_{\texttt{c}}^{i} is the element selector matrix that maps elements from vector ji𝐮gj\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j} to vector 𝐮g,Copyi\mathbf{u}^{i}_{\texttt{g,Copy}} based on a consensus constraint (24). Each row of 𝐈ci\mathbf{I}_{\texttt{c}}^{i} contains a single 1 at a position that corresponds to the desired element from ji𝐮gj\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j} and 0s elsewhere. Therefore, 𝐈ci[ji𝐮gj]\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j}] represents the vector of global variables that are required to be imitated by agent ii.

Let 𝒮Locali\mathcal{S}_{\texttt{Local}}^{i} be the set of local constraints associated with agent ii, i.e., (25b)-(25c). Therefore, the compact form of decentralized formulation at the agent-level as defined in (2).

II-E Updating Rules Within Agents

The standard form of ADMM solves problem (25) (or (2)) by dealing with the augmented Lagrangian function L\mathrm{L}:

minL=\displaystyle\min\mathrm{L}\!\!=\!\!\! i(fi(𝐮i,𝐮g,Copyi,𝐱i)+𝝀iT(𝐮g,Copyi𝐈ci[ji𝐮gj])\displaystyle\sum_{i}\!\!\left(f^{i}(\mathbf{u}^{i},\mathbf{u}^{i}_{\texttt{g,Copy}},\mathbf{x}^{i})\!\!+\!\!{\boldsymbol{\lambda}}^{i\texttt{T}}(\mathbf{u}^{i}_{\texttt{g,Copy}}\!\!\!\!-\!\!\mathbf{I}^{i}_{\texttt{c}}[\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j}])\right.
+ρ𝐮g,Copyi𝐈ci[ji𝐮j]22)\displaystyle\left.+\rho\left\|\mathbf{u}^{i}_{\texttt{g,Copy}}-\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}^{j}]\right\|_{2}^{2}\right) (26a)
[𝐮i𝐮g,Copyi]𝒮Locali(𝐱i),i\displaystyle\begin{bmatrix}\mathbf{u}^{i}\\ \mathbf{u}^{i}_{\texttt{g,Copy}}\end{bmatrix}\in\mathcal{S}_{\texttt{Local}}^{i}(\mathbf{x}^{i}),\forall i (26b)

where ρ>0\rho>0 is a positive constant. 𝝀i\boldsymbol{\lambda}^{i} denotes the vector of all Lagrangian multipliers for the corresponding consensus equality relationship between agent ii’s copy and neighboring agent jj’s global variable.

The search for a solution to (26) is performed through an iterative process (indexed by [k],k=1,,NK[k],k=1,...,N_{\texttt{K}}). All NAN_{\texttt{A}} agents will execute this process simultaneously and independently before communicating with neighboring agents. At the agent level, these updates manifest themselves as follows,

𝝀i[k]=𝝀i[k1]+ρ(𝐮g,Copyi[k1]𝐈ci[ji𝐮gj[k1]])\displaystyle\boldsymbol{\lambda}^{i^{[k]}}=\boldsymbol{\lambda}^{i^{[k-1]}}+\rho(\mathbf{u}^{i^{[k-1]}}_{\texttt{g,Copy}}-\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j^{[k-1]}}]) (27)
[𝐮i[k]𝐮g,Copyi[k]]=argminL(𝝀i[k],ji(𝐮j[k1],𝐮g,Copyj[k1],𝝀j[k])),\displaystyle\begin{bmatrix}\!\!\!\!\mathbf{u}^{i^{[k]}}\\ \mathbf{u}^{i^{[k]}}_{\texttt{g,Copy}}\!\!\end{bmatrix}\!\!\!=\!\!\arg\min\mathrm{L}\!\!\left(\!\!\boldsymbol{\lambda}^{i^{[k]}}\!\!\!,\bigoplus_{j\neq i}\!\!\left(\mathbf{u}^{j^{[k-1]}}\!\!\!,\mathbf{u}^{j^{[k-1]}}_{\texttt{g,Copy}},\boldsymbol{\lambda}^{j^{[k]}}\right)\!\!\!\right)\!\!,
s.t.[𝐮i[k]𝐮g,Copyi[k]]𝒮Locali(𝐱i)\displaystyle\texttt{s.t.}\begin{bmatrix}\mathbf{u}^{i^{[k]}}\\ \mathbf{u}^{i^{[k]}}_{\texttt{g,Copy}}\end{bmatrix}\in\mathcal{S}_{\texttt{Local}}^{i}(\mathbf{x}^{i}) (28)

The dual update equation, i.e., (27), modifies the Lagrangian multipliers to estimate the discrepancies between an agent’s local copy of variables (designed to emulate the global variables of its neighbors) and the actual global variables held by those neighbors. Subsequently, (28) provides an optimization solution leveraging prior iteration data from other agents.

It’s essential to note that agent ii doesn’t require all the updated values from other agents to update equations (27) and (28). Agent ii primarily needs:

  • Neighboring agents’ global variables: 𝐈ci[ji𝐮gj]\mathbf{I}_{\texttt{c}}^{i}[\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g}}^{j}]. In the context of the distributed DER problem, agent ii requires values of POj,t[k1]P_{\texttt{O}}^{j,t^{[k-1]}} from their neighboring agent jj.

  • Neighboring agents’ local copies mirroring agent ii’s global variables: 𝐈gi[ji𝐮g,Copyj[k1]]\mathbf{I}_{\texttt{g}}^{i}[\bigoplus_{j\neq i}\mathbf{u}_{\texttt{g,Copy}}^{j^{[k-1]}}], where 𝐈gi\mathbf{I}_{\texttt{g}}^{i} functions as a selector matrix. In the distributed DER context, agent ii requires POCopyj,i,t[k]P_{\texttt{OCopy}}^{{j,i,t}^{[k]}} from their neighboring agent jj.

We use 𝐮Otheri[k1]\mathbf{u}^{i^{[k-1]}}_{\texttt{Other}} to represent the set of variables owned by other agents but are needed by agent ii to update (27) and (28). Finally, the intra-agent updates are represented by (4) and (3).

The standard form of ADMM guarantees the feasibility of local constraints by (4) and penalizes violations of consensus constraints by iteratively updating Lagrangian multipliers as (3). In what follows, we will propose a ML-based method to accelerate ADMM for decentralized DER coordination. The ADMM iterations will guide the consensus protocol, while the gauge map [39] is adopted to enforce hard local constraints.

III Proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  Methodology

III-A Overview of the Method

This section provides a high-level overview of the 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method to incorporate ML to accelerate the ADMM algorithm. As shown in Fig. 2, instead of solving agent-level local optimization problems (4) by an iterative solver, we will train NA{N}_{\texttt{A}} agent-level neural approximators ξi,i𝒩A\xi^{i},i\in\mathcal{N}_{\texttt{A}} to directly map inputs to optimized value of agent’s optimization variables in a single feed-forward. The resulting prediction of each agent ii, denoted as 𝐮i[k]\mathbf{u}^{i^{[k]}}, will be trained to approximate the optimal solution of (2).

𝐮i[k],𝐮g,Copyi[k]=ξi(𝐱i,𝐮Otheri[k1])\displaystyle\mathbf{u}^{i^{[k]}},\mathbf{u}^{i^{[k]}}_{\texttt{g,Copy}}=\xi^{i}\left(\mathbf{x}^{i},\mathbf{u}^{i^{[k-1]}}_{\texttt{Other}}\right) (29)
Refer to caption
Figure 2: Comparison between ADMM and the proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method. The ADMM approach is comprised of two key components: the dual update and the primal optimization. The dual update guides the consensus protocol, while the primal optimization leverages this estimation to adjust the local decision-making process through iterative solvers. In contrast, our 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method replaces these two procedures with a single-step data infusion and mapping input parameters and other agents’ values to agent-level optimized values, directly.

Pseudo code of the proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method is given in Algorithm 1. 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method includes two steps for each iteration. First, each agent receives variables of prior iteration from neighboring agents. Second, each agent uses a neural approximator to predict its optimal values.

Algorithm 1 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method
Input: DER coordination problem parameters, e.g., NA{N}_{\texttt{A}} neural approximators ξi,i\xi^{i},\forall i, input parameters 𝐱i,i\mathbf{x}^{i},\forall i; Initial value of 𝐮i\mathbf{u}^{i} and 𝝀i,i\boldsymbol{\lambda}^{i},\forall i.
Output: Distributed solution 𝐮i[k+1]\mathbf{u}^{i^{[k+1]}} to DER coordination problem.
while Convergence criteria unmet do
     for ii in range(𝒩A\mathcal{N}_{\texttt{A}}do
         \bullet Send previous global variable 𝐮gi[k1]\mathbf{u}^{i^{[k-1]}}_{\texttt{g}} and local copy 𝐮g,Copyi[k1]\mathbf{u}^{i^{[k-1]}}_{\texttt{g,Copy}} to neighboring agents, and receive 𝐮Otheri[k1]\mathbf{u}^{i^{[k-1]}}_{\texttt{Other}}
         \bullet Generate prediction 𝐮i[k],𝐮g,Copyi[k]=ξi(𝐱i,𝐮Otheri[k1])\mathbf{u}^{i^{[k]}},\mathbf{u}^{i^{[k]}}_{\texttt{g,Copy}}\!=\!\xi^{i}(\mathbf{x}^{i},\mathbf{u}^{i^{[k-1]}}_{\texttt{Other}})
         k++k++
     end for
end while

III-B Design of Neural Approximators Structures

Violations of consensus constraints could be penalized by ADMM iterations. Further, we will design each neural approximator’s structure to guarantee that its output satisfies the local constraints, i.e., ξi𝒮Locali(𝐱i)\xi^{i}\in\mathcal{S}_{\texttt{Local}}^{i}(\mathbf{x}^{i}). We adopt the 𝒪𝒪𝒫𝒞\mathcal{LOOP-LC} (Learning to Optimize the Optimization Process with Linear Constraints) model proposed in [39] to develop each neural approximator ξi\xi^{i}. The 𝒪𝒪𝒫𝒞\mathcal{LOOP-LC} model learns to solve optimization problems with hard linear constraints. It applies variable elimination and gauge mapping for equality and inequality completions, respectively. The 𝒪𝒪𝒫𝒞\mathcal{LOOP-LC} model produces a feasible and near-optimal solution. In what follows, we will present the main components of 𝒪𝒪𝒫𝒞\mathcal{LOOP-LC} and how it applies to the VPP coordination problem.

Refer to caption
Figure 3: The structure and building blocks of Neural Approximator ξi\xi^{i}.

III-B1 Variable Elimination

Based on the equality constraints given in (25b), the variables 𝐮i\mathbf{u}^{i} and 𝐮g,Copyi\mathbf{u}^{i}_{\texttt{g,Copy}} can be categorized into two sets: the dependent variables 𝐮Depi\mathbf{u}^{i}_{\texttt{Dep}} and the independent variables 𝐮Indi\mathbf{u}^{i}_{\texttt{Ind}}. The dependent variables are inherently determined by the independent variables. For instance in (9), the variable THVACi,t+1T_{\texttt{HVAC}}^{i,t+1} is dependent on PHVACi,tP_{\texttt{HVAC}}^{i,t}; hence, once PHVACi,tP_{\texttt{HVAC}}^{i,t} is derived, THVACi,t+1T_{\texttt{HVAC}}^{i,t+1} can be caculated.

The function 𝔽i\mathbb{F}^{i} is introduced to establish the relationship between 𝐮Depi\mathbf{u}^{i}_{\texttt{Dep}} and 𝐮Indi\mathbf{u}^{i}_{\texttt{Ind}}, such that 𝐮Depi=𝔽i(𝐮Indi)\mathbf{u}^{i}_{\texttt{Dep}}=\mathbb{F}^{i}(\mathbf{u}^{i}_{\texttt{Ind}}), shown in Fig. 3. A comprehensive derivation of 𝔽i\mathbb{F}^{i} can be found in [39]. By integrating 𝔽i\mathbb{F}^{i} into the definition of 𝒮Locali\mathcal{S}_{\texttt{Local}}^{i} and substituting 𝐮Depi\mathbf{u}^{i}_{\texttt{Dep}}, the optimization problem of (4) can be restructured as a reduced-dimensional problem with 𝐮Indi\mathbf{u}^{i}_{\texttt{Ind}} as the primary variable. The corresponding constraint set for this reformulated problem is denoted by 𝒮Local,Refi\mathcal{S}_{\texttt{Local,Ref}}^{i} and presented as,

𝒮Local,Refi={𝐀ineqi[𝐮Indi𝔽i(𝐮Indi)]+𝐁ineqi𝐱i+𝐛ineqi𝟎}\displaystyle\mathcal{S}_{\texttt{Local,Ref}}^{i}=\begin{Bmatrix}\mathbf{A}^{i}_{\texttt{ineq}}\begin{bmatrix}\mathbf{u}^{i}_{\texttt{Ind}}\\ \mathbb{F}^{i}(\mathbf{u}^{i}_{\texttt{Ind}})\end{bmatrix}+\mathbf{B}^{i}_{\texttt{ineq}}\mathbf{x}^{i}+\mathbf{b}^{i}_{\texttt{ineq}}\leq\mathbf{0}\end{Bmatrix} (30)

Therefore, as long as the prediction of the reformulated problem ensures 𝐮Indi𝒮Local,Refi\mathbf{u}^{i}_{\texttt{Ind}}\in\mathcal{S}_{\texttt{Local,Ref}}^{i}, 𝔽i\mathbb{F}^{i} will produce the full-size 𝐮i,𝐮g,Copyi\mathbf{u}^{i},\mathbf{u}^{i}_{\texttt{g,Copy}} vectors satisfying local constraints 𝒮Locali(𝐱i)\mathcal{S}_{\texttt{Local}}^{i}(\mathbf{x}^{i}) by concatenating 𝐮Depi\mathbf{u}^{i}_{\texttt{Dep}} and 𝐮Depi,𝐮Indi\mathbf{u}^{i}_{\texttt{Dep}},\mathbf{u}^{i}_{\texttt{Ind}}.

III-B2 Gauge Map

After variable elimination, our primary objective is to predict 𝐮Indi\mathbf{u}^{i}_{\texttt{Ind}} such that it satisfies the constraint set 𝒮Local,Refi\mathcal{S}_{\texttt{Local,Ref}}^{i}. Instead of directly solving this problem, we will utilize a neural network that finds a virtual prediction 𝐮^Indi\mathbf{\hat{u}}^{i}_{\texttt{Ind}} which lies within the \ell_{\infty}-norm unit ball (denoted as \mathcal{B}) a set constrained by upper and lower bounds. The architecture of the neural network is designed to ensure that the resulting 𝐮^Indi\mathbf{\hat{u}}^{i}_{\texttt{Ind}} remains confined within \mathcal{B}. Subsequently, we introduce a bijective gauge mapping, represented as 𝕋i\mathbb{T}^{i}, to transform 𝐮^Indi\mathbf{\hat{u}}^{i}_{\texttt{Ind}} from \mathcal{B} to 𝒮Local,Refi\mathcal{S}_{\texttt{Local,Ref}}^{i}. As presented in [39], 𝕋i\mathbb{T}^{i} is a predefined function with an explicit closed-form representation as below,

𝐮Indi=𝕋i(𝐮^Indi)=ψ(𝐮^Indi)ψ𝒮Local,Ref0i(𝐮^Indi)𝐮^Indi+𝐮Ind,0i\displaystyle\mathbf{u}^{i}_{\texttt{Ind}}=\mathbb{T}^{i}(\mathbf{\hat{u}}^{i}_{\texttt{Ind}})=\frac{\psi_{\mathcal{B}}(\mathbf{\hat{u}}^{i}_{\texttt{Ind}})}{\psi_{\mathcal{S}_{\texttt{Local,Ref0}}^{i}}(\mathbf{\hat{u}}^{i}_{\texttt{Ind}})}\mathbf{\hat{u}}^{i}_{\texttt{Ind}}+\mathbf{u}^{i}_{\texttt{Ind,0}} (31)

The function ψ\psi_{\mathcal{B}} is the Minkowski gauge of the set \mathcal{B}, while 𝐮Ind,0i\mathbf{u}^{i}_{\texttt{Ind,0}} represents an interior point of 𝒮Local,Refi\mathcal{S}_{\texttt{Local,Ref}}^{i}. Moreover, the shifted set, 𝒮Local,Ref0i\mathcal{S}_{\texttt{Local,Ref0}}^{i}, is defined as,

𝒮Local,Ref0i={𝐮¯Indi(𝐮Ind,0i+𝐮¯Indi)𝒮Local,Refi}\displaystyle\mathcal{S}_{\texttt{Local,Ref0}}^{i}=\left\{\mathbf{\bar{u}}^{i}_{\texttt{Ind}}\mid\left(\mathbf{u}^{i}_{\texttt{Ind,0}}+\mathbf{\bar{u}}^{i}_{\texttt{Ind}}\right)\in\mathcal{S}_{\texttt{Local,Ref}}^{i}\right\} (32)

with ψ𝒮Local,Ref0i\psi_{\mathcal{S}_{\texttt{Local,Ref0}}^{i}} representing the Minkowski gauge on this set.

III-C Training the Neural Approximators

We use the historical trajectories of ADMM (i.e. applied on historical power demands) for training purposes. Note that predicting the converged ADMM values is a time-series prediction challenge. Specifically, outputs from a given iteration are requisites for the subsequent iterations. This relationship implies that 𝐮i[k],𝐮g,Copyi[k],i\mathbf{u}^{i^{[k]}},\mathbf{u}^{i^{[k]}}_{\texttt{g,Copy}},\forall i are contingent upon 𝐮Otheri[k1]\mathbf{u}^{i^{[k-1]}}_{\texttt{Other}}, derived from other agents’ outputs 𝐮j[k1],𝐮g,Copyj[k1],ji\mathbf{u}^{j^{[k-1]}},\mathbf{u}^{j^{[k-1]}}_{\texttt{g,Copy}},\forall j\neq i from the prior iteration. To encapsulate this temporal dependency, our training approach adopts a look-ahead format, facilitating the joint training of all neural approximators in a recurrent manner, which ensures that prior outputs from different agents are seamlessly integrated as current inputs (see Fig. 4).

Suppose there are NDN_{\texttt{D}} training data points, indexed and associated with their respective output by the superscript (d)(d). As an initial step, ADMM is employed to generate all values of optimization variables required for training. Concurrently, the optimal solution 𝐮i(d)\mathbf{u}^{i*(d)} pertaining to (2) is calculated. Subsequently, for NRN_{\texttt{R}} recurrent steps, the loss function fLf_{\texttt{L}} is defined as the cumulative distance dd between the prediction 𝐮i[k+r](d)\mathbf{u}^{i^{[k+r](d)}} and the optimal solution 𝐮i(d)\mathbf{u}^{i*(d)}. This summation spans all agents, every recurrent step, every iteration (k=1,..NKk=1,..N_{\texttt{K}}), and all data points, as delineated in (33).

fL=d=1NDk=1NKr=1NRi𝒩Ad(𝐮i[k+r](d),𝐮i(d))\displaystyle f_{\texttt{L}}=\sum_{d=1}^{N_{\texttt{D}}}\sum_{k=1}^{N_{\texttt{K}}}\sum_{r=1}^{N_{\texttt{R}}}\sum_{i\in\mathcal{N}_{\texttt{A}}}d(\mathbf{u}^{i^{[k+r](d)}},\mathbf{u}^{i*(d)}) (33)
Refer to caption
Figure 4: ADMM convergence values present a time-series prediction challenge, with outputs from one iteration feeding into the next. Our training approach uses a look-ahead format, enabling recurrent joint training of neural approximators, integrating prior outputs as current inputs.

IV Experimental Results

IV-A Experiment Setup

IV-A1 Test Systems

We examine a VPP consisting of three distinct agents, as illustrated in Fig. 5.

  • Agent 1 manages inflexible loads, flexible loads, and energy storage systems.

  • Agent 2 is responsible for inflexible loads and the operations of plug-in electric vehicles.

  • Agent 3 oversees inflexible loads, heating, ventilation, and air conditioning systems, in addition to photovoltaics.

We derive the load profile from data recorded in central New York on July 24th, 2023 [44]. Both preferred flexible and inflexible loads typically range between 10 to 25 kW. The production schedule range is set between 45 to 115 kW.

For plug-in electric vehicles, our reference is the average hourly public L2 charging station utilization on weekdays in March 2022 as presented by Borlaug et al. [45]. In [45] the profile range for EPEVi,τ,τ[0,24h]E_{\texttt{PEV}}^{i,\tau},\tau\in[0,24h] between 10 and 22 kW.

With regards to the heating, ventilation, and air conditioning systems, the target indoor temperature TRefi,tT_{\texttt{Ref}}^{i,t} is maintained at 77F{77}^{\circ}F. Guided by the ASHRAE(American Society of Heating, Refrigerating, and Air-Conditioning Engineers) standards [46], the acceptable summer comfort range is determined as Tmini=75FT^{i}_{\texttt{min}}={75}^{\circ}F and Tmaxi=79FT^{i}_{\texttt{max}}={79}^{\circ}F. External temperature readings for New York City’s Central Park on July 24th, 2023 were obtained from the National Weather Service [47].

Also, the Global CMP22 dataset from July 24th, 2023 [48] is used to calculate the regional solar radiation intensity RPVtR_{\texttt{PV}}^{t}. Supplementary parameters are presented in Table I.

IV-A2 Training Data

A total of 20 ADMM iterations are considered, i.e., NK=20N_{\texttt{K}}=20. This results in a dataset of 24×12×2024\times 12\times 20 data points. For model validation, data from odd time steps is designated for training, whereas even time steps are reserved for testing. The DER coordination problem includes 192 optimization variables alongside 111 input variables.

IV-A3 ADMM Configuration

The ADMM initialization values are set to zero. In our ADMM implementation, the parameter ρ\rho is set to 0.00050.0005. Optimization computations are carried out using the widely-accepted commercial solver, Gurobi [49].

IV-A4 Neural Network Configuration

Our neural network models consist of a single hidden layer, incorporating 500 hidden units. The Rectified Linear Unit (ReLU) activation function is employed for introducing non-linearity. To ensure that 𝐮^Indi\mathbf{\hat{u}}^{i}_{\texttt{Ind}} resides within \mathcal{B} (the \ell_{\infty} unit ball), the output layer utilizes the Hyperbolic Tangent (TanH) activation. Furthermore, 3 recurrent steps are considered, represented by NR=3N_{\texttt{R}}=3.

Refer to caption
Figure 5: Our proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method to coordinate DERs.
TABLE I: Parameters of DERs that are controlled by agents.
parameter value parameter value
αFLi,t\alpha_{\texttt{FL}}^{i,t} 0.1 [43] PESSmaxiP_{\texttt{ESSmax}}^{i} 80kW [43]
αESSi\alpha_{\texttt{ESS}}^{i} 0.01 [50] ηESSCi\eta_{\texttt{ESSC}}^{i} 0.94 [43]
ηESSDi\eta_{\texttt{ESSD}}^{i} 1.06 [43] EESSNiE_{\texttt{ESSN}}^{i} 300kW [43]
RSoCminiR_{\texttt{SoCmin}}^{i} 0.15 [43] RSoCmaxiR_{\texttt{SoCmax}}^{i} 0.85 [43]
RSoCi(τ=0)R_{\texttt{SoC}}^{i}(\tau=0) 0.2 [43] αESSi\alpha_{\texttt{ESS}}^{i} 0.01 [50]
ηHVACi\eta_{\texttt{HVAC}}^{i} 2.5 [41] AHVACiA_{\texttt{HVAC}}^{i} 0.25[41]
εHVACi\varepsilon^{i}_{\texttt{HVAC}} 0.93 [41] PHVACmaxiP^{i}_{\texttt{HVACmax}} 11.5kW [41]
βHVACi,t\beta_{\texttt{HVAC}}^{i,t} 1 αHVACi,t\alpha_{\texttt{HVAC}}^{i,t} 1
APVA_{\texttt{PV}} 1000 m2m^{2} [51] ηPV\eta_{\texttt{PV}} 0.2[51]

IV-B Runtime Results

Fig. 6 illustrates the cumulative computation time across all agents and test data points over iterations. The performance comparison is conducted among the decentralized setup employing ADMM solvers, our proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method, and traditional centralized solvers.

From the case study, it is observed that the computational time required by the classical ADMM solver exceeds the centralized solvers solution time after approximately five iterations. Remarkably, our proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} method significantly outperforms the classical ADMM, achieving 500x speed up. Also, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  even surpasses the efficiency of the centralized solver in terms of computation speed.

Refer to caption
Figure 6: Comparison of cumulative computational time across all agents and test data points. The classical ADMM approach closely matches the computational efficiency of centralized solvers after approximately five iterations. In contrast, the 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  solves the problem 500 times faster than the classical ADMM, even surpassing the centralized solvers in performance.

Table II provides the average computational time for a single iteration on a single data point. An insightful observation from the results suggests that the 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method would require around 3300 iterations to match the computational time of centralized solvers. However, based on the convergence analysis that will be provided later, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method demonstrates convergence in a mere 10 iterations.

TABLE II: Average running time for a single ADMM iteration over the decision-making time horizon .
Method Time(Millisecond)
𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method 0.0060
ADMM using Gurobi[49] 3.2966
Centralized formulation using Gurobi [49] 19.4496/NKN_{\texttt{K}}
Refer to caption
Figure 7: Comparison of the optimality deviation (error) rate for the classical ADMM versus 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}. The rate quantifies deviation (error) from the optimal DER operation profile using 𝐮i[k](d)𝐮i(d)22/𝐮i(d)22\left\|\mathbf{u}^{i^{[k](d)}}-\mathbf{u}^{i*(d)}\right\|^{2}_{2}/\left\|\mathbf{u}^{i*(d)}\right\|^{2}_{2}. 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} method distinctly reduces the iterations needed for convergence and demonstrates decreased post-convergence variance and peak deviation, demonstrating its superior capabilities in handling repetitive optimization scenarios.
Refer to caption
Figure 8: VPP output deviation rate represents the variance between the VPP’s output and planned production schedules, measured by the feasibility gap of coupled constraints (17) as t=ττ+τw(iPOi,tPScht)/PScht\sum_{t=\tau}^{\tau+\tau_{\texttt{w}}}\left(\sum_{i}P_{\texttt{O}}^{i,t}-P_{\texttt{Sch}}^{t}\right)/P_{\texttt{Sch}}^{t}. Compared to the ADMM, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  exhibits a faster and more stable convergence.

IV-C Optimality and Feasibility Results

Fig. 7 presents the optimality deviation rate for both the traditional ADMM algorithm and 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method. The deviation rate metric quantifies the degree to which the operational profiles of the DERs deviate from the optimal (derived from solving the centralized problem). It is evident that 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method achieves faster convergence. Moreover, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  showcases faster reduction of the deviation rate compared to the standard ADMM approach.

Similarly, Fig. 8 depicts the deviation rate of the VPP schedule for both the ADMM approach and 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method. This rate sheds light on the difference between the actual VPP production schedule and its planned output. In the context of our optimization problem, the deviation rate is equivalent to the feasibility gap rate of the coupled constraints, as shown in (17). Notably, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  excels in convergence speed and stability. The VPP schedule deviation rate declines more rapidly and remains stable using 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} method, whereas the traditional ADMM method results in more oscillations and converges at a slower pace.

Table III summarizes post-convergence metrics for both algorithms across all agents, iterations, and test data points. While the minimum optimality deviation rate achieved by 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  is slightly higher than that of the classical ADMM, our approach showcases a much lower variance and a significantly reduced maximum deviation. These results highlight 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method’s efficacy, especially when tasked with recurrently solving similar optimization problems. The observed improvements in variance and maximum deviation highlight the versatility and robustness of 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  in varied problem scenarios. To sum up, the proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} solution speeds up the solution time of each ADMM iteration by up to 500X. Also, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  needs fewer iterations to converge, hence, the overall run time will be significantly shorter.

TABLE III: Post-convergence statistics.
ADMM 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}
Optimality deviation rate Average 0.0527 0.0492
Variance 0.0012 0.0003
Maximum 0.1396 0.1278
Minimum 0.0023 0.0099
VPP schedule deviation rate Average 0.0611 0.0512
Variance 0.0026 0.0008
Maximum 0.2471 0.1363
Minimum 0.0097 0.0142

V Conclusion

In this work, we introduced a novel ML-based method, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}, to significantly enhance the performance of the distributed optimization techniques and discussed its performance in addressing challenges of the DER coordination problem (solved by VPP). Our multi-agent framework for VPP decision-making allows each agent to manage multiple DERs. Key to our proposed 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC} approach is the capability of each agent to predict their local power profiles and strategically communication with neighboring agents. The collective problem-solving efforts of these agents result in a near-optimal solution for power dispatching, ensuring compliance with both local and system-level constraints.

A key contribution of our work is developing and incorporating neural network approximators in the process of distributed decision-making. This novelty significantly accelerates the solution search and reduces the iterations required for convergence. Uniquely, in contrast to restoration-centric methodologies, 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  bypasses the need for auxiliary post-processing steps to achieve feasibility using a two-pronged solution approach, where local constraints are inherently satisfied through the gauge mapping technique, and coupled constraints are penalized over ADMM iterations.

The 𝒪𝒪𝒫𝒜𝒞\mathcal{LOOP-MAC}  method reduces the solution time per iteration by up to 500%. Coupled with requiring fewer iterations for convergence, the net result is a drastic reduction of overall convergence time while respecting the problem constraints and maintaining the quality of the resulting solution.

Acknowledgement

Thanks to Dr. Erik Blasch (Fellow member) for concept discussion. This research is funded under AFOSR grants #FA9550-24-1-0099 and FA9550-23-1-0203.

References

  • [1] Q. Wang, C. Zhang, Y. Ding, G. Xydis, J. Wang, and J. Østergaard, “Review of real-time electricity markets for integrating distributed energy resources and demand response,” Applied Energy, vol. 138, pp. 695–706, 2015.
  • [2] E. B.C. and A. Somani, “Impact of ferc order 2222 on der participation rules in us electricity markets,” tech. rep., Pacific Northwest National Laboratory, Richland, WA, 2022.
  • [3] B. Goia, T. Cioara, and I. Anghel, “Virtual power plant optimization in smart grids: A narrative review,” Future Internet, vol. 14, no. 5, p. 128, 2022.
  • [4] T. Navidi, A. El Gamal, and R. Rajagopal, “Coordinating distributed energy resources for reliability can significantly reduce future distribution grid upgrades and peak load,” Joule.
  • [5] H. Pandžić, J. M. Morales, A. J. Conejo, and I. Kuzle, “Offering model for a virtual power plant based on stochastic programming,” Applied Energy, vol. 105, pp. 282–292, 2013.
  • [6] E. Dall’Anese, S. S. Guggilam, A. Simonetto, Y. C. Chen, and S. V. Dhople, “Optimal regulation of virtual power plants,” IEEE transactions on power systems, vol. 33, no. 2, pp. 1868–1881, 2017.
  • [7] M. Vasirani, R. Kota, R. L. Cavalcante, S. Ossowski, and N. R. Jennings, “An agent-based approach to virtual power plants of wind power generators and electric vehicles,” IEEE Transactions on Smart Grid, vol. 4, no. 3, pp. 1314–1322, 2013.
  • [8] M. Mohammadi, J. Thornburg, and J. Mohammadi, “Towards an energy future with ubiquitous electric vehicles: Barriers and opportunities,” Energies, vol. 16, no. 17, p. 6379, 2023.
  • [9] M. Mohammadi and A. Mohammadi, “Empowering distributed solutions in renewable energy systems and grid optimization,” in Distributed Machine Learning and Optimization: Theory and Applications, pp. 1–17, Springer, 2023.
  • [10] N. Ruiz, I. Cobelo, and J. Oyarzabal, “A direct load control model for virtual power plant management,” IEEE Transactions on Power Systems, vol. 24, no. 2, pp. 959–966, 2009.
  • [11] A. Bagchi, L. Goel, and P. Wang, “Adequacy assessment of generating systems incorporating storage integrated virtual power plants,” IEEE Transactions on Smart Grid, vol. 10, no. 3, pp. 3440–3451, 2018.
  • [12] A. Mnatsakanyan and S. W. Kennedy, “A novel demand response model with an application for a virtual power plant,” IEEE Transactions on Smart Grid, vol. 6, no. 1, pp. 230–237, 2014.
  • [13] A. Thavlov and H. W. Bindner, “Utilization of flexible demand in a virtual power plant set-up,” IEEE Transactions on Smart Grid, vol. 6, no. 2, pp. 640–647, 2014.
  • [14] A. Cherukuri and J. Cortés, “Distributed coordination of ders with storage for dynamic economic dispatch,” IEEE transactions on automatic control, vol. 63, no. 3, pp. 835–842, 2017.
  • [15] E. G. Kardakos, C. K. Simoglou, and A. G. Bakirtzis, “Optimal offering strategy of a virtual power plant: A stochastic bi-level approach,” IEEE Transactions on Smart Grid, vol. 7, no. 2, pp. 794–806, 2015.
  • [16] M. Giuntoli and D. Poli, “Optimized thermal and electrical scheduling of a large scale virtual power plant in the presence of energy storages,” IEEE Transactions on Smart Grid, vol. 4, no. 2, pp. 942–955, 2013.
  • [17] A. G. Zamani, A. Zakariazadeh, and S. Jadid, “Day-ahead resource scheduling of a renewable energy based virtual power plant,” Applied Energy, vol. 169, pp. 324–340, 2016.
  • [18] H. Wang, Y. Jia, C. S. Lai, and K. Li, “Optimal virtual power plant operational regime under reserve uncertainty,” IEEE Transactions on Smart Grid, vol. 13, no. 4, pp. 2973–2985, 2022.
  • [19] S. Hadayeghparast, A. S. Farsangi, and H. Shayanfar, “Day-ahead stochastic multi-objective economic/emission operational scheduling of a large scale virtual power plant,” Energy, vol. 172, pp. 630–646, 2019.
  • [20] S. R. Dabbagh and M. K. Sheikh-El-Eslami, “Risk assessment of virtual power plants offering in energy and reserve markets,” IEEE Transactions on Power Systems, vol. 31, no. 5, pp. 3572–3582, 2015.
  • [21] G. Chen and J. Li, “A fully distributed admm-based dispatch approach for virtual power plant problems,” Applied Mathematical Modelling, vol. 58, pp. 300–312, 2018.
  • [22] D. K. Molzahn, F. Dörfler, H. Sandberg, S. H. Low, S. Chakrabarti, R. Baldick, and J. Lavaei, “A survey of distributed optimization and control algorithms for electric power systems,” IEEE Transactions on Smart Grid, vol. 8, no. 6, pp. 2941–2962, 2017.
  • [23] T. Yang, X. Yi, J. Wu, Y. Yuan, D. Wu, Z. Meng, Y. Hong, H. Wang, Z. Lin, and K. H. Johansson, “A survey of distributed optimization,” Annual Reviews in Control, vol. 47, pp. 278–305, 2019.
  • [24] Y. Wang, S. Wang, and L. Wu, “Distributed optimization approaches for emerging power systems operation: A review,” Electric Power Systems Research, vol. 144, pp. 127–135, 2017.
  • [25] A. H. Fitwi, D. Nagothu, Y. Chen, and E. Blasch, “A distributed agent-based framework for a constellation of drones in a military operation,” in 2019 Winter Simulation Conference (WSC), pp. 2548–2559, IEEE, 2019.
  • [26] Z. Li, Q. Guo, H. Sun, and H. Su, “Admm-based decentralized demand response method in electric vehicle virtual power plant,” in 2016 IEEE Power and Energy Society General Meeting (PESGM), pp. 1–5, IEEE, 2016.
  • [27] L. Dong, S. Fan, Z. Wang, J. Xiao, H. Zhou, Z. Li, and G. He, “An adaptive decentralized economic dispatch method for virtual power plant,” Applied Energy, vol. 300, p. 117347, 2021.
  • [28] D. E. R. T. Force, “Der integration into wholesale markets and operations,” tech. rep., Energy Systems Integration Group, Reston, VA, 2022.
  • [29] F. Darema, E. P. Blasch, S. Ravela, and A. J. Aved, “The dynamic data driven applications systems (dddas) paradigm and emerging directions,” Handbook of Dynamic Data Driven Applications Systems: Volume 2, pp. 1–51, 2023.
  • [30] E. Blasch, H. Li, Z. Ma, and Y. Weng, “The powerful use of ai in the energy sector: Intelligent forecasting,” arXiv preprint arXiv:2111.02026, 2021.
  • [31] D. Biagioni, P. Graf, X. Zhang, A. S. Zamzam, K. Baker, and J. King, “Learning-accelerated admm for distributed dc optimal power flow,” IEEE Control Systems Letters, vol. 6, pp. 1–6, 2020.
  • [32] M. Li, S. Kolouri, and J. Mohammadi, “Learning to optimize distributed optimization: Admm-based dc-opf case study,” in 2023 IEEE Power & Energy Society General Meeting (PESGM), pp. 1–5, IEEE, 2023.
  • [33] T. W. Mak, M. Chatzos, M. Tanneau, and P. Van Hentenryck, “Learning regionally decentralized ac optimal power flows with admm,” IEEE Transactions on Smart Grid, 2023.
  • [34] G. Tsaousoglou, P. Ellinas, and E. Varvarigos, “Operating peer-to-peer electricity markets under uncertainty via learning-based, distributed optimal control,” Applied Energy, vol. 343, p. 121234, 2023.
  • [35] A. Mohammadi and A. Kargarian, “Learning-aided asynchronous admm for optimal power flow,” IEEE Transactions on Power Systems, vol. 37, no. 3, pp. 1671–1681, 2021.
  • [36] W. Cui, J. Li, and B. Zhang, “Decentralized safe reinforcement learning for inverter-based voltage control,” Electric Power Systems Research, vol. 211, p. 108609, 2022.
  • [37] W. Cui, G. Shi, Y. Shi, and B. Zhang, “Leveraging predictions in power system frequency control: an adaptive approach,” arXiv e-prints, pp. arXiv–2305, 2023.
  • [38] M. Al-Saffar and P. Musilek, “Distributed optimization for distribution grids with stochastic der using multi-agent deep reinforcement learning,” IEEE access, vol. 9, pp. 63059–63072, 2021.
  • [39] M. Li, S. Kolouri, and J. Mohammadi, “Learning to solve optimization problems with hard linear constraints,” IEEE Access, 2023.
  • [40] T. Cioara, M. Antal, and C. Pop, “Deliverable d3.3-consumption flexibility models and aggregation techniques,” tech. rep., H2020 eDREAM, 2019.
  • [41] Y.-Y. Hong, J.-K. Lin, C.-P. Wu, and C.-C. Chuang, “Multi-objective air-conditioning control considering fuzzy parameters using immune clonal selection programming,” IEEE Transactions on Smart Grid, vol. 3, no. 4, pp. 1603–1610, 2012.
  • [42] Y. Wang, Z. Yang, M. Mourshed, Y. Guo, Q. Niu, and X. Zhu, “Demand side management of plug-in electric vehicles and coordinated unit commitment: A novel parallel competitive swarm optimization method,” Energy conversion and management, vol. 196, pp. 935–949, 2019.
  • [43] S. Cui, Y.-W. Wang, and J.-W. Xiao, “Peer-to-peer energy sharing among smart energy buildings by distributed transaction,” IEEE Transactions on Smart Grid, vol. 10, no. 6, pp. 6491–6501, 2019.
  • [44] N. Y. I. S. Operator, “Real-time load data for new york city’s central park,” 2023.
  • [45] B. Borlaug, F. Yang, E. Pritchard, E. Wood, and J. Gonder, “Public electric vehicle charging station utilization in the united states,” Transportation Research Part D: Transport and Environment, vol. 114, p. 103564, 2023.
  • [46] A. Standard, “Thermal environmental conditions for human occupancy,” ANSI/ASHRAE, 55, vol. 5, 1992.
  • [47] N. W. Service, “Weather data for new york city’s central park,” 2023.
  • [48] T. Stoffel and A. Andreas, “Nrel solar radiation research laboratory (srrl): Baseline measurement system (bms); golden, colorado (data),” tech. rep., National Renewable Energy Lab.(NREL), Golden, CO (United States), 1981.
  • [49] Gurobi Optimization, LLC, “Gurobi Optimizer Reference Manual.”
  • [50] H. Wang and J. Huang, “Incentivizing energy trading for interconnected microgrids,” IEEE Transactions on Smart Grid, vol. 9, no. 4, pp. 2647–2657, 2016.
  • [51] G. Li, D. Wu, J. Hu, Y. Li, M. S. Hossain, and A. Ghoneim, “Helos: Heterogeneous load scheduling for electric vehicle-integrated microgrids,” IEEE Transactions on Vehicular Technology, vol. 66, no. 7, pp. 5785–5796, 2016.