Loading…

Loading grant details…

Active STANDARD GRANT National Science Foundation (US)

Optimizing Power Distribution Grids on a Data and Computational Budget

$4.1M USD

Funder National Science Foundation (US)
Recipient Organization Purdue University
Country United States
Start Date Apr 15, 2025
End Date Mar 31, 2028
Duration 1,081 days
Number of Grantees 2
Roles Principal Investigator; Co-Principal Investigator
Data Source National Science Foundation (US)
Grant ID 2434502
Grant Description

This project aims to develop novel algorithmic solutions to advance the optimal scheduling of distributed energy resources (DERs) in power distribution grids. The project will bring potentially transformative changes in real-time DER orchestration as the proposed solutions waive low-observability challenges in distribution grids and unlock the full potential of DERs controlled by machine learning (ML) models.

These two goals are achieved by identifying the most critical data streams for orchestrating DERs in real time and by integrating ML models into grid scheduling. Regarding intellectual merits, this project: i) introduces feature selection into grid optimization, ii) leverages contemporary differentiation tools, iii) develops algorithms for grid optimization using generalized load models; and iv) draws parallels between grid optimization and empirical risk minimization.

Regarding broader impacts, this project: i) explores converging ideas at the intersection of power systems, optimization, and ML, which could bring potentially significant technological innovation in the power industry ecosystem; ii) advances knowledge in sensitivity analysis, feature selection for optimization and control, and resilience against data attacks to optimization input parameters; and iii) reaches out to undergraduate students through hands-on activities, seminars, and discussions on career opportunities in STEM.

Optimally scheduling distributed energy resources (DERs) in distribution grids entails communicating with thousands of customers in near real-time to read load demands and pass them as input data to the optimal power flow (OPF) problem. At the same time, DERs increasingly deviate from traditional constant-power models as their injections are specified by data-driven control rules or policies, which often give rise to dynamics by interacting with the power grid.

In this context, this project aspires to develop OPF formulations that are data-frugal and seamlessly accommodate grid-adaptive power injection models for DERs. The particular objectives of the projects are to: 1) Reduce data communication for a grid operator to orchestrate DERs, by strategically identifying subsets of customer or grid meters that are most influential in making effective OPF decisions; 2) Use machine learning tools to transform distilled data so that when fed into the OPF, they yield near-optimal near-feasible decisions; 3) Expedite the calculation of sensitivities (partial derivatives) for the inverse power flow (PF) mapping and equilibrium states under DER-induced grid dynamics; and 4) Devise novel algorithms for solving the OPF under generalized power injection models, when convex relaxations are deemed inadequate.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

All Grantees

Purdue University

Advertisement
Apply for grants with GrantFunds
Advertisement
Browse Grants on GrantFunds
Interested in applying for this grant?

Complete our application form to express your interest and we'll guide you through the process.

Apply for This Grant