Smart Microgrids Can Restore Power More Efficiently and Reliably in an Outage

“Essentially, we want to bring the power generation closer to the demand side in order to get rid of the long transmission lines,” Zhang said. “This can improve the power quality and reduce the power losses over the lines. In this way, we will make the grid smaller, but stronger and more resilient.”

To optimally operate microgrids, Zhang’s lab developed an AI-based technique called deep reinforcement learning, the same concept that underpins large language models, to create an efficient framework  that includes models of many components of the power system. Reinforcement learning depends on rewarding the algorithm for successfully responding to the changing environment  —  so an agent  is rewarded when it is able to successfully restore the demanded power of all components of the network. They explicitly model the practical constraints of the real-world system, such as the branch flows that power lines can handle.

“We’re modeling a whole bunch of things — solar, wind, small generators, batteries, and we’re also modeling when people’s electricity demand changes,” Bose said. “The novelty is that this specific flavor of reinforcement learning, which we call constrained policy optimization (CPO), is being used for the first time.”

Their CPO approach takes into account real-time conditions and uses machine learning to find long-term patterns that affect the output of renewables, such as the varying demand on the grid at a given time and intermittent weather factors that affect renewable sources. This is unlike traditional systems which often use a technique called model predictive control (MPC) that bases decisions simply on the available conditions at the time of optimization. 

For example, if the CPO method predicts that the sun will shine brightly in an hour, it would use up its supply of solar energy with the knowledge that it will later be replenished — a different strategy than it might take if the day was cloudy. It can also learn about the system based on long-term patterns of how the grid uses solar.

The researchers found that their CPO technique significantly outperforms traditional MPC methods when the forecasts of renewable sources are lower than the reality because of its better understanding of all the possible solar profiles throughout any given day. 

They also found that the reinforcement learning controller is able respond much faster than traditional optimization methods in the moment of a power outage. 

The research team recently proved the success of their method when they placed first in a global competition that invited participants to use reinforcement learning or similar techniques to operate a power grid. The competition, called L2RPN Delft 2023, was co-sponsored by France’s electricity transmission system operator (Réseau de Transport d’Électricité), which the UC Santa Cruz researchers see as an indicator that now large-scale grid operators may start moving toward AI and renewable energy techniques.

Now that they’ve developed a successful algorithm in simulations, the research team is working to test their model on microgrids in their lab. In the long-term, the researchers hope to implement their solution on the UC Santa Cruz campus’s energy system to address outage issues that the residential campus community faces. They also hope to see further interest and collaboration from industry.

Emily Cerf is engineering writer and media relations officer at the Baskin School of Engineering, University of California-Santa Cruz. The article was originally posted to the website of UC Santa Cruz.