Voltage Control based on ISS Neural Certificates

We Developed methods for stabilizing large scale power systems based on Input-to-State Stability (ISS) Lyapunov-based neural certificate, by treating a large system as an interconnection of smaller subsystems. Each ISS Lyapunov function of subsystem could be collected to prove the global stability of power system.

We demonstrate NeuralISS using three examples - Power systems, Platoon, and Drone formation control, and show that NeuralISS can find certifiably stable controllers for networks of size up to 100 subsystems. Compared with centralized neural certificate approaches, NeuralISS reaches similar results in small-scale systems, and can generalize to large scale systems that centralized approaches cannot scale up to. Compared with LQR, NeuralISS can deal with strong coupled networked systems like the microgrids, and reaches smaller tracking errors on both small and large scale systems. Compared with RL (PPO, LYPPO, MAPPO), our algorithm achieves similar or smaller tracking errors in small systems, and can hugely reduce the tracking errors in large systems (up to 75%).

In our experiments, we consider a power distribution system that consists of 8 buses. The buses are arranged in a line, each one is connected to a static generator. The nominal controller is similar to the droop control and is a proportion controller on the voltage deviation, u_i(t) = c_i(v_i(t)-1) (where c_i is a constant). This proportion controller is a standard controller used in practice.

See the related Paper if you are interested.

The example of 8 bus GridVoltage.
To control the bus voltage to 1kv. Left is the performance of Neural Controller. Middle, we take a simple practical Propotional controller for reference. Right is the figure of lyapunov values.
We compare NeurISS with both centralized and decentralized baselines. For centralized ones, we compare with the state-of-the-art RL algorithm PPO, the RL-with-Lyapunov-critic algorithm LYPPO, and the centralized Neural CLF controller (NCLF). For decentralized ones, we compare with the classical LQR controller and the multi-agent RL algorithm MAPPO. We hand-craft reward functions based on the common way of designing reward functions of tracking problems for the RL algorithms. For LQR, since the agents only have local observation, we calculate the goal point for the LQR controller based on local observations in each time step.
Experiment Results of NeuralISS