Comparison of PPO-DRL and A2C-DRL Algorithms for MPPT in Photovoltaic Systems via Buck-Boost Converter

Wiwat Jeungthanasirigool, Thanyaphob Sirimaskasem, Terapong Boonraksa, Promphak Boonraksa

Abstract

This research investigates the effectiveness of two deep reinforcement learning algorithms, Proximal Policy Optimization (PPO) and Advantage Actor-Critic (A2C), in achieving the MPPT for PV systems implemented via a Buck-Boost converter. The algorithms were trained and evaluated under varying environmental conditions, including different levels of irradiance and temperature. The results are presented through duty cycle heatmaps, power output heatmaps, and performance curves for power, voltage, and current. The PPO algorithm demonstrated stable and consistent control across all scenarios, maintaining a nearly constant duty cycle and achieving high power output. In contrast, A2C exhibited more adaptive control behavior, adjusting the duty cycle based on environmental changes, but showed lower power output under weak irradiance. Overall, PPO outperformed A2C in terms of stability, accuracy, and ability to reach the optimal operating point, making it a more suitable choice for MPPT applications in PV systems under dynamic conditions.

Authors

Wiwat Jeungthanasirigool
Thanyaphob Sirimaskasem
Terapong Boonraksa
Promphak Boonraksa
promphak.b@rmutsb.ac.th (Primary Contact)
Jeungthanasirigool, W. ., Sirimaskasem, T. ., Boonraksa, T. ., & Boonraksa, P. . (2025). Comparison of PPO-DRL and A2C-DRL Algorithms for MPPT in Photovoltaic Systems via Buck-Boost Converter. International Journal of Innovative Research and Scientific Studies, 8(3), 2438–2453. https://doi.org/10.53894/ijirss.v8i3.7022

Article Details

Most read articles by the same author(s)

Similar Articles

You may also start an advanced similarity search for this article.

No Related Submission Found