2023
DOI: 10.3390/app13042470
|View full text |Cite
|
Sign up to set email alerts
|

Label-Free Fault Detection Scheme for Inverters of PV Systems: Deep Reinforcement Learning-Based Dynamic Threshold

Abstract: Generally, photovoltaic (PV) fault detection approaches can be divided into two groups: end-to-end and threshold methods. The end-to-end method typically uses a deep neural network (DNN) to learn fault patterns from labeled datasets, which directly detect whether faults occur or not. The threshold method first estimates power generation and uses thresholds to detect atypical deviations of measured values from estimated ones. The former method heavily relies on fault-labeled data and, therefore, requires the co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 53 publications
0
0
0
Order By: Relevance
“…Conversely, DDPG outshines DQN in stability, attributed to target networks and a deterministic policy, enhancing the learning process. 31 DDPG excels in scenarios with continuous action spaces, demonstrating better generalization from finite experiences and offering superior convergence properties compared with Q-learning-based methods.…”
Section: Basics Of Drlmentioning
confidence: 99%
See 1 more Smart Citation
“…Conversely, DDPG outshines DQN in stability, attributed to target networks and a deterministic policy, enhancing the learning process. 31 DDPG excels in scenarios with continuous action spaces, demonstrating better generalization from finite experiences and offering superior convergence properties compared with Q-learning-based methods.…”
Section: Basics Of Drlmentioning
confidence: 99%
“…Despite its effectiveness, DQN demands significant computational resources and time, especially in environments with high‐dimensional state spaces, leading to potential instability during training. Conversely, DDPG outshines DQN in stability, attributed to target networks and a deterministic policy, enhancing the learning process 31 . DDPG excels in scenarios with continuous action spaces, demonstrating better generalization from finite experiences and offering superior convergence properties compared with Q ‐learning‐based methods.…”
Section: System Configuration and Basics Of Drlmentioning
confidence: 99%