Hosseini, Mohammad MehdiParvania, Masood2021-12-242021-12-242022-01-04978-0-9981331-5-7http://hdl.handle.net/10125/79761Capabilities of deep reinforcement learning (DRL) in obtaining fast decision policies in high dimensional and stochastic environments have led to its extensive use in operational research, including the operation of distribution grids with high penetration of distributed energy resources (DER). However, the feasibility and robustness of DRL solutions are not guaranteed for the system operator, and hence, those solutions may be of limited practical value. This paper proposes an analytical method to find feasibility ellipsoids that represent the range of multi-dimensional system states in which the DRL solution is guaranteed to be feasible. Empirical studies and stochastic sampling determine the ratio of the discovered to the actual feasible space as a function of the sample size. In addition, the performance of logarithmic, linear, and exponential penalization of infeasibility during the DRL training are studied and compared in order to reduce the number of infeasible solutions.9 pagesengAttribution-NonCommercial-NoDerivatives 4.0 InternationalMonitoring, Control, and Protectiondeep reinforcement learningdistributed energy resourcespower distribution systemsolution verificationOn the Verification of Deep Reinforcement Learning Solution for Intelligent Operation of Distribution Gridstext10.24251/HICSS.2022.426