DSAC: Distributional Soft Actor-Critic for Risk-Sensitive Reinforcement Learning

Main Article Content

Xiaoteng Ma
Junyao Chen
Li Xia
Jun Yang
Qianchuan Zhao
Zhengyuan Zhou

Abstract

We present Distributional Soft Actor-Critic (DSAC), a distributional reinforcement learning (RL) algorithm that combines the strengths of distributional information of accumulated rewards and entropy-driven exploration from Soft Actor-Critic (SAC) algorithm. DSAC models the randomness in both action and rewards, surpassing baseline performances on various continuous control tasks. Unlike standard approaches that solely maximize expected rewards, we propose a unified framework for risk-sensitive learning, one that optimizes the risk-related objective while balancing entropy to encourage exploration. Extensive experiments demonstrate DSAC’s effectiveness in enhancing agent performances for both risk-neutral and risk-sensitive control tasks.

Article Details

Section
Articles