Compound Asynchronous Exploration and Exploitation
Download as PDF
Jie Bai, Li Liu, Yaobing Wang, Haoyu Zhang, Jianfei Li
Data efficiency has always been a significant key topic for deep reinforcement learning. The main progress has been on sufficient exploration and effective exploitation. However, the two are often discussed separately. Profit from distributed systems, we propose an asynchronous approach to deep reinforcement learning by combining exploration and exploitation. We apply our framework to off-the-shelf deep reinforcement learning algorithms, and experimental results show that our algorithm is superior in final performance and efficiency.
Deep Reinforcement Learning, Exploration And Exploitation, Asynchronous Methods