A Lightweight PyBullet-Based Framework for Fast Reinforcement Learning Prototyping on 6-DOF Robotic Arms
- 1 Faculty of Basic Sciences, Van Lang University, Ho Chi Minh City, Vietnam
- 2 eWalk Co. Ltd., Ho Chi Minh City, Vietnam
- 3 Department of Physics, International University, VNU-HCM, Ho Chi Minh City, Vietnam
- 4 Viet Nam National University, Ho Chi Minh City, Vietnam
Abstract
Controlling 6-Degree-of-Freedom (6-DOF) robotic arms for precise manipulation tasks is challenging due to kinematic redundancy and the complexity of existing simulation environments like MuJoCo or ROS-Gazebo. This paper presents ArmReach6DOFEnv, a lightweight, open-source simulation framework built on PyBullet for rapid Reinforcement Learning (RL) prototyping on 6-DOF robotic arms. Using a Universal Robot Description Format (URDF) model, the environment supports a continuous state-action space for a 3D reaching task, with a reward function balancing accuracy and control effort. We evaluate two state-of-the-art RL algorithms, Proximal Policy Optimization (PPO) and Deep Deterministic Policy Gradient (DDPG), implemented via Stable-Baselines3, comparing their convergence, success rate, and motion smoothness. Experimental results demonstrate DDPG’s superior performance (69% success rate vs. PPO’s 34%) and smoother trajectories, despite PPO’s faster convergence. This framework enables accessible RL experimentation on resource-constrained systems, with potential for future sim-to-real transfer.
DOI: https://doi.org/10.3844/jmrsp.2025.35.39
Copyright: © 2025 Ngoc Kim Khanh Nguyen, Anh Thu Mang and Quang Nguyen. This is an open access article distributed under the terms of the
Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
- 32 Views
- 6 Downloads
- 0 Citations
Download
Keywords
- Reinforcement Learning
- 6-DOF Robotic Arm
- PyBullet
- PPO
- DDPG
- Rapid Prototyping