@article {10.3844/jmrsp.2025.35.39, article_type = {journal}, title = {A Lightweight PyBullet-Based Framework for Fast Reinforcement Learning Prototyping on 6-DOF Robotic Arms}, author = {Nguyen, Ngoc Kim Khanh and Mang, Anh Thu and Nguyen, Quang}, volume = {9}, year = {2026}, month = {Feb}, pages = {35-39}, doi = {10.3844/jmrsp.2025.35.39}, url = {https://thescipub.com/abstract/jmrsp.2025.35.39}, abstract = {Controlling 6-Degree-of-Freedom (6-DOF) robotic arms for precise manipulation tasks is challenging due to kinematic redundancy and the complexity of existing simulation environments like MuJoCo or ROS-Gazebo. This paper presents ArmReach6DOFEnv, a lightweight, open-source simulation framework built on PyBullet for rapid Reinforcement Learning (RL) prototyping on 6-DOF robotic arms. Using a Universal Robot Description Format (URDF) model, the environment supports a continuous state-action space for a 3D reaching task, with a reward function balancing accuracy and control effort. We evaluate two state-of-the-art RL algorithms, Proximal Policy Optimization (PPO) and Deep Deterministic Policy Gradient (DDPG), implemented via Stable-Baselines3, comparing their convergence, success rate, and motion smoothness. Experimental results demonstrate DDPG’s superior performance (69% success rate vs. PPO’s 34%) and smoother trajectories, despite PPO’s faster convergence. This framework enables accessible RL experimentation on resource-constrained systems, with potential for future sim-to-real transfer.}, journal = {Journal of Mechatronics and Robotics}, publisher = {Science Publications} }