DRL-robot-navigation and DRL-Robot-Navigation-ROS2

DRL-robot-navigation
54
Established
Maintenance 6/25
Adoption 10/25
Maturity 16/25
Community 22/25
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 15/25
Stars: 1,262
Forks: 189
Downloads:
Commits (30d): 0
Language: Python
License: MIT
Stars: 194
Forks: 22
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No Package No Dependents
Stale 6m No Package No Dependents

About DRL-robot-navigation

reiniscimurs/DRL-robot-navigation

Deep Reinforcement Learning for mobile robot navigation in ROS Gazebo simulator. Using Twin Delayed Deep Deterministic Policy Gradient (TD3) neural network, a robot learns to navigate to a random goal point in a simulated environment while avoiding obstacles.

This project enables a simulated mobile robot to learn how to navigate to a target destination while avoiding obstacles. It takes laser sensor readings and a goal in polar coordinates as input, and outputs trained navigation policies. This is useful for robotics researchers and engineers who are developing autonomous navigation systems.

robot-navigation autonomous-systems robotics-research simulation-training mobile-robotics

About DRL-Robot-Navigation-ROS2

reiniscimurs/DRL-Robot-Navigation-ROS2

Deep Reinforcement Learning for mobile robot navigation in ROS2 Gazebo simulator. Using DRL (SAC, TD3) neural networks, a robot learns to navigate to a random goal point in a simulated environment while avoiding obstacles.

This project helps robotics engineers and researchers train mobile robots to navigate safely and efficiently in simulated environments. It takes sensor data (like laser readings for obstacles and goal coordinates) as input and outputs a trained deep reinforcement learning model. This model enables a robot to reach a specified goal while autonomously avoiding collisions with obstacles.

mobile-robotics robot-navigation robot-simulation autonomous-systems robot-training

Scores updated daily from GitHub, PyPI, and npm data. How scores work