🤖
Robotics Handbook
HomeConnect
  • Welcome
    • Authors Note
  • Computer Aided Designs and Simulations
    • Computer Aided Design and Simulations
    • 3D Modelling CAD
      • SolidWorks
    • Simulations
    • PCB Design
  • ROS (Advanced)
    • ROS
    • ROS
      • Concepts and Packages
      • Manual and Quick Setup
    • Some Important packages
  • Hardware
    • Design Processes
      • Materials Selection
      • Build and Prototyping
      • 3D Printing and Machining
    • Fabrication Parts
  • Common Mechanisms
    • Wheels and Drives
    • Power Transmission
  • Career Paths & Research Opportunities
    • Career in Robotics
    • Job Roles In Robotics
    • Conferences and Journals
  • Companies Hiring for Robotics
  • Leading Institutes
  • Mathematical and Programming Foundations
    • Linear Algebra for Robotics
    • Calculus
  • Programming for Robotics
    • Common Languages
    • Algorithms
    • Digital Twin
  • Embedded Systems for Robotics
    • Embedded Systems
    • Microcontrollers
      • Microcontrollers (Advanced Theory)
      • Choosing a Microcontroller
    • Sensors and Actuators
      • Sensors for Robotics
      • Actuators for Robotics
    • Communication
      • Communication Protocols
    • RTOS
    • Power Systems
      • Battery Charging and Storage Best Practices
  • ML and Perception
    • ML and Perception
    • Reinforcement Learning
    • Cameras, Depth Sensors and LiDAR
    • Image Processing Basics (OpenCV)
    • Object Detection and Tracking
    • Example of a Vision Pipeline
  • Mobile Robotics
    • Mobile Robotics
    • SLAM and Navigation
    • Robot Kinematics and Dynamics
      • Some Kinematic Models
    • Trajectory Planning
    • AMR's and AGV's
    • MH633 : Mobile Robotics
      • Geometric Foundations
      • Kinematics
  • Frontiers and Emerging Fields
    • Frontiers and Emerging Fields
    • Humanoids
    • Autonomous Navigation
    • Bio-inspired and Soft Robotics
    • Space Robotics
    • Cobots
    • Edge Robotics
    • Medical Robotics
  • Drones, Rocketry and Aviation
    • Drones
      • Drone Anatomy
    • Rocketry
Powered by GitBook
On this page
  • Core RL Algorithms and Resources
  • Robotics Applications
  • Software Frameworks & Toolkits
  • Online Courses & Tutorials
  • Key Survey Papers
  1. ML and Perception

Reinforcement Learning

PreviousML and PerceptionNextCameras, Depth Sensors and LiDAR

Last updated 1 day ago

Reinforcement Learning (RL) endows robots with the ability to learn control policies through trial-and-error interactions rather than hand-coding behaviors. This page surveys core RL approaches, their robotic applications, and a curated set of learning resources and software tools.

Core RL Algorithms and Resources

  • Value-Based Methods

    • Q-Learning & SARSA – Tabular methods for discrete state–action spaces

    • Deep Q-Networks (DQN) & Variants (Double DQN, Dueling DQN) – Neural-network approximators for high-dimensional inputs

  • Policy-Gradient Methods

    • REINFORCE – Monte-Carlo policy search

    • Trust Region Policy Optimization (TRPO) and Proximal Policy Optimization (PPO) – Stable on-policy updates

    • Actor-Critic (A2C, A3C) – Combines policy gradient with value estimates

  • Continuous-Control Algorithms

    • Deep Deterministic Policy Gradient (DDPG) & Twin Delayed DDPG (TD3) – Off-policy actor-critic for continuous actions

    • Soft Actor-Critic (SAC) – Maximum-entropy RL for robustness

  • Model-Based and Hybrid Methods

    • Model-Based Policy Optimization (MBPO) – Leverages learned dynamics models

    • Guided Policy Search – Uses trajectory optimization to supervise policy learning

  • Multi-Agent and Hierarchical RL

    • Multi-Agent Deep Q-Learning (MADDPG) – Cooperative and competitive settings

    • Hierarchical RL (options framework) – Temporal abstractions for long-horizon tasks

Robotics Applications

  • Locomotion & Legged Control

    • Learning stable walking, running gaits on quadrupeds and bipeds

  • Manipulation & Grasping

    • End-to-end policies for pick-and-place, tool use, and dexterous in-hand manipulation

  • Navigation & Mobile Robotics

    • Maze solving, obstacle avoidance, and mapless navigation with deep RL

  • Sim-to-Real Transfer

  • Aerial Robotics

    • Autonomous flight control for drones via RL

Software Frameworks & Toolkits

Online Courses & Tutorials

Key Survey Papers

By blending these algorithms, platforms, and learning pathways, practitioners can accelerate the deployment of RL-powered robots-from simulated prototypes to real-world autonomy.

Sutton & Barto’s “Reinforcement Learning: An Introduction” ()

OpenAI Baselines DQN implementation ()

OpenAI Spinning Up tutorials ()

Stable Baselines3 implementations ()

Survey: “Reinforcement Learning in Robotic Applications” ()

NVIDIA’s Legged Gym environments ()

Dex-Net grasp planner with RL integration ()

ROS-Gazebo RL tutorials ()

Domain Randomization and Sim-to-Real pipelines in NVIDIA Isaac Sim ()

Microsoft AirSim environments ()

OpenAI Gym & Gym-Robotics ()

ROS RL Packages & ROS-Gym Bridges ()

NVIDIA Isaac RL & Isaac Gym ()

Ray RLlib: Scalable RL library ()

Unity ML-Agents: Game-engine–based RL ()

Intel Coach: Research RL framework ()

Coursera “Reinforcement Learning Specialization” by University of Alberta ()

Udacity “Deep Reinforcement Learning Nanodegree” ()

The Construct Academy “Reinforcement Learning for Robotics” ()

30 Days Coding “RL for Robotics: Locomotion & Navigation” ()

Kober, Bagnell & Peters (2013), “Reinforcement Learning in Robotics: A Survey” ()

Deisenroth, Neumann & Peters (2011), “A Survey on Policy Search for Robotics” ()

Singh et al. (2021), “Reinforcement Learning in Robotic Applications: A Comprehensive Survey” ()

http://incompleteideas.net/book/the-book.html
https://github.com/openai/baselines
https://spinningup.openai.com
https://github.com/DLR-RM/stable-baselines3
https://doi.org/10.1007/s10462-021-09997-9
https://developer.nvidia.com/isaac-legged-gym
https://berkeleyautomation.github.io/dex-net
http://wiki.ros.org/gym_gazebo
https://developer.nvidia.com/isaac-sim
https://github.com/microsoft/AirSim
https://gym.openai.com/envs/#robotics
https://github.com/ros-gym
https://developer.nvidia.com/isaac-gym
https://docs.ray.io/en/latest/rllib.html
https://github.com/Unity-Technologies/ml-agents
https://github.com/intel/coach
https://www.coursera.org/specializations/reinforcement-learning
https://www.udacity.com/course/deep-reinforcement-learning-nanodegree--nd893
https://www.theconstruct.ai/robotigniteacademy_learnros/ros-courses-library/reinforcement-learning-for-robotics/
https://30dayscoding.com/blog/reinforcement-learning-for-robotics-locomotion-and-navigation
https://www.ias.informatik.tu-darmstadt.de/uploads/Publications/Kober_IJRR_2013.pdf
https://spiral.imperial.ac.uk/bitstream/10044/1/12051/7/fnt_corrected_2014-8-22.pdf
https://doi.org/10.1007/s10462-021-09997-9