Deep Q-Learning Aircraft Landing Control

Machine Learning Deep Q-Learning Python Gymnasium PyTorch

This project explores how different hyperparameters affect deep q-learning performance through systematic experimentation. Using a custom gymnasium environment simulating aircraft descent, I investigated the complex relationships between batch size, learning rate, discount factor, and exploration strategies.

The neural network architecture consisted of an input layer, 3 hidden layers (256 neurons each), and an output layer, utilizing ReLu activation functions. This setup enabled the agent to learn control policies for managing descent rate, angle, and positioning while maintaining stability throughout the landing sequence.

Read the Full Study

Key Achievements

  • Custom Environment: Developed a gymnasium environment for aircraft descent simulation with physics-based reward shaping
  • Parameter Optimization: Conducted comprehensive studies across batch sizes, learning rates, discount factors, and exploration rates
  • Performance Improvement: Achieved 51% above average performance (397.9 vs 263.5) through optimal parameter selection
  • Stability Analysis: Utilized surface plots and statistical analysis to evaluate parameter influence on model stability

Key Learnings

  • Deep Q-Learning Implementation: Gained hands-on experience in implementing and optimizing deep q-learning algorithms
  • Reward Engineering: Mastered the art of reward shaping for complex control tasks, learning through multiple iterations
  • Parameter Sensitivity: Understood the delicate balance between learning stability and performance through hyperparameter tuning
  • Data Analysis: Developed skills in analyzing experimental results and creating meaningful visualizations