1- # Project Navigation - Udacity Deep Reinforcement Learning
21[ // ] : # ( Image References )
32
4- [ video_random ] : https://github.com/dalmia/udacity-deep-reinforcement-learning /blob/master/2%20-%20Value-based%20methods/Project-Navigation/results /random_agent.gif " Random Agent "
3+ [ video_random ] : https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning /blob/master/Training-Results /random_agent.gif " Random Agent "
54
6- [ video_trained ] : https://github.com/dalmia/udacity-deep-reinforcement-learning /blob/master/2%20-%20Value-based%20methods/Project-Navigation/results /trained_agent.gif " Trained Agent "
5+ [ video_trained ] : https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning /blob/master/Training-Results /trained_agent.gif " Trained Agent "
76
8- # Project 1: Navigation
7+ # Project 1: Navigation - Udacity Deep Reinforcement Learning
98
109### Introduction
1110
@@ -53,9 +52,7 @@ env = env = UnityEnvironment(file_name="Banana.app")
5352- ` dqn.pth ` : saved model weights for the original DQN model
5453- ` ddqn.pth ` : saved model weights for the Double DQN model
5554- ` ddqn.pth ` : saved model weights for the Dueling Double DQN model
56- - ` Navigation_exploration.ipynb ` : explore the unity environment
57- - ` Navigation_solution.ipynb ` : notebook containing the solution
58- - ` Navigation_Pixels.ipynb ` : notebook containing the code for the pixel-action problem (see below)
55+ - ` Navigation.ipynb ` : notebook containing the solution
5956
6057### Instructions
6158
@@ -71,16 +68,16 @@ To watch a trained smart agent, follow the instructions below:
7168Several enhancements to the original DQN algorithm have also been incorporated:
7269
7370- Double DQN [[ Paper] ( https://arxiv.org/abs/1509.06461 )] [[ Code] ( https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/dqn_agent.py#L94 )]
74- - Prioritized Experience Replay [[ Paper] ( https://arxiv.org/abs/1511.05952 )] [[ Code] ( )] (WIP )
71+ - Prioritized Experience Replay [[ Paper] ( https://arxiv.org/abs/1511.05952 )] [[ Code] ( )] (To be worked out )
7572- Dueling DQN [[ Paper] ( https://arxiv.org/abs/1511.06581 )] [[ Code] ( https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/model.py )]
7673
7774### Results
7875
79- Plot showing the score per episode over all the episodes. The environment was solved in ** 377 ** episodes (currently).
76+ Plot showing the score per episode over all the episodes. The environment was solved in ** 361 ** episodes (currently).
8077
8178| Double DQN | DQN | Dueling DQN |
8279:-------------------------:|:-------------------------:|:-------------------------:
83- ![ double-dqn-scores] ( https://github.com/dalmia/udacity-deep-reinforcement-learning /blob/master/2%20-%20Value-based%20methods/Project-Navigation/results/ ddqn_new_scores.png ) | ![ dqn-scores] ( https://github.com/dalmia/udacity-deep-reinforcement-learning /blob/master/2%20-%20Value-based%20methods/Project-Navigation/results/ dqn_new_scores.png ) | ![ dueling-double-dqn-scores] ( https://github.com/dalmia/udacity-deep-reinforcement-learning /blob/master/2%20-%20Value-based%20methods/Project-Navigation/results /dddqn_new_scores.png )
80+ ![ double-dqn-scores] ( https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning /blob/master/Training-Results/ ddqn_new_scores.png ) | ![ dqn-scores] ( https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning /blob/master/Training-Results/ dqn_new_scores.png ) | ![ dueling-double-dqn-scores] ( https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning /blob/master/Training-Results /dddqn_new_scores.png )
8481
8582
8683### Challenge: Learning from Pixels
0 commit comments