Skip to content

Commit 34a8d38

Browse files
authored
Update README.md
1 parent f3099df commit 34a8d38

File tree

1 file changed

+7
-10
lines changed

1 file changed

+7
-10
lines changed

README.md

Lines changed: 7 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,10 @@
1-
# Project Navigation - Udacity Deep Reinforcement Learning
21
[//]: # (Image References)
32

4-
[video_random]: https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/results/random_agent.gif "Random Agent"
3+
[video_random]: https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning/blob/master/Training-Results/random_agent.gif "Random Agent"
54

6-
[video_trained]: https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/results/trained_agent.gif "Trained Agent"
5+
[video_trained]: https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning/blob/master/Training-Results/trained_agent.gif "Trained Agent"
76

8-
# Project 1: Navigation
7+
# Project 1: Navigation - Udacity Deep Reinforcement Learning
98

109
### Introduction
1110

@@ -53,9 +52,7 @@ env = env = UnityEnvironment(file_name="Banana.app")
5352
- `dqn.pth`: saved model weights for the original DQN model
5453
- `ddqn.pth`: saved model weights for the Double DQN model
5554
- `ddqn.pth`: saved model weights for the Dueling Double DQN model
56-
- `Navigation_exploration.ipynb`: explore the unity environment
57-
- `Navigation_solution.ipynb`: notebook containing the solution
58-
- `Navigation_Pixels.ipynb`: notebook containing the code for the pixel-action problem (see below)
55+
- `Navigation.ipynb`: notebook containing the solution
5956

6057
### Instructions
6158

@@ -71,16 +68,16 @@ To watch a trained smart agent, follow the instructions below:
7168
Several enhancements to the original DQN algorithm have also been incorporated:
7269

7370
- Double DQN [[Paper](https://arxiv.org/abs/1509.06461)] [[Code](https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/dqn_agent.py#L94)]
74-
- Prioritized Experience Replay [[Paper](https://arxiv.org/abs/1511.05952)] [[Code]()] (WIP)
71+
- Prioritized Experience Replay [[Paper](https://arxiv.org/abs/1511.05952)] [[Code]()] (To be worked out)
7572
- Dueling DQN [[Paper](https://arxiv.org/abs/1511.06581)] [[Code](https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/model.py)]
7673

7774
### Results
7875

79-
Plot showing the score per episode over all the episodes. The environment was solved in **377** episodes (currently).
76+
Plot showing the score per episode over all the episodes. The environment was solved in **361** episodes (currently).
8077

8178
| Double DQN | DQN | Dueling DQN |
8279
:-------------------------:|:-------------------------:|:-------------------------:
83-
![double-dqn-scores](https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/results/ddqn_new_scores.png) | ![dqn-scores](https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/results/dqn_new_scores.png) | ![dueling-double-dqn-scores](https://github.com/dalmia/udacity-deep-reinforcement-learning/blob/master/2%20-%20Value-based%20methods/Project-Navigation/results/dddqn_new_scores.png)
80+
![double-dqn-scores](https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning/blob/master/Training-Results/ddqn_new_scores.png) | ![dqn-scores](https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning/blob/master/Training-Results/dqn_new_scores.png) | ![dueling-double-dqn-scores](https://github.com/vivekthota16/Project-Navigation-Udacity-Deep-Reinforcement-Learning/blob/master/Training-Results/dddqn_new_scores.png)
8481

8582

8683
### Challenge: Learning from Pixels

0 commit comments

Comments
 (0)