Skip to content

Commit 7a31a2b

Browse files
authored
Merge pull request dennybritz#120 from jonahweissman/master
Fix links in all the `README.md`s
2 parents 90b0a4a + 7017f9e commit 7a31a2b

File tree

7 files changed

+40
-40
lines changed

7 files changed

+40
-40
lines changed

DP/README.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -34,13 +34,13 @@
3434
### Exercises
3535

3636
- Implement Policy Evaluation in Python (Gridworld)
37-
- [Exercise](Policy Evaluation.ipynb)
38-
- [Solution](Policy Evaluation Solution.ipynb)
37+
- [Exercise](Policy%20Evaluation.ipynb)
38+
- [Solution](Policy%20Evaluation%20Solution.ipynb)
3939

4040
- Implement Policy Iteration in Python (Gridworld)
41-
- [Exercise](Policy Iteration.ipynb)
42-
- [Solution](Policy Iteration Solution.ipynb)
41+
- [Exercise](Policy%20Iteration.ipynb)
42+
- [Solution](Policy%20Iteration%20Solution.ipynb)
4343

4444
- Implement Value Iteration in Python (Gridworld)
45-
- [Exercise](Value Iteration.ipynb)
46-
- [Solution](Value Iteration Solution.ipynb)
45+
- [Exercise](Value%20Iteration.ipynb)
46+
- [Solution](Value%20Iteration%20Solution.ipynb)

DQN/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -39,11 +39,11 @@
3939

4040
### Exercises
4141

42-
- [OpenAI Gym Atari Environment Playground](Breakout Playground.ipynb)
42+
- [OpenAI Gym Atari Environment Playground](Breakout%20Playground.ipynb)
4343
- Deep-Q Learning for Atari Games
44-
- [Exercise](Deep Q Learning.ipynb)
45-
- [Solution](Deep Q Learning Solution.ipynb)
44+
- [Exercise](Deep%20Q%20Learning.ipynb)
45+
- [Solution](Deep%20Q%20Learning%20Solution.ipynb)
4646
- Double-Q Learning
4747
- This is a minimal change to Q-Learning so use the same exercise as above
48-
- [Solution](Double DQN Solution.ipynb)
48+
- [Solution](Double%20DQN%20Solution.ipynb)
4949
- Prioritized Experience Replay (WIP)

FA/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,5 +36,5 @@
3636
### Exercises
3737

3838
- Solve Mountain Car Problem using Q-Learning with Linear Function Approximation
39-
- [Exercise](Q-Learning with Value Function Approximation.ipynb)
40-
- [Solution](Q-Learning with Value Function Approximation Solution.ipynb)
39+
- [Exercise](Q-Learning%20with%20Value%20Function%20Approximation.ipynb)
40+
- [Solution](Q-Learning%20with%20Value%20Function%20Approximation%20Solution.ipynb)

MC/README.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -37,13 +37,13 @@
3737

3838
### Exercises
3939

40-
- [Get familiar with the Blackjack environment (Blackjack-v0)](Blackjack Playground.ipynb)
40+
- [Get familiar with the Blackjack environment (Blackjack-v0)](Blackjack%20Playground.ipynb)
4141
- Implement the Monte Carlo Prediction to estimate state-action values
42-
- [Exercise](MC Prediction.ipynb)
43-
- [Solution](MC Prediction Solution.ipynb)
42+
- [Exercise](MC%20Prediction.ipynb)
43+
- [Solution](MC%20Prediction%20Solution.ipynb)
4444
- Implement the on-policy first-visit Monte Carlo Control algorithm
45-
- [Exercise](MC Control with Epsilon-Greedy Policies.ipynb)
46-
- [Solution](MC Control with Epsilon-Greedy Policies Solution.ipynb)
45+
- [Exercise](MC%20Control%20with%20Epsilon-Greedy%20Policies.ipynb)
46+
- [Solution](MC%20Control%20with%20Epsilon-Greedy%20Policies%20Solution.ipynb)
4747
- Implement the off-policy every-visit Monte Carlo Control using Weighted Important Sampling algorithm
48-
- [Exercise](Off-Policy MC Control with Weighted Importance Sampling.ipynb)
49-
- [Solution](Off-Policy MC Control with Weighted Importance Sampling Solution.ipynb)
48+
- [Exercise](Off-Policy%20MC%20Control%20with%20Weighted%20Importance%20Sampling.ipynb)
49+
- [Solution](Off-Policy%20MC%20Control%20with%20Weighted%20Importance%20Sampling%20Solution.ipynb)

PolicyGradient/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -50,13 +50,13 @@
5050

5151
- REINFORCE with Baseline
5252
- Exercise
53-
- [Solution](CliffWalk REINFORCE with Baseline Solution.ipynb)
53+
- [Solution](CliffWalk%20REINFORCE%20with%20Baseline%20Solution.ipynb)
5454
- Actor-Critic with Baseline
5555
- Exercise
56-
- [Solution](CliffWalk Actor-Critic Solution.ipynb)
56+
- [Solution](CliffWalk%20Actor-Critic%20Solution.ipynb)
5757
- Actor-Critic with Baseline for Continuous Action Spaces
5858
- Exercise
59-
- [Solution](Continuous MountainCar Actor-Critic Solution.ipynb)
59+
- [Solution](Continuous%20MountainCar%20Actor-Critic%20Solution.ipynb)
6060
- Deterministic Policy Gradients for Continuous Action Spaces (WIP)
6161
- Deep Deterministic Policy Gradients (WIP)
6262
- Asynchronous Advantage Actor-Critic (A3C)

README.md

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -26,21 +26,21 @@ All code is written in Python 3 and uses RL environments from [OpenAI Gym](https
2626

2727
### List of Implemented Algorithms
2828

29-
- [Dynamic Programming Policy Evaluation](DP/Policy Evaluation Solution.ipynb)
30-
- [Dynamic Programming Policy Iteration](DP/Policy Iteration Solution.ipynb)
31-
- [Dynamic Programming Value Iteration](DP/Value Iteration Solution.ipynb)
32-
- [Monte Carlo Prediction](MC/MC Prediction Solution.ipynb)
33-
- [Monte Carlo Control with Epsilon-Greedy Policies](MC/MC Control with Epsilon-Greedy Policies Solution.ipynb)
34-
- [Monte Carlo Off-Policy Control with Importance Sampling](MC/Off-Policy MC Control with Weighted Importance Sampling Solution.ipynb)
35-
- [SARSA (On Policy TD Learning)](TD/SARSA Solution.ipynb)
36-
- [Q-Learning (Off Policy TD Learning)](TD/Q-Learning Solution.ipynb)
37-
- [Q-Learning with Linear Function Approximation](FA/Q-Learning with Value Function Approximation Solution.ipynb)
38-
- [Deep Q-Learning for Atari Games](DQN/Deep Q Learning Solution.ipynb)
39-
- [Double Deep-Q Learning for Atari Games](DQN/Double DQN Solution.ipynb)
29+
- [Dynamic Programming Policy Evaluation](DP/Policy%20Evaluation%20Solution.ipynb)
30+
- [Dynamic Programming Policy Iteration](DP/Policy%20Iteration%20Solution.ipynb)
31+
- [Dynamic Programming Value Iteration](DP/Value%20Iteration%20Solution.ipynb)
32+
- [Monte Carlo Prediction](MC/MC%20Prediction%20Solution.ipynb)
33+
- [Monte Carlo Control with Epsilon-Greedy Policies](MC/MC%20Control%20with%20Epsilon-Greedy%20Policies%20Solution.ipynb)
34+
- [Monte Carlo Off-Policy Control with Importance Sampling](MC/Off-Policy%20MC%20Control%20with%20Weighted%20Importance%20Sampling%20Solution.ipynb)
35+
- [SARSA (On Policy TD Learning)](TD/SARSA%20Solution.ipynb)
36+
- [Q-Learning (Off Policy TD Learning)](TD/Q-Learning%20Solution.ipynb)
37+
- [Q-Learning with Linear Function Approximation](FA/Q-Learning%20with%20Value%20Function%20Approximation%20Solution.ipynb)
38+
- [Deep Q-Learning for Atari Games](DQN/Deep%20Q%20Learning%20Solution.ipynb)
39+
- [Double Deep-Q Learning for Atari Games](DQN/Double%20DQN%20Solution.ipynb)
4040
- Deep Q-Learning with Prioritized Experience Replay (WIP)
41-
- [Policy Gradient: REINFORCE with Baseline](PolicyGradient/CliffWalk REINFORCE with Baseline Solution.ipynb)
42-
- [Policy Gradient: Actor Critic with Baseline](PolicyGradient/CliffWalk Actor Critic Solution.ipynb)
43-
- [Policy Gradient: Actor Critic with Baseline for Continuous Action Spaces](PolicyGradient/Continuous MountainCar Actor Critic Solution.ipynb)
41+
- [Policy Gradient: REINFORCE with Baseline](PolicyGradient/CliffWalk%20REINFORCE%20with%20Baseline%20Solution.ipynb)
42+
- [Policy Gradient: Actor Critic with Baseline](PolicyGradient/CliffWalk%20Actor%20Critic%20Solution.ipynb)
43+
- [Policy Gradient: Actor Critic with Baseline for Continuous Action Spaces](PolicyGradient/Continuous%20MountainCar%20Actor%20Critic%20Solution.ipynb)
4444
- Deterministic Policy Gradients for Continuous Action Spaces (WIP)
4545
- Deep Deterministic Policy Gradients (DDPG) (WIP)
4646
- [Asynchronous Advantage Actor Critic (A3C)](PolicyGradient/a3c)

TD/README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -40,11 +40,11 @@
4040

4141
### Exercises
4242

43-
- [Windy Gridworld Playground](Windy Gridworld Playground.ipynb)
43+
- [Windy Gridworld Playground](Windy%20Gridworld%20Playground.ipynb)
4444
- Implement SARSA
4545
- [Exercise](SARSA.ipynb)
46-
- [Solution](SARSA Solution.ipynb)
47-
- [Cliff Environment Playground](Cliff Environment Playground.ipynb)
46+
- [Solution](SARSA%20Solution.ipynb)
47+
- [Cliff Environment Playground](Cliff%20Environment%20Playground.ipynb)
4848
- Implement Q-Learning in Python
4949
- [Exercise](Q-Learning.ipynb)
50-
- [Solution](Q-Learning Solution.ipynb)
50+
- [Solution](Q-Learning%20Solution.ipynb)

0 commit comments

Comments
 (0)