Skip to content

Commit 697e7cb

Browse files
author
“yezhenbin
committed
代码小调,增加readme
1 parent 01ba6d7 commit 697e7cb

File tree

10 files changed

+29
-1
lines changed

10 files changed

+29
-1
lines changed

PER-and-ACQ/net.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
import torch
22
from torch import nn
33

4+
45
class AtariNet(nn.Module):
56

67
def __init__(self, num_actions):

PER-and-ACQ/util.py

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,13 @@
11
import cv2
22
import numpy as np
3+
4+
35
def preprocess(observation):
6+
"""
7+
image preprocess
8+
:param observation:
9+
:return:
10+
"""
411
observation = cv2.cvtColor(cv2.resize(observation, (84, 110)), cv2.COLOR_BGR2GRAY)
512
observation = observation[26:110,:]
613
ret, observation = cv2.threshold(observation,1,255,cv2.THRESH_BINARY)

README.md

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,4 +26,10 @@ ddpg.py
2626

2727
## 10.1.2 A3C DDPG:
2828

29-
A3C-DDPG.py
29+
A3C-DDPG.py
30+
31+
## 联系作者
32+
33+
### 邮箱:[email protected]
34+
35+
### 代码持续更新,您若有改进建议或者问题请联系作者

ddpg.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,10 @@ def sample(self, batch_size):
3030
d.append(np.array(D, copy=False))
3131
return np.array(x), np.array(y), np.array(u), np.array(r).reshape(-1, 1), np.array(d).reshape(-1, 1)
3232

33+
3334
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
3435

36+
3537
class Actor(nn.Module):
3638
def __init__(self, state_dim, action_dim, max_action):
3739
super(Actor, self).__init__()

double-DQN.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -71,6 +71,7 @@ def learn(self, state, action, reward, next_state, done):
7171
loss.backward()
7272
self.optimizer.step()
7373

74+
7475
agent = Agent()
7576

7677
for i_episode in range(TOTAL_EPISODES):

dueling-DQN.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -124,6 +124,7 @@ def learn(self, state, action, reward, next_state, done):
124124
loss.backward()
125125
self.optimizer.step()
126126

127+
127128
agent = Agent()
128129

129130
for i_episode in range(TOTAL_EPISODES):

nature-DQN.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -69,6 +69,7 @@ def learn(self, state, action, reward, next_state, done):
6969
loss.backward()
7070
self.optimizer.step()
7171

72+
7273
agent = Agent()
7374

7475
for i_episode in range(TOTAL_EPISODES):

net.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
import torch
22
from torch import nn
33

4+
45
class AtariNet(nn.Module):
56

67
def __init__(self, num_actions):

nips-DQN.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -63,6 +63,7 @@ def learn(self, state, action, reward, next_state, done):
6363
loss.backward()
6464
self.optimizer.step()
6565

66+
6667
agent = Agent()
6768

6869
for i_episode in range(TOTAL_EPISODES):

util.py

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,13 @@
11
import cv2
22
import numpy as np
3+
4+
35
def preprocess(observation):
6+
"""
7+
image preprocess
8+
:param observation:
9+
:return:
10+
"""
411
observation = cv2.cvtColor(cv2.resize(observation, (84, 110)), cv2.COLOR_BGR2GRAY)
512
observation = observation[26:110,:]
613
ret, observation = cv2.threshold(observation,1,255,cv2.THRESH_BINARY)

0 commit comments

Comments
 (0)