Skip to content

Commit c729f79

Browse files
authored
Update README.md
1 parent 9f5cb26 commit c729f79

File tree

1 file changed

+19
-3
lines changed

1 file changed

+19
-3
lines changed

README.md

Lines changed: 19 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,18 @@
11

22
## About
33

4-
SRU is a recurrent unit that can run over 10 times faster than cuDNN LSTM, without loss of accuracy tested on many tasks.
4+
**SRU** is a recurrent unit that can run over 10 times faster than cuDNN LSTM, without loss of accuracy tested on many tasks.
55

66
<p align="center">
7-
<img width=650 src="imgs/speed.png"><br>
8-
<i>Average processing time of cuDNN LSTM, conv2d and SRU, tested on GTX 1070</i>
7+
<img width=620 src="imgs/speed.png"><br>
8+
<i>Average processing time of LSTM, conv2d and SRU, tested on GTX 1070</i><br>
9+
</p>
10+
11+
For example, the figures above and below present the processing time of a single mini-batch and the training time for sentence-level classification tasks. SRU achieve 10 to 16 times speed-up compared to LSTM, and operates as fast as (or faster than) word-level convolutional model (CNNs).
12+
13+
<p align="center">
14+
<img width=580 src="imgs/classification.png"><br>
15+
<i>Training time (x-axis) vs valid accuracies (y-axis) on classification benchmarks</i><br>
916
</p>
1017

1118
<br>
@@ -19,6 +26,15 @@ CuPy and pynvrtc needed to compile the CUDA code into a callable function at run
1926

2027
<br>
2128

29+
## Tasks
30+
- classification
31+
- question answering on SQuAD
32+
- language modelling on PTB
33+
- machine translation
34+
- speech recognition
35+
36+
<br>
37+
2238
## Contributors
2339
- **Tao Lei** ([email protected])
2440
- **Yu Zhang** ([email protected])

0 commit comments

Comments
 (0)