You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* Ultimate Guide to [Understand & Implement Natural Language Processing](https://www.analyticsvidhya.com/blog/2017/01/ultimate-guide-to-understand-implement-natural-language-processing-codes-in-python/)
56
57
*[Introduction to NLP at Hackernoon](https://hackernoon.com/learning-ai-if-you-suck-at-math-p7-the-magic-of-natural-language-processing-f3819a689386) is for people who suck at math - in their own words
*[Deep Learning for NLP with Pytorch](http://pytorch.org/tutorials/beginner/deep_learning_nlp_tutorial.html)
58
60
59
61
Specialized Blogs
60
62
*[Deep Learning, NLP, and Representations](http://colah.github.io/posts/2014-07-NLP-RNNs-Representations/)
@@ -84,6 +86,11 @@ Bayesian, statistics and Linguistics approaches for Natural Language Processing
84
86
*[NLTK with Python 3 for Natural Language Processing](https://www.youtube.com/playlist?list=PLQVvvaa0QuDf2JswnfiGkliBInZnIC4HL) by Harrison Kinsley(sentdex). Good tutorials with NLTK code implementation
85
87
*[Computational Linguistics I](https://www.youtube.com/playlist?list=PLQVvvaa0QuDf2JswnfiGkliBInZnIC4HL) by Jordan Boyd-Graber, Lectures from University of Maryland
86
88
89
+
### Books
90
+
91
+
*[Text Mining in R](https://www.tidytextmining.com)
92
+
*[Natural Language Processing with Python](http://www.nltk.org/book/)
93
+
87
94
## Libraries
88
95
89
96
[Back to Top](#contents)
@@ -276,8 +283,10 @@ Bahdanau, Cho 2014 introduced the **attention mechanism** in NLP
276
283
[Back to Top](#contents)
277
284
278
285
*[A Neural Network Approach to Context-Sensitive Generation of Conversational Responses](http://arxiv.org/pdf/1506.06714v1.pdf)
279
-
Sordoni 2015. Generates responses to tweets. Uses [Recurrent Neural Network Language Model (RLM) architecture
280
-
of (Mikolov et al., 2010).](http://www.fit.vutbr.cz/research/groups/speech/publi/2010/mikolov_interspeech2010_IS100722.pdf) source code: [RNNLM Toolkit](http://www.rnnlm.org/)
286
+
Sordoni 2015. Generates responses to tweets.
287
+
* Uses [Recurrent Neural Network Language Model (RLM) architecture
288
+
of (Mikolov et al., 2010).](http://www.fit.vutbr.cz/research/groups/speech/publi/2010/mikolov_interspeech2010_IS100722.pdf). Source code: [RNNLM Toolkit](http://www.fit.vutbr.cz/~imikolov/rnnlm/index.html)
289
+
* RNNLM Tutorial: [Implementing RNN Language Models by Denny Britz](http://www.wildml.com/2015/09/recurrent-neural-networks-tutorial-part-2-implementing-a-language-model-rnn-with-python-numpy-and-theano/)
281
290
282
291
*[Neural Responding Machine for Short-Text Conversation](http://arxiv.org/pdf/1503.02364v2.pdf)
283
292
Shang et al. 2015 Uses Neural Responding Machine. Trained on Weibo dataset. Achieves one round conversations with 75% appropriate responses.
0 commit comments