Difference between revisions of "Winter 2018 CS291A Syllabus"
From courses
Line 14: | Line 14: | ||
** : [https://nlp.stanford.edu/~socherr/EMNLP2013_RNTN.pdf Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank, Socher et al., EMNLP 2013] | ** : [https://nlp.stanford.edu/~socherr/EMNLP2013_RNTN.pdf Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank, Socher et al., EMNLP 2013] | ||
*02/06 RNNs | *02/06 RNNs | ||
− | ** [https://pdfs.semanticscholar.org/8adb/8257a423f55b1f20ba62c8b20118d76a25c7.pdf A Learning Algorithm for Continually Running Fully Recurrent Neural Networks, Ronald J. Williams and David Zipser, 1989] | + | ** : [https://pdfs.semanticscholar.org/8adb/8257a423f55b1f20ba62c8b20118d76a25c7.pdf A Learning Algorithm for Continually Running Fully Recurrent Neural Networks, Ronald J. Williams and David Zipser, 1989] |
** : [http://www.fit.vutbr.cz/research/groups/speech/publi/2010/mikolov_interspeech2010_IS100722.pdf Recurrent neural network based language model] | ** : [http://www.fit.vutbr.cz/research/groups/speech/publi/2010/mikolov_interspeech2010_IS100722.pdf Recurrent neural network based language model] | ||
** : [https://arxiv.org/pdf/1308.0850.pdf Generating Sequences With Recurrent Neural Networks, Alex Graves, 2013 arxiv] | ** : [https://arxiv.org/pdf/1308.0850.pdf Generating Sequences With Recurrent Neural Networks, Alex Graves, 2013 arxiv] | ||
Line 26: | Line 26: | ||
** : [http://www.aclweb.org/anthology/P16-1100 Achieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models, Luong and Manning, ACL 2016] | ** : [http://www.aclweb.org/anthology/P16-1100 Achieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models, Luong and Manning, ACL 2016] | ||
*02/15 Attention mechanisms | *02/15 Attention mechanisms | ||
− | ** : | + | ** : [https://arxiv.org/pdf/1409.0473.pdf NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE, Bahdanau et al., ICLR 2015] |
** : [https://arxiv.org/abs/1506.03340 Teaching Machines to Read and Comprehend, NIPS 2015] | ** : [https://arxiv.org/abs/1506.03340 Teaching Machines to Read and Comprehend, NIPS 2015] | ||
** : [http://papers.nips.cc/paper/5846-end-to-end-memory-networks.pdf End-to-end memory networks, NIPS 2015] | ** : [http://papers.nips.cc/paper/5846-end-to-end-memory-networks.pdf End-to-end memory networks, NIPS 2015] |
Revision as of 15:24, 16 January 2018
- 01/16 Introduction, logistics, NLP, and deep learning.
- 01/18 Tips for a successful class project
- 01/23 NLP Tasks
- 01/25 Word embeddings
- : Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space, Neelakantan et al., EMNLP 2014
- : Glove: Global Vectors for Word Representation, J Pennington, R Socher, CD Manning - EMNLP, 2014
- : AutoExtend: Extending Word Embeddings to Embeddings for Synsets and Lexemes, Rothe and Schutze, ACL 2015
- 01/30 Neural network basics (Project proposal due to Grader: Ke Ni < ke00@ucsb.edu> , HW1 out)
- 02/01 Recursive Neural Networks
- 02/06 RNNs
- 02/08 LSTMs/GRUs
- 02/13 Sequence-to-sequence models and neural machine translation (HW1 due and HW2 out)
- : Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation, Cho et al., EMNLP 2014
- : Sequence to Sequence Learning with Neural Networks, Sutskever et al., NIPS 2014
- : Achieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models, Luong and Manning, ACL 2016
- 02/15 Attention mechanisms
- 02/20 Convolutional Neural Networks (Mid-term report due to Grader: Ke Ni <ke00@ucsb.edu>)
- : Natural Language Processing (Almost) from Scratch, Collobert et al., JMLR 2011
- : A Sensitivity Analysis of (and Practitioners’ Guide to) Convolutional Neural Networks for Sentence Classification, Zhang and Wallace, Arxiv 2015
- : Convolutional Neural Network Architectures for Matching Natural Language Sentences, Hu et al., NIPS 2014
- 02/22 Language and vision
- 02/27 Deep Reinforcement Learning 1 (HW2 due)
- 03/01 Deep Reinforcement Learning 2
- 03/06 Unsupervised Learning
- 03/08 Project: final presentation (1)
- 03/13 Project: final presentation (2)
- 03/15 Project: final presentation (3)
- 03/23 23:59PM PT Project Final Report Due. Grader: Ke Ni <ke00@ucsb.edu>