Difference between revisions of "Spring 2017 CS292F Syllabus"

From courses
Jump to: navigation, search
Line 10: Line 10:
 
** Shayan Sadigh: [https://arxiv.org/abs/1609.04747 An overview of gradient descent optimization algorithms, Sebastian Ruder, Arxiv 2016]
 
** Shayan Sadigh: [https://arxiv.org/abs/1609.04747 An overview of gradient descent optimization algorithms, Sebastian Ruder, Arxiv 2016]
 
*04/20 Recursive Neural Networks  
 
*04/20 Recursive Neural Networks  
** [https://nlp.stanford.edu/pubs/SocherBauerManningNg_ACL2013.pdf Parsing with Compositional Vector Grammars, Socher et al., ACL 2013]
+
** Yun Zhao: [https://nlp.stanford.edu/pubs/SocherBauerManningNg_ACL2013.pdf Parsing with Compositional Vector Grammars, Socher et al., ACL 2013]
** [https://nlp.stanford.edu/~socherr/EMNLP2013_RNTN.pdf Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank, Socher et al., EMNLP 2013]
+
** Rachel Redberg: [https://nlp.stanford.edu/~socherr/EMNLP2013_RNTN.pdf Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank, Socher et al., EMNLP 2013]
 
*04/25 RNNs (NLP seminar: Stanford NLP's Jiwei Li 04/26)
 
*04/25 RNNs (NLP seminar: Stanford NLP's Jiwei Li 04/26)
 
** Adam Ibrahim: [http://www.fit.vutbr.cz/research/groups/speech/publi/2010/mikolov_interspeech2010_IS100722.pdf Recurrent neural network based language model]  
 
** Adam Ibrahim: [http://www.fit.vutbr.cz/research/groups/speech/publi/2010/mikolov_interspeech2010_IS100722.pdf Recurrent neural network based language model]  
Line 17: Line 17:
 
*04/27 LSTMs/GRUs
 
*04/27 LSTMs/GRUs
 
** Omid Askarisichani: [http://www.bioinf.jku.at/publications/older/2604.pdf Long short term memory, S. Hochreiter and J. Schmidhuber, Neural Computation, 1997]
 
** Omid Askarisichani: [http://www.bioinf.jku.at/publications/older/2604.pdf Long short term memory, S. Hochreiter and J. Schmidhuber, Neural Computation, 1997]
** [https://arxiv.org/pdf/1409.1259.pdf On the Properties of Neural Machine Translation: Encoder–Decoder Approaches, Cho et al., 2014]
+
** Brandon Huyh: [https://arxiv.org/pdf/1409.1259.pdf On the Properties of Neural Machine Translation: Encoder–Decoder Approaches, Cho et al., 2014]
 
** Daniel Spokoyny: [https://arxiv.org/pdf/1502.02367v3.pdf Gated Feedback Recurrent Neural Networks, Chung et al., ICML 2015]
 
** Daniel Spokoyny: [https://arxiv.org/pdf/1502.02367v3.pdf Gated Feedback Recurrent Neural Networks, Chung et al., ICML 2015]
 
*05/02 Sequence-to-sequence models and neural machine translation (HW1 due and HW2 out)
 
*05/02 Sequence-to-sequence models and neural machine translation (HW1 due and HW2 out)
Line 25: Line 25:
 
** Xinyi Zhang:  [https://arxiv.org/pdf/1409.0473.pdf NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE, Bahdanau et al., ICLR 2015]
 
** Xinyi Zhang:  [https://arxiv.org/pdf/1409.0473.pdf NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE, Bahdanau et al., ICLR 2015]
 
** Hanwen Zha: [https://arxiv.org/abs/1506.03340 Teaching Machines to Read and Comprehend, NIPS 2015]
 
** Hanwen Zha: [https://arxiv.org/abs/1506.03340 Teaching Machines to Read and Comprehend, NIPS 2015]
** [http://papers.nips.cc/paper/5846-end-to-end-memory-networks.pdf End-to-end memory networks, NIPS 2015]
+
** Zhujun Xiao: [http://papers.nips.cc/paper/5846-end-to-end-memory-networks.pdf End-to-end memory networks, NIPS 2015]
 
*05/09 Project: mid-term presentation (1)
 
*05/09 Project: mid-term presentation (1)
 
*05/11 Project: mid-term presentation (2)
 
*05/11 Project: mid-term presentation (2)
 
*05/16 Convolutional Neural Networks  (HW2 due)
 
*05/16 Convolutional Neural Networks  (HW2 due)
** [http://ronan.collobert.com/pub/matos/2011_nlp_jmlr.pdf Natural Language Processing (Almost) from Scratch, Collobert et al., JMLR 2011]
+
** Zachary Feinn: [http://ronan.collobert.com/pub/matos/2011_nlp_jmlr.pdf Natural Language Processing (Almost) from Scratch, Collobert et al., JMLR 2011]
** [https://arxiv.org/pdf/1510.03820.pdf A Sensitivity Analysis of (and Practitioners’ Guide to) Convolutional Neural Networks for Sentence Classification, Zhang and Wallace, Arxiv 2015]
+
** Fangjun Zhang: [https://arxiv.org/pdf/1510.03820.pdf A Sensitivity Analysis of (and Practitioners’ Guide to) Convolutional Neural Networks for Sentence Classification, Zhang and Wallace, Arxiv 2015]
 
*05/18 Language and vision
 
*05/18 Language and vision
 
** Shiliang Tang: [https://arxiv.org/pdf/1411.4555.pdf Show and Tell: A Neural Image Caption Generator, CVPR 2015]
 
** Shiliang Tang: [https://arxiv.org/pdf/1411.4555.pdf Show and Tell: A Neural Image Caption Generator, CVPR 2015]
Line 40: Line 40:
 
*05/25 Deep Reinforcement Learning 2
 
*05/25 Deep Reinforcement Learning 2
 
** Xin Wang: [https://www.cs.toronto.edu/~vmnih/docs/dqn.pdf Playing Atari with Deep Reinforcement Learning, Mnih et al., NIPS workshop 2013]
 
** Xin Wang: [https://www.cs.toronto.edu/~vmnih/docs/dqn.pdf Playing Atari with Deep Reinforcement Learning, Mnih et al., NIPS workshop 2013]
** [https://arxiv.org/pdf/1509.02971.pdf Continuous control with deep reinforcement learning, Lillicrap et al, ICLR 2016]
+
** Zhijing Li: [https://arxiv.org/pdf/1509.02971.pdf Continuous control with deep reinforcement learning, Lillicrap et al, ICLR 2016]
 
*05/30 Unsupervised Learning
 
*05/30 Unsupervised Learning
 
** [https://arxiv.org/abs/1312.6114 Auto-encoding variational Bayes, Kingma and Welling, ICLR 2014]
 
** [https://arxiv.org/abs/1312.6114 Auto-encoding variational Bayes, Kingma and Welling, ICLR 2014]

Revision as of 13:04, 7 April 2017