Difference between revisions of "Winter 2018 CS595I Advanced NLP/ML Seminar"

From courses
Jump to: navigation, search
(Reinforcement Learning)
(Reinforcement Learning)
Line 21: Line 21:
 
* Robust Imitation of Diverse Behaviors, Wang et al. 2017, https://arxiv.org/pdf/1707.02747.pdf
 
* Robust Imitation of Diverse Behaviors, Wang et al. 2017, https://arxiv.org/pdf/1707.02747.pdf
 
* Programmable Agents, Denil et al., https://arxiv.org/pdf/1706.06383v1.pdf
 
* Programmable Agents, Denil et al., https://arxiv.org/pdf/1706.06383v1.pdf
 +
* Compatible Reward Inverse Reinforcement Learning Metelli et al., NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=8993
 +
* Scalable trust-region method for deep reinforcement learning using Kronecker-factored approximation, Wu et al., NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=10087
 
* Expected Policy Gradients, Kamil Ciosek, Shimon Whiteson, https://arxiv.org/abs/1706.05374
 
* Expected Policy Gradients, Kamil Ciosek, Shimon Whiteson, https://arxiv.org/abs/1706.05374
 
* Reinforcement Learning with Deep Energy-Based Policies  Haarnoja et al, ICML 2017 http://proceedings.mlr.press/v70/haarnoja17a/haarnoja17a.pdf
 
* Reinforcement Learning with Deep Energy-Based Policies  Haarnoja et al, ICML 2017 http://proceedings.mlr.press/v70/haarnoja17a/haarnoja17a.pdf
 
* Interpolated Policy Gradient: Merging On-Policy and Off-Policy Gradient Estimation for Deep Reinforcement Learning, Gu et al., NIPS 2017 https://arxiv.org/abs/1706.00387
 
* Interpolated Policy Gradient: Merging On-Policy and Off-Policy Gradient Estimation for Deep Reinforcement Learning, Gu et al., NIPS 2017 https://arxiv.org/abs/1706.00387
 
* Sequence Level Training with Recurrent Neural Networks https://arxiv.org/pdf/1511.06732.pdf
 
* Sequence Level Training with Recurrent Neural Networks https://arxiv.org/pdf/1511.06732.pdf
 +
* Distral: Robust multitask reinforcement learning, Teh et al., NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=9227
 +
* Repeated Inverse Reinforcement Learning, Amin et al., NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=10107
 
* Hybrid Reward Architecture for Reinforcement Learning, Van Seijen et al., NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=9314
 
* Hybrid Reward Architecture for Reinforcement Learning, Van Seijen et al., NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=9314
 
* Cold-Start Reinforcement Learning with Softmax Policy Gradient, Ding and Soirut, NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=9067
 
* Cold-Start Reinforcement Learning with Softmax Policy Gradient, Ding and Soirut, NIPS 2017 https://nips.cc/Conferences/2017/Schedule?showEvent=9067

Revision as of 00:38, 4 January 2018

Time: TBD Location: HFH 1132.

If you registered this class, you should contact the instructor to present one paper *and* be the discussant of two papers below.

  • Presenter: prepare a short summary of no more than 15 mins of presentation.
  • Discussant: prepare two questions for discussion about the paper.

If you don't present or lead the discussion, you will then need to write a 2-page final report in ICML 2018 style, comparing any two of the papers below. Due: TBD to william@cs.ucsb.edu.

Word Embeddings

Relational Learning and Reasoning

Reinforcement Learning

Generation

Dialog

Learning

NLP for Computational Social Science