MediaWiki API result

This is the HTML representation of the JSON format. HTML is good for debugging, but is unsuitable for application use.

Specify the format parameter to change the output format. To see the non-HTML representation of the JSON format, set format=json.

See the complete documentation, or the API help for more information.

{
    "batchcomplete": "",
    "continue": {
        "gapcontinue": "Spring_2018_CS165B_Machine_Learning",
        "continue": "gapcontinue||"
    },
    "query": {
        "pages": {
            "5": {
                "pageid": 5,
                "ns": 0,
                "title": "Spring 2017 CS292F Deep Learning for NLP",
                "revisions": [
                    {
                        "contentformat": "text/x-wiki",
                        "contentmodel": "wikitext",
                        "*": "====Instructor and Venue====\n* Instructor: [http://www.cs.ucsb.edu/~william William Wang]\n* Time: T R 1:00pm - 2:50pm \n* Location: PHELPS 2510\n* Instructor Office Hours: Tu 3-4pm HFH 1115 starting the second week.\n* Prerequisites:\n** Machine Learning (CS 165B) or equivalent\n** Good programming skills and knowledge of data structure (e.g., CS 130A)\n** Solid background in machine learning, linear algebra, probability, and calculus.\n** Comfortable with deep learning platforms such as TensorFlow, (Py)Torch, Theano, MXNet, Caffe etc.\n** Prior experiences with AWS is not required, but could be very useful.\n\n====Course Objective====\nAt the end of the quarter, students should have a good understanding about basic deep learning models, and \nshould be able to implement some fundamental algorithms for simple problems in deep learning. \nStudents will also develop an understanding of the open research problems in deep learning,\nand be able to conduct cutting-edge research with novel contributions to improve existing solutions.\n\n====Piazza====\n\n====Syllabus====\n[[Spring 2017 CS292F Syllabus]]\n\n====Course Description====\nDeep learning has revolutionized many subfields within AI. DeepMind's AlphaGo combined convolutional neural networks together with deep reinforcement learning and MCTS, and won many games against top human Go players. In computer vision, most of the leading systems in ImageNet competitions are based on deep neural networks. \nDeep learning has also changed the game in NLP: for example, Google has recently replaced their phrase-based machine translation system with neural machine translation system.\nThroughout the quarter, we will go over some of the basics in neural networks, and we will also go through the deep learning revolution after 2006. \nIn this graduate class, we will also emphasize on the development of graduate student's paper reading and presentation abilities: each student will need to present research papers related to this topic.\nLast but not least, the most important aspect of this course is for students to work on a novel research project in open problems related to NLP and deep learning, and gain hands-on experiences of doing cutting-edge research.\n\n====Text Book====\nNo textbook is required, but the following optional textbook is recommended:\n* Deep Learning, An MIT Press book, Ian Goodfellow and Yoshua Bengio and Aaron Courville\n* HTML version of the book: http://www.deeplearningbook.org/\n\n====Project====\nOne key aspect of this class is to have students to gain hands-on experiences in open research problems.\nTo do this, each student will need to propose a research project. \nThe teaching staff will provide the feedback on the proposal, and track the progress of each student.\nComputing resources will be provided: each team will be provided with sufficient amount of AWS credits for their projects.\n\nIn the project proposal, each team must clearly mention the following aspects of their project:\n* What is the motivation of the problem?\n* What is the exact definition of the problem? How do we formulate the problem in machine learning?\n* What are some existing approaches to this problem?\n* What are some existing datasets that you can work on?\n* What is the novelty in your project? New problem? New approach? New dataset?\n* How are you going to implement your approach and verify the idea?\n\nGood places to look for project inspirations:\n* Recent papers from ACL, EMNLP, NAACL from ACL Anthology:  http://aclweb.org/anthology/\n* Recent papers from ICML, NIPS, and ICLR conferences: http://jmlr.org/proceedings/ http://papers.nips.cc/\n\nFAQ: Can I use my existing research projects / thesis research project as the project in this class?\nA: I would prefer students to get out of their comfort zone and try something new in this class.\nIf you are using existing techniques from your existing project, it is unlikely that you will be able to learn anything new during the course project.\nHowever, you may still draw the inspiration from your research problem to formulate your class project.\n\n=====Available Datasets=====\n* Wikipedia Harassment/Personal Attack Dataset https://figshare.com/projects/Wikipedia_Talk/16731 Ex Machina: Personal Attacks Seen at Scale, https://arxiv.org/abs/1610.08914 \n* Stance Detection / Fake News Detection / Automated Fact-Checking, email William.\n* Deep learning for abstractive humor generation. Dataset: https://www.cs.ucsb.edu/~william/papers/meme.pdf\n* NELL Knowledge Graph http://rtw.ml.cmu.edu/\n* Relation Prediction / Reasoning FB15K-237 https://www.microsoft.com/en-us/download/details.aspx?id=52312\n* Abstractive summarization datasets https://www.aclweb.org/anthology/D/D15/D15-1044.pdf\n* WikiHow: learning processes from lists and free text https://github.com/paolo7/KnowHowDataset\n\n====Grading====\nThere will be two homework assignments (20%), one project (65%), and an in-class paper presentation (15%). The breakdown of project grading includes: 1-page proposal (10%), mid-term presentation (10%), final presentation (15%), and a final report (30%). Four late days are allowed with no penalty. After that 50% will be deducted if it is within 4 days after the due day, unless you have a note from the doctors' office. Homework assignment submissions that are five days late will receive zero credits.\n\n====Final Report Format====\nYou must use the [https://nips.cc/Conferences/2017/PaperInformation/StyleFiles NIPS 2017 latex style files] for writing the report.\nThe final report must be 6-8 pages long including references. It is encouraged to include the following components in your report (not necessarily this order): \nabstract, introduction (motivation, task definition, your novel contributions), related work, your technical approach, such as math formulation of the problem, algorithms, theorems (if any),\nexperiments, discussion, and conclusion.\n\n====Academic Integrity====\nWe follow UCSB's academic integrity policy from UCSB Campus Regulations, Chapter VII:``Student Conduct and Discipline\"):\n\n* ''It is expected that students attending the University of California understand and subscribe to the ideal of academic integrity, and are willing to bear individual responsibility for their work. Any work (written or otherwise) submitted to fulfill an academic requirement must represent a student\u2019s original work. Any act of academic dishonesty, such as cheating or plagiarism, will subject a person to University disciplinary action. Using or attempting to use materials, information, study aids, or commercial \u201cresearch\u201d services not authorized by the instructor of the course constitutes cheating. Representing the words, ideas, or concepts of another person without appropriate attribution is plagiarism. Whenever another person\u2019s written work is utilized, whether it be a single phrase or longer, quotation marks must be used and sources cited. Paraphrasing another\u2019s work, i.e., borrowing the ideas or concepts and putting them into one\u2019s \u201cown\u201d words, must also be acknowledged. Although a person\u2019s state of mind and intention will be considered in determining the University response to an act of academic dishonesty, this in no way lessens the responsibility of the student.''\n\nMore specifically, we follow Stefano Tessaro and William Cohen's policy in this class:\n\nYou cannot copy the code or answers to homework questions or exams from your classmates or from other sources;\nYou may discuss course materials and assignments with your classmate, but you '''cannot write anything down'''. You must write down the answers / code independently. The presence or absence of any form of help or collaboration, whether given or received, must be explicitly stated and disclosed in full by all involved, on the first page of their assignment. Specifically, each assignment solution must start by answering the following questions:\n* (1) Did you receive any help whatsoever from anyone in solving this assignment? Yes / No.\n** If you answered 'yes', give full details:  (e.g. ``Jane explained to me what is asked in Question 3.4\")\n* (2) Did you give any help whatsoever to anyone in solving this assignment? Yes / No.\n** If you answered 'yes', give full details: (e.g. ``I pointed Joe to section 2.3 to help him with Question 2\".\n* No electronics are allowed during exams, but you may prepare an A4-sized note and bring it the exam.\n* If you have questions, often ask the teaching staff.\n\nAcademic dishonesty will be reported to the highest line of command at UCSB. Students who engage in such activities will receive an F grade automatically. \n\n====Accessibility====\nStudents with documented disability are asked to contact the DSP office to arrange the necessary academic accommodations.\n\n====Discussions====\nAll discussions and questions should be posted on our course Piazza site."
                    }
                ]
            },
            "8": {
                "pageid": 8,
                "ns": 0,
                "title": "Spring 2017 CS292F Syllabus",
                "revisions": [
                    {
                        "contentformat": "text/x-wiki",
                        "contentmodel": "wikitext",
                        "*": "*04/04 Introduction, logistics, NLP, and deep learning.\n*04/06 Tips for a successful class project\n*04/11 NLP Tasks\n*04/13 Word embeddings \n** Christian Bueno: [https://people.cs.umass.edu/~arvind/emnlp2014.pdf Efficient Non-parametric Estimation of Multiple Embeddings per Word in Vector Space, Neelakantan et al., EMNLP 2014]\n** Keqian Li: [http://www.anthology.aclweb.org/D/D14/D14-1162.pdf Glove: Global Vectors for Word Representation, J Pennington, R Socher, CD Manning - EMNLP, 2014]\n** Mengya Tao: [http://www.aclweb.org/anthology/P15-1173 AutoExtend: Extending Word Embeddings to Embeddings for Synsets and Lexemes, Rothe and Schutze, ACL 2015]\n*04/18 Neural network basics (Project proposal due, HW1 out)\n** Arturo Deza: [http://www.iro.umontreal.ca/~vincentp/ift3395/lectures/backprop_old.pdf Learning representations by back-propagating errors, Nature, 1986]\n** Shayan Sadigh: [https://arxiv.org/abs/1609.04747 An overview of gradient descent optimization algorithms, Sebastian Ruder, Arxiv 2016]\n*04/20 Recursive Neural Networks \n** Yun Zhao: [https://nlp.stanford.edu/pubs/SocherBauerManningNg_ACL2013.pdf Parsing with Compositional Vector Grammars, Socher et al., ACL 2013]\n** Rachel Redberg: [https://nlp.stanford.edu/~socherr/EMNLP2013_RNTN.pdf Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank, Socher et al., EMNLP 2013]\n*04/25 RNNs (NLP seminar: Stanford NLP's Jiwei Li 04/26)\n** [http://www.fit.vutbr.cz/research/groups/speech/publi/2010/mikolov_interspeech2010_IS100722.pdf Recurrent neural network based language model] \n** Yuanshun Yao: [https://arxiv.org/pdf/1308.0850.pdf Generating Sequences With Recurrent Neural Networks, Alex Graves, 2013 arxiv]\n*04/27 LSTMs/GRUs\n** [http://www.bioinf.jku.at/publications/older/2604.pdf Long short term memory, S. Hochreiter and J. Schmidhuber, Neural Computation, 1997]\n** [https://arxiv.org/pdf/1409.1259.pdf On the Properties of Neural Machine Translation: Encoder\u2013Decoder Approaches, Cho et al., 2014]\n** Daniel Spokoyny: [https://arxiv.org/pdf/1502.02367v3.pdf Gated Feedback Recurrent Neural Networks, Chung et al., ICML 2015]\n*05/02 Sequence-to-sequence models and neural machine translation (HW1 due and HW2 out)\t\n** Wenhan Xiong: [https://arxiv.org/pdf/1406.1078.pdf Learning Phrase Representations using RNN Encoder\u2013Decoder for Statistical Machine Translation, Cho et al., EMNLP 2014]\n** Xiaoyong Jin: [https://papers.nips.cc/paper/5346-sequence-to-sequence-learning-with-neural-networks.pdf Sequence to Sequence Learning with Neural Networks, Sutskever et al., NIPS 2014]\n*05/04 Attention mechanisms\n** Xinyi Zhang:  [https://arxiv.org/pdf/1409.0473.pdf NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE, Bahdanau et al., ICLR 2015]\n** Hanwen Zha: [https://arxiv.org/abs/1506.03340 Teaching Machines to Read and Comprehend, NIPS 2015]\n** Zhujun Xiao: [http://papers.nips.cc/paper/5846-end-to-end-memory-networks.pdf End-to-end memory networks, NIPS 2015]\n*05/09 Project: mid-term presentation (1)\n** JONNALAGADDA, ADITYA\n** ZHA, HANWEN\n** AGHAKHANI, HOJJAT\n** JAIN, ROHAN\n** WANG, XIN\n** KOUPAEE, MAHNAZ\n** YAO, YUANSHUN\n** LI, ZHIJING\n*05/11 Project: mid-term presentation (2)\n** SPOKOYNY, DANIEL\n** ZHANG, FANGJUN\n** FEINN, ZACHARY\n** JIN, XIAOYONG\n** REDBERG, RACHEL\n** XIONG, WENHAN\n** ZHAO, YUN\n** SADIGH, SHAYAN\n** XIAO, ZHUJUN\n** ZHANG, XINYI\n*05/16 Convolutional Neural Networks  (HW2 due)\n** Zachary Feinn: [http://ronan.collobert.com/pub/matos/2011_nlp_jmlr.pdf Natural Language Processing (Almost) from Scratch, Collobert et al., JMLR 2011]\n** Fangjun Zhang: [https://arxiv.org/pdf/1510.03820.pdf A Sensitivity Analysis of (and Practitioners\u2019 Guide to) Convolutional Neural Networks for Sentence Classification, Zhang and Wallace, Arxiv 2015]\n*05/18 Language and vision\n** Shiliang Tang: [https://arxiv.org/pdf/1411.4555.pdf Show and Tell: A Neural Image Caption Generator, CVPR 2015]\n** Aditya Jonnalagadda: [http://www.cv-foundation.org/openaccess/content_cvpr_2015/papers/Karpathy_Deep_Visual-Semantic_Alignments_2015_CVPR_paper.pdf Deep Visual-Semantic Alignments for Generating Image Descriptions, Andrej Karpathy and Li Fei-Fei, CVPR 2015]\n** : [http://www.cv-foundation.org/openaccess/content_iccv_2015/papers/Zhu_Aligning_Books_and_ICCV_2015_paper.pdf Aligning Books and Movies: Towards Story-like Visual Explanations by Watching Movies and Reading Books, Zhu et al., ICCV 2015]\n*05/23 Deep Reinforcement Learning 1\n** Rohan Jain: [https://aclweb.org/anthology/D16-1127, Deep Reinforcement Learning for Dialogue Generation, Li et al., EMNLP 2016]\n** Mahnaz Koupaee: [https://arxiv.org/abs/1603.07954 Improving Information Extraction by Acquiring External Evidence with Reinforcement Learning, Narasimh et al., EMNLP 2016]\n*05/25 Deep Reinforcement Learning 2\n** Xin Wang: [https://www.cs.toronto.edu/~vmnih/docs/dqn.pdf Playing Atari with Deep Reinforcement Learning, Mnih et al., NIPS workshop 2013]\n** Zhijing Li: [https://arxiv.org/pdf/1509.02971.pdf Continuous control with deep reinforcement learning, Lillicrap et al, ICLR 2016]\n*05/30 Unsupervised Learning\n** : [https://arxiv.org/abs/1312.6114 Auto-encoding variational Bayes, Kingma and Welling, ICLR 2014]\n** Hojjat Aghakhani: [https://arxiv.org/pdf/1511.06434.pdf%C3%AF%C2%BC%E2%80%B0 Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks, Redford et al., 2015]\n*06/01 Project: final presentation (1)\n**ADITYA \n**XINYI\n**RACHEL\n**YUANSHUN\n**HANWEN\n**ROHAN\n*06/06 Project: final presentation (2)\n**MAHNAZ\n**YUN\n**XIN\n**SHAYAN\n**ZHIJING\n**ZHUJUN\n**ZACHARY\n*06/08 Project: final presentation (3)\n**XIAOYONG\n**DANIEL \n**WENHAN\n**HOJJAT\n**SHILIANG\n**FANGJUN \n*06/10 23:59PM PT Project Final Report Due."
                    }
                ]
            }
        }
    }
}