Awesome-NLP-Research (ANLP)

Overview

Awesome-NLP-Research (ANLP)

(Update on 2020-01-10: we have also added the presentations from the Fall 2020 installment of the course. Check for them under "slides2020".)

As part of the Fall 2018 course CPSC 677 "Advanced Natural Language Processing" at Yale, we developed, with the help of the students, a corpus of useful resources for NLP research. Bibliographies and Powerpoint Presentations for each topic are found below, in addition to several blog posts. We asked the students to also list relevant and prerequisite concepts for each topic, and these keywords are found here.

If you have any questions, would like to contribute further to this project or feel we are missing an important citation, please contact Alex Fabbri at alexander[dot]fabbri[at]yale.[first three letters of education]

Overview of papers presented in class

  • Capsule Networks for NLP by Will Merrill - BIB BLOG SLIDES
  • Commonsense Learning by Michihiro Yasunaga - BIB SLIDES
  • Dialogue Systems by Suyi Li - BIB SLIDES
  • Multilingual-Word-Embeddings by Davey Proctor - BIB SLIDES
  • Neural Embeddings By John Brandt - BIB SLIDES
  • Temporal and Dynamic Embeddings by Yavuz Nuzumlali - BIB SLIDES
  • NLP in Finance by Gaurav Pathak BIB SLIDES
  • Natural Language Generation by Tianwei She - BIB SLIDES
  • Knowledge Graphs by Tomoe Mizutani - BIB SLIDES
  • Cross-Lingual Information Retrieval by Rui Zhang - BIB BLOG SLIDES
  • Neural Information Retrieval by Danny Keller - BIB SLIDES
  • Character-Level Language Modeling by Angus Fong - BIB SLIDES
  • Latent Variable Models in NLP by Brian Kitano - BIB SLIDES
  • Unsupervised Machine Translation By Yongjie Lin - BIB SLIDES
  • Neural Computational Morphology by Garrett Bingham - BIB SLIDES
  • Network Methods by Noah Amsel - BIB SLIDES
  • Neural Semi-Supervised Learning by Alex Fabbri - BIB SLIDES
  • Question Answering by Talley Amir - BIB SLIDES
  • Attribute-Level Sentiment Analaysis by Ishita Chakraborty and Davey Proctor - BIB BLOG SLIDES
  • Semantic Parsing by Bo Pang - BIB SLIDES
  • Sequence2Sequence by Jack Koch - BIB SLIDES
  • Seq2SQL by Tao Yu - BIB SLIDES
  • Spectral Learning by Hannah Lawrence - BIB SLIDES
  • Single Document Summarization by Yi Chern Tan - BIB SLIDES
  • Transfer Learning by Irene Li - BIB SLIDES

Additionally, students from the class made blog posts on the following topics:

  • DARTS - BLOG
  • OpenAI Transformer - BLOG
You might also like...
The Internet Archive Research Assistant - Daily search Internet Archive for new items matching your keywords
The Internet Archive Research Assistant - Daily search Internet Archive for new items matching your keywords

The Internet Archive Research Assistant - Daily search Internet Archive for new items matching your keywords

Flexible interface for high-performance research using SOTA Transformers leveraging Pytorch Lightning, Transformers, and Hydra.
Flexible interface for high-performance research using SOTA Transformers leveraging Pytorch Lightning, Transformers, and Hydra.

Flexible interface for high performance research using SOTA Transformers leveraging Pytorch Lightning, Transformers, and Hydra. What is Lightning Tran

Ongoing research training transformer language models at scale, including: BERT & GPT-2

What is this fork of Megatron-LM and Megatron-DeepSpeed This is a detached fork of https://github.com/microsoft/Megatron-DeepSpeed, which in itself is

Ongoing research training transformer language models at scale, including: BERT & GPT-2
Ongoing research training transformer language models at scale, including: BERT & GPT-2

Megatron (1 and 2) is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA.

A Non-Autoregressive Transformer based TTS, supporting a family of SOTA transformers with supervised and unsupervised duration modelings. This project grows with the research community, aiming to achieve the ultimate TTS.
A Non-Autoregressive Transformer based TTS, supporting a family of SOTA transformers with supervised and unsupervised duration modelings. This project grows with the research community, aiming to achieve the ultimate TTS.

A Non-Autoregressive Transformer based TTS, supporting a family of SOTA transformers with supervised and unsupervised duration modelings. This project grows with the research community, aiming to achieve the ultimate TTS.

Python bindings to the dutch NLP tool Frog (pos tagger, lemmatiser, NER tagger, morphological analysis, shallow parser, dependency parser)

Frog for Python This is a Python binding to the Natural Language Processing suite Frog. Frog is intended for Dutch and performs part-of-speech tagging

💫 Industrial-strength Natural Language Processing (NLP) in Python

spaCy: Industrial-strength NLP spaCy is a library for advanced Natural Language Processing in Python and Cython. It's built on the very latest researc

NLP, before and after spaCy

textacy: NLP, before and after spaCy textacy is a Python library for performing a variety of natural language processing (NLP) tasks, built on the hig

Owner
Language, Information, and Learning at Yale
Language, Information, and Learning at Yale
An open-source NLP research library, built on PyTorch.

An Apache 2.0 NLP research library, built on PyTorch, for developing state-of-the-art deep learning models on a wide variety of linguistic tasks. Quic

AI2 11.4k Jan 1, 2023
An open-source NLP research library, built on PyTorch.

An Apache 2.0 NLP research library, built on PyTorch, for developing state-of-the-art deep learning models on a wide variety of linguistic tasks. Quic

AI2 9.7k Feb 18, 2021
Machine learning models from Singapore's NLP research community

SG-NLP Machine learning models from Singapore's natural language processing (NLP) research community. sgnlp is a Python package that allows you to eas

AI Singapore | AI Makerspace 21 Dec 17, 2022
Grading tools for Advanced NLP (11-711)Grading tools for Advanced NLP (11-711)

Grading tools for Advanced NLP (11-711) Installation You'll need docker and unzip to use this repo. For docker, visit the official guide to get starte

Hao Zhu 2 Sep 27, 2022
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language mod

null 20.5k Jan 8, 2023
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production

Provides an implementation of today's most used tokenizers, with a focus on performance and versatility. Main features: Train new vocabularies and tok

Hugging Face 6.2k Dec 31, 2022
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language mod

null 11.3k Feb 18, 2021
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production

Provides an implementation of today's most used tokenizers, with a focus on performance and versatility. Main features: Train new vocabularies and tok

Hugging Face 4.3k Feb 18, 2021
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language mod

null 13.2k Jul 7, 2021