jinmang2 / Awesome-Papers

:snowflake: All about my interest Papers and Review :)
3 stars 2 forks source link
attention-mechanism conversational-ai deep-learning deepmind financial-engineering generative-model mathematics meta-learning natural-language-understanding nlp reinforcement-learning schmidhuber transformers vision

Awesome-Papers

:question: Objective of jinmang2/Awesome-Papers Repo.

:bulb: To be AI Researcher, Artist and Good Person...!!

2021 Papers to Read

2020 Reading Papers

Reinforcement Learning

Meta-Learning

NLP

Graph

Conversational AI

Fundamental

ETC

:office: NLP

Tokenization

Wordpiece

SentencePiece

Morphological

Word Vector Representation

NLP Tasks

A large annotated corpus for learning natural language inference, Bowman et al., 2015 (EMNLP)

A board-coverage challenge corpus for sentence understanding through inference, Williams et al., 2018

SQuad: 100,000+ questions for machine comprehension of text, Rajpurkar et al., 2016

introduction to th conll-2003 shared task: language-independent named entity recognition, Tjong Kim Sang and De Meulder, 2003

Dependency Parsing

Neural Machine Translation

Text Classification

Question Answering

Stochastic Answer Networks for Machine Reading Comprehension https://arxiv.org/abs/1712.03556

Textual Entailment

Enhanced LSTM for Natural Language Inference https://arxiv.org/abs/1609.06038

Semantic Role Labeling

Deep Semantic Role Labeling: What Works and What’s Next https://www.aclweb.org/anthology/P17-1044/

Summarization

Extractive

  • [ ] BertSum; Fine-tune BERT for Extractive Summarization (19.03.25, arxiv) paper
  • [ ] BertSum-Full Paper; Text Summarization with Pretrained Encoders (19.08.22, arxiv) paper

Pre-trained NLP Architecture

Word Representations: A Simple and General Method for Semi-Supervised Learning

institute subtitle title journal published etc
AllenAI ELMo Deep contextualized word representations ACL 2018 paper
AllenAI LongFormer Longformer: The Long-Document Transformer arxiv 20.04.10 paper
GoogleAI BERT BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding ACL 2018 paper
GoogleAI ALBERT ALBERT: A LITE BERT FOR SELF-SUPERVISED LEARNING OF LANGUAGE REPRESENTATIONS ICLR 19.09.26 paper
GoogleAI T5 Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer JMLR 19.10.23 paper
GoogleAI PEGASUS PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization ICML 2020 paper
GoogleAI ELECTRA ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS ICLR 2020 paper
DeepMind Compressive Transformers COMPRESSIVE TRANSFORMERS FOR LONG-RANGE SEQUENCE MODELLING arxiv 19.11.13 paper
UNC Chapel Hill LXMERT LXMERT: Learning Cross-Modality Encoder Representations from Transformers arxiv 19.08.20 paper
OpenAI GPT-1 Improving language understanding with unsupervised learning OpenAI 2018 paper
OpenAI GPT-2 Language Models are Unsupervised Multitask Learners OpenAI 2019 paper
OpenAI GPT-3 Language Models are Few-Shot Learners OpenAI 2020 paper
FAIR FastText Advances in Pre-Training Distributed Word Representations arxiv 17.12.26 paper
FAIR XLM Cross-lingual Language Model Pretraining arxiv 19.01.22 paper
FAIR FSMT Facebook FAIR's WMT19 News Translation Task Submission arxiv 19.07.15 paper
FAIR RoBERTa RoBERTa: A Robustly Optimized BERT Pretraining Approach arxiv 19.07.26 paper
FAIR MMBT Supervised Multimodal Bitransformers for Classifying Images and Text arxiv 19.09.06 paper
FAIR BART BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension arxiv 19.10.29 paper
FAIR CamemBERT CamemBERT: a Tasty French Language Model arxiv 19.11.10 paper
FAIR mBART Multilingual Denoising Pre-training for Neural Machine Translation arxiv 20.01.22 paper
FAIR RAG Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks arxiv 20.05.22 paper
Hugging Face DistilBERT DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter arxiv 19.10.02 paper
Microsoft Marian Marian: Cost-effective High-Quality Neural Machine Translation in C++ ACL 2018 paper
Microsoft MT-DNN Multi-Task Deep Neural Networks for Natural Language Understanding arxiv 19.05.30 paper
Microsoft LayoutLM LayoutLM: Pre-training of Text and Layout for Document Image Understanding arxiv 19.12.31 paper
NVIDIA MegatronLM Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism arxiv 19.09.17 paper
Univ. of Washington Grover-Mega Defending Against Neural Fake News arxiv 19.10.29 paper
Carnegie Mellon GoogleBrain Transformer-XL Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context arxiv 19.06.02 paper
Carnegie Mellon GoogleBrain XLNet XLNet: Generalized Autoregressive Pretraining for Language Understanding arxiv 19.06.19 paper
Carnegie Mellon GoogleBrain Funnel Funnel-Transformer: Filtering out Sequential Redundancy for Efficient Language Processing arxiv 20.06.05 paper
Salesforce CTRL CTRL: A CONDITIONAL TRANSFORMER LANGUAGE MODEL FOR CONTROLLABLE GENERATION arxiv 19.09.11 paper
Anonymous authors MobileBERT MobileBERT: Task-Agnostic Compression of BERT by Progressive Knowledge Transfer ICLR 2020 paper

:sparkles: Attention Mechanism

:massage: Conversational AI

Memory-Based Research

Open-Domain

:art: Generative Model

GAN

:monkey_face: Meta Learning

Curiosity Algorithms

Road to General Intelligence

:brain: Reinforcement Learning

RL.start() 오늘의 논문 series

:chart_with_upwards_trend: Financial Mathematics & Engineer

:art: Neuromorphic

:cat2: Theoretical Deep Learning

Batch Normalization

Lipschitz gradient

Global Batch Normalization

Input Covariate Shift

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

How Does Batch Normalization Help Optimization?

Layer Normalization https://arxiv.org/abs/1607.06450

LeCun Initialization Efficient BackProp

Xavier initialization Understanding the difficulty of training deep feedforward neural networks

He Initialization Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

Nesterov Optimizer (Optimization류 논문들)

weight_standardization

:heart_eyes: Schmidhuber

Juergen Schmidhuber's Google Scholar

  • [x] Long short-term memory (Neural Computation 1997) paper
  • [ ] LSTM: A Search Space Odyssey (IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2017) paper
  • [x] Highway Networks (15.05.03, arxiv) paper
  • Full Paper: Training Very Deep Networks link
  • [x] Recurrent Highway Networks (ICML 2017) paper
  • [ ] Gradient flow in recurrent nets: the difficulty of learning long-term dependencies (IEEE 2001) paper paper
  • [ ] Bidirectional LSTM networks for improved phoneme classification and recognition (International Conference on Artificial Neural Networks 05.09.11)
  • [ ] Sequential neural text compression (IEEE 1996) paper
  • [ ] Neural expectation maximazation (NIPS 2017) paper
  • [ ] Accelerated Neural Evolution through Cooperatively Coevolved Synapses (JMLR 2008) paper
  • [ ] World Models (18.05.09, arxiv) paper

ETC

LSTM-SAE Unsupervised Pre-training of a Deep LSTM-based Stacked Autoencoder for Multivariate Time Series Forecasting Problems

C3D Learning Spatiotemporal Features with 3D Convolutional Networks

n-gram 관련 논문

Pointing the Unknown Words (몬트리홀 대학)

Seq2Seq Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation

Real-World Anomaly Detection in Surveillance Videos

self-attention on classification - A Structured Self-Attentive Sentence Embedding