- [2011 JMLR] Natural Language Processing (Almost) from Scratch, cover Tagging, Chunking, Parsing, NER, SRL and etc. tasks, [paper], [bibtex], sources: [attardi/deepnl].
- [2017 ArXiv] An Overview of Multi-Task Learning in Deep Neural Networks, [paper], [bibtex].
- [2017 EMNLP] A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks, cover Tagging, Chunking, Parsing, Relatedness, Entailment tasks, [paper], [bibtex], [blog], sources: [rubythonode/joint-many-task-model], [hassyGo/charNgram2vec].
- [2018 ICLR] Bi-Directional Block Self-Attention for Fast and Memory-Efficient Sequence Modeling, [paper], [bibtex], sources: [taoshen58/BiBloSA].
- [2018 CoNLL] Sequence Classification with Human Attention, [paper], [bibtex], sources: [coastalcph/Sequence_classification_with_human_attention].
- [2019 ArXiv] Multi-Task Deep Neural Networks for Natural Language Understanding, [paper], [bibtex].
- [2019 ArXiv] BERT and PALs: Projected Attention Layers for Efficient Adaptation in Multi-Task Learning, [paper], [bibtex].
- [2019 AAAI] Latent Multi-task Architecture Learning, [paper], [bibtex], sources: [sebastianruder/sluice-networks].