Skip to content

Instantly share code, notes, and snippets.

@vahbuna
Created April 5, 2023 23:05
Show Gist options
  • Save vahbuna/d6b33b06edf63fbc451851d3ef40e342 to your computer and use it in GitHub Desktop.
Save vahbuna/d6b33b06edf63fbc451851d3ef40e342 to your computer and use it in GitHub Desktop.
references from Learning with not Enough Data - Lilian Weng

Semi-Supervised Learning

[1] Ouali, Hudelot & Tami. “An Overview of Deep Semi-Supervised Learning” arXiv preprint arXiv:2006.05278 (2020).

[2] Sajjadi, Javanmardi & Tasdizen “Regularization With Stochastic Transformations and Perturbations for Deep Semi-Supervised Learning.” arXiv preprint arXiv:1606.04586 (2016).

[3] Pham et al. “Meta Pseudo Labels.” CVPR 2021.

[4] Laine & Aila. “Temporal Ensembling for Semi-Supervised Learning” ICLR 2017.

[5] Tarvaninen & Valpola. “Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results.” NeuriPS 2017

[6] Xie et al. “Unsupervised Data Augmentation for Consistency Training.” NeuriPS 2020.

[7] Miyato et al. “Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning.” IEEE transactions on pattern analysis and machine intelligence 41.8 (2018).

[8] Verma et al. “Interpolation consistency training for semi-supervised learning.” IJCAI 2019

[9] Lee. “Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks.” ICML 2013 Workshop: Challenges in Representation Learning.

[10] Iscen et al. “Label propagation for deep semi-supervised learning.” CVPR 2019.

[11] Xie et al. “Self-training with Noisy Student improves ImageNet classification” CVPR 2020.

[12] Jingfei Du et al. “Self-training Improves Pre-training for Natural Language Understanding.” 2020

[13] Iscen et al. “Label propagation for deep semi-supervised learning.” CVPR 2019

[14] Arazo et al. “Pseudo-labeling and confirmation bias in deep semi-supervised learning.” IJCNN 2020.

[15] Berthelot et al. “MixMatch: A holistic approach to semi-supervised learning.” NeuriPS 2019

[16] Berthelot et al. “ReMixMatch: Semi-supervised learning with distribution alignment and augmentation anchoring.” ICLR 2020

[17] Sohn et al. “FixMatch: Simplifying semi-supervised learning with consistency and confidence.” CVPR 2020

[18] Junnan Li et al. “DivideMix: Learning with Noisy Labels as Semi-supervised Learning.” 2020 [code]

[19] Zoph et al. “Rethinking pre-training and self-training.” 2020.

[20] Chen et al. “Big Self-Supervised Models are Strong Semi-Supervised Learners” 2020

Active Learning

[1] Burr Settles. Active learning literature survey. University of Wisconsin, Madison, 52(55-66):11, 2010.

[2] https://jacobgil.github.io/deeplearning/activelearning

[3] Yang et al. “Cost-effective active learning for deep image classification” TCSVT 2016.

[4] Yarin Gal et al. “Dropout as a Bayesian Approximation: representing model uncertainty in deep learning." ICML 2016.

[5] Blundell et al. “Weight uncertainty in neural networks (Bayes-by-Backprop)" ICML 2015.

[6] Settles et al. “Multiple-Instance Active Learning." NIPS 2007.

[7] Houlsby et al. Bayesian Active Learning for Classification and Preference Learning." arXiv preprint arXiv:1112.5745 (2020).

[8] Kirsch et al. “BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning." NeurIPS 2019.

[9] Beluch et al. “The power of ensembles for active learning in image classification." CVPR 2018.

[10] Sener & Savarese. “Active learning for convolutional neural networks: A core-set approach." ICLR 2018.

[11] Donggeun Yoo & In So Kweon. “Learning Loss for Active Learning." CVPR 2019.

[12] Margatina et al. “Active Learning by Acquiring Contrastive Examples." EMNLP 2021.

[13] Sinha et al. “Variational Adversarial Active Learning” ICCV 2019

[14] Ebrahimiet al. “Minmax Active Learning” arXiv preprint arXiv:2012.10467 (2021).

[15] Mariya Toneva et al. “An empirical study of example forgetting during deep neural network learning." ICLR 2019.

[16] Javad Zolfaghari Bengar et al. “When Deep Learners Change Their Mind: Learning Dynamics for Active Learning." CAIP 2021.

[17] Yang et al. “Suggestive annotation: A deep active learning framework for biomedical image segmentation." MICCAI 2017.

[18] Fedor Zhdanov. “Diverse mini-batch Active Learning” arXiv preprint arXiv:1901.05954 (2019).

Data Augmentation

[1] Zhang et al. “Adversarial AutoAgument” ICLR 2020.

[2] Kumar et al. “Data Augmentation using Pre-trained Transformer Models." AACL 2020 Workshop.

[3] Anaby-Tavor et al. “Not enough data? Deep learning to rescue!" AAAI 2020.

[4] Wang et al. “Want To Reduce Labeling Cost? GPT-3 Can Help." EMNLP 2021.

[5] Wang et al. “Towards Zero-Label Language Learning." arXiv preprint arXiv:2109.09193 (2021).

[6] Schick & Schutze. Generating Datasets with Pretrained Language Models." EMNLP 2021.

[7] Han et al. “Unsupervised Neural Machine Translation with Generative Language Models Only." arXiv preprint arXiv:2110.05448 (2021).

[8] Guo et al. “Augmenting data with mixup for sentence classification: An empirical study." arXiv preprint arXiv:1905.08941 (2019).

[9] Ekin D. Cubuk et al. “AutoAugment: Learning augmentation policies from data." arXiv preprint arXiv:1805.09501 (2018).

[10] Daniel Ho et al. “Population Based Augmentation: Efficient Learning of Augmentation Policy Schedules." ICML 2019.

[11] Cubuk & Zoph et al. “RandAugment: Practical automated data augmentation with a reduced search space." arXiv preprint arXiv:1909.13719 (2019).

[12] Zhang et al. “mixup: Beyond Empirical Risk Minimization." ICLR 2017.

[13] Yun et al. “CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features." ICCV 2019.

[14] Kalantidis et al. “Mixing of Contrastive Hard Negatives” NeuriPS 2020.

[15] Wei & Zou. “EDA: Easy data augmentation techniques for boosting performance on text classification tasks." EMNLP-IJCNLP 2019.

[16] Kobayashi. “Contextual Augmentation: Data Augmentation by Words with Paradigmatic Relations." NAACL 2018

[17] Fang et al. “CERT: Contrastive self-supervised learning for language understanding." arXiv preprint arXiv:2005.12766 (2020).

[18] Gao et al. “SimCSE: Simple Contrastive Learning of Sentence Embeddings." arXiv preprint arXiv:2104.08821 (2020). [code]

[19] Shen et al. “A Simple but Tough-to-Beat Data Augmentation Approach for Natural Language Understanding and Generation." arXiv preprint arXiv:2009.13818 (2020) [code]

[20] Wang & van den Oord. “Multi-Format Contrastive Learning of Audio Representations." NeuriPS Workshop 2020.

[21] Wu et al. “Conditional BERT Contextual Augmentation” arXiv preprint arXiv:1812.06705 (2018).

[22 Zhu et al. “FreeLB: Enhanced Adversarial Training for Natural Language Understanding." ICLR 2020.

[23] Affinity and Diversity: Quantifying Mechanisms of Data Augmentation Gontijo-Lopes et al. 2020 (https://arxiv.org/abs/2002.08973)

[24] Song et al. “Learning from Noisy Labels with Deep Neural Networks: A Survey." TNNLS 2020.

[25] Zhang & Sabuncu. “Generalized cross entropy loss for training deep neural networks with noisy labels." NeuriPS 2018.

[26] Goldberger & Ben-Reuven. “Training deep neural-networks using a noise adaptation layer." ICLR 2017.

[27] Sukhbaatar et al. “Training convolutional networks with noisy labels." ICLR Workshop 2015.

[28] Patrini et al. “Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach” CVPR 2017.

[29] Hendrycks et al. “Using trusted data to train deep networks on labels corrupted by severe noise." NeuriPS 2018.

[30] Zhang & Sabuncu. “Generalized cross entropy loss for training deep neural networks with noisy labels." NeuriPS 2018.

[31] Lyu & Tsang. “Curriculum loss: Robust learning and generalization against label corruption." ICLR 2020.

[32] Han et al. “Co-teaching: Robust training of deep neural networks with extremely noisy labels." NeuriPS 2018. (code)

[33] Ren et al. “Learning to reweight examples for robust deep learning." ICML 2018.

[34] Jiang et al. “MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels." ICML 2018.

[35] Li et al. “Learning from noisy labels with distillation." ICCV 2017.

[36] Liu & Tao. “Classification with noisy labels by importance reweighting." TPAMI 2015.

[37] Ghosh, et al. “Robust loss functions under label noise for deep neural networks." AAAI 2017.

[38] Hu et al. “Does Distributionally Robust Supervised Learning Give Robust Classifiers? “ ICML 2018.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment