Showing results 1 to 4 of 4
Co-training and Co-distillation for Quality Improvement and Compression of Language Models Lee, Hayeon; Hou, Rui; Kim, Jongphil; Liang, Davis; Zhang, Hongbo; Hwang, Sung Ju; Min, Alexander, Empirical Methods in Natural Language Processing, EMNLP 2023, Association of Computational Linguistics, 2023-12-06 |
Meta-prediction Model for Distillation-Aware NAS on Unseen Datasets Lee, Hayeon; An, Sohyung; Kim, Minseon; Hwang, Sung Ju, The Eleventh International Conference on Learning Representations, ICLR 2023, International Conference on Learning Representations, 2023-05-01 |
Online Hyperparameter Meta-Learning with Hypergradient Distillation Lee, Hae Beom; Lee, Hayeon; Shin, Jae Woong; Yang, Eunho; Hospedales, Timothy; Hwang, Sung Ju, 10th International Conference on Learning Representations, ICLR 2022, International Conference on Learning Representations, ICLR, 2022-04-25 |
Task-Adaptive Neural Network Search with Meta-Contrastive Learning Jeong, Wonyong; Lee, Hayeon; Park, Geon; Hyung, Eunyoung; Baek, Jinheon; Hwang, Sung Ju, 35th Conference on Neural Information Processing Systems (NeurIPS) 2021, Conference on Neural Information Processing Systems (NeurIPS), 2021-12-09 |
Discover