Hayeon Lee
Hayeon Lee
Postdoctoral Researcher, Meta AI (FAIR)
Verified email at - Homepage
Cited by
Cited by
Learning to balance: Bayesian meta-learning for imbalanced and out-of-distribution tasks
HB Lee*, H Lee*, D Na*, S Kim, M Park, E Yang, SJ Hwang
ICLR 2020, 2019
Help: Hardware-adaptive efficient latency prediction for nas via meta-learning
H Lee, S Lee, S Chong, SJ Hwang
arXiv preprint arXiv:2106.08630, 2021
Rapid neural architecture search by learning to generate graphs from datasets
H Lee*, E Hyung*, SJ Hwang
ICLR 2021, 2021
A survey of supernet optimization and its applications: Spatial and temporal optimization for neural architecture search
S Cha, T Kim, H Lee, SY Yun
arXiv preprint arXiv:2204.03916, 2022
Task-Adaptive Neural Network Search with Meta-Contrastive Learning
W Jeong, H Lee, G Park, E Hyung, J Baek, SJ Hwang
Advances in Neural Information Processing Systems 34, 21310-21324, 2021
Online Hyperparameter Meta-Learning with Hypergradient Distillation
HB Lee, H Lee, J Shin, E Yang, T Hospedales, SJ Hwang
arXiv preprint arXiv:2110.02508, 2021
DiffusionNAG: Predictor-guided Neural Architecture Generation with Diffusion Models
S An, H Lee, J Jo, S Lee, SJ Hwang
arXiv preprint arXiv:2305.16943, 2023
Meta-prediction Model for Distillation-Aware NAS on Unseen Datasets
H Lee, S An, M Kim, SJ Hwang
The Eleventh International Conference on Learning Representations, 0
A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models
H Lee, R Hou, J Kim, D Liang, SJ Hwang, A Min
arXiv preprint arXiv:2305.18239, 2023
Lightweight Neural Architecture Search with Parameter Remapping and Knowledge Distillation
H Lee, S An, M Kim, SJ Hwang
First Conference on Automated Machine Learning (Late-Breaking Workshop), 2022
Co-training and Co-distillation for Quality Improvement and Compression of Language Models
H Lee, R Hou, J Kim, D Liang, H Zhang, SJ Hwang, A Min
arXiv preprint arXiv:2311.02849, 2023
The system can't perform the operation now. Try again later.
Articles 1–11