Sihwa Lee
Sihwa Lee
Verified email at
Cited by
Cited by
NN-LUT: Neural approximation of non-linear operations for efficient transformer inference
J Yu, J Park, S Park, M Kim, S Lee, DH Lee, J Choi
Proceedings of the 59th ACM/IEEE Design Automation Conference, 577-582, 2022
Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
M Kim, S Lee, S Hong, DS Chang, J Choi
Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022
Token-scaled logit distillation for ternary weight generative language models
M Kim, S Lee, J Lee, S Hong, DS Chang, W Sung, J Choi
Advances in Neural Information Processing Systems 36, 2024
The system can't perform the operation now. Try again later.
Articles 1–3