sjquan / 2022-Study

56 stars 8 forks source link

[11/29] 김민수, Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders #12

Open MarsJacobs opened 1 year ago

MarsJacobs commented 1 year ago

Date

Who

Keywords

What

Preliminary

Transformer 계열 모델 (주로 Encoder) 에 QAT 가 적용된 최신 연구들을 차례차례 살펴볼 예정입니다.