Mobile QR Code
Title Layer-adaptive Gradient Clipping in Quantization-Aware Training
Authors 박지훈(Jihoon Park) ; 송병철(Byung Cheol Song)
DOI https://doi.org/10.5573/ieie.2023.60.8.37
Page pp.37-40
ISSN 2287-5026
Keywords Quantization-Aware training; Gradient clipping; Layer-adaptive
Abstract Quantization in neural network folds the batch normalization layer, so the normalization effect cannot be used and accuracy can be degraded. To avoid the accuracy degradation, this paper achieves stabilized training by applying Adaptive Gradient Clipping (AGC) to Quantization-Aware Training (QAT). Also, we propose a neural network with MLP that takes random clipping vector as inputs for gradient clipping adaptively for each layer.