OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models
트윗하기
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.