Update docs/source/conceptual/quantization.md

Co-authored-by: Nicolas Patry <patry.nicolas@protonmail.com>
This commit is contained in:
Merve Noyan 2023-08-25 11:57:14 +03:00 committed by GitHub
parent 937e4269e1
commit 764d946607
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -41,7 +41,7 @@ You can learn more about GPTQ from the [paper](https://arxiv.org/pdf/2210.17323.
bitsandbytes is a library used to apply 8-bit and 4-bit quantization to models. It can be used during training for mixed-precision training or before inference to make the model smaller. bitsandbytes is a library used to apply 8-bit and 4-bit quantization to models. It can be used during training for mixed-precision training or before inference to make the model smaller.
8-bit quantization enables multi-billion parameter scale models to fit in smaller hardware without degrading performance. 8bit quantization works as follows 👇 8-bit quantization enables multi-billion parameter scale models to fit in smaller hardware without degrading performance too much. 8bit quantization works as follows 👇
1. Extract the larger values (outliers) columnwise from the input hidden states. 1. Extract the larger values (outliers) columnwise from the input hidden states.
2. Perform the matrix multiplication of the outliers in FP16 and the non-outliers in int8. 2. Perform the matrix multiplication of the outliers in FP16 and the non-outliers in int8.