AWQ (Activation-aware Weight Quantization)

A quantization technique that preserves model quality by considering activation outliers when quantizing weights.