https://ai.google.dev/edge/litert/conversion/tensorflow/quantization/quantization_spec
LiteRT 8-bit quantization specification | Google AI Edge | Google AI for Developers
google ai edge8 bitfor developerslitertquantization
https://huggingface.co/quantization
quantization (Quantization)
quantization
https://arxiv.org/abs/2310.17723
[2310.17723] ZeroQuant-HERO: Hardware-Enhanced Robust Optimized Post-Training Quantization...
Abstract page for arXiv paper 2310.17723: ZeroQuant-HERO: Hardware-Enhanced Robust Optimized Post-Training Quantization Framework for W8A8 Transformers
herohardwareenhancedrobustoptimized
https://30fps.net/pages/hyab-kmeans/
HyAB k-means for color quantization
meanscolorquantization
https://ai.google.dev/edge/litert/conversion/tensorflow/quantization/post_training_quantization
Post-training quantization | Google AI Edge | Google AI for Developers
google ai edgefor developersposttrainingquantization
https://ai.google.dev/edge/litert/conversion/tensorflow/quantization/post_training_float16_quant
Post-training float16 quantization | Google AI Edge | Google AI for Developers
google ai edgefor developersposttrainingquantization
https://ai.google.dev/edge/litert/conversion/tensorflow/quantization/quantization_debugger
Inspecting Quantization Errors with Quantization Debugger | Google AI Edge | Google AI for...
google ai edgequantizationerrorsdebugger
Sponsored https://wannahookup.com/
WannaHookUp - WannaHookUp
Join our online social adult community WannaHookUp
https://arxiv.org/abs/1812.09162
[1812.09162] Quicker ADC : Unlocking the hidden potential of Product Quantization with SIMD
Abstract page for arXiv paper 1812.09162: Quicker ADC : Unlocking the hidden potential of Product Quantization with SIMD
quickeradchiddenpotentialproduct
https://ai.google.dev/edge/litert/conversion/tensorflow/quantization/post_training_quant
Post-training dynamic range quantization | Google AI Edge | Google AI for Developers
google ai edgedynamic rangefor developersposttraining
Sponsored https://bellesaplus.co/
Join Bellesa Plus. The Netflix of Porn.
https://arxiv.org/abs/1908.10396
[1908.10396] Accelerating Large-Scale Inference with Anisotropic Vector Quantization
Abstract page for arXiv paper 1908.10396: Accelerating Large-Scale Inference with Anisotropic Vector Quantization
large scaleinferencevectorquantization
https://www.ibm.com/think/topics/quantization
What is Quantization? | IBM
Nov 17, 2025 - Quantization is the process of reducing the precision of a digital signal, typically from a higher-precision format to a lower-precision format.
what isquantizationibm
https://arxiv.org/abs/2206.01861
[2206.01861] ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale...
Abstract page for arXiv paper 2206.01861: ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
large scaleefficientaffordableposttraining
https://www.tensorflow.org/api_docs/python/tf/quantization/experimental/TfRecordRepresentativeDatasetSaver
tf.quantization.experimental.TfRecordRepresentativeDatasetSaver | TensorFlow v2.16.1
Representative dataset saver in TFRecord format.
tfquantizationexperimentaltensorflowv2
https://arxiv.org/abs/2303.08302
[2303.08302] ZeroQuant-V2: Exploring Post-training Quantization in LLMs from Comprehensive Study to...
Abstract page for arXiv paper 2303.08302: ZeroQuant-V2: Exploring Post-training Quantization in LLMs from Comprehensive Study to Low Rank Compensation
v2exploringposttrainingquantization
Sponsored https://www.cheekycrush.com/
CheekyCrush
https://llm-quantization-attack.org/
Exploiting LLM Quantization
We show that large language model quantization can be exploited to introduce malicious behavior (only) in quantized LLMs.
llmquantization
https://arxiv.org/abs/2308.13137
[2308.13137] OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models
Abstract page for arXiv paper 2308.13137: OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models
large language modelsquantization
https://www.tensorflow.org/api_docs/python/tf/quantization
Module: tf.quantization | TensorFlow v2.16.1
Public API for tf._api.v2.quantization namespace
moduletfquantizationtensorflowv2
https://ai.google.dev/edge/litert/conversion/tensorflow/quantization/post_training_integer_quant_16x8
Post-training integer quantization with int16 activations | Google AI Edge | Google AI for...
google ai edgeposttrainingintegerquantization
https://ai.google.dev/edge/litert/conversion/tensorflow/quantization/post_training_integer_quant
Post-training integer quantization | Google AI Edge | Google AI for Developers
google ai edgefor developersposttraininginteger
https://arxiv.org/abs/2301.12017
[2301.12017] Understanding INT4 Quantization for Transformer Models: Latency Speedup,...
Abstract page for arXiv paper 2301.12017: Understanding INT4 Quantization for Transformer Models: Latency Speedup, Composability, and Failure Cases
transformer modelsunderstandingquantizationlatency
https://pngquant.org/lib/
libimagequant (LIQ) — Image Quantization Library
imagequantizationlibrary
https://ngrok.com/blog/quantization
Quantization from the ground up | ngrok blog
Mar 25, 2026 - A complete guide to what quantization is, how it works, and how it's used to compress large language models
ground upquantizationngrokblog