🏠 Home
Benchmark Hub
📊 All Benchmarks 🦖 Dinosaur v1 🦖 Dinosaur v2 ✅ To-Do List Applications 🎨 Creative Free Pages 🎯 FSACB - Ultimate Showcase 🌍 Translation Benchmark
Models
🏆 Top 10 Models 🆓 Free Models 📋 All Models ⚙️ Kilo Code
Resources
💬 Prompts Library 📖 AI Glossary 🔗 Useful Links

AI Glossary

The complete dictionary of Artificial Intelligence

162
categories
2,032
subcategories
23,060
terms
📖
terms

Extreme Quantization

Precision reduction technique pushing model parameters down to 1-2 bits for maximum compression, partially sacrificing accuracy for efficiency.

📖
terms

Binary Quantization

Quantization method where each weight and activation is represented by a single bit (-1 or +1), drastically reducing memory and accelerating computations.

📖
terms

Ternary Quantization

Technique using three values typically (-1, 0, +1) to represent weights, offering a better trade-off between compression and performance than pure binarization.

📖
terms

1-bit Quantization

Extreme form of quantization where each model parameter is stored on a single bit, enabling a 32x reduction compared to standard 32-bit models.

📖
terms

2-bit Quantization

Representation of weights and activations on two bits, allowing four quantization levels (-3, -1, +1, +3) with a better accuracy/efficiency balance.

📖
terms

Weight Binarization

Process of converting neural network weights into binary values while preserving activations at higher precision to maintain performance.

📖
terms

Post-Training Extreme Quantization

Technique applied after training to reduce parameter precision to 1-2 bits without requiring complete model retraining.

📖
terms

Extreme Quantization-Aware Quantization

Advanced method taking into account the impact of extreme quantization during the calibration process to minimize performance degradation.

📖
terms

Extreme Quantization with Extreme Learning

Approach where the model is fine-tuned specifically to adapt to extreme quantization constraints, better preserving final accuracy.

📖
terms

Binary Neural Network

Architecture where weights and activations are fully binarized, using XNOR and popcount operations for ultra-optimized computations.

📖
terms

Ternary Neural Network

Variant of binary networks using three states, allowing better expressivity while maintaining strong compression and computational efficiency.

📖
terms

Extreme Asymmetric Quantization

1-2 bit quantization method using asymmetric value ranges to better adapt to non-centered weight distributions.

📖
terms

Extreme Symmetric Quantization

Quantization approach where the value range is centered on zero, simplifying calculations but potentially less effective for certain distributions.

📖
terms

Model Compression via Extreme Quantization

Global technique combining extreme quantization with other compression methods to achieve compression rates exceeding 100x.

📖
terms

Minimal Precision Optimization

Process aiming to determine the minimum bit precision required for each layer of the model while maintaining acceptable performance levels.

📖
terms

Calibration for Extreme Quantization

Critical phase where quantization parameters are optimized using a small dataset to minimize the impact of extreme precision reduction.

📖
terms

Quantification Adaptative Extrême

Technique ajustant dynamiquement le niveau de quantification (1 ou 2 bits) par couche ou par neurone en fonction de leur sensibilité à la réduction de précision.

📖
terms

Stabilité de la Quantification Extrême

Propriété mesurant la robustesse d'un modèle face à la quantification extrême, essentielle pour garantir des performances fiables en déploiement.

🔍

No results found