🏠 Strona Główna
Benchmarki
📊 Wszystkie benchmarki 🦖 Dinozaur v1 🦖 Dinozaur v2 ✅ Aplikacje To-Do List 🎨 Kreatywne wolne strony 🎯 FSACB - Ostateczny pokaz 🌍 Benchmark tłumaczeń
Modele
🏆 Top 10 modeli 🆓 Darmowe modele 📋 Wszystkie modele ⚙️ Kilo Code
Zasoby
💬 Biblioteka promptów 📖 Słownik AI 🔗 Przydatne linki

Słownik AI

Kompletny słownik sztucznej inteligencji

162
kategorie
2 032
podkategorie
23 060
pojęcia
📖
pojęcia

Quantization

Process of reducing the numerical precision of AI model weights and activations to optimize inference and reduce memory footprint.

📖
pojęcia

8-bit Quantization

Compression technique reducing model weights from 32 bits to 8 bits, offering an optimal trade-off between performance and accuracy for LLMs.

📖
pojęcia

4-bit Quantization

Extreme compression method reducing weights to 4 bits, allowing significant memory gains but with potential quality loss.

📖
pojęcia

Post-Training Quantization (PTQ)

Technique applied after model training, converting weights to reduced precision without requiring full retraining.

📖
pojęcia

Quantization Aware Training (QAT)

Training approach simulating quantization effects during the learning process to minimize accuracy loss.

📖
pojęcia

Dynamic Quantization

Method applied during inference where activations are quantized on-the-fly, offering flexibility but with computational overhead.

📖
pojęcia

Static Quantization

Approach precomputing quantization parameters before inference, optimizing speed at the expense of flexibility.

📖
pojęcia

Quantization Calibration

Process of determining optimal quantization parameters (scale, zero-point) from a sample of representative data.

📖
pojęcia

GPTQ

Gradient-based Post Training Quantization, an advanced technique that iteratively optimizes quantized weights to minimize reconstruction error.

📖
pojęcia

AWQ

Activation-aware Weight Quantization, a method that weights the importance of weights according to the amplitude of corresponding activations.

📖
pojęcia

Zero-shot Quantification

Technique requiring no calibration data, using heuristics based on weight distribution to quantify the model.

📖
pojęcia

Mixed Precision Quantification

Strategy applying different quantification precisions according to model layers to optimize the performance/accuracy trade-off.

📖
pojęcia

Symmetric Quantification

Quantification scheme where the value range is centered around zero, simplifying calculations but potentially underutilizing the dynamic range.

📖
pojęcia

Asymmetric Quantification

Approach allowing value ranges not centered on zero, optimizing the use of the quantized range for asymmetric distributions.

📖
pojęcia

Scale Factor

Multiplicative parameter used to map continuous values into the quantized range, crucial for quantification accuracy.

📖
pojęcia

Zero Point

Offset added during asymmetric quantification to align the floating-point zero value with the quantized representation.

📖
pojęcia

Quantization Noise

Error introduced by precision reduction, manifesting as model performance degradation due to weight approximation.

📖
pojęcia

Quantization-aware Fine-tuning

Post-quantization fine-tuning process aimed at recovering accuracy lost during model compression.

📖
pojęcia

SmoothQuant

Quantization technique equalizing quantization difficulty between weights and activations through prior mathematical transformation.

📖
pojęcia

LLM.int8()

Specific 8-bit quantization method for large language models, combining matrix decomposition and hybrid quantization.

🔍

Nie znaleziono wyników