🏠 Startseite
Vergleiche
📊 Alle Benchmarks 🦖 Dinosaurier v1 🦖 Dinosaurier v2 ✅ To-Do-Listen-Apps 🎨 Kreative freie Seiten 🎯 FSACB - Ultimatives Showcase 🌍 Übersetzungs-Benchmark
Modelle
🏆 Top 10 Modelle 🆓 Kostenlose Modelle 📋 Alle Modelle ⚙️ Kilo Code
Ressourcen
💬 Prompt-Bibliothek 📖 KI-Glossar 🔗 Nützliche Links

KI-Glossar

Das vollständige Wörterbuch der Künstlichen Intelligenz

162
Kategorien
2.032
Unterkategorien
23.060
Begriffe
📖
Begriffe

Quantization

Process of reducing the numerical precision of AI model weights and activations to optimize inference and reduce memory footprint.

📖
Begriffe

8-bit Quantization

Compression technique reducing model weights from 32 bits to 8 bits, offering an optimal trade-off between performance and accuracy for LLMs.

📖
Begriffe

4-bit Quantization

Extreme compression method reducing weights to 4 bits, allowing significant memory gains but with potential quality loss.

📖
Begriffe

Post-Training Quantization (PTQ)

Technique applied after model training, converting weights to reduced precision without requiring full retraining.

📖
Begriffe

Quantization Aware Training (QAT)

Training approach simulating quantization effects during the learning process to minimize accuracy loss.

📖
Begriffe

Dynamic Quantization

Method applied during inference where activations are quantized on-the-fly, offering flexibility but with computational overhead.

📖
Begriffe

Static Quantization

Approach precomputing quantization parameters before inference, optimizing speed at the expense of flexibility.

📖
Begriffe

Quantization Calibration

Process of determining optimal quantization parameters (scale, zero-point) from a sample of representative data.

📖
Begriffe

GPTQ

Gradient-based Post Training Quantization, an advanced technique that iteratively optimizes quantized weights to minimize reconstruction error.

📖
Begriffe

AWQ

Activation-aware Weight Quantization, a method that weights the importance of weights according to the amplitude of corresponding activations.

📖
Begriffe

Zero-shot Quantification

Technique requiring no calibration data, using heuristics based on weight distribution to quantify the model.

📖
Begriffe

Mixed Precision Quantification

Strategy applying different quantification precisions according to model layers to optimize the performance/accuracy trade-off.

📖
Begriffe

Symmetric Quantification

Quantification scheme where the value range is centered around zero, simplifying calculations but potentially underutilizing the dynamic range.

📖
Begriffe

Asymmetric Quantification

Approach allowing value ranges not centered on zero, optimizing the use of the quantized range for asymmetric distributions.

📖
Begriffe

Scale Factor

Multiplicative parameter used to map continuous values into the quantized range, crucial for quantification accuracy.

📖
Begriffe

Zero Point

Offset added during asymmetric quantification to align the floating-point zero value with the quantized representation.

📖
Begriffe

Quantization Noise

Error introduced by precision reduction, manifesting as model performance degradation due to weight approximation.

📖
Begriffe

Quantization-aware Fine-tuning

Post-quantization fine-tuning process aimed at recovering accuracy lost during model compression.

📖
Begriffe

SmoothQuant

Quantization technique equalizing quantization difficulty between weights and activations through prior mathematical transformation.

📖
Begriffe

LLM.int8()

Specific 8-bit quantization method for large language models, combining matrix decomposition and hybrid quantization.

🔍

Keine Ergebnisse gefunden