🏠 Beranda
Benchmark
📊 Semua Benchmark 🦖 Dinosaurus v1 🦖 Dinosaurus v2 ✅ Aplikasi To-Do List 🎨 Halaman Bebas Kreatif 🎯 FSACB - Showcase Utama 🌍 Benchmark Terjemahan
Model
🏆 Top 10 Model 🆓 Model Gratis 📋 Semua Model ⚙️ Kilo Code
Sumber Daya
💬 Perpustakaan Prompt 📖 Glosarium AI 🔗 Tautan Berguna

Glosarium AI

Kamus lengkap Kecerdasan Buatan

162
kategori
2.032
subkategori
23.060
istilah
📖
istilah

Quantization

Process of reducing the numerical precision of AI model weights and activations to optimize inference and reduce memory footprint.

📖
istilah

8-bit Quantization

Compression technique reducing model weights from 32 bits to 8 bits, offering an optimal trade-off between performance and accuracy for LLMs.

📖
istilah

4-bit Quantization

Extreme compression method reducing weights to 4 bits, allowing significant memory gains but with potential quality loss.

📖
istilah

Post-Training Quantization (PTQ)

Technique applied after model training, converting weights to reduced precision without requiring full retraining.

📖
istilah

Quantization Aware Training (QAT)

Training approach simulating quantization effects during the learning process to minimize accuracy loss.

📖
istilah

Dynamic Quantization

Method applied during inference where activations are quantized on-the-fly, offering flexibility but with computational overhead.

📖
istilah

Static Quantization

Approach precomputing quantization parameters before inference, optimizing speed at the expense of flexibility.

📖
istilah

Quantization Calibration

Process of determining optimal quantization parameters (scale, zero-point) from a sample of representative data.

📖
istilah

GPTQ

Gradient-based Post Training Quantization, an advanced technique that iteratively optimizes quantized weights to minimize reconstruction error.

📖
istilah

AWQ

Activation-aware Weight Quantization, a method that weights the importance of weights according to the amplitude of corresponding activations.

📖
istilah

Zero-shot Quantification

Technique requiring no calibration data, using heuristics based on weight distribution to quantify the model.

📖
istilah

Mixed Precision Quantification

Strategy applying different quantification precisions according to model layers to optimize the performance/accuracy trade-off.

📖
istilah

Symmetric Quantification

Quantification scheme where the value range is centered around zero, simplifying calculations but potentially underutilizing the dynamic range.

📖
istilah

Asymmetric Quantification

Approach allowing value ranges not centered on zero, optimizing the use of the quantized range for asymmetric distributions.

📖
istilah

Scale Factor

Multiplicative parameter used to map continuous values into the quantized range, crucial for quantification accuracy.

📖
istilah

Zero Point

Offset added during asymmetric quantification to align the floating-point zero value with the quantized representation.

📖
istilah

Quantization Noise

Error introduced by precision reduction, manifesting as model performance degradation due to weight approximation.

📖
istilah

Quantization-aware Fine-tuning

Post-quantization fine-tuning process aimed at recovering accuracy lost during model compression.

📖
istilah

SmoothQuant

Quantization technique equalizing quantization difficulty between weights and activations through prior mathematical transformation.

📖
istilah

LLM.int8()

Specific 8-bit quantization method for large language models, combining matrix decomposition and hybrid quantization.

🔍

Tidak ada hasil ditemukan