🏠 Beranda
Benchmark
📊 Semua Benchmark 🦖 Dinosaurus v1 🦖 Dinosaurus v2 ✅ Aplikasi To-Do List 🎨 Halaman Bebas Kreatif 🎯 FSACB - Showcase Utama 🌍 Benchmark Terjemahan
Model
🏆 Top 10 Model 🆓 Model Gratis 📋 Semua Model ⚙️ Kilo Code
Sumber Daya
💬 Perpustakaan Prompt 📖 Glosarium AI 🔗 Tautan Berguna

Glosarium AI

Kamus lengkap Kecerdasan Buatan

162
kategori
2.032
subkategori
23.060
istilah
📖
istilah

Additive Attention

Proposed by Bahdanau, this method combines the decoder's hidden state and encoder outputs through a feed-forward network to calculate attention weights.

📖
istilah

Multiplicative Attention

Introduced by Luong, calculates attention scores by dot product between the decoder state and encoder outputs, offering a more efficient implementation.

📖
istilah

Multi-Head Attention

Extension of self-attention using multiple attention heads in parallel to capture different types of relationships in the data.

📖
istilah

Context Vector

Weighted representation of encoder outputs, calculated using attention weights and provided to the decoder as contextual information.

📖
istilah

Scaled Dot-Product Attention

Attention variant used in Transformers where the dot product is divided by the square root of the dimension to stabilize training.

📖
istilah

Global Attention

Attention mechanism considering all positions of the source sequence to calculate the context vector at each decoding step.

📖
istilah

Local Attention

Attention variant considering only a subset of predicted positions around a central position, reducing computational complexity.

📖
istilah

Hierarchical Attention

Multi-level architecture applying attention at different granularities, first at the word level then at the sentence or document level.

📖
istilah

Query, Key, Value

Triple of fundamental vectors in attention: Query represents the current request, Key the available keys, and Value the values to retrieve.

📖
istilah

Temporal Attention

Mechanism specialized in capturing temporal dependencies in time series by weighting relevant time steps.

📖
istilah

Spatial Attention

Application of attention to spatial data (images, videos) to focus on the most informative regions in space.

📖
istilah

Adaptive Attention

Approach where the attention mechanism dynamically adjusts during training to optimize its parameters according to the task.

📖
istilah

Sparse Attention

Attention variant that computes weights only for a subset of positions, enabling efficient processing of longer sequences.

📖
istilah

Attention Mask

Technique that masks certain positions to prevent attention on irrelevant tokens such as padding or future tokens.

📖
istilah

Linear Attention

Approximation of standard attention with linear complexity rather than quadratic, enabling processing of much longer sequences.

📖
istilah

Performer Attention

Variant using feature mapping kernels to approximate attention with efficient linear complexity in memory and computation.

🔍

Tidak ada hasil ditemukan