🏠 Home
Benchmark
📊 Tutti i benchmark 🦖 Dinosauro v1 🦖 Dinosauro v2 ✅ App To-Do List 🎨 Pagine libere creative 🎯 FSACB - Ultimate Showcase 🌍 Benchmark traduzione
Modelli
🏆 Top 10 modelli 🆓 Modelli gratuiti 📋 Tutti i modelli ⚙️ Kilo Code
Risorse
💬 Libreria di prompt 📖 Glossario IA 🔗 Link utili

Glossario IA

Il dizionario completo dell'Intelligenza Artificiale

162
categorie
2.032
sottocategorie
23.060
termini
📖
termini

Additive Attention

Proposed by Bahdanau, this method combines the decoder's hidden state and encoder outputs through a feed-forward network to calculate attention weights.

📖
termini

Multiplicative Attention

Introduced by Luong, calculates attention scores by dot product between the decoder state and encoder outputs, offering a more efficient implementation.

📖
termini

Multi-Head Attention

Extension of self-attention using multiple attention heads in parallel to capture different types of relationships in the data.

📖
termini

Context Vector

Weighted representation of encoder outputs, calculated using attention weights and provided to the decoder as contextual information.

📖
termini

Scaled Dot-Product Attention

Attention variant used in Transformers where the dot product is divided by the square root of the dimension to stabilize training.

📖
termini

Global Attention

Attention mechanism considering all positions of the source sequence to calculate the context vector at each decoding step.

📖
termini

Local Attention

Attention variant considering only a subset of predicted positions around a central position, reducing computational complexity.

📖
termini

Hierarchical Attention

Multi-level architecture applying attention at different granularities, first at the word level then at the sentence or document level.

📖
termini

Query, Key, Value

Triple of fundamental vectors in attention: Query represents the current request, Key the available keys, and Value the values to retrieve.

📖
termini

Temporal Attention

Mechanism specialized in capturing temporal dependencies in time series by weighting relevant time steps.

📖
termini

Spatial Attention

Application of attention to spatial data (images, videos) to focus on the most informative regions in space.

📖
termini

Adaptive Attention

Approach where the attention mechanism dynamically adjusts during training to optimize its parameters according to the task.

📖
termini

Sparse Attention

Attention variant that computes weights only for a subset of positions, enabling efficient processing of longer sequences.

📖
termini

Attention Mask

Technique that masks certain positions to prevent attention on irrelevant tokens such as padding or future tokens.

📖
termini

Linear Attention

Approximation of standard attention with linear complexity rather than quadratic, enabling processing of much longer sequences.

📖
termini

Performer Attention

Variant using feature mapping kernels to approximate attention with efficient linear complexity in memory and computation.

🔍

Nessun risultato trovato