🏠 Startseite
Vergleiche
📊 Alle Benchmarks 🦖 Dinosaurier v1 🦖 Dinosaurier v2 ✅ To-Do-Listen-Apps 🎨 Kreative freie Seiten 🎯 FSACB - Ultimatives Showcase 🌍 Übersetzungs-Benchmark
Modelle
🏆 Top 10 Modelle 🆓 Kostenlose Modelle 📋 Alle Modelle ⚙️ Kilo Code
Ressourcen
💬 Prompt-Bibliothek 📖 KI-Glossar 🔗 Nützliche Links

KI-Glossar

Das vollständige Wörterbuch der Künstlichen Intelligenz

162
Kategorien
2.032
Unterkategorien
23.060
Begriffe
📖
Begriffe

Textual Adversarial Attack

Technique consisting of subtly modifying an input text to mislead an NLP model while preserving semantics for a human reader.

📖
Begriffe

Character-Level Perturbation

Modification of individual characters in text (insertion, deletion, substitution) to create adversarial examples that are difficult to detect.

📖
Begriffe

Lexical Substitution Attack

Replacement of words with semantically close synonyms that change the NLP model's prediction in a targeted manner.

📖
Begriffe

Universal Adversarial Triggers

Specific sequences of words or characters that, when inserted into any text, systematically cause a classification error by the model.

📖
Begriffe

Black-Box Attack

Attack conducted without knowledge of the model's internal parameters, using only the model's predictions to construct adversarial examples.

📖
Begriffe

White-Box Attack

Attack exploiting complete knowledge of the model's architecture and gradients to generate optimal perturbations.

📖
Begriffe

Transfer Attack

Generation of adversarial examples on a source model that retain their effectiveness on unknown target models.

📖
Begriffe

Semantic Preservation

Constraint ensuring that textual perturbations do not alter the overall meaning of the text for a human reader.

📖
Begriffe

Data Poisoning Attack

Malicious insertion of corrupted examples into the training set to degrade model performance during its learning phase.

📖
Begriffe

Syntactic Perturbation

Modification of the grammatical or syntactic structure of a sentence while preserving its semantic meaning to deceive NLP models.

📖
Begriffe

Gradient Masking

Defense technique that modifies the model's gradient to prevent optimization-based attacks, without necessarily improving actual robustness.

📖
Begriffe

Query Attack

Black-box attack that optimizes perturbations by iteratively querying the model and analyzing its responses.

📖
Begriffe

Semantic Robustness

Ability of an NLP model to maintain consistent predictions in the face of textual variations preserving meaning but altering form.

📖
Begriffe

Adversarial Search Space

Set of all possible text modifications that can be applied to generate valid adversarial examples.

📖
Begriffe

Perturbation Score

Quantitative metric evaluating the magnitude of modification applied to the original text to create an adversarial example.

📖
Begriffe

Multi-objective Attack

Adversarial attack seeking simultaneously to deceive the model while optimizing multiple constraints such as readability or semantic preservation.

📖
Begriffe

Adversarial Attack Detection

Defensive mechanism identifying potentially adversarial inputs based on statistical or behavioral anomalies in predictions.

🔍

Keine Ergebnisse gefunden