🏠 Startseite
Vergleiche
📊 Alle Benchmarks 🦖 Dinosaurier v1 🦖 Dinosaurier v2 ✅ To-Do-Listen-Apps 🎨 Kreative freie Seiten 🎯 FSACB - Ultimatives Showcase 🌍 Übersetzungs-Benchmark
Modelle
🏆 Top 10 Modelle 🆓 Kostenlose Modelle 📋 Alle Modelle ⚙️ Kilo Code
Ressourcen
💬 Prompt-Bibliothek 📖 KI-Glossar 🔗 Nützliche Links

KI-Glossar

Das vollständige Wörterbuch der Künstlichen Intelligenz

162
Kategorien
2.032
Unterkategorien
23.060
Begriffe
📖
Begriffe

PCIe Bandwidth

Maximum data transfer rate through the PCIe bus, crucial for communication speed between CPU and GPU in AI workloads.

📖
Begriffe

NVMe over Fabrics

Protocol allowing access to NVMe storage over a network, reducing latency for massive datasets in AI.

📖
Begriffe

GPUDirect Storage

NVIDIA technology enabling direct data transfer from storage to GPU memory, bypassing the CPU and RAM.

📖
Begriffe

Memory Pinning

Process of locking memory pages in RAM to ensure continuous and fast DMA (Direct Memory Access) by the GPU.

📖
Begriffe

Zero-Copy

Optimization technique where data is transferred directly between devices without intermediate copying in CPU memory.

📖
Begriffe

Tensor Core Throughput

Computing capacity of GPU Tensor Cores, often limited by the data feeding speed from memory.

📖
Begriffe

Data Pipeline Parallelism

Strategy where data loading, preprocessing, and transfer execute in parallel with GPU computation to hide latencies.

📖
Begriffe

Prefetching

Loading data into GPU memory in advance before it is needed by computation, to keep the GPU busy.

📖
Begriffe

Host-to-Device Latency

Time required to initiate and complete a data transfer from the CPU (host) to the GPU (device).

📖
Begriffe

CUDA Stream

Sequence of operations executed on the GPU in a specific order, allowing transfers and computations to be concurrent.

📖
Begriffe

NUMA Awareness

Optimization of memory allocations to respect the NUMA topology of multi-CPU servers, reducing access latencies.

📖
Begriffe

GPUDirect RDMA

Technology enabling direct data transfer between GPU memory of different nodes via RDMA, without CPU copying.

📖
Begriffe

Asynchronous Data Transfer

Data transfer executed in parallel with GPU computations, using CUDA streams to hide latencies.

📖
Begriffe

Page-Locked Memory

Non-pageable system memory, required for high-bandwidth asynchronous DMA transfers to the GPU.

🔍

Keine Ergebnisse gefunden