🏠 Home
Benchmark
📊 Tutti i benchmark 🦖 Dinosauro v1 🦖 Dinosauro v2 ✅ App To-Do List 🎨 Pagine libere creative 🎯 FSACB - Ultimate Showcase 🌍 Benchmark traduzione
Modelli
🏆 Top 10 modelli 🆓 Modelli gratuiti 📋 Tutti i modelli ⚙️ Kilo Code
Risorse
💬 Libreria di prompt 📖 Glossario IA 🔗 Link utili

Glossario IA

Il dizionario completo dell'Intelligenza Artificiale

162
categorie
2.032
sottocategorie
23.060
termini
📖
termini

Embedded AutoML

Subfield of AutoML specialized in the automatic generation of models optimized for the specific constraints of embedded devices, including limited memory, low computational power, and energy constraints.

📖
termini

Model Quantization

Optimization technique that reduces the numerical precision of a neural network's weights and activations (typically from 32-bit to 8-bit or less) to decrease model size and accelerate inference on constrained hardware.

📖
termini

Neural Pruning

Process of selectively removing redundant weights or neurons in a neural network to reduce its computational complexity and memory footprint while preserving its accuracy.

📖
termini

Knowledge Distillation

A transfer learning method where a large teacher model trains a more compact student model, allowing the performance of the large model to be retained in an architecture suitable for Edge devices.

📖
termini

Inference Optimization

Set of techniques aimed at reducing the time and resources required to execute a trained model, including operator fusion, efficient memory allocation, and hardware parallelism exploitation.

📖
termini

NAS for Edge

Constrained Neural Architecture Search that automatically optimizes network structures by specifically considering the hardware limitations of Edge devices, such as target latency and power consumption.

📖
termini

Model Compiler

Tool that transforms AI computational graphs into optimized machine code for specific target architectures, incorporating optimizations like quantization and operator fusion.

📖
termini

TensorRT

NVIDIA's optimization and runtime SDK for deploying AI models in production, using quantization, layer fusion, and kernel optimization to maximize performance on NVIDIA GPUs.

📖
termini

TinyML

Field of machine learning focused on running AI models on microcontrollers and ultra-low-power devices, typically with less than 1MB of memory and operating at less than 1mW.

📖
termini

Edge TPU

ASIC hardware accelerator developed by Google specifically for edge AI inference, optimized to run quantized TensorFlow Lite models with high energy efficiency.

📖
termini

Memory optimization

Techniques for reducing the memory footprint of models including weight sharing, compression, and dynamic allocation to adapt to embedded device constraints.

📖
termini

Inference latency

Time elapsed between data input into a model and obtaining its prediction, a critical parameter in real-time Edge applications where typical target values are below 10ms.

📖
termini

Lightweight model

Neural network architecture specifically designed to minimize parameters and computational operations, such as MobileNet or EfficientNet, optimized for mobile and Edge deployments.

📖
termini

Distributed deployment

Strategy of distributing AI workloads across multiple Edge devices to optimize overall resources and improve scalability of distributed AI applications.

📖
termini

Energy optimization

Process of minimizing power consumption of AI models on Edge devices, crucial for battery-powered applications and large-scale deployments.

📖
termini

Edge AI

Paradigm of processing artificial intelligence directly on edge devices, eliminating the need to communicate with the cloud for critical inference tasks.

📖
termini

AI Microcontroller

Ultra-low-power system-on-chip integrating dedicated hardware accelerators for AI inference, enabling the execution of TinyML models with a consumption of a few microwatts.

📖
termini

Hardware-aware optimization

AutoML approach that integrates the specific characteristics of the target hardware into the automatic model design process, ensuring optimal compatibility and performance.

📖
termini

Operator fusion

Compilation technique that combines several adjacent layers or operations into a single kernel operation, reducing memory overhead and improving computational efficiency on the Edge.

🔍

Nessun risultato trovato