🏠 Trang chủ
Benchmark
📊 Tất cả benchmark 🦖 Khủng long v1 🦖 Khủng long v2 ✅ Ứng dụng To-Do List 🎨 Trang tự do sáng tạo 🎯 FSACB - Trình diễn cuối cùng 🌍 Benchmark dịch thuật
Mô hình
🏆 Top 10 mô hình 🆓 Mô hình miễn phí 📋 Tất cả mô hình ⚙️ Kilo Code
Tài nguyên
💬 Thư viện prompt 📖 Thuật ngữ AI 🔗 Liên kết hữu ích

Thuật ngữ AI

Từ điển đầy đủ về Trí tuệ nhân tạo

162
danh mục
2.032
danh mục con
23.060
thuật ngữ
📖
thuật ngữ

Embedded AutoML

Subfield of AutoML specialized in the automatic generation of models optimized for the specific constraints of embedded devices, including limited memory, low computational power, and energy constraints.

📖
thuật ngữ

Model Quantization

Optimization technique that reduces the numerical precision of a neural network's weights and activations (typically from 32-bit to 8-bit or less) to decrease model size and accelerate inference on constrained hardware.

📖
thuật ngữ

Neural Pruning

Process of selectively removing redundant weights or neurons in a neural network to reduce its computational complexity and memory footprint while preserving its accuracy.

📖
thuật ngữ

Knowledge Distillation

A transfer learning method where a large teacher model trains a more compact student model, allowing the performance of the large model to be retained in an architecture suitable for Edge devices.

📖
thuật ngữ

Inference Optimization

Set of techniques aimed at reducing the time and resources required to execute a trained model, including operator fusion, efficient memory allocation, and hardware parallelism exploitation.

📖
thuật ngữ

NAS for Edge

Constrained Neural Architecture Search that automatically optimizes network structures by specifically considering the hardware limitations of Edge devices, such as target latency and power consumption.

📖
thuật ngữ

Model Compiler

Tool that transforms AI computational graphs into optimized machine code for specific target architectures, incorporating optimizations like quantization and operator fusion.

📖
thuật ngữ

TensorRT

NVIDIA's optimization and runtime SDK for deploying AI models in production, using quantization, layer fusion, and kernel optimization to maximize performance on NVIDIA GPUs.

📖
thuật ngữ

TinyML

Field of machine learning focused on running AI models on microcontrollers and ultra-low-power devices, typically with less than 1MB of memory and operating at less than 1mW.

📖
thuật ngữ

Edge TPU

ASIC hardware accelerator developed by Google specifically for edge AI inference, optimized to run quantized TensorFlow Lite models with high energy efficiency.

📖
thuật ngữ

Memory optimization

Techniques for reducing the memory footprint of models including weight sharing, compression, and dynamic allocation to adapt to embedded device constraints.

📖
thuật ngữ

Inference latency

Time elapsed between data input into a model and obtaining its prediction, a critical parameter in real-time Edge applications where typical target values are below 10ms.

📖
thuật ngữ

Lightweight model

Neural network architecture specifically designed to minimize parameters and computational operations, such as MobileNet or EfficientNet, optimized for mobile and Edge deployments.

📖
thuật ngữ

Distributed deployment

Strategy of distributing AI workloads across multiple Edge devices to optimize overall resources and improve scalability of distributed AI applications.

📖
thuật ngữ

Energy optimization

Process of minimizing power consumption of AI models on Edge devices, crucial for battery-powered applications and large-scale deployments.

📖
thuật ngữ

Edge AI

Paradigm of processing artificial intelligence directly on edge devices, eliminating the need to communicate with the cloud for critical inference tasks.

📖
thuật ngữ

AI Microcontroller

Ultra-low-power system-on-chip integrating dedicated hardware accelerators for AI inference, enabling the execution of TinyML models with a consumption of a few microwatts.

📖
thuật ngữ

Hardware-aware optimization

AutoML approach that integrates the specific characteristics of the target hardware into the automatic model design process, ensuring optimal compatibility and performance.

📖
thuật ngữ

Operator fusion

Compilation technique that combines several adjacent layers or operations into a single kernel operation, reducing memory overhead and improving computational efficiency on the Edge.

🔍

Không tìm thấy kết quả