🏠 Trang chủ
Benchmark
📊 Tất cả benchmark 🦖 Khủng long v1 🦖 Khủng long v2 ✅ Ứng dụng To-Do List 🎨 Trang tự do sáng tạo 🎯 FSACB - Trình diễn cuối cùng 🌍 Benchmark dịch thuật
Mô hình
🏆 Top 10 mô hình 🆓 Mô hình miễn phí 📋 Tất cả mô hình ⚙️ Kilo Code
Tài nguyên
💬 Thư viện prompt 📖 Thuật ngữ AI 🔗 Liên kết hữu ích

Thuật ngữ AI

Từ điển đầy đủ về Trí tuệ nhân tạo

162
danh mục
2.032
danh mục con
23.060
thuật ngữ
📖
thuật ngữ

Multimodal Attribute Fusion

Process of combining features from different modalities (text, image, sound) into a unified representation for a learning model, aiming to capture complex interactions between data sources.

📖
thuật ngữ

Explanation by Projection

Interpretability method that involves projecting the contribution of a complex modality (e.g., an image) onto a simpler and interpretable space (e.g., keywords or concepts) to explain its influence on the model's prediction.

📖
thuật ngữ

Multimodal Salience Map

Visualization that highlights the most influential regions or segments of each modality (pixels of an image, words of a text, audio segments) for a specific model decision, often by overlaying contributions on the original data.

📖
thuật ngữ

Inter-modality Semantic Alignment

Technique aimed at establishing semantic correspondences between elements of different modalities (e.g., linking a word to an image region or a sound to an action), crucial for the model to understand relationships and provide coherent explanations.

📖
thuật ngữ

Modality-wise Decomposition

Explainability approach that isolates and quantifies the individual contribution of each input modality to the final prediction, allowing understanding of whether a decision is primarily guided by text, image, or sound.

📖
thuật ngữ

Multimodal Concept Bottleneck

Model architecture where the final prediction is conditioned on a set of interpretable concepts, themselves inferred from the fusion of modalities, offering clear traceability from raw data to concepts then to decision.

📖
thuật ngữ

Orthogonality Regularization

Constraint applied during training to force the representations of different modalities in the shared latent space to be as independent as possible, avoiding redundancy and improving clarity of modality-specific explanations.

📖
thuật ngữ

Multimodal Counterfactual Explanation

Generation of modified examples (by changing one or more modalities) that are sufficient to reverse the model's prediction, helping to understand the minimal conditions and interactions between modalities necessary for a decision.

📖
thuật ngữ

Late Fusion for Interpretability

Strategy where each modality is processed by a specialized model up to an intermediate decision, with results then being fused. This approach facilitates interpretation by isolating the logic of each modality before the final combination.

📖
thuật ngữ

Foundational VISION-LANGUAGE Model

Large-scale model pre-trained on massive corpora of textual and visual data, capable of understanding and generating content from these two modalities, whose interpretability is a major challenge due to its intrinsic complexity.

📖
thuật ngữ

Modal Role Analysis

Systematic evaluation of the role played by each modality in different tasks or contexts, determining whether a modality acts as contextual support, a primary information source, or a modifier for others.

📖
thuật ngữ

Visuo-Linguistic Grounding

Process of anchoring linguistic symbols (words, phrases) to concrete entities or concepts in visual data, fundamental for model explanations linking text and image to be semantically correct and understandable.

📖
thuật ngữ

Fusion Node Interpretability

Method that focuses on analyzing the specific neurons or layers where the fusion of multimodal information occurs, to understand how interactions are encoded and how they influence the model's output.

📖
thuật ngữ

Cross-Modal Gradient Explanation

Interpretability technique that calculates the gradient of the model's output with respect to the features of one modality, while conditioning this calculation on the features of another modality, thus revealing inter-modal dependencies.

🔍

Không tìm thấy kết quả