🏠 Ana Sayfa
Benchmarklar
📊 Tüm Benchmarklar 🦖 Dinozor v1 🦖 Dinozor v2 ✅ To-Do List Uygulamaları 🎨 Yaratıcı Serbest Sayfalar 🎯 FSACB - Nihai Gösteri 🌍 Çeviri Benchmarkı
Modeller
🏆 En İyi 10 Model 🆓 Ücretsiz Modeller 📋 Tüm Modeller ⚙️ Kilo Code
Kaynaklar
💬 Prompt Kütüphanesi 📖 YZ Sözlüğü 🔗 Faydalı Bağlantılar

YZ Sözlüğü

Yapay Zekanın tam sözlüğü

162
kategoriler
2.032
alt kategoriler
23.060
terimler
📖
terimler

Faithfulness Score

Metric evaluating the degree of faithfulness of the generated response to the provided context, measuring whether the statements are factually supported by the retrieved sources.

📖
terimler

Context Relevance

Indicator quantifying the relevance of retrieved documents or passages to the initial query, essential for evaluating the quality of the RAG retrieval component.

📖
terimler

Answer Relevance

Score measuring the extent to which the generated response directly and completely answers the posed question, regardless of factual accuracy.

📖
terimler

Retrieval Precision

Proportion of relevant documents among all retrieved documents, evaluating the system's effectiveness in returning only useful information.

📖
terimler

Retrieval Recall

Ratio of relevant documents retrieved compared to the total number of relevant documents available in the knowledge base.

📖
terimler

Knowledge F1 Score

Harmonic mean between precision and recall of retrieved knowledge, providing a balanced measure of the overall RAG system performance.

📖
terimler

Context Utilization Rate

Percentage of relevant information from the retrieved context that is actually used in the final response, measuring the efficiency of source utilization.

📖
terimler

Hallucination Rate

Frequency at which the model generates information not supported by the provided context, a critical indicator of RAG system reliability.

📖
terimler

Semantic Similarity Score

Semantic similarity measure between the generated response and a reference response, using embeddings to capture meaning nuances.

📖
terimler

Answer Completeness

Evaluation of the coverage of all relevant aspects of the question in the generated response, ensuring a comprehensive answer.

📖
terimler

Retrieval Latency

Time required to retrieve relevant documents from the knowledge base, a crucial criterion for user experience in production.

📖
terimler

Token Efficiency Ratio

Ratio between the number of relevant tokens used and the total number of tokens generated, measuring the economic efficiency of the RAG system.

📖
terimler

Groundedness Score

Metric assessing the extent to which each statement in the response is supported by explicit evidence in the retrieved sources.

📖
terimler

Source Attribution Accuracy

Accuracy with which the system correctly attributes each part of the response to its appropriate documentary source in the retrieved context.

📖
terimler

Response Consistency

Measure of the internal consistency of the generated response, assessing the absence of contradictions between different parts of the response.

📖
terimler

Query Ambiguity Resolution

Ability of the RAG system to interpret and resolve ambiguities in the user query to retrieve the most relevant information.

📖
terimler

Information Overlap Score

Measure of the overlap between information present in the response and that available in the retrieved context, avoiding redundancies.

📖
terimler

Answer Accuracy

Evaluation of the factual truthfulness of the generated response compared to a ground truth or validated reference sources.

📖
terimler

Retrieval Coverage

Extent of the knowledge base actually accessible by the retrieval system, impacting the ability to answer diverse questions.

📖
terimler

Response Coherence

Quality of the logical structure and narrative flow of the generated response, ensuring clear and understandable presentation of information.

🔍

Sonuç bulunamadı