Interpretability Evaluation Metrics
Inter-Annotator Agreement
Statistical measure (e.g., Cohen's Kappa score) assessing the level of consensus among different human experts on the quality or correctness of an explanation, validating its subjectivity.
← Terug