Measures of Agreement PDF
Document Details
Uploaded by FondMonkey75
King Khalid University, Abha
Saeed Mastour
Tags
Related
- Principles of Plant Quarantine, WTO Regime & SPS Measures Lecture 02 PDF
- Principles of Plant Quarantine, WTO Regime & SPS Measures PDF
- Principles of Plant Quarantine, WTO Regime & SPS Measures Lecture Notes PDF
- Convenio Colectivo para el PAS Laboral de las Universidades Públicas Canarias 2013 PDF
- Paris Agreement (UNFCCC) PDF
- Resolución do 10 de xaneiro de 2024, da Dirección Xeral de Centros e Recursos Humanos, Galicia - PDF
Summary
This document provides an overview of measures of agreement, commonly used in health sciences research to assess consistency and reliability of data. Different methods, including Cohen's Kappa, Intra-class correlation, and Youden's J, are discussed.
Full Transcript
Saeed Mastour Measures of Agreement Saeed Mastour Measures of Agreement • In health sciences research, several measures are commonly used to assess agreement among observers, raters, or methods. • These measures help researchers understand the consistency and reliability of data. Saeed Mastour...
Saeed Mastour Measures of Agreement Saeed Mastour Measures of Agreement • In health sciences research, several measures are commonly used to assess agreement among observers, raters, or methods. • These measures help researchers understand the consistency and reliability of data. Saeed Mastour Types • Cohen's Kappa • • • • (k) Intra-Class Correlation (ICC) Test-Retest Reliability Percentage Agreement Lin's Concordance Correlation Coefficient • Krippendorff's Alpha • Youden's J (J statistic) Saeed Mastour Cohen's Kappa (k) * Description: Cohen's Kappa assesses the agreement between two raters beyond chance. * Formula: where Po is the observed agreement, and Pe is the expected agreement by chance. * Example: In a study comparing the diagnoses of two psychiatrists regarding the presence or absence of a specific mental health disorder in patients, Cohen's Kappa would measure the level of agreement beyond what would be expected by chance. ’ Application: Used when assessing agreement on categorical data, such as diagnoses or classifications. ’ Interpretation: Values range from -1 (complete disagreement) to 1 (perfect agreement), with 0 indicating agreement equivalent to chance. Saeed Mastour Intra-Class Correlation (ICC) • Description: ICC assesses the reliability of measurements made by multiple observers on a continuous scale. • Formula: Various formulas are used depending on the model chosen. A commonly used formula for a two-way random-effects model is: _ MSB-MSW ~ MSB+(k-l)MSW where MSB is the mean square between, MSW is the mean square within, and fc is the number of measurements. • Example: Researchers might use ICC to assess the reliability of blood pressure measurements taken by different nurses across multiple clinic visits. • Application: Useful for continuous data, such as measurements of blood pressure or laboratory values. • Interpretation: Values range from 0 to 1, with higher values indicating greater reliability. Saeed Mastour Test-Retest Reliability * Description: Involves measuring the same subjects at two different points in time and correlating the two sets of measurements. ' Formula: Typically calculated using correlation coefficients such as Pearson’s correlation. * Example: A study assessing the reliability of a depression scale administered to patients at two different time points to determine the stability of their reported symptoms. ’ Application: Commonly used to assess the stability of measurements overtime. * Interpretation: Higher correlations indicate greater stability and reliability. Saeed Mastour Percentage Agreement • Description: Calculates the percentage of agreement between raters. • Formula: Percentage Agreement - x 100 • Example: Evaluating the agreement between different healthcare providers in assessing the severity of pain in patients using a visual analog scale. • Application: Simple and intuitive, often used for binary outcomes. • Interpretation: Provides a straightforward measure of agreement but does not account for chance agreement. Observer B Positive Observer B Negative Observer A Positive Observer A Negative A Positive and B Positive A Positive and B Negative A Negative and B Positive A Negative and B Negative Saeed Mastour Lin's Concordance Correlation Coefficient ' Description: Measures the agreement between two variables by assessing both precision and accuracy. ’ Formula: 2*cov(X,K) ~ Var(X)+Var(F)+(Jf-F)2 * Example: Comparing two laboratory methods for measuring the concentration of a specific biomarker in blood samplesand assessing both how closely the measurements agree and how far they deviate from a perfect line of agreement. * Application: Useful when assessing how closely measurements agree and how far they deviate from a perfect correlation. * Interpretation: Values range from -1 to 1, with 1 indicating perfect agreement. Saeed Mastour Kri ppendorff’s Alpha * Description: A measure of agreement that accommodates various types of scales. ■ Formula: It varies based on the type of data being analyzed (nominal, ordinal, interval, ratio). ■ Example: In a qualitative content analysis of patient interviews, multiple coders might use Krippendorff’s Alpha to assess the reliability of coding patient statements into different thematic categories. • Application: Applied to assess the reliability of content analysis or coding in qualitative research. ’ Interpretation: Values range from 0 to 1, with higher values indicating greater agreement. Saeed Mastour Youden's J (J statistic) • Youden's J, also known as the Youden's Index, is a single statistic that combines both sensitivity and specificity to assess the overall performance of a diagnostic test. • It is particularly useful when there is an optimal threshold for test results to classify individuals or items into binary categories (e.g., positive or negative). The formula for Youden's J is as follows: Actual Positive • Youden's J = Sensitivity + Specificity -1 Test Positive Test Negative Sensitivity (True Positive Rate) = TP / (TP + FN) Specificity (True Negative Rate) = TN / (TN + FP) Actual Negative Saeed Mastour Interpretation of Youden's J • Youden's J ranges from -1 to 1. • 1 indicates perfect test performance (perfect sensitivity and specificity). • 0 indicates a test with no discriminative power (similar to random chance) • Negative values indicate worse than random performance. • Higher Youden's J values suggest better diagnostic test performance. • The optimal threshold for the test can often be identified by maximizing Youden's J.