Evaluating the relationship between variables is one of the most common goals of statistical analysis.
In this webinar, we cover the methods for analysis and sample size determination for correlation measures, interrater agreement measures and diagnostic measures.
Correlation measures are widely used to summarise the strength of association between variables.
Commonly seen in areas such as regression analysis, the most widely used version is the Pearson correlation for a linear relationship.
However other correlations may be more suitable in certain contexts such as rank correlations like Spearman’s correlation for dealing with ordinal rank data.
Assessing the reliability of different “raters” is vital in areas where multiple assessors criteria or methods are available to evaluate a disease or condition.
Cohen’s Kappa statistic is a widely used approach to quantify the degree of agreement between multiple raters and provides a basis for the testing and estimation of interrater reliability.
The statistical evaluation of diagnostic testing is a vital component of ensuring that proposed screening or testing procedures have the appropriate accuracy for clinical usage.
A plethora of measures exist to evaluate the performance of a diagnostic test but one of the most common is Receiver Operating Characteristic (ROC) curve analysis where the Area Under the Curve (AUC) provides a snapshot of how well a test performs over the entire range of discrimination boundaries.
So join us for this webinar as we cover the methods for analysis and sample size determination.
Duration: 60 minutes
|Explore The Data
||Get The Deck
Download and explore the data yourself. Data files include: