- Sensitivity and specificity
**Sensitivity and specificity**are statistical measures of the performance of abinary classification test. The**sensitivity**or the**recall rate**measures the proportion of actual positives which are correctly identified as such (i.e. the percentage of sick people who are identified as having the condition); and the**specificity**measures the proportion of negatives which are correctly identified (i.e. the percentage of well people who are identified as not having the condition). They are closely related to the concepts oftype I and type II errors .For any test, there is usually a trade-off between each measure. For example in a manufacturing setting in which one is testing for faults, one may be willing to risk discarding functioning components (low specificity), in order to increase the chance of identifying nearly all faulty components (high sensitivity). This trade-off can be represented graphically using a

ROC curve .**Definitions**Imagine a scenario where people are tested for a disease. The outcome can be positive (sick) or negative (healthy).

**ensitivity**:$\{\; m\; sensitivity\}=frac\{\; m\; number\; of\; True\; Positives\; called\; as\; Positives\}$ m number of True Positives (called as Positives}+{ m called as False Negatives)

A sensitivity of 100% means that the test recognizes all sick people as such. Thus in a high sensitivity test, a negative result is used to rule out the disease.

Sensitivity alone does not tell us how well the test predicts other classes (that is, about the negative cases). In the binary classification, as illustrated above, this is the corresponding specificity test, or equivalently, the sensitivity for the other classes.

Sensitivity is not the same as the

positive predictive value (ratio of true positives to combined true and false positives), which is as much a statement about the proportion of actual positives in the population being tested as it is about the test.The calculation of sensitivity does not take into account indeterminate test results. If a test cannot be repeated, the options are to exclude indeterminate samples from analyses (but the number of exclusions should be stated when quoting sensitivity), or, alternatively, indeterminate samples can be treated as false negatives (which gives the worst-case value for sensitivity and may therefore underestimate it).

**pecificity**:$\{\; m\; specificity\}=frac\{\; m\; number\; of\; True\; Negatives\; called\; as\; Negatives\}$ m number of True Negatives (called as Negatives}+{ m called as False Positives)

A specificity of

**100%**means that the test recognizes all healthy people as healthy. Thus a positive result in a high specificity test is used to rule in the disease. The maximum is trivially achieved by a test that claims everybody healthy regardless of the true condition. Therefore, the specificity alone does not tell us how well the test recognizes positive cases. We also need to know the sensitivity of the test to the class, or equivalently, the specificities to the other classes.A test with a high specificity has a low Type I error rate.

Specificity is sometimes confused with the precision or the

positive predictive value , both of which refer to the fraction of returned positives that are true positives. The distinction is critical when the classes are different sizes. A test with very high specificity can have very low precision if there are far more true negatives than true positives, and vice versa.**Worked example****Terminology in information retrieval**In

information retrieval positive predictive value is called**precision**, and sensitivity is called**recall**.The

F-measure can be used as a single measure of performance of the test. The F-measure is theharmonic mean of precision and recall::$F\; =\; 2\; imes\; (\{\; m\; precision\}\; imes\; \{\; m\; recall\})\; /\; (\{\; m\; precision\}\; +\; \{\; m\; recall\}).$

In the traditional language of

statistical hypothesis testing , the sensitivity of a test is called thestatistical power of the test, although the word "power" in that context has a more general usage that is not applicable in the present context. A sensitive test will have fewer Type II errors.**See also***

binary classification

*receiver operating characteristic

*statistical significance

*Type I and type II errors

*Selectivity

*Negative predictive value

*Positive predictive value

*statistical significance

*Youden's J statistic **References***

**External links*** [

*http://www.musc.edu/dc/icrebm/sensitivity.html Sensitivity and Specificity*] Medical University of South Carolina

* Calculators:

** [*http://faculty.vassar.edu/lowry/clin1.html Vassar College's Sensitivity/Specificity Calculator*]

**OpenEpi software program

*Wikimedia Foundation.
2010.*