1. Sampling Validity (similar to content validity) ensures that the measure covers the broad range of areas within the concept under study.
2. Formative Validity when applied to outcomes assessment it is used to assess how well a measure is able to provide information to help improve the program under study
3. Criterion-Related Validity is used to predict future or current performance - it correlates test results with another criterion of interest.
4. Construct Validity is used to ensure that the measure is actually measure what it is intended to measure (i.e. the construct), and not other variables
5. Face Validity ascertains that the measure appears to be assessing the intended construct under study
6. Validity refers to how well a test measures what it is purported to measure.
7. Internal consistency reliability is a measure of reliability used to evaluate the degree to which different test items that probe the same construct produce similar results.
7.1. Average inter-item correlation is a subtype of internal consistency reliability. It is obtained by taking all of the items on a test that probe the same construct (e.g., reading comprehension), determining the correlation coefficient for each pair of items, and finally taking the average of all of these correlation coefficients.
7.2. Split-half reliability is another subtype of internal consistency reliability. The process of obtaining split-half reliability is begun by “splitting in half” all items of a test that are intended to probe the same area of knowledge (e.g., World War II) in order to form two “sets” of items.
8. Inter-rater reliability is a measure of reliability used to assess the degree to which different judges or raters agree in their assessment decisions.
9. Parallel forms reliability is a measure of reliability obtained by administering different versions of an assessment tool (both versions must contain items that probe the same construct, skill, knowledge base, etc.) to the same group of individuals.
10. Test-retest reliability is a measure of reliability obtained by administering the same test twice over a period of time to a group of individuals
11. Reliability is the degree to which an assessment tool produces stable and consistent results.