Term
|
Definition
| Consistency in measurement |
|
|
Term
|
Definition
| Index of reliability, indicates the ratio between the true score variance on a test and the total variance [ Observer score = true score + Error] X= T=E |
|
|
Term
|
Definition
| Component of the observed score that does not have to do with a test taker's true ability or the trait being measured |
|
|
Term
|
Definition
| Standard deviation Squared |
|
|
Term
|
Definition
| all of the factors associated with the process of measuring some variable, other than the variable being measured |
|
|
Term
|
Definition
| A source of error in measuring a targeted variable caused by unpredictable fluctuations and inconsistencies of other variables in the measurement process |
|
|
Term
|
Definition
| A source of error in measuring a variable that is constant ot proportionate to what is presumed to be the true value of the vairable being measured [variance = true variance + Error Variance] Reliability id the proportions of the total cariance attributed to true variance |
|
|
Term
|
Definition
| Variation may exist within items in a test or between tests |
|
|
Term
|
Definition
| Sources of error may stem from the testing environment |
|
|
Term
| Test scoring and interpretation |
|
Definition
| Computer testing reduces error in test scoring, but many tests still require expert interpretation (e.g., projective tests); subjectivity in scoring can enter into behavioral assessment. |
|
|
Term
|
Definition
| The extent to which the population of voters in the study actually was representative of voters in the election. |
|
|
Term
|
Definition
| Interviewers may not have been trained properly, the wording in the questionnaire may have been ambiguous, or the items may have somehow been biased to favor one or another of the candidates. |
|
|
Term
|
Definition
| Obtained by correlating pairs of scores from the same people on two different administrations of the same test. |
|
|
Term
|
Definition
| With intervals greater than 6 months, the estimate of test-retest reliability |
|
|
Term
| Standard Error of Measurement |
|
Definition
| Provides a measure of Precision of an observed test score; an estimate of the amount of error inherent in an observed score or measurement |
|
|
Term
|
Definition
| A range or band of test scores that is likelu to contain the true score |
|
|
Term
| Standard Error of Difference |
|
Definition
| A measure that can aid a test user in determining how large a difference in test scores should be before it is considered statistically significant |
|
|
Term
|
Definition
| A value that according to CCT genuinely reflects an individuas ability level as measured by particular tests |
|
|
Term
|
Definition
| Estimates the extent to which specific sources of variation under defined conditions are contributing to the test score |
|
|
Term
|
Definition
| Based on the Idea that a person's test score vary from testing to testing because of variables in the testing situation. |
|
|
Term
|
Definition
Provides a way to model the probability that a person with X ability will be able to perform at a level of Y. [IRT refers to a family methods and techniques used to distinguish specific approaches. IRT incorporates considerations of an item’s level of difficulty and discrimination. Difficulty relates to an item not being easily accomplished, solved, or comprehended.] |
|
|
Term
|
Definition
| The degree to which an item dofferentation among people with higher or lower levels of trait,ability, or other variables being measured. |
|
|
Term
| Variance in a hypothetical Test |
|
Definition
| The nature of the test will often determine the reliability metric; some considerations include: (1) The test items are homogeneous or heterogeneous in nature (2) The characteristic, ability, or trait being measured is presumed to be dynamic or static (3) The range of a test scores is or is not restricted (4) The test is a speed or a power test (5) The test is or is not criterion-referenced |
|
|
Term
|
Definition
| The degree of relatedness of item on a scale (to gauge the homogenity of a test) |
|
|
Term
| Kuder-richardson-formula 20 |
|
Definition
| Statistic choice for determining the inter-item consistency of Dichotomous items |
|
|
Term
|
Definition
| Mean of all possible split-half correlations, corrected by the spearman-brown formula; it is the most popular approach for internal consistency and the calues range from 0 to 1 (non-dicho items) |
|
|
Term
| Average proportional Distance |
|
Definition
| Focuses on the degree of difference between score on test items; it involves averaging the diference between scores on all of the items dividing by the number of response options on the test, and then subtracting by 1 |
|
|
Term
|
Definition
| The degree of a relationship between various forms of a test |
|
|
Term
|
Definition
| For each form of a test the means and the variances of observed test scores are equal |
|
|
Term
|
Definition
| Different versions of the test that have been constructed so as to be parallel; they do not meet the strict requirements of parallel forms but item content and difficulty are similar between tests |
|
|
Term
|
Definition
Obtained by corelating two pairs of scores obtained from equivalent halves of a single test administered once, entails three steps: STEP 1 – Divide the test into equivalent halves STEP 2 – Calculate a Pearson r between scores on the two halves of the test STEP 3 – Adjust the half-test reliability using the Spearman-Brown formula |
|
|
Term
|
Definition
| Allows a test developer or user to estimate internal consistency reliability from a correlation of two-halves of a test. |
|
|