This study compares the relative accuracy of two methods of estimating employment test validity, (a) expert judgment and (b) small sample criterion-related validation studies The study was based on Navy data with samples of over 3,000 for each of nine jobs, with validity results on six tests for each job Twenty experienced psychologists estimated the observed validity for each of the 54 test-job combinations Both the random and systematic error in the expert judgments were evaluated Psychologists typically underestimated the validity by a small amount (an average systematic error of 019) On the average, in order to equal the accuracy of a single judge, the sample size of a criterion-related validation study would have to be 92 To match the accuracy of an average across four judges, the sample size must be 326 The sample size must be 1,164 to match the accuracy of the pooled judgment of 30 judges These results indicate that, given highly trained and experienced judges, expert judgment may provide more accurate estimates of validity for cognitive tests than do local criterion-related validation studies 0
<4 [2X1-1]= fl-p