A common suggestion made in the psychometric literature for fixed-length classification tests is that one should design tests so that they have maximum information at the cut score. Designing tests in this way is believed to maximize the classification accuracy and consistency of the assessment. This article uses simulated examples to illustrate that one can obtain higher classification accuracy and consistency by designing tests that have maximum test information at locations other than at the cut score. We show that the location where one should maximize the test information is dependent on the length of the test, the mean of the ability distribution in comparison to the cut score, and, to a lesser degree, whether or not one wants to optimize classification accuracy or consistency. Analyses also suggested that the differences in classification performance between designing tests optimally versus maximizing information at the cut score tended to be greatest when tests were short and the mean of ability distribution was further away from the cut score. Larger differences were also found in the simulated examples that used the 3PL model compared to the examples that used the Rasch model.An important function of many educational and psychological tests is to classify examinees into different performance categories. In K-12 settings, examinees might be classified into multiple performance categories, such as advanced, proficient, partially proficient, and basic, or simply proficient and not proficient. In licensure and certification (credentialing) testing, examinees are classified as having passed or failed an exam. Key considerations in these contexts are the classification accuracy and consistency of the assessment. Classification accuracy reports the extent to which observed classification agrees with "true" classification, and classification consistency is the proportion of examinees that would be classified into the same performance category over parallel replications of the assessment (Lee, 2010).Measurement experts have expressed a simple, intuitive, and widely accepted approach for how to develop fixed-length tests for maximizing classification accuracy and consistency of exam scores when using item response theory (IRT) models. That approach is to maximize test information near the cut score used to classify examinees (