Several modes of inference are currently used in practice to evaluate the confidence of putative peptide identifications resulting from database scoring algorithms such as Mascot, SEQUEST or X!Tandem. The approaches include parametric methods, such as classic PeptideProphet, and distribution free methods, such as methods based on reverse or decoy databases. Due to its parametric nature classic PeptideProphet, although highly robust, was not highly flexible and was difficult to apply to new search algorithms or classification scores. While commonly applied, the decoy approach has not yet been fully formalized and standardized. And, although they are distribution-free, they like other approaches are not free of assumptions. Recent manuscripts by Kall et al, Choi and Nesvizhskii and Choi et al help advance these methods, specifically by formalizing an alternative formulation of decoy databases approaches and extending the PeptideProphet methods to make explicit use of decoy databases, respectively. Taken together with standardized decoy database methods, and expectation scores computed by search engines like Tandem, there exist at least four different modes of inference used to assign confidence levels to individual peptides or groups of peptides. We overview and compare the assumptions of each of these approach and summarize some interpretation issues. We also discuss some suggestions, which may make the use of decoy databases more computationally efficient in practice.
PerspectiveInterpreting tandem mass spectrometry (MS/MS) experiments involves a large number of computational steps, the first of which is to assign to each of thousands of spectra a single putative sequence and an associated score related to the accuracy of the assignment. The second step is that of assigning an interpretable measure of confidence to one or a group of those identifications. Because all subsequent results depend highly on these confidence measures this step may be the most fundamental of all. Several different modes of inference are now used to assign confidence measures, each of which relies on the same underlying model yet proceeds using a different class of assumptions. In our discussion we will follow the language suggested by Kall et al [1] with one exception: rather than the term False Discovery Rate (FDR) we suggest the use of False Identification Rate (FIR). Although we recognize that they are derived from the same underlying statistical foundations it is