RAId_DbS: Peptide Identification using Database Searches with Realistic Statistics
- Gelio Alves^{1},
- Aleksey Y Ogurtsov^{1} and
- Yi-Kuo Yu^{1}Email author
DOI: 10.1186/1745-6150-2-25
© Alves et al; licensee BioMed Central Ltd. 2007
Received: 05 October 2007
Accepted: 25 October 2007
Published: 25 October 2007
Abstract
Background
The key to mass-spectrometry-based proteomics is peptide identification. A major challenge in peptide identification is to obtain realistic E-values when assigning statistical significance to candidate peptides.
Results
Using a simple scoring scheme, we propose a database search method with theoretically characterized statistics. Taking into account possible skewness in the random variable distribution and the effect of finite sampling, we provide a theoretical derivation for the tail of the score distribution. For every experimental spectrum examined, we collect the scores of peptides in the database, and find good agreement between the collected score statistics and our theoretical distribution. Using Student's t-tests, we quantify the degree of agreement between the theoretical distribution and the score statistics collected. The T-tests may be used to measure the reliability of reported statistics. When combined with reported P-value for a peptide hit using a score distribution model, this new measure prevents exaggerated statistics. Another feature of RAId_DbS is its capability of detecting multiple co-eluted peptides. The peptide identification performance and statistical accuracy of RAId_DbS are assessed and compared with several other search tools. The executables and data related to RAId_DbS are freely available upon request.
Open peer review
Reviewed by Frank Eisenhaber, Wing-Cheong Wong (co-reviewer invited by Frank Eisenhaber), Dongxiao Zhu (nominated by Arcady Mushegian) and Shamil Sunyaev. For the full reviews, please go to the Reviewers' comments section.
Introduction
Protein identification is the key to proteomics. As an indispensable component in mass spectrometry (MS) based protein identification, peptide identification through tandem MS (MS^{2}) is usually aided by automated data analysis. Among available data analysis tools, methods based on database searches are most frequently used. Methods using database searches may be roughly classified into two categories, depending on whether or not they provide E-values (or P-values) for candidate peptides. Methods – using either correlation, posterior probabilities, score, or Z-score – include, but are not limited to, SEQUEST [1], MS-Tag [2], Scope [3], CIDentify [4], Popitam [5], ProbID [6], and PepSearch [7]. Examples of database search methods directly reporting P- or E-values include, but are not limited to, Mascot [8], Sonar [9], InsPecT [10], OMSSA [11], and X!Tandem [12]. A comprehensive survey may be found in [13] and a performance evaluation of several of the methods mentioned can be found in [14].
For a given quality score cutoff S, E-value is defined as the expected number of hits, in a random database, with quality score being the same as or higher than the cutoff. (Similarly, P-value refers to the probability of finding a random hit with quality score being the same as or higher than the cutoff.) A realistic E-value assignment thus provides the user with the number of false positives to anticipate. Our goal in developing RAId_DbS (Robust Accurate Identification of Peptides in Database Search) is to provide a database search method with realistic E-value assignments. Among methods that report E-values, we find the approach employed by [15] is closest to what we have developed. Basically, both methods use the real score histogram from scoring database peptides against a query spectrum to form the basis of score statistics. The difference, however, lies in the fact RAId_DbS has its score statistics founded on a theoretical distribution, while the method of [15] assumes an exponential distribution pdf(S) ≈ exp(-λS) for large score S.
To illustrate an important aspect of peptide score statistics, let us note that the noise in an MS^{2} spectrum is spectrum-specific. That is, it is not yet possible to predict spectral noise, which nonetheless influences the scoring of candidate peptides for a given spectrum. One of our goals in developing RAId_DbS is to take into account the fact that noise is spectrum-specific. This goal is achieved by using a scoring scheme whose statistics can be theoretically characterized. Our scoring scheme is largely similar to that proposed by [16]. However, in addition to the introduction of weight factors to encourage mass accuracy, we have taken into account the effect of finite sampling and finite skewness and have derived a new score distribution function replacing the Gaussian distribution assumed by [16].
Typical database search methods usually ask the user to set a maximum number of allowed enzymatic miscleavages. Not only do we lift this constraint, we even allow for non-canonical N-terminal cleavages (NNTC), also referred to as "incorrect N-terminal cleavages" by [17]. These are handled efficiently by first scoring exhaustively all the four-letter C- and N-terminal tags to produce two high-scoring tag lists, one for each terminal. A candidate peptide with NNTC will be scored only when one of its two terminal tags ranks high enough in its respective list.
Unless otherwise mentioned, we limit our discussions to methods that directly report E-values (or P-values). This is because converting scores, correlation coefficients, etc. into E-values is non-trivial and is method-dependent. In fact, even converting E-values reported by one method to E-values reported by other method is already non-trivial. This important task of standardizing E-values, although beyond the scope of the current paper, will be addressed in a forthcoming publication [18].
To better highlight the main points of this paper, we have relegated to the appendix details such as m/z peak filtering, tag scoring, and a detailed description of implementation. Throughout the paper, we use the dalton (Da) as the unit for molecular weight. In the following, we first provide a brief description of the two different types of data (centroid and profile) used, and the experimental protocol used to obtain the profile spectra. We will then describe RAId_DbS's scoring scheme, followed by a detailed description of the mathematical underpinning of the score statistics. The E-value test and the performance test will then be described followed by a section discussing the importance of quantifying the goodness of score distribution modeling in statistical inference. We conclude in the last section with some relevant remarks.
Experiment
Two data sets were used in this study. The centroid mode data set developed by the Institute for Systems Biology [19] was used only for performance comparison, while the profile mode data set was used for both statistical assessment and performance comparison. Because RAId_DbS is designed to take profile mode data and most of the published data are collected in centroid mode, it was necessary to generate profile data for this study. The profile mode data we used was provided by Dr. R.-F. Shen, the director of the mass spectrometry core facility at the National Heart, Lung, and Blood Institute (NHLBI). The acquisition of those profile mode spectra is described below.
A mixture of 7 proteins (Sigma) containing equimolar levels of α-lactalbumin (LALBA_BOVIN, P00711), lysozyme (LYSC_CHICK, P00698), β-lactoglobulin B (LACB_BOVIN, P02754), hemoglobin (HBA_HUMAN and HBB_HUMAN, P69905 and P68871), bovine serum albumin (ALBU_BOVIN, P02769), apotransferrin (TRFE_HUMAN, P02787), and β-galactosidase (BGAL_ECOLI, P00722) was used for all experiments. Note that both the α chain and the β chain of hemoglobin are included. The protein mixture in 50 mM ammonium bicarbonate buffer was reduced with 10 mM DTT at 60°C for 1 hr, alkylated with 55 mM iodoacetamide at room temperature in the dark for 30 min, and digested with trypsin (Promega) at 50:1 mass ratio at 37°C overnight, as described in [20]. Three different levels of protein mixture (50 fmols, 500 fmols, and 5 pmols of each protein) were then injected into LC/MS/MS. Two different kinds of mass spectrometers were utilized in this study, nanospray (NSI)/LTQ FT (Thermo Finnigan) and matrix assisted laser desorption ionization (MALDI)/TOF/TOF (Applied Biosystems). For NSI/LTQ FT, following [21], peptides were first loaded onto a trap cartridge (Agilent) at a flow rate of 2 μ l/min. Trapped peptides were then eluted onto a reversed-phase PicoFrit column (New Objective) using a linear gradient of acetonitrile (0–60%) containing 0.1% FA. The duration of the gradient was 20 min at a flow rate of 0.25 μ l/min, which was followed by 80% acetonitrile washing for 5 minutes. The eluted peptides from the PicoFrit column were nano-sprayed into an LTQ FT mass spectrometer. The data-dependent acquisition mode was enabled, and each survey MS scan was followed by five MS/MS scans with the dynamic exclusion option on. The spray voltage and ion transfer tube temperature were set at 1.8 kV and 160°C, respectively. The normalized collision energy was set at 35%. Three different combinations of mass analyzers (LTQ LTQ, LTQ FT, and FT FT) were used to acquire protein mixtures at each level. For MALDI/TOF/TOF, following [22], peptide separation was performed on a Famos/Switchos/Ultimate chromatography system (Dionex/LC Packings) equipped with a Probot (MALDI-plate spotting device). Peptides were injected and captured onto a trap column (PepMap C18, 5 μ m, 100 A, 300 μ m i.d. × 5 mm) at 10 μ l/min. Peptide separation was achieved on an analytical nano-column (PepMap C18, 3 μ m, 100 A, 75 μ m i.d. × 15 cm) using a gradient of 5 to 60% solvent B in A over 90 min (solvent A: 100% water, 0.1% TFA; solvent B: 80% acetonitrile/20% water, 0.1% TFA), 60 to 95% solvent B in A for 1 min, and then 95% solvent B for 19 min at a flow rate of 0.16 μ l/min. The HPLC eluant was supplemented with 5 mg/ml α-cyano-4-hydroxycinnamic acid (in 50/50 acetonitrile/water containing 0.1% TFA) from a syringe pump at a flow rate of 1 μ l/min, and spotted directly onto the ABI 4700 576-well target plates using the Probot. MALDI/TOF/TOF data were acquired in batch mode.
Scoring scheme
Like many other peptide analysis methods, RAId_DbS uses primarily b- and y-series peaks to score a candidate peptide or a tag. As will be seen, it is simple to include more evidence peaks, and it is also straightforward to switch to a different peak series for scoring. For example, in place of b- and y-series, one may use c- and z-series for scoring. This will be useful for analyzing spectra generated by the electron transfer dissociation (ETD) method [23].
where b(π)(y(π)) represents the set of theoretical b (y) peaks of peptide π, T(π) is the total number of peaks when one unites b(π) and y(π), $\mathcal{I}$(i) = max{I_{ i }, 1} with I_{ i }being the intensity found in the processed query spectrum (see appendix) for the peak labeled i with m/z value m_{ i }in the set {b(π) ∪ y (π)}. The weight factor w_{ i }is introduced to emphasize peaks with less mass error. The default is w_{ i }(m_{ i }) = exp(-|Δm_{ i }|) with Δm_{ i }= m_{ i }- t_{ i }being the difference between the observed m/z value m_{ i }and the theoretical value t_{ i }.
In the absence of weighting, our scoring scheme is the same as that of [16]. The difference between our method and that proposed by [16] lies in the theoretical distribution being derived as opposed to assumed. When w_{ i }= 1, one will pick the strongest intensity in range i ([t_{ i }- 1, t_{ i }+ 1]) as I_{ i }. With weighting, RAId_DbS first multiplies the log intensity of each candidate peak in the mass range i by w_{ i }(m_{ i }) and then pick the maximum w_{ i }(m_{ i }) ln [$\mathcal{I}$(i)] from each of the mass ranges ${\{[{t}_{i}-1,{t}_{i}+1]\}}_{i=1}^{T(\pi )}$. The same scoring method is used for sequence tags.
When the number T(π) in Eq. (1) is fixed and large, one would anticipate the central limit theorem to hold and the score distribution be a Gaussian. However, for a typical search T(π) is not fixed and not necessarily large enough to guarantee that the skewness is negligible. In the next section, we will derive a new probability distribution function, whose end results is given in Eq. (17), to accommodate the finite sampling and skewness. It turns out that even with the weights included to encourage better m/z matching, the score statistics of Eq. (1) still follow the general form given by Eq. (17).
where [c] represents the integer part of c > 0, p ≡ ⟨c⟩/(L_{eff}) with ⟨c⟩ representing c averaged over all peptides entering the final scoring, and L_{eff} ≡ Molecular weight/110Da. Formula (2), heuristic in nature, is invoked if ⟨c⟩ ≤ 2 to provide a conservative E-value. There is apparently room for improvement in scoring spectrum with little information.
Theory
In this section, we will provide a heuristic derivation to solve a generic problem that may occur in various scientific fields. Specifically, we will address how the Gaussian distribution assured by the central limit theorem can be modified in the presence of skewness and finite sampling. To deal with skewness and finite sampling size is by no means a new front of attack. There exist many well written literatures [24, 25] touching upon this subject. However, the deviations (due to finite size and/or skewness) from the Gaussian distribution are usually dealt with through computing the difference between the pre-Gaussian and the Gaussian using Hermite polynomials [24, 25]. Adding only a few or finite number of those correction terms sometimes roughens the tail of the distribution function. We provide a different way to derive the distribution function, incorporating the finite size effect and the skewness, that has a smooth tail and has the correct asymptotics in a closed form.
The random variable x corresponds to the logarithm of m/z peak intensity whose distribution g(x) is governed by the experimental spectrum under consideration. Because of the use of a logarithm, a rescaling of peak intensity can only result in a constant shift of the mean of the variable distribution, but not the shape of the distribution. With this understanding in mind, we now proceed with the theory in a rather general setting.
Given a distribution function g(x) with ∫g(x)dx = 1, the k th moment is given by ⟨x^{ k }⟩ ≡ ∫x^{ k }g(x)dx. The first moment is the mean, and the difference between the second moment and first moment squared, ⟨x^{2}⟩ – ⟨x⟩^{2}, is the variance. The central limit theorem may be stated as follows [24]: If one samples independently n numbers, say x_{1}, x_{2}, ..., x_{ n }, from a given distribution function g(x) with mean $\overline{x}$ and variance σ^{2}, the distribution of the quantity ${y}^{\prime}\equiv \sqrt{n}[({\displaystyle {\sum}_{i=1}^{n}{x}_{i}})/n-\overline{x}]$, a random number itself, will approach a Gaussian as n approaches infinity with zero mean and variance σ^{2} provided that $\overline{x}$ and σ^{2} are finite. When dealing with finite n, one may simply consider $y\equiv ({\displaystyle {\sum}_{i=1}^{n}{x}_{i}})/n-\overline{x}$, and anticipate the distribution function of y to be close to Gaussian with zero mean and variance σ^{2}/n.
The situation we wish to study is when n is not too large and when ⟨(x - $\overline{x}$)^{3}) ⟩ ≡ ∫(x - $\overline{x}$)^{3})g(x)dx is not small, i.e., when the skewness of the distribution function is nonnegligible in the sense that the condition |⟨(x - $\overline{x}$)^{3} ⟩| ≪ ∫ n⟨(x - $\overline{x}$)^{2})⟩ is not true due to finite n. Inclusion of this term and other higher order odd moments introduces skewness to the distribution function of the y variable. The simplest choice, however, is to keep only the third moment. As for the higher order even moments, they contribute symmetrically to the probability distribution function of y, thus suppressing the skewness observed in the score distribution. We therefore choose to ignore all moments fourth or higher. To provide an analytical expression for the probability distribution function for y with nonnegligible skewness, we first show how the central limit theorem can be derived heuristically and how such a heuristic approach can be readily employed to give correct asymptotics that one needs. For simplicity, we will proceed under the assumption that $\overline{x}$ ≡ ⟨x⟩ = ∫ xg(x)dx = 0. Extending the results obtained here to the case of nonzero but finite $\overline{x}$ is straightforward. Note that by definition, we have g(x) ≥ 0 ∀ x. Also, we further assume that g(x) > 0 over only a finite range of x, and define X ≡ max{abs(x) |g(x) > 0}.
which is the celebrated Gaussian distribution of the central limit theorem.
In principle, one may choose to retain the Gaussian part and expand the rest of the exponents in powers of (k – k*). We did not pursue this route, however, since such expansion will require information about higher cumulants of g(x) and experimentally g(x) may already contain some uncertainties. Instead, we simply treat $\mathcal{C}$ as an integration constant.
approaches as expected from central limit theorem.
In the case where the first moment $\overline{x}$ ≡ ∫xg(x)dx ≠ 0, one may simply replace y by (y - $\overline{x}$) in Eq. (17).
Brief description of implementation
The operation of RAId_DbS consists of three stages. The first step includes centroidizing m/z peaks followed by peak filtering. After this crucial step, RAId_DbS exhaustively scores all possible C- and N-tags of four amino acids. This helps RAId_DbS in filtering peptide candidates with NNTC before full scoring. More details for those parts can be found in the appendix. In the third stage, RAId_DbS uses primarily the b- and y-series peaks for scoring. For each query spectrum, the collection of scores from all candidate peptides constitutes a score histogram, that is then used to determine the constant $\mathcal{C}$ (and other parameters) of theoretical distribution, see Eq. (17). Once all parameters are determined, one then integrates the pdf from infinitely large score back to a finite score S to obtain the spectrum-specific P-value for score S. The goodness of the theoretical distribution is then assessed. These information are then used in conjunction with the effective database size to provide the E-value.
Results
When scoring peptides against a query spectrum, peptides within the given mass range will not have an identical number of b (or y) peaks. Separating the candidate peaks into different groups, each with a fixed number of b (or y) peaks is not practical. Further, we also wish to encourage mass accuracy and thus score candidate peptides using Eq. (1) with weights w_{ i }turned on. We still need characterizable statistics even with all of those additional complications. Fortunately, in this case all we need to do is consider β and γ ≡ n/(6⟨x^{2}⟩β^{2}) as two additional variables to be determined from fitting the score histogram. In Fig. 1(B), using a typical query spectrum, the red staircase is the score histogram with scores from database peptides with allowed molecular weights; each peptide was scored using Eq. (1) with weights w_{ i }(m_{ i }) = exp(-| Δ m_{ i }|). The solid curve is obtained from fitting the histogram with β, γ and log $\mathcal{C}$ as variables using a least squares procedure. As one may see, the statistics provided using Eq. (17) seems to capture the nature of score distribution reasonably well. The goodness of the fitting to the theoretical distribution may be quantified by a Student's t-test. The importance of such a test and its implication will be discussed in detail in the next section.
For a given search method and database, a list of candidate peptides is obtained for every spectrum analyzed. A peptide in the reported list will be classified as a false positive if it is not a subsequence of any of the seven standard proteins used to generate the spectra. For a given E-value cutoff, we count cumulatively the total number of false positive peptides assigned with E-values less than or equal to that cutoff. Dividing by the total number of spectra, we obtain the average cumulative count of false positives for that E-value cutoff. There are in total 6,734 spectra obtained through LTQ/LTQ, LTQ/FT, TOF/TOF, and FT/FT. Therefore, the usable region of this E-value accuracy test is limited to E ≥ 1/6734 ≈ 1.5 × 10^{-4}. Fig. 2 shows that RAId_DbS has better statistical accuracy than other methods. In particular, the results for nr after cluster removal seem to reflect well the behavior expected from a random protein database. That is, the resulting curve from RAId_DbS tracks well with the theoretical curve.
Example output of RAId_DbS containing multiple significant peptide hits. The contents in the "DEFINITION" and "GI-LIST" columns have been shortened to fit the page. The first two hits correspond to the same peptide MYLGYEYVTAIR, while the third to the fifth hits correspond to the same peptide LGEYGFQNALLVR if we follow the mass spectrometry convention not to distinguish Leucine from Isoleucine. After that, the next peptide has an E-value 1.5, indicating a false hit. One thing worth noticing is that there is a clean separation between significant hits and the rest of peptide hits
E-VALUE | PEPTIDE | MASS | DEFINITION | GI-LIST |
---|---|---|---|---|
4.423375e-05 | KMYLGYEYVTAIR | 1478.720 | ..|ref|NP 001054.1| transferrin [Homo sapiens] | [4557871,94717618,15021381,31415705,...... |
4.423375e-05 | RMYLGYEYVTAIR | 1478.720 | ..|emb|CAH91543.1| hypothetical protein [Pongo | [55729628] |
1.488740e-04 | KLGEYGFQNAILVR | 1479.780 | ..|ref|NP 033784.1| albumin 1 [Mus musculus] | [33859506,55391508,191765,19353306, ....... |
1.488740e-04 | KLGEYGFQNALLVR | 1479.780 | ..|emb|CAA59279.1| albumin precursor [Felis | [886485,57977283,633938,30962111, ...... |
1.488740e-04 | KLGEYGFQNALIVR | 1479.780 | ..|gb|AAT98610.1| albumin [Sus scrofa] | [51235682,52353352,15808978,76445989,...... |
1.526504e+00 | KTTLALQFLMEGVR | 1478.800 | ..|ref|YP 466151.1| putative circadian | [86159366] |
3.710973e+00 | [MFKANMKQLIVR | 1478.820 | ..|dbj|BAD64473.1| cell wall lytic activ | [56909946] |
. | . | . | . | . |
. | . | . | . | . |
. | . | . | . | . |
Theoretical peaks of two peptides MYLGYEYVTAIR and LGEYGFQNALLVR. Both peptides are found to be significant by RAId_DbS for a given query spectrum and were found to be partial sequences of proteins originally put in for the experiment. The right column lists the b ∪ y peaks of both peptides in ascending m/z order. The two sets of theoretical peaks only have two pairs that are within three daltons of each other. They are (175.12, 175.12) and (1019.45, 1017.58). This negligible overlap between theoretical peaks reinforces the possibility of co-elution of the two peptides during the experiment
Peptide/Mass | b ∪ y peaks (in ascending order) |
---|---|
MYLGYEYVTAIR 1478.72 | 132.04, 175.12, 288.2, 295.11, 359.24, 408.2, 460.29, 465.22, 559.36, 628.28, 722.42, 757.32, 851.46, 920.39, 1014.53, 1019.45, 1071.55, 1120.5, 1184.63, 1304.62, 1347.69 |
LGEYGFQNALLVR 1479.79 | 114.08, 171.11, 175.12, 274.19, 300.16, 387.27, 463.22, 500.36, 520.24, 571.39, 667.31, 685.44, 795.37, 813.38, 90 9.41, 960.56, 980.45, 1017.58, 1093.53, 1180.65, 1206.62, 1305.68, 1309.69, 1366.71 |
Accuracy of score pdf modeling
To address the accuracy of score pdf modeling, we define two spectrum-specific pdfs, data-derived pdf (Dpdf) and model pdf (Mpdf). For a given query spectrum, the former, Dpdf, is the normalized score histogram including contributions from both the true positive peptides and the false positive peptides; the latter, Mpdf, represents the pdf of only the false positives in the limit of very large number of qualified peptides. For example, we have derived the model Mpdf (eq. (17)) in this paper for the scoring function we used. However, in most cases, the forms of the Mpdf are assumed because analytical results for the Mpdf are difficult to obtain in general.
Ideally, the Mpdf should resemble very much the re-normalized score histogram after removing the true positives, at least in the region where the fluctuations are negligible compared to the corresponding Mpdf value. At the very high scoring tail, one typically does not have enough data to suppress the fluctuations and there may exist true positives that should not be counted towards Mpdf. Thus, one cannot use the tail region of the Dpdf to assign the statistical significance for peptides, an Mpdf extrapolated from high but not very high scoring region is needed for this purpose. This underscores the importance of the accuracy of the Mpdf as it heavily influences the statistical significance assignment. Note that one may wish to have only the large score part modeled faithfully as it is the region of primary interest. However, good agreement between the Dpdf and the Mpdf over a wider range of score does increase the confidence in the validity of the Mpdf. Furthermore, if one were to include a large range of score in Dpdf when fitting to Mpdf, the fluctuations from high scoring tail of Dpdf will not be sufficient to distort the overall Mpdf fit and one may just fit over the entire medium to large score region to obtain the Mpdf.
Because of its importance, for each search engine the accuracy of the Mpdf employed should be reported along with the E- or P-values for peptide hits when reporting the search results from a query spectrum. For a given query spectrum, if the Mpdf agrees well with the Dpdf, the reported statistics can be taken with confidence. On the other hand, if the agreement between the Mpdf and the Dpdf is poor, one may avoid taking the reported statistics literally. A quantification of fitting quality between the Dpdf and the Mpdf may therefore provide the users with valuable information in data interpretation. In this section, we will attempt to quantify the accuracy of the Mpdf in terms of how well it reflects the Dpdf.
Although there exist standard methods for characterizing the goodness/badness of fitting distribution function, not all of them have similar sensitivity or intuitive appeal. For example, as documented in the literature [27], χ^{2} tests often results in very small goodness numbers even for good models and one often needs to set the rejection threshold very low to avoid rejecting decent models. The Dpdf, derived from the score histogram, is discrete in nature and may be expressed as a list of pairs {S_{ i }, Dpdf(S_{ i })}_{ i }To emphasize the region of medium score to large score, it is better to work with the log-scale. That is, we will transform the list into {S_{ i }, ln [Dpdf(S_{ i })]}_{ i }. We introduce a short hand notation here: LDpdf(S) represents ln [Dpdf(S)] and similarly LMpdf(S) represents ln [Mpdf(S)]
The Mpdf, when taking values at {S_{ i }} will also form a list of pairs {S_{ i }, LMpdf(S_{ i })}_{ i }. If the Mpdf reproduces exactly Dpdf at those points, the pairs Γ ≡ {(LDpdf(S_{ i }), LMpdf(S_{ i }))}_{ i }when plotted on a plane will fall on the straight line x = y exactly. It is thus natural to ask how well the points in Γ fall on the x = y line and how strongly are the two sets {LDpdf(S_{ i })}_{ i }and {LMpdf(S_{ i })}_{ i }correlated. Fortunately, there exist two Student's t-tests that may serve these purposes [28]. We must emphasize that although these two t-tests are useful, there is definitely room for improvement in terms of quantification of the accuracy of Mpdf.
where B(α, ν) is the Beta function. This measure of goodness is intuitive and will allow the user to set a cutoff to prevent from using corrupted fitting results. We suggest to accept the Mpdf only if the goodness number is larger than 0.1. This should be contrasted with popular χ^{2} test where setting a goodness threshold at 10^{-3} or smaller is common [27].
with ν being the number of points in Γ less the number of fitting parameters of the Mpdf. The probability to arrive at correlation r, assuming that {Dpdf(S_{ i })}_{ i }and {Mpdf(S_{ i })}_{ i }are drawn from random, is given by
P_{ M }= 1 - A(t_{2}|ν).
In a way, P_{ M }may also be viewed as the probability that the Mpdf to be wrong. This observation has a nontrivial consequence in assigning statistical significance to peptide hits. It sets a limit on the lowest P-value one can get for a peptide hit, which we elaborate below.
If we have full confidence in a Mpdf, for a given peptide with score S, one may infer from the Mpdf a P-value (and consequently an E-value) for this hit. However, if our confidence in the Mpdf is not 100 percent, the statistics reported by the Mpdf may need adjustment. We propose below a simple way to do so. Let the P-value reported by the Mpdf for a peptide hit be P_{ h }, one may then view 1 - P_{ h }as the probability of correct identification. We may also view 1 - P_{ M }as the probability for the Mpdf to be correct. Thus, the probability of correct identification confidently supported by the Mpdf becomes (1 - P_{ h })(1 - P_{ M }). And the final P -value becomes
P_{h|M}= 1 - (1 - P_{ h })(1 - P_{ M }) = P_{ h }+ P_{ M }- P_{ h }P_{ M }.
Apparently, when P_{ M }approaches zero, that is, we have full confidence in the Mpdf, the final P-value reduces to P_{ h }. As an example of how this formulation may prevent exaggerated statistics, let us consider the case where P_{ h }= 10^{-50} and P_{ M }= 10^{-8}. Without eq. (23), one will infer a hit of very small P-value (10^{-50}). With eq. (23), we find that the final P-value, 10^{-50} + 10^{-8} - 10^{-58} = 10^{-8} + 10^{-50}(1 - 10^{-8}), to be greater than 10^{-8}. That is, one will not get a smaller P-value than P_{ M }.
However, one has to pay attention to that 1 - P_{h|M}represents the probability of correct identification supported by confident Mpdf. It is definitely possible that a method may identify the true peptide as the top hit but the Mpdf used may be very off. However, if this happens frequently for a given search engine, then it becomes hard to pool its search results due to the lack of a common statistical standard. That is, one can't set a priori an E-value cutoff that should represent the expected number of false positives found per spectrum. If one were to take just the top hit from each spectrum, depending on the spectral quality, one may ended up having many more true/false positives in one experiment than the others.
An example for computing fitting confidence. A randomly chosen spectrum is used to demonstrate the computation of the fitting confidence in detail. In each of the N = 28 numerical rows, the first entry is the score, the second entry records the LDpdf and the third entry corresponds to the LMpdf. Using the LDpdf as the x-coordinate and the Mpdf as the y-coordinate, we perform least square linear regression and find: an intercept value a = -0.00421 and a slope b = 0.9992. Eq. (18) is then used to compute t_{1} (t_{1} = 0.0421) and the goodness number, 1 - A(t_{1}|N -2), is found to be 0.96674 through (19). To test the strength of correlation between the second column and the third column, we use (20) to compute r and through (21) we find the t_{2} value to be 0.99567. Given r = 0.99567 and = 25, through (22) we find the P_{ M }value to be 2.58 × 10^{-27.}
S | ln [Dpdf(S)] | ln [Mpdf(S)] |
---|---|---|
0.0284661 | 0.479518 | 0.438266 |
0.0691319 | 0.431753 | 0.407608 |
0.109798 | 0.369235 | 0.351511 |
0.150463 | 0.2708 | 0.270076 |
0.191129 | 0.163419 | 0.163403 |
0.231795 | 0.014358 | 0.031592 |
0.272461 | -0.156812 | -0.125259 |
0.313127 | -0.340242 | -0.307054 |
0.353792 | -0.551264 | -0.513698 |
0.394458 | -0.79275 | -0.745095 |
0.435124 | -1.04746 | -1.00115 |
0.47579 | -1.34063 | -1.28178 |
0.516456 | -1.63587 | -1.58688 |
0.557121 | -1.96251 | -1.91636 |
0.597787 | -2.2322 | -2.27015 |
0.638453 | -2.72001 | -2.64814 |
0.679119 | -3.00809 | -3.05025 |
0.719785 | -3.52319 | -3.4764 |
0.76045 | -3.94211 | -3.92649 |
0.801116 | -4.31754 | -4.40045 |
0.841782 | -4.72005 | -4.89819 |
0.882448 | -5.27305 | -5.41962 |
0.923114 | -5.73387 | -5.96467 |
0.963779 | -7.04955 | -6.53326 |
1.00445 | -6.55707 | -7.1253 |
1.04511 | -7.368 | -7.74071 |
1.08578 | -9.44744 | -8.37942 |
1.12644 | -8.75429 | -9.04134 |
Concluding summary and outlook
We have designed a peptide identification method (RAId_DbS) using database searches. By taking into account the skewness in the peak intensity distribution of processed data, we have provided a theoretical derivation for the tail of the score distribution in the context of RAId_DbS's scoring scheme. The theoretical distribution agrees well with score statistics collected from each experimental spectrum. The E-value test performed indicates that RAId_DbS indeed provides realistic statistics. Quantitative tests on the agreement of our theoretical distribution and data-derived histogram have shown that RAId_DbS assigns accurate spectrum-specific statistical significance to peptide hits. The P-value obtained through (23) prevents exaggerated statistics in peptide identification, and thus may reduce protein misidentification for identification methods founded on peptide identification.
It seems that using RAId_DbS allows for theoretically characterized peptide score statistics without losing sensitivity, see Fig. 3. In addition, the use of profile mode in data acquisition seems to be valuable because of the higher probability of correct peptide identification. We have also found evidence that during an experiment it is possible for two charged peptides to be co-eluted and fragmented together, and their m/z peaks are logged in a mixed spectrum. In this context, RAId_DbS seems to be able to identify both peptides. This phenomenon actually discourages the use of heuristics that boost the separation between the best and the second best candidate peptides. This is because any method attempting such heuristics may be deprived of the possibility to capture two true peptides in a single spectrum.
Finally, we would like to say that there is room for improvement in RAId_DbS. For example, in the future, we would like to improve on the scoring scheme to enhance the sensitivity of RAId_DbS while keeping the characterizable statistics. In addition to improving the detecting power of RAId_DbS, we will also look at the possibility of combining RAId_DbS with other search methods. However, to be able to appropriately combine results from different methods, it is essential to build a common ground for score statistics. This important task will be performed and will be described in a separate publication.
Appendix – RAId_DbS implementation detail
The operation of RAId_DbS consists of three stages. The first step includes centroidizing m/z peaks followed by peak filtering. After this crucial step, RAId_DbS exhaustively scores all possible C- and N-tags of four amino acids. This helps RAId_DbS in filtering peptide candidates with NNTC before full scoring. In the third stage, as in many other MS^{2} analysis methods (be they the de novo type or database search type), RAId_DbS uses primarily the b- and y-series peaks for scoring. For each query spectrum, the collection of scores from all candidate peptides constitutes a score histogram, that is then used to determine the constant $\mathcal{C}$ of theoretical distribution, see Eq. (17). Once $\mathcal{C}$ is determined, one then integrates the pdf from infinitely large score back to a finite score S to obtain the spectrum-specific P-value for score S. This information is then used in conjunction with the effective database size to provide the E-value. In the following subsections, we describe each individual component, the sum of which constitutes RAId_DbS, followed by some details of implementation.
Peak processing
Peak processing can be roughly divided into three steps. In first step, precursor ion peaks and their associated one-dalton-cluster ions are removed from spectrum data. One-dalton-cluster ions associated with a peptide fragment of mass m' are members of a list of ions having masses given by {m' + Hyd, m' + 2Hyd, ...}, with Hyd being the mass of hydrogen (1.007825035 Da). For a parent peptide with mass m and charge q, the precursor and cluster ion peaks to be removed from the spectrum are those having their mass/charge peaks within 0.05 Da of [m + (q_{ i }- 1 + k) × Hyd]/q_{ i }for every q_{ i }= 1, ..., q, and k = 0, 1, ..., q_{ i }- 1.
Peak centroidizing is the second step of RAId_DbS's peak processing. In the centroidizing procedure, RAId_DbS first identifies what we term ε-clusters, then distills from each cluster either a single or multiple representative peaks depending on the noise level that we shall define shortly. An ε-cluster consists of a list of peaks, ordered according to their m/z values, for which any two neighboring elements have m/z difference no more than ε Da. The ε value usually depends on the instrument type used. The current default for ε value is 0.2 Da for low resolution spectra such as those produced from Linear Quadrupole Ion Trap (LTQ)/LQT experiment and is 0.05 Da for high resolution spectra such as those produced from Time of Flight (TOF)/TOF or Fourier Transform (FT)/FT. The noise level is currently defined heuristically. For each ε-cluster of p_{ ε }peaks, RAId_DbS uses the least intense 2p_{ ε }/3 peaks to compute the average intensity as well as the standard deviations. The noise level is then defined as the average intensity plus three standard deviations. A separate subcluster (a hill) is a subsequence of peaks whose intensities are greater than the noise level. Each subcluster is transformed to a separate peak: with m/z at the center of mass of the subcluster, and with intensity being the intensity of the strongest peak in the subcluster. The m/z peaks inside an ε-cluster with intensities less than the noise level are disregarded. When there are no hills present in an ε-cluster, one treat that ε-cluster as a single hill. This step is rather heuristic: we are still investigating possible avenues to improve this.
The third step is peak filtering. The idea is to keep only a finite number of informative peaks within a specified mass range, say ± x Da, regardless of where the center is. To be specific, RAId_DbS orders all the peaks produced from the centroidizing steps in two ways: in descending order of intensities and in ascending order of m/z. Going first to the strongest peak, RAId_DbS first makes sure that within 2ε Da, only one peak is retained. After that, depending on the charge state of the parent ion, RAId_DbS uses either x = 27 for single and doubly charged precursor ions or x = 27/(q - 1) for precursor ion with charge state q ≥ 3. RAId_DbS further normalize the peak intensity by a user-selected cutoff I_{ c }. Each peak intensity will then be multiplied by 1/I_{ c }and m/z peaks with normalized intensities less than one are removed. The current default is I_{ c }= 1. That is, no rescaling of the peak intensities.
De novo tag scoring
Besides allowing for any number of miscleavages, we also designed RAId_DbS to accommodate NNTC [17]. Allowing NNTC, however, introduces a huge excess number of peptides to be scored when searching in a database. In order to filter out peptides with higher chance to be the correct peptide, we implement a full de novo tag scoring to rank all possible de novo tags and only allow peptides with a high-scoring tag to enter scoring routine provided that the peptide considered has NNTC.
Using sequence tags to aid peptide identification is not a new concept. There exist, for example, several known methods [10, 29] that use sequence tags to mine candidate peptides in a database. Our use of sequence tags is distinguished from other methods by the following points. First, our sequence tag is used for the purpose of filtering out potential peptide candidates with NNTC [17], not used as a criterion for pooling candidate peptides. Second, for each spectrum we score all possible four amino acid tags (20^{4} for each terminal) and we keep many more tags, of order several thousands for each terminal, when compared with other tag-based method. Another reason for us to score tags is to provide a different foundation for de novo peptide sequencing using low resolution data. This direction, however, will be addressed in a separate publication.
All possible four amino acid tags are generated on the fly and scored (see scoring section of the paper for details) using m/z peaks after peak processing. RAId_DbS then ranks all the tags according to their score. However, it should be noted that in some low resolution experiments, the parent ion mass of a peptide reported by a mass analyzer can be as far off as two Da. To tolerate such a mass uncertainty, RAId_DbS actually scores each tag seven times, assuming the parent ion mass to be respectively -m_{ E }- 3, m_{ E }- 2, m_{ E }- 1, m_{ E }, m_{ E }, + 1, m_{ E }+ 2, m_{ E }+ 3- with m_{ E }being the parent ion mass provided by the mass analyzer from experiment. High-scoring tags, from each of the seven parent ion mass used, are pooled together to form two separate tag lists: one for each terminal. Note that it is possible that the highest scoring N-terminal tag is obtained by assuming parent ion mass to be m_{ E }+ 2 while the second best N-terminal tag is obtained by assuming parent ion mass to be m_{ E }-3, etc. With care, RAId_DbS can achieve this task in a few seconds.
Statistical assessment and implementation
For a given MS^{2} spectrum, RAId_DbS first scores all the possible de novo tags as described earlier. This step provides two high-scoring tag lists, one for C-terminal and one for N-terminal. After the tag scoring is done, RAId_DbS scans either a user-chosen or the default protein database for peptides with correct C-terminal cleavage and with matching molecular weights within 3 Da. When a qualified peptide appears multiple times while scanning through the database, RAId_DbS will combine them and only score the peptide once. A peptide with correct N-terminal cleavage will be automatically scored regardless of how many miscleavages are present. On the other hand, peptides with NNTC will be scored only if they contain a high-scoring tag, either from C-terminal or from N-terminal.
has the advantage of over estimating the effective database size for peptides with NNTC to compensate for the excessively small P-values. This may provide more accurate E-values for peptides with NNTC and good tag score. We leave the use of Eq. (25) as an option while keeping Eq. (24) as the default of RAId_DbS.
provided that β > 0 and y ≥ 0, the situation we encounter here. Consequently, Eq. (17) dictates that the maximum of the histogram occurs at y = 0, corresponding to S_{ u }= ⟨ln $\mathcal{I}$⟩. Therefore, RAId_DbS will leave the number of peaks of intensity one in the processed data as a parameter determined by the S_{u s}= ⟨ln I⟩. Note that in addition to its dependence on the spectrum considered, S_{u} may also depend on the database used. Thus the statistics provided by Eq. (17) will be spectrum-specific and may also be database-specific. Once the number of intensity one peaks is fixed, one may continue to compute the second and third cumulants of the ln I distribution from the processed spectrum. The constants β and ⟨x^{2}⟩ in Eq. (17) are thus fixed. Note that this procedure is applied regardless of whether the peak accuracy weight w_{ i }is turned on or off. However, when the number of theoretical peaks are variable, such as in the case of limiting only the molecular weights to be in a certain range, RAId_DbS treats both β and γ ≡ n/(6⟨x^{2⟩} β^{2}) as two additional variables to be determined from fitting the score histogram.
A peptide with correct N-terminal cleavage and with k miscleavages will be assigned an effective database N_{ c }(k). Similarly, a peptide with NNTC and with k miscleavages will be assigned an effective database size N_{ n }(k).
Reviewers' comments
Reviewer's report 1, first review comments
sent to the reviewers on July 26, 2007. Review received on September 11th, 2007.
Review by Wong Wing Cheong and Frank Eisenhaber, Bioinformatics Institute, Agency for Science, Technology and Research (A*STAR)
Matching of measured MS/MS spectra with theoretically calculated spectra for database peptides is the standard procedure for interpreting mass spectrometric data from protein samples. Alves et al. propose an alternative procedure for calculating E-values for peptide matches based on the expansion of the central limit theorem up to the third moment (skewness; not new though but with an elegant analytical derivation) to better fit the score distribution of a query spectrum.
General comments:
1) It should be noted that all methods that calculate scores and e-values with the actual intensities (here: I_{ i }in equation 1) suffer from the inaccuracies in the intensity measurements. The fragmentation and ionization efficiency for each peptide species given the same concentration can be very different from one another in terms of abundance in fragment ions. So it is difficult to decide whether the actual skewness in match scores' distribution is due to its natural form or due to insufficient sampling. To attempt to model skewness can be a double-edged sword in the sense that true hits for weaker spectra (for example, very incomplete b-y ion series) might be given higher p-values and marked as insignificant. As a result, one might lose true hits as false negatives though more noise spectra will also be thrown out. This effect should be examined more closely.
2) The existing methods are generally well performing for the identification of peptide hits in the case of spectra with sufficient clear signals. New methods can gain recognition in the field only for their ability to find significant hits even in more noisy or scarce spectra, not just for analytical elegance. For example, it was shown by Mujezinovic et al. (2006 Proteomics 6, 5117) show that the scores of hits dramatically improve if the amount of noise is reduced. The authors of this work attempt to demonstrate the performance of RAId_DbS to be more superior to others in a comparison but fall short of proving his point. Figures 2 and 3 do not provide the necessary arguments
3) In Fig. 2, the E-values of the various programs are directly compared, although the authors have previously declared that the E-values have to be transformed for this purpose (page 3 second paragraph). Thus, the figure is misleading. Curve clustering suggests that all the older methods calculate one kind of E-value whereas RAId_DbS belongs to another class of methods. Further, it would be sensible to consider only the reasonable E-value range << 1. Instead of using the nr-CR as described on page 11, it would be more appropriate to use nr with 90% or 80% similarity thresholds since there are many different classes of sequentially similar entries.
4) Figure 3 does not represent a true ROC curve. The absolute numbers of true and false positives appear to be calculated either with method-specific score or with method-specific E-value thresholds (note that SEQUEST does not provide E-values!!). One would expect the axes to show sensitivity (= TP/(TP + FN)) and specificity (= TN/(TN + FP)) instead. It should be possible to calculate and compare the area below the curves. For example at first glance, X!Tandem seems much inferior to RAId_DbS but when we look at Figure 3, their sensitivity and specificity performance appears quite similar. Especially in the practically important region of small false-positive assignments, the curves essentially overlay.
Minor
a) Will the newly derived pdf works better than other existing distributions (for example, the gamma distribution that can accommodate skewness)? With only a few measurements available, it is not clear how to decide what the naturally underlying distribution is in the far tail region. Understandably, the increase of the number of fitting parameters will improve the match with real score distributions.
b) The authors say that Fig. 1 was generated from a randomly picked spectrum. It appears rather that this spectrum is especially clean with large, almost complete b- and y-ion rows. To obtain an almost ideal fit of the score histogram and the fitted pdf is not a great surprise.
c) Summary: It would be advisable to provide a WWW-site for download of data and programs.
d) page 3, last paragraph: Please provide accession numbers instead of identifiers since only the latter are stable in time.
e) page 5, second paragraph: The weighting factors are purely heuristic without any fundamental justification.
f) page 5, penultimate paragraph: It is not good practice to refer to equation 17 after equation 1. This equation should be number 2.
g) page 6, 1st paragraph, equation (2) : This equation appears derived from a binomial distribution. Its information value at this point is unclear?!
h) page 14, reference list: It is not a good practice to refer to yet unpublished work; here, the authors have done it even twice.
Author's response to first review comments
Response and revised manuscript sent back to reviewers on September 17, 2007.
1) We believe there might be some misunderstanding here. First, we don't model the skewness as the skewness in the final score distribution is not determined based on the intensity profile of a given MS/MS spectrum. Note that the number of unit intensity peaks is used as a fitting parameter allowing for dynamic adjustment of the size of the skewness. We extended the central limit theorem to accommodate the case of small sample size. If the score distribution turned out to be symmetric, in our protocol, the skewness parameter β will have zero value as the best fit. That is, the statistics used in our method is intrinsically spectrum-specific, designed to take into account those effects mentioned by the reviewers. Regarding spectra of weak signals, it has been the major challenge of the field to extract the true hits from them. Our current method does not guarantee a better extraction, however, it will provide us with faithful E-values for each peptide hit given the main information content (b- and y-peaks) present in those spectra.
2) The fundamental reason to develop RAId_DbS is not just to develop another MS/MS database search tool. Rather, we want to provide a method capable of assigning correct E-values. The other existing methods might be well performing, but may not have accurate enough or even do not have E-value assignments. Fig. 2 demonstrates the E-value deviations of other search methods. Fig. 3 shows that RAId_DbS does not perform worse than other methods on top of having accurate E-value assignment.
3) The main point of Figure two is not to decide whether the E-values found among different methods are similar or not, instead, we are asking whether the E-value found by various methods are accurate or not. The standard definition of E-value gives rise to the theoretical curve in the figure. Agreement with the theoretical curve indicates faithful E-value assignment. Disagreement with the theoretical curve indicates inaccurate E-values. Since the number of spectra tested is only of order 10,000, one can't assess the E-value accuracy in the range of E < 10^{-4}. However, it is still possible to test the E-value range with E > 10^{-4} and E < 10^{2}. The large E-value portion is of course of low practical interest for reporting results from a single spectrum, but it does show the level of consistency between the reported E-values and the standard definition of E-value. This becomes particularly important when one needs to pool the results from various spectra together. Regarding the use of a database with similar entries purged, it is indeed a good suggestion for it mimicks better a random database. However, we find removing entries similar to target proteins sufficient to support our theoretical statistics, providing a larger range of usability of our statistics in real applications.
4) To emphasize the small false-positive assignment region, we replotted the ROC curves with the number of false positives plotted in log-scale while the number of true positives plotted in the linear-scale. Note that the ROC curve is a parametric plot of false positives versus true positives. The internal parameter used can be either the E-value or any monotonic function of the E-value. For methods not reporting E-values but scores, one assumes the score to be a monotonic function of the E-value and use the score as the internal parameter accordingly. Per the request of the reviewers, we have added the specificity versus sensitivity plots in Figure 3. Contrary to what the reviewers anticipated, we find that there is an even larger difference in performance between RAId_DbS and X! Tandem. In the figure legend, we also explain why this magnified difference may be an artifact due to low number of negatives reported by a specific method.
a) We anticipated Gaussian-type of distribution from our scoring construction. Theoretically speaking, if the number of theoretical peaks is large enough and the spectrum contains enough information, the score distribution will be very close to a Gaussian. The gamma distribution, although accommodates skewness, decaying with e^{-S}does not have the right tail. Our pdf, on the other hand, naturally interpolate (see our theory section) between the Gaussian and the case of finite skewness.
b) We are afraid there might be some misunderstanding here. Looking at the score S, corresponding to maximum pdf in the plot, we find S ~ 1. This implies that the majority of peptide entering the score histogram, if they are to have a complete b- and y-peaks present in the spectrum, will have peak intensity of only 2.718. Thus, it is highly unlikely that the spectrum has a complete b- and y- series for the candidate peptides. We had looked into a large number of score histograms from scoring various spectra, and all of them show excellent theoretical fits. Besides, if what is shown in Fig. 1 is a single best example, we won't be able to obtain E-value that traces theoretical curve so well (see Fig. 2).
c) We are in the process of setting up the website for RAId_DbS that allows the users to test run a few data set of their own and provide the options for code download.
d) We have added the accession numbers accordingly.
e) That is correct. This heuristic is to encourage fragment peaks with better m/z matching rather than just picking the m/z peaks of the largest intensities within the mass errors.
f) We agree with the opinion. However, since eq. 17 is the final results after going through eq.3 to eq. 16, it is rather hard to make it eq. 2. We therefore choose to keep the text as is and added a short text to remind the readers that it is the final results in the theory section that we are referring to.
g) Eq. (2) is another heuristic introduced to handle the case when the information level is too low, say, when there are too few peaks. When the weighted average number of evidence peaks found in the candidate list is below 2, there is simply not enough information to distinguish them well. In that situation, we switch to eq. (2) to obtain conservative p-values
h) We have removed one reference to the future and updated the other one.
Reviewer's report 1, second review comments
sent to reviewers on July 26, 2007. first review received on September 11th, 2007; first authors' response sent to the reviewers on September 17th; second review comments received on September 21st, 2007.
Please find attached the reply to your comments. Although we still have some reservations, we think that the MS has improved as a result of the revision, especially with regard to Figure 3. Therefore, we think that the MS might be published in Biology Direct given the comment on the "major challenge in MS" being reduced to E-value computation is removed. Other comments might be considered valuable by the authors, too.
1) There is no way around that any value calculated as a function of intensities in the MS/MS spectrum depends on the accuracy of the intensity measurements. Therefore, the proposed method possibly results in more accurate expectation values for sequence hits but the information in the underlying spectrum might be not fully reliable.
2) The authors claim that accurate E-value calculation is the major challenge in mass spectrometry. This is certainly not the case (see Hernandez et al. 2006 Mass. Spectrometry Rev. v. 25, 235–254), especially in the practically not relevant range of large E-values. At the same time, we acknowledge that the authors have provided an analytical method for calculating them. In our reply, we have discussed to which extent this specific finding might change the practice of protein mass spectrometry and we are in doubt.
3) It is obvious that E-values have been derived for many methods with less fundamental considerations and they were never thought to be more than a rough guide for assessing false positive numbers; here, the authors have made progress. Nevertheless, it remains to be seen to which extent the type of MS data (profile versus centroid) puts some of the methods into an inferior position. It would also be of interest to analyze reasons for the non-linearity of the RAId_DbS curve in the range of E-values at about 1 and below.
4) We appreciate that the authors have added true ROC plots. Again, the usage of profile data instead of centroid data might have put some algorithms into a difficult position since some operate only with a limited number of peaks for protein identification. Also, the great difference between the two plots (3c and 3d, which differ only in the data format and the size of the dataset) is not discussed. In the range of small scores, OMSSA is superior to RAId_DbS in 3d but much inferior in 3c.
f) Nevertheless, you force the reader to lookup the end of the article when he just as started reading it.
Author's response to second review comments
1) The reliability of the information contained in the spectrum may still be judged from the correct E-value reported for the best hit. If even the best hit has a poor E-value, it is most likely that the information content in the spectrum is hard to retrieve due to high noise level.
2) We thank the reviewer for acknowledging our progress in producing more accurate E-values in peptide search. The full power of having correct E-values indeed awaits to be seen. In anticipation that the issue of having correct statistics will become more and more important, we have added a section discussing the accuracy of fitting the score pdf in general and have put RAId_DbS to test in this regard.
3) From what we can see in this study, it seems that every method's detecting power increases when using the profile data. Perhaps the question one may ask is that how much can each individual method gain by using profile data. The developers of the search methods may wish to investigate this problem for their own sake. Regarding the minor horizontal leap in RAId_DbS's E-value near E = 1, it is mainly due to that RAId_DbS reports only the top 250 hits and this may cause a slight under representation of high E-value hits.
4) As we said earlier, in general each method gains some when using profile data. However, it is not guaranteed that the gain will be of similar size. Consequently wishing to observe a consistent performance across different data type may not be a correct anticipation.
f) We must apologize for the inconvenience that this may have caused the readers.
Reviewer's report 2
sent to the editorial board member/reviewer on July 23rd, 2007. Review received on August 15th, 2007. Review by Dongxiao Zhu (nominated by Arcady Mushegian) with contribution from Arcady Mushegian, Stowers Institute for Medical Research
1. While the aim of this paper is to derive score pdf for finite-size samples, it appears that infinite-size samples were assumed in eqs (7)-(9) and elsewhere. Please explain why this is so.
2. Is the result that the method performs when y is slightly more than 0, not only when y ≫ 1, consistent with the expectations?
3. Page 6, second paragraph from bottom, possible definition error: the variance of a random variable is its second central moment, not the difference between first and second moment. Ibid., should it not be y instead of y' to be consistent to the rest of the article?
4. Authors may ve advised to keep closer to standard statistical representations in their derivation, for example, in the second paragraph from bottom, E [X^{ k }] is preferable as a definition of the kth moment of X, and μ should be used to represent population mean instead of $\overline{X}$. (The latter is usually used for representing sample mean NOT population mean, and may be confusing).
5. There is no particular reason to have supplementary materials for this paper: the article will be much easier to read if they were incorporated into the main text.
Author's response
1. We only assume that n is large, not infinite. Note that if n is infinite, the second moment ⟨x^{2}⟩ will need to be infinite as well in order for (7)-(9) to make sense. And we apparently did not assume that.
2. Yes. As explained in the last paragraph of the theory section, the validity of the derivation is lost when 1 + 2βy becomes near zero. The accuracy of the derived expression, however, becomes better when y becomes larger. What is said there is that when y ≫ 1, in the high scoring tail, the statistical significance assigned through our derived formula will work better.
3. We are sure it is a misunderstanding. The variance of a random variable is its second moment only if the random variable has zero mean. In general, the variance is the difference between the second moment and the second power of the first moment.
4. We thank the reviewer for the suggestion. However, we decide to keep our original notation since both type of notations are well accepted.
5. A very good suggestion indeed. We have moved the supplementary materials into the appendix of the paper to form a single article.
Reviewer's report 3
sent to the reviewer on August 2nd, 2007.
Review received on October 2nd, 2007. Review by Shamil Sunyaev, Harvard Medical School
This manuscript introduces a new method for peptide identification for mass spectrometry proteomics. There is a growing need for more accurate computational methods for shotgun proteomics and the presented method has a high chance of becoming a popular tool in the field. The manuscript presents a derivation of an accurate p-value for the search using an asymptotic extension of the central limit theorem for the case of non-negligible skewness. Although the manuscript is technical in nature, due to the potential importance for applications, it fits Biology Direct. The manuscript is very well written and the presentation is clear.
Declarations
Acknowledgements
We thank Dr. Ning Zhang for providing the data in [19]. We thank the administrative group of the NIH biowulf clusters, where all the computational tasks were carried out. GA wishes to thank Susan Chacko for her help with running Mascot. We also thank the personnel of Proteomics core facility of NHLBI for the experimental support. This work was supported by the Intramural Research Program of the National Library of Medicine at National Institutes of Health. Funding to pay the Open Access publication charges for this article was provided by the NIH.
Authors’ Affiliations
References
- Eng JK, McCormack AL, Yates JR III: An approach to correlate tandem mass spectral data of peptides with amino acid sequences in a protein database. J Amer Soc Mass Spectrom 1994, 5: 976-989. 10.1016/1044-0305(94)80016-2View ArticleGoogle Scholar
- Clauser KR, Baker PR, Burlingame AL: Peptide fragment-ion tags from maldi/psd for error tolerant searching of genomic databases. Proceedings of the 44th ASMS Conference on Mass Spectrometry and Allied Topics: 12–16 May 1996; Portland, Oregan 1996, 365.Google Scholar
- Bafna V, Edwards N: SCOPE: a probabilistic model for scoring tandem mass spectra against a peptide database. Bioinformatics 2001,17(Suppl 1):S13-S21.PubMedView ArticleGoogle Scholar
- Johnson RS, Taylor JA: Searching sequence databases via de novo peptide sequencing by tandem mass spectrometry. Mol Biotech 2002, 22: 301-315. 10.1385/MB:22:3:301View ArticleGoogle Scholar
- Hernandez P, Gras R, Frey J, Appel RD: Popitam: Towards new heuristic strategies to improve protein identification from tandem mass spectrometry data. Proteomics 2003, 3: 870-878. 10.1002/pmic.200300402PubMedView ArticleGoogle Scholar
- Zhang N, Aebersold R, Schwikowski B: A probabilistic algorithm to identify peptides through sequence database searching using tandem mass spectral data. Proteomics 2002, 2: 1406-1412. 10.1002/1615-9861(200210)2:10<1406::AID-PROT1406>3.0.CO;2-9PubMedView ArticleGoogle Scholar
- Sheng QH, Tang HX, Tao X, Wang LS, Ding DF: A Novel Approach for Peptide Identification by Tandem Mass Spectrometry. Acta Biochem Biophys Sinica 2003, 35: 735-740.Google Scholar
- Perkins DN, Pappin DJC, Creasy DM, Cottrell JS: Probability-based protein identification by searching sequence database using mass spectrometry data. Electrophoresis 1999, 20: 3551-3567. 10.1002/(SICI)1522-2683(19991201)20:18<3551::AID-ELPS3551>3.0.CO;2-2PubMedView ArticleGoogle Scholar
- Field HI, Fenyo D, Beavis RC: RADARS, a bioinformatics solution that automates proteome mass spectral analysis. Proteomics 2002, 2: 36-47. 10.1002/1615-9861(200201)2:1<36::AID-PROT36>3.0.CO;2-WPubMedView ArticleGoogle Scholar
- Tanner S, Shu H, Frank A, Wang LC, Zandi E, Mumby M, A PP, Bafna V: InsPecT: Identification of Posttranslationally Modified Peptides from Tandem Mass Spectra. Anal Chem 2005, 77: 4629-4639. 10.1021/ac050102dView ArticleGoogle Scholar
- Geer LY, Markey SP, Kowalak JA, Wagner L, Xu M, Maynard DM, Yang XWS, Bryant SH: Open Mass Spectrometry Search Algorithm. J Proteome Res 2004, 3: 958-964. 10.1021/pr0499491PubMedView ArticleGoogle Scholar
- Craig R, Beavis RC: TANDEM: matching proteins with tandem mass spectra. Bioinformatics 2004, 20: 1466-1467. 10.1093/bioinformatics/bth092PubMedView ArticleGoogle Scholar
- Cristoni S, Bernardi LR: Bioinformatics in mass spectrometry data analysis for proteomics studies. Expert Rev Proteomics 2004, 1: 469-483. 10.1586/14789450.1.4.469PubMedView ArticleGoogle Scholar
- Kapp EA, Schütz F, Connolly LM, Chakel JA, Meza JE, Miller CA, Fenyo D, Eng JK, Adkins JN, Omenn GS, Simpson RJ: An evaluation, comparison, and accurate benchmarking of several publicly available MS/MS search algorithms: sensitivity and specificity analysis. Proteomics 2005, 5: 3475-3490. 10.1002/pmic.200500126PubMedView ArticleGoogle Scholar
- Fenyo D, Beavis RC: A Method for Assessing the Statistical Significance of Mass Spectrometry-Based Protein Identification Using General Scoring Schemes. Anal Chem 2003, 75: 768-774. 10.1021/ac0258709PubMedView ArticleGoogle Scholar
- Sadygov R, Wohlschlegel J, Paul SK, Yates JRI: Central Limit Theorem as an Approximation for Intensity-Based Scoring Function. Anal Chem 2006, 78: 89-95. 10.1021/ac051206rPubMedView ArticleGoogle Scholar
- Alves G, Yu YK: Robust accurate identification of peptides (RAId): deciphering MS^{ 2 }data using a structured library search with de novo based statistics. Bioinformatics 2005, 21: 3726-3732. 10.1093/bioinformatics/bti620PubMedView ArticleGoogle Scholar
- Alves G, Ogurtsov A, Wu WW, Wang G, Shen RF, Yu YK: Calibrating E-values for MS^{ 2 }Library Search Methods. [Submitted to Biology Direct].
- Keller A, Samuel P, Nesvizhskii AI, Stolyar S, Goodlett DR, Kolker E: Experimental Protein Mixture for Validating Tandem Mass Spectral Analysis. OMICS 2002, 6: 207-212. 10.1089/153623102760092805PubMedView ArticleGoogle Scholar
- Wu WW, Wang G, Yu MJ, Knepper MA, Shen RF: Identification and Quantification of Basic and Acidic Proteins Using Solution-Based Two Dimensional Protein Fractionation and Label-free or 180-labeling Mass Spectrometry. J Proteome Res, in press.
- Wang G, WW W, Zeng W, Chou CL, Shen RF: Label-Free Protein Quantification Using LC-Coupled Ion Trap or FT Mass Spectrometry: Reproducibility, Linearity, and Application with Complex Proteomes. J Proteome Res 2006, 5: 1214-1223. 10.1021/pr050406gPubMedView ArticleGoogle Scholar
- Wu WW, Wang G, Baek SJ, Shen RF: Comparative Study of Three Proteomic Quantitative Methods, DIGE, cICAT, and iTRAQ, Using 2D Gel- or LC-MALDI TOF/TOF. J Proteome Res 2006, 5: 651-658. 10.1021/pr050405oPubMedView ArticleGoogle Scholar
- Syka JE, Coon JJ, Schroeder MJ, Shabanowitz J, Hunt DF: Peptide and protein sequence analysis by electron transfer dissociation mass spectrometry. Proc Natl Acad Sci USA 2004, 101: 9528-9533. 10.1073/pnas.0402700101PubMedPubMed CentralView ArticleGoogle Scholar
- Feller W: An Introduction to Probability Theory and Its Application. New York: John Wiley & Sons; 1957.Google Scholar
- Petrov VV: Limit Theorems of Probability Theory. Oxford: Claredon Press; 1995.Google Scholar
- Masselon C, Pasa-Tolic L, Lee SW, Li L, Anderson GA, Harkewicz R, Smith RD: Identification of trypic peptides from large databases using multiplexed tandem mass spectrometry: simulations and experimental results. Proteomics 2006, 3: 1279-1286. 10.1002/pmic.200300448View ArticleGoogle Scholar
- Press WH, Teukolsky SA, Vetterling WT, Flannery BP: Numerical Recipes in. 2nd edition. C. Cambridge: Cambridge University Press; 1999.Google Scholar
- Hoel PG: Introduction to Mathematical Statistics. 5th edition. New York: John Wiley & Sons; 1984.Google Scholar
- Mann M, Wilm M: Error-Tolerant Identification of Peptides in Sequence Database by Peptide Sequence Tags. Anal Chem 1994, 66: 4390-4399. 10.1021/ac00096a002PubMedView ArticleGoogle Scholar
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.