 Research
 Open Access
 Published:
A distributionfree convolution model for background correction of oligonucleotide microarray data
BMC Genomics volume 10, Article number: S19 (2009)
Abstract
Introduction
Affymetrix GeneChip^{®} highdensity oligonucleotide arrays are widely used in biological and medical research because of production reproducibility, which facilitates the comparison of results between experiment runs. In order to obtain highlevel classification and cluster analysis that can be trusted, it is important to perform various preprocessing steps on the probelevel data to control for variability in sample processing and array hybridization. Many proposed preprocessing methods are parametric, in that they assume that the background noise generated by microarray data is a random sample from a statistical distribution, typically a normal distribution. The quality of the final results depends on the validity of such assumptions.
Results
We propose a Distribution Free Convolution Model (DFCM) to circumvent observed deficiencies in meeting and validating distribution assumptions of parametric methods. Knowledge of array structure and the biological function of the probes indicate that the intensities of mismatched (MM) probes that correspond to the smallest perfect match (PM) intensities can be used to estimate the background noise. Specifically, we obtain the smallest q2 percent of the MM intensities that are associated with the lowest q1 percent PM intensities, and use these intensities to estimate background.
Conclusion
Using the Affymetrix Latin Square spikein experiments, we show that the background noise generated by microarray experiments typically is not well modeled by a single overall normal distribution. We further show that the signal is not exponentially distributed, as is also commonly assumed. Therefore, DFCM has better sensitivity and specificity, as measured by ROC curves and area under the curve (AUC) than MAS 5.0, RMA, RMA with no background correction (RMAnoBG), GCRMA, PLIER, and dChip (MBEI) for preprocessing of Affymetrix microarray data. These results hold for two spikein data sets and one real data set that were analyzed. Comparisons with other methods on two spikein data sets and one real data set show that our nonparametric methods are a superior alternative for background correction of Affymetrix data.
Introduction
Affymetrix GeneChip^{®} arrays are widely used in biological and medical research to estimate gene expression levels. Each gene is interrogated using 11–20 probe pairs (depending on the platform), each of which consists of a perfect match (PM) and a mismatch (MM) probe. PM probes are sequences of 25 nucleotides that are intended to be a perfect complement to a subsequence of the target transcript of interest (gene). A MM probe is also 25 nucleotides in length, with the same composition as the corresponding PM probe, except that the middle base (13th) is changed to its WatsonCrick complement. The MM probes were originally designed to be different at one base pair so that their intensities could be subtracted from those of the PM as a measure of nonspecific hybridization.
In order to estimate gene expression values and perform highlevel analyses, such as classification and clustering, probelevel preprocessing of the data is necessary. Typically, there are three steps of preprocessing: background correction, normalization and summarization, although not necessarily in that order. It has been argued that background correction is the most crucial step for probe level processing [1, 2]. Thus, it is important to understand the assumptions underlying background correction methods, and test those assumptions, before blindly applying any preprocessing method.
One popular method, Robust Multichip Average (RMA) uses an exponentialnormal convolution model for background correction, quantile normalization for the normalization step, and a median polish algorithm to summarize probe level values into a single expression value per gene [3]. Some software packages allow the user to interchange background correction methods with the normalization and summarization methods (e.g. Bioconductor [4]).
The exponentialnormal convolution model is given by X = S + Y, where X is the observed PM intensity for a probe on the array, S is the true signal, assumed to have an exponential distribution with rate parameter α, and Y is normally distributed background noise [3]. The normal noise distribution is truncated at zero so that the model does not return negative intensity values. Background correction involves estimating the parameters μ and σ of the normal distribution and the rate parameter α of the exponential distribution. In practice these parameters cannot be estimated by conventional methods, such as maximum likelihood [1]; therefore, the implementation of RMA background correction in Bioconductor [4] uses an ad hoc method. We have previously shown that this method returns poor parameter estimates [5].
The exponentialnormal convolution model is built on the reasonable assumption that fluorescence intensities from a microarray experiment are composed of both signal and noise, and that the noise is ubiquitous throughout the signal distribution. A convolution model of a signal distribution and a noise distribution is a natural choice in such a situation. The choice of a normal distribution for the background noise and an exponential distribution for the signal was likely made for two reasons. First, density estimates of raw PM intensities from the Affymetrix Latin Square spikein data sets show a rightskewed curve with what looks like a long exponential tail (see Figure 1). Second, the normal and exponential distributions are easy to manipulate mathematically in order to obtain a closed form for the expectation of the signal given the observed values, which is necessary for parameter estimation. However, Figures 2 and 3 in the next section show that the convolution of a normal and an exponential distribution is not generally a good fit for microarray data. These observations, plus the difficulty of checking assumptions and estimating parameters, motivate a nonparametric background correction method.
The objective of this paper is to introduce a new background correction method, called Distribution Free Convolution Model (DFCM). The model has the same form as the exponentialnormal convolution model (X = S + Y), except that no distributional assumptions are made on the signal (S) of the noise (Y). The mean and variance for the noise distribution are estimated using MM probe intensities in a novel way that is more consistent with their biological and structural characteristics. The signal is given by the PM intensities once the estimated background has been subtracted (as explained in Methods); therefore, there is no need to estimate parameters for the signal. We compare DFCM against RMA, RMA with no background correction (RMA – noBG) [6], GCRMA [7], MAS 5.0 [8], dChip [9, 10], and PLIER [11]. In general, DFCM outperforms these other methods for two different spikein data sets and a real data set involving the role of CD40 in regulatory networks in human B cells [12].
Methods
Estimating the distributionfree convolution model
Let X = S + Y, where X = observed PM intensities, S = true intensity signal, and Y = background noise. The DFCM algorithm for background correction proceeds as follows:

1.
Obtain the lowest q1 percent PM intensities. q1 is typically a small number (less than 30%). Information on the estimation of q1 is given in the next section.

2.
Obtain lowest q2 percent (typically 90% or 95%) of MM intensities associated with the PMs obtained in step 1. These MM intensities are a measure of background noise, and will be termed "noise" in the sequel.

3.
Use a nonparametric density estimate of the lowest q2 percent of the MM intensities to find the mode of the noise distribution. By default, the DFCM algorithm uses an Epanechnikov kernel estimate. Consider this mode as an estimate of the mean of the noise distribution. We denote this estimate as .

4.
Estimate the standard deviation of the background noise by calculating the sample standard deviation of the noise for values that are smaller than . Then is the sample standard deviation multiplied by . The square root of 2 enters into the estimation of sigma due to the use of only one side of the noise distribution (those values less than the mean) to estimate the standard deviation.

5.
Obtain backgroundcorrected PM intensity values for the k^{th}probe (k = 1,..., K) in the i^{th}probe set (gene), , using the following formula, where min denotes the minimum intensity (PM or MM). Here, x_{ ki }denotes noncorrected PM intensity values. Let
Therefore, for large enough values of x_{ ki }, we correct for background by subtracting the estimated mean of the background noise. For small values of x_{ ki }, the background corrected PM intensities are determined by a linear interpolation, where the slope is a function of the background mean and standard deviation. The two equations will give the same result if .
After background correction, any method of normalization or summarization can be used. For the purposes of this paper, quantile normalization and median polish summarization are used for DFCM.
Choosing q1 and q2
Background noise is estimated using a select set of MM probe signals that are not likely to include effects of nonspecific hybridization to the PM target transcript or crosshybridization to a related target transcript. In choosing q_{1} and q_{2}, we are attempting to choose PM values that are small enough so that nonspecific MM hybridization to a PM target is not a problem, and then choosing a subset of MM values that are unlikely to be affected by crosshybridization.
The parameter q_{1}can be thought of as a measure of the percentage of PM probes that recognize genes that are not expressed in the data set, based on the assumption that in any given biological sample there will be a subset of genes that are not expressed at a detectable level. Any signal detected for MM probes for these genes cannot be due to nonspecific hybridization because the gene is not expressed, based on the low PM values.
To choose the parameter q_{1}, we developed an algorithm which calculates q_{1} such that the proportion of MM intensities greater than the PM intensities for the smallest q_{1}% of the data is approximately 50%. We believe that one of the reasons that MM intensities are sometimes greater than their corresponding PMs is nonspecific hybridization. Therefore, in estimating q_{1}, we obtain a measure the percentage of nonspecific hybridization in the data set.
Data sets used for comparison
The two Affymetrix LatinSquare spikein data sets (HGU113A and HGU95Av2) each contain several spikedin transcripts in known locations on a set of chips. These data sets, and a detailed description of the Latin Square design are available at http://www.affymetrix.com/support/technical/sample_data/datasets.affx. Affymetrix has reported that certain probe pairs for transcripts 407_at and 36889_at had been found to perform poorly in the HGU95Av2 spikein data. In addition, other researchers have found that the number of spikein probe sets should be 16 instead of 14. Two articles [6, 13] report that probe set 546_at should be considered with the same concentration as 36202_at since both of them were designed against the target Unigene ID Hs. 75209. Further, probe set 33818_at should be included as a spiked transcript in the 12th column of the Latin square design. Our definition of spikeins for the HGU95Av2 data includes all four of the above mentioned probes, resulting in a total of 18 spikedin transcripts.
The HGU133A experiment differs from the HGU95Av2 experiment in several important ways. First, the HGU133A experiment consists of 42 specific transcripts that are spiked in at 14 concentrations ranging from 0 pM to 512 pM, again arranged in a Latin Square design. Therefore, there is a finer gradation of concentrations used than in the HGU95Av2 experiment. Also, there are three transcripts spikedin at each concentration and three replicate arrays for each experiment, thus a total of 42 arrays. For convenience, we will call the triples of probe sets that recognize transcripts spikedin at the same concentration "groups".
Recently, the HGU133A data has also been examined for the presence of additional spikeins [14]. Twentytwo additional spikedin transcripts were found. Most of the "new" spikeins are variants of the original spikein probe sets, or share a large percentage of probe sequences in common with original spikeins. For example, the probe sets initially described as recognizing bacterial controls (e.g. AFFXLysX3_at, AFFXDapX3_at, AFFXPheX3_at) are targeted at the 3' end of the gene (hence the notation "3" in the name of the probe set). It makes sense that the probe sets recognizing the 5' and middle sections of the same genes would behave as spikeins, since the target RNA mixture for hybridization is likely to be made up sequences covering the 5' end and middle regions of the genes. Indeed, the use of 22 additional spikeins in Receiver Operating Characteristic (ROC) curve plots and Area Under the Curve (AUC) calculations improved the sensitivity and specificity of RMA, RMA with no background correction (RMAnoBG), MAS 5.0, PLIER, and dChip. GCRMA performed slightly worse with the use of all 64 spikeins. For the ROC curves and AUC calculations that follow, we use a total of 64 spikeins for the HGU133A data (42 original spikeins plus 22 "new" spikeins)..
We use a third spikein experiment to examine the distributional assumptions of the exponentialnormal convolution model [2]. This series of spikein experiments was run on the DrosGenome1 chip, and has been named the GoldenSpike experiment. In addition to targeting a different organism than the Affymetrix spikein data, the GoldenSpike experiment contains 1331 spikedin transcripts whose levels are varied and 2,551 RNA species whose levels are held constant between the control and test array sets. The large number of spikedin transcripts allows for more accurate estimates of the false positive and false negative rates and provides an RNA mix that more closely resembles total cellular RNA. Furthermore, no transcript targets were included for approximately twothirds of the probe sets, allowing for an accurate definition of background data. In contrast, Affymetrix uses an uncharacterized RNA background for their spikein data sets. Lastly, the fold differences between the test and control array sets for some of the spikein transcripts are very low (1.2 fold), which allows an estimate of the reliability and sensitivity of detection of small fold differences.
While this data set was used to examine the distributional assumptions of the exponentialnormal convolution model, these data were not used for evaluation of the relative performance of DFCM versus other algorithms due to controversy surrounding the use of the GoldenSpike dataset for method comparison. It has been observed that the GoldenSpike experiment uses technical replicates of a single experiment, rather than biological replicates. Thus, random variability in the experiment is confounded with real signal [15]. Others have found that features spikedin at a 1:1 ratio tend to have different behavior for the control and spikein experiments [16]. For these reasons, we restricted our comparisons of ROC curves and AUC calculations to the two Affymetrix Latin Square data sets.
Examining distributional assumptions
In order to test the validity of the noise and signal distributional assumptions, we compared background noise distribution estimated by the exponentialnormal convolution model with the standard normal distributions in both quantilequantile (QQ) plots and density plots using the Affymetrix Latin Square spikein data sets. All calculations were done using the Bioconductor suite in the R software package for statistical analysis [4].
Quantilequantile (QQ) plots are designed to compare the distributions of two data sets usually a "gold standard" and a test data set. Sometimes, the gold standard consists of simulated values from a distribution of interest (e.g. the normal distribution), and sometimes it is simply data observed from another experiment. If the gold standard is simulated from a known distribution, the purpose of the plot is to see if the observed data have that particular distribution. The sorted values for one data set (quantiles) are plotted on the horizontal axis, and the sorted values of the other data set on the vertical axis. If the plot results in a straight line, then this is evidence that the two data sets have the same distribution.
We also examined the assumption of a normal background distribution using three normality tests: ShapiroWilk, AndersonDarling, and KolmogorovSmirnov [17, 18], as implemented by the R software package [19]. For each of the spikein data sets, the background noise was estimated using the Bioconductor implementation of RMA background correction [4]. Once the noise vector was estimated, a random sample of length 100 was taken and the tests were applied to this vector. This was done because normality tests can be extremely sensitive to sample size, often rejecting the null hypothesis of normality just because the sample size is extremely large. A sample size of 100 is large enough to have reasonable power against some alternatives, but not so large that the tests would reject in error [20]. The samples were submitted to each of the three tests 1000 times, and the pvalues for each iteration recorded. The results are given in Table 1.
ROC curves and Area Under the Curve (AUC)
In order to compare the performance of DFCM versus currently available methods, we examined ROC curves and AUC for the two Latin Square spikein data sets mentioned previously. We tested the performance of DFCM against RMA, RMAnoBG, GCRMA, MAS, dChip, and PLIER. All data files were preprocessed together for each method. For the Affymetrix data sets, we compared pairs of experiments that were separated by the same number of permutations of the Latin Square (where d = number of permutations), and obtained average true and false positive rates for each preprocessing method for each value of d, d = 1,..., 7. In these Latin Square designs, d can be thought of as the log2 fold difference in spikein transcript levels for a majority of the transcripts. For example, for the HGU133A data set, experiments 1 and 2, 2 and 3, 3 and 4, etc. are separated by one shift in the Latin Square design; therefore, d = 1 for these pairs. For twelve groups of spikedin transcripts (there are three spikein transcripts per concentration group in the HGU133A experiment) in each of these fourteen pairs of experiments, there is a 2fold difference in concentration. Similarly, experiments 3 and 5, 4 and 6, and 5 and 7 are separated by two permutations in the Latin Square design; therefore, d = 2. Eleven spikein groups have fold changes of 2 on the log base 2 scale between pairs of experiments.
We compared experiments with d = 1 through d = 7, since d = 8 is equivalent to d = 6, d = 9 equivalent to d = 5, and so on. AUC calculations were done for a cutoff of 100 false positives for the HGU95Av2 experiment, and 200 false positives for the HGU133A experiment. These cutoff points correspond to a false positive rate of approximately 0.8% for both experiments.
Again, the GoldenSpike data was not used for methods comparison due to serious design flaws, described fully in [15, 16].
Results
Testing distributional assumptions for the convolution model
In order to test the validity of the noise and signal distributional assumptions, we compared background noise distribution estimated by the exponentialnormal convolution model with the standard normal distribution in both quantilequantile (QQ) plots and density plots using the Affymetrix spikein data sets. All calculations were done using the Bioconductor suite in the R software package for statistical analysis [4]. R code is provided in Additional file 1.
Figure 2a shows a QQ plot of the estimated background noise for four randomly selected experiments (and their replicates, for a total of 12 arrays) from the HGU95A spikein data. The plot is given on the original scale, since the assumption of normal background noise is applied to the probelevel intensities on the original scale. The background was estimated using the RMA background correction method as coded in the affy package of Bioconductor [4]. According to the assumptions of the exponentialnormal convolution model, the background noise should have a truncated normal distribution. Therefore, a plot of the background noise estimated using the convolution model versus values simulated from a truncated normal distribution should produce a straight line. In Figure 2a, there are several values deviating from a straight line in the lower left corner of the graph, and the line is bent slightly. However, both of these departures are small. For this data set, assumption of normality for the background noise seems to be reasonable.
Figure 2b is a QQ plot of the background corrected PM intensities (on the original scale) versus quantiles from an exponential distribution for the same data set. The rate parameter used for the exponential distribution is equal to the estimated rate parameter of the signal given by the affy package. The QQ plot for the background corrected (signal) intensities does not show a straight line; in fact, it shows that the distribution of the signal is much heavier tailed than one would expect if the data were exponentially distributed. This suggests that either the exponential model is not a good one for the signal from the PM intensities, or the background correction algorithm is flawed. Indeed, given the heterogeneity of the variances for the intensity level of each gene, we would not expect a clean fit to any distribution, which further bolsters our argument for the application of a nonparametric background correction method.
Figure 2c shows density estimates of the observed log base 2 PM intensities (solid line) and the same intensities after background correction with the exponentialnormal convolution model (dashed line). The background corrected intensities should exhibit an exponential distribution. However, the signal from these data has two modes, suggesting that the estimated signal is composed of a mixture of two or more distributions rather than a single exponential distribution, at least for this data set. This density estimate suggests that there are two groups of genes in this data set – genes that are expressed at low levels, and fewer genes expressed at higher levels.
Figure 2d shows the same density estimate of the original PM intensities that was seen in plot 4c, but now this density is plotted against a density consisting of a simulated convolution of a truncated normal and an exponential, using parameters estimated by the background correction algorithm given in Bioconductor. The parameters for the normal and exponential distributions were obtained using estimates given by the Bioconductor implementation of RMA. The estimation procedure for the convolution model produces a decent estimate of the mean, but is not accurate for the rate parameter.
The results shown for the HGU95Av2 spikein data apply to the HGU133A spikein data, with one notable exception. Figure 3a shows the QQ plot for 3 randomly selected experiments (and their replicates, for a total of nine arrays) from the HGU133A experiment. Clearly, the background as estimated by the exponentialnormal model does not have a normal distribution, since the QQ plot does not display a straight line. In addition, once the data are background corrected, the resulting distribution is not exponential (Figure 3b). Figure 3c shows a QQ plot of estimated background noise data from the GoldenSpike experiment [2]. This plot seems to support a normally distributed background, but not an exponentially distributed signal (Figure 3d).
Table 1 gives the results of the three tests of normality for the estimated background noise for all three data sets. Tests were done using probelevel data on the original scale. We calculated the number of pvalues that were less than 0.05, in order to ascertain how often each test rejected. If the null hypothesis were true, we would expect rejections approximately 5% of the time. For all of the results, the rejection rates are much higher, indicating that the data are not at all normally distributed. We also give the minimum, median, and maximum of the 1000 pvalues calculated for each test. The minimum pvalues are all 0 to at least five decimal places. Median pvalues are typically less than 0.001, again indicating that the tests reject often. The fact that the low power KS test at n = 100 yields such a preponderance of small pvalues is convincing evidence against normality.
Downstream performance of DFCM
The quantilequantile plots provide evidence that the exponentialnormal convolution model does not fit the data. These observations lead us to develop the DFCM as a means of background correction that does not rely on specific distributional assumptions. The estimation of noise and signal using DFCM is described in the Methods section. In order to determine if DFCM leads to an improvement of background correction and signal estimation, we applied DFCM and the other methods to each of the Affymetrix Latin Square data sets, and evaluated their performance characteristics using ROC/AUC analysis. The GoldenSpike data set was not used for this comparison because of the controversy around this data set [15, 16]. Once the background is corrected using DFCM, the data are normalized using quantile normalization and summarized with median polish.
In order to compare the performance of DFCM with other commonly used methods, we examined ROC curves and AUC analyses for the two Latin Square spikein data sets mentioned previously. We tested the performance of DFCM against RMA, RMAnoBG, GCRMA, MAS, dChip, and PLIER. All data files were preprocessed together for each method using a PowerMac G5 running R Cocoa GUI with R version 2.8.1 [21]. For the Affymetrix data sets, we compared pairs of experiments that were separated by the same number of permutations of the Latin Square (where d = number of permutations), and obtained average true and false positive rates for each preprocessing method for each value of d, d = 1,..., 7. A more detailed description of the parameter d is given in the methods section.
Figure 4 shows the ROC curves generated from results of analysis to identify differentially expressed genes using various methods on the HGU95Av2 spikein data. For Figure 4a, d = 1, and in Figure 4b, d = 2. In this case, q1 = 30% and q2 = 90% for DFCM. For these spikein data sets, true positive and false positive results can be determined based on the nature of the Latin square design. DFCM and GCRMA both perform well for this data set based on AUC analyses.
Figure 5 shows the ROC curves generated from the HGU133 data for d = 1 (a) and d = 2 (b). For these data, DFCM outperforms all versions of RMA. Recall that the normal distribution is not a good fit to the background noise as estimated by RMA (Figure 3a). In this case, a nonparametric approach works better because there is no distributional assumption on the background. Other contributing factors could be larger number of the spikein transcripts (64 for the HGU133 data versus 16 for the HGU95 data), and the different chip platform. Since the exponentialnormal convolution model was developed before the HGU133 spikein data was available, it may be the case that the model was optimized to perform well on the HGU95 spikein data.
Table 2 reports the average AUC values for these methods based on two spikein data sets and different values of d. The results show that DFCM performs best for most of the time. For example, except for d = 2, DFCM outperforms all other methods on the HGU95 spikein data. For HGU133 spikein data, DFCM has the largest AUC values for all situations except d = 4.
Clustering and classification comparisons using real data
It has been argued that comparisons based on spikein data do not necessarily translate to data derived from real biological specimens [22]. Therefore, we applied Gene Ontology to validate our result based on the premise that any improvement during the microarray data analysis process should result in tighter clustering of functionally related genes [23]. For example, in a gene list of size g, suppose that f number of genes are annotated with a given GO term. Suppose further that, after clustering the gene list using an accepted clustering method, n numbers of genes annotated with the given GO term are coclustered together in a cluster with c number of genes. The probability of this specific GO term coclustering can be calculated based on a hypergeometric distribution [24], and has been implemented in the CLASSIFI website http://pathcuric1.swmed.edu/pathdb/classifi.html. Smaller probabilities indicate that the clustering is less likely to be due to chance. Therefore, we would expect that the preprocessing method producing the smallest GO term coclustering Pvalues would be the method that most effectively reduces noise in the data.
Table 3 shows the number of GO terms with pvalues less than 1010 for each of the combinations of background correction and normalization algorithms tested. The data used were selected from the GSE2350 series [12], downloaded from the NCBI GEO database http://www.ncbi.nih.gov/projects/geo. In the comparison, the first three samples from both "control" (GSM44051, GSM44052 and GSM44053) and "CD40L treatment" (GSM44057, GSM44058 and GSM44059) groups are used. DFCM outperforms the others when paired with scale normalization, and performs comparably to the others when paired with loess normalization. The zonal background adjustment as given in MAS 5.0 has the overall worst performance. We can also make a case that quantile normalization gives the worst results of the normalization methods presented here.
Discussion
The RMA convolution model for background correction of microarray data from Affymetrix platforms is very popular. This model assumes that the observed value of fluorescence intensities is composed of an exponentially distributed signal with underlying normally distributed noise. This idea of a combination of signal and noise is quite reasonable, but the analysis presented here indicates that the distributional assumptions are not always correct. In order to examine the assumption of normally distributed background noise, we performed background correction using the convolution model and plotted the estimated background intensities versus a normal distribution using a quantilequantile plot for three spikein data sets. The plots indicate that the normality assumption may not hold for all of the spikein data sets examined. To confirm this, we examined the data with three wellknown goodnessoffit tests. The KS test, in particular, is known to have extremely low power [20]. The fact that the test rejects so often is quite strong evidence against normality.
We also examined the background corrected intensities, which are purported to represent the true signal, against the exponential distribution. QQ plots and goodness of fit tests show that the background corrected signal is clearly not exponentially distributed for any of the data studied here. These observations lend credibility to the notion that preprocessing approaches should not rely heavily on distributional assumptions.
There is some evidence that the gene distributions within groups are normally distributed after preprocessing with MBEI and MAS 5.0 [25]; and these distributions are indeed relevant for the purposes of testing the differential expression of genes with parametric methods such as the ttest. However, we are concerned in this paper with the distribution of the background noise, and not with the distribution of individual probes. In our framework, the background noise results from a combination of autoflourescence (a constant) and nonspecific hybridization. As nonspecific hybridization is not likely to be gene (and thus probe) specific, it is reasonable to model it with a global distribution [26].
Recently, it has been argued that the assumption that intensity values from a microarray study are random samples from any statistical distribution is seriously flawed [27]. The notion of a random sample implies independence of the intensity values, or at least that the dependence structure is sufficiently weak so that the random sample assumption is plausible. However, the dependence structure among genes, and the probe sets that interrogate them is quite complicated and, in some cases, strong. In this light, a nonparametric approach to background correction is a good alternative. DFCM does not make any assumptions on the dependency structure of the PM or MM intensities. This is manifested in the fact that linear interpolation is used to correct for background with small intensities. With larger intensities, the estimated background mean is simply subtracted because the impact of background noise is minor for the larger intensities.
DFCM uses q2th percentile of the MM signal corresponding to the smallest q1 percentage of PM intensities to estimate background noise. The original intent of the MM probes was to provide a measure of nonspecific hybridization that could be subtracted from the PM intensities, leaving the true signal. MAS 5.0 was developed under this assumption. It should be noted that the use of PM values alone could be justified by the noisiness and lack of validity of the MM measurement [3]. For example, approximately onethird of the MM intensities are greater than their corresponding PM intensities, and this tends to be constant across all Affymetrix platforms [3]. Furthermore, the MM measurements tend to be highly correlated with the corresponding PM measurements, indicating that the MM probes are either cross hybridizing to the incorrect gene or nonspecifically hybridizing to the correct gene. Therefore MM probes are imperfect predictors of nonspecific binding [28]. Thus, PM signal correction through MM subtraction has been largely rejected in the field.
There is a biochemically defensible rationale for estimation of background noise using DFCM. By selecting the lowest q1 percent of PM, we ensure that nonspecific hybridization will not be an issue. However, there could still be some crosshybridization, which is eliminated by taking only the smallest q2 percentage of MM. Having said this, one could simply use the lowest q1 percent of the PM; however, selecting the value for q1 could be difficult and somewhat arbitrary, and if a relatively high value for q1 is chosen might include some real signal for some of the PMs. The corresponding MMs in this case should be less and should be closer to background since in theory they should not be hybridizing to the real target.
The algorithm for choosing the value of q1 is very stable (see Methods), almost always choosing the same value of q1 for a given platform. For example, two experiments completed on the HGU95 platform will have very similar values of q1 (approximately 0.25). In other words, the values of q1 are more platform dependent than they are experimentdependent. This fact supports the notion that different normalization procedures are required for different platforms [22].
One way to think of q2 is as an estimate of the chance that an MM probe is crosshybridizing to another target transcript (or that most of its signal is from nonspecific hybridization). A reasonable estimate of q2 is given by subtracting this estimate of crosshybridization potential from 100%. In the examples that follow, q2 = 90%. In practice, the value of q2 was found to have little effect on the background correction (see Figure 6). This is understandable because the probe sets have been preselected to avoid crosshybridization of both the PM and MM probes.
Conclusion
We have shown that microarray data from three different Affymetrix platforms (GoldenSpike, HGU95Av2, and HGU133A) do not meet the assumptions of the exponentialnormal convolution model for background correction. This model is used in the Bioconductor software package in conjunction with quantile normalization and median polish summarization to comprise the RMA method. In all cases examined, estimated background noise did not follow a normal distribution, nor did the resulting estimated signal follow a simple exponential distribution. To circumvent these problems, we devised a distributionfree method to subtract background noise (DFCM). This method tended to perform better than many popular algorithms across a variety of experiments and array platforms.
This finding has four important implications. First, it is important to account for nonspecific hybridization. We attempted to do so by using MM intensities to obtain an estimate of background noise. MAS 5.0 uses ideal mismatch to account for nonspecific hybridization, but given the strong correlation between PM and MM values, the method is likely subtracting signal from the PM intensities, resulting in poor sensitivity and specificity. In other words, the method does not really account for nonspecific hybridization, since the MM values do not perform as designed.
GCRMA uses the probe sequence information given by MM probes and it works well for the HGU95Av2 data, but not for the HGU133A data. The performance discrepancy may be explained in part by the improved technology and better knowledge of the human genome at the time of the creation of the HGU133A chip. The other part of the explanation lies with deficiencies in estimating the various components of the GCRMA model. For example, parameter estimates for nonspecific hybridization are difficult to estimate reliably since the signal and noise from an observed intensity cannot be distinguished for most data. Estimating probe affinity is also quite difficult in practice. It has been reported that the top 2% probes will contain up to 50% of total signals [3]. If there are not enough arrays from enough different conditions the estimated affinity will be very biased towards probes with high intensities. In addition, the probe affinity relating to nonspecific hybridization should be investigated instead of that belonging to the whole signal since we want to know the effect of nonspecific hybridization between PM and MM within a probe pair. Therefore, it would be better to use "nonspecific hybridization" rather than the observed intensities. The difference between the observed intensity and the unknown nonspecific hybridization rate might be of practical importance.
Second, any background correction method based on assumptions that the background noise is normally distributed and that the real signal is exponentially distributed may not be valid for any given array platform. Testing the distributional assumptions for real data is impossible, since we cannot know what is background and what is signal. Third, it is clear that we need to develop an understanding of the reasons certain methods perform better on certain platforms, and the role that nonspecific hybridization and crosshybridization play in the observed intensities from microarray data. Finally, the fact that different methods perform better (or worse) on different platforms indicates that no one method may be a panacea for all preprocessing needs. However, in order to test this conjecture, more spikein data sets on a variety of platforms are necessary, as well as performance measures for use on real data sets. Automated methods for choosing the best method to analyze a particular microarray data set would be an important contribution.
Appendix
Please see Additional file 1
References
 1.
Bolstad BM: Low Level Analysis of Highdensity oligonucleotide array data: Background, normalization and summarization. PhD Thesis. 2004, University of California at Berkeley, Department of Statistics
 2.
Choe SE, Boutros M, Michelson AM, et al: Preferred analysis methods for Affymetrix genechips revealed by a wholly defined control dataset. Genome Biol. 2005, 6: R1610.1186/gb200562r16.
 3.
Irizarry RA, Hobbs B, Collin F, et al: Exploration, normalization, and summaries of high density oligonucleotide array probe level data. Biostatistics. 2003, 4: 249264. 10.1093/biostatistics/4.2.249.
 4.
Gentleman RC, Carey VJ, Bates DM, et al: Bioconductor: open software development for computational biology and bioinformatics. Genome Biol. 2004, 5: R8010.1186/gb2004510r80.
 5.
McGee M, Chen Z: Parameter Estimation for the ExponentialNormal Convolution Model for Background Correction of Affymetrix GeneChip Data. Statistical Applications in Genetics and Molecular Biology. 2006, 5: Article 2410.2202/15446115.1237.
 6.
Cope LM, Irizarry RA, Jaffee H, et al: A benchmark for Affymetrix GeneChip expression measures. Bioinformatics. 2003, 20: 323331. 10.1093/bioinformatics/btg410.
 7.
Wu Z, Irizarry RA, Gentleman R, et al: A modelbased background adjustment for oligonucleotide expression arrays. Journal of the American Statistical Association. 2004, 99: 909917. 10.1198/016214504000000683.
 8.
Affymetrix: Statistical algorithms description document. 2002
 9.
Li C, Wong HW: Modelbased analysis of oligonucleotide arrays: expression index computation and outlier detection. Proceedings of the National Academy of Sciences. 2001, 98: 3136. 10.1073/pnas.011404098.
 10.
Li C, Wong HW: Modelbased analysis of oligonucleotide arrays: model validation, design issues and standard error application. Genome Biology. 2001, 2: research0032.10032.11. 10.1186/gb200128research0032.
 11.
Affymetrix I: Technical note: guide to probe logarithmic intensity error (PLIER) estimation. 2005
 12.
Basso K, Margolin AA, Stolovitzky G, Klein U, DallaFavera R, Califano A: Reverse engineering of regulatory networks in human B cells. Nat Genet. 2005, 37: 382390. 10.1038/ng1532.
 13.
Hseih WP, Chu TZ, Wolfinger R: Who are those strangers in the Latin Square?. Methods of Microarray Data Analysis III. Edited by: Johnson KF, Lin SM. 2003, New York: SpringerVerlag, 199208.
 14.
McGee M, Chen Z: New spikedin probe sets for the Affymetrix hgu133a Latin square experiment. COBRA Preprint Series. 2006, Article 5
 15.
Dabney AR, Storey JD: A reanalysis of a published Affymetrix GeneChip control data set. Genome Biology. 2006, 7: 40110.1186/gb200673401.
 16.
Irizarry R, Cope L, Wu Z: FeatureLevel exploration of a published Affymetrix genechip control data set. Genome Biology. 2006, 7: 8404. 10.1186/gb200678404.
 17.
D'Agostino RB, Stephens MA, editors: Tests based on EDF statistics. 1986, New York: Marcel Dekker
 18.
Thode HC: Testing for Normality. 2002, New York: Marcel Dekker
 19.
Ihaka R, Gentleman RC: R: a language for data analysis and graphics. Journal of Computational and Graphical Statistics. 1996, 5: 299314. 10.2307/1390807.
 20.
Shapiro SS, Wilk MB, Chen HJ: A comparative study of various tests for normality. Journal of the American Statistical Association. 1968, 63: 13431372. 10.2307/2285889.
 21.
Iacus SM, Urbanek S: R Cocoa GUI 1.14 (2129). 2005, R Foundation for Statistical Computing
 22.
Allison DB, Cui X, Page GP, et al: Microarray data analysis: from disarray to consolidation and consensus. Nature Reviews Genetics. 2005, 7: 5565. 10.1038/nrg1749.
 23.
Kong M, Chen Z, Qian Y, Cai J, Lee J, Rab E, McGee M, Scheuermann RH: Use of Gene Ontotology as a Tool for Assessment of Analytical Algorithms with Real Data Sets: Impact of Revised Affymetrix CDF Annotation. Proceedings of the Seventh International Workshop on Data Mining in Bioinformatics (BIOKDD 2007). 2007, ACM Press
 24.
Lee JA, Sinkovits RS, Mock D, Rab EL, Cai J, Yang P, Saunders B, Hsueh RC, Choi S, Subramaniam S, Scheuerman RH: Components of the antigen processing and presentation pathway revealed by gene expression microarray analysis following B cell antigen receptor (BCR) stimulation. BMC Bioinformatics. 2006, 7: 23710.1186/147121057237.
 25.
Giles PJ, Kipling D: Normality of oligonucleotide microarrays and implications for parametric statistical analyses. Bioinformatics. 2003, 19: 22542262. 10.1093/bioinformatics/btg311.
 26.
Hein AMK, Richardson S, Causton HC, Ambler GK, Green PJ: BGX: a fully Bayesian integrated approach to the analysis of Affymetrix GeneChip data. Biostatistics. 2005, 6: 349373. 10.1093/biostatistics/kxi016.
 27.
Klebanov L, Yakovlev A: Treating expression levels of different genes as a sample in microarray data analysis: is it worth a risk?. Stat Appl Genet Mol Biol. 2006, 5: Article9
 28.
Wu Z, Irizarry RA: Preprocessing of oligonucleotide array data. Nature Biotechnology. 2004, 22: 656658. 10.1038/nbt0604656b.
Acknowledgements
The authors wish to thank William R. Schucany for a careful reading of this manuscript during its preparation, and gratefully acknowledge support for this work from Dr. Milton Packer, Director, Department of Clinical Sciences, University of Texas Southwestern Medical Center. This research was supported by the National Institutes of Health contracts N01AI40076 and N01AI40041 to RHS and by grant R15AG16192 to Monnie McGee. The authors have no potential conflicts of interest directly relevant to the contents of the manuscript.
This article has been published as part of BMC Genomics Volume 10 Supplement 1, 2009: The 2008 International Conference on Bioinformatics & Computational Biology (BIOCOMP'08). The full contents of the supplement are available online at http://www.biomedcentral.com/14712164/10?issue=S1.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors' contributions
ZC devised the algorithms and performed the study; MM and RHS supervised the study, obtained supports and drafted the manuscript; QL helped to design the algorithms; YD assisted in the study; MK developed and ran the code for the CLASSIFI algorithm. All authors have read and approved the final manuscript.
Electronic supplementary material
Rights and permissions
About this article
Cite this article
Chen, Z., McGee, M., Liu, Q. et al. A distributionfree convolution model for background correction of oligonucleotide microarray data. BMC Genomics 10, S19 (2009). https://doi.org/10.1186/1471216410S1S19
Published:
Keywords
 Receiver Operating Characteristic Curve
 Perfect Match
 Background Correction
 Distributional Assumption
 Robust Multichip Average