Volume 9 Supplement 2

## IEEE 7th International Conference on Bioinformatics and Bioengineering at Harvard Medical School

# Array CGH data modeling and smoothing in Stationary Wavelet Packet Transform domain

- Heng Huang
^{1}Email author, - Nha Nguyen
^{2}, - Soontorn Oraintara
^{2}and - An Vo
^{2}

**9(Suppl 2)**:S17

**DOI: **10.1186/1471-2164-9-S2-S17

© Huang et al; licensee BioMed Central Ltd. 2008

**Published: **16 September 2008

## Abstract

### Background

Array-based comparative genomic hybridization (array CGH) is a highly efficient technique, allowing the simultaneous measurement of genomic DNA copy number at hundreds or thousands of loci and the reliable detection of local one-copy-level variations. Characterization of these DNA copy number changes is important for both the basic understanding of cancer and its diagnosis. In order to develop effective methods to identify aberration regions from array CGH data, many recent research work focus on both smoothing-based and segmentation-based data processing. In this paper, we propose stationary packet wavelet transform based approach to smooth array CGH data. Our purpose is to remove CGH noise in whole frequency while keeping true signal by using bivariate model.

### Results

In both synthetic and real CGH data, Stationary Wavelet Packet Transform (SWPT) is the best wavelet transform to analyze CGH signal in whole frequency. We also introduce a new bivariate shrinkage model which shows the relationship of CGH noisy coefficients of two scales in SWPT. Before smoothing, the symmetric extension is considered as a preprocessing step to save information at the border.

### Conclusion

We have designed the SWTP and the SWPT-Bi which are using the stationary wavelet packet transform with the hard thresholding and the new bivariate shrinkage estimator respectively to smooth the array CGH data. We demonstrate the effectiveness of our approach through theoretical and experimental exploration of a set of array CGH data, including both synthetic data and real data. The comparison results show that our method outperforms the previous approaches.

## Background

Gene amplifications or deletions frequently contribute to tumorigenesis. When part or all of a chromosome is amplified or deleted, a change in DNA copy number results. Characterization of these DNA copy number changes is important for both the basic understanding of cancer and its diagnosis. Cancer researchers currently use array comparative genomic hybridization (array CGH) to identify sets of copy number changes associated with the particular cancer or its congenital and developmental disorders. In array CGH, because the clones contain sequences information directly connecting with the genome database, array CGH offers rapid genome-wide analysis at high resolution and the information it provides is directly linked to the physical and genetic maps of the human genome. Bacterial Artificial Chromosomes (BAC) based CGH arrays were amongst the first genomic arrays to be introduced [1] and are routinely used to detect single copy changes in the genome, owing to their high resolution in the order of 1 Mb [1, 2]. More recently Oligonucleotide aCGH [3, 4] was developed to allow flexibility in probe design, greater coverage, and much higher resolution in the order of 35–100 Kb [5].

In order to develop effective methods to identify aberration regions from array CGH data, the previous research works focus on both smoothing-based [5–9] and segmentation-based data processing [10–14]. The array CGH is very noisy. For example, in cDNA array CGH data, the signal to noise ratio is often approximately 1 (0 dB) [15]. Research in this area has been active in the last few years. Beheshti *et al.* proposed to use the robust locally weighted regression and smoothing scatterplots (lowess) method in [6]. Eilers and Menezes [7] perform a quantile smoothing method based on the minimization of the sum of absolute errors to create sharper boundaries between segments. Hsu *et al.* [8] investigated the usage of maximal overlap discrete wavelet transform (MODWT) in the analysis of array CGH data. They have shown translation invariant wavelets are promising methods for array CGH data smoothing and also observed that the denoising techniques may miss singleton clones that have small changes but somehow are consistent across tumors. In 2005, Lai [16] compared 11 different algorithms for analyzing array CGH data. Many smoothing and estimation methods were included in [16] such as CGHseg (2005) [17], Quantreg (2005) [7], CLAC (2005) [18], GLAD (2004) [11], CBS (2004) [14], HMM (2004) [19], MODWT (2005) [8], Lowess [6], ChARM (2004) [13], GA (2004) [12], ACE (2005) [20]. Lai concluded that Wavelet, Quantreg and Lowess method gave better detection results (higher true position rate and lower false position rate) than other methods. So, the wavelet based smooth was considered as the promising approach. More recently Y. Wang and S. Wang [5] extended the stationary wavelet (SWT) denoising and regression for nonequispaced data, because the physical distance between adjacent probes along a chromosome are not uniform, even vary drastically. However, if a signal is decomposed by using SWT or MODWT, we get unequal sub-bands and a long high frequency sub-bands. Because true CGH signals include many step functions, they contain important information in high frequency. If long high frequency is used to remove noise, maybe, some high frequency true information of CGH will be loosen.

In this paper, we propose to use the Stationary Wavelet Packet Transform (SWPT) to denoise the array CGH data. Because, in SWPT, all sub-bands are also shift invariant, each sub-band provides a shiftable description of signal in a specific scale as the same SWT or MODWT. SWPT analyzes signal to many equally frequency sub-bands. So, information in both of low and high frequency sub-band are saved. Moreover, new bivariate shrinkage function is used in SWPT instead of universal thresholding at the first time, soft thresholding [21–23] and BayesShrink [24]. We demonstrate the effectiveness of our approach through theoretical and experimental exploration of a set of array CGH data, including both synthetic data and real array CGH data. The comparison results show that our method outperforms the previous approaches about 6.4% – 57.9%. Let's see detail results in next section.

## Results and discussion

In this section, results of our proposed methods such as the SWPT and the SWPT-Bi will be compared to the other efficient smooth methods such as the Lowess [16], the Quantreg [7, 25], the SWTi [5], the DTCWTi-bi [26]. In our experiments, the artificial chromosomes are generated using the methods proposed in [27] and [5]. Finally, real data examples are showed to make sure that our methods are still better the others.

### Synthetic data

First, we describe how to create synthesis data as follow.

#### Artificial chromosome generation

*c*= 0, 1, 2, 3, 4 and 5 are generated with probability 0.01, 0.08, 0.81, 0.07, 0.02 and 0.01. The lengths for segments are picked up randomly from the corresponding empirical length distribution given in [27]. Each sample is a mixture of tumor cells and normal cells. A proportion of tumor cells is

*P*

_{ t }, whose value is from a uniform distribution between 0.3 and 0.7. As in paper [27], the

*log*2

*ratio*is calculated by

where *c* is the assigned copy number. The expected *log* 2*ratio* value is then the latent true signal.

Gaussian noises with zero mean and variance ${\sigma}_{n}^{2}$ are added to the latent true signal. Till now, we get the equally spaced CGH signal. Because the distances between two probes are randomly, the best way to get these distances is from the UCSF HumArray2 BAC array. Thus, we create a real CGH signal from the equally spaced CGH signal when the unequally spaced probes are placed on the chromosome. Now, we have many artificial chromosomes of length 200 *Mbase* which are created by many noise levels *σ*_{
n
}= 0.1, 0.125, 0.15, 0.175, 0.2, 0.25 and 0.275.

#### Comparison by RMSE

In this section, we will present the results when applying six methods such as the Lowess [16], the Quantreg [7, 25], the SWTi [5], the DTCWTi-bi [26] and our methods the SWPT and the SWPT-Bi. One thousand artificial chromosomes with seven different noise levels *σ*_{
n
}= 0.1, 0.125, 0.15, 0.175, 0.2, 0.25 and 0.275 are denoised.

*σ*= 0.2. In those Figures, the black solid lines represent the latent true signals, the blue points stand for the noisy DNA copy data

*log*

_{2}

*ratio*at the probe loci and the red lines correspond to the denoised data. We should note that the line connecting the denoised data points is only for visualization purpose.

At the copy three *c* = 3 (from 1 *kbase* to 1.4 × 10^{4} *kbase*) as shown in Figure 2, the *log*_{2}*ratio* value of the latent true signal is 0.3598, but these values of the Quantreg, the SWTi and the DTCWTi-bi based denoised signal in Figure 2 are from 0.2262 to 0.4966, from 0.1774 to 0.3828 and from 0.09233 to 0.6182 respectably. These values can cause a mistake when we segment the DNA copy number data. However, the denoised data using the Lowess, the SWPT and the SWPT-Bi will be segmented correctly as the copy three (from 0.2 to 0.4) because the *log*_{2}*ratio* values are from 0.2129 to 0.3619, 0.2794 to 0.3649 and from 0.2565 to 0.3964. At the copy two *c* = 2 (from 1.4 × 10^{4} *kbase* to 1.2 × 10^{5} *kbase*), the denoised data in the second sub-figure (denoised by Quantreg) of Figure 2 has an amplitude of 0.2262 which will make an error in segmentation process, while the denoised data in other sub-figures of Figure 2 will give a correct segmentation. In this copy, the denoised signals using DTCWTi-bi, the SWPT and SWPT-Bi are approximately the latent true signals, while the denoised data using the Lowess, the Quantreg and the SWTi have many ripples. At the copy zero *c* = 0 (from 1.2 × 10^{5} *kbase* to 1.79 × 10^{5} *kbase*), if we use TPR (true position rate = number of denoised probes belove -0.4/total of true probes), the Lowess, the Quantreg and our methods gave a ratio of 22 over 34 instead of 17/34 and 14/34 of the SWTi and DTCWTi-bi. However, the denoised signals of the Lowess, the SWPT and the SWPT-Bi look better than of the Quantreg. At the copy two *c* = 2 (form 1.79 × 10^{5} *kbase* to the end of the chromosome), the fourth, fifth and sixth sub-figures's signal (denoised by DTCWTi-bi, SWPT the and SWPT-Bi) of Figure 2 look smoother than the others. Furthermore, the denoised signals at the first sub-figure (the Lowess's) and the second sub-figure (the Quantreg's) may cause error when segmentation because denoised signals change from -0.3133 to 0.101 (Lowess) and from -0.2119 to 0.2084 (Quantreg).

From above results, we can see that our proposed SWPT and SWPT-Bi methods with the stationary wavelet packet transform are better than the others. Now, real data will be used to test five smoothing methods as follow.

### Real data examples

*kbase*to 1.56 × 10

^{5}

*kbase*), the SWPT and SWPT-Bi based smoothed signals are smoother than the others. With the copy four, from 1.56 × 10

^{5}

*kbase*to the end of this chromosome, the performance of the Lowess, the SWPT and the SWPT-Bi based denoising methods are the better than of the Quantreg, the SWTi and the DTCWTi-bi. From the above figures, we can believe that our methods perform better than the others in denoising of real CGH data.

## Conclusion

In this paper, we explored the stationary wavelet packet transform method with the new bivariate shrinkage estimator in array CGH data denoising study. In the simulation situations, the denoising results from the SWPT and the SWPT-Bi are much better (improve 6.4% – 57.9%) than the previous methods in terms of the root mean squared error measurement at different noise levels. Furthermore, we also demonstrate our method by using the real array CGH data. In our future work, we will develop a smoothing and segmentation combinatorial algorithm to improve the aberration regions identification from DNA copy number data.

## Methods

Our methods named the SWPT (SWPT and universal shrinkage function) and the SWPT-Bi (SWPT and bivariate shrinkage function) will be introduced. First, lets review wavelet transform and see how SWPT operates.

### Wavelet methods

We will provide a brief review of wavelet transforms which were used for array CGH data smoothing and is used by this paper. We should note that the simple wavelet transform will be introduced firstly and the SWPT will be mentioned finally.

#### Discrete wavelet transform

*N*sequence, and generates an output sequence of length

*N*using a set of lowpass and highpass fiters followed by a decimator. It has

*N/*2 values at the highest resolution,

*N/*4 values at the next resolution, and

*N/*2

^{ L }at the level

*L*. Because of decimation, the DWT is a critically sampled decomposition. However, the drawback of DWT is the shift variant property. In signal denoising, the DWT creates artifacts around the discontinuities of the input signal [30]. These artifacts degrade the performance of the threshold-based denoising algorithm.

#### Stationary wavelet transform

The stationary wavelet transform (SWT) [30], showed in Figure 4, is similar to the DWT except that it does not employ a decimator after filtering, and each level's filters are upsampled versions of the previous ones. The SWT is also known as the shift invariant DWT. The absence of a decimator leads to a full rate decomposition. Each subband contains the same number of samples as the input. So for a decomposition of *L* levels, there is a redundant ratio of (*L* + 1): 1. However, the shift invariant property of the SWT makes it preferable for the usage in various signal processing applications such as denoising and classification because it relies heavily on spatial information. It has been shown that many of the artifacts could be suppressed by a redundant representation of the signal [30].

#### Dual-tree complex wavelet transform

A dual-tree structure that produces a dyadic complex DWT, showed in Figure 4, is proposed by Kingsbury [31, 32]. In the case of 1-D signals, the structure consists of two binary trees of multi-resolution decomposition of the same signal. It is therefore an overcomplete representation with a redundant ratio of 2:1. In the two trees, the filters are designed in such a way that the aliasing in one branch in the first tree is approximately canceled by the corresponding branch in the second tree. The relation between the wavelet filters of the two trees yields shift-invariant property [31].

The analysis FB for the DTCWT is an iterative multi-scale FB. Each resolution level consists of a pair of two-channel FBs. The purpose of the dual-tree CWT is to provide a shiftable and scalable multiresolution decomposition. The input signal is passed through the first level of a multiresolution FB. The low frequency component, after decimation by 2, is fed into the second level decomposition for the second resolution. The outputs of the two trees are the real and imaginary parts of complex-valued subbands. To reconstruct the signal, the real part and imaginary part are inverted to obtain two real signals, respectively. These two real signals are then averaged to obtained the final output. For more details of the construction of the dual-tree, the reader is referred to [33].

#### Discrete wavelet packet transform

We continue with an another basic othornormal wavelet transform. Discrete wavelet packet transform (DWPT), which can be readily computed by using a very simple adjustment of the pyramid algorithm for DWT, will be mentioned. All of DWPT scales are performed at the same level *j*. The *jth* level DWPT decomposes the frequency interval [0, 1/2] into 2^{
j
}equal and individual intervals, each of which has *N/* 2^{
j
}values if taking a length *N* sequence. DWPT still keeps a shift variant property.

#### Stationary wavelet packet transform

Stationary Wavelet Packet Transform (SWPT), showed in Figure 4, still keeps two important properties of SWT such as shift invariance and redundancy. In the SWPT, both scaling and wavelet coefficients are subjected to the high-pass and low-pass filter when computing the next level coefficients. At the given level *L*, there are 2^{
L
}scales with the same length as the input signal's. The redundant ratio is (2^{
L
}): 1 for a decomposition of *L* levels. SWPT is really combination of SWT and DWPT. So, it is very useful in denoising of DCN data. After wavelet transform, reader should be introduced a new shrinkage function to remove noise of CGH data in SWPT domain as follows.

### New vivariate shrinkage function for SWPT-based denoising

In this sub-section, the bivariate shrinkage function which describes the relationship of child and parent (Figure 4) coefficients will be reminded. Because SWPT, which decomposes a signal into many subbands at the same scale, just has child and cousin coefficients (Figure 4) at the same level, new bivariate shrinkage function will be developed to exploit the relationship between child and cousin coefficients. A simple denoising algorithm via wavelet transform consists of three steps: decompose the noisy signal by wavelet transform, denoise the noisy wavelet coefficients according to some rules and take the inverse wavelet transform from the denoised coefficients. To estimate wavelet coefficients, some of the most well-known rules are universal thresholding, soft thresholding [21–23] and BayesShrink [24]. In these algorithms, the authors assumed that wavelet coefficients are independent. Sendur and Selesnick [34] has recently exploited the dependency between coefficients and proposed a non-Gaussian bivariate pdf for the child coefficient *w*_{
c
}and its parent *w*_{
p
}. Nguyen *et el* [26, 35] applied that function to recover CGH data successfully and got some promising results.

Now basing on the idea in [34], we try to discover the connection of child and cousin coefficients in SWPT with CGH data. We assume that we get the DNA copy number data *Y* which includes the deterministic signal *D* and the independent and identically distributed (IID) Gaussian noise *n*. This Gaussian noise has zero mean and variance ${\sigma}_{n}^{2}$.

*Y* = *D* + *n*.

*Y*by the SWPT, we get the coefficients

**y**

_{ k }. In the wavelet domain, those coefficients can be formulated as

*y*

_{1}and

*y*

_{2}are noisy wavelet coefficients,

*w*

_{1}and

*w*

_{2}are true coefficients,

*w*

_{2}represents the cousin of

*w*

_{1}(child),

*n*

_{1}and

*n*

_{2}are independent Gaussian noise coefficients. If the cousin scale

*y*

_{2}continue being decomposed, we will get detail and approximation coefficients. Let's call

*y*

_{3}as approximation coefficients of

*y*

_{2}. We can calculate

*y*

_{3}from

*y*

_{2}by the follow equation:

where *h* [*n*] is the low pass filter and *N* is the length of signal *y*_{2}. In general, we can write

**y** = **w** + **n**,

**y**= (

*y*

_{1},

*y*

_{3}),

**w**= (

*w*

_{1},

*w*

_{3}) and

**n**= (

*n*

_{1},

*n*

_{3}). The noise pdf of two next scales should be followed as

**w**from

**y**is followed as

*w*

_{1}and

*w*

_{3}as

*w*

_{1}and

*w*

_{3}are really dependent. Let us define:

*f*(

*w*) with respect to

*w*

_{1}and

*w*

_{3}, respectively. We can get ${f}_{{w}_{1}}$ and ${f}_{{w}_{3}}$ from (9)

*r*from (15):

*r*by (16) into (15), the MAP estimator can be written as:

*u*)

_{+}is defined by

*y*

_{3}from (4) to (17), we can rewrite the MAP estimator as

*σ*can be estimated by

*M*is the size of the neighborhood

*N*(

*k*). In the packet wavelet transform, the cousin scales have not any parent scale. In this case, we can use hard thresholding estimator [21] to recover cousin coefficients ${\widehat{w}}_{cs}$:

Comparison table of our new bivariate shrinkage function and function in [34]

Comparison Table | ||
---|---|---|

Method | New bivariate shrinkage function. | Bivariate shrinkage function in [34]. |

Applying to Relationship
Transform | CGH data. child and cousin coefficient.
SWPT and DWPT. | image. child and parent coefficient.
DWT, SWT and DTCWT. |

### Signal extension

CGH data is finite signal. If we apply wavelet smooth method directly, we may get error at the border of denoised signal. So, extension step is a very important preprocessing step before denoising. There are three main extension methods. According to the book [36] (chapter 8), symmetric extension is the best if applied to a filtered image because we can save information at the border better. With CGH data, we also need save the information at the border. So, we recommend that symmetric extension method should be used as a preprocessing step before denoising. Let's assume that the length of the CGH signal is *N*. In order to get the best performance in the wavelet denoising algorithm, the length of the input signal is required to be a power of two [37]. If *N* is not a power of two, we can extend our signal to make sure *N* = 2^{
j
}by using symmetric extension method. Finally, the SWPT-Bi will be detailed in next part.

### Proposed method

The DWT with the redundant ratio of 1:1 is efficient for the denoising applications. However, the DWT creates artifacts around the discontinuities of the input signal [30] because it is shift-variant. To overcome this problem, SWT [5] or MODWT [8] and DTCWT [26, 35] with translation invariant property was proposed for signal denoising. It has been shown that many of the artifacts could be suppressed by a redundant representation of the signal [30]. One important thing is that CGH data contains many step functions which their information is in both low frequency and high frequency. The above wavelet methods have one disadvantage which some high frequency components of CGH data were removed. In this paper, the SWPT will be used to overcome some above problems because it keeps shift invariant property and looks for signal both in low frequency and in high frequency band for denoising operation. Several methods were proposed for selecting thresholding values such as hard universal [21, 22] and un-universal thresholding [23]. However, the dependency between wavelet coefficients are not exploited in these methods. Thus, we propose the usage of shift invariant SWPT and new bivariate shrinkage estimator which takes advantage of the dependency between wavelet coefficient and its cousin for array-based DNA copy number data denoising.

*Y*so that the root mean squared error (RMSE) (24) is the smallest.

and *N* is the number of input samples, *D* = {*D*_{
i
}} and $\widehat{D}=\left\{{\widehat{D}}_{i}\right\}$.

We propose a stationary wavelet packet transform and new bivariate shrinkage function based smooth method (SWPT-Bi). The SWPT-Bi can be summarized as follows:

**Step 1**: *Extend Y by using symmetric extension method and decompose new data Y' by the SWPT to L levels as (25). The numbers of decomposition levels* [38]*(at the remark* 11) *can be computed by*

*L* = log_{2}(*N*) - *J*.

*where J* = 3, 4, 5, 6. *This is a perfect number of levels* [38]*which yields the best denoising result. In this paper, we use J* = 4 *as the same in* [8]*and* [5].

**Step 2**: *Calculate the noise variance* ${\widehat{\sigma}}_{n}^{2}$*and the marginal variance* ${\widehat{\sigma}}^{2}$*for wavelet coefficient y*_{
k
}*by using (21), (22) and (20)*.

**Step 3**: *Estimate the child coefficients* ${\widehat{w}}_{c}={\widehat{w}}_{1}$*as in (19) and estimate the counsin coefficients* ${\widehat{w}}_{cs}$*as in (23). In this case, k* = 1.45 *should be chosen*.

**Step 4**: *Reconstruct data* $\widehat{D}$*from the denoised coefficients* ${\widehat{w}}_{c}$*and* ${\widehat{w}}_{cs}$*by taking inverse SWPT*.

We also propose one simple method SWPT. In the SWPT method, hard thresholding [22] method is used. The SWPT method can be summarized as follows:

**Step 1**: *Extend Y by using symmetric extension and decompose new data using the SWPT*.

**Step 2**: *Estimate the noise variance* ${\sigma}_{n}^{2}$*by (21)*.

**Step 3**:

*Find the denoised coefficients from noisy coefficients as follow*

*where N is length of y*.

**Step 4**: *Reconstruct data* $\widehat{D}$*from the denoised coefficients* ${\widehat{w}}_{i}$*by taking inverse SWPT*.

## Declarations

### Acknowledgements

Funding was provided by UTA-UTSW collaborative research grant.

This article has been published as part of *BMC Genomics* Volume 9 Supplement 2, 2008: IEEE 7^{th} International Conference on Bioinformatics and Bioengineering at Harvard Medical School. The full contents of the supplement are available online at http://www.biomedcentral.com/1471-2164/9?issue=S2

## Authors’ Affiliations

## References

- Pinkel D, Segraves R, Sudar D, Clark S: High resolution analysis of DNA copy number variation using comparative genomic hybridization to microarrays. Nat Genet. 1998, 20: 207-211. 10.1038/2524.PubMedView Article
- Snijders AM, Nowak N, Segraves R, Blackwood S: Assembly of microarrays for genome-wide measurement of DNA copy number. Nat Genet. 2001, 29 (3): 263-264. 10.1038/ng754.PubMedView Article
- Brennan C, Zhang Y, Leo C, Fenga B: High-resolution global profiling of genomic alterations with long oligonucleotide microarray. Cancer Res. 2004, 64: 4744-4748. 10.1158/0008-5472.CAN-04-1241.PubMedView Article
- Pollack J, Perou C, Alizadeh A, Eisen M: Genome-wide analysis of DNA copy-number changes using cDNA microarrays. Nat Genet. 1999, 23: 41-46. 10.1038/14385.PubMedView Article
- Wang Y, Wang S: A Novel Stationary Wavelet Denoising Algorithm for Array-Based DNA Copy Number Data. International Journal of Bioinformatics Research and Applications. 2007, 3 (2): 206-222. 10.1504/IJBRA.2007.013603.PubMedView Article
- Beheshti B, Braude I, Marrano P, Thorner P, Zielenska M, Squire J: Chromosomal localization of DNA amplifications in neuroblastoma tumors using cDNA microarray comparative genomic hybridization. Neoplasia. 2003, 5: 53-62.PubMedPubMed CentralView Article
- Eilers P, de Menezes R: Quantile smoothing of array CGH data. Bioinformatics. 2005, 21: 1146-1153. 10.1093/bioinformatics/bti148.PubMedView Article
- Hsu L, Self SG, Grove D, Randolph T, Wang K, Delrow JJ, Loo L, Porter P: Denoising array-based comparative genomic hybridization data using wavelets. Biostatistics(Oxford, England). 2005, 6 (2): 211-226.View Article
- Pollack J, Sorlie T, Perou C, Rees C: Microarray analysis reveals a major direct role of DNA copy number alteration in the transcriptional program of human breast tumors. Proc Natl Acad Sci. 2002, 99: 12963-12968. 10.1073/pnas.162471999.PubMedPubMed CentralView Article
- Daruwala R, Rudra A, Ostrer H, Lucito R, Wigler M, Mishra B: A versatile statistical analysis algorithm to detect genome copy number variation. Proc Natl Acad Sci. 2004, 101: 16292-16297. 10.1073/pnas.0407247101.PubMedPubMed CentralView Article
- Hupe P, Stransky N, Thiery JP, Radvanyi F, Barillot E: Analysis of array CGH data: from signal ratio to gain and loss of DNA regions. Bioinformatics. 2004, 20: 3413-3422. 10.1093/bioinformatics/bth418.PubMedView Article
- Jong K, Marchiori E, Meijer G, Vaart A, Ylstra B: Breakpoint identification and smoothing of array comparative genomic hybridization data. Bioinformatics. 2004, 20: 3636-3637. 10.1093/bioinformatics/bth355.PubMedView Article
- Myers C, Dunham M, Kung S, Troyanskaya O: Accurate detection of aneuploidies in array CGH and gene expression microarray data. Bioinformatics. 2004, 20: 3533-3543. 10.1093/bioinformatics/bth440.PubMedView Article
- Olshen A, Venkatraman E, Lucito R, Wigler M: Circular binary segmentation for the analysis of array-based DNA copy number data. Biostatistics. 2004, 5: 557-572. 10.1093/biostatistics/kxh008.PubMedView Article
- Bilke S, Chen QR, Whiteford CC, Khan J: Detection of low level genomic alterations by comparative genomic hybridization based on cDNA microarrays. Bioinformatics. 2005, 21 (7): 1138-1145. 10.1093/bioinformatics/bti133.PubMedView Article
- Lai W, Johnson M, Kucherlapati R, Park P: Comparative analysis of algorithms for identifying amplifications and deletions in array CGH data. Bioinformatics. 2005, 21: 3763-3770. 10.1093/bioinformatics/bti611.PubMedPubMed CentralView Article
- Picard F: A statistical approach for array CGH data analysis. BMC Bioinformatics. 2005, 6: 27-10.1186/1471-2105-6-27.PubMedPubMed CentralView Article
- Wang P, Kim Y, Pollack J, Narasimhan B, Tinshirani R: A method for calling gains and losses in arrray CGH data. Biostatistics. 2005, 6: 45-58. 10.1093/biostatistics/kxh017.PubMedView Article
- Fridkyand J: Hidden markov models approach to the analysis of arrray CGH data. J Multivariate Anal. 2004, 90: 132-153. 10.1016/j.jmva.2004.02.008.View Article
- Lingjaerde O: CGH-Exploxer: a program for analysis of array-CGH data. Bioinformatics. 2005, 21: 821-822. 10.1093/bioinformatics/bti113.PubMedView Article
- Donoho D, Johnstone I: Ideal spatial adaptation by wavelet shrinkage. Biometrika. 1994, 81: 425-455. 10.1093/biomet/81.3.425.View Article
- Donoho D: De-Noising by soft-thresholding. IEEE Trans on Inf Theory. 1995, 41 (3): 613-627. 10.1109/18.382009.View Article
- Johnstone I, Silverman B: Wavelet Threshold Estimators for Data with Correlated Noise. Journal of the Royal Statistical Society. 1997, 319-351. 59
- Chang S, Yu B, Vetterli M: Adaptive wavelet thresholding for image denoising and compression. IEEE Trans Image processing. 2000, 9: 1532-1546. 10.1109/83.862633.View Article
- Li Y, Zhu J: Analysis of array CGH data for cancer studies using fused quantile regression. Bioinformatics. 2007, 23: 2470-2476. 10.1093/bioinformatics/btm364.PubMedView Article
- Nguyen N, Huang H, Oraintara S, Vo A: A New Smoothing Model for Analyzing Array CGH Data. IEEE BIBE. 2007
- Willenbrock H, Fridlyand J: A comparison study: applying segmentation to array CGH data for downstream analyses. Bioinformatics. 2005, 21 (22): 4084-4091. 10.1093/bioinformatics/bti677.PubMedView Article
- Snijders AM, Nowak N, Segraves R, Blackwood S: Assembly of microarrays for genome wide measurement of DNA copy number by CGH. Nature Genetics. 2001, 29: 263-264. 10.1038/ng754.PubMedView Article
- University S: Assembly of microarrays for genome-wide measurement of DNA copy number. [http://www.nature.com/ng/journal/v29/n3/suppinfo/ng754_S1.html]
- Coifman R, Donoho D: Translation-invariant de-noising. Wavelets and Statistics, 103 of Lecture Notes in Statistics. 1995, 125-150.
- Kingsbury NG: Image Processing with Complex Wavelets. Phil Trans Royal Society London A. 1999, 357 (1760): 2543-2560.View Article
- Kingsbury NG: Complex wavelets for shift invariant analysis and filtering of signals. Journal of Applied and Computational Harmonic Analysis. 2001, 10 (3): 234-253. 10.1006/acha.2000.0343.View Article
- Selesnick IW, Baraniuk RG, Kingsbury NC: The dual-tree complex wavelet transform. IEEE Signal Processing Magazine. 2005, 22 (6): 123-151. 10.1109/MSP.2005.1550194.View Article
- Sendur L, Selesnick I: Bivariate Shinkage Function for Wavelet-Based Denoising exploiting Interscale Dependency. IEEE Transaction on Signal Processing. 2002, 50 (11):
- Nguyen N, Huang H, Oraintara S, Wang Y: Denoising of Array-Based DNA Copy Number Data Using The Dual-tree Complex Wavelet Transform. IEEE BIBE. 2007
- Strang G, Nguyen T: Wavelets and filter banks. 1996, Wellesley-Cambridge Press
- Coifman R, Wickerhauser M: Entropy-based Algorithms for best basis selection. IEEE Trans on Inf Theory. 1992, 38: 713-718. 10.1109/18.119732.View Article
- Bruce A, Gao H: Understanding waveshrink: Variance and bias estimation. Biometrika. 1996, 83: 727-745. 10.1093/biomet/83.4.727.View Article

## Copyright

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.