Abstract: We started offering an introduction to very basic aspects of molecular biology, for the reader coming from computer sciences, information technology, mathematics. Similarly we offered a minimum of information about pathways and networks in graph theory, for a reader coming from the bio-medical sector. At the crossover about the two different types of expertise, we offered some definition about Systems Biology. The core of the article deals with a Molecular Interaction Map (MIM), a network of biochemical interactions involved in a small signaling-network sub-region relevant in breast cancer. We explored robustness/sensitivity to random perturbations. It turns out that our MIM is a non-isomorphic directed graph. For non physiological directions of propagation of the signal the network is quite resistant to perturbations. The opposite happens for biologically significant directions of signal propagation. In these cases we can have no signal attenuation, and even signal amplification. Signal propagation along a given pathway is highly unidirectional, with the exception of signal-feedbacks, that again have a specific biological role and significance. In conclusion, even a relatively small network like our present MIM reveals the preponderance of specific biological functions over unspecific isomorphic behaviors. This is perhaps the consequence of hundreds of millions of years of biological evolution.
Abstract: High-throughput (HT) technologies provide huge amount of gene expression data that can be used to identify biomarkers useful in the clinical practice. The most frequently used approaches first select a set of genes (i.e. gene signature) able to characterize differences between two or more phenotypical conditions, and then provide a functional assessment of the selected genes with an a posteriori enrichment analysis, based on biological knowledge. However, this approach comes with some drawbacks. First, gene selection procedure often requires tunable parameters that affect the outcome, typically producing many false hits. Second, a posteriori enrichment analysis is based on mapping between biological concepts and gene expression measurements, which is hard to compute because of constant changes in biological knowledge and genome analysis. Third, such mapping is typically used in the assessment of the coverage of gene signature by biological concepts, that is either score-based or requires tunable parameters as well, limiting its power.
Abstract: The identification of robust lists of molecular biomarkers related to a disease is a fundamental step for early diagnosis and treatment. However, methodologies for the discovery of biomarkers using microarray data often provide results with limited overlap. These differences are imputable to 1) dataset size (few subjects with respect to the number of features); 2) heterogeneity of the disease; 3) heterogeneity of experimental protocols and computational pipelines employed in the analysis. In this paper, we focus on the first two issues and assess, both on simulated (through an in silico regulation network model) and real clinical datasets, the consistency of candidate biomarkers provided by a number of different methods.
Abstract: PURPOSE: To evaluate survival and clinical outcome for patients with a large uveal melanoma treated by either enucleation or proton beam radiotherapy (PBRT). ProceduRES: This retrospective non-randomized study evaluated 132 consecutive patients with T3 and T4 choroidal melanoma classified according to TNM stage grouping. RESULTS: Cumulative all-cause mortality, melanoma-related mortality and metastasis-free survival were not statistically different between the two groups (log-rank test, p = 0.56, p = 0.99 and p = 0.25, respectively). Eye retention of the tumours treated with PBRT at 5 years was 74% (SD 6.2%). In these patients at diagnosis, 73% of eyes had a best-corrected visual acuity (BCVA) of 0.1 or better. After 12 and 60 months, BCVA of 0.1 or better was observed in 47.5 and 32%, respectively. CONCLUSION AND MESSAGE: Although enucleation is the most common primary treatment for large uveal melanomas, PBRT is an eye-preserving option that may be considered for some patients.
Abstract: In this paper we present a framework for structured variable selection (SVS). The main concept of the proposed schema is to take a step towards the integration of two different aspects of data mining: database and machine learning perspective. The framework is flexible enough to use not only microarray data, but other high-throughput data of choice (e.g. from mass spectrometry, microarray, next generation sequencing). Moreover, the feature selection phase incorporates prior biological knowledge in a modular way from various repositories and is ready to host different statistical learning techniques. We present a proof of concept of SVS, illustrating some implementation details and describing current results on high-throughput microarray data.
Abstract: A molecular characterization of Alzheimer's Disease (AD) is the key to the identification of altered gene sets that lead to AD progression. We rely on the assumption that candidate marker genes for a given disease belong to specific pathogenic pathways, and we aim at unveiling those pathways stable across tissues, treatments and measurement systems. In this context, we analyzed three heterogeneous datasets, two microarray gene expression sets and one protein abundance set, applying a recently proposed feature selection method based on regularization.
Abstract: Hypoxia is a condition of low oxygen tension occurring in the tumor and negatively correlated with the progression of the disease. We studied the gene expression profiles of nine neuroblastoma cell lines grown under hypoxic conditions to define gene signatures that characterize hypoxic neuroblastoma. The l(1)-l(2) regularization applied to the entire transcriptome identified a single signature of 11 probesets discriminating the hypoxic state. We demonstrate that new hypoxia signatures, with similar discriminatory power, can be generated by a prior knowledge-based filtering in which a much smaller number of probesets, characterizing hypoxia-related biochemical pathways, are analyzed. l(1)-l(2) regularization identified novel and robust hypoxia signatures within apoptosis, glycolysis, and oxidative phosphorylation Gene Ontology classes. We conclude that the filtering approach overcomes the noisy nature of the microarray data and allows generating robust signatures suitable for biomarker discovery and patients risk assessment in a fraction of computer time.
Abstract: Hypoxia is a condition of low oxygen tension occurring in the tumor microenvironment and it is related to poor prognosis in human cancer. To examine the relationship between hypoxia and neuroblastoma, we generated and tested an in vitro derived hypoxia gene signature for its ability to predict patients' outcome.
Abstract: To evaluate the results of 15 years of experience with proton beam radiotherapy in the treatment of intraocular melanoma, and to determine univariate and multivariate risk factors for local failure, eye retention, and survival.
Abstract: Gene expression signatures are clusters of genes discriminating different statuses of the cells and their definition is critical for understanding the molecular bases of diseases. The identification of a gene signature is complicated by the high dimensional nature of the data and by the genetic heterogeneity of the responding cells. The l1-l2 regularization is an embedded feature selection technique that fulfills all the desirable properties of a variable selection algorithm and has the potential to generate a specific signature even in biologically complex settings. We studied the application of this algorithm to detect the signature characterizing the transcriptional response of neuroblastoma tumor cell lines to hypoxia, a condition of low oxygen tension that occurs in the tumor microenvironment.
Abstract: The search for predictive biomarkers of disease from high-throughput mass spectrometry (MS) data requires a complex analysis path. Preprocessing and machine-learning modules are pipelined, starting from raw spectra, to set up a predictive classifier based on a shortlist of candidate features. As a machine-learning problem, proteomic profiling on MS data needs caution like the microarray case. The risk of overfitting and of selection bias effects is pervasive: not only potential features easily outnumber samples by 10(3) times, but it is easy to neglect information-leakage effects during preprocessing from spectra to peaks. The aim of this review is to explain how to build a general purpose design analysis protocol (DAP) for predictive proteomic profiling: we show how to limit leakage due to parameter tuning and how to organize classification and ranking on large numbers of replicate versions of the original data to avoid selection bias. The DAP can be used with alternative components, i.e. with different preprocessing methods (peak clustering or wavelet based), classifiers e.g. Support Vector Machine (SVM) or feature ranking methods (recursive feature elimination or I-Relief). A procedure for assessing stability and predictive value of the resulting biomarkers' list is also provided. The approach is exemplified with experiments on synthetic datasets (from the Cromwell MS simulator) and with publicly available datasets from cancer studies.
Abstract: We propose a method for studying the stability of biomarker lists obtained from functional genomics studies. It is common to adopt resampling methods to tune and evaluate marker-based diagnostic and prognostic systems in order to prevent selection bias. Such caution promotes honest estimation of class prediction, but leads to alternative sets of solutions. In microarray studies, the difference in lists may be bewildering, also due to the presence of modules of functionally related genes. Methods for assessing stability understand the dependency of the markers on the data or on the predictor's type and help selecting solutions.
Abstract: We connect in a grid-enabled pipeline an ontology-based environment for proteomics spectra management with a machine learning platform for unbiased predictive analysis. We exploit two existing software platforms (MS-Analyzer and BioDCV), the emerging proteomics standards, and the middleware and computing resources of the EGEE Biomed VO grid infrastructure. In the setup, BioDCV is accessed by the MS-Analyzer workflow as a Web service, thus providing a complete grid environment for proteomics data analysis. Predictive classification studies on MALDI-TOF data based on this environment are presented.
Abstract: In the statistical learning framework, the use of appropriate kernels may be the key for substantial improvement in solving a given problem. In essence, a kernel is a similarity measure between input points satisfying some mathematical requirements and possibly capturing the domain knowledge. In this paper, we focus on kernels for images: we represent the image information content with binary strings and discuss various bitwise manipulations obtained using logical operators and convolution with nonbinary stencils. In the theoretical contribution of our work, we show that histogram intersection is a Mercer's kernel and we determine the modifications under which a similarity measure based on the notion of Hausdorff distance is also a Mercer's kernel. In both cases, we determine explicitly the mapping from input to feature space. The presented experimental results support the relevance of our analysis for developing effective trainable systems.