Multi-omics is one of the most discussed topics in molecular research. The promise is straightforward: by measuring multiple molecular layers simultaneously, you get a more complete picture of biology. The practice is messier.
The promise
Genomics tells you what variants are present. Transcriptomics tells you which genes are being expressed. Proteomics tells you which proteins are actually being produced. Each layer gives a different view, and combining them could reveal patterns that no single layer shows on its own.
In cancer research, for example, a genomic variant might increase risk, but only if the gene is actually expressed (transcriptomics) and the protein is functional (proteomics). Looking at all three layers together could identify patients whose risk is driven by the full molecular chain, not just one part of it.
The appeal for precision medicine is clear. If you can measure a patient's molecular profile across multiple layers, you can potentially stratify them into groups that respond differently to treatment, identify biomarkers that predict treatment resistance, or discover molecular subtypes that have different prognoses. Multi-omics integration is the analytical method that turns these possibilities into testable hypotheses.
The reality
Most multi-omics studies have small sample sizes, often fewer than 100 patients per group. Each omics layer produces thousands of variables. Combining three layers means searching for patterns across tens of thousands of features in a small sample. The statistical power to find real associations is low, and the risk of finding false associations is high.
Data harmonisation is another challenge. Different platforms measure different things at different scales with different error profiles. Aligning genomic variants with expression levels with protein abundances requires careful normalisation, batch correction, and quality control at each layer. Mistakes at any step propagate through the integration.
Most published multi-omics integration methods are still being benchmarked. There is no consensus on the best approach for combining layers, handling missing data across platforms, or validating integrated results.
A practical example illustrates the challenge. A research team measures genomic variants in 80 breast cancer patients, gene expression in the same 80 patients, and protein levels in 65 of the 80 patients (15 samples failed quality control for proteomics). The genomic data includes 50,000 variants, the expression data includes 20,000 genes, and the proteomic data includes 5,000 proteins. Integrating across all three layers on 65 patients means fitting a model with 75,000 potential features on 65 observations. No statistical method can reliably identify real associations under those conditions.
The most honest approach is to use each omics layer separately to answer layer-specific questions, then look for convergent evidence across layers. If a genomic variant is associated with expression changes in the same gene, and that gene's protein shows differential abundance between groups, the convergent evidence is stronger than any single-layer finding. This is less elegant than a joint model, but it is more robust statistically and more interpretable biologically.
- Sample sizes are usually too small for reliable multi-omics discovery
- Each omics layer has its own error profile and normalisation requirements
- Integration methods are still evolving, with no established standard
- Most multi-omics findings are exploratory and require independent validation
What works today versus what is aspirational
Some multi-omics applications have matured to the point of clinical utility. Tumour molecular profiling that combines mutation panels with expression data is used in clinical oncology to guide treatment selection. Pharmacogenomics panels that integrate variant data with known drug-gene interactions have established clinical workflows. These are not speculative; they are deployed and validated.
Other applications are still firmly in the research phase. Integrating metabolomics with genomics to predict drug response, combining proteomics with transcriptomics to identify disease subtypes, or using epigenomics alongside expression data to infer regulatory mechanisms are all active research areas with promising results but no established clinical workflows. The methods work in well-controlled research settings with carefully curated samples and generous computational budgets. Whether they work in routine clinical settings with heterogeneous samples and limited bioinformatics support is an open question.
We help partners distinguish between these categories when planning projects. A multi-omics integration project that aims to use an established workflow (like tumour profiling) has a clear path to actionable results. A project that aims to discover novel multi-omics biomarkers should be framed as exploratory from the start, with realistic expectations about the likelihood of findings that replicate independently.
Where we stand
We include bioinformatics in our platform because molecular data is increasingly part of health analytics. But we describe multi-omics capabilities honestly. Where integration adds real analytical value, we support it. Where it is still exploratory, we say so. Partners deserve to know the difference between a validated workflow and a research prototype.
Our approach to multi-omics projects follows a staged model. Stage one is single-omics analysis: process each layer independently, answer layer-specific questions, and assess data quality. Stage two is cross-layer exploration: look for convergent signals across layers using methods that are robust to the sample size constraints. Stage three, if warranted by the stage two findings, is formal integration using methods appropriate to the data structure and sample size. Many projects produce their most useful findings at stage one or two. Stage three is not always necessary or possible.
This staged approach protects partners from overinvesting in integration methods that their data cannot support. It also produces interim results at each stage, so the project delivers value even if the full integration is not feasible. A partner who receives validated single-omics findings and convergent evidence across layers has actionable results, even if the formal integration model does not converge or produces unstable results.