Introduction

The 50 years since the first successful solid organ transplantation in humans in 19531,2 have been marked by an exponential increase in our understanding of the basic immunological processes involved in organ transplantation, rejection and allograft tolerance.3 This knowledge has resulted in the identification of immunogenic drug targets and resultant rational drug design. Transplantation is now a well-established routine treatment strategy for end-stage kidney, heart, liver and lung diseases.4 Nevertheless, new diagnostic and predictive monitoring tools are desperately needed to customize the delivery of immunosuppressive drugs for optimal patient and organ care, such that graft life is extended and patient morbidity minimized.

The past 50 years have also seen an explosive evolution in the fields of genetics and molecular biology: the discovery of the DNA double-stranded helix by Watson and Crick in 1953;5 the completion of the Human Genome Project in 2003;6 and the further refinement of the analytical tools that enable high-resolution sequence mapping of human structural variation.7 Using these insights and with the advent of high-throughput technical devices, molecular biology has undergone an impressive transformation, with the recent advent of informative fields of biology such as the elucidation of the following: the importance of epigenetics on the DNA transcription process (most importantly DNA methylation and histone modifications);8 alternative messenger (m)RNA splicing;9 RNA interference (RNAi; that is, RNA-mediated post-transcriptional regulation through noncoding microRNA [miRNA] or small interfering RNA [siRNA]);8,10,11 post-translational modifications;12,13 and regulation at the metabolic level14,15 (Figure 1).

Figure 1: Current paradigm of molecular biology.
figure 1

DNA is packed in chromosomes in tertiary structures defined by the histones. Histone modifications and DNA methylation are epigenetic factors that can influence the transcription process. DNA is transcribed into pre-mRNA, which is spliced into different mRNA molecules by alternative splicing. Noncoding RNAs, such as siRNA and miRNA, are also transcribed from DNA, and these small RNA fragments affect mRNA stability and regulate the translation of mRNA into proteins. In addition, post-translational modifications of proteins (for example, folding, cleavage, and chemical modifications) contribute greatly to the diversity of the proteome. These proteins can function as enzymes, affecting DNA repair or replication, the transcription process or translation, or can act as a catalyst for metabolic reactions. Proteins are also involved in cell signaling and ligand binding, and as structural elements. Each level interacts closely with the other level: protein–DNA and RNA–DNA interactions and chemical modifications of DNA that do not affect the primary DNA sequence also affect the molecular biological processes in cell systems. Each of these molecular biological processes can be studied with specialized omics tools—genomics/epigenomics, transcriptomics, proteomics and metabolomics—which study the complete sets of genes, transcripts, proteins and metabolites. Abbreviations: dsRNA, double-stranded RNA; Me, methylation; miRNA, microRNA; mRNA, messenger RNA; pre-mRNA, precursor messenger RNA; siRNA, small interfering RNA.

“It is the responsibility of those of us involved in today's biomedical research enterprise to translate the remarkable scientific innovations we are witnessing into health gains”, stated Dr Elias Zerhouni, former Director of the NIH, in 2005.16 In this Review, we discuss some of the opportunities that molecular diagnostic tools have to offer both basic scientists and translational researchers for bench-to-bedside clinical applications in transplantation. Terms used in the text are defined in Box 1.

General concepts in molecular diagnostics

Single-molecule versus high-throughput strategies

The classic paradigm that the DNA of the cell will determine its fate is too simplistic. Variations in the genome sequences of humans or other species only partly explain the fundamental nature of the cells and organisms, and are often only marginally implicated in disease processes.17 The different levels of biomolecular organization and control, such as regulation of the transcriptome (the catalogue of RNA transcripts, including mRNA, siRNA and miRNA), the proteome (the complete set of proteins) and the metabolome (the complete set of small-molecule metabolites) (Figure 1) are interdependent and are greatly affected by environmental events and stresses throughout life. An analysis of the full spectrum of cellular organization is therefore necessary to gain insight into the mechanisms of disease.18

Each of these aspects can be studied at the level of individual molecules, but advances in engineering have allowed for increased data throughput, enabling the study of complete sets of molecules (omics) with improved speed, accuracy, and cost-effectiveness compared with previous experimental models of repeated single-molecule experiments, and with potential for the generation of new hypotheses. The omics approach has impressively accelerated biomarker discovery and our understanding of multifactorial biological processes and diseases. The major differences between omics strategies and single-molecule approaches obviously lies in throughput (hundreds of thousands versus one or a few), but also in experimental design. Classic single-molecule biomedical research is based on hypothesis testing, building new experiments based on prior observations and theory, such that classic scientific research tends towards a reductionist approach in understanding disease and disease processes, owing to the limitations of most technologies and the complex nature of pathological systems. This approach, which drills deeply into a focused area and eliminates confounding variables, had great importance in advancing our biological insights, and will continue to do so.

The application of omics technologies and the analysis of the genome, transcriptome, proteome or metabolome, however, creates unprecedented opportunities for improving understanding of complex systems and diseases at a global level, and enables hypothesis generation with fewer a priori assumptions. Therefore, the technological progress in genome-wide measurements has shifted the scientific discovery process from hypothesis-driven approaches towards more data-driven approaches. Nevertheless, omics approaches are also based on assumptions about the nature of the experiments (the theory underlying the central paradigms in molecular biology, the phenotypes of the studied samples and sometimes the pathways deduced from previous experiments), as well as assumptions about the functioning of the instruments used to perform the experiments. After a first omics hypothesis-generating approach, classic single-molecule experiments should ensue, in order to delve further into the mechanisms of the newly generated hypotheses.

Biomarker versus therapeutic target discovery

Omics technologies are currently used in two closely related areas: biomarker discovery and the elucidation of pathophysiologic processes to identify novel therapeutic targets. However, clear differences exist between these aspects of translational research in terms of the ultimate goals, optimal matrix, optimal platforms and preferred study design (Table 1).

Table 1 Differences in omics study design for biomarker identification versus therapeutic target discovery

Determining the immunologic threshold

The 1st International Conference on Transplantomics and Biomarkers in Organ Transplantation was held in San Francisco in 2010. Hosted by The Transplantation Society, the conference brought together myriad disciplines in transplantation research including genomics, transcriptomics, proteomics, metabolomics, informatics, next-generation sequencing technologies, imaging and clinical transplantation. Individualized therapy was the focus of the conference, and the prominent topic was the search for both diagnostic and predictive biomarkers for allograft dysfunction that could be used in personalizing treatment. Other themes highlighted throughout the conference included the use of protocol and for-cause biopsies for microarray and histological analysis, the use of noninvasive methods for biomarker discovery, the pitfalls of relying solely on gold standard criteria for biopsy classification, low-cost and time-efficient diagnostic and analytic tools, defining and identifying causative versus correlative biomarkers, and compiling and harnessing information in the public domain.

The term 'immunosuppression' derives from the general sense that the immune system is solely oriented in an offensive antiallograft response and thus needs continuous suppression. However, it has become increasingly apparent that alloimmunity is far more complex and diverse. For example, T regulatory (TREG) cells control the immune response in such a way that their suppression per se would actually augment rejection. Similarly, allospecific T cells are now recognized as having widely variable effector capabilities and activation requirements based on prior history: naive T cells seem to be exceptionally dependent on antigen presentation and co-stimulation, whereas antigen-experienced, or memory, T cells are triggered primarily through the T-cell receptor without substantial requirements for accessory or co-stimulatory molecule engagement.19 Additionally, nonalloantigen exposure is capable of influencing alloimmune reactivity through heterologous immunity, a process that has been postulated as an explanation for the difficulties in tolerance induction.20

The net result of these findings is that alloimmune competence is now recognized as being fluid, changing with ongoing immune experience, age, and disease state. Even donor characteristics, such as differing degrees of human leukocyte antigen (HLA) mismatch, donor age, cold ischemia time and organ type, change the degree to which antigens and accessory and co-stimulatory molecules are presented and thus alter the effective allospecific precursor frequency. Thus, each patient can be considered a mosaic of alloimmune potential that is specific for a particular donor–recipient pair at a particular time and situation. Simply declaring the immune system as requiring suppression defies the complexity inherent in the relevant biology; the concept of 'immunosuppression' should give way to deliberate and individualized 'immune modulation'.

The complexity of the immune responses after allotransplantation can initially seem daunting. Development of an immune repertoire is fundamentally a random process and responsiveness toward any particular complement of alloantigens can be considered equally stochastic. However, the aggregate threshold of alloimmune responsiveness changes in reasonably predictable ways. In general, alloimmune responsiveness is greatest at the time of transplantation, and wanes with time. As the need for more intense immunosuppression wanes, the overall need for maintenance immunosuppression varies with two predominant variables—precursor frequency and memory. Precursor frequency relates to T-cell repertoire compatibility and is influenced most notably by HLA mismatch.21 Memory is a function of prior exposures, either to alloantigens or to sufficiently cross-reactive environmental antigens. These standard variables help to establish rational choices for therapeutic agents based on increasingly well-defined susceptibilities that individual immune cell types have to commonly available immunosuppressive drugs.

Therapeutic decisions based on individualized immune modulation can be distilled down to a balanced appraisal of risk versus benefit. In transplantation, an unmet need exists for specific and sensitive, noninvasive biomarkers that can track graft injury and graft survival and replace the renal transplant biopsy as the gold standard. The application of molecular diagnostics may uncover novel biomarkers. Importantly, the discovery of such biomarkers would allow the dose (or load) of immunosuppression to be customized for individual patients, to identify the particular threshold between rejection and infection that avoids allorecognition yet retains normal immunity to infection and cancer.

Molecular diagnostics in transplantation

Long-term outcomes in transplantation are hampered by an incomplete understanding of the mechanisms driving acute and chronic alloimmune injury, immunological quiescence, operational tolerance and tissue injury. The absence of sufficiently robust biomarkers further complicates the clinical management of allograft recipients; better diagnostic biomarkers could potentially improve outcomes. Here, we provide an overview of transplantation research in humans that used different applied molecular cell biology tools for improving mechanistic insight and biomarker discovery (Figure 2).

Figure 2: Application of omics in renal transplantation.
figure 2

The outcome of renal transplantation is determined by pretransplantation, peritransplantation and post-transplantation factors. Omics tools can be used either for biomarker discovery or for elucidation of the molecular mechanisms underlying pathophysiologic processes. Peripheral blood contains almost exclusively recipient-derived cells, and genetic profiles are recipient-driven. Transcriptomics, proteomics and metabolomics research in peripheral blood is promising for both biomarker and therapeutic target discovery. Genotype analysis in kidney biopsy samples gives information on the donor genotype, although extensive graft inflammation with recipient-derived cells could also lead to the presence of recipient's DNA in biopsy samples. Omics analyses in biopsy specimens can be used for both therapeutic target studies and for biomarker studies, although the invasive procedure represents a major drawback of potential tissue-derived biomarkers. Finally, urine analysis using omics tools offers a great window for biomarker discovery, but the low amount of DNA in urine samples makes it unsuitable for genomic studies. Abbreviations: BKV, BK virus; CMV, cytomegalovirus; HLA, human leukocyte antigen; UTI, urinary tract infection.

Genetics, genomics and epigenetics

Candidate gene approach for DNA polymorphisms

DNA polymorphisms are variations in the DNA sequence that are present in the population at a frequency >1%. These polymorphisms can be single nucleotide polymorphisms (SNPs), a variable number of tandem repeats (microsatellites and minisatellites), deletions or insertions. If such polymorphisms occur in the coding region of a gene, they may or may not alter the structure of the mature protein, depending on the impact of the gene sequence variation on the amino acid sequence. Polymorphisms in noncoding regions may also influence gene splicing, transcription and translation. Although the degree of health and disease determined by inherited DNA sequences outside of all the currently known genetic diseases is not known, the inherited genetic diseases do tell us that single gene molecular changes encoded for in DNA can have remarkable implications for health and can lead to multisystemic disease.

As in other medical disciplines, association studies between genetic variation and disease conditions have also been performed in the field of transplantation. Most of these studies have been performed using a candidate gene approach.

Over the past few decades, knowledge about the highly polymorphic HLA system, an important determinant of outcomes after transplantation, has increased greatly.22 This knowledge has been translated into routine clinical practice, with many laboratories routinely using DNA-based HLA typing (instead of serological typing),23 and has proven especially valuable in bone marrow transplantation.24,25 However, current HLA genotyping is limited to a small portion of the HLA region in the genome. Detailed, more comprehensive HLA genotyping might be important for immunology and transplantation, although this idea has not been studied to date.

The majority of the other studies addressing genetic variation in transplantation have been performed in the fields of rejection,26,27 ischemia–reperfusion injury,28,29,30,31 chronic allograft damage32,33 and pharmacogenetics.34 The well-known association between polymorphisms in the gene encoding thiopurine methyltransferase (TPMT) and azathioprine toxicity35 illustrates that genetic variation research using the candidate gene approach can be of high clinical importance; the FDA recommends that consideration is given to determining a patient's TPMT genotype or phenotype before azathioprine is used. Other pharmacogenetic markers (for example, ABCB1 and CYP3A5) also have potential value in the clinical setting.34,36,37,38 Validation of these early results is needed, however, and the long-term benefits of genotyping for these additional markers on patient outcome remains to be shown.

In contrast to the clinical success of HLA genotyping and pharmacogenetics, data linking clinical outcome parameters in transplantation with SNPs in genes encoding cytokines and cytokine receptors, chemokines and chemokine receptors, complement genes, co-stimulation molecules, growth factors and adhesion molecules, innate immune genes and genes implicated in ischemia–reperfusion injury and chronic allograft damage, have been inconsistent and remain without clinical application.26,27,33 No candidate genes are ready for their clinical value to be tested in prospective trials, mainly because studies to date have been relatively small, have exhibited selection bias, have insufficiently accounted for confounding factors and have had ill-defined and short-term outcome variables; such limitations have led to conflicting results and poor reproducibility. Better designed and adequately powered studies are urgently needed to overcome these issues and bring genotyping into the transplantation clinic.

Genome-wide association studies

Genetic analyses using the candidate gene approach have not been very successful in translating genotype analyses into the clinic. Apart from flaws with study design and underpowering in the majority of the studies, the manual selection of potentially interesting genes is not the most efficient way to link variation in genome sequence with outcome parameters.39,40 The number of SNPs recorded in the public HapMap database41,42 and the Single Nucleotide Polymorphism Database (dbSNP)43 has reached several millions, and upcoming massive parallel sequencing tools will enable the screening of these SNPs and many more, as will be discussed. Genome-wide association (GWA) studies, which use accessible tools for high-throughput genotyping, have now become a standard method for disease gene discovery,44 and the catalogue of published GWA studies is growing rapidly.45

One drawback with GWA studies is that a substantial number of recent GWA studies (albeit in fields other than transplantation) found that only a few common variants were implicated in human disorders, and that the associated SNPs explain only a small fraction of the genetic risk.46,47 The relevance of findings from standard GWA studies is indeed based on a 'common disease, common gene' hypothesis: the results are relevant when the polymorphisms associated with a disease are present sufficiently frequently in a population that is sufficiently large. In recent years, many GWA studies have successfully elucidated such polymorphisms,40 but many more GWA studies are necessary to elucidate the true influence of common DNA variants on disease susceptibility. However, some researchers argue that although many genes have been linked with disease processes, many have too low an impact on the risk of the disease to be biologically relevant.46,48 With this reasoning, the limited role of common variation in many heritable diseases would suggest that many rare or less frequent variants remain to be found. More in-depth study of these disease-associated rare variants might better reveal novel therapeutic targets or improve the tailoring of treatment regimens to an individual's risk profile, compared with the 'common disease, common gene' approach.39,48 In this light, the advent of newer tools such as exome arrays and whole-genome sequencing (through next-generation sequencing techniques, as will be discussed), and expanded reference panels such as the 1,000 Genomes project,49 will better determine the degree of complex diseases caused by alterations in DNA sequence. Surely GWA studies are a first but very important step in the process of identifying genes to improve patient management.

At least one multicenter GWA study is being conducted in clinical transplantation (a UK–Irish consortium project funded by the Wellcome Trust Case–Control Consortium),50 and is likely to help us to define the genetic variation in recipients' genomes that determines transplantation outcome. As each transplantation involves two genomes, donors' genomes are also likely to affect transplantation outcome, and interactions between donors' and recipients' genomes will need to be taken into account in GWA studies in transplantation.

Epigenetics and epigenomics

Epigenetic alterations encompass heritable regulation of gene expression that does not involve DNA nucleotide sequence variation. The epigenome constitutes an array of molecular modifications to both DNA and chromatin, the most extensively investigated of which are DNA methylation, and changes to the chromatin packaging of DNA by post-translational histone modifications.8,51 Interestingly, environmental exposure to nutritional, chemical and physical factors has the potential to affect gene expression and alter disease susceptibility through changes in the epigenome.

Epigenetics and epigenomics are exciting and critically important new areas of research. Recent years have seen development of novel techniques for studying chromatin state on a genome scale. Most of these techniques are based on chromatin immunoprecipitation (ChIP), with fragment identification by hybridization to a microarray (ChIP-chip). Nowadays, ChIP sequencing (ChIP-Seq) is becoming available, whereby ChIP DNA is directly sequenced using next-generation massively parallel sequencing tools.52 Aside from the use of ChIP-Seq in cancer research,53 one of the biological systems most studied using ChIP-Seq is T cells,54,55 which is of special interest in the field of transplantation. The first studies demonstrated the importance of histone modifications for T-cell responses54,56,57 and ischemia–reperfusion injury.58 T cells are not the only immune cells regulated by epigenetic phenomena: natural killer cell characteristics59 and B-cell development60 are also governed by epigenetic mechanisms.

Approaches for the analysis of genome-scale DNA methylation have also been developed, and the different approaches (for example, array-based versus massively parallel sequencing-based approaches) have competing strengths and weaknesses.61 DNA methylation analysis is also of interest in the field of immunology and transplantation. Methylation of CpG residues has been shown to repress expression of Foxp3, which is involved in TREG cell development and function, whereas complete demethylation is necessary for stable Foxp3 expression.62,63 Many more studies have investigated the influence of differential DNA methylation on T-cell responses, as is reviewed elsewhere.57 Another proof-of-concept study for the potential usefulness of DNA methylation for transplantation research demonstrated that modification of methylated cytosines at a regulatory site within the promoter of complement component C3 may occur as a result of ischemia–reperfusion injury.64

As the importance of the epigenome for gene expression, cell development and cell function is only beginning to be recognized, epigenetic and epigenomic research is expected to become important in the future; improved accessibility of newer tools for screening the epigenome will likely contribute to this success.52,61 However, epigenetics research in the field of transplantation and immunology has so far exclusively been in the form of basic research. Translational studies with more direct clinical relevance should be initiated to transfer the successes of epigenetics to the clinic.

Gene expression and transcriptomics

Candidate gene approach for gene expression

DNA is transcribed to RNA, and the complete set and quantity of RNA transcripts in a cell at a specific developmental stage or physiological condition is called the transcriptome. The transcriptome is dynamic and changes in different circumstances owing to different patterns of gene expression. Hypothesis-driven candidate mRNA real-time polymerase chain reaction (RT-PCR) studies have been performed in experimental and clinical transplantation and have contributed to current insight into the functioning of the immune system and rejection, the mechanisms of action of immunosuppressive drugs, the mechanisms involved in tissue injury and repair, cell senescence, epithelial–mesenchymal transition (EMT), and the study of chimerism. No field in transplantation research has been untouched by RNA studies, and quantitative (RT-PCR) or spatial (in situ hybridization) molecular analysis of transcripts in allograft biopsy samples provides a sensitive and powerful means of evaluating these specimens.

Gene expression studies are not only useful in the detection of underlying pathogenic processes. The concept that gene expression in transplanted tissue alters before histological damage is visible has been demonstrated for ischemia–reperfusion injury,65,66 acute rejection67 and chronic allograft damage,68,69 and could represent a means to foresee and perhaps prevent future injury by timely intervention.70 In addition, if validated in larger ongoing multicenter studies, analysis of mRNA in urine71 or peripheral blood72,73 could become a useful and, importantly, noninvasive biomarker for ongoing intragraft damage such as rejection or progressive chronic histological injury, as is reviewed elsewhere.74 The advent of high-throughput techniques does not mean that standard techniques are irrelevant. On the contrary, the targeted gene approach will probably be more cost-effective than the high-throughput microarray techniques for gene expression analysis in a clinical setting, as just a few well-selected gene targets should suffice as clinical biomarkers; the interpretation of microarray data is more complex.

Gene expression analysis with microarrays

Microarray technology has revolutionized gene expression research by allowing the screening of gene expression of thousands of genes simultaneously.75,76 Several commercial gene expression microarrays are currently available for researchers. Among these widely used platforms are Affymetrix's GeneChip®, Agilent's oligonucleotide microarrays, Illumina's microarrays, and Roche's Nimblegen® microarrays. Each of these platforms have strengths and weaknesses, and researchers must choose the appropriate platform based on their priorities.77

As the technical and biostatistical analytical tools are more complicated than candidate gene approaches, and because they must be controlled for multiple testing, gene expression microarrays have been used by only a few groups in the field of transplantation. Nevertheless, studies performed over the past decade have covered the full spectrum of transplantation biology—including acute rejection, operational tolerance, chronic allograft rejection, and the identification of biomarkers, in liver, heart, pancreas, kidney, and lung transplantation, as reviewed elsewhere.78,79,80 For example, the use of gene expression microarray technology on renal allograft biopsy samples has demonstrated the molecular heterogeneity of allografts at the time of transplantation,81,82 acute rejection83,84 and after chronic histological damage.85,86,87,88

Microarray expression analysis has also been used to study the periphery. This approach offers insight into the pathways of acute and chronic rejection, and of induced and operational tolerance.89,90,91 Additionally, the application of microarray technology on peripheral blood samples is a suitable platform for the efficient discovery of new noninvasive biomarkers for rejection or chronic allograft damage.92,93,94,95 However, gene expression microarray technology has its own disadvantages, such as a lower dynamic range than RT-PCR, a high cost, decreased specificity owing to cross-hybridization, errors in probe design, differences in concept and results between the different microarray platforms, insufficient coverage of splice variation and annotation difficulties, as is reviewed elsewhere.80 For implementation in clinical practice, a good compromise will likely be found in custom-made arrays or in commercial high-density oligoarrays with fewer but well-chosen probes.

The splice variant issues have been partly overcome by the development of exon expression or 'whole genome tiling' arrays. This approach has uncovered the immense complexity of the transcriptome: gene expression varies much more greatly than was previously believed.96 However, many of the limitations of standard DNA microarrays remain, including reliance upon existing knowledge about the genome sequence, high background levels owing to cross-hybridization, and a limited dynamic range.97 Moreover, akin to what is seen with standard DNA microarrays, comparing expression levels across different experiments is often difficult and can require complicated normalization methods.

Newer technologies using high-throughput (next-generation) sequencing tools, such as RNA sequencing (RNA-Seq), aim to overcome many of the above problems by revealing unprecedented detail while simultaneously improving quantification.97,98 The faster, cheaper and more reliable analysis of gene expression will potentially advance transcriptomics, and RNA-Seq is expected to replace microarrays for many applications that involve determining the structure and dynamics of the transcriptome. However, although these new technologies might improve the quality of transcriptome profiling, the greater challenge associated with omics research—how best to generate biologically meaningful interpretations of complex datasets that are sufficiently interesting to drive follow-up experimentation—will remain (Box 2).

Noncoding RNA (miRNA and siRNA)

The transcriptome contains many more RNA fragments than the mRNA that is translated into peptides and proteins. Among these noncoding RNAs, miRNAs are well known and form a populous class of evolutionarily conserved, naturally occurring RNA fragments (22 nucleotides long). Apart from miRNAs, other types of small RNAs have also been found, including endogenous siRNAs and Piwi-interacting RNAs (piRNAs). All of these small RNAs function as guide RNAs within the broad phenomenon known as RNA silencing (or RNAi), through their interaction with Argonaute proteins that regulate the translation process (translational repression) or guarantee mRNA cleavage (mRNA degradation).10,11,99,100,101 Since their discovery, hundreds of small noncoding RNAs have been identified, and it is predicted that the actual diversity is a plurality of this.

It is believed that miRNAs are involved in human disease;102 noncoding RNA fragments are also likely to be of importance in the field of transplantation and immunity. In addition, miRNAs are implicated in the regulation of immune cell development and in the modulation of innate and adaptive immune responses.103,104,105 Moreover, in clinical transplantation, an internally validated miRNA expression profiling study demonstrated that miRNA expression in renal allograft biopsy samples correlates with acute T-cell-mediated rejection and renal allograft function, with circumstantial evidence that the differentially expressed miRNAs are largely derived from infiltrating immune cells.106,107,108 Although this result confirms previous findings that different cell types and different cell activation states are associated with specific miRNAs,109 it remains to be elucidated whether these miRNA expression alterations are pathogenic or whether they represent only the infiltration of the transplanted organ with activated immune cells.

The ultimate goal of small noncoding RNA research is to discover disease-causing or disease-modifying small RNA fragments that could subsequently be used as biomarkers or for specific targeted therapy. Indeed, siRNAs have the potential to silence disease-relevant genes that cannot be shut down with available drugs. Moreover, specific miRNAs that promote disease processes could also themselves be targets for shut-down. Despite the relatively recent discovery of small noncoding RNA, RNAi is already being used therapeutically in human clinical trials, and biotechnology companies that focus on RNAi therapeutics are emerging.110 One of the most problematic issues is how to move small RNAs efficiently and specifically to their target sites in the human body; in addition, safety, effectiveness and reliability issues need to be addressed.110

Regardless of its great therapeutic potential, RNAi has already revolutionized basic biomedical research, allowing researchers to rapidly, efficiently and accurately repress the function of a gene of interest. The coming years promise to be exciting in the field of small noncoding RNA, both in basic and in translational research, and possibly even in clinical trials.

Proteins and antibodies

Gel-based and mass spectrometric approaches

The development of genomics and transcriptomics notwithstanding, gene polymorphisms and transcript levels correlate incompletely with the expression level of the functionally active proteins, which more accurately reflect actual cellular events than genomics or transcriptomics.111 This poor correlation between genotype, gene expression and the localization or activity of the ensuing proteins is caused by the complex regulation of the transcriptome and the broad range of post-translational modifications that change the properties of proteins (for example, activity, localization, turnover, kinetics, and interactions with other proteins).112,113 Proteins therefore provide a better picture of events that occur inside an organism and provide ideal biomarkers for disease conditions, as has been extensively demonstrated for proteins such as prostate-specific antigen, CA125, CA19-9 and carcinoembryonic antigen.114 Proteomics and peptidomics aim to identify, quantify, analyze and functionally define a large number of proteins and peptides in biological processes or disease states. However, as genomics and transcriptomics techniques and analysis tools are well established and have matured over a couple of decades, their strengths and weaknesses are relatively well addressed in study designs and analysis methods. By contrast, proteomics and peptidomics tools have only just begun to gain widespread use. These methods still suffer from issues such as a lack of widely accepted standards in quality control, normalization, data analysis, and protein identification criteria, which makes comparison of data from different groups difficult. More robust proteomics and peptidomics methods are evolving and the number of publications in this field is increasing exponentially.115,116,117

Over the past few decades, new tools have revolutionized proteomics and peptidomics. Proteomics tools include gel electrophoresis (for example, two-dimensional gel electrophoresis [2DE] and two-dimensional difference in gel electrophoresis [2D DIGE]) and gel-free methods using mass spectrometry (for example, surface-enhanced laser desorption/ionization with time-of-flight mass spectrometry [SELDI-TOF-MS], liquid chromatography mass spectrometry [LC-MS] and matrix-assisted laser desorption/ionization [MALDI]) are increasingly being used for proteomics and peptidomics.113,114,115,117 Mass spectrometry was initially used to identify and characterize isolated proteins and to profile the mass of components in biological or clinical samples, but has evolved into a technique for interrogating complex proteomes by matching mass spectra to sequence databases to identify proteins.118

Proteomics methods are also increasingly being used in the field of organ transplantation. Because urine is the ideal noninvasive specimen for renal diseases,119 the number of proteomic studies of urine has surged, and urine proteomics are a promising tool for the noninvasive diagnosis of acute rejection120,121,122,123,124 and chronic allograft histological damage.125,126,127 However, although sensitivity and accuracy of these tools have improved over the last decade, the complexities of tissues and biological fluid proteomes often exceed the capacity of proteomic analysis tools. Even the very best technologies available can interrogate only a small fraction of the total proteome in any sample and the more complex the sample, the more this limitation is a problem for discovery and interpretation of results. These detected proteins are also usually the most abundant, and efforts to resolve these issues and improve sensitivity and detection are underway. In proteomics and peptidomics, a compromise needs to be reached between in-depth analysis and identification of individual protein and peptide components on the one hand, and broad proteome coverage on the other hand. The most challenging step in proteomics and peptidomics research is the integration of seemingly unrelated findings of various protein fragments into a rational pathogenic pathway or into a reasonable biomarker. In addition, the value of urinary biomarkers is limited in anuric patients and in patients in whom urine may come from native rather than transplant kidneys.

Protein and antibody microarrays

Enzyme-linked immunosorbent assays (ELISAs) have long been the gold standard for quantifying protein expression in tissue or body fluids, and remain popular today. Multiplexed assays based on ELISA have become available and are especially useful when samples only need to be analyzed for a handful of molecules (for example, in biomarker qualification). Building on the same principle, high-density protein arrays have been introduced in recent years, and encompass any platform in which a large number of proteins are immobilized on a solid support in a spatially addressable manner (for example, planar arrays such as Protoarray® by Invitrogen or ProteoScan Cancer Lysate Arrays by Origene) or a spectrally addressable manner (for example, bead-based arrays such as Luminex® platform by Luminex Corporation).128 High-density microarrays enable the analysis of protein interaction networks, the generation of protein phosphorylation maps, and the identification of disease-specific autoantibodies. Protein arrays are therefore usually used for high-throughput profiling of secreted factors such as cytokines, chemokines and growth factors, as well as for screening for autoantibodies129 or cancer antibodies.130,131 As the diversity of proteins is greater than that of nucleic acids, true proteomic antibody arrays, targeting a significant portion of a tissue proteome, are not practical, unlike genomic or transcriptomic level arrays. Protein arrays will probably be developed for specific hypotheses in specific tissue types.

Antibody microarrays are also of special interest in transplantation research because pathogenic donor-directed alloantibodies can form. These antibodies can target HLA molecules and lead to antibody-mediated rejection, but non-HLA antibodies can also be implicated in allograft rejection. First screening experiments demonstrated the potential usefulness of the Protoarray® platform in human transplantation, and revealed increased levels of a large set of non-HLA antibodies in the post-transplantation serum of renal allograft recipients.132,133,134 In 2009, another study evaluated antibody responses using Protoarray® in tolerant recipients of combined kidney and bone marrow transplants from HLA haplo-identical donors, and showed the development of B-cell alloimmunity and autoimmunity in patients with T-cell tolerance to the donor graft.135 Although the data from these studies need confirmation and the pathogenic capacity of these antibodies needs to be examined, and although it is not known why so many autoreactive antibodies exist in the circulation, protein microarray technology offers an interesting approach for antibody research in transplantation.

Metabolomics

Mass spectrometry and NMR

Rather than characterizing and quantifying large macromolecules (DNA, RNA, proteins, and peptides), as is done in genomics, transcriptomics and proteomics, metabolomics focuses on small molecules (<1,000 Da), catabolic and metabolic products arising from the interactions of these large macromolecules.136 The correlation between transcriptomics or proteomics and metabolic changes is poorer than expected as the relationship between mRNA levels and protein expression is not simply linear, and even if the protein is expressed, its concentration may not correlate with its activity. The metabolome is a more direct reflection of cell functionality.

Researchers have already catalogued more than 7,900 metabolites in the human body; these molecules can be either endogenous or exogenous, and include minerals, elements, carbohydrates, hormones, and phospholipids, among others.137 Metabolites are not inert and are more than end products of metabolism. They have many biological functions and represent accurate reporters of an organism's physiology, being involved in processes such as glycolysis, gluconeogenesis and lipid metabolism. Metabolites such as citrate, lactate, and glucose can reflect situations such as apoptotic alterations, hypoxia and oxidative stress. Therefore, metabolite concentrations often track with specific disease conditions, turning metabolomics into an efficient tool for distinguishing health from disease.

The study of the whole metabolome (metabolomics, also known as metabonomics) was made possible by technological breakthroughs in small molecule separation and identification. Such breakthroughs include mass spectrometry, nuclear magnetic resonance (NMR), high-pressure liquid chromatography, capillary electrophoresis and ultrahigh pressure liquid chromatography. As outlined in a 2008 review,136 metabolomics can be approached in two very distinct ways: in a targeted profiling manner with identification of single metabolites, or in a pattern recognition fashion where spectral patterns and intensities are recorded rather than individual molecules. In order to promote and speed up targeted profiling studies, the Human Metabolome Database (HMDB) was created to provide reference mass spectrometry and NMR spectra, metabolite–disease associations, metabolic pathway data, and reference metabolite concentrations (for diagnostic comparisons) for hundreds of human metabolites.137,138

Metabolomics has been used by very few groups in transplantation, for the evaluation of metabolic processes involved in ischemia–reperfusion injury,139 acute rejection,140 drug toxicity,141 graft dysfunction and graft damage.141,142,143 Although these first studies showed some promising results, all were hampered by small sample sizes and insufficient validation. More systematic and collaborative research is necessary to bring metabolomics closer to the clinic.141

Chemical genetics and chemical genomics

Another area of critical future importance is chemical genetics and chemical genomics (also known as chemogenomics or chemical biology), the study of the influence of small molecules (rather than genetic intervention) on biological processes; that is, research on the cross-section between chemical space and biological activity. Chemical genomics can be seen as the omics counterpart of chemical genetics, being performed less on an ad hoc basis and more systematically than chemical genetics.144 In the past, the study of small molecules was hampered by the complexity and high cost of synthesizing and purifying these compounds. Nowadays, the advent of combinatorial chemistry has enabled the synthetic preparation of large collections of diverse compounds, using relatively few chemical steps. In addition, advances in robotic technology and informatics now allow hundreds of thousands of compounds to be screened in a single day, orders of magnitude greater than was possible a decade ago.

The integration of large collections of compounds with advanced high-throughput omics technologies (such as genomics, transcriptomics, proteomics, metabolomics) represents a powerful way to accelerate the discovery of potential drug targets.145 In this light, Chembank, the public, web-based informatics environment that contains freely available data derived from small-molecule screens performed by the National Cancer Institute's Initiative for Chemical Genetics,146 is likely to have a major role in future drug target development, both in academic and in pharmaceutical industry research. With respect to transplantation, it is interesting to note that many chemical genetics studies have been performed with sirolimus and tacrolimus, yielding novel insights into the downstream signaling pathways of the targeted proteins.144

Nanoparticles and functional imaging

Molecular imaging reporters of specific processes and activities are becoming available and offer exciting opportunities for the noninvasive tracking of disease processes. Fluorescent, nuclear or magnetic probes can be bound to metabolites involved in metabolic processes, and can act as in vivo markers of metabolic activity.147,148 Studies in cancer and atherosclerosis show that these techniques are easily translatable into the clinic.149,150 Targeted molecular imaging tools can also track immune cells, as was shown for lymphocytes with specific targeting of the deoxyribonucleotide salvage pathway151 and for monocytes,152 and this could be especially interesting in transplantation research. In 2009, molecular imaging reporters of either phagocytosis or protease activity detected cardiac allograft rejection noninvasively by providing a functional three-dimensional map of macrophage accumulation, at least in mice.153 Whether these targeted molecular imaging tools could be of interest for clinical transplantation has not yet been studied.

Future challenges and opportunities

Next-generation sequencing

The application of the newer genome-wide molecular diagnostics has led to a rapid increase in the complexity of translational biomedical research. Upcoming high-throughput techniques such as massively parallel sequencing for genotyping, epigenomics and transcriptomics are expected to further accelerate the pace of translational research.154,155,156,157,158,159 These high-throughput sequencing technologies ('next-generation sequencing' tools) include Illumina's Genome Analyzer systems, Applied Biosystems' SOLiD system, Roche's 454 sequencing system, Helicos Biosciences' tSMS system156 and Pacific Biosciences' SMRT® system.160 With these new technologies, the cost of omics technologies is fast approaching a threshold at which omics will become a routine part of the diagnostic armamentarium, and with a modest investment, data acquisition for any genomic project can become comprehensive, integrating multiple genome-wide data sets with great depth and resolution.

Clinical application of omics data

The flood of research data concerning the genetic basis of health and disease, and the analytical and interpretative problems that come with this wealth of data, risk the further alienation of clinicians from basic sciences. True translational research should be multidimensional and collaborative, to ensure adequate data retrieval, correct data flow analysis and relevant clinical interpretation. The complexity of the clinical parameters, the complex physiology and immunology of many disease processes, the multitude of data per sample and the complex data analyses can only lead to biologically and clinically meaningful results with a multidisciplinary, holistic approach involving clinicians, molecular biologists, bioinformaticists and biostatisticians. Truly showing a 'return on investment' from functional genomics will require findings to be taken beyond the hypothesis-generating stage and integrated in downstream research applications, in discovery pipelines, or as biomarkers in clinical practice (Figure 3).

Figure 3: An integrative omics approach for translational research.
figure 3

A growing gap exists between the generation of high-throughput molecular data and the translation of omics features into clinically meaningful concepts. Standard clinical research is performed by associating clinical features with pathophysiologic processes, histological alterations and identification of risk factors or comorbidities. The integration of these clinical data leads to the identification of disease networks of sometimes related phenomena such as acute and chronic ABMR or chronic ischemia with calcineurin-inhibitor nephrotoxicity. Omics research aims to integrate molecular diagnostic data into molecular networks. The great challenge in translational research today is to integrate the complex clinical networks with the even more complex molecular networks, in order to identify clinically relevant molecular networks and reclassify diseases on the basis of the underlying molecular alterations. Abbreviation: ABMR, antibody-mediated rejection.

A first step to bridge clinical (evidence-based) medicine and genomics-based translational science would be to collect and store DNA from patients enrolled in clinical trials, which would provide a resource for interrogating the role of genomic variation in treatment response and disease evolution. Trials that use genomic data for guiding clinical decision making would also be needed. As is outlined by the Evaluation of Genomic Applications in Practice and Prevention (EGAPP) Working Group,161 the analytical validity, clinical validity and clinical utility of genomic tests will need assessing extensively before these tests can be adopted into clinical practice.

As outlined above, the currently accepted paradigms in molecular biology offer an integrated view on health and disease processes, with sophisticated regulation at each molecular level and indispensable feedback regulation between the different molecular levels. This complex interplay between the genome, transcriptome, proteome and metabolome18 has mostly been overlooked in traditional translational research and in the field of transplantation. Research groups often have access to few molecular diagnostic tools and collect information from single types of genomics data. The integration of different omics tools (integrative genomics), however, offers great advances in understanding human disease: each technology interrogates different aspects of gene function and each technology tends to produce noisy data and can be associated with its own inherent experimental limitations.114,162 Previous studies have demonstrated that data integration can improve the sensitivity and specificity for detecting true functional relationships among genes,163,164 and the use of integrated approaches is increasing rapidly, also in transplantation research.88,95,132

Bioinformatics and biostatistics

Multiple testing corrections

The important discrepancy between the vast amount of data obtained per sample and the often limited number of samples included in genomics studies is, however, a major problem for the translation of omics technologies to clinical practice (Box 2). This problem is known as the 'curse of dimensionality' (too many features in one sample), combined with the 'curse of dataset sparsity' (too few clinical samples).165 Bioinformatics and biostatistics have provided new tools to partly overcome this problem; false discovery rates are now recognized as marker of significance in omics studies, while standard P values should be reserved for testing individual hypotheses in classic experiments.166,167 Proper and sufficiently large validation sets will, however, still be necessary to obtain reliable and interpretable data. Many easy-to-use analysis packages have now been developed to avoid the need for tiresome programming that is out of the reach of many laboratories. Important tools for bioinformatics, such as Significance Analysis of Microarrays (SAM),168 GenePattern,169 GenMAPP,170 R software facilities and Bioconductor,171 are often free to download and are widely used in translational science.

Confounder analysis

Omics studies in transplantation typically include only a small number of transplant recipients. If the complexity of the patient population and the number of potential confounding factors is taken into account, the statistical power of many of these studies must be considered insufficient.141 This problem could be partly overcome by compiling and analyzing publicly available data in a sort of meta-analysis. Indeed, most omics data are included in publicly available databases such as the Gene Expression Omnibus (GEO),172 which offers unique opportunities for a more systematic meta-study of combined smaller studies, with unprecedented power for revealing commonalities and differences across different organ systems and different diseases.173,174 However, although expression and sequencing data are often publicly available, clinical data are often not, which obviates the integration of the molecular measurements with complex clinical features and phenotypes.

Molecule annotation and pathway analysis

The evolving annotation of probesets (that is, the mapping of probes in an omics platform to established gene identifiers), is a second major issue in bioinformatics for omics data. Probeset annotation often changes as the annotation for the underlying sequences changes over time, and care should be taken to use updated annotation files for analysis.175,176,177 It is not only the annotation of probes that constitute an omics platform that change over time. In pathway analysis, results depend on the validity of the pathways under consideration, and inconsistencies or evolution of the annotations of different molecules and their biological processes, location or function, will erratically be adopted in the final conclusions of pathway analysis.178,179

In addition, the integration of single-molecule data into biologically relevant networks is needed (Figure 3). Molecular networks and genesets can be manually curated, based on prior literature findings, or can be newly assembled using data obtained from omics techniques. The gene sets used for network analysis can be divided into five broad categories: curated gene sets (canonical pathways defined by domain experts); locational gene sets (based on chromosomal location); motif gene sets (based on the sharing of regulatory motifs); computational gene sets (based on prior experimental data); and Gene Ontology gene sets (based on Gene Ontology annotations for cellular components, biological processes or molecular functions). Many tools have been developed to map these functional networks, to assess the importance of candidate networks in complex disease processes and to evaluate single molecules as potential therapeutic targets within these networks. For this purpose, pathway analysis tools have been developed, enabling easy data mining without labor-intensive customized programming: GenePattern and the included Gene Set Enrichment Analysis tool (GSEA); the Database for Annotation, Visualization and Integrated Discovery (DAVID); Onto-Tools' PathwayExpress; GeneGo Metacore; Ingenuity Pathway Analysis; and BIOBASE ExPlain Analysis.

Absence of anatomical detail and cell specificity

Finally, typical genomics experiments collapse the cells in a given tissue, be it peripheral blood, biopsy tissue, urine or other samples. Although the tools seem to be sensitive for differential molecular regulation, they are not specific for localization within tissues. This lack of specificity is a major drawback of many omics studies, and spatial/anatomical information will only be reached with additional sample visualization techniques, microdissection and cell sorting.180,181 Obviously, these strategies are time-consuming and expensive. Recently, a novel method—cell-specific significance analysis of microarrays (csSAM)—was used to show that statistical deconvolution approaches can estimate cell-type-specific gene expression from heterogeneous samples, by weighted computational analysis of specific white cell compartment compositions available from an individual's differential white cell count.182 This approach helps in the virtual dissection of gene expression localization, and illustrates how the integration of clinical and laboratory values with molecular diagnostics can enhance understanding without additional effort.

Conclusions

Improved cross-linking and public repository of available data sets, together with in-depth, integrative data mining and smart biological interpretation, will ultimately lead to the more rapid development of translational research, easier validation of small studies in larger cohorts and a more efficient pipeline from hypothesis generation to validation and application. In 2009, Hirschhorn wrote: “Each discovery of a biologically relevant locus is a potential first step in a translational journey, and some journeys will be shorter than others. With a more complete collection of relevant genes and pathways, we can hope to shorten the interval between biologic knowledge and improved patient care.”40 We fully agree with this statement.

Review criteria

We searched PubMed and Google Scholar using the following terms: “transplantation”, “immune system”, “genome” and “genomics”, “genome-wide association”, “DNA polymorphisms”, “epigenetics” and “epigenomics”, “DNA methylation”, “gene expression” and “transcriptomics”, “mRNA degradation”, “translational repression”, “RNA interference”, “micro-array”, “next-generation sequencing”, “proteomics” and “peptidomics”, “metabolomics” and “chemical genomics”. Full-text publications relevant to the field were selected and relevant articles from reference lists and relevant review articles were also included. The search was limited to papers published in the English language.