Clinical decision-making and secondary findings in systems medicine
© Fischer et al. 2016
Received: 13 December 2015
Accepted: 10 May 2016
Published: 21 May 2016
Systems medicine is the name for an assemblage of scientific strategies and practices that include bioinformatics approaches to human biology (especially systems biology); “big data” statistical analysis; and medical informatics tools. Whereas personalized and precision medicine involve similar analytical methods applied to genomic and medical record data, systems medicine draws on these as well as other sources of data. Given this distinction, the clinical translation of systems medicine poses a number of important ethical and epistemological challenges for researchers working to generate systems medicine knowledge and clinicians working to apply it.
This article focuses on three key challenges: First, we will discuss the conflicts in decision-making that can arise when healthcare providers committed to principles of experimental medicine or evidence-based medicine encounter individualized recommendations derived from computer algorithms. We will explore in particular whether controlled experiments, such as comparative effectiveness trials, should mediate the translation of systems medicine, or if instead individualized findings generated through “big data” approaches can be applied directly in clinical decision-making. Second, we will examine the case of the Riyadh Intensive Care Program Mortality Prediction Algorithm, pejoratively referred to as the “death computer,” to demonstrate the ethical challenges that can arise when big-data-driven scoring systems are applied in clinical contexts. We argue that the uncritical use of predictive clinical algorithms, including those envisioned for systems medicine, challenge basic understandings of the doctor-patient relationship. Third, we will build on the recent discourse on secondary findings in genomics and imaging to draw attention to the important implications of secondary findings derived from the joint analysis of data from diverse sources, including data recorded by patients in an attempt to realize their “quantified self.”
This paper examines possible ethical challenges that are likely to be raised as systems medicine to be translated into clinical medicine. These include the epistemological challenges for clinical decision-making, the use of scoring systems optimized by big data techniques and the risk that incidental and secondary findings will significantly increase. While some ethical implications remain still hypothetical we should use the opportunity to prospectively identify challenges to avoid making foreseeable mistakes when systems medicine inevitably arrives in routine care.
KeywordsSystems medicine Big data Epistemology Translation Secondary findings Electronic health records Scoring systems Clinical decision-making
What is Systems Medicine?
The distinction between bottom-up simulation and top-down modelling is a subtle one, but one that is key to our analysis. Bottom-up simulation, utilizing analytical methods from bioinformatics, systems biology, and network theory, is intended to create computer simulations of biological processes at the level of constituent mechanisms, with the hope that such simulations can be used to predict how systems will respond to interventions or perturbations [4, 5]. These methods depend on a causal understanding of the biological mechanisms of interest. For this reason, research groups working within this strain of systems medicine tend to focus on clarifying signaling pathways and molecular pathways that connect, for example, genotypes with phenotypes .
Top-down modelling, on the other hand, does not depend on a detailed understanding of biological mechanisms. Instead, top-down models are developed through the statistical analysis of large datasets. This approach is analogous to Amazon’s recommendation engine, which utilizes a huge database of past purchasing behavior to predict which items individuals might want to purchase in the future. This approach, typically referred to as the big data approach, relies on statistical associations rather than mechanistic understandings . Amazon’s recommendation engine does not, for example, require an understanding of consumers’ motivations. The recommendations it provides to customers are based instead on the assumption that customers’ future purchasing behavior can be predicted by comparing their past purchasing behavior with that of customers who have purchased similar items. Similarly, top-down models in systems medicine do not begin with signaling pathways or molecular pathways, but rather with large datasets reflecting biological and clinical parameters.
As we will examine in this paper, the relationship between top-down modelling and clinical translation is a matter of some controversy within systems medicine. Some enthusiasts look forward to a day when top-down modelling can be incorporated into clinical information systems to generate useful predictions about individual patients’ health and their response to clinical interventions . Others understand top-down modelling as a hypothesis-generating step that might precede mechanistic research [16, 17] or the identification of promising interventions whose effectiveness would require empirical verification .
Within the vision for systems medicine, however, the ultimate goal for both bottom-up simulation and top-down modelling is the development of medical informatics tools. These tools would depend on the collection of large sets of clinical data, including, perhaps, data collected by patients using fitness trackers or other mobile devices. This data would then be used to support clinical decisions, create a long-term electronic “memory” of co-morbidities and other relevant information that could be applied to ongoing treatment, and even be utilized for future research that could be used to optimize these approaches [9, 10].
In short, then, systems medicine is not a single research model or clinical approach. It is a heterogeneous set of practices that have been promoted in recent years under a single name. Perhaps one explanation for this complexity, then, is the nascent state of these sciences. For the most part, applications of systems medicine have not reached clinical contexts with the exception of pilot projects. It is likely that as efforts to apply these ideas to clinical care continue, more clarity and focus will develop.
It would be premature, then, to suggest that we can anticipate or address all of the significant ethical challenges raised by systems medicine. Instead, we will focus here on a few key issues especially relevant to specific components of the field of systems medicine. First, we will examine the compatibility of big data methods with responsible clinical decision-making. This is not an ethical issue proper, but rather an epistemological concern that is likely to give rise to important ethical challenges. Second, we will examine a case study that highlights pitfalls in the application of probabilistic scoring systems to clinical practice. Third, we will examine the challenge of unintended findings that are likely to be raised as the types of medical informatics tools envisioned for systems medicine proliferate.
Systems medicine and clinical decision-making: the problem of epistemology
Epistemology is the field of philosophy that addresses questions related to knowledge: How do we know what we know? When should we be convinced, and how can we convince others, that new knowledge we have discovered is sound? How can we apply knowledge to real-life problems, such as the application of scientific knowledge to medicine?
Although these questions may seem far removed from research on applying omics-based technologies to clinical care, in truth they lie at the heart of ongoing debates within systems medicine. As we have observed, systems medicine is comprised of scientific approaches that are based on mechanistic understanding, which we have called bottom-up simulation, as well as those that are agnostic to mechanisms, which we call top-down modelling. In order to see why this distinction is important, and how it creates barriers for clinical applications of systems medicine, we can examine the same trends within a more mature movement in medicine: personalized medicine.
As we observed earlier, systems medicine overlaps somewhat with the vision for personalized medicine. Both approaches share a focus on identifying statistical associations using large datasets. But while personalized medicine focuses on genotype-phenotype associations, systems medicine additionally draws on other omics-based technologies, such as proteomics and metabolomics, as well as broad sources for health data, such as personal mobile devices. One important scientific tool within personalized medicine is the genome-wide association study (GWAS). This type of research involves the identification of genotypes anywhere in the genome that are statistically associated with health-related phenotypes, such as those related to pharmaceuticals dosing. Advocates for personalized medicine disagree about how to apply the knowledge generated by GWAS studies to clinical care . On the one hand, many personalized medicine researchers view genome-wide association study “hits” as the starting point for research into molecular mechanisms; they view GWAS as a hypothesis generating technology . Only once these mechanisms are elucidated could GWAS findings be applied to clinical care. This emphasis on molecular mechanisms can even determine which GWAS findings are considered valid. Some investigators, for example, use understanding of molecular mechanisms to aid in the analysis of GWAS hits, focusing on genetic variants in genes that are thought to be most likely to be causally-related to the phenotype under study.
On the other hand, some personalized medicine researchers do not insist on corroborating evidence derived from research on molecular mechanisms. They advance the idea that genotype-phenotype associations with a large effect size are directly applicable to clinical care. This assumption is most apparent in recent debates on secondary findings, in which some argue that genetic variants can be “actionable” even if the only evidence supporting an association with a phenotype was derived from GWAS studies . This perspective is closely linked with the big data approach, which suggests that top-down models can be effective for addressing real world problems, including recommendations for online retailers as well as health-related interventions.
For those with such a variant in CYP2C19, we would predict that clopidogrel can provide no benefit. If doctors tested for this variant prior to starting clopidogrel, it was assumed, they could identify patients who would not respond to clopidogrel and instead use a different medication.
In this case, scientists working on the pharmacogenetics of clopidogrel made a prediction based on a bottom-up understanding of the mechanisms through which clopidogrel acts in the body. It turned out, however, that this component of the human body, and the way it functions in its complex environment, was more complex than anticipated. It took clinical trials, however, to recognize that our mechanistic understanding was inadequate to guide clinical practice.
This is not to say, however, that bottom-up simulations and top-down models are always unsuitable for application to clinical care. Modern bottom-up simulations simply represents a more mathematical approach to understanding molecular mechanisms, and thus build on a longstanding tradition in medicine that bases clinical care on an understanding of human biology and physiology. This tradition extends back to 19th century physician and laboratory scientist Claude Bernard, and continues today in intensive care units and rare disease clinics around the world . Top-down models, however, have no true precedent in medicine, and thus require closer examination. In the next section, we will examine an example from outside medicine both to further demonstrate how top-down models are different from bottom-up simulations, and to demonstrate why the clinical application of top-down models should be treated with extreme caution.
The Enigma Machine
In many ways, systems medicine is based on Turing’s work. Later in his career, Turing grew interested in the idea of simulating one computing device using another computing device, even if the underlying mechanisms of the original device were not understood. That this is possible is often referred to as the “Turing-Church Thesis,” and the Turing Test refers to his ideas about how such a simulation could be evaluated.
He argued that a computer could be built to simulate the ability of a human to engage in conversation. He was not suggesting that the entire mind of a human would need to be understood in order to create such a simulation. Instead, he argued that a suitable simulation could be constructed by observing typical human conversations, and would be judged to be effective if it could pass what came to be known as the Turing Test. In this test, a human would engage in a dialog with the simulation. If the human could be fooled into thinking he was conversing with another human rather than a simulation, the simulation had passed the test .
Turing’s work in both stages of his career serves as an important inspiration for computation-based sciences like systems medicine. In fact, systems medicine includes both a modern equivalent of Turing’s mechanism-based work with the Enigma machine (bottom-up simulation), and the biological equivalent of his later work on simulating the output of the human mind (the Turing Test).
For systems medicine, however, the task of creating a computation-based simulation of the body is not as simple as “opening the box” in the way Turing and other British cryptologists were able to do at Bletchley Park. There are certainly scientists who have this vision for medicine – the vision of understanding the mechanisms of the human body is the focus of all experimental medicine from the time of Claude Bernard to today. But the modern, “big data” approach in systems biology is inspired by the later Turing’s vision. In the top-down approach, the goal is to examine the inputs and the outputs of the human body. The hope is that with enough data, systems biology will be able to generate a computer algorithm that is able to predict how the body will respond to inputs, without actually understanding the mechanisms of the body. For such a model, the internal workings of the body would remain a “black box.” But such an algorithm could, it is hoped, pass a type of clinical Turing test – to imitate the body well enough to guide therapy. Data about an individual patient could be gathered using clinical tests, electronic medical records, and self-monitoring data, and predict a specific output, such as the patient’s response to a variety of possible therapies.
The challenge with this “black box” approach to the human body, however, is that it precludes responsible decisionmaking by patients and healthcare providers.
In the tradition of evidence-based medicine (Fig. 5b), clinicians depend primarily on a very particular type of scientific knowledge – empirical findings from clinical research – and particular types of knowledge about the patient. They then utilize a very well-defined process of analysis to apply those clinical research findings to particular patients. Out of this deliberation comes a decision about a clinical plan.
Another prominent approach is the experimental medicine approach (Fig. 5c). In this model, clinicians combine understanding of the mechanisms of the human body with specific observations about the physiological condition of the patient. Using a process of reasoning based on mechanistic cause-and-effect, he or she is then able to develop a clinical plan.
With this in mind, we can begin to see the deep challenge that the top-down dimension of systems medicine poses – and indeed the challenge that many industries face as they work to apply big data analytics to their conventional practices (Fig. 5d). Top-down modelling provides clinicians with a computer algorithm, but that algorithm represents the patient’s body as a black box. It is not based on mechanistic science, and it is not based on clinical research. It simply takes the patient’s data, and generates an output. Neither clinicians nor patients have a way to evaluate its predictions. We are faced with the choice of following its predictions, or ignoring them. But this choice is arbitrary, and based purely on confidence in the black box.
We find this way of applying systems biology to medical practice to be untenable. In order for medical care to be successful, clinicians must be able to be accountable for their clinical decisions, and patients must be given grounds for placing trust in the expertise of their providers. Medical practice must therefore be based on clinical decisions rooted in scientific knowledge, not on the capricious predictions of a black box. Based on this analysis, then, we argue that the top-down dimension of systems biology must be viewed as a basic science – as an approach for generating hypotheses. Whatever potential this approach may hold, it must be mediated by more responsible approaches to medical care, and must therefore not be applied prematurely to clinical care.
In the next section, we will take this critique even further. Taking the Riyadh Intensive Care Program Mortality Prediction Algorithm as an example, we will demonstrate how provider responsibility and patient trust remain significant challenges for systems biology, even when evidence-based medicine has been utilized to validate the effectiveness of predictions based on top-down modelling. In other words, even if this approach works, we are still faced with a complex ethical decision about how and when to use it.
Ethical aspects of applying scoring systems to clinical care
Within the systems medicine approach, both top-down models and bottom-up simulations are intended for use in clinical care. Specifically, they are designed to combine data about a patient from a variety of sources (medical records, biomarkers, etc.) and apply this data to an algorithm designed to make predictions about how that patient’s illness will progress or how he or she will respond to a particular clinical intervention. Even when these algorithms are based on bottom-up study of biological mechanisms, they are still fundamentally statistical in nature. They are not based on a comprehensive understanding of the patient’s biological status and all mechanisms at play. Instead, they depend on a set of factors that can be measured relatively easily, combined with observations in other patients, to predict which outcomes are most likely.
This approach is fundamentally similar to earlier efforts to develop risk scores based on clinical indicators [16, 17]. Both approaches involve identifying (preferably strong) statistical links between clinical features that can be measured and outcomes of interest, using this data to quantify the likelihood that certain outcomes will occur, and using this score to inform clinical decision-making. Systems medicine simply optimizes this approach by incorporating a larger number of clinical measurements and allowing for more complex scoring algorithms. From an ethical perspective, however, both systems medicine tools and earlier scoring systems carry important implications for clinical practice. In this way, they are not simply pieces of information that are “good to know”. After all, when such scores are generated in clinical settings, an obligation is created for clinicians to respond to them.
Risk scores in ICUs – The example of APACHE II
As we have seen, an important focus of systems medicine is the generation of statistically-based clinical predictions, predictions that promise to inform clinicians of what is likely to happen, even if they can cast no light on the question of why this will happen. One general example of this in a clinical setting is a risk scoring system intended to predict the chances that a patient will survive: “Given all the information we have about patient (P) and correlated with our experience about the outcome of all the patients in similar circumstances for whom we have data in our database, the estimated chance of survival for patient (P) equals X %.” Framed this way, scoring systems can seem quite jarring and impersonal. But in fact, this approach to prognosis is already an important factor of care in intensive care units (ICUs) around the world; and has been for more than three decades. These scoring systems for classifying the prognosis of patients are not based on big data, but rather on more mundane measurements like a patients’ age and routine physiological measurements (temperature, mean arterial pressure, heart rate, respiratory rate, creatinine, hematocrit etc.). While a variety of such scoring systems have been used, one of the best known the APACHE II (Acute Physiology and Chronic Health Evaluation II), which was designed to measure the severity of disease for adult patients admitted to intensive care units . It is designed to be calculated on the day an adult patient is admitted to the ICU, and results in a computer-based score between 0 to 71. Higher scores indicate more severe disease and a higher risk for death. The link between high APACHE II scores, influenced primarily by the presence of multiple physiologic derangements, and risk for dying has been shown repeatedly in independent studies. When applied to the care of a single patient, an APACHE II score essentially generates a prediction about a patient’s risk for morbidity and mortality that is based on the observed rate of outcomes among the group of patients involved in earlier studies who had the same score. APACHE II scores can be used to inform decisions about which treatments should be offered, as well as for creating effectiveness and quality controls .
APACHE II is not really intended to generate predictions about an individual patient’s chance for survival. Methods have certainly been developed to support such predictions, but given the wide variety of factors that influence this type of outcome, but are not accounted for in the calculations, such methods are rather imprecise [29–31]. This is just one important reason that it is highly questionable to base decisions about discontinuation of therapy or withholding of interventions based on APACHE II scores and similar scoring systems. Not because these scores are irrelevant to prognosis, but because they generate a prognostic score that is not very precise.
As long as the big data-based scoring systems of the future are regarded as merely updated methods that use more data or more advanced algorithms to make predictions that can be used for similar applications, we would consider it as an improvement with no ethical concerns. However, a brief view back in recent history can provide a preview of the discussions that lay ahead if those in the systems medicine movement come to regard big data-based scoring systems as something more than just a more precise evolution of earlier scoring systems. Specifically, serious ethical questions are likely to arise if such methods are regarded as so precise they should be used to make decisions about withdrawing treatment or withholding life-extending interventions, based solely on the predictive score.
The Riyadh Intensive Care Unit Program
The Riyadh Intensive Care Unit Program was an effort developed in England in the 1990s that drew its name from the capital of Saudi Arabia, where data from several hundred thousand people had been collected and which formed the statistical basis for the program [32–34]. Unlike other scoring systems like APACHE II, the Riyadh software incorporated not only clinical observations, but also a measure of the intensity of treatment and nursing care that a patient required called the Therapeutic Intervention Scoring System (TISS). The software used these and other measures to generate a so-called “cost-performance profile.” This analysis combined prognostic information (like the APACHE II) and TISS with the intention of evaluating the therapeutic effort (which reflected the cost of treatment) in relation to the chances that an individual patient would survive. By including information outside of that relevant to a patient’s prognosis, the Riyadh program became a cost-benefit analysis. It was intended to suggest when the costs of a given therapy were out of proportion to the patient’s chance of survival; it generated an economic rationale for termination of treatment. Moreover, the Riyadh algorithm was considered to be highly accurate, with an accuracy in its predictions of 99.9 % [32–34].
Following the introduction of the Riyadh algorithm in German ICUs, a debate among physicians, professional societies, ethicists and politicians about the “death computer” ensued. The program was widely critiqued, especially for including economic rationale into the decisions to withdraw treatment [35, 36]. The authors themselves recognized the dangers, and warned: “We consider that the data on outcome predictions with the Riyadh Intensive Care Program have always been presented in a responsible manner. However, we advocate extreme caution in the use of any computer system which predicts death, and advise a careful explanation of its functioning and dangers” . Still, they stuck by the ideas behind their project: “Because of the problems of triage on admission, it is inevitable that a number of patients will be admitted to the ICU with a hopeless prognosis. Treatment of the critically ill with a hopeless prognosis is wasteful of precious medical resources, as well as having a demoralizing effect on the nursing staff and the patients’ relatives” .
Should an (expensive) therapy be withheld from patients who are unlikely to survive? Should a computer algorithm be allowed to determine, or even influence, the fate of the critically ill? For this examination, it is not necessary to delve deeper into the ethical debates around the allocation of medical resources or the role cost benefit analyses in medical care. Instead, we simply want to stress that these types of discussions are inevitable once the systems medicine approaches begin to incorporate a greater number of data sources, and lead to the development of algorithms with a higher predictive value. The utilization of non-medical data, up to and including economic data, open the door for abuse or, at the very least, the conflation of medical and non-medical considerations in decision-making. And the perception of greater predictive power inevitably increases the likelihood that such scores will be used as justifications apart from other important considerations, like provider judgments.
Ultimately the Riyadh algorithm failed. The beginning of the end came when further research called the reliability of its predictions into question. The authors of Riyadh found no false predictions of death, but several other investigations identified a survival rate of up to 41 % among those predicted to die . At least one German study has shown that the prognoses of experienced clinicians are at least as good as those generated by the Riyadh algorithm, and another study showed that of 53 patients with “high probability of death,” one third survived. Other important studies have identified a range of problems associated with the application of score-based mortality predictions to individual patients .
The Riyadh program never achieved the 99.9 % reliability that was promised. For a time, however, the belief that it could provide highly accurate predictions using all relevant data (including economic data) drove a significant amount of excitement. If, in the coming years, a big data-based scoring system attains the accuracy that was promised for the Riyadh algorithm, we expect these ethical issues will arise again. What role should patient’s wishes play? How do we respect the individuality of the dying life? What weight should be given to considerations of quality of life? What chance for survival should be accepted as for continuing treatment (after all, even if the chance of survival is at 2 %, this still is that 2 out of 100 patients will survive)? What effect should scoring systems have on the doctor-patient relationship, and vice versa? What justifications could be given for rejecting the “objective” prediction generated by an algorithm? What if the score indicates a good chance for survival but the patient does not want to continue treatment?
Even if we were to have access to flawless prognostic algorithms, we would still be left with these larger ethical questions about what to do with this prognostic information. In order to translate prognosis into recommendation, we need to know about the patients’ values . Even in the evidence-based medicine tradition, with its emphasis on basing decisions on empirical data, there is still an emphasis on taking patient preferences and perspectives into account when making clinical decisions . In what ways should systems medicine understand the balance between big data-informed predictions and individual patients’ values, lifestyles, and declarations of will?
At a more fundamental level, we even need to ask whether there is an appropriate role for big data predictions in medical care. Should serious decisions like those related to a change in therapeutic approach or the discontinuation of a therapy be linked with purely statistical interpretations of probability and risks? Or perhaps it is better to assume that the bigger the data and the better the score, the stronger would be the argument for not using it as a sole predictor? Not because of the data for which it can account, but because of the data it cannot. In the final analysis, we are likely to conclude that algorithm-based prediction is never the end, but only the beginning of a decision-making-process. This was even acknowledged by René Chang, one of the key developers of the Riyadh algorithm, defending his software: “If the computer makes a prediction of death, check that there have been no errors in data collection or entry; then determine if there is any treatment which has not yet been tried that might make a substantial difference to the patient’s outcome. If, in the clinicians’ judgment, there is such a treatment, then that treatment should be tried and the patient observed for a significant improvement. If there is no improvement or if in the first instance there were no further treatment regimen available, discussions concerning the withdrawal of treatment should commence with relatives and staff before a final decision is made” .
The ethical dimension of incidental or secondary findings in systems medicine
Based on current evidence, two types of medical studies seem especially prone to creating secondary findings in a future medical practice driven by analyses of EHR data: whole body imaging techniques and genomic sequencing [42–47]. Extensive evidence from the research context makes it clear that both of these technologies generate a huge amount of information that is of potential diagnostic, therapeutic or prognostic relevance to patients, or which indicates a need for further examination to determine its relevance [42, 46, 47].
Presently, neither whole body imaging nor genomic sequencing is widely used in clinical care. However, both are already available in research and direct-to-consumer contexts. Someday soon, however, these examinations are likely to be incorporated into clinical care and deposited in patient EHRs, assuming that either (a) data generated in the research or direct-to-consumer context will find its way into the clinical context or (b) these technologies will enter use in clinical settings to inform big data analyses. If this occurs, it is inevitable that this type of data will be applied to analyses not driven by active clinical questions – and thus generate findings that are incidental or secondary to the original reasons the tests were performed. Although these findings might be beneficial to patients, they can also create harms. In one study, for example, we studied participants taking part in a population based-study in Germany that involved whole body MRIs. When this research examination generated suspicious findings, some participants experienced significant harm and distress [42, 43, 46].
If systems medicine proceeds as envisioned, it will not just be the amount of stored patient data that will increase. The vision for systems medicine also includes the analysis of this data for research purposes, leading to an increase in the number of patient characteristics and biomarkers considered to have clinical relevance. Going further, a major goal of big data research is to identify associations among these factors that, in combination, can be used to make clinical predictions, creating a whole additional level of associations and correlations through which secondary findings can be generated.
Who is to decide which findings need to be communicated to providers and patients?
By what methods, in a world of big data and patient empowerment, should secondary findings be communicated?
How can secondary findings be prioritized responsibly within the treatment context so as to help patients and providers handle a large volume of such findings?
The discourse within human genetics over secondary findings gives us some idea about how these questions will be answered [44, 45]. It is already well-recognized, for example, that it can be difficult for patients and providers to anticipate such findings and prospectively develop a plan for addressing them. When data on the scale of whole genome sequence is generated, such a discussion would require the patient to be informed about the genetic variants that could potentially be identified beyond those relevant to the clinical question motivating the testing, but also understand the various implications for this information and the wide variability in the uncertainty accompanying such findings. Even in Germany, where legal standards are established for the design of such a genetic consultation process, the approaches used for addressing the challenge of secondary findings remains highly inconsistent overall.
The debate over incidental and secondary findings is certainly not new . In human genetics, for example, debates over best practices for informed consent, risk communication, and assessment of clinical validity, among others, predate the current discourse around Personalized or Individualized Medicine [49–51]. In fact, similar debates took place with the field of public health genetics in the 1990s . Nevertheless, systems medicine could provide the context for a range of technological developments that will increase the urgency for identifying effective methods for handling secondary findings and solving some of the difficult questions faced by bioethicists during the last decade and beyond.
In this paper we examine three possible ethical challenges that are likely to be raised as systems medicine comes to be translated into clinical medicine. These include the epistemological challenges for clinical decision-making created by black box algorithms, the use of scoring systems optimized by big data techniques capable of integrating non-medical data into routine healthcare and healthcare systems, and the risk that incidental and secondary findings will significantly increase as big data-driven EHRs are implemented in clinical care. At this point in time, all three of these concerns are somewhat conjectural. We might also have added other predictable issues with big data, including concerns related to privacy, data protection, and ownership of data . While such conjectural work runs the risk of leading to an ethics of “false alarms,” it also provides an important opportunity to prospectively identify challenges, gather empirical evidence earlier rather than later, and, hopefully, avoid making foreseeable mistakes. To be sure, excessive hype should have no place in either the science of systems medicine or the study of its ethical, legal and societal implications. However, in both of these domains it makes sense to continuously look ahead so we are ready for the future when it inevitably arrives.
Ethics approval and consent to participate
Consent for publication
Availability of data and materials
We thank Sally Werner for her contribution to harmonize the manuscript and references in form and style.
This publication originates from within the framework of the research network MENON - medical theoretical, ethical and economical evaluation of System Medicine. The MENON – consortium is financed by the Bundesministerium für Bildung und Forschung (federal ministry of education and research) (01GP1408A and 01GP1408B).
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
- Capobianco E. Ten challenges for systems medicine. Front Genet. 2012;3:193.Google Scholar
- Capobianco E. Dynamic networks in systems medicine. Front Genet. 2013;3:185.Google Scholar
- Hood L. Systems biology and P4 Medicine: Past, Present, and Future. RMMJ. 2013;4:e0012.View ArticleGoogle Scholar
- Price ND, Edelman LB, Lee I, Yoo H, Hwang D, Carlson G, et al. Systems Biology and Systems Medicine. In: Ginsburg J, Willard H, editors. Genomic and Personalized Medicine. London: Elsevier; 2013. p. 60–72.View ArticleGoogle Scholar
- Wolkenhauer O, Auffray C, Jaster R, Steinhoff G, Dammann O. The road from systems biology to systems medicine. Pediatr Res. 2013;73:502–7.View ArticleGoogle Scholar
- Schmutzler R, Huster S, Wasem J, Dabrock P. Risikoprädiktion: Vom Umgang mit dem Krankheitsrisiko. Dtsch Arztebl. 2015;112:A-910/B-762/C-738.Google Scholar
- Langanke M, Lieb W, Erdmann P, Dörr M, Fischer T, Kroemer H, et al. The Meaning of “Individualized Medicine”: A Terminological Adjustment of a Perplexing Term. In: Fischer T, Langanke M, Marschall P, Michl S, editors. Individualized medicine, ethical, economical and historical perspectives. Cham Heidelberg New York Dordrecht London: Springer; 2015. p. 11–28.Google Scholar
- Schleidgen S, Klingler C, Bertram T, et al. What is personalized medicine: sharpening a vague term based on a systematic literature review. BMC Med Ethics. 2013;14:55.View ArticleGoogle Scholar
- Zhao Z, Shen B, Lu X, Vongsangnak W. Translational biomedical informatics and computational systems medicine. Biomed Res Int. 2013;2013:237465.Google Scholar
- Sarkar IN. Biomedical informatics and translational medicine. J Transl Med. 2010;8:22.View ArticleGoogle Scholar
- Krüger-Brand HE. 2014. Forschungsinfrastrukturen: Herausforderungen durch Big Data. Dtsch Arztebl. 2015;111:A-1735/B-1497/C-1429.Google Scholar
- Erdmann P, Fischer T, Raths S, Flessa S, Langanke M. Systemmedizin: Herausforderungen eines aktuellen Ansatzes. Dtsch Arztebl. 2015;112:A-1330/B-1116/C-1091.Google Scholar
- Gomez-Cabrero D, Menche J, Cano I, Abugessaisa I, Huertas-Migueláñez M, Tenyi A, et al. Systems Medicine: from molecular features and models to the clinic in COPD. J Transl Med. 2014;12 Suppl 2:S4.View ArticleGoogle Scholar
- Mayer-Schönberger V, Cukier K. Big Data: A Revolution That Will Transform How We Live, Work, and Think. Boston: Houghton Mifflin Harcourt; 2013.Google Scholar
- Flores M, Glusman G, Brogaard K, Price ND, Hood L. P4 medicine: how systems medicine will transform the healthcare sector and society. Pers Med. 2013;10(6):565–76.View ArticleGoogle Scholar
- Binder H, Blettner M. Big Data in der wissenschaftlichen Medizin – eine biostatistische Perspektive: Teil 21 der Serie zur Bewertung wissenschaftlicher Publikationen. Dtsch Arztebl Int. 2015;112:137–42.Google Scholar
- Mainzer K. Algorithms are powerful and useful. But by themselves they are blind. Faszination Forschung. 2014;15:64–75.Google Scholar
- Khoury MJ, Ioannidis JP. Big Data Meets Public Health. Science (New York, NY). 2014;346(6213):1054–5.View ArticleGoogle Scholar
- MacArthur D. Genome-wide association studies: failure or success? 2009. http://www.wired.com/2009/04/genome-wide-association-studies-failure-or-success/. Accessed 7 May 2015.
- Roden DM, Pulley JM, Basford MA, et al. Development of a large-scale de-identified DNA biobank to enable personalized medicine. Clin Pharmacol Ther. 2008;84(3):362–9.View ArticleGoogle Scholar
- Brothers KB. The Genome on the Horizon: Practical Reasoning in the Age of Personalized Medicine. 2015. http://etd.library.vanderbilt.edu/available/etd-03232015-115000/unrestricted/Brothers.pdf. (Accessed 17 May 2016).
- Mao L, Jian C, Changzhi L, et al. Cytochrome CYP2C19 polymorphism and risk of adverse clinical events in clopidogrel-treated patients: a meta-analysis based on 23,035 subjects. Arch Cardiovasc Dis. 2013;106(10):517–27.View ArticleGoogle Scholar
- Trenk D, Hochholzer W. Genetics of platelet inhibitor treatment. Br J Clin Pharmacol. 2014;77(4):642–53.View ArticleGoogle Scholar
- Osnabrugge RL, Head SJ, Zijlstra F, et al. A systematic review and critical assessment of 11 discordant meta-analyses on reduced-function CYP2C19 genotype and risk of adverse clinical outcomes in clopidogrel users. Genet Med. 2015;17(1):3–11.View ArticleGoogle Scholar
- Hodges A. Alan Turing: The Enigma. Princeton, N.J.: Princeton University Press; 2014.View ArticleGoogle Scholar
- Turing AM. Computing machinery and intelligence. Mind. 1950;59:433–60.View ArticleGoogle Scholar
- Knaus WA, Harrell Jr FE, Lynn J, Goldman L, Phillips RS, Connors Jr AF, et al. The SUPPORT Prognostic Model. Objective Estimates of Survival for Seriously Ill Hospitalized Adults. Ann Intern Med. 1995;122:191–203.View ArticleGoogle Scholar
- Moreno R, Afonso S. Ethical, legal and organizational issues in the ICU: prediction of outcome. Curr Opin Crit Care. 2006;12:619–23.View ArticleGoogle Scholar
- Schuster HP. Die Bedeutung von Score-Systemen für die Voraussage von Behandlungsergebnissen in der Intensivmedizin. Internist. 1996;37:1237–43.Google Scholar
- Murdoch IA, Bihari DJ. Predictions in the intensive care unit – in search of an “unholy grail”! Pediatr Nephrol. 1993;7:708–10.View ArticleGoogle Scholar
- Lefering R, Wolfrum B, Wauer H, Neugebauer EA. Limitations of score-based daily outcome predictions in the individual intensive care patient. An example of the RIAHDH algorithm. Inflamm Res. 2004;53 Suppl 2:169–74.Google Scholar
- Chang RW, Jacobs S, Lee B. Predicting outcome among intensive care unit patients using computerized trend analysis of daily Apache II scores corrected for organ system failure. Intensive Care Med. 1988;14:558–66.View ArticleGoogle Scholar
- Chang RW. Individual outcome prediction models for intensive care units. Lancet. 1989;2:143–6.View ArticleGoogle Scholar
- Chang RW, Bihari DJ. Outcome prediction for the individual patient in the ICU. Unfallchirurg. 1994;97:199–204.Google Scholar
- Weber R, Vogt-Weber B. Computerprognose – Ärztliche Entscheidungen zwischen Statistik und Intuition. Juristische Aspekte von Prognose und Behandlungspflicht. MedR. 1999;5:204–9.View ArticleGoogle Scholar
- Zylka-Menhorn V.“Todescomputer”: Risikobewertung von Intensivpatienten. Dtsch Arztebl. 1997;94:A-621/B-523/C-499.Google Scholar
- Jacobs S, Arnold A, Clyburn PA, Willis BA. The Riyadh Intensive Care Program applied to a mortality analysis of a teaching hospital intensive care unit. Anaesthesia. 1992;47:775–80.View ArticleGoogle Scholar
- Grundmann RT. Prognostizierbarkeit des Todes – Ärztliche Beurteilung oder Scores? In: Junginger T, Perneczky A, Vahl C-F, Werner C, editors. Grenzsituationen in der Intensivmedizin. Berlin Heidelberg: Springer; 2008. p. 153–63.View ArticleGoogle Scholar
- Downar J. Even without our biases, the outlook for prognostication is grim. Crit Care. 2009;13:168.View ArticleGoogle Scholar
- Evidence-Based Medicine Working Group. Evidence-based medicine. A new approach to teaching the practice of medicine. JAMA. 1992;268:2420–5.View ArticleGoogle Scholar
- Chang RW, Bihari DJ. The Riyadh Intensive Care Programme. Anaesthesia. 1993;48(4):345–6.Google Scholar
- Erdmann P. Zufallsbefunde aus bildgebenden Verfahren in populationsbasierter Forschung. Münster: mentis; 2015.Google Scholar
- Erdmann P. Handling Incidental Findings from Imaging within IM related Research – Results from an Empirical-Ethical Study. In: Fischer T, Langanke M, Marschall P, Michl S, editors. Individualized medicine, ethical, economical and historical perspectives. Cham Heidelberg New York Dordrecht London: Springer; 2015. p. 231–50.Google Scholar
- Rudnik-Schöneborn S, Langanke M, Erdmann P, Robienski J. Ethische und rechtliche Aspekte im Umgang mit genetischen Zufallsbefunden – Herausforderungen und Lösungsansätze. Ethik Med. 2014;26:105–19.View ArticleGoogle Scholar
- Langanke M, Erdmann P, Robienski J, Rudnik-Schöneborn S, editors. Zufallsbefunde bei molekulargenetischen Untersuchungen. Medizinische, juristische und ethische Perspektiven. Heidelberg Berlin: Springer-Verlag; 2015.Google Scholar
- Schmidt CO, Hegenscheid K, Erdmann P, Kohlmann T, Langanke M, Völzke H, et al. Psychosocial consequences and severity of disclosed incidental findings from whole-body MRI in a general population study. Eur Radiol. 2013;23:1343–51.View ArticleGoogle Scholar
- Brothers KB, Langanke M, Erdmann P. Implications of the incidentalome for clinical pharmacogenomics. Pharmacogenomics. 2013;14:1353–62.View ArticleGoogle Scholar
- Heinemann T, Hoppe C, Listl S, Spickhoff A, Elger CE. Zufallsbefunde bei bildgebenden Verfahren in der Hirnforschung. Ethische Überlegungen und Lösungsvorschläge. Dtsch Arztebl. 2007;104:A-1982/B-1748/C-1684.Google Scholar
- Grabe H, Assel H, Bahls T, Dörr M, Endlich K, Endlich N, et al. Cohort profile: Greifswald approach to individualized medicine (GANI_MED). J Transl Med. 2014;12:144.View ArticleGoogle Scholar
- Fischer T, Langanke M, Marschall P, Michl S, editors. Individualized medicine, ethical, economical and historical perspectives. Cham Heidelberg New York Dordrecht London: Springer; 2015.Google Scholar
- Fischer T, Dörr M, Haring R, Langanke M. Alarming symptoms of a paradigm shift? An approach to bridge the gap between hypothetical ethics and the current status of IM research. In: Vollmann J, Schildmann J, editors. Personalised Medicine: Ethical, Medical, Economic and Legal Critical Perspectives. Fulham: Ashgate Publishing Limited; 2015. p. 25–40.Google Scholar
- Dabrock P. Public Health Genetics and Social Justice. Community Genet. 2006;9:34–9.View ArticleGoogle Scholar
- Mittelstadt BD, Luciano F. The Ethics of Big Data: Current and Foreseeable Issues in Biomedical Contexts. Sci Eng Ethics. 2015;29:1–39.Google Scholar