Policy Significance Statement
The tension between data protection and using data for the public good, such as in health and medical research, is an ongoing policy issue. There are a growing number of data platforms, which aim to either ease data sharing between institutions and researchers or give greater control to individuals over their data. These platforms reflect a growing shift in data governance away from institutional control toward greater individual control. Looking into the future, there is a growing interest in collective data governance of data for the public good. These emerging data platforms and the parallel shift in data governance pose broader challenges to both the health system and health and biomedical research with regard to data control.
1 Introduction
Health data, especially big health data, have been publicized as having the potential to improve health and medical research, health systems, and healthcare (Hripcsak et al., Reference Hripcsak, Bloomrosen, FlatelyBrennan, Chute, Cimino, Detmer, Edmunds, Embi, Goldstein, Ed Hammond, Keenan, Labkoff, Murphy, Safran, Speedie, Strasberg, Temple and Wilcox2014; Kruse et al., Reference Kruse, Goswamy, Raval and Marawi2016; Van Dijck and Poell, Reference Van Dijck and Poell2016; Sonja et al., Reference Sonja, Ioana, Miaoqing and Anna2018). However, because health data are often considered to be a particularly sensitive type of personal data, its use beyond individual care has been limited (Hafen et al., Reference Hafen, Kossmann and Brand2014; Hripcsak et al., Reference Hripcsak, Bloomrosen, FlatelyBrennan, Chute, Cimino, Detmer, Edmunds, Embi, Goldstein, Ed Hammond, Keenan, Labkoff, Murphy, Safran, Speedie, Strasberg, Temple and Wilcox2014; Presser et al., Reference Presser, Hruskova, Rowbottom and Kancir2015; Van Roessel et al., Reference Van Roessel, Reumann and Brand2017). In addition to data collected in healthcare institutions (e.g., hospitals), there are also collections of biomedical research data (e.g., genomic data) that sit in data silos (Brand et al., Reference Brand, Evangelatos and Satyamoorthy2016; Carbon et al., Reference Carbon, Champieux, McMurry, Winfree, Wyatt and Haendel2018). Finally, there are new sources of data related to individual health, such as person-generated health data from wearables and health devices or inferred from other data (e.g., social media posts). Equally, these new data sources are not readily accessible to individuals, health systems, or researchers (Van Kleek and OHara, Reference Van Kleek and OHara2014; Hodson, Reference Hodson2016; Verborgh, Reference Verborgh2019). In this context, there is a growing number of platforms, which propose to “break down” the traditional data silos promoting both greater access to data for researchers, while also giving control of data back to individuals (Mählmann et al., Reference Mählmann, Reumann, Evangelatos and Brand2017; Riso et al., Reference Riso, Tupasela, Vears, Felzmann, Cockbain, Loi and Rakic2017; Starkbaum and Felt, Reference Starkbaum and Felt2019). Alongside these platforms are new data governance strategies posing opportunities and challenges to current established practice.
This paper uses four case studies to explore the shifts in the control of health data (“Emerging Data Platforms” and “Data Control Dimensions” sections). Emerging models of data governance will also be considered, along with the potential impacts of these shifts on health systems (“Discussion” and “Conclusions”).
1.1 Health data
Traditionally, health data included data collected within specific health services, such as tests, medical record, or biomedical research data. In recent years, what is considered as health data have expanded to include data as genomic data and patient-generated health data (Wood et al., Reference Wood, Bennett and Basch2015; Maglaveras et al., Reference Maglaveras, Kilintzis, Koutkias and Chouvarda2016; O’Doherty et al., Reference O’Doherty, Christofides, Yen, Bentzen, Burke, Hallowell, Koenig and Willison2016; Vayena et al., Reference Vayenal, Brownsword, Jane Edwards, Greshake, Kahn, Ladher, Montgomery, O’Connor, O’Neill, Richards, Rid, Sheehan, Wicks and Tasioulas2016). Patient-generated health data have been defined by Shapiro et al. (Reference Shapiro, Johnston, Wald and Mon2012, p. 2) as “health-related data—including health history, symptoms, biometric data, treatment history, lifestyle choices, and other information-created, recorded, gathered, or inferred by or from patients or their designees (i.e., care partners or those who assist them) to help address a health concern.” Purtova (Reference Purtova2017) further makes the argument, in the age of data analytics, all data should be considered health data. This statement resonates due to the potential for health information to be gathered from any data. There is also a closing of the gap between research data and clinical care data. Traditionally, these two data sources have been treated in distinctly different ways. New data platforms are making this distinction porous, challenging how current regulations are applied to these platforms (Deverka et al., Reference Deverka, Majumder, Villanueva, Anderson, Bakker, Bardill, Boerwinkle, Bubela, Evans, Garrison, Gibbs, Gentleman, Glazer, Goldstein, Greely, Harris, Knoppers, Koenig, Kohane, La Rosa, Mattison, O’Donnell, Rai, Rehm, Rodriguez, Shelton, Simoncelli, Terry, Watson, Wilbanks, Cook-Deegan and McGuire2017).
1.2 Big data, data protection, and protecting health
Healthcare and health research have gradually moved toward greater data dependence (Weber et al., Reference Weber, Mandl and Kohane2014; Starkbaum and Felt, Reference Starkbaum and Felt2019). The shift toward data-driven models of care and research is happening in tension with a greater focus on privacy and data protection, as captured by the European Union’s General Data Protection Regulation (EU) 2016/679 (GDPR) or the United States Health Insurance Portability and Accountability Act (HIPAA). Although privacy and data protection are common policy themes, there are no agreed upon policy direction for the use of big data in healthcare (Blasimme et al., Reference Blasimme, Fadda, Schneider and Vayena2018). Mills (Reference Mills2019, p.14) describes this as a laissez-faire model of data governance, in which “individuals provide their data on a pseudo-transactional basis in exchange for ‘free’ services.” External regulation (e.g., HIPPA, GDPR) acts as a constraint on these transactions, setting standards regarding the transaction (e.g., notice and consent). However, beyond these constraints, there is little in the way of a data governance framework that involves individuals after the initial transaction. Driving toward a more data-driven health sector is the rise of big data and data analytic, which facilitates the collection, analysis, and linking of large sets of health data (Luo et al., Reference Luo, Wu, Gopukumar and Zhao2016; Wang et al., Reference Wang, Kung and Byrd2018). There is great interest in linking together the different types of health data outlined above to provide new insights. For example, what social media activity can tell us about health behaviors (Young, Reference Young2014). Interestingly, much of these data are not under the control of those who generate it, as the quote below from the Obama’s White House report into Big Data highlights (Executive Office of the President, 2014, p. 9).
“Certainly data is freely shared and duplicated more than ever before…The technological trajectory, however, is clear: more and more data will be generated about individuals and will persist under the control of others.”
2 Emerging Data Platforms
There are several existing platforms that promote giving control of health data back to individuals or dissolution of data silos. Rather than attempt to provide a comprehensive overview of these platforms and their features, this paper explores four case studies that provide insight into the challenges and opportunities these platforms raise. Each represent a way to approach the control of data generated in a traditional health institution and person-generated health data.
2.1 Case study 1: Open Humans and Nightscout Data Commons
Open Humans (www.openhumans.org) is a community-centric ecosystem. Launched in 2015, Open Humans now has over 8,000 members on the platform. Delivered as an online platform, it supports individuals importing and aggregating various types of personal data. These include personal health data such as wearable device data and genomic data—from various sources. Members registered with the platform are encouraged to decide how their data are used, for personal data exploration, as well as sharing with participant-led or academic research projects (Greshake Tzovaras et al., Reference Greshake Tzovaras, Angrist, Arvai, Dulaney, Estrada-Galiñanes, Gunderson, Head, Lewis, Nov, Shaer, Tzovara, Bobe and Price Ball2019). By aggregating personal health data from a variety of sources and putting the individual in control of the data, Open Humans aims to provide individuals access to their own data—breaking down silos—and increasing control over their data. Individuals can make granular decisions if and how to share those data.
The governance of the data held in Open Humans is handled at multiple levels. At the basic level, governance of data is performed by the individual, who has full control over who accesses any part of their data. Additionally, the larger community performs governance functions for the platform; individual projects invite individual members to share their data, these need to be approved before they are publicly listed to solicit data donations. The approval process is performed by community members reviewing prospective new projects. In addition, community members vote for the board members of the nonprofit foundation running the Open Humans platform.
When a member shares data, they give permission for diverse data re-use. This infers that Open Humans facilitate, the formation of "data commons" for specific data types—including by patient communities. Noteworthy examples of this are projects aggregating continuous glucose monitor (CGM) data, associated with the type 1 diabetes community: the Nightscout Data Commons (managed by the Nightscout Foundation) invites data donation from individuals using Nightscout open-source software for CGM data; users of Open Artificial Pancreas System (OpenAPS) for automated basal insulin regulation are invited to donate data to the OpenAPS Data Commons (managed by Dana Lewis, a lead OpenAPS developer). Lewis has managed the re-use of these data in further research and academic collaborators, for example, to assess glycemic control among OpenAPS users (Melmer et al., Reference Melmer, Züger, Lewis, Leibrand, Stettler and Laimer2019).
2.2 Case study 2: Health Record Banking
Health Record Bank (HRB) is a patient-centric health information repository (Yasnoff, Reference Yasnoff2016). HRB has been described by the Health Record Banking Alliance (2012) as an “independent organizations that provide a secure electronic repository for storing and maintaining an individual’s lifetime health and medical records from multiple sources and assure that the individual always has complete control over who accesses their information.” The repository is governed by three key principles: (a) each patient’s record should be functionally stored in one place (currently, not all patient records in the same place), (b) each patient should control who accesses to their medical records, and (c) medical records should be stored under the patient’s control by a trusted organization. These principles facilitate and safeguard HRB storage, access, quality, control, security, and privacy. HRB proposes that an independent entity allows individuals to set up a HRB, which health institutions can then deposit health information into. The individual decides who can withdraw information from their account. While privacy has culturally dependent variables (Sahama et al., Reference Sahama, Simpson and Lane2013), the HRB data model enhances patient empowerment, including sharing electronic health records with functional data interoperability and meaningful use to minimize the security and privacy concerns. The HRB data architecture does not support all patient records in a single repository. Thus, it is less vulnerable to loss of the entire dataset from a single unauthorized intrusion (Yasnoff, Reference Yasnoff2016). HRB is still in its infancy, but has garnered support from a number of stakeholders in the healthcare system. There have been pilots in the United States, which was discontinued due to low enrolment of people (Yasnoff and Shortliffe, Reference Yasnoff and Shortliffe2014).
2.3 Case study 3: Unforgettable.me
Unforgettable Research Services Pty Ltd (www.unforgettable.me) provides a memory prosthetic and personal data brokerage service (Dennis et al., Reference Dennis, Yim, Garrett, Sreekumar and Stone2019). Unforgettable.me (2020) “has in excess of 2,675 registered participants… some of whom have been collecting for over four years. In November 2019 [the platform recorded] over 1.3 million events.” It is currently in use as a research tool and not widely available to consumers. Participants can connect over 600 data streams to their account from: smartphones, wearables, online services, and internet of things devices. These data streams derive from Unforgettable.me, If This Then That (IFTTT, ifttt.com), email servers and the SEMA3 ecological momentary assessment system (https://sema3.com/).
Using keyword search and visualization interfaces, users are able to retrieve specific events, reflect on patterns across their daily activities and reminisce about their experiences. For compensation, they can also participate in projects posted by researchers. Users provide authorization for the use of their data on a project by project basis. This is considered a form of dynamic consent (Kaye et al., Reference Kaye, Whitley, Lund, Morrison, Teare and Melham2015). Projects posted to the system must have ethical clearance from approved research institutions, the clearance documentation is posted with the project. All data provided by users, either through passive data streams or through the completion of surveys or experiment, remain the property of the user. This is in contracts to typical university processes the data becomes the property of the research institution, allowing them to accumulate a personal data asset that they are free to license multiple times.
To protect users’ privacy and preserve the value of their data assets, researchers employ a privacy-preserving language called Private. This language allows analysis to be conducted without revealing raw data thus preventing researchers from seeing or copying it. All results provided by the language are tested to ensure that they are not too sensitive to the data provided by any given individual before they are released to the researcher (Dennis et al., Reference Dennis, Yim, Garrett, Sreekumar and Stone2019).
2.4 Case study 4: PatientsLikeMe
PatientsLikeMe (PLM; www.patientslikeme.com) is an online patient community with the vision of facilitating online interaction between individuals living with chronic health concerns (Frost and Massagli, Reference Frost and Massagli2008; PatientsLikeMe, 2019). PLM was founded in 2004 and remains in high use today. The platform has expanded to include thousands of patients and communities living with many different conditions. The main component of PLM, that participants interact with, is their profile. This is where individuals can input and visualize their health data. PLM collects a variety of data, which is specific for each condition they have communities for. It includes data on disease progression, treatments, side-effects, medication, and physical activity; data is self-entered by the individual (Eichler et al., Reference Eichler, Cochin, Han, Hu, Vaughan, Wicks, Barr and Devenport2016). The data are transformed into visualizations, such as charts to track medications related to side-effects, or progression of disease. Data from all participants with the same condition are aggregated, allowing participants to relate their progression with other participants living with the same condition. Participants can also print out their data to share with their health providers (Wicks et al., Reference Wicks, Massagli, Frost, Brownstein, Okun, Vaughan, Bradley and Heywood2010). Research is an important aspect of the PLM model, much of their funding comes from the selling of de-identified data back to industry (Morgan, Reference Morgan2019). PLM shares de-identified data with a number of groups, including the PLM community, PLM as a company, partners of PLM, and vendors (Patients Like Me, 2018). PLM has become a platform for crowdsourced research and patient-driven research with researchers engaging PLM participants in research trials, using PLM data in research, and using PLM data as the basis of research studies (Bradley et al., Reference Bradley, Braverman, Harrington and Wicks2016).
3 Data Control Dimensions
To explore the different approaches to data control the four cases have taken, the authors draw on three dimensions of data control proposed by Vayena and Blasimme (Reference Vayena and Blasimme2017). These are data use, data access, and governance. It is pertinent to mention at this point, the legal framework the four cases operate under only tell us the requirments for data processing, not how to implement the values of privacy and data protection in specific contexts. Thus, what the four cases have provided are examples of data governance models, which align with regulatory frameworks and expand the idea of data governance within a specific context.
3.1 Data use
Data use relates to both the primary and secondary uses of data, who decides these uses, and how they are legitimized and approved through data governance strategies. Traditionally implied or informed consent models are applied in healthcare in addition to data anonymization, so data can be used for secondary purposes (Vayena and Blasimme, Reference Vayena and Blasimme2017). Regulatory frameworks set the amount of control that an individual has over their data. In the GDPR, for example, consent is the primary mechanism for protection of personal data, and it must be freely given, specific, informed, and unambiguous. In the context of processing of health data, the data subject must have given explicit consent. However, both consent and anonymization approaches have come under scrutiny due to the ease by which data can be de-anonymized (Ohm, Reference Ohm2009), and more recently the plethora of uses of data beyond which individuals initially consent (Mittelstadt and Floridi, Reference Mittelstadt and Floridi2016). As already established, the current Laissez-Fair approach (Mills, 2017) to data governance does not give participants a voice in how their data are used, beyond the notice and consent procedure. This leaves the majority of control with the health system, healthcare providers, and clinicians. In relation to health data collected outside the health system, which usually resides either with the individual or with the companies that collect it, through notice and consent models. In the case studies we have outlined, some platforms increase individual’s control over how data are used, and others decrease it. Unforgettable.me and HRB give greater control to individuals about how their data are used. For example, in unforgettable.me, people can choose exactly what projects can use their data, what types of data are used and retain ownership over their data. On the other hand, platforms such as PLM follow a traditional terms and conditions model, whereby signing-up to the platform individuals give away control over their data, as such it may be sold on to PLM partners. Finally, Open Humans shares control between the individual and the collective. However, individuals can decide how their data are used, the opportunities for data use are decided by the collective of people on the platform.
We have focused on individuals readily sharing their own data, but it is pertinent to consider what legal and ethical obligations parents and carers have in sharing their charge’s health data, for example, the CGM data (mentioned “Case study 1: Open Humans and Nightscout Data Commons”). The term “sharenting” is used to describe parents sharing information about their children on social media, but any sharing of another’s data comes with a moral obligation to act with appropriate discretion (Steinberg, Reference Steinberg2016). If data can be linked to an individual, this leaves a digital footprint, which can have foreseen (health insurance) and perhaps unforeseen implications.
3.2 Data access
Current challenges with both health data held in institutions and health data collected through other platforms, such as social media, is much of it is not easily accessible by the individual (Li et al., Reference Li, Yu, Ren, Lou, Jajodia and Zhou2010). Thus, individuals have little ability outside institutional channels to gain benefit from that data. A potential benefit of the four case examples is that they aim to give individuals greater access to their data. HRB aims to ensure individuals can access and take their data with them. This represents what Vayena and Blassime (Reference Vayena and Blasimme2017) descibe as a shift toward the individual as a distributor of their health data. In Open Humans, individuals and groups are also leading their own research but “traditional” health services and researchers will need to decide how to discern their findings (Vayena et al., Reference Vayena, Dzenowagis, Brownstein and Sheikh2018). An interesting point raised by Tempini and Del Savio (Reference Tempini and Del Savio2019) is the risk of “digital orphans,” individuals provide platforms with access to their data, however, this does not lead to broader accessibility for researchers. PLM data, for example, are accessible by the individual and PLM, but due to their business model, PLM became gatekeepers for others hoping to access that data (Tempini and Del Savio, Reference Tempini and Del Savio2019). A similar example is personal health apps, which though might provide some sharing options for data collected, usually putting constraints on the amount and granularity of data that can be accessed and shared (Kim et al., Reference Kim, Lee and Choe2019).
Finally, does control mean active engagement? There is a risk that people do not, will not, or cannot actively engage with these platforms and it may be no better than leaving control in the hands of clinicians and health services (Steinsbekk et al., Reference Steinsbekk, Myskja and Solberg2013). This phenomenon has been seen in the implementation of patient portals. Engagement with such portals is not guaranteed and may hinge on multiple elements, such as the design of the platform (Fylan et al., Reference Fylan, Caveney, Cartwright and Fylan2018; Dendere et al., Reference Dendere, Slade, Burton-Jones, Sullivan, Staib and Janda2019). Further, there is an implicit assumption, in some systems; people can take control of their health data. This assumption is made regardless of the fact that there are different levels of literacy, health literacy and digital literacy across any community. These factors can and will be a barrier to using such platforms (Bodie and Dutta, Reference Bodie and Dutta2008).
3.3 Governance of system
Data governance is defined by Rosenbaum (Reference Rosenbaum2010, p. 1444) as “the process by which stewardship responsibilities are conceptualized and carried out, that is, the policies and approaches that enable stewardship.” With the increasing potential of health data, data stewards and data governance approaches are under increasing scrutiny to define the appropriate uses of data in these new contexts (Hripcsak et al., Reference Hripcsak, Bloomrosen, FlatelyBrennan, Chute, Cimino, Detmer, Edmunds, Embi, Goldstein, Ed Hammond, Keenan, Labkoff, Murphy, Safran, Speedie, Strasberg, Temple and Wilcox2014). Data governance is also one of the main challenges to the use of big data in healthcare (Kruse et al., Reference Kruse, Goswamy, Raval and Marawi2016). There are a few broad ways these emerging platforms appear to address data governance. Platforms such as PLM require consent to the initial use of the platform, this then allows secondary use of de-identified data without further consent; thus, PLM becomes a data steward. HRBs propose a banking model, control sits with the individual and allows them to make decisions regarding who can withdraw health data from their individual “bank.” This approach shifts the current control away from clinicians who traditionally tend to, with consent, make decisions on what data to share and when a person needs data about them shared (Kariotis et al., Reference Kariotis, Prictor, Chang and Gray2019). More broadly, what is seen in models, such as Open Humans is a move away from individuals just providing researchers access to their data, they are also able to define the projects that their data might contribute toward.
4 Discussion
4.1 What does the future look like?
In addition to the case studies, there are a number of emerging data governance approaches. These raise new considerations for health data; three examples which require further exploration are data trusts, data marketplaces, and data collectives.
4.1.1 Data trusts
Data trusts defined by the (Open Data Institute (2018, p. 1) are “a repeatable framework of terms and mechanisms that is mandated for use (or subject to scrutiny, or certification) in particular contexts to provide oversight of data access.” Like any legal trust, it is envisioned that a trustee would be legally obliged to make decisions on data access and use in the best interest of stakeholders in a specific context (Open Data Institute, 2019). Data trusts have burst into popular policy discussion in recent years in part due to Alphabet (Google’s parent company) proposing a data trust as a solution to concerns over data collected as part of a smart city project in Toronto (McDonald, Reference McDonald2019). There are examples, such as NightScout (outlined in “Case study 1: Open Humans and Nightscout Data Commons”), that act as a Data Trust as per the Open Data Institute (ODI) definition, even though it may not be labelled as such.
4.1.2 Data marketplaces
Data marketplaces put sole focus on the monetization of personal data. Examples include DataCoup (www.datacoup.com/) and CitizenMe (www.citizenme.com/). Both examples currently include the collection of health data, mainly collected from health devices, and personal wearable and readable devices. Data marketplaces, though they give control over data access to individuals, they do not, from the examples available, promote voice to individuals over how data should be used or serve any specific public good (Elvy, Reference Elvy2017). It is also questionable, who has the power to decide how valuable data are? and whether individuals will be able to negotiate this price. Individual data points may not seem valuable until they are linked together with other data points, many individuals may be unaware of this fact (Cooper and LaSalle, Reference Cooper and LaSalle2016).
4.1.3 Data cooperatives
Data cooperatives are membership organizations, individuals pay a fee to join and then can store their data in the cooperative, while also becoming part-owner and thus having a say in how the data are used (Hafen et al., Reference Hafen, Kossmann and Brand2014). One example of data cooperatives is MiDATA (https://www.midata.coop/), established in Switzerland. Individual account holders have logged access to their data, data cooperatives give complete decision-making control back to the individual whose data are stored, as well as how the revenues generated by sharing their data used (Hafe, Reference Hafen, Krutzinna and Floridi2019). Global and clinical research could be facilitated through secure national cooperatives (Hafen et al., Reference Hafen, Kossmann and Brand2014).
4.2 Shift from institutional to individual to collective and public good
It appears that there is a shift toward collective models of governance, as in Open Humans. Collective autonomy is an evolution of previous waves of data governance, which prioritizes individual autonomy (Evans, Reference Evans2017). This reflects the evidence available on public attitudes for the use of health data beyond care; there is support for its use to further the common good (Skovgaard et al., Reference Skovgaard, Wadmann and Hoeyer2019). Evans (Reference Evans2017, p. 1) suggests, we may be seeing “a popular uprising of regular people seeking a meaningful voice in establishing citizen-led ethical and privacy standards to advance big-data science while addressing the concerns people feel about the privacy of their health data.” This shift toward collective control poses questions for privacy policies, which tend to focus on the individual right to privacy, rather than viewing privacy and control over data as a social good (Kasper, Reference Kasper2007).
The different approaches to data governance outlined in the case studies above are situated in a broader societal narrative around privacy of health data. There a number of recent examples of government supported electronic health records (e.g., My Health Record in Australia) or secondary data platforms (e.g., the proposed care.data in the Unite Kingdom), which have raised community concern of privacy and expectations of government compliance. Carter et al. (Reference Carter, Laurie and Dixon-Woods2015) consider that care.data experienced ongoing “challenge and contestation” resulting in its failed implementation because it did not meet societal expectations in relation to its proposed activities. Privacy of medical data is an important issue due to the sensitive nature of some information and the risks posed if the data “leaks” into other areas of life (e.g., insurance or the workplace). The risks of breaking confidentiality of medical data is especially great for some health conditions, which remain highly stigmatized (e.g., mental health conditions). However, at the same time, many service users and clinicians face barriers to receiving and providing good care because of the lack of information sharing.
The platforms considered have outlined three broad waves of data governance, that of control sitting with institutions, control sitting with individuals, and control being shared between individuals and the collective (Figure 1). The changes these data governance strategies pose go beyond the control of data and present challenges to the “traditional” health system.
4.3 How will this change healthcare?
The platforms described sit in contrast to the current workings of a traditional health system, where data within the system remains siloed, and data outside the health system lack legitimacy. These platforms can provide individuals with access to new data and new knowledge. This infers individuals are now entering a new era for health services. Clinicians and healthcare providers can be provided with more data than they may expect. The health system may not be prepared to use data collected from outside the health system, both in terms of individuals knowing how to bring their data into the health system and clinicians knowing what to do with that data (Chung et al., Reference Chung, Dew, Cole, Zia, Fogarty, Kientz and Munson2016; Zhu et al., Reference Zhu, Colgan, Reddy and Choe2017). There is also a risk that the shift toward data-driven healthcare facilitated by these platforms, may not be to the benefit of greater person-centred care which relies on the therapeutic relationship (Greenhalgh et al., Reference Greenhalgh, Snow, Ryan, Rees and Salisbury2015). Further, the OpenAPS glucose monitoring system (outlined in “Case study 1: Open Humans And Nightscout Data Commons”) raises the issue of regulation in a health system where individuals and groups are sharing and learning new skills to manage their health. The rise of “do-it-yourself” systems, such as OpenAPS pose questions as to how will health services and health regulators manage this crowdsourced data and knowledge? Which, quite fairly, individuals may view as meeting their needs more than other options in the health system (Farrington, Reference Farrington2017).
This paper has explored the way the different platforms and how their associated data governance strategies may change healthcare in relation to information norms. Nissenbaum (Reference Nissenbaum2009) in her book “Privacy in Context” has pioneered an alternative way of considering why the implementation of new technologies into certain contexts leads to privacy concerns. Historically, privacy has been considered the individuals ability to limit the collection of personal information about themselves. Nissenbaum (Reference Nissenbaum2019) argues that privacy is the appropriate flow of information, what is appropriate is defined by contextual information norms. Nissenbaum (Reference Nissenbaum2019) outlines, in any context there are specific information norms relating to: the different actors (sender, receiver, subject); the type of information (e.g., health information or administrative information); and transmission principles (e.g., consent is required, by law, online/offline). New technologies, which breach these norms can lead to privacy backlashes and/or workarounds by individuals attempting to realign to their expected norms. Although a technology, such as the proposed Care.data may align with legal requirements, it may not necessarily align with expected contextual norms (Carter et al., Reference Carter, Laurie and Dixon-Woods2015). This identified a key weakness of the current Laissez-Faire model of data governance described by Mills (2017), it does not have any mechanisms built-in to grapple with contextual norms.
5 Conclusions
Health data are powerful resource that have the potential to improve healthcare, empower individuals, and advance health research. However, traditional control and access to health data have been limited, held in siloes by institutions. An emerging array of health data platforms are changing this siloed situation by opening up data to individual and collective uses. These new models pose new data governance opportunities and challenges. The four case studies explored represent a shift from institutional to individual control. However, there is also a current shift toward more collective approaches to data governance, such as data trusts and data collectives. These new emerging data platforms raise broader concerns regarding the management of privacy and integration into the traditional health system. A particular obstacle is how the knowledge accessed through the platforms, that is not part of the “traditional” health system, may create challenges for clinicians and health services in the translational use of this information.
Funding Statement.
No specific funding to declare.
Competing Interests.
Simon Dennis is CEO of Unforgettable Research Services Pty Ltd. Mad Price Ball is Executive Director and President of Open Humans Foundation and co-founder of Open Humans. Bastian Greshake Tzovaras is Director of Research at Open Humans. Timothy Kariotis, Helen Almond, Ann Borda, Tony Sahama, and Carolyn Johnson declare none.
Author Contributions.
Conceptualization, T.K.; Investigation, T.K, M.P.B, B.G.T., S.D., T.S., and A.B.; Writing-original draft, T.K., M.P.B., B.G.T., S.D., and T.S.; Writing-review & editing, T.K., M.P.B., B.G.T., H.A., A.B., and C.J.
Comments
No Comments have been published for this article.