Abstract
The emergence of a global industry of digital health platforms operated by Big Tech corporations, and its growing entanglements with academic and pharmaceutical research networks, raise pressing questions on the capacity of current data governance models, regulatory and legal frameworks to safeguard the sustainability of the health research ecosystem. In this article, we direct our attention toward the challenges faced by the European General Data Protection Regulation in regulating the potentially disruptive engagement of Big Tech platforms in health research. The General Data Protection Regulation upholds a rather flexible regime for scientific research through a number of derogations to otherwise stricter data protection requirements, while providing a very broad interpretation of the notion of “scientific research”. Precisely the breadth of these exemptions combined with the ample scope of this notion could provide unintended leeway to the health data processing activities of Big Tech platforms, which have not been immune from carrying out privacy-infringing and socially disruptive practices in the health domain. We thus discuss further finer-grained demarcations to be traced within the broadly construed notion of scientific research, geared to implementing use-based data governance frameworks that distinguish health research activities that should benefit from a facilitated data protection regime from those that should not. We conclude that a “re-purposing” of big data governance approaches in health research is needed if European nations are to promote research activities within a framework of high safeguards for both individual citizens and society.
Introduction
In November 2019, the Wall Street Journal broke the news that the second-largest healthcare provider in the US, run by the non-profit organization Ascension, had entered a deal with Google that could eventually see detailed health information on 50 million American patients moving into the company’s cloud-computing system (Copeland and Needleman, 2019). The massive data transfer is part of an agreement—code name: “Project Nightingale”—whose stated aim has been to employ Google Cloud’s machine learning tools for developing new diagnostic tests and technologies for exchange of patient health information (Cohen, 2019). At the outbreak of the news, the two parties involved were swift in publicly claiming no legal wrongdoing and allay fears over repurposing of patient data, such as for targeted advertising (Conrado, 2019; Shaukat, 2019). Nonetheless, the news of Ascension’s sharing with Google of identifiable information on millions of patients, without notification to either patients or doctors, sparked an intense public outcry, in the US and also beyond (Copeland and Needleman, 2019; Prainsack, 2020).
Whilst much scrutinized, Google’s deal with Ascension represents but one of the latest examples of an increasing trend that, over the space of a decade, has seen the health sector being flooded by a growing stream of digital technology corporations—Google, but also Amazon, Microsoft, IBM, and Apple, to name but the largest—jockeying for position in this rapidly expanding market (Sharon, 2016). Google itself, in response to the scandal, has been plain in acknowledging that similar agreements have been established “with dozens of other healthcare providers” (Shaukat, 2019). In fact, the company is said to have gained access to tens of millions of patient health records in more than three-quarters of US states, consisting in many instances of personally identifiable health information (Copeland et al., 2020). And Google is not alone in this endeavor. All its competitors have moved equally aggressively to claim their stakes in the ever-expanding field of digital health.
Not surprisingly, the emergence of a global industry of digital health platforms1 operated by Big Tech corporations has raised pressing questions on the capacity of current data governance models, regulatory and legal frameworks to safeguard the sustainability of the digital health ecosystem, through the promotion of mutually beneficial interactions between digital health platforms and society (Taylor and Purtova, 2019; Van Dijck et al., 2018).
In this article, we direct our attention toward the data governance regime unfolding in the European Union (EU) around Regulation (EU) 2016/679, the General Data Protection Regulation (GDPR), which became applicable since 25 May 2018. Establishing a comprehensive and harmonized data protection framework across the EU, the GDPR has been lauded in many quarters as the cornerstone of Europe’s value-friendly “third way” to navigate the digital world, in between the “techno-libertarianism” of US Silicon Valley and the “digital authoritarianism” of China (Thornhill, 2018). Yet, as we reviewed in recent work (Marelli et al., 2020), the GDPR faces mounting challenges in regulating the entanglements of digital health technologies and related big data practices that, as means for “intensified data sourcing” (Hoeyer, 2016: 74), are increasingly pervading our digitized European societies.

A stylized representation of how the research exemption could apply in practice.
Within this scenario, this article specifically focuses on the GDPR’s provisions that directly target health research. Notwithstanding its “omnibus” rather than sectoral approach to data protection, intended to cover a wide scope of processing areas, the GDPR devotes a specific set of provisions to regulate the processing of personal data for scientific research activities. Notably, the GDPR ushers in a rather permissive and flexible regime for scientific research through a number of derogations to otherwise stricter data protection requirements, while providing a very broad interpretation of the notion of “scientific research”. In fact, we contend, the breadth of these derogations combined with the ample scope of this notion can provide unintended leeway to the health data processing activities of Big Tech platforms, which have not been immune from carrying out privacy-infringing and socially disruptive practices in the health domain. Accordingly, it is the contention of this article, further finer-grained demarcations need to be traced in order to distinguish research activities that can legitimately avail themselves of data protection facilitations from those that mandate for stricter data protection standards.
The article is structured as follows. First, we reconstruct the policy trajectory that inscribed facilitations for scientific research into the GDPR. Next, we outline the bundle of provisions that make up the so-called “research exemption” and critically discuss the notion of scientific research in the GDPR. Then, we review the challenges and potential disruptions elicited by the influence of Big Tech platforms in the health domain, partly owing to the data protection facilitations originally intended to streamline and de-bureaucratize established and well-regulated research practices. Finally, we provide an overview of and discuss proposals for deploying use-based data governance frameworks to demarcate research activities to be carried out within a facilitated data protection regime from those to subject to more stringent data protection scrutiny. This discussion is intended as the basis for the further development of more detailed policy guidance—such as the one presently in elaboration at both national and European level (European Data Protection Board (EDPB), 2021).
Making the GDPR into law: Policy struggles toward a special derogatory regime for scientific research
The approval of the GDPR has been characterized by a lengthy and contentious policy process. “Trilogue” negotiations amongst the European Commission, the European Parliament, and the Council of Europe2 were shaped by intense lobbying efforts, mostly by digital corporations, and brought to the surface underlying lines of tension among EU Member States diverging in economic interests and cultural understandings on privacy and data protection (Mager, 2017). Throughout the process, almost 4000 amendments had been tabled, making the GDPR the most lobbied legislation in EU history.
A number of actors and organizations associated with (data-intensive) health research (from the Wellcome Trust to Science Europe) were prominently involved in the negotiating phases, as they set out to shift the policy discourse from data- to health-protection through sustained generation and application of scientific knowledge (Starkbaum and Felt, 2019). In so doing, they took aim at drafts of the legislation, most notably the one approved in March 2014 by the Committee on Civil Liberties, Justice and Home Affairs of the European Parliament, deemed to adopt a too restrictive approach to regulating research (Casali, 2014; Ploem et al., 2013). Overall, the aim of these converging lobbying efforts from the part of individual researchers and research agencies was to ensure “that the needs of scientific research with respect to accessibility and processing of personal data are considered, that the provisions and derogations that facilitate scientific research are maintained, and that amendments which would dramatically weaken these provisions are rejected” (Science Europe, 2013: 2).
A settlement in Trilogue negotiations was reached in December 2015, and, notwithstanding some critical remarks (e.g. LERU, 2016), the definitive text of the GDPR was met with a general sense of escaped danger from the part of the European health research community (Abbott, 2015; Dove et al., 2016). For, the outcome of these negotiations resulted in the GDPR arguably providing a significant degree of flexibility in the use of personal data for scientific research purposes, stemming from the implicit recognition of the
The research exemption in the GDPR
The “special derogatory regime” (EDPB, 2021) for scientific research in the GDPR—also known as the “research exemption”—contains four distinct sets of provisions. The first, set out in Art. 9(2)(j), provides an alternative to the explicit consent requirement for personal data processing outlined in Art. 9(2)(a). Pursuant to Art. 9(2)(j), scientific research is identified as one of the processing activities for which the general prohibition otherwise imposed in Art. 9(1) on the processing of “special categories of personal data” (i.e. sensitive data, such as health-related and genetic data) is withdrawn; what is more, scientific research is identified as one of the lawful grounds for processing sensitive data, even in absence of consent from the data subject (Art. 9(2)(j) read in conjunction with Art. 6(1)(f)).3 To ensure the lawfulness of the processing, scientific research activities should meet two additional conditions. First, they should be subject to appropriate organizational and technical safeguards for data subjects, as set out in Art. 89(1), in particular to ensure the principle of data minimization (which limits the use of data to what is necessary to achieve the stated purpose of the processing). In addition, the research should be undertaken on the basis of a specific national or EU law providing a relevant legal basis. Such law should be proportionate to the aim pursued, respect the essence of the right to data protection, and must be interpreted in consideration of the jurisprudence of the European Court of Justice (ECJ) (EDPB, 2020).
Second, meeting the demands of research organizations, the Regulation introduces much called-for provisions geared to facilitate the re-use of data in secondary research. The GDPR provides that secondary processing of previously collected data for scientific research purposes is “not incompatible” with the initial purpose for which data has been gathered (Art. 5(1)(b)), thus making possible to re-use data without consent from the data subject. This “presumption of compatibility” (EDPS, 2020: 22)—however—does not provide a blanket authorization to secondary research. Rather, as recently observed by the “
Additionally, in those cases in which consent is used as a legal basis, the GDPR allows data controllers (the entities legally in control of data processing) to avail themselves of “broad consent”, whenever this is required by the intended research purposes, and especially when it is not possible to fully identify the purposes of personal data processing at the time of data collection, as eminently in the case of biobanking (recital 33; cf. also EDPB, 2021; Hallinan, 2020; Marelli and Testa, 2018; Shabani et al., 2021). In such instances, data controllers are required to provide adequate safeguards to data subjects and describe the purposes of data processing in “a high-level way, for instance in terms of (types of) research questions and/or fields of research to be explored” (EDPB, 2021: 7).
The GDPR also relaxes limitations in the amount of time that data can be stored (Art. 5(1)(e)), and recognizes the importance to take into account “the legitimate expectations of society for an increase of knowledge” (Recital 113) where assessing whether the legitimate interest of the data controller can be invoked for (limited) cross-border data transfers outside of the EU.4
Third, the GDPR provides for a number of derogations to data subjects’ rights where the exercise of such rights is likely to render impossible or seriously impair the accomplishment of research objectives.5 A first subset of derogations to data subjects’ rights is directly inscribed in the GDPR, which withholds the exercise of the “right to be forgotten” (which allows data subjects to have their data erased), and the right to receive information from data controllers which have not themselves collected the data, as it could be the case with data retrieved from patients’ records (Art. 17(3)(d) and Art. 14(5)(b)). Pursuant to Art. 89(2), a second subset of derogations related to scientific research undertakings—namely to the rights of access (Art. 15), rectification (Art. 16), restriction of processing (Art. 18), and objection (Art. 21)—can instead be introduced by Member States in their national legislations. However, as noted by the EDPB, these provisions should be interpreted in the light of the jurisprudence of the ECJ, meaning that all restrictions of the rights of data subjects must apply only insofar as strictly necessary (EDPB, 2020: 11).
Finally, in a move that has been regarded as potentially leading to the further fragmentation of the European health research landscape (LERU, 2016), Member States are also endowed with the prerogative to introduce further provisions for the processing of genetic, health, and biometric data (Art. 9(4)). Whilst these could also include limitations to research, countries such as Italy have been prompt in passing legislation that can been seen as highly favorable for health research, for instance by providing that sensitive data collected for clinical purposes by Italian research hospitals—the 51
The notion of “scientific research” in the GDPR
In defining the scope of the special derogatory regime for scientific research, the legislator has apparently confronted a contemporary uptake of a foundational epistemic problem—what should qualify as genuine science? In that regard, the GDPR advances a remarkably broad interpretation of activities falling within the remit of what should count as scientific research proper. As stated in Recital 159:
At any rate, such a broad construe extends the scope and reach of the special derogatory regime for research. Also depending on additional provisions concerning scientific research enacted at the national level (for a comprehensive review, see European Commission, 2021), the bundle of facilitations traced above stands to apply not only to public and private academic or clinical research centers, but also to actors such as pharmaceutical, medical device, and—as per the focus of this article—digital technologies companies undertaking health research activities.
How broadly should the GDPR’s notion of scientific research be interpreted?
In the view of the EU legislator, public officials, and data protection bodies, “the definition of scientific research must not be stretched beyond its intended limits” (Buttarelli, 2018: 2), and processing operations linked to scientific research must “respect the essence of the right to data protection” (Art. 9(2)(j) GDPR). Likewise, as claimed by the EDPS, the EU’s independent data protection supervisory authority, “performing an activity deemed to be research cannot be a
And yet the above points of caution appear, in a number of respects, of difficult implementation. For one thing, it is questionable whether—and especially
It is also questionable whether current ethical standards and modes of ethical oversight provide sufficient bite and are adequately tailored for governing current developments in health research. For instance, as preliminary data shows,6 research ethics committees—which have long since played a central role for upholding ethical standards in research—have to confront the lack of well-established guidelines and regulatory tools for providing adequate ethics assessment for non-traditional types of data-intensive research (e.g. AI, social companion robots). In parallel, inasmuch as they have been traditionally appointed with the aim to review clinical research on human subjects, they are in many cases devoid of the epistemic as well as ethical expertise required to master recently established research fields and deal with the ethical quandaries raised by new data-intensive technologies and research practices. This, as a consequence, can lower standards of ethical oversight in health research, to the potential detriment of research participants (Staunton et al., 2019).
Likewise, the reference to “sector-related methodological standards” seems a rather ineffective if not counter-productive move to avoid over-stretching the limits of what should count as scientific research. In contemporary health research, which increasingly aligns heterogeneous technologies and expertise, what is regarded as sound “methodological standards” is the result of ongoing and open-ended processes of coordination and convention-making (Cambrosio et al., 2006; Keating and Cambrosio, 2003). As the latter involve a multiplicity of actors spanning distinct sectors, the very notion of “scientific sector” appears something increasingly challenged and in flux. Moreover, in the case of digital health, a sizeable portion of actors that play a central role in standard-setting (e.g. digital technology companies) are primarily involved in activities
As to the remarks about the overall primary orientation of research toward “collective knowledge and wellbeing” (rather than private interests), the difficulty here is twofold. First, consistently with what we have just observed, scientific research in the health domain has increasingly come to be carried out in a “transdisciplinary” manner in so-called “contexts of application”, where it is difficult to neatly discriminate between undertakings that primarily pursue collective knowledge and wellbeing as opposed to private interests (Carrier and Nordmann, 2011; Nowothny et al., 2001; Verschraegen, 2018). Second, from a technical legal point, it should be observed that the notion of “collective knowledge and wellbeing” is arguably subsumed in the GDPR under the notion of “public interest”. The scope of this notion, which can apply for instance to epidemiological studies (Article 29 Working Party, 2014), has to be defined in specific and precise enough terms in European and national legislation (Art. 6(3)(a)/(b) GDPR, Article 29 Working Party, 2014), and thus maintains a narrower connotation with respect to what, as a society, we would normally be prepared to identify, in lay terms, as the “general interest” or the “public good”. From the same legal perspective, it should be noted that “public interest” represents but
As a consequence of these two points, in absence of a better definition of what the pursuit of collective knowledge and wellbeing should amount to, the observation that research should be geared to some forms of public good appears of difficult actionability.
As it stands, and notwithstanding claims to the contrary by EU officials and data protection bodies, the notion of scientific research advanced in the GDPR appears exceedingly broad-construed. Consequently, questions remain as to whether the ad hoc regime for scientific research maintains an excessive reach, in ways perhaps unintended by the legislator.
The relaxation of data protection requirements was foreseen to streamline and de-bureaucratize research practices traditionally subject to high ethical and regulatory standards, in view of the potential societal benefits accruing from (health) research. Still, depending on how Member States modulate the research exemption in their own legislation (see European Commission, 2021), and considering especially fields of research involving the processing of highly sensitive and identifiable data such as genomics (Shabani and Marelli, 2019), it can be argued that the GDPR risks reaching a paradoxical outcome. While wanting to carve out a special status for scientific research (thereby implicitly stating that not every data processing activity can fall under its rubric and that the sheer pervasiveness of the digital has not yet eroded its contours), it has broadened its definition so much that it may end up enabling, in practice, the erosion of the very status that it strived to protect. This may possibly lead to a major regulatory leeway in favor of data controllers over data subjects (Pormeister, 2017), reducing the accountability of research organizations, while depriving research participants of decisional autonomy and adequate safeguards.
Whilst the argument about the GDPR providing excessive regulatory leeway could be plausibly applied to the processing activities of different types of data controllers, including pharmaceutical corporations (Marelli and Testa, 2018), the reach of such leeway raises concerns especially with regard to the research undertakings of a specific type of organizations, namely Big Tech platforms. The latter have moved decisively to set foothold in the health domain—even through the attempted provision of healthcare services (e.g. Wingfield et al., 2018)—and have emerged as key enablers of data-intensive health research (Sharon, 2016, 2018). In so doing, however, they have proved to raise distinctive regulatory, ethical and socio-political challenges—which we review in the next section—which makes it highly questionable whether they should be allowed to benefit from the full spectrum of data protection facilitations afforded to scientific research undertakings.
Big Tech platforms in health research
At the present day, pointing to the growing dominance of Big Tech corporations within an increasingly digitized health sector has arguably become a sort of truism. From research funding (e.g. Chan Zuckerberg Initiative, 2020) to the provision of research infrastructures (e.g. Broad Institute, 2015), from direct involvement in big data research (e.g. Verily, 2020) to the provision of technologies for recruitment of research participants, Big Tech has become a force to be reckoned with in health research.
The drawbacks stemming from the apparently inexorable rise of these powerful corporate actors in health research become visible at different levels. For one thing, Big Tech has long since being associated with dubious privacy-infringing practices. Aside from sanctioned data breaches and public scandals that have surfaced over the years, such as those involving Alphabet-controlled DeepMind in the UK (Powles and Hodson, 2017) and IBM Watson in Italy (Barbacetto, 2017), these corporations are known to routinely engage in ethically suboptimal yet legal practices predicated on large-scale data harvesting and the capacity to combine health, lifestyle, and mundane data types (e.g. search data) in myriad alarming ways (Fussell, 2019). These, in turn, can have a number of detrimental consequences for individual people as well as societies, ranging from breach of individual and group privacy (Mittelstadt, 2017) to social sorting (Hogle, 2016), from profiling (Hildebrandt and Gutwirth, 2008) to algorithmic discrimination (O’Neil, 2017), from increased inequality and social exclusion (Eubanks, 2018) to commercial “dataveillance” (Pasquale, 2015; Zuboff, 2019).
These “quasi legal” practices, and the ethical and social disruptions they entail, can be often traced to the existence of legal gray areas in data protection regimes, which makes it difficult to ascertain and/or clearly demonstrate breaches (McMahon et al., 2019). More in general, regulatory regimes, including the GDPR in the EU, reveal shortcomings in adequately governing digital health technologies and, notably, the data processing activities of digital health platforms, due to the challenges faced by data protection standards dating back to a pre-digital world (Marelli et al., 2020; Price and Cohen, 2019; Wachter, 2019).
Yet, while privacy and regulatory concerns have come to the fore in public debates, equally pressing are the broader societal issues raised by the very core architecture and business models of these organizations, which are able to achieve dominant positions in the markets in which they operate (Srnicek, 2017; Van Dijck et al., 2018). In particular, as they become integral part of leading research undertakings across the globe, Big Tech corporations are able to convert the advantage acquired in the digital domain—through their technical expertise in terms of data collection, data analytics, and infrastructure development—into epistemic and governance advantages in the domain of health and medicine (Sharon, 2020; Shaw, 2020). As they increasingly engage established actors (academia, pharma) in research activities, these corporations not only manage to bring precious epistemic expertise in-house, thus emerging as prominent actors in research; also, they render their own digital expertise an indispensable “entry ticket” to new data-intensive fields of health research (Sharon, 2020). Moreover, by investing in the construction of repositories of public health, fitness, genomics, and health record data, they can easily become dominant data brokers controlling and establishing the rule of access to large-scale databases (Shaw, 2020). In turn, this gatekeeping function in health research translates into a governance advantage, as it grants these actors the power to reshape this domain according to their values and interests, while being able to define the future directions that the sector will take (Sharon, 2020). The agenda-setting prerogatives of Big Tech platforms are further reinforced by their capacity to target desired research areas for funding (Prainsack, 2020), establish close ties with key opinion leaders in the field (Shaw, 2020), and provide support to major science policy initiatives (e.g. Robbins, 2020)—thus exerting significant influence on standard- and policy-making processes. In sum, as Big Tech platforms strive in health research, they are increasingly poised to become obligatory passage points in the digitized biomedical landscape of the coming decades.
For all of the above, it appears questionable whether these actors should be made to benefit from the full spectrum of facilitations that the current data governance regime in Europe, centered around the GDPR, but involving also a relevant role by individual Member States (Donnelly and McDonagh, 2019; European Commission, 2021), renders available for scientific research. When derogations to stringent data protection requirements for data retention, secondary research, or information to data subjects are availed by Big Tech corporations—for instance, as it is routinely the case, within collaborative research networks—they could lead to a heightened risk of infringements to the rights and interests of individual people or vulnerable social groups. At the same time, these dynamics can have a lasting impact on the very social contract that underpins science–society relations, as well as reconfiguring the values in which European health systems are rooted.
Toward a use-based and functionalist framework for scientific research
In this section, we review proposals and sketch out a possible path to unpack the GDPR’s broadly construed notion of “scientific research” and functionally demarcate types of scientific research activities and data uses that can legitimately aspire to the regulatory leeway afforded by the GDPR and national legislation, and those instead to subject to higher data protection requirements. We do not intend to provide a full-fledged “governance fix” to address the issues identified above, yet we aim to outline points to consider that can be then further translated into more defined policy guidance for data-intensive health research undertakings (involving Big Tech platforms).
Moving from concerns similar to the ones articulated here, proposals have been made to steer the data governance approach toward
First, a use-based approach could enable to shift the focus of the data protection machinery from the informational self-determination of data subjects9 toward the ex ante identification of desirable and undesirable personal data
The adoption of a use-based framework, centered on the concrete assessment of data uses, would also be conducive to the elaboration of a more fine-grained definition of scientific research and, relatedly, to arrive to a functionalist demarcation between research activities that should be made amenable to data protection incentives and those that should not.
Moving largely in this vein, Floridi and colleagues (2019) have proposed the notion of “
The proposal by Floridi and colleagues is valuable in a number of respects. First, it explicitly recognizes the need for inscribing further specifications in the category of “scientific research”. In addition, by maintaining a
Yet, such a definition faces challenges in applying to data-intensive research activities involving Big Tech platforms. First, the framing of its goal as the “discovery of new knowledge” seems to corroborate, albeit even inadvertently, enduring distinctions between discovery and application, basic and translational that have increasingly lost sharpness in contemporary research practice, most certainly in data-intensive biomedicine. Additionally, the much welcome functionalist turn in “certifying” the
Fourth, such a broad-construed framing of “general interest” fails to explicitly capture the multiple “normative repertoires” (from accelerating scientific innovation, to improving public health, to enhancing wealth creation, etc.) that drive, and provide social legitimacy and ethical justification to, contemporary data-intensive health research undertakings (Sharon, 2018). In turn, disregarding these different and in fact competing articulations of the public good can lead to overlook the fact that different actors involved in health research, as e.g. research participants or data custodians, may each hold different plausible conceptions of what constitutes “socially acceptable” types of research and data uses.11 The problem is thus that if
Accordingly, the implementation of a narrower yet sufficiently actionable notion of research undertakings and data uses eligible for data protection facilitations seems to require the consideration of the following additional elements.
In the first place, use-based frameworks should incorporate, and ensure the application of, strong mechanisms for preventing violations of “contextual integrity”12 (Nissenbaum, 2004) throughout the full life cycle of personal data. This entails not only avoiding that health data is repurposed
In practice, implementing such mechanisms would require data custodians and data governance bodies (see below) to elaborate, on a context-sensitive basis, well-defined criteria for permitting data re-purposing while preserving contextual integrity, to be enforced through binding data transfer agreements (DTAs). In particular, DTAs should contain clauses permitting re-use of data within the facilitated data protection regime for types of research activities that pursue (context-related) forms of common good and do not involve increased risks for either research participants, larger social groups, or society as a whole.
An additional requirement for effective implementation of a finer-grained definition of scientific research undertakings amenable to data protection facilitations within a use-based model is the establishment of a new layer of regulatory bodies in the guise of
First, to enforce and carry out
Second, such bodies would be tasked with implementing effective mechanisms of
Whilst unlikely to exhaust all issues triggered by the involvement of Big Tech in health research, these policy measures could help tame some of their most disruptive impacts in personal data processing while chaperoning an orderly entry into the health research ecosystem.
Conclusions
Since its entry into force in 2018, the GDPR has had an increasingly structuring effect on health research, in Europe and beyond. In this contribution, we analyzed the GDPR’s special derogatory regime for research in relation to Big Tech platforms inasmuch as, in light of the governance and social challenges they raise, these entities can be seen as representing the ultimate litmus test for a regulation designed with the explicit intent of achieving “the right balance” between a high level of protection for the fundamental right to data protection and the enhancement of digital innovation in the continent (Albrecht, 2016).
As it is by now evident, Big Tech platforms pose specific governance challenges and raise distinctive socio-political concerns in the health domain, revolving around monopoly/market-dominance, enclosure, and privatization dynamics. For sure, these dynamics do not arise out of, and do not owe exclusively to, the data protection derogations foreseen for scientific research. Nor amending and tightening the latter will automatically tame the disruptive impact of such platforms. Still, the special derogatory regime for research is poised to amplify—rather than contain—these concerns, and can thus risk providing a “regulatory backdoor” that exacerbates some of the challenges faced by the GDPR in regulating digital health undertakings.
Moving from these premises, this contribution has built on a growing discussion in data protection and science policy debates to outline points to consider for the implementation of use-based data governance frameworks that tailor the scope and reach of the facilitations mechanisms for research to the ethical and societal desirability that different types of health research undertakings entail. Such a use-based approach is meant to move upstream the assessment of potential data protection criticalities, while also accounting for the broader societal implications of data-intensive health research undertakings. Specifically, it unpacks the abstract and broadly construed notion of “scientific research” advanced in the GDPR in order to enable the tracing of new taxonomies and demarcations within this fuzzy category. At the same time, we have stressed the importance of foreseeing mechanisms for preserving contextual integrity and providing benefit-sharing, so as to let the promissory research involving Big Tech platforms unleash its full potential while remaining tightly accountable to the multiple societal expectations associated with it.
In conclusion, all of the above may require a
Footnotes
Acknowledgements
The authors would like to thank the Editors and three anonymous reviewers for their insightful and constructive comments that helped strengthening the argument of the article. We would also like to acknowledge organizers and participants to the “
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
This work was supported by the European Union’s Horizon 2020 research and innovation program, under the Marie Sklodowska-Curie grant agreement number 753531 (LM).
