Abstract
The global Covid-19 pandemic has resulted in social and economic disruption unprecedented in the modern era. Many countries have introduced severe measures to contain the virus, including travel restrictions, public event bans, non-essential business closures and remote work policies. While digital technologies help governments and organizations to enforce protection measures, such as contact tracing, their rushed deployment and adoption also raises profound concerns about surveillance, privacy and data protection. This article presents two critical cases on digital surveillance technologies implemented during the Covid-19 pandemic and delineates the privacy implications thereof. We explain the contextual nature of privacy trade-offs during a pandemic and explore how regulatory and technical responses are needed to protect privacy in such circumstances. By providing a multi-disciplinary conversation on the value of privacy and data protection during a global pandemic, this article reflects on the implications digital solutions have for the future and raises the question of whether there is a way to have expedited privacy assessments that could anticipate and help mitigate adverse privacy implications these may have on society.
Keywords
Introduction
The Covid-19 pandemic has resulted in unprecedented social and economic disruption worldwide. Racing against an invisible enemy, a virus simultaneously foreign and ubiquitous, most governments have introduced drastic measures to curb its spread and flatten the curve. Time is of the essence during a pandemic (French et al., 2018; Milne and Costa, 2020), providing both governments and organizations with the moral license to take extraordinary measures, such as travel restrictions, bans of public gatherings, closures of non-essential businesses and transitions to remote work and education. A particularly striking characteristic of the Covid-19 pandemic has been the rapid deployment of novel digital technologies, such as contact-tracing apps (Ferretti et al., 2020). Society has also witnessed intensification in the use of pre-existing digital products, such as video-conferencing software. Receiving only limited exposure pre-pandemic, Zoom has become a household name and an essential component for parties (Matyszczyk, 2020), weddings (Pajer, 2020), school and work.
Governments, private companies and public organizations have traditionally all been involved in shaping norms for digital surveillance in Europe. Yet, during the Covid-19 pandemic, these digital technologies have been deployed ad-hoc without proper impact assessment, stakeholder consultation or evaluation (Frith and Saker, 2020; Madianou, 2020; Milan and Treré, 2020). Since many of these digital technologies are data-intensive, generating vast amounts of intimate biometric and location data (Taylor, 2020), concerns about their impact on privacy and data protection have been raised from civil rights advocates, policymakers and the news media (Dubov and Shoptaw, 2020; French and Monahan, 2020; Kitchin, 2020; Maalsen and Dowling, 2020; Morley et al., 2020).
Concerns centre around how the pandemic could be exploited as an opportunity to normalize governmental surveillance (French and Monahan, 2020), particularly into the domestic and biopolitical sphere (Algorithmwatch, 2020; Maalsen and Dowling, 2020; Vallee, 2020). There is a particular concern about ‘surveillance creep’ (Andrejevic and Gates, 2014; Cheung, 2020) and that rushed decisions may stay in place in the long run, as ‘temporary measures have a nasty habit of outlasting emergencies’ (Harari, 2020). Structural inequalities in data access and exposure also mean that there may be disproportionate privacy invasion on certain, disadvantaged, societal groups (Taylor, 2020). Such concerns are justified, since states of emergency often result in increased surveillance and a bypassing of individual rights (Bigo, 2006; Lyon, 2003; Topak, 2017; York and McSherry, 2020). A primary concern is the potential transfer of such information from telecommunication companies to law enforcement agencies (FDPIC, 2020a; Ghosh, 2020; Polish Government, 2020). Organizations such as Human Rights Watch fear that some of the technology being used to seize the pandemic could be equally used to trace rioters in massive protests, such as the protests arising from George Floyd’s murder (Toh and Brown, 2020).
The deployment of new applications and the intensified use of existing technologies to tackle the Covid-19 pandemic do not happen in a regulatory vacuum and therefore, compliance with data protection law, in particular the EU General Data Protection Regulation (GDPR), and the general data protection principles (GDPR Article 5) is key to the successful deployment and adoption of these technologies. General data protection principles such as the principle of proportionality, the principle of purpose limitation and the principle of transparency set the stage for compliance of contact-tracing applications and video-conferencing tools with data protection law. These principles require developers of new or existing technologies to implement privacy-friendly options from the beginning (‘privacy-by-design’) and ensure that not more personal data is processed than necessary to trace contact with a person infected with Covid-19, that the data is not used for other purposes than curbing the spread of Covid-19, and that the use of any data collected with these new and innovative technologies is transparent to the users. Finally, the GDPR requires a person’s consent to the processing of their health data (GDPR Article 9). These general privacy principles do not prohibit the surveillance measures currently taken by governments or organizations, but rather ensure the correct handling of the data once processed under the current circumstances (GPA, 2020).
Research has started to explore the spatiality of Covid-19, in terms of data collection and privacy (Frith and Saker, 2020; Poom et al., 2020), but has so far mostly overlooked the temporal nature of how such privacy-negating technologies are deployed and normalized. Given the speed and complexity of the pandemic’s societal impacts, it is equally as challenging as it is essential to track the development and use of digital surveillance and review the impact of such technologies on society to ensure compliance with data privacy law in all contexts. In this paper, we, therefore, look at the temporality of the Covid-19 pandemic and how speed interacts with regulatory protection of privacy in Europe.
This article is divided into two main parts. After the introduction, we discuss the notion of rushed innovation and argue how rushed innovation can have negative implications in general and in terms of privacy protection. This second section, which focuses on the socio-technological aspects of rushed innovation, is divided into two subsections. The first subsection revolves around the rushed deployment of novel technological solutions, with contact tracing apps as a central case. The second subsection focuses on the rushed adoption of pre-existing technologies, with Zoom as the central case. These cases were, and remain, topics that all governments, as well as industrial players in Europe, have to deal with in connection with the curbing of the spread of Covid-19, while at the same time ensuring compliance with privacy law and related expectations. The second key section of the paper then focuses on regulatory aspects of rushed innovation, focusing on the EU general data protection framework (GDPR). This section examines whether this legal framework accelerates or slows down rushed innovation. We bring forward concerns relating to how the law allows introducing products and services in the market that have not followed appropriate assessment and do not demonstrate full legal compliance. The article ends with a collection of lessons learned and paths forward.
The socio-technological aspects of rushed innovation during the Covid-19 pandemic
We live in an age of hypertemporality and hypercompetition (Carillo, 2005). Acceleration, as theorized by sociologists such as Hartmut Rosa, is a key aspect of modernity (Rosa, 2013) and technical acceleration, describing the increased speed of technological innovation and goal-directed processes, goes hand in hand with the acceleration of social change (e.g. cultural and institutional norms) and the pace of life. Such acceleration is particularly visible in the tech industry, where speed acts as an important part of the business model, both among large tech companies and start-ups (Tromble and McGregor, 2019; Wajcman, 2019). Speed in technological innovation often leads to competitive advantage and network effects (Eisenhardt and Martin, 2000; Griffin et al., 2019; Kessler and Chakrabarti, 1996). Accordingly, many tech firms have embraced a technological solutionism and social engineering approach that is characterized by ‘release early, release often’ in its softer and more benign version (Raymond, 1999) and ‘move fast and break things’ in its more aggressive version (Taneja, 2019). Addressing the critical importance of speed, innovation frameworks such as the lean start-up approach explicitly aim at shortening development cycles through the quick development of prototypes or minimum viable products and have enjoyed widespread success in the business community (Ries, 2011).
Rapid innovation is particularly important in a crisis, not only so that companies can exploit market advantages in what Klein (2007) refers to as ‘disaster capitalism’, but so that life-, society- and economy-preserving solutions can be rolled out where needed. Acknowledging the subjective experience of time (Orlikowski and Yates, 2002), Calhoun (2010) emphasizes how urgency shapes our temporal framework in a crisis. Governments worldwide face criticism for acting too slowly and costing lives. The constant update of Covid-19 death statistics online, like a ticking doomsday clock, offers a stark reminder that nothing can ever happen quickly enough in a crisis. There is a growing societal anxiety for quick responses and quick action to solve some of the greatest world challenges (Leslie, 2020).
Experts have stressed how innovation in the field of pharmaceutical technology (Bryan et al., 2020) should be inspired by fast and frugal innovations (Harris et al., 2020). Indeed, within research and policy fields, there is currently a lively discussion about fast-tracking vaccine development. News of the Russian authorization of Sputnik V and the Chinese authorization of CanSino’s vaccine for military use (Liu and Woo, 2020), as well as Sinovac’s and Sinopharm vaccines for emergency use (Liu and Kim, 2020), have sparked widespread criticism among Western experts (Callaway, 2020). However, daily media coverage in Europe centres around when, not if, a vaccine will be ready, to the point that a temporary pause on the Oxford University/AstraZeneca vaccine trials in September 2020 led to front-page news.
However, rushing innovation comes with its own problems. For instance, Tromble and McGregor (2019) discuss a speed-oriented focus on Facebook as one reason for some of the adverse developments in recent years. Facebook encourages its employees ‘to be creative, to experiment, to move quickly from one idea to the next. If something fails, that is no problem. They keep trying until something works’ (p. 325). However, such a speed-oriented process prohibits a slower but more methodical approach – one which could anticipate better the vast negative social impacts of social media in recent years.
Rapid innovation and regulatory compliance also often make for poor bedfellows. As Harris et al. (2020) remark concerning fast innovation for the Covid-19 pandemic, ‘[i]t is sometimes necessary to forego high regulatory standards in order to rapidly address new demands at low cost’. Coined by Hermosilla (2020) to refer to pharmaceutical companies rushing through licensing processes to the detriment of the product, ‘rushed innovation’ is a useful concept to explain the trade-offs between innovation, market advantage and regulatory compliance.
While rushed innovation does not have to mean that privacy rights are infringed upon, it is indeed likely that other concerns such as functionality and usability will be prioritized. Public health tools, in particular, will require access to certain information and can impinge on one’s privacy. Yet, as Gasser et al. (2020) state: ‘[p]rivacy risks vary depending on the purpose and data types used by a digital tool’. The privacy-invasiveness will depend on the granularity of the data obtained, the centralized access versus decentralized access to said data, the duration of access to said data, and the risks of identification. Because ‘privacy risks can change and accumulate over time’, it is critical to have a strong regulatory framework in place to address privacy issues caused by rushed innovation (Gasser et al., 2020: 5). However, at the same time during this pandemic, we can see a shift towards protecting public health over privacy across many levels while relation on different justifications for this shift. The overall goal – protecting public health – seems to coincide on all levels as long as the principles of purpose limitation, proportionality and transparency are ensured (European Commission, 2020).
In the following, we distinguish two forms of rushed innovation, which we call
Rushed deployment of new solutions
Rising infection rates and spreading uncertainty have led government officials and private companies on an urgent search for technical solutions that could be deployed promptly and efficiently. In an effort to monitor the spread of the virus, government officials turned towards Big Data solutions, which promise to analyze aggregated mobility data to provide information about when and near whom individuals were at a given moment. Multiple countries have implemented such practices (Buckee, 2020). Some countries, including Germany (Dalg, 2020), Switzerland (Wyss, 2020) and Norway (Balsari et al., 2020), have mandated national emergency cooperation with telecommunication providers to track the flow of individuals, mainly to ensure that people are not gathering in large groups or to track persons who have tested positive for Covid-19. In Italy, authorities analyzed citizens’ location data from their phones to determine how many people are following the government’s lockdown order. They could also identify the average distances in which citizens move every day (Corriere Della Sera, 2020).
Two technical approaches have, so far, dominated the dash towards the development of contact tracing technologies: GPS methods of co-localization tracing and Bluetooth-based methods of proximity tracing (Leslie, 2020). These practices have been scrutinized by privacy scholars since governments and health authorities have not provided enough transparency about how the tracking works (Ghosh, 2020; Steiger, 2020). Although the accumulation of data within telecommunication companies is not new, and its privacy implications were predictable, disclosure to the government is a novel consequence. This raises concern about the so-called private/public surveillance partnership (Schneier, 2015), referring to government and corporations’ cooperative effort to collect mass amounts of personal data. Such attempts can be observed in the UK, where Prime Minister Boris Johnson called on the technology industry to contribute in the fight against Covid-19 (Waterson, 2020).
Beyond monitoring, many countries have launched contact-tracing apps to limit the spread of Covid-19. Mobile applications offer various ways to collect and distribute Big Data (Lai and Flensburg, 2020: 1). Even though contact tracing apps were used before to track epidemics of Ebola (Sacks et al., 2015) and Tuberculosis (Ha et al., 2016), their adoption in a pandemic is new. Such voluntary applications (Cho et al., 2020; Parker et al., 2020) have surged across Europe, giving rise to a broad discussion about the benefits and risks of such contact tracing apps. Scholars have warned about the risk of growing function creep and have argued: that apps […] should be evaluated in terms of the contextual integrity of information flows; in other words, the appropriateness of sharing health and location data will be contextually dependent on factors such as who will have access to data, as well as the transmission principles underlying data transfer (Vitak and Zimmer, 2020: 1).
Even though, on aggregate, European countries want to be role models by aligning rushed deployment and privacy rights, individual countries do not always achieve this goal (Algorithmwatch, 2020). In Norway, for instance, technology-oriented solutionism, specific cultural forms of trust, and the Norwegian dugnaðr concept (which refers to collective unpaid and voluntary work), all fostered the rapid deployment of the Norwegian Smittestopp app to fight Covid-19 (Norwegian Institute of Public Health, 2020; Sandvik, 2020: 2). Yet, due to privacy concerns, the Norwegian Data Protection Authority quickly imposed a temporary ban on the Smittestopp contact tracing app (Datatilsynet, 2020).
Rushed adoption of existing technologies
In addition to the rushed deployment of novel technologies, we also witness an expansion in the use of pre-existing technologies. Institutions (such as universities, schools and corporations) were forced to rapidly implement solutions to conduct their daily businesses online. A quick fix that is typically applied in such situations is to obtain consent of students or employees even though public institutions and employers should rely on the legal ground of compliance rather than consent if possible. Institutions rushed to obtain licenses for platforms that otherwise would have undergone more scrutiny from the data protection departments. At a university level, this included software solutions to conduct online teaching, but also online exams such as Proctorio and ExamSoft (Patil and Bromwich, 2020).
The introduction of social distancing has also led to reduced face-to-face interaction between individuals, forcing people to turn to digital alternatives. According to a Gallup poll of US-based adults in April 2020, 25% use social media more frequently since the beginning of the Covid-19 pandemic, 57% use them the same and 7% use them less frequently (10% do not use social media at all) (Ritter, 2020). Zoom has been used for birthday parties (Matyszczyk, 2020) and weddings (Pajer, 2020), while informal get-togethers on apps such as Houseparty are increasingly common (McIntosh, 2020). The forced transition to online socializing raises the question of whether consenting to use social communication platforms can still be considered voluntary during a pandemic.
Rapid adoption of third-party video-conferencing platforms such as Zoom or Skype creates new vulnerabilities in terms of privacy and data protection. Hosts of Zoom calls can see the IP address, location data and device information of participants. Although Zoom made changes to its code in late March 2020 to stop sending data to Facebook (Cox, 2020), privacy concerns about the software continue. One severe social privacy and data security infringement has been Zoombombing (Wakefield, 2020). Zoom has seen much criticism for its privacy policy and security standards during this pandemic (e.g. Leitschuh, 2019; Searls, 2020). This continued pressure forced Zoom to react and thoroughly revise its privacy policy (Zoom, 2020a) (the EDPB as well as the Swiss Federal Data Protection and Information Commissioner (FDPIC) even issued guidance on the safe use of video-conferencing tools as a reaction to this rushed adoption of existing technologies (EDPB, 2020b; FDPIC, 2020b). Zoom claims to have not changed their practices, implying that they have always been compliant. Rather, they claim to have only updated their privacy policy ‘to be more clear, explicit, and transparent’ (Zoom, 2020b) and to have issued a 90-day feature freeze to address security issues (Zoom, 2020c). In any case, the use of clearer language in their privacy policy improves transparency and strengthens user decision making, according to the principle of informed consent.
The widespread transition to remote work because of Covid-19 has also increased the use of digital surveillance measures in the home (Maalsen and Dowling, 2020). Many professionals currently working from home are now subject to greater surveillance than previously experienced. One example is screen capturing video-services like Sneek, which automatically takes photos of employees through their webcams every five minutes (Holmes, 2020).
Remote work has traditionally raised concerns about the loss of managerial oversight (Sewell and Taskin, 2015), leading to potential worker misbehaviour or slacking. Although recent research has emphasized remote workers’ own desire to maintain visibility so they do not feel ‘exiled’ (Hafermalz, 2020), there are unresolved privacy concerns when workers willingly or unwillingly submit to remote digital surveillance. Although the sales of workplace surveillance tools have increased since the start of the pandemic, privacy concerns about how employers monitor the performance of employees have also grown, raising questions of what form of surveillance is necessary and what forms of surveillance are merely intrusive, such as continuous desktop and webcam sharing (Holmes, 2020; Morrison, 2020).
The regulatory aspects of rushed innovation during the Covid-19 pandemic
Rushed deployment and rushed adoption do not occur in a regulatory vacuum. On the contrary, industry players and governments are well aware of the comprehensive regulatory frameworks protecting individuals’ personal data and privacy. The EU GDPR sets out ample provisions on what must be considered when implementing new processes and adopting new technologies that impinge on individuals’ privacy. For instance, private persons processing personal data are required to adopt a privacy-by-design approach (GDPR Article 25), which means that they must consider the privacy implications their new processes and technologies may have on their users and address these implications in their design.
Anyone processing personal data must comply with the general data protection principles (article 5 EU GDPR), particularly the principle of proportionality (which entails the principle of data minimization), the principle of transparency and the principle of purpose limitation. These principles have been discussed in connection with the rushed deployment of new solutions and the rushed adoption of existing technologies. Companies and governments must ensure that only personal data necessary for assessing whether someone may be infected with Covid-19 is processed in their solutions and that access to the collected data was restricted to a strictly need-to-know basis. Moreover, personal information must be deleted once the data is no longer needed and users of these technologies should be transparently informed on how their data is processed. Furthermore, the data processed within these technologies must be adequately protected by technical and organizational safety measures to prevent the accidental and unlawful destruction, loss alteration and the unauthorized disclosure of or access to the collected data (GDPR Article 32). Health data (such as information on flu symptoms or whether someone has tested positive for Covid-19) is particularly strictly regulated, as it is considered as sensitive personal data that requires higher protection than ‘regular’ personal data (GDPR Article 9).
In general, one relies on the individual’s explicit consent to process health data. The rushed and sometimes enforced implementation and use of video-conferencing tools or contact tracing applications raises the question of whether consent to the use of rushed innovations can still be considered voluntary during a pandemic. In any case, whether consent is an adequate tool (Hirsch, 2020; Mayer-Schönberger, 2010) and how it should be implemented digitally (Edenberg and Jones, 2019), have been debated in privacy literature. For those situations, European data protection law requires the data controller to demonstrate that it is possible to refuse or withdraw consent without detriment (GDPR, Recital 42). Thus, the argument can be made that consent currently given online in emergency circumstances might be considered invalid.
Rushed market entrance versus slow regulatory frameworks
From a business perspective, implementing data protection requirements is burdensome and time consuming (Tikkinen-Piri et al., 2018), especially so if there is a pressure to act fast during an emergency. Existing privacy law does not prevent companies from rapidly deploying technologies that are not compliant with data protection law. Companies can conduct initial risk assessments and decide whether it makes more commercial sense to rush deployment, even if the technology is not fully compliant. Companies can choose to act first, and then try to clear-up the privacy-mess later. However, if companies deploy a product that is not compliant with the GDPR, they risk large fines (GDPR Article 83). Even if they are not fined, a supervisory authority may decide to inspect, suspend or completely ban the use of non-compliant technologies (GDPR Article 58).
Introducing products or services to the market that do not demonstrate full legal compliance is not a new phenomenon caused by the Covid-19 pandemic. Medical devices, such as pacemakers, do not always undergo appropriate testing to demonstrate they are safe and effective before entering the market (Van Norman, 2016). In Europe, non-implantable and low-risk devices are self-marked, i.e., the manufacturer itself certifies the compliance and applies for a Conformité Européenne (CE) mark. High-risk devices, however, need to undergo an outside revision process. This involves a notified body which checks compliance with the relevant legislation and issues a CE mark if the device meets the requirements. CE marks across Europe were authorized ‘without further controls and no further evaluation’ until 2010 when new regulations obliged the approval of devices that are similar to already legally marketed devices (predicate devices) (Jefferys, 2001). The notified-body European system was designed to promote innovation, similar to how the EU GDPR exists purely as a legal framework which favours innovation and allows companies to introduce technologies that process personal data to the market.
However, the goal of the data protection supervisory bodies in Europe is clearly the protection of an individual’s privacy interests, which can lead to the supervisory authority stopping, suspending or banning certain technologies if they are not compliant. In this case, the GDPR does provide a certain balance, insofar as companies wanting to introduce new technologies to process personal data may have to conduct a data protection impact assessment that takes into account ‘the nature, scope, context and purposes of the processing’ (GDPR Article 35). This especially applies if the envisaged technology entails processing sensitive personal data, such as health data, on a large scale (GDPR Article 35, Para. 3, Letter b).
Governments operate in a tightly regulated environment. In Switzerland, for example, government bodies may process personal data only if a legal basis allows them to do so (Article 17 of the Swiss Federal Act on Data Protection). They cannot take a risk-based approach as companies can. The contact and proximity tracing applications depicted above are a good example of how, although some governments took much longer to deploy these new technologies, they worked together with privacy scholars and data protection authorities to ensure compliance with privacy laws. The Norwegian example shows that if you go to market with a non-compliant tool, your product may be banned or its use may be suspended until it is compliant which is even more detrimental than just slowing down the process to be fully compliant.
The difference between the rushed deployment and adoption of medical devices and privacy-harvesting digital technologies is that ‘interference with data protection rights does not depend on whether there has been any harm or inconvenience to an individual’ (Kuner et al., 2015). The example of Zoom shows that it is possible to adopt existing technologies and scale their use in a very short time, even though they are very privacy unfriendly, and then the company can clean their ‘privacy-mess’ up retrospectively.
Given that some of these technologies process information that may have an ulterior impact on the population’s health, however, such an interference could be more salient. At least, that is how some data protection authorities have interpreted it, halting Covid-19 solutions due to data processing concerns (Datatilsynet, 2020). The state of force majeure and the demand for a hurried response to ensure the safety of the population’s life and health allowed the deployment of ‘a highly invasive and technically unfinished app on their population’ that did not go through a tendering process or appropriate risk assessments (Sandvik, 2020). Indeed, after a thorough balance-of-interest exercise between health benefits and adverse privacy impacts, the Norwegian Data Protection Authority imposed a temporary ban on the Smittestopp contact tracing app (Datatilsynet, 2020).
Emergency laws and exemptions to foster rushed innovation
What are companies and governments supposed to do in emergencies when the protection of public health is of utmost importance, and individuals are willing to risk their privacy for the greater good? The GDPR provides a semi-useful solution. As it stands, processing health data is lawful if it is necessary for reasons of public interest in the area of public health, such as protecting against serious cross-border threats to health […] on the basis of Union or Member State law which provides for suitable and specific measures to safeguard the rights and freedoms of the data subject.
In light of this, governmental authorities have created guidelines to tackle data analysis practices during the Covid-19 pandemic, arguing that privacy and health protection can go hand in hand. The Committee of the Global Privacy Assembly (GPA) issued a guiding statement in March 2020, supporting governments and organizations to share personal information to fight the spread of the virus. The GPA (2020) states: We are confident that data protection requirements will not stop the critical sharing of information to support efforts to tackle this global pandemic. The universal data protection principles in all our laws will enable the use of data in the public interest and still provide the protections the public expects. Data protection authorities stand ready to help facilitate swift and safe data sharing to fight COVID-19.
On 8 April 2020, the European Commission adopted the
As a legitimate basis for data processing, the European Commission reminds that GDPR Article 5 e-Privacy Directive requires the consent of the user to store or gain access to information already stored on the user’s device, unless the storage is necessary for the app, and the user has explicitly requested it. GDPR Article 6(1)(e) also allows the processing of personal data to perform a task of public interest. The general privacy principles of purpose limitation, proportionality and transparency do not prohibit the surveillance measures currently taken by governments or organizations, but rather ensure the correct handling of the data once processed under the current circumstances (GPA, 2020).
While there seems to be some consensus on allowing rushed innovation in emergency situations and the adoption of emergency law may even let the processing of personal data even in non-compliance with applicable data protection law, at the same time, neither governments nor industry players took this direction (maybe because they were too afraid of the fines under the EU GDPR). This illustrates that there is no proper or comprehensive exception in privacy law itself that would explicitly allow the rushed deployment of new solutions or the rushed adoption of existing technologies which do not fully comply with the general data protection principles, as there is, for example, with regard to processing personal data for research and scientific purposes. The existing law was not tailored to handle emergencies, and this may be due to the governmental powers in such situations that can be used to override regulatory frameworks to rush innovation in breach of data protection law for the greater good if it came to it.
What all our examples show is that, in reality, privacy becomes the core value to be protected regardless of the public health benefits that privacy-invasive technologies may bring about. What we need if we want more legal certainty may be something more clear than the public health exemption which explicitly allows rushed innovation that is not compliant with privacy laws in situations of emergency. A solution could be an exemption similar to the GDPR research exemption, which states that data processing for research purposes is not considered incompatible with the initial purpose as long as safeguards are put in place (Articles 5 (1) b and 89).
While this mainly alleviates the burden regarding the legal basis for data processing, it does not help shorten the data protection impact assessment (DPIA), which might have been one of the main reasons for slower deployment. While the GDPR already allows for weighing of different interests, DPIA provisions would need to explicitly allow for interests other than privacy to be considered. Similar to the research exemption, the reasoning would also be to benefit the greater good. If higher-weighted interests, such as public or individual health, could allow for lower burden of proof, this could provide the much needed legal certainty for states and corporations in such emergency situations, where decisions have to be made based on scarce and unclear information. Such a process could enable actors to anticipate and mitigate data protection risks in emergency situations to the population, including vulnerable groups.
Conclusion: Lessons learned and steps ahead
The global spread of Covid-19 shows that, in an emergency, governments are willing and permitted to implement extraordinary measures to combat such a pandemic even if these measures directly affect the rights and freedoms of individuals. On a personal level, citizens seem to be shifting towards protecting public health over privacy across different contexts. However, ‘when choosing between alternatives, we should ask ourselves not only how to overcome the immediate threat, but also what kind of world we will inhabit once the storm passes’ (Harari, 2020).
In this sense, a first lesson learned is that the urgency of states of emergency is not always compelling enough to suppress one right against another in a balance-of-interest exercise. On the contrary, fundamental rights need to be protected equally. This pandemic, and the reactions of public authorities as well as society, show that rushed measures implemented during a state of emergency are accepted only to a certain extent, and only if they are in line with other fundamental rights. This does not hinder the implementation of measures that interfere with the privacy of citizens, but it requires governments and the industry to be mindful and compliant with existing regulations, beyond raising awareness to society. Employers, for instance, are required to assist governments with the implementation of such measures. On the other hand, they have a duty of care to protect their employees’ health, thus permitting employees to work from home and requiring employees to report symptoms connected to Covid-19. Moreover, employers must also ensure their operative business, which is why some have resorted to far-reaching employee surveillance. In this sense, a further key lesson learned is that designing privacy-friendly solutions may dissipate the concerns that the pandemic could be exploited as an opportunity to normalize governmental surveillance to a certain degree.
The banning of different applications in the fight against Covid-19, due to privacy concerns, poses the additional question of whether these ex-post temporary measures are a call out in favour of privacy or whether they elucidate the weakness of data protection regulation effectiveness. Banning COVID-related applications may be an opportunity to stress the importance of enforcing data protection principles, even in emergency states. However, it also illustrates the potential waste of resources that non-compliant solutions may entail. What remains unclear is what happens to the user’s personal information and the data protection issues arising from the use and deployment of such solutions.
For users, one lesson learned is that they should inform themselves about the privacy risks involved with newly deployed technologies such as contact tracing apps, rather than rushing to adopt them in the interest of the public good. While scholarship has warned against putting the onus of responsibility squarely on the individual when it comes to privacy (Obar, 2015), users still have power to vote with their feet, particularly for technologies that rely on network effects such as contact tracing apps. Of course, voluntary adoption decisions are not always possible, as governments and employers either aggressively nudge (Sandvik, 2020) or even force citizens to rapidly adopt problematic solutions in terms of privacy (Algorithmwatch, 2020, where the Slovenia country report by Lenart Kučić describes the ‘looming spectre of a mandatory tracking app’). Here, a critical public sphere is key that brings privacy and tech scandals to the surface, holding those responsible accountable through reputational pressure (Kolkman, 2020). Rushed adoption, and technology-infused rush more generally, can be counteracted on the user level by slow computing (Kitchin and Fraser, 2020), where users seek out more balanced digital lives through conscious initiatives and reflection.
From a research perspective, the investigation of data technologies within the pandemic is only at the beginning. Research has started to reflect on the privacy implications of the Covid-19 pandemic (Vitak and Zimmer, 2020), especially regarding the spatial dimensions and location-based data (Frith and Saker, 2020; Poom et al., 2020). It is not surprising that spatial aspects have been prioritized over temporal ones, given the data being collected within many applications to mitigate the pandemic and the strong spatial considerations that come with the pandemic (e.g. quarantine and social distancing as key spatial strategies). However, such data is often also temporal and we have argued that scholars should pay close attention to the temporal dynamics of the pandemic when it comes to privacy. For instance, to what extent can we use privacy assessments before the deployment of technological solutions to anticipate and help mitigate adverse privacy implications? More conceptual analysis and synthesis are needed to connect the first steps made in this paper to established theories of time and technology (Rosa, 2013; Wajcman, 2015) and newer work (Kitchin and Fraser, 2020).
Rushed innovation in disaster capitalism (Klein, 2007) should be investigated empirically. Case studies, such as the one discussed by Sandvik (2020), offer a fruitful approach to analyze the deployment of new solutions. We encourage qualitative researchers to document and critically investigate the temporalities of technology projects relating to the pandemic. On the adoption side, case studies could accompany the decision-making processes within organizations and public institutions that lead to rushed adoption, spotlighting privacy implications for those affected. Recent examples that ask for critical inquiry are the introduction of proctoring software at universities (Patil and Bromwich, 2020), the application of grading algorithms for A levels in UK schools (Hern, 2020; Kolkman, 2020), and the increased use of workplace surveillance software such as Hubstaff (Jones, 2020). These examples also show the intertwinement of privacy risks with intersectional concerns for rushed innovation. Those more vulnerable and marginalized in society face disproportionate privacy repercussions and more general risks from Big Data in the wake of the pandemic (Milan and Treré, 2020).
Digital inequalities research has looked into digital communication changes that come with the pandemic and how these changes are unequally distributed, favouring individuals with advanced Internet skills (Nguyen et al., 2020). Future research should study the disparate intersectional implications of rushed innovation. Quantitative surveys could be combined with qualitative and ethnographic studies with those at the margins (Marwick and boyd, 2018), giving distinct voice to their concerns. Action research from a social justice-based perspective is particularly promising, as it involves marginalized communities and emphasizes practical change through design initiatives (Costanza-Chock, 2020). Outside of organizations and institutions, rushed innovation can be investigated with a variety of user-oriented methods, including digital methods such as walkthroughs (Light et al., 2018), glitch studies analyses (Menkman, 2011), and sentiment analysis, social network analysis, or topic models of social media coverage (Sloan and Quan-Haase, 2017).
Pandemics and other world catastrophes push for immediate responses. Still, these responses will have many consequences for society in the immediate and long run that require a thorough understanding of how responsible rushed innovation can be. Greater effort in incorporating privacy considerations beforehand in the design of digital solutions is very much needed, as afterthought privacy reflections risk exposing the health of citizens, wasting public resources and worsen the consequences that the state of emergency already has for society.
Footnotes
Acknowledgements
We would like to thank the editorial team of
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Gemma Newlands and Christoph Lutz received funding from the Research Council of Norway under grant agreement 275347 ‘Future Ways of Working in the Digital Economy’. Eduard Fosch Villaronga received funding from the LEaDing Fellows Marie Curie COFUND fellowship, a project funded by the European Union’s Horizon 2020 research and innovation program under the Marie Skłodowska-Curie Grant Agreement No. 707404. Aurelia Tamò-Larrieux received funding from the Digital Society Initiative fellowship (University of Zurich, Switzerland) and the International Postdoctoral Fellowship grant (University of St. Gallen, Switzerland; project number 1031564).
