Abstract
Data-driven technologies have come to pervade almost every aspect of business life, extending to employee monitoring and algorithmic management. How can employee privacy be protected in the age of datafication? This article surveys the potential and shortcomings of a number of legal and technical solutions to show the advantages of human rights-based approaches in addressing corporate responsibility to respect privacy and strengthen human agency. Based on this notion, we develop a process-oriented model of Privacy Due Diligence to complement existing frameworks for safeguarding employee privacy in an era of Big Data surveillance.
Introduction
Data-driven technologies are increasingly gathering and processing data across the full spectrum of contemporary society and human activity. This datafication touches on most areas of life (Cukier and Mayer-Schoenberger, 2013; Neff and Nafus, 2016). It is not surprising, then, that Human Resource Management has similarly begun to embrace datafication for its core processes. Companies are increasingly attracted by the promise offered by data analytics to monitor the behaviour and performance of their employees in the workplace, sometimes even extending to non-job-related behaviour. ‘People Analytics’ vow to take human resource management practices to a new level. Often, there is also an underlying belief that technology might make people management decisions more objective, efficient and less prone to serving individuals’ preferences (Finlay, 2014). As a result, the monitoring of employees on a minute-by-minute basis increasingly affects employees across a range of industries (Ajunwa, 2020; Ball, 2010; Mateescu and Nguyen, 2019; Prassl, 2018).
The object of monitoring/surveillance also extends across employment grades and pay levels, from call centre workers to senior managers, and increasingly affects ‘thinking work’ (Phan et al., 2017). For instance, banks in the City of London rely on surveillance technology to see whether employees are present or not (Morris et al., 2017). Some companies even use more physically invasive methods of surveillance, such as microchip implants that connect employees to the company network (Astor, 2017).
The global spread of COVID-19 in the spring of 2020 has dramatically accelerated the use of workplace analytics, for at least two reasons: first, because many of the technologies involved in datafying the workplace are now being deployed and/or repurposed for public health protection, such as monitoring workers’ social distancing in factories and warehouses (Vincent, 2020). Second, the dramatic growth of home working has significantly increased demand for software solutions that offer remote surveillance and management possibilities, taking the datafication of work right into individual homes (Collins, 2020; Frantziou, 2020).
These dynamics raise major ethical questions, including notably the broader shift from human to algorithmic decision-making (Mittelstadt et al., 2016). In addition to these ethical challenges, companies face significant legal risks, as they struggle to map and mitigate the legal implications new technologies used for workplace monitoring might have on their employment relationships (Ajunwa et al., 2016). Yet, at the same time, workplace monitoring is becoming the new normal (Kellogg et al., 2020). Emerging scholarship has demonstrated that workers are increasingly confronted with a ‘black box’ at work, lacking transparency, accountability, or explanation about monitoring practices (Ajunwa, 2020; Pasquale, 2015). A major concern emerging from the current literature centres on the protection of privacy (Bhave et al., 2019) and related questions surrounding accountability structures, transparency about information sharing, and potential discrimination (Ajunwa et al., 2016; Boyd and Crawford, 2012).
Whilst expressed in a number of distinct ways across different jurisdictions, the concept of privacy ultimately protects the right to respect for private life, family life, home and correspondence (for an influential illustration, see European Convention on Human Rights, Art. 8). Big Data analytics in human resource management heavily impact employee privacy and can lead to privacy breaches, infringements and violations (Mateescu and Nguyen, 2019). The right to privacy underpins, and is closely connected to, other fundamental rights at work and beyond, such as freedom of association and speech (Grabenwarter, 2014). To respond to this multi-layered challenge of managing employee privacy at the workplace, we suggest solutions might be found in both legal and ethical scholarship to address transnational challenges for the ‘data citizen’ (Guild, 2019). A multi-disciplinary and global approach is needed to address privacy protection in a world where workplace monitoring is quickly becoming the new normal (Kellogg et al., 2020).
The privacy implications of this significant increase in the uptake of workplace monitoring technologies, as well as its managerial and legal implications, have so far been relatively underexplored in a solution-oriented analysis which looks beyond a particular regulatory regime or jurisdiction (Collins, 2020; Frantziou, 2020). In this article, we address this gap by conducting a critical inquiry into privacy issues in workplace monitoring, as well as exploring why a human rights-based due diligence approach is suitable to protect employee privacy. This approach serves as a complementary, holistic framework to existing legal and technical approaches, including data protection and Privacy by Design (PbD). We propose a Privacy Due Diligence approach that allows companies to develop a systematic mechanism to handle privacy issues in the workplace as an on-going practice, tailored to individual business models and workplace settings. The proposed Privacy Due Diligence model goes beyond a purely legal or technological solution: our model offers a dynamic managerial process to address privacy issues at the data-driven workplace as they arise, and empowers workers’ data autonomy. By taking a multi-disciplinary stance anchored in strong stakeholder engagement mechanisms, our model furthermore ensures that the perspectives and needs of all affected groups are included in dialogue at the managerial level.
To this end, we argue that a set of mechanisms from the ‘Business & Human Rights’ (B&HR) literature can address corporate responsibility to respect privacy at the workplace (Ruggie, 2007, 2013; Wettstein, 2015, 2016). Human rights due diligence is widely discussed in the B&HR scholarship and offers a rightsholder-centric approach for corporate management (Ebert et al., 2020; OHCHR B-Tech, 2021). A B&HR perspective is enhanced by ethical demands in addition to legal compliance. Its benefit lies in multi-disciplinary, process-oriented managerial tools and implementation strategies based on clearly defined human rights norms, namely the UN Guiding Principles on Business & Human Rights (UNGPs; United Nations Human Rights Council, 2011). Rather than offering static solutions, such as design options, one-off risk assessments, or
Our discussion is structured as follows. A first section provides illustrations of workplace surveillance and algorithmic management techniques, highlighting employee privacy issues along the data life cycle, from collection to erasure. We then survey existing frameworks for employee protection, from legal to design-based approaches, highlighting their strengths and identifying a number of weaknesses. It is on the basis of that discussion that we then turn to the B&HR approach as a way of providing a structured process to map risks, identify privacy gaps and anchor privacy due diligence in corporate practice.
Employee privacy issues along the life cycle of data
Algorithmic management has come to augment, or even replace, the full gambit of traditional employer functions (Trade Union Congress, 2018): whilst hiring is perhaps the most visible use of algorithmic management to date, the use of Big Data HR extends to scoring workers’ productivity (Heaven, 2020), tracking day-to-day work behaviour and even terminating employment relationships by firing workers with low ‘rates’ as determined algorithmically (Steele, 2020). Whilst present space limitations prohibit a detailed descriptive account of these technologies (Neff et al., 2020), suffice is to say that the rapid growth and expansion of algorithmic surveillance and management at work is bringing about a significant shift in work organization.
It is not difficult to imagine the ensuing risks of privacy infringement. A large range of industries aim to monitor and, to a certain extent, predict individual future behaviour using data analytics, e.g., to determine the employees’ mood and willingness to exert a task (Eubanks, 2018; Waddell, 2016). Some companies, for example, use neural networks to connect and analyse large data sets (Cheekoty, 2019). These techniques can convey profound insights about individual preferences and behaviour, but are often criticized as not being fully retraceable (Monahan, 2016; Pasquale, 2015). Employee privacy is at stake throughout the entire life cycle of data (European Parliament Position, 2014: Recitals 71a, 71b, Art. 23 para. 1, Art. 33 para. 3), which can be broken up in four phases with regard to privacy concerns resulting from data processing (Tamò-Larrieux, 2018):
During Employees might not be informed about During
Misconduct with regard to the use of data can result in a so-called ‘function creep’, meaning that the data collected is used for other purposes than previously communicated (Christl, 2017). Furthermore, data information might be an issue, as often employees find themselves in a weak position to demand transparency or insight into certain analytics practices that use their personal data.
The inherent conflict between datafication and privacy
The more data is collected about individual employees, the more valuable it gets for predictions based on these techniques. Whilst datafication technologies, such as AI, build on large amounts of data for increased accuracy of results, many privacy provisions would call for alignment with the data minimization principle (e.g. taken up in GDPR Art. 5) that stands fundamentally at odds with Big Data techniques. This is a tightrope walk for any organization: Whereas compliance or corporate governance departments might call for privacy as a high priority, business intelligence and HR management might be highly interested in collecting and processing as much data as possible (Koops and Leenes, 2014).
Existing frameworks for the protection of employee privacy at the workplace
In the following, we explore selected current frameworks for the protection of employee privacy in the workplace, highlighting the potential of both legal and technological solutions in resolving the inherent conflict of interest between Datafication and Privacy. We also identify a number of potential shortcomings in both legal and technical solutions, as set against the background of broader socio-technical notions of privacy and workplace monitoring.
Legal protection of employee privacy at the workplace
The applicable law to an employment contract and to an individual employee always relates to a specific jurisdiction. Whilst companies may be held accountable for privacy infringements in the workplace based on national labour or data protection laws, records of personal data seem to float freely across jurisdictions. Companies perceive increased pressure to deal with privacy issues at the international level, partly due to new legislation in Europe, the US (California Consumer Privacy Act (CCPA), 2018) or also Brazil (CCPA, 2018; Singer, 2019; Thomas, 2019). Therefore, data-driven workplace monitoring is a phenomenon affecting data sharing practices beyond a nation’s border. Multinational companies operating in several jurisdictions with multiple privacy standards will strive to find a solution that protects privacy across geographies (Bhave et al., 2019; Guild, 2019).
A growing body of scholarship is exploring the role of privacy and data protection in the context of work, with a particular emphasis on European regulatory regimes (Brassart Olsen, 2020; Otto, 2019; Simitis, 1999). Given the EU’s early regulatory invention, there is clear potential that the so-called ‘Brussels Effect’ (Bradford, 2020) will lead to spill-overs of similar legislation in jurisdictions beyond European borders. A number of distinct yet overlapping and closely intertwined legal regimes in Europe aim to protect aspects of employee privacy, including the European Convention on Human Rights (1953), the European Union Data Protection Regulation (GDPR, 2016), the EU Charter of Fundamental Rights (2012) and national employment law regimes. One aspect of privacy, the respect for private life is required by the European Convention on Human Rights of the Council of Europe (1953) as well as national legislation: The respect for private life also extends to privacy in the workplace, as recognized by the European Court of Human Rights in Niemitz v. Germany (1992; Grabenwarter, 2014). Moreover, data protection law in the form of the GDPR addresses privacy issues resulting from datafication in the workplace (see also EU Directive (EU) 2019/1152; see further Otto, 2019). The GDPR is directly applicable to private actors within the EU member states and even has some extraterritorial effects (GDPR). Information about the workforce can only be ‘collected for specified, explicit and legitimate purposes’ (GDPR, Art. 5(1)(b)); there are further safeguards in place for sensitive data, including ‘racial or ethnic origin, political opinions, religious or philosophical beliefs, or trade union membership’ (GDPR, Art. 9(1)).
Perhaps the strongest safeguard can be found in GDPR Article 22(1), which grants workers the ‘right not to be subject to a decision based solely on automated processing … which produces legal effects concerning [them] or similarly significantly affects [them].’ As the WP29 Guidelines EU General Data Protection Regulation (GDPR) Regulation (EU) (2016) make clear, this provision is to be interpreted widely: it will not be sufficient, for example, merely to ‘fabricat[e] human involvement’. In an employment context, it would appear that the deployment of automated scheduling software and other management tools falls within the Article 22 prohibition, as they will frequently involve ‘decisions that deny someone an employment opportunity or put them at a serious disadvantage’ (WP29, WP 251rev.01: 22). That said, Article 22 also provides for a number of exceptions.
Other scholars have also elucidated a ‘Strasbourg Effect’ similar to the ‘Brussels Effect’ by the GDPR (Bygrave, 2020), tending from the regional level (Council of Europe Recommendation No. R (89) 2 and Rec (2015)5; WP29, 2001, 2002, 2017) to the international level. The right to privacy in the workplace has also been recognized in concrete terms at a universal level (International Labour Organization, 1997) efforts to extend the responsibility for respecting human rights by obliging private companies to ensure compliance with international human rights norms are increasingly becoming visible at the international level (Kälin and Künzli, 2019). One reason for this is that private companies, including tech industry companies, have gained powerful agenda-setting power ever more resembling a state actor: private companies using datafication at the workplace have a strong stake in determining the ways in which an individual’s life is transposed into quantifiable data (Keats et al., 2014) – and they need to do so responsibly to not violate employees’ human rights in the workplace context. Overall, however, no legal regime to date has laid down comprehensive protective standards internationally.
The promises and pitfalls of design-based approaches to uphold privacy at the workplace
A number of solutions based on tech-design have also been proposed to complement existing legal frameworks for the deployment of people analytics software – though they are not necessarily able to provide a complete solution to the concerns we have identified. One such tech-based approach to provide better privacy protection across borders by embedding design specifications of information technologies, accountable business practices and networked infrastructures is PbD (Cavoukian, 2012; Koops and Leenes, 2014; Rubinstein, 2012). PbD has the potential to protect personal data and prevent legal proceedings. It departs from a decentralized set-up of safeguards against privacy issues, potentially suitable for privacy protection of transnational people management practices (Koops and Leenes, 2014).
The expectations for PbD from scholars from different continents and disciplines are high: PbD is described as a pragmatic compliance enabler to guarantee important elements of procedural regularity (Kroll et al., 2017; McQuay and Cavoukian, 2010). Some scholars claim that it will be difficult if not impossible to achieve meaningful privacy protection in the 21st century without PbD (Dix, 2010). Indeed, design-based approaches are popular even beyond the tech industry and have increasingly found their way into legal frameworks. According to the GDPR, data protection by design is an ‘appropriate measure’ to comply with data protection law (GDPR, Art. 25 para. 1). The European Court of Human Rights (ECHR) has been relatively early in embracing ideas similar to those of PbD, such as in I v. Finland (2008).
Yet, the claim that design-based solutions within the current legal structures alone can provide sufficient privacy protection has been contested in the light of on-going technological progress (Koops and Leenes, 2014; Rubinstein and Good, 2013; Spiekermann, 2012). There are three overarching reasons why existing legal frameworks and/or tech-design approaches can offer only limited protection to prevent an infringement of privacy in the context of workplace monitoring.
Missing contextuality for proportionality & consent
At present, it is not conceivable that a specific design can respond to legal situations where
The principle of
In addition to proportionality, consent is a factor of major concern in the workplace setting.
Blurry legal terminology translated in code
A second obstacle for design-based approaches is the lack of consensus about the meaning of legal terms. It is difficult for a computer programmer to translate legal concepts into code if the respective legal rule builds on vague legal terms. The concept of PbD consists of not one, but two unclear terms:
Technological progress outpacing design-based privacy protection measures
The current state of technological development can overhaul the possibilities of PbD over time. On the one hand, the technology meant to protect privacy today is developing more slowly than newly invented privacy-invading technology (Montjoye et al., 2013). As a result, gaps in protection are omnipresent. However, the lack of technological advancement might only be a temporary problem. On the other hand, the technology of today might not protect against privacy issues caused by the technology of tomorrow, such as the risk of anonymization not being effective due to potential re-identification in Big Data environments (Rocher et al., 2019): So, there is a leapfrog gap between privacy-invading technologies and privacy-protecting technologies.
In sum, just as we saw the promises of particular legal solutions, PbD is an appealing and widely accepted strategy of the digital age to protect employee data. However, we have identified why PbD offers only limited privacy protection at the workplace. We must therefore search for an additional protective approach which will be complementary to, rather than exclusive of, existing approaches.
A broader privacy approach is necessary
We have shown that PbD with its tech-based solutions is
An organization has to consider the socio-technical notion of datafication to grasp how and when employees are subject to monitoring technology and how they react to this (Neff et al., 2020).
Organizational measures need to be systemically integrated and follow an on-going, consistent evaluation of potential privacy risks for the entire data life cycle. The key drivers behind the concept of PbD include ‘accompanying’ organizational measures for accountable business practices but only to a bare minimum. Faced with the scale and intrusive nature of Big Data techniques, these accompanying measures appear insufficient. Privacy-related decision-making cannot happen in an executive management or legal compliance silo. A broader privacy approach is necessary that can also encompass ethical expectations towards the fair treatment of employees by the management deciding about workplace monitoring measures. Representatives from all affected stakeholder groups need to be strategically involved.
The holistic model to uphold privacy at the workplace offered by business & human rights
Understanding how an organization can deploy technology without violating privacy demands knowing how organizational stakeholders make sense of the technology in use and how much agency they have (Nagy and Neff, 2015; Wagner, 2019). A major asset of a human rights-based approach is that the rightsholder serves as the focal point of attention and is not neglected in a ‘passive’ role as data subject. At the same time, however, it is important to recognize the potential pitfalls of a fundamental rights approach to workplace protection – including in particular the charge of its atomistic nature, reducing worker solidarity, and thus potentially exacerbating the very inequality of bargaining power which triggers the need for protection in the first place (Youngdahl, 2009). One way of addressing that challenge is to ensure that collective as well as individual employee voices are brought back into the conversation through on-going stakeholder engagement, which can include trade unions, works councils, or other worker representation bodies.
The rights-based approach of B&HR calls for the prevention, mitigation and remediation of negative impacts on human rights through all business operations and is applicable to the workplace and towards companies’ own employees. It has the following three overarching benefits to protect privacy at the workplace (Alston, 2005; Ruggie, 2007, 2013; Wettstein, 2015, 2016): Firstly, it refers to a universally defined frame of reference with the Universal Declaration of Human Rights and the UNGPs (2011). Secondly, it proposes concrete managerial proposals and processes through human rights due diligence that can connect to existing risk assessment processes within the business to achieve human rights respecting business conduct. Thirdly, the notion of B&HR restates the state duty to protect human rights, also in technology (OHCHR B-Tech, 2021), whilst perceiving the state in a non-static manner, and emphasizing the responsibility of business to respect human rights, such as privacy at the workplace and provide ‘human agency’ to all affected stakeholder groups by stakeholder engagement (Wagner, 2019).
In the line with UNGPs, all businesses have the corporate responsibility to respect human rights across their business activities. This notion of corporate responsibility under B&HR is distinct from the conceptualization of corporate responsibility in the academic discourse on ‘Corporate Social Responsibility’ (CSR) or ‘AI Ethics’: CSR and AI Ethics have no common reference framework and definitions vary from company to company, ranging from voluntary efforts to industry self-regulation (Smuha, 2020). The consequence is that both CSR and AI Ethics have been critiqued as volatile for hiding unpleasant facts (‘whitewashing’), rather than addressing root causes or mitigating actual risks (Wagner, 2018). At the same time, the B&HR approach does not neglect ethical considerations but rather ties them to the UNGPs as the baseline to depart from: During the process of due diligence, considerations from the various approaches within the AI Ethics discourse can be integrated (Smuha, 2020).
Due diligence in line with B&HR is not solely a legal or technical process but also a multi-disciplinary managerial stance to uphold ethical values by respecting human rights across company operations and integrating rightsholders’ voices (McCorquodale et al., 2017; OHCHR B-Tech, 2021). It can give back agency to humans rather than making them a ‘basic rubber-stamping mechanism in an otherwise completely automated decision-making system’ (Wagner, 2019). Following the B&HR rationale, companies should carry out due diligence regarding the impact of their business on human rights, including the privacy of employees. Private employers should therefore respect the privacy, along with connected human rights, of their employees. The right to privacy remains closely inter-connected with other fundamental rights and cannot be discussed in isolation from other human rights at the workplace. The protection of personal data is a specific aspect of the right to respect for private life (Grabenwarter, 2014). The essential aim of the Privacy Due Diligence process is, in line with arguments made by Ajunwa et al. (2017: 775), to emphasize the right to privacy of the employee in the light of the employer and allow for employee autonomy over their data and provide for greater data autonomy. Additionally, a solid understanding of the technological state-of-the-art and its analytical capacity is necessary to grasp the dangers of workplace monitoring (Ball, 2010). Based on the requirements of the UNGPs, companies need to formulate policies on privacy at the workplace and implement them using a due diligence process (UNGPs, 2011: Number 15). Human rights due diligence is to be integrated as an on-going process, aiming at on continuous improvement. The focus lies on the rightsholder(s) and is naturally context-dependent, given the space it creates for stakeholder engagement and representation of voices from most impacted. This means that every business has the responsibility to protect its employees from privacy infringements across its operations.
The devil is in the detail. There is no ‘one size fits all’ for human rights due diligence. This means that a due diligence model focussing on privacy needs to be implemented as part of a wider conceptual human rights understanding at the company level. For the context of workplace monitoring, it is important to emphasize that an intrusion into the private sphere of an individual lays bare the very data that people analytics might use, in particular in data-driven organizations. Hence, upholding privacy can be seen as one of the gateways for human rights protection in the data economy. A violation of privacy can impact other human rights. For instance, workers can be prevented from carrying out a strike or other form of collective form of resistance (right to freedom of association) as they were monitored even whilst still mobilizing for a certain political-social cause (Peterson, 2020). Hence, upholding privacy can be crucial to protect against cascading infringements of inter-connected human rights. For instance, companies need to be aware of their responsibility to protect employees from risks of negative repercussions of privacy invasions on their mental health (Hillmann, 2015).
Introducing privacy due diligence
In the following, we describe how a company can enact privacy protection at the workplace, following a due diligence approach. Our
The Privacy Due Diligence model we suggest follows a
Mapping the ‘privacy footprint’
The first step is to map the scope of privacy concerns. A solid understanding of the technological state-of-the-art and its potential from a technical side is necessary to grasp the privacy footprint. To analyse adverse impacts, it is important to engage with a wide variety of different audiences to understand all privacy implications of a company’s workforce monitoring practices. Before processing data in the workplace context, it is vital to ask what the purpose is (purpose specification) and consider data minimization practices. Asking the purpose question should tease out a lot of privacy concerns from the outset (cf. also considerations in step 3).
It is questionable whether an employer should collect vast amounts of data in the first place. Yet, if an employer decides to do so, due diligence should focus not only on data minimization, but also on data quality. To this end, we can apply the
Yet the model of Privacy Due Diligence goes beyond a solid interpretation and application of data protection law: Its checks and balances are oriented towards ensuring the requirements of the UNGPs
Companies need to grasp where their operations affect employees’ privacy most. Key questions include: Which groups are affected by privacy issues and in what ways? Who might be particularly vulnerable? The privacy impact assessment as part of the Privacy Due Diligence should include a hybrid model that consists of both the engagement with internal stakeholders, as well as the strategic involvement of additional external stakeholders.
Whilst, for example, the GDPR proposes a data protection impact assessment (DPIA) to assess how personally identifiable information is collected, used, shared and maintained within an organization (Hartzog, 2018), it neither expressly stipulates an obligation to take the expressed opinions into account, nor includes potentially affected stakeholders, besides data protection officers, the employees and the supervisory authority (GDPR, Art. 35 para. 2, GDPR, Art. 35 para. 9). GDPR DPIAs can be integrated into Privacy Due Diligence if amended by strategic engagement with potentially affected stakeholders. A solely internal process is at risk of being biased to the companies’ interests, whereas a process purely targeted as a reporting exercise to a supervisory authority misses the point of the on-going character of due diligence. Hence, the Privacy Due Diligence approach builds on both external and internal stakeholder involvement.
2. Privacy gap analysis: Identifying existing processes and potential disparities
In this second step, the business sets up an inventory of privacy-protective measures in place in the company to determine where gaps exist with regard to privacy protection at the workplace in data-based management processes. As stressed previously, workplace monitoring needs to be limited to its appropriate context, the actual workplace and the actual work tasks and this prerequisite should not be able to be waived away with notice-and-consent mechanisms (Ajunwa et al., 2017:774f). In this step, some companies might discover design-based solutions to address emerging privacy issues. Yet, as sketched above, tech-solutions only protect against privacy infringements to a certain extent. From a B&HR perspective, a company enters a Are all necessary legal requirements met? This includes taking into account the context, proportionality, consent plus establishing clarity of the meaning of legal terms and the technological state of the art of protective measures. Are those ethical challenges with regard to privacy also addressed (legal grey areas) which might lead managers or employees into a socio-technical dilemma?
This gap analysis goes beyond the legal framework and addresses issues arising from regulatory gaps or different legal notions across jurisdictions for a sound company policy on privacy across jurisdictions. There are well-established gap analyses that focus exclusively on the legal dimension and deal with elementary issues such as lawful basis and transparency, or data security. Such gap analyses can easily be integrated into a Privacy Due Diligence approach. However, the systematic handling of such regulatory gaps is more challenging. The gaps identified are highly dependent on the respective business activity, sector, or employee groups.
Privacy Due Diligence can identify and address arising privacy gaps better than a purely legal or technical assessment. For example, ‘Hubstaff’ offers software recording employees’ keyboard strokes, mouse movements and visited websites or ‘Time Doctor’ takes videos of employees’ screens and/pictures through a webcam each 10 minutes to check that employees are at their computer (Heaven, 2020). Many decisions here remain subject to the decision-making inside companies, but should be subject to wider stakeholder engagement practices. Such decisions need to be made in line with human rights requirements, e.g., it is hardly justifiable inferring political opinions, sexual orientation or information about an individual’s health by analysing clicking and browsing patterns at the workplace.
Often, rightsholders might not be fully able to anticipate future privacy risks: Particularly in dealing with novel technology, the potential consequences cannot be fully estimated today. It is important to ensure ethical acceptability by considering normative issues emerging from the use of workplace analytics. A legally and ethically demanding gap analysis therefore cannot process an existing catalogue of standardized, formulated items, but requires a deeper reflection based on ethical key questions around the interpretation of what constitutes the right to privacy.
Prioritizing measures, impact mitigation and management
Privacy dilemmas result from a complex interplay of interests and hence do require human judgement and weighing of interests. This allows for coping with situations where design-based approaches fail to deliver proper privacy protection. Room for human judgment is required where the proportionality of an employment relationship needs to be assessed.
The most severe impacts from a rightsholder perspective (salient privacy issues) need to be addressed and acted upon first. Companies need to identify what next steps to take in order to mitigate risks ranked by salience of the privacy risks. For impact mitigation and management, a company needs to outline how the gaps identified in step 2 can be closed for salient privacy issues. For example, Deloitte and Bank of America workers allegedly had to wear badges that recorded everything the workers saw and heard, by analysing the speech of the person wearing the device, its volume and pitch, length of time span spent in a place, and mapping the daily paths enabled by beacons through the office space (Steele, 2020). Whilst all this might sound promising for delivering insights to the people management department, such invasive methods are often barely justified and might not be connected to the actual output that they are supposed to measure (purpose limitation). Asking the purpose question ‘I would like to analyse my employees’ productivity, so I need this data’ should prompt the answer ‘Why? On what basis? Why do you need to know about this, isn’t that a breach of their privacy?’. For some technologies, the purpose seems clear at first sight but with a closer look embodies salient privacy gaps: For example, smart jackets for first responders can be equipped with modules that monitor the heart rate, temperature, motion and geo location (Steele, 2020). Some modules, such as body cams are active beyond the emergency moment and allow to track the completion of tasks and monitor workflow. Due Diligence might show that this is not an appropriate use of data-driven monitoring when balancing the necessity for monitoring work performance when juxtaposed with the privacy intrusion.
In contrast, a geolocation tracking system that tracks a van delivering parcels to send notifications to customers when a parcel arrives seems less controversial at first sight. It can become controversial, however, if the van movement measuring can be used to instruct, e.g. when an employee is allowed to take toilet or lunch breaks (Schafheitle et al., 2020). Adding to this, in a workplace setting, privacy protection stands in tension with the power relationship between employer and employee, and the potential drawbacks of a consent-based approach, as discussed earlier – employees should not be forced to waive away their privacy rights in exchange for work (Ajunwa et al., 2017).
Anchoring Privacy Due Diligence in business practice – Reporting, evaluating, learning
To anchor Privacy Due Diligence into business practice, management needs to find a way to make the continuous reporting, evaluation and learning about the privacy impacts of its business matter within the company. For example, are there dedicated mechanisms for accountability and oversight for workplace monitoring in consultation with affected stakeholders? Diverse membership and composition in accountability governance structures with a clear, transparent process is key – with an emphasis on taking the view of potentially marginalized voices into account. Key elements could entail measures such as a policy commitment at the highest level, setting out a company’s privacy standards, awareness raising measures about data processing practices, or grievance mechanisms for employees and workers to speak out against intrusive measures. The UNGPs suggest operational grievance mechanisms to be accessible directly to stakeholders who may be adversely impacted. Anchoring Privacy Due Diligence in business practice should involve the integration of preventive and remedial mechanisms to act against adverse privacy impacts. The remedial rights to data subjects required by the GDPR can deliver complementarity (see GDPR, Arts. 15, 16, 17, 18, 20). A feedback loop should ensure learning from past mistakes and improve privacy conduct: The management needs to continuously evaluate accountability mechanisms, based on robust stakeholder engagement, rather than doing a static one-time assessment. The individuals conducting the review need to be empowered to change the data models and particular algorithmic decisions, and indeed do so on a regular basis, if needed. Through such structural measures, the management takes ownership of arising privacy dilemma, rather than ‘outsourcing’ it to the data protection officers. Such management ownership to deal with privacy issues pays justice to the increasing threats posed to privacy in a data-driven workplace.
Conclusion
In this article, we set out to explore the promises of a B&HR approach to tackle the privacy challenges brought on by the rise of algorithmic management. Our review of different legal and technical approaches tackling these challenges revealed a number of promising avenues, but also significant gaps. Despite its popularity in industry circles, for example, design-based approaches do not suffice to protect employees’ privacy at the workplace. Legal approaches also fall short, notably given the difficulty of applying jurisdiction-specific norms to a truly global phenomenon.
Privacy Due Diligence vows to play a fruitful role in closing these gaps. The balancing exercise between managerial prerogative and worker protection required by employment law and data protection law cannot be appropriately carried out through tech solutions alone – nor is
Interests need to be weighed before intrusive surveillance begins, and continue to be scrutinized over the course of the data life cycle. By building on existing models of corporate due diligence processes, in combination with key insights from data protection, legal frameworks and ethical considerations, the Privacy Due Diligence model establishes a company-wide process for responsible business conduct towards privacy at work.
Footnotes
Acknowledgements
We would like to thank Gabriel Kasper for useful discussions, and the Swiss National Science Foundation, and Economic and Social Research Council (UKRI) for their support.
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This article is partly funded by a grant of the Swiss National Science Foundation, NFP 75 “Big Brother or Big Data”? HR control practices and employee trust, and a grant by the Economic and Social Research Council (UKRI).
