Abstract
Children and young people’s online privacy is increasingly challenged by the datafication of the digital world, and this is an increasingly important area of policy concern. Understanding what young people understand online privacy to be, and what they want done to protect it, is key to creating effective and rights-realizing policy responses. This article explores young people’s perceptions across four countries and finds they have nuanced understandings about online privacy and clear, robust ideas about how to improve it. Context mattered, and their online privacy concerns and ideal protections were often informed by their socio-political context and awareness of and trust in datafication.
Introduction
The experience of growing up today is imbued with digital technologies in ways that previous generations could have only imagined. Digital technologies provide access to vital education, socialization, participation, wellbeing, and entertainment opportunities. Young people’s rights are now realized — or violated — as much in the digital world as they are in the analogue world.
Young People’s Online Privacy and Datafication
This rapid adoption and increasing reliance on the online world have raised concerns about the effects on children and young people’s online privacy and digital experiences (Livingstone & Third, 2017; Mascheroni, 2018). Children and young people’s digital experiences are pervasively and aggressively recorded, tracked, aggregated, analyzed, and exploited by many commercial digital products (Zuboff, 2015). This is often described as the “datafication” of childhood (Barassi, 2020).
Datafication is an extremely common experience (Wang et al., 2022). Barassi (2020), for example, describes how children are now datafied even before they are born through the collection and harvesting of pre-natal data from parents; Mascheroni (2020) describes how this datafication continues right throughout childhood, through, for example, the use of connected baby devices and toys; and Jarke and Breiter (2019) describe how these practices continue in school through digital teaching tools.
Datafication can be a privacy harm in itself. It violates young people’s reasonable expectations to be let alone and right to be free from arbitrary interference with their privacy (United Nations [UN] General Assembly, 1989). But subsequent uses of this data create additional risks to children’s autonomy and safety, such as fuelling potentially harmful products (Twenge et al., 2020); enabling the delivery of potentially harmful content through content recommender algorithms (e.g., see the study by Harriger et al., 2022); and enabling inappropriate or potentially manipulative behavioral advertising (Nairn & Fine, 2008; Verdoodt et al., 2016), leading to harm through automated decision-making (Willson, 2019).
The invisibility and frequency of datafication practices—the gathering, computation, and sharing and selling people’s online data—may have become normalized and taken-for-granted conditions (Mascheroni & Siibak, 2021; van Dijck, 2014). The normalization of datafication could affect young people’s understandings and expectations of privacy. In a context where datafication en masse is widely expected and normalized, it becomes challenging to anticipate privacy as it has traditionally been understood.
Datafication may be especially normalized among children and young people for two reasons. First, by virtue of their emerging capacities, young people might not have critically reflected on digital privacy concerns; they may still be developing the “citizen agency” (Kennedy et al., 2015) necessary to do so. Further complicating the matter, today’s children and young people have grown up datafied from before birth (Barassi, 2020). They have not experienced a world without everyday datafication. Thus, young people’s perspectives about online privacy may be profoundly different to older people’s perspectives.
Three comparative factors surface in the literature as potentially important in understanding the normalization of datafication among the young. First, the level of datafication itself may be important, as young people’s experience with datafication may be culture-dependent. Second, young people’s awareness of datafication may differ depending on their levels of digital literacy and “citizen agency,” and finally, young people’s trust in datafication. Datafication can appear normalized under conditions of trust (van Dijck, 2014). Yet, there is limited understanding of how the cultural and socio-political context may influence the way young people perceive datafication.
Online Privacy Policy and Young People
A growing “techlash” has challenged trust in digital actors over recent years (Weiss-Blatt, 2021), encouraging a range of different social and political responses. Many of these attempts aim to counter datafication with a range of individually-focused privacy initiatives, such as advances in critical media curriculum (e.g., Bernd et al., 2015 or Zeichner, 2019) or attempts to improve young people’s online privacy skills (Wang et al., 2023). Many also attempt to mitigate the impact of datafication through regulation and legislation.
Regulations and legislations aiming to improve children’s online privacy are being implemented across multiple jurisdictions. Some are international initiatives, such as the UN’s General Comment (UN Committee on the Rights of the Child, 2021), while others are supra-national, such as the European Union (EU) Better internet for Kids + strategy (European Commission, 2022); national, such as Australia’s proposed Children’s Online Privacy Code (Australian Attorney General’s Office, 2023); or state-based, such as California’s Age Appropriate Design Code (State of California, 2022).
These recent policy attempts to improve children’s online privacy rest on a body of regulation already in place. This includes the EU’s GDPR-K (the special protections for children’s data within the General Data Protection Regulation), which affords European children some additional protections (European Union [EU], 2016), and the United States’ (US’s) Child Online Privacy Protection Act (COPPA) (US, 1998) which provides some protections for children under 13 years of age for example.
Many of these new regulations and legislation have only recently been implemented or are yet to be implemented, and evidence shows that compliance with older existing legislation has been patchy (Irish Council for Civil Liberties, 2023; Sirur et al., 2018), leaving children largely facing datafied childhoods.
Young People’s Perceptions and Expectations
Understanding young people’s perceptions of datafication, privacy and ideal privacy solutions is an important part of developing effective policy responses—especially understanding how these perceptions may vary across contexts (e.g., media contexts or national/(sub-)cultures) and what this means for domestic policy. Indeed, the engagement of young people has been a cornerstone in the development of various legislations related to children, such as Ireland (Data Protection Commission, 2019) and the UK (Information Commissioner’s Office [ICO] and Revealing Reality, 2019). Realizing young people’s rights to be heard in these emerging and imminent policy debates is also a democratic imperative across the globe Hartung (2020).
There is increased research interest related to how children perceive their data privacy online, and many studies have shown that children and young people understand and value their privacy as a concept (Livingstone et al., 2019; Third & Moody, 2021; Wang et al., 2022). Some studies find that while valued, young people may struggle to understand the “harms” from privacy incursions. For example, research has found that young people struggle to describe potential negative consequences that may emerge from personal data collection, use, or inference (Mandell & Farthing, 2023; Sefton-Green et al., 2022; Stoilova et al., 2020) and to conceptualize the potential risks around how data might shape their future experiences and behavior (Acker & Bowler, 2017; Pangrazio & Selwyn, 2017). Research has also found that young people’s perceptions of privacy are sensitive to the different actors involved in their privacy (Third & Moody, 2021)—be they commercial actors, civil society organizations, or public institutions, such as their schools—but that these understandings may be incomplete (Bowler et al., 2017; Kumar et al., 2017). (We note that this may also be true for adults in general.) Confounding this however, there is an acknowledged research gap when it comes to exploring young people’s perceptions about data collection and use in commercial contexts (Stoilova et al., 2021).
Much of the research exploring young people’s perspectives of privacy has been carried out with children from Western, Educated, Industrial, Rich, Democratic (WEIRD) countries. Studies about how datafication is perceived by children from non-WEIRD countries are scarce. This limitation could potentially limit the effectiveness of global regulatory initiatives. Recent research has shown that children from different cultural and social political contexts could exhibit different approaches toward data ownership and privacy incursions. For example, a recent study with UK and Chinese young children has identified that Chinese children are more likely to take a more pragmatic stance while children from the UK exhibited higher demand for data autonomy (Zhou et al., 2023). Similarly, international research exploring young people’s perceptions about their rights in relation to the digital environment noted regional variations, especially around privacy (Third & Moody, 2021). While privacy was widely valued, understandings differed. For example, “children in high-income countries such as Canada and New Zealand tended to demonstrate a higher level of privacy literacy than those in low-income countries, though not exclusively nor uniformly,” and higher levels of concerns around commercial actors’ use of personal information (Third & Moody, 2021).
Within WEIRD contexts too, children’s perceptions about online privacy have been found to vary. For example, a Europe-wide study found variations in children’s self-reported experiences of “data misuses,” such as “using personal information in a way children do not like,” across Europe (Smahel et al., 2020).
Research Questions
We recognize that having a deeper understanding about how children and young people recognize and problematise the implications associated with datafication is crucial for informing effective legislative and regulative agendas. Furthermore, these agendas need to be sensitive to domestic context where regulation may be implemented in practice, as well as the global context that unites many of the experiences of online privacy in principle. This research aims to address these challenges by exploring:
How young people define privacy;
To what extent young people problematise their online privacy, with a particular focus on datafication and commercial relationship, and;
What changes young people want to see to improve their online privacy.
This article makes a critical contribution to our understanding of these issues by taking a comparative perspective and focuses on identifying the importance of social and political context for us to consider datafication in shaping young people’s perspectives about online privacy and privacy protections.
Methods
This research used mixed methods to unpack young people’s perceptions of online privacy and adopted both design-based (Crippen & Brown, 2018) and action-research (McNiff, 2017) techniques. Young people’s perspectives on privacy and privacy protections were investigated in four real-world settings, Antigua & Barbuda, Australia, Ghana, and Slovenia, using focus groups and surveys, and participants (aged 10–18 years) then engaged with change-making policy discussions as part of the research process.
Sites were selected to allow for insightful comparative analysis and to maximize potential policy relevance in the action phase. Countries where relevant policy dialogues were active or emerging were long-listed, including, for example, countries anticipating a review of children’s privacy or related regulations (such as Australia, Slovenia via the EU, various US States, Canada, etc.) or emerging conversation among civil society and/or policy-makers about potential reforms (such as Antigua & Barbuda, Ghana, South Africa, Turkey, etc.). The long list was narrowed down to maintain geographic diversity; including countries from the global North and the global South, both hemispheres, different continents, and both WEIRD and non-WEIRD countries (Figure 1). This study received ethical approval from the Reset.Tech internal ethics review process on January 10, 2022.

Research sites selection; the socioeconomic status is based on OECD (2022).
A series of deliberative focus groups with young people aged 10–18 years and a survey of additional young people aged 10–18 years was held at each site. These were adapted to meet the needs of each context, including methods that might maximize relevance to domestic policy-makers (such as requiring large-scale research to inform policy development at an EU level, versus more facilitated focus groups helpful to the civil service in Antigua & Barbuda), and to accommodate different local restrictions (such as location availability, school days vs. holidays, availability of survey partners, etc.).
Embracing this variability was a principled decision. Learning from design-based research principles, this research aimed to focus on the requirements of each domestic context and the participants themselves (Crippen & Brown, 2018). This research did not attempt to replicate the same research methods across four sites, rather we set out instead to answer the same research questions in an adaptive and responsive way in four sites. The focus was less on generating “generalisations” but to instead focus on the “particular” (Bertelsen et al., 2018). This variability complicates methods and analysis, and this is addressed below. However, embracing variability allowed us to create meaningful engagement with young people in different circumstances, produce domestically relevant research, and maintain our ability to consistently address core research questions.
In total, in:
Describing the detailed session plans of each activity used in the focus groups is not possible within the space allowed, especially given the variation, so we have published them online for researchers to access. As an overview, each focus group followed four sequential phases (see Figure 2 also):
Exploring and discussing young people’s data footprint. Following an ice-breaker and warm up activity, each group collectively brainstormed the types of personal data they believed were collected about young people by existing online platforms. Participants were then encouraged to unpack who they believed these data were accessible to. Information about what data participants believed was being collected, and which actors accessed it was then used as a prompt to open up discussions about the levels of comfort or discomfort participants felt around online privacy.
Deliberations about the acceptability of these data flows, and discussions about whether young people felt these data processes and processors were trustworthy. Following this exploration session, each group developed a definition of privacy that related to their online experiences. Using this definition as a prompt, groups then discussed whether they felt they were private online or not, and if they trusted those involved in the data flow.
Developing a list of guiding principles for protecting young people’s privacy. As a whole group, participants discussed what they wanted to happen about their online privacy, addressing any gaps or issues identified in Phase 2. This broad discussion moved on to a process of attempting to state their specific desires in the form of ‘dos and don’ts.’ Groups were provided with a set of 10 or so ‘dos and don’ts’ prompt cards to start this process. Building on the principles of co-design (e.g., Bevan Jones et al., 2020), these prompt cards included suggestions that young people had made in the surveys and/or in other research sites. The groups then worked to flesh out their own extensive lists of ‘dos and don’ts’ (see examples in Figure 3). This first long list of ‘dos and don’ts’ was refined into a collated list of principles by the young people (or in Slovenia, the researchers).
A small action-research phase, where young participants were supported to share their thoughts and privacy principles with relevant decision-makers. Young people brainstormed who they might want to share their guiding principles with. These discussions were also deliberative and informed by the facilitators’ suggestions around policy possibilities in each location (i.e., if there was a regulatory review coming up, or who might be “responsible” for what change). Due to the pace of the Slovenian workshops, young people were simply offered the opportunity to connect with decision-makers via an upcoming consultation if they wanted. Young people were supported to undertake the actions they deliberated and decided on, leading to:
Antigua & Barbuda: participants gave a list of suggestions to teachers about things they believe young people should be taught about privacy and called for the publication of a “magazine” about the issue for decision-makers (which researchers then published and distributed);
Australia: participants wrote a submission to a government inquiry and gave oral evidence to a government department about Australia’s privacy law review;
Ghana: participants wrote a letter to multiple stakeholders they identified, sharing their guiding principles;
Slovenia: participants’ privacy principles were sent to relevant regulators and ministers and will be sent to the EU regulators inquiry when it opens.
Our qualitative analysis involved exploring the “artefacts” created by each group, including their definitions of privacy, their lists of dos and don’ts, and their privacy principles. We focused on analyzing these artifacts first because they provided consistency across the groups and represent a sort of “experiential statement” for the whole-of-group experience. For example, in one site, 15 different definitions of privacy were drafted by small groups (pairs or triads), but through discussion and deliberation, the whole group agreed on a single definition. This analysis focused on this agreed-upon definition. These artifacts were coded using a close reading method. Close reading involves sustained and careful reading to interpret the meaning of a text and generate any dialogic insights (Ruiz De Castilla, 2017). This approach allowed us to group concepts and uncover themes in these artifacts. These themes, uncovered from a close reading of the artifacts, are supplemented in this article by quotes from focus group transcripts and field notes from researchers.

A diagrammatic representation of the research phases.

A small group of young people presenting their list of ‘dos and don’ts’ to the whole group. (Photo credits to Antigua and Barbuda Ministry of Education Broadcast Unit).
Privacy as a Protective, Contextual, and Enabling Right
Participants held nuanced views about what online privacy meant to them, but four key ingredients emerged from their definitions and descriptions of privacy. First and most centrally, privacy involved the ability to conceal personal information. Young people spoke about wanting to “protect and conceal our information” (Antigua & Barbuda) or the ability to “protect and conceal our personal information” (Australia). This concept of concealing information and data was always the first idea to emerge among the four groups, aligning with Westin (1968)’s traditional conceptualization of privacy as “control over” information.
The nature of the personal information that young people spoke of wanting to conceal was diverse and highlighted an awareness of datafication as an issue. It included traditional identity markers such as name and date of birth, as well as categorical descriptors such as religion, gender, or your hometown, but also included types of data associated with commercial actors and online surveillance (e.g., metadata). Data about all the things “you have posted online,” photographs posted online, and especially geolocation data were noted as types of data that were personal. Geolocation data in particular were problematised across all sites. As one young Australian put it, “I understand that there is an economic side to it, but morally, you don’t need to know a 17-year-old’s location.” Other digital datasets were thought to be particularly deserving of privacy too, including all the “naughty things you have done” and details about the people “you don’t like,” noted in Ghana as particularly sensitive.
Second, privacy protects personal information from “others” who might want to interfere. Privacy was described as a protective factor that kept information and personal business away from potentially interfering “others.” The Antiguan and Barbudan group described privacy as “helping to create boundaries to protect us from interference in our lives” or as safeguarding information “that you don’t want others to get,” for example. The nature of the “others,” from whom personal information needed to be protected, was equally diverse. In Slovenia and Antigua & Barbuda, hackers were frequently discussed as a malicious “other” deserving of particular attention, while they were present but less of a focus in Australia and Ghana. Focus groups spoke about protecting data from states or state institutions like schools or education boards but also wanting interpersonal privacy from family and friends. Notably, commercial companies were also frequently discussed, with young people expressing a desire for privacy from individual companies like Google and Facebook to broader commercial concepts like “online game companies” or “social media” in general. The third-sector too was not immune, as one young person outlined, “(it’s) not just advertisers, but any companies. Even not-for-profits will get up in your face sometimes.”
Young people spoke about different “others” as warranting different levels of protection, particularly with different types of personal information. For example, in Ghana, as a quick activity to help unpack the concept of privacy, we ran a “straw poll” asking the young people who they thought should see the different types of personal information they identified in Phase 1 of the methods (brainstorming about data footprints). When it came to data about “all the things you have done online,” 33% of the group said that they would be okay for best friends to see that, 44% for online friends to see that, and only 22% said their parents or carers should see that. But they wanted complete protection from online companies; nobody wanted them to see that information. When it came to a specific piece of content they posted online, 22% thought it was okay for their best friends to see it, 67% for online friends to see it, but nobody wanted their parents, carers, or online companies to see it.
While this may seem somewhat confusing, especially to older generations that do not differentiate between online and analogue friends, or to online companies trying not to “see” content posted on their own platforms, it is loaded with comprehensible intent. Young people wanted differential protections from others and saw online companies as among the least trustworthy actors when it came to their privacy. This is much more in line with Nissenbaum (2004)’s contextual definition of privacy, which describes privacy as the appropriate flow of information as assessed by each individual based on their preferences and perceptions. Who the data is flowing to is important for young people when it comes to their preferences and perceptions.
Third, privacy was seen as an enabling factor that creates a sense of security, safety, and wellbeing. It was described in Antigua & Barbuda as “set(ing) boundaries making your life personal and more comfortable” or as “mak(ing) you feel a sense of safety and comfortability knowing your info is safe with you.” In Ghana, young people described feelings of being “worried” or “upset” because of privacy invasions. In Australia, young people described feeling “unsafe” about some common data practices, such as broadcasting live location details on social media platforms. One young person described it as “really bad, like, actually so bad. Like, I think that’s crazy” when we talked about broadcasting maps of where young people are. Privacy—when realized—helps to enable young people’s sense of security, safety, and wellbeing, and when violated, it creates a sense of ill-ease and concern.
Finally, privacy was described as a right or expectation in its own right. While privacy also helped secure young people’s wellbeing, it was also valued as a right or a legitimate expectation young people hold regardless. In Australia, privacy was described as “a right to protect and conceal”; in Ghana, young people felt they had “the right to keep certain information to themselves away from others”; and in Antigua & Barbuda, young people felt “no one should be allowed to look at (private information).” Curiously, while the idea that privacy was a right or legitimate expectation was uncontested in each country, the expectations about what the realization of this right looked like differed widely. We discuss the implications of this through a lens of datafication below.
The Normalization or Problematization of Datafication
Datafication is a risk to young people’s online privacy as they defined it in this research. Widespread datafication left young people unable to conceal personal information (e.g., geolocation data and what they posted online, as mentioned by young people from Ghana and Australia for example) from “others” who they wanted to conceal it from (e.g., commercial companies, as mentioned in Antigua & Barbuda and Australia for example). It was a frequent and arbitrary interference with young people’s expectations to be and feel private online. However, this does not necessarily mean that datafication would automatically be problematised or indeed normalized. Normalization refers to the process of general acceptance of certain social behavior which becomes a social norm (Petzold and Peter, 2015); the pervasive practice of collecting and processing personal data, particularly among young people, is normalized and becomes societal norm, whereas problematization pertains to the critical examination and questioning of the impact and implications of datafication on young people’s online privacy. Problematization, unlike normalization, suggests “transformative engagement” (Stengers, 2021), and therefore, conscious awareness and questioning of the ethical and social dimensions of datafication.
The literature suggests that the extent to which young people problematise/normalize datafication depends on (at least) three culturally located experiences: the level of datafication of young people, young people’s awareness of datafication, and young people’s trust in datafication.
First, it is worth recognizing that problematisation or normalization of datafication depends on the extent to which young people’s lives are indeed datafied. While datafication is rife, there are still many young people in communities or parts of the world with limited connectivity where datafication might not be such a significant feature of their lives (including in WEIRD countries, like outback Australia, and non-WEIRD countries, like parts of east Antigua). However, this research was not intended to investigate the perspectives of unconnected young people. The ice-break activity—asking young people to introduce themselves and their favorite app for example—doubled as a check to ensure that all participants were indeed connected.
For connected young people, who this research focussed on, understanding the extent to which datafication is problematized or normalized requires unpacking two socially located experiences. Young people’s awareness of datafication is crucial; where young people are unaware that they are datafied (such as when they are unaware that apps track their location data), the very invisibility of the practice makes it difficult to problematize and allows the practice to operate as an unspoken social norm. Where there is awareness of datafication, trust—the second socially located practice—comes into play. Datafication only becomes normalized under conditions of trust (van Dijck, 2014). Without trust in those actors involved in datafication, the practice would be strongly rejected. Both trust and awareness are explored below.
Awareness of Datafication
Awareness of datafication was mixed in these research sites, with a split between WEIRD and non-WEIRD research sites. In Australia and Slovenia, young participants appeared to be largely aware of the practice of datafication, that is, all the young people we spoke to were aware that datafication happened, although with different levels of understanding about the scale of the practice, that is, some thought that very little data were collected while others understood the potential scale of the practice.
In Ghana and Antigua & Barbuda, while many young people were deeply aware of the datafication of their lives and the scale of it, some were not. A description of a quick, pre-lunch activity in Antigua & Barbuda highlights this perfectly. Trying to fill in a 10-minute gap after a discussion about “data footprints,” before lunch arrived, we ran a quick quiz & movement activity to activate the “hungry teen energy.” Giving each young person a post-it note, we asked them to guess how many data points they thought online companies held about each young person and to get up and try and stick them on a whiteboard in order of lowest to highest guess. The breadth of guesses gave us an unexpected glimpse into the participant’s awareness of datafication. Ten young people made guesses on the very low end, around 50 or 100 data points, while eight young people added ginormous estimates of 400 million or 10 billion to the board, and other guesses were scattered in between. We did not discuss what a “datapoint” was to understand what young people meant by 50 or 10 billion, but it is not the numbers themselves that matter here. Two guesses stood out to the research team. Even after the activities around data footprints, with small group discussions among peers and many young people checking which app was collecting what for an entire morning, two young people guessed zero. A descriptive field note taken after the session highlights this: I watched one young woman who had written “0” on her post-it note walk up to the board in the maelstrom of activity, ready to stick her guess up. She stopped three steps in front of the board, and watched her peers sticking up guesses that said thousands, millions or billions. She looked down at her post-it note again, then up at the board, watching the post-it note guesses spread further and further to the right hand side (where the high numbers were being stuck up). She looked at her own post-it note again and froze. I could see her trying to work out why her peers thought that anyone held any data about young people, she was just completely unable to reconcile this with her guess of “0.”
This young person was not disconnected from the digital world and had spoken about multiple online services and apps they used regularly, from free games to Google products. It appears that for some young people, the datafication of their lives truly was invisible to them. This could suggest it was an unobserved norm for them. But this was not true for all young people in Antigua & Barbuda, and many young people were deeply aware. We asked one young Antiguan if he was aware of the practice, for example, and he said “a couple of years ago, I read the (Terms of Service) for the first time. Yeah, you can’t go back after that.” To most young people, datafication was not invisible, and they were aware to various extents that it was happening.
Trust in Those Who Datafy
Awareness of datafication does not mean that the practice is accepted and normalized, nor challenged and problematised. van Dijck (2014) describes the importance of trust in those who collect and use data as a prerequisite for the normalization of datafication. van Dijck (2014) posits that for the questionable ideological grounds for datafication to become normalized, people must implicitly trust in the agents that collect, analyze, and share their data.
In the surveys and polls, we asked young people very broadly if they “trusted” that digital products and services, like apps and websites, handled their personal data carefully (Figure 4). What we found appeared to be broadly speaking equal levels of trust and distrust, and a lot of young people who felt ambivalent. The four-country average for those who trusted that their data was handled carefully was 26%, compared to 27% who distrusted and 43% who neither trusted nor distrusted.

The percentage of young people who reported trusting or distrusting that their data were handled carefully by digital products and services.
We asked some young Australians to help us understand what this might mean, and their descriptions suggested a strong relationship between normalization and trust as van Dijck (2014) described. But this was not necessarily trust in the “dictionary” sense of the word. Young people spoke about trusting that companies were handling their data carefully because they had to. Trust did not seem to be unquestionably given, nor earned, but offered as a sort of “deal with the devil.” One young woman said “if I was given that question [in the survey] I’d say ‘Sweet Jesus, like no, I don’t think I trust them with my privacy.’ But at the same time, you know, I’m on every social media that there is, so . . ..” Another young man outlined “because you rely on it. So it’s not even about whether or not you can, you don’t really have the choice to trust it or not. You just have to use it because everyone else is on it. It isn’t about whether or not you believe in your privacy.”
Trust as a function of necessity appeared in almost all our research sites. One young Antiguan surveyed explained that they had trust “because they are things I use on daily,” and one Ghanaian youngster said they trusted their data were being used carefully on platforms “because it is what we use in our everyday life” but added that “I also do not trust it because it is easy for hackers to access your personal information.” Normalizing the questionable way the digital world exploits data requires trust (van Dijck, 2014), but for some young people, this appears to have been a forced normalization. They appear to be actively deciding to trust those involved in data processing simply because they have to.
This idea of “having to trust” those involved in data processing resulted in a lot of discussions around trust as a personal balancing act, where young people felt they were personally balancing their right to privacy with their right to access the digital world. One young man in Australia described this as basically a trade off. And the whole world is full of different perspectives and views, just like the internet. So if we have a look at the two, it’s big. Your privacy for something else, or that fun for just a few, like a little bit of information. But I think what makes most people willing to share that information is they think, “Oh, who would be interested in me, like, I’m just one drop out of the ocean. You know, there’s millions of other people who do the same thing.” So they’re pretty trusting.
The commerce involved in this trade-off was a significant consideration when it came to trusting or distrusting digital actors. Young people in each research site described the monetisation of their data as the opposite of their privacy. For example, one young Ghanaian spoke about their ambivalence in trusting that their data were handled carefully by explaining “I trust them because they keep my data safe, and I mostly have control over it. But, I don’t trust them because of the way they gather my data and somethings sell it to companies like Google or Facebook to push ads to me.” An Australian who did not trust that their data was handled carefully explained that this was “because big companies only care about money and will do anything to get more money, including disrespecting privacy,” and a young Antiguan & Barbudan said they distrusted digital services with their personal data because “I believe that they will sell my account.” Wherever commercialisation of data was described, it appeared to be a cause of distrust. While the commercialisation–distrust nexus was universal across all four research sites, the frequency with which it was mentioned varied. It was frequently discussed in Australia but rarely in Antigua & Barbuda for example.
As interesting as these variations and nuances are, there is a more obvious conclusion that warrants stating explicitly. Only a minority of young people unquestionably trust that their data are handled carefully, ranging from 21% in Ghana, 27% in Australia and Slovenia, to 31% in Antigua & Barbuda.
Trust is a multifaceted and complex concept intricately linked with various contextual factors, such as the actors involved, the tasks being performed, and the cultural or digital literacy background of the stakeholders (Sztompka, 1999). Our findings underscore the intricate nature of trust development among young people, highlighting the significant roles played by familiarity, knowledge about actors, and past experiences in shaping trust dynamics (Hameleers and van der Meer, 2023). It is crucial to acknowledge that young individuals may sometimes feel compelled to “express” or “normalize” trust, even when they harbor reservations about digital products, driven by a perceived lack of viable alternatives (Bryce and Fraser, 2014; Ghaiumy Anaraky et al., 2021). Prior research has consistently shown the pivotal role of transparency as a key factor influencing user trust development (Kizilcec, 2016; Schmidt et al., 2020). While our young participants exhibited a heightened awareness of datafication practices, it is essential to recognize that they are not consistently provided with transparent information about how these technologies operate Wang et al. (2023). This lack of transparency, coupled with a lack of support for empowerment, poses significant challenges to their informed decision-making and raises concerns about the actual capability of trust development by young people in the digital landscape.
Privacy Principles
Each group developed a list of “dos and don’ts” about what young people thought should and should not happen with their data. These lists were created with deliberation, and participants were informed of survey data, as well as principles that young people in other research sites had created. In Antigua & Barbuda, Australia, and Ghana, young people themselves analyzed, collated, and prioritized their list of “dos and don’ts” into a final set of “privacy principles,” but in Slovenia, the analysis and collation was done by researchers because of the compressed session time. Groups came up with between 9 and 14 distinct principles, sometimes with subcategories or multiple principles that could be meaningfully combined. (This is another example that variation may create difficulties in sensemaking. But even if the variable length of their lists makes comparison difficult, the order of their priorities provides crucial indications of their importance to the young people.)
The final lists of principles varied across each site but bore many commonalities. Our analysis suggests that there were 15 key concepts across all the groups, with only three that were entirely unique to one site (all three in Slovenia, which followed a compressed method). These include:
Requirements around transparency and meaningful consent around data collection and use. For example, renaming cookies as “data grabbers” so young people are not “pushed or tricked” into handing over data, or that “apps must not process or ‘eavesdrop’ on the content of messages exchanged through them.”
Providing young users with more control of their data. Suggestions were sometimes broad, like “let us be more controllable of our data,” which should be specific, like “young people should have the right to request it be deleted.”
Requirements for stronger data security. Young people wanted to “make security stronger for young people’s data” for example.
Requirements for data minimization or collecting less data, such as requirements to “only collect the information about young people that they actually really need to run their app.” This especially applied to location data, with suggestions to prevent products “Collect as much location data about young people as they want.”
Requiring data to be processed only where it is in young people’s best interests, which was described as a meta-principle framing the other concepts. For example, suggestions that young people’s data are “only collected and used in ways that advance their best interests, but this needs specifics about what it means. Young people need to decide what young people’s best interests are.”
Requirements to prevent excessive data sharing or selling. Popular principles include “for data to not be resold” or “not be sold or traded to other companies.”
Restricting or ending targeted advertising to young people. Suggestions ranged from blanket calls to “stop advertising” to more targeted calls for “don’t have advertising turned on by default for young people.”
Requirements around data retention and obligations to delete data when it is not needed. Suggestions ranged from requirements for data only to “be kept for as long as is it needed only” to requirements that “when we log out, all our data is deleted.”
Requiring companies to provide adequate help and support. For example, it was suggested that “companies that collect and use young people’s data should be accountable to them. If something goes wrong, it should be the company’s responsibility to provide help and support and fix it.”
Requirements for stronger content moderation or mitigation of algorithmic recommendation of harmful content. These ranged from calls for content controls, such as “banning the posting of videos that encourage children to take up dangerous challenges,” to calls for controls on how harmful content is shared, such as not “encourag(ing) harmful content in ‘for you’ feeds.”
Improving the use of data for good, or using it in ways to benefit young people. This ranged from simple calls for things like “free games” to the more transactional “If you take my data, at least make the app better” or a request “to use my data to do things that would benefit me, and let me know.”
Requirements to provide better education for young people. There were calls for young people to be “supported and educated about privacy, their rights and risks,” or for the “safe use of the internet (to become) a school subject.”
Restrictions or limitations for young people, such as time limits or age limits to own devices and so on. These ranged from soft requests like “For there to be a day without phones,” to stronger suggestions “for under 16s to not use the internet” (Slovenia only).
Requirements for parental supervision, or additional parental supervision when using the digital world (Slovenia only).
Others. The scale of the Slovenian workshops generated some suggestions that fit no other categories, from clarion calls like “for children to be in a decision-making role instead of experts and politicians” to specific issues such as “for it to be easier to change the phone number associated with your account” (Slovenia only).
Figure 5 outlines the concepts that young people embedded into their privacy principles, in order of priority. Below, we expand on requirements about behavioral advertising to showcase divergence and “providing users with more control” to highlight convergence.

The 15 broad categories of principles listed by young people in each country, in order of priority, where 1 is the most important priority identified by the young people.
Privacy and Advertising
It was widely understood in all four sites that behavioral advertising—involving tracking and collecting young people’s personal data and online activities, and using this to target them with advertising—is the business behind much of the digital world. It was also a widely held belief that this was not great, with descriptions ranging from being annoying to an outright violation of young people’s privacy. Despite this convergence, how advertising was discussed by young people across four sites was particularly insightful and suggests a strong role for political mediation of young people’s expectations.
Taking Australia as an example, there was a discernible “chilling effect” where young people limited their principles to what they believed was possible within the current policy climate. In their submission to a Senate inquiry, the young participants opened up with “fundamentally, young people do not want their data used to sell them things.” They quickly went on to moderate this statement by calling for advertising to “not be turned on by default for young people. Young people should be able to opt-in and choose to have advertising overall, and also be able to choose if they want their data used to personalise these ads or not.” This was not because they felt young people wanted or needed a choice about receiving advertisements, but because, in their words, they wanted to be “realistic” in their discussions with policymakers. They “support(ed) a ban on behavioural advertising, but (are) aware it might be unpopular or difficult to implement.”
We unpacked this desire to be “realistic” with the group, as it came up multiple times. During the deliberations around privacy principles, young people expressed genuine concerns that what they really wanted might not be “too much” to ask for. There appeared to be a belief that young people needed to be sources of profit for technology companies to access the digital world. Notes and transcripts from the discussions included multiple comments like “but they won’t do that, so don’t add it (to the list),” “but that won’t make them a profit,” “if they don’t profit, they won’t do it.” Or as one young person we interviewed said: We can’t expect the government to, you know, make (digital products and services) default to “no, you can’t share my data.” . . . Because like that wouldn’t get passed, like no matter what. Because it’s just like, it’s really unrealistic for them to be able to do that and then make profit at the same time.
Uniquely, these Australian young people had experienced firsthand what can happen when tech companies go head-to-head with governments. In 2021, Facebook responded to the Australian Government’s attempts to implement regulation by withdrawing news from its Australian platform, causing “havoc” during bushfire season and the COVID vaccine rollout (Hagey et al., 2022). These young people had also grown up with effective privacy or data-protection regulations; Australia’s Privacy Act was passed in 1988, 5 years before the internet was made available to the public and does not protect metadata for example. Growing up in this climate could potentially explain why these young people were inclined to “chill” their expectations, in contract to the principles of young people in Slovenia, Ghana, and Antigua & Barbuda, all of whom called for a more straightforward ban on targeted advertising.
In a welcome but tragic twist, the tempering dose of realism that these young people swallowed was not necessary. Five months after they developed their principles, a review into the Privacy Act—which these young people ultimately gave evidence to—called for a complete prohibition on direct marketing to a child under 18 years of age in Australia, with exceptions only if the personal information used for the direct marketing was collected directly from the child, and the marketing is in the child’s best interests. It appears that for these young people, datafication of the digital world was so normalized and unavoidable that they could not even imagine making a plea for what turned out to be thoroughly possible policy reform.
Control of Your Own Data
For young people in each site, having control over their own data was a central expectation and demand that emerged across privacy principles. While control is implicit in almost all the privacy principles young people developed (from data retention to calling for more support where things go wrong), explicit calls for control took many forms. For example, broader calls like “let us be more controllable of our data” were made in Antigua to specific calls like the “option to set permission settings for posts” from Slovenia.
Ghana presented a unique version of this principle, which called for a complete reframing in the relationship between digital platforms and services and young people. In a move that would restore control over data to young people, they asked for companies to consider themselves as mere “caretakers” of young people’s data where they use it, rather than as data owners who could collect it, use it, and trade it however they wanted.
In Ghana, the young people were acutely alive to the relationship between data and control and discussed with researchers how regaining control of their data could be an act of agency. They were aware that data could be used to wielding power over young people, with one young person noting: assuming a child’s information since they were 8 years’ old was online, by 12 years, advertisers and data collectors would have enough information to about the person, to determine likes, dislikes, social affiliations and much more which gives them the ability to determine how best to influence the child towards their desired outcome.
The ability for data to push children to “their desired outcome”—emphasis added—is loaded. “Their” means advertisers and data collectors, or people generally involved in the commercial aspects of the online world. This was not what the young people in Ghana wanted. They wanted control over their own data themselves. In the language of power (Lukes, 2021), they wanted to curtail data’s power over them (currently wielded by advertisers and data collectors) and replace it with power to control their own data (putting young people in the driving seat).
To them, this was not as an emancipatory act of freedom rather it was described as an act of fairness. Fairness in the relationship between young people and digital products and services, as they described it, meant young people having power and control over their own information. In a culture—like many others—where young people are often “seen and not heard,” control over data was talked about with researchers as an act of fairness for the younger generations.
Conclusions
The young people involved in this research held nuanced and sophisticated understandings of what privacy online should mean for them. They were largely aware of the ways the online world datafied them, and what this meant for their privacy. However, given the prevalence of the practice, datafication did not lead to a straightforward rejection or strong plea for greater privacy. Rather, it appeared to produce difficulties in imagining viable or realistic alternatives when it came to their privacy online. Regardless, at each site, young people were able to articulate a strong set of privacy principles they believed should govern the online world.
Comparatively speaking, these principles showed extensive convergence. Only three principles emerged that were completed unique to one country, and all in Slovenia, including (a) restrictions or limitations for young people, (b) requirements for parental supervision, and (c) a broad category of other disconnected suggestions, which are expected given the sample size. This may reflect the collective experience these young people shared in a “global” online environment. However, even within this convergence, unique socially located differences were apparent.
Understanding young people’s perspectives and desires for privacy protections is an increasingly important task. As policy makers around the world activate to protect young people’s privacy, young people’s voices must be heard, and these can help to inform more effective policy outcomes. This research suggests that young people can and do want to be engaged in these deliberations and highlights the potential for and importance of local deliberations to inform these policy deliberations.
Footnotes
Acknowledgements
We thank all the young people who participated in this research, whose generosity and expertise made this research possible.
Declaration of Conflicting Interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was supported by the Internet Society Foundation.
