Document Details

PicturesqueSpessartine4940

Uploaded by PicturesqueSpessartine4940

Ontario Tech University

Jennifer Laffier, Aalyia Rehman

Tags

deepfakes gender-based violence online abuse technology

Summary

This paper explores the harm women face in the digital age due to deepfake technologies. The use of deepfakes provides a new method for perpetrating gender-based violence and erodes women's autonomy. The study highlights the unique challenges faced by women online and the need for further research.

Full Transcript

Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Deepfakes and Harm to Women JENNIFER LAFFIER1, AALYIA REHMAN1 1...

Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Deepfakes and Harm to Women JENNIFER LAFFIER1, AALYIA REHMAN1 1 Ontario Tech University Abstract As deepfake technologies become more sophisticated and accessible to the broader online community, their use puts women participating in digital spaces at increased risk of experiencing violence online and abuse. In a ‘post-truth’ era, the ability to discern what is real and what is fake allows malevolent actors to manipulate public opinion or ruin the social reputation of individuals to wider audiences. While the scholarly research on the topic is sparse, this study explored the harm women have experienced in technology and deepfakes. Results of the study suggest that deepfakes are a relatively new method to deploy gender-based violence and erode women’s autonomy in their on-and-offline world. This study highlights the unique harms for women that are felt on both an individual and systemic level and the necessity for further inquiry into online harm through deepfakes and victims’ experiences. Keywords: Deepfakes, Women, Synthetic media, Image-based abuse, Harm INTRODUCTION O ver the past 20 years, gender-based violence has increased, affecting one in three women in their lifetime (The World Bank, 2019; World Health Organization, 2014). Violence against women refers to a myriad of gendered-based acts almost exclusively directed at women and girls that can cause direct or indirect harm over their lifespan. The World Health Organization (2014) describes violence against women as, An act of gender-based violence that results in, or is likely to result in, physical sexual, or psychological harm or suffering to women, including threats of such acts, coercion, or arbitrary deprivation of liberty, whether occurring in public or private life. (para. 1) While rates of police-reported gendered violence have decreased in Canada, nearly 70% of violence remains unreported to the police and women continue to live at greater risk for domestic violence, sexual assault, and harassment (Department of Justice Canada, 2019). Demonstrations of harm and abuse include physical aggression, psychological violence, neglect, and control over speech and actions. In severe instances, these acts may result in death (Dawson et al., 2020). Given the secondary status of women within families, states, and societies, it is not uncommon for gendered violence to be observed as ordinary or, in some cases, socially sanctioned (Kishore, 2011; -1- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 The World Bank, 2017). While subordination may not be the perpetrator’s motivation, it may result as an indirect consequence of their actions (Namy et al., 2017). These actions may remain concealed by victims in fear of being further attacked or violated. Thus, violence against women is not only rooted in sex inequalities present in one’s social world but also serves as a mechanism to sustain unequal powers and reinforce the status of vulnerability that fuels the abuse (Kaur & Garg, 2009; Namy et al., 2017; Vyas & Jansen, 2018). The advent of technology has allowed new mechanisms of gender-based violence to emerge that may allow users to inflict harm against women through online hate and cyberviolence. While individuals of all genders experience technology-facilitated abuse, the Canadian Women’s Foundation (2019) has indicated that women are at heightened risk of experiencing online violence, harassment, and sexualised abuse. Research on technology-facilitated abuse has shown that digital technologies are used routinely for online stalking, impersonation, harassment, and abuse, sometimes under a concealed identity (Duggan, 2017; Vogels, 2021). Scholars have raised concerns over the growing use of technology; a diverse range of sophisticated tools to abuse, control, and intimidate victims are widely accessible within the digital field – often free of cost (Albahar & Almalki, 2019). Progress in the field of artificial intelligence (AI), specifically machine learning, has allowed users to readily create altered media content online, amassing large and often global audiences in on-and-offline spaces (Kreps, 2020). Amongst these advancements is the emergence of deepfake technology, henceforth referred to as deepfakes. Broadly, deepfakes are machine learning-based software tools that produce realistic synthetic media content. Deepfakes replace the likeness of one individual with another, making the target individual appear to say things or perform actions that the source individual does (Albahar & Almalki, 2019; Nguyen et al., 2019). The generative adversarial networks (GANs) had been mainly restricted to the AI research community until 2017, when a Reddit user, using the anonymous handle Deepfakes, began posting digitally altered pornographic videos through a free, open-source machine learning software (Schwartz, 2018). The Reddit user built GANs to superimpose celebrities’ faces on adult film stars for pornographic purposes (Spivak, 2018). In 2023, Twitch streamer Brandon "Atrioc" Ewing's inadvertent disclosure of his previously viewed deepfake pornography of well-known female Twitch streamers further highlighted this problem (Mercante, 2023). This incident exposed the creators' exploitation of women's deepfakes for profit, as Ewing had inadvertently revealed the pay-to-view subscription site that hosted such content to his live audience (Mercante, 2023). In recent years, the use of deepfakes has risen in popularity due to the quality of synthetic media content and the ease of their applications for users with limited or sophisticated computer skills (Chesney & Citron, 2019a). Although deepfakes can be produced with the approval of the featured individuals, they are commonly created without consent. New developments have allowed for sophisticated deepfake videos to be produced with a still image, permitting any digital human trace to be fabricated and further threatening privacy (Galston, 2020). Moreover, in a post-truth era, deepfakes are often utilised with malicious intent to influence popular beliefs or damage the -2- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 reputation of individuals (Bayer et al., 2019; Westerlund, 2019). The sophistication of deepfakes is anticipated to pose a significant challenge in maintaining their credibility, which may result in negative consequences for vulnerable groups in both their online and offline lives (Strömbäck et al., 2020). Deepfakes can be a unique instrument for technology abuse and violence against women (Chesney & Citron, 2019a, 2019b). However, the impact on deepfake victims, particularly women, remains unclear and understudied (Chesney & Citron, 2019a). The lives of individuals in both the digital and physical realms are not perceived as distinct entities, but are rather interconnected, such that actions taken in one sphere can carry considerable repercussions in the other. The purpose of this review is to explore the harms experienced by women through technology and deepfakes. METHODOLOGY The researchers of this review examined the literature exploring the harms women have experienced through technology and deepfakes. A narrative review was conducted, as it allows for descriptive and critical investigation without simplifying results into standard metrics (Greenhalgh et al., 2018). This review involves emerging scholarly literature, news and media articles, and peer- reviewed studies from technological sectors that may provide insight into the characteristics of threats brought forth by digital technologies and deepfakes. A search was conducted for scholarly literature and peer-reviewed articles on Google Scholar and in the Omni database. Google search engine was used to scan news and web articles exploring cases of deepfake and deepfake technology. Searches included the following search terms (with Boolean operations): cyber, gendered image-based, video-based, media, abuse, harassment, threats, bullying, cyber, online, deepfakes, harm, threats, and synthetic. These search terms were informed by the focus of the research, which was to explore the harms experienced by women in technology and deepfakes. The researchers selected search terms that were relevant to this topic, and which could provide insight into the characteristics of threats brought forth by digital technologies and deepfakes. For example, the terms "gendered image-based" and "video-based" were included to capture the specific types of deepfake content that are often used to harm women. The terms "abuse," "harassment," "threats," and "bullying" were included to capture the various ways in which women can be harmed online. The terms "cyber," "online," and "media" were included to capture the different types of platforms and technologies that are used to spread harmful content. The terms "deepfakes," "harm," and "synthetic" were included to specifically capture content related to deepfakes and their potential harm to women. By combining these search terms using Boolean operators, the researchers were able to generate a comprehensive list of articles that addressed the specific issues of interest. Data extraction should be relevant to the review question, and thus results were further filtered according to the inclusion and exclusion criteria (Popay et al., 2006). The inclusion criteria for this study included literature and articles that focused on technology's role in online abuse and women, were written in English, and published between 2001-2022. Exclusion criteria for this study included: non-English language articles, articles not focused on technology's role in online abuse and women, articles published before 2001, and articles that are not accessible to the -3- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 researchers. Various terms were used to describe cyber abuse within the literature, including cyber- bullying, victimisation, and harassment (Grigg, 2010). Thus, cyber abuse was used as the superset term to reference intentional digital harm or threats or harassment or bullying or stalking through electronic and digital media through forms of media to which an individual feels uncomfortable, threatened, or abused. Eighty-five sources satisfied the inclusion and exclusion criteria, and ten examined specific cases of five deepfake victims. The data was collected and inputted in an Excel matrix to be further examined, coded for emergent themes, and analysed through qualitative content analysis to identify the types of abuse and threats posed by digital technologies and deepfakes. Qualitative content analysis can be used as a data reduction method to identify fundamental patterns and interpretations that arise from text or images, enabling researchers to gain a more scientific understanding of social reality (Zhang & Wildemuth, 2009). Qualitative content analysis aims to systematically convert larger quantities of data into highly organised and effective results (Erlingsson & Brysiewicz, 2017). Following Juntenen and Lehenkari (2021), the researchers of this study began by reading and revising literature from the data collection. Next, information relevant to the aim of this review was systematically extracted. Three analytical questions were developed as researchers read and discussed the literature together: 1. What type of harm was inflicted by technology? 2. How was this technology used, and what was the impact? 3. How has harm disrupted participatory on-and-offline behaviours? These questions aided in maintaining focus on the purpose of the review. Responses to the questions mentioned above were found across the literature and aided the synthesis of the research collection. The data collected in the Excel matrix were then reviewed for recurring inductive themes and presented in the following section. Lastly, interpretation was conducted through abstraction, based on discussions between the researchers on results related to other gendered cyber-abuse research. These interpretations are presented in the discussion section. Given the recent emergence of deepfake technologies and threats examining individual accounts, there is the likelihood that the researchers did not observe literature that may have fit the study's selection criteria. In the preliminary search, some publications provided fragmentary insights into the theme and questions guiding the analysis. Only literature and media articles primarily focusing on the gendered issues related to cyber abuse, deepfakes, their characteristics, and their harms were presented to preserve sources at a feasible quantity and increase the efficiency of the review process. Despite these limitations, the researchers are confident this review is representative of the literature exploring gendered cyber abuse in the digital age. LITERATURE REVIEW -4- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 This literature review examines the harm women have experienced through technology and deepfakes. This review begins with a brief introduction to cyber abuse of women. Next, literature exploring image-based sexual abuse in women is explored. Lastly, deepfakes and related harm are presented. Cyber Abuse in Women Willard (2005) defines abuse within cyberspace as the “intent to consistently harass another to harm via any electronic method, including social media, and includes defamation, public disclosure of private information, and intentional emotional distress” (p. 1). Citron (2014) states, “threats of violence, privacy invasions, reputation harming lies, calls for strangers to harm victims physically and technological attacks” (p. 3) are volitional, not simply innocuous commentary. Cyber abuse may manifest directly or indirectly (Langos, 2012). Direct cyber abuse may occur exclusively between the perpetrator and the victim. In contrast, indirect cyber abuse can occur when the perpetrator participates in the online space, posting content about the victim in various social formats that are accessible to larger audiences (Langos, 2012). Within these contexts, cyber- abuse components include harassment, repetition, power differential, and aggression/intention (Watts et al., 2017). However, some researchers do not consider the repetitive aspect as essential to the definition of cyber abuse (Grigg, 2010). Abuse within cyberspace features unique qualities that magnify inflicted harm and make it increasingly difficult to detect (Vaillancourt et al., 2017). The globalised nature of digital technologies allows malicious content to be created anonymously and circulated across the internet rapidly reaching global audiences (Vaillancourt et al., 2017). More than half of victims do not know the identity of their perpetrator, further contributing to increased fear, distress, and powerlessness in cyber victims (Dennehy et al., 2020; Dooley et al., 2009; Kowalski & Limber, 2007). Compared to men, women were three times more likely to face various forms of abuse online (i.e., online harassment, threats, sexualization, stalking) and more than two times more likely to experience severe forms of abuse (i.e., doxing, revenge porn, threats of physical harm; Veletsianos et al., 2018). Eckert (2018) found that young women were significantly affected, and the abuse had taken an emotionally significant toll. Barak (2005) states that “gender harassment in cyberspace is very common” (p. 78), through which cyber worlds become charged with diverse modes of abuse and violence against women (Poland, 2016; Veletsianos et al., 2018). Feminist scholars have argued that cyber abuse echoes broader cultural and societal understandings presented within one’s on-and-offline worlds regarding gender and status (Jane, 2016; Nadim & Fladmoe, 2019). Nadim and Fladmoe (2019) refer to gendered cyber abuse as offline misogyny evolving into social sectors. According to Jane (2014), gendered cyber abuse rests upon misogynistic practices that insist on women’s inferiority. Inflicted abuse of women is an attempt to reassert the ordained power structures of men when they feel threatened by the loss of their relative position (Bareket & Shnabel, 2020; Berdahl, 2007). Thus, the act of abuse against women is intended to restore and reinforce differences between gender (Nadim & Fladmoe, 2019). -5- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Sexism, misogyny, and abuse are apparent within cyberspace, and new media forms can further exacerbate and complicate outcomes surrounding abuse and violence (Abidin, 2013; Abrahams, 2017; Mantilla, 2013; Megarry, 2014; Piner, 2016; Rivers, 2017). Veletsianos et al. (2018) indicated that identifying as a woman online increases the risk of intentional abuse. The risk of cyber abuse is even greater for women in the public sphere or working within the digital space (Sobieraj, 2018). The most prominent forms of cyber abuse displayed and encountered by women were found to be slut-shaming, rape threats, death threats, and doxing (Gosse, 2022; Mantilla, 2013; Piner, 2016). In addition to the ill-psychological symptoms and distress brought forth by cyber abuse, women respond by self-censoring their presence online and withdrawing from the digital world (Citron, 2014; Vitak et al., 2017). Amnesty International (2017) has investigated the effects of abuse against women on social media platforms, specifically concerning how they act online after abusive incidents. Their study found that abuse created the silencing effect, whereby women feel discouraged when participating online. A report by Gender, Women and Democracy (2017) provides further illumination on the impacts silencing and censoring has on women victimised by online abuse, stating: By silencing and excluding the voices of women and other marginalized groups, online harassment fundamentally challenges both women’s political engagement and the integrity of the information space. … In these circumstances, women judge that the costs and danger of participation outweigh the benefits and withdraw from or choose not to enter the political arena at all. (p. 14) Vera-Gray (2017) argues that, with online usage by women and “the continual development of new tools for online social networking” (p. 5), more significant instances of cyber abuse can occur. Victims of cyber abuse are often left fearful of their continued online use with limited opportunities to pursue justice. For those wishing to pursue legal action, victims may lose time and pay from work, their reputation, and their employability due to online posts (Citron, 2014). The extent to which perpetrators persecute and damage women online infringe on their offline worlds, threatening the safety of these victims within physical, social, psychological, and financial sectors (Hess, 2014). Image-Based Sexual Abuse in Women The discourse on technology's role in cyber abuse has primarily concentrated on image- based sexual abuse (IBSA) or the exploitation of intimate content (Powell & Henry, 2017). IBSA refers to the act or threat of distributing intimate content without the consent of the targeted individual (Sparks, 2021). While anyone can be the target of IBSA, research suggests women are more likely to be victimised by IBSA, whereas men are more commonly the perpetrators (Powell et al., 2018). The most frequent forms of online creation and distribution are social media, email, and mobile messages (Silver et al., 2019). Henry and colleagues (2017) suggest that the growing concerns with IBSA and new technologies can be attributed to the ease with which harmful media -6- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 can be generated, shared, and downloaded, the permanence of generated content in online spaces, and the variety of accessible software and tools across multiple platforms. Efforts to normalise and popularise IBSA can be traced back to 2010 with the creation of IsAnyoneUp.com, a website that showcased stolen and hacked pornographic media (Henry & Flynn, 2019). Internet users were encouraged by the site's creator, Hunter Moore, and site viewers to anonymously submit non-consensual content of women, as well as share personal and identifying information (i.e., victim's social media accounts, name, occupation, and address; Henry & Flynn, 2019). Following an influx of media attention and legal actions by victims, Hunter Moore was prosecuted and found guilty of extortion and identity theft (Citron & Franks, 2014). Few large- scale research studies have examined the dissemination and outcomes of IBSA, yet more than 3000 websites are dedicated to the submission of non-consensual sexual media content (Powell & Henry, 2019). While these websites host images of women and sometimes men, images of women are more commonly viewed, commented on, and shared (Powell & Henry, 2019). The evidence on prevalence implies high levels of abuse and a gendered impact (Powell & Henry, 2019). With recent advancements in modern digital technologies, it is projected that non-consensual sharing and threats will continue to increase due to the feasibility of new assaults (Dodge & Johnstone, 2018). Specifically, AI tools and deepfake applications make it easier for perpetrators to anonymously engage in abusive non-consensual behaviour on various platforms without impunity for their actions (United States Department of Homeland Security, 2019; Westerlund, 2019). ISBA victims are reported to have significant feelings of helplessness, hopelessness, psychological distress, and social disruption (O'Malley & Holt, 2022). Bates (2017) reported that women victimised by IBSA suffered from trauma, post-traumatic syndrome disorder, anxiety, depression, feelings of loss of control, impaired sense of self, and difficulties developing trust in others. Women victims were more likely than men to report feelings of vulnerability and threats to overall safety, suggesting that some forms of IBSA are related to stalking and violence (Powell et al., 2018). Threats to safety are further exacerbated as victims are easily identifiable through image-based content and personal information posted online (Powell et al., 2018). Women victimised by IBSA can be contacted by on-and-offline actors who may harass, coerce, further distribute, or extort unless behavioural requests are met (Dunn, 2020). Cyber sextortion is a part of a larger continuum of IBSA, which broadly refers to making threats to share explicit content online to garner a victim's compliance with specific demands (O'Malley & Holt, 2020; Patchin & Hinduja, 2020). Demands are varied and are not always sexual (Liggett, 2019). Content is used to leverage control over victims, and the perpetrator's power to possibly harm victims is central to this crime (Liggett, 2019; Nadim & Fladmoe, 2019). Perpetrators of cyber sextortion garner the victim's images in various ways but most commonly online through manipulation or coercion (Liggett, 2019). Cyber sextortion is unique to other forms of abuse as it is ambiguous whether the images will be shared, which gives rise to fear and desperation in targeted victims (Nadim & Fladmoe, 2019). -7- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 While sextortion is used to control the target's behaviour, digital technologies and software have improved how the perpetrators commit these crimes (O'Malley & Holt, 2020). Motivations and methodology underpinning sextortion are associated with intimate partner violence, violence against women, cybercrime, and exploitation (O'Malley & Holt, 2020). Acar (2016) suggests three primary differences between cyber sextortion and other crimes: (a) cyber sextortion is exclusively online; (b) the perpetrator is in possession of threatening content; and (c) victims are forced into acts through threats to distribute content. Most IBSA victims prefer to stay unidentified, which results in a lack of public attention on issues related to cyber abuse (Wittes et al., 2016). Due to fear of personal safety and key figures being alerted, victims are forced to withdraw from online spaces, disclose the incident to employers, change contact information, and move physical locations and jobs (Powell et al., 2018). As a result, many victims have reported losing relationships with family, friends, and romantic partners (Henry et al., 2017). In some cases, victims may refuse to disclose threats due to shame and victim blaming, as women are commonly blamed for the content produced (Wolak et al., 2018). Hence, while the impact of media distribution is significant, its outcomes are emotionally harmful, particularly with the loss of close social networks. Deepfakes and Women The availability of technological devices has resulted in substantial growth of multimedia content in cyberspace, leading to increased content production and greater accessibility (Masood et al., 2021). Recently, digital impersonation has become more realistic and convincing (Chesney & Citron, 2019b). Jane (2016) argues that women have traditionally been the primary victims of nefarious adaptations in media and online content. While deepfake technology has been put to a wide range of benign uses, the technology has also been deployed to abuse, debase, and humiliate women victimised by IBSA (Westerlund, 2019). Widely overlooked in public debate, deepfakes offer a new way to exercise domination and power (Eaton & McGlynn, 2020; Jane, 2016). As effective weapons of abuse, deepfakes are most often deployed to compromise a woman's identity, inflict damage on her reputation, intimidate, and enforce silence (Dunn, 2020). By using a woman's face and body as a digital resource in the absence of consent, the distribution of deepfakes further presumes, breeds, and reinforces an online space in which women's images are forged for the overall pleasure of men who occupy these digital worlds (van der Nagel, 2020). Of the nearly 85,000 deepfakes circulating online, 96% depict non-consensual and sexually explicit content featuring women (Wang, 2019). For example, the online application, DeepNude, uses GANs to create natural looking depictions of nude women’s bodies (Cole, 2019). Over 104,000 fake nudes of women have been created and shared without the consent of victims, many of whom were under 18 years old (Burgess, 2020). Few women who have been targeted likely know that their deepfakes exist (Burgess, 2020). Often, there is very little that can be done even when deepfakes are known to victims (Ayyub, 2018; Maddocks, 2020). The prevalence of deepfakes and the limited recourse available to victims -8- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 is a complex issue, rooted in several factors. One primary challenge is the nascent nature of laws and regulations pertaining to deepfakes, which may not be sufficiently comprehensive or effective in providing adequate legal protection or avenues for redress (Chesney & Citron, 2019a). Moreover, the anonymity of deepfake creators, who often distribute manipulated content through anonymous or fake accounts, poses significant challenges in identifying and holding them accountable (Chesney & Citron, 2019a, 2019b; Citron, 2014). Compounding this difficulty is the persistent nature of deepfakes once they have been widely disseminated, as they can be readily replicated and circulated through multiple online channels, including social media and messaging applications, thereby exacerbating the challenges of mitigation and control (Chesney & Citron, 2019a; Citron, 2014). In 2018, pseudonymous users circulated a viral deepfake sex video featuring Rana Ayyub, an Indian investigative journalist, after she had made political comments regarding the child rape of a Kashmiri girl (Maddocks, 2020). Deepfake videos were distributed among various social media platforms, alongside her home address, phone number, and the phrase, “I am available” (Maddocks, 2020; Ayyub, 2018). Her Twitter and Facebook feed were overwhelmed with screenshots of the video, alongside death and rape threats (Ayyub, 2018). The video was further popularised when a member of the Bharatiya Janata Party quote tweeted the fake content and responded with "WTH [What the hell]....are you out of your mind????", giving the fake content further legitimacy to Twitter users (Maddocks, 2020). Despite calls for protection at national and international levels, Ayyub continued receiving obscene deepfake videos that attempted to slut shame her (Ayyub, 2018). In an article detailing her experiences, Ayyub (2018) stated that she now practices self-censorship, has withdrawn from various digital spaces, and fears people taking pictures of her as they may be used to create more deepfakes. Internet users can efficiently conceive deepfake videos subjecting targets to violent and humiliating sex acts, although not all deepfakes will be designed for the creator's sexual gratification (Chesney & Citron, 2019a). The techno-social landscape of online environments provides deepfakes with the capability to incite reputational damage (Morales, 2021). In 2019, a doctored video of US Speaker of the House Nancy Pelosi was circulated on social media, which made it appear as though she was drunk and slurring her words (Paris & Donovan, 2019). The video was shared widely, leading to concerns about Pelosi's health and fitness for office (Paris & Donovan, 2019). Chesney and Citron (2019a) note that deepfakes can be created merely to inflict fear, pain, and silence on women participating in public and private spaces. Deepfakes depict women partaking in non-consensual activities without ever having engaged in them, inciting abuse on those who do not want the attention and do not have adequate resources to address these fabrications (Chesney & Citron, 2019b). In turn, the effects of deepfakes on women can have social, professional, and personal ramifications (TRT World, 2021). For example, Noelle Martin's face and personal information were used to create deepfake pornographic content when she was a high school student (Paris & Donovan, 2019). She faced death threats, rape threats, extortion, stalking, and unwanted sexual advances as a result. Although she sought help from authorities and -9- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 government agencies, there was no resolution. The deepfake attacks had a lasting impact on her social life, law school prospects, and comfort in public settings (Paris & Donovan, 2019). Significant psychological distress has been observed when victims discover their media content has been deepfaked, reporting feelings of humiliation, violation, fear, helplessness, and powerlessness (Jaiman, 2020; Molloy, 2019). Research has found that victims of deepfakes and IBSA suffered from psycho-affective harm, online-and-offline harassment, mental health issues, suicidal ideation, damaged professional and character reputations, and personal and physical violation despite never participating in the observed actions (Bates, 2017; Davidson et al., 2019; HM Government, 2019; Sparks, 2021). In 2019, British poet and novelist Helen Mort had her images superimposed onto a series of deepfakes on pornographic websites (Mort, 2022). These deepfakes appeared to have shown Mort engaging in extreme acts of sexual violence (Mort, 2022). She reported recurrent nightmares, in which these images replayed themselves repeatedly, and was "incredibly anxious about leaving the house” (Royle, 2021, para. 4). The threats of sexual violence "literally, albeit not physically, penetrates the body" (Chesney & Citron, 2019a, p. 1773). Citron (2019) suggests that deepfake videos force women into non-consensual activities, dehumanising them to mere objects of gratification. For victims, these videos can transform virtual threats into terrifying and physical realities; and to online users within cyberspace, it suggests that women can be abused at whim (Chesney & Citron, 2019a). Strong stigma is attributed to women's bodies and actions, so deepfake victims are forced to suffer the collateral consequences and reputational damage invoked by their perpetrators (Chesney & Citron, 2019a). DISCUSSION AND RECOMMENDATIONS Cyber abuse is rooted in unequal powers of control, resources, and social rules that seek to preserve a man's dominance over their female counterpart (Ayyub, 2018; Eckert, 2018; Morales, 2021; Powell & Henry, 2017; Powell et al., 2018; Sobieraj, 2018; Vera-Gray, 2017; Vitak et al., 2017; Wolak et al., 2017). Despite their recent emergence, deepfakes have become an effective tool in inflicting gender-based cyber-abuse, eroding a women's autonomy through shame, humiliation, and extortion (Chesney & Citron, 2019a; Maddocks, 2020; Molloy, 2019; Morales, 2021; Powell et al., 2018; Westerlund, 2019). While the harm from deepfakes is related to those found in IBSA victims, this study has highlighted the unique harms for women that are felt both at individual and systemic levels. The gendered nature of deepfakes, especially when sexual in design, must not be minimised. The nature of deepfakes fosters opportunities for nefarious actions, as perpetrators can easily superimpose a woman's face onto another individual without the victim having ever committed the act depicted in the image or video (Ciancaglini et al., 2020; United States Department of Homeland Security, 2019; Westerlund, 2019). Often, women are shown engaged in humiliating acts threatening their reputation, safety, and offline worlds (O'Malley & Holt, 2020). The inflicted harm is magnified when deepfake content is anonymously created and shared across -10- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 various internet platforms, allowing audiences to re-watch, download, and distribute (Westerlund, 2019). Findings suggest that using deepfakes to perpetuate abuse follows the underlying themes of power, control, and sexuality (Bates, 2017; Eaton & McGlynn, 2020; Jane, 2016; Liggett, 2019; O'Malley & Holt, 2020). The creation and use of deepfakes can represent a perception of regaining control or what can be described as compensatory manhood acts (van der Nagel, 2020). On individual levels, targeted women of deepfakes may experience fear, psychological and social maladjustments, mental health issues, reputational sabotage, and face long-lasting social, professional, and economic consequences (Ayyub, 2018; Chesney & Citron, 2019b; Paris & Donovan, 2019; TRT World, 2021). Deepfake abuse may rupture their lives, their sense of self and identity and their abilities to form healthy relationships with themselves and others (Bates, 2017; Davidson et al., 2019; HM Government, 2019; Sparks, 2021). For some, the harm and abuse may seem relentless and constant, whereby victims live each day in fear of having their images or deepfakes (re)discovered. Once uploaded, deepfake content becomes extremely difficult to remove from the Internet permanently; the availability and dissemination of this content become another iteration of abuse (Chesney & Citron, 2019a). As this is out of the control of victims, they are rendered powerless, further contributing to and magnifying feelings of helplessness and hopelessness (Jaiman, 2020; Molloy, 2019; O'Malley & Holt, 2022). The use of deepfakes challenges traditional claims and knowledge hierarchies; it thrives on tactics that have been historically used to sabotage women's claims of truth and justice (Ayyub, 2018; Jane, 2016; Mort, 2022). As the impacts of deepfakes carry an insurmountable impact on their targets, women may involuntarily withdraw from their on-and-offline worlds and be silenced. The silencing effect observed by victims further contributes to a profound sense of seclusion and mistrust at micro-, meso-, and macro-levels (Ayyub, 2018; Chesney & Citron, 2019a, 2019b; Mort, 2022). This is concerning as online harms often target women finding their voices and building their identities online and offline. The lack of recognition and extent of deepfake harm may be a form of disquisitional injustice for victims (Nadim & Fladmoe, 2019). Deepfake abuse fortifies issues related to gender disparities and preserves discriminatory practices that erode women's sovereignty and human rights (Nadim & Fladmoe, 2019). Deepfakes operate to discredit victims of their truth and draw from correspondence and coherence theories of truth to advance their credibility (Westerlund, 2019). First, AI technologies make deepfakes look real and correspond with the viewer's observed reality (Chesney & Citron, 2019b; Citron, 2019; Ngyuen et al., 2020). Second, deepfakes cohere with assumptions about women's sexual availability (Maddocks, 2020; Wang, 2019). Within this framing, discussion about fake porn appears to concretise existing trust flows and heighten distrust towards women; thus, the expected cyber abuse provides accountability (Chesney & Citron, 2019a; Citron, 2019; Maddocks, 2020). When speaking out, women become vulnerable to further abuse that endangers their physical safety. This automated distribution and silencing contribute to the normalisation of this type of harm (Ayyub, 2018; Chesney & Citron, 2019a; Citron, 2019; -11- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Mort, 2022). Without victim testimony or critical news coverage, this newest incarnation of abuse could become collectively unintelligible to the public. This study highlights the need for further examination into cyber abuse with deepfakes, the lived experiences of victims on-and-offline, and how they may cope with such abuse (Henry & Powell, 2015). The impact of cyber abuse is both critical and extensive; interventions that aim to empower victims should be explored and included (Jeffrey et al., 2020). Furthermore, psycho- educational programs that aim to promote digital literacy and awareness of image-based and deepfake cyber abuse through various technological programs and mediums must be developed and promoted in educational spaces. REFERENCES Abidin, C. (2013). Cyber-BFFs*: Assessing women’s ‘perceived interconnectedness’ in Singapore’s commercial lifestyle blog industry *Best Friends Forever. Global Media Journal: Australian Edition, 7(1), 1-20. https://www.hca.westernsydney.edu.au/gmjau/wp-content/uploads/2013/09/GMJAU_V7- 1_Cyber-BFFs.pdf Abrahams, J. (2017, August 14). Everything you wanted to know about fourth wave feminism— but were afraid to ask. Prospect Magazine. https://www.prospectmagazine.co.uk/magazine/everything-wanted-know-fourth-wave- feminism Acar, K. V. (2016). Sexual extortion of children in cyberspace. International Journal of Cyber Criminology, 10(2). https://doi.org/10.5281/zenodo.163398 Albahar, M., & Almalki, J. (2019). Deepfakes: Threats and countermeasures systematic review. Journal of Theoretical and Applied Information Technology, 97(22), 3242-3250. http://www.jatit.org/volumes/Vol97No22/7Vol97No22.pdf Amnesty International. (2017, November 20). Amnesty reveals alarming impact of online abuse against women. https://www.amnesty.org/en/latest/press-release/2017/11/amnesty- reveals-alarming-impact-of-online-abuse-against-women/ Ayyub, R. (2018, November 21). I was the victim of a deepfake porn plot intended to silence me. HuffPost. https://www.huffingtonpost.co.uk/entry/deepfake- porn_uk_5bf2c126e4b0f32bd58ba316 Bates, S. (2017). Revenge porn and mental health: A qualitative analysis of the mental health effects of revenge porn on female survivors. Feminist Criminology, 12(1), 22-42. https://doi.org/10.1177/1557085116654565 -12- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Bareket, O., & Shnabel, N. (2020). Domination and objectification: Men’s motivation for dominance over women affects their tendency to sexually objectify women. Psychology of Women Quarterly, 44(1), 28-49. https://doi.org/10.1177/0361684319871913 Bayer, J., Bitiukova, N., Bard, P., Szakács, J., Alemanno, A., & Uszkiewicz, E. (2019). Disinformation and propaganda–impact on the functioning of the rule of law in the EU and its Member States. Policy Department for Citizens' Rights and Constitutional Affairs. https://www.europarl.europa.eu/RegData/etudes/STUD/2019/608864/IPOL_STU(2019)6 08864_EN.pdf Berdahl, J. L. (2007). Harassment based on sex: Protecting social status in the context of gender hierarchy. Academy of Management Review, 32(2), 641-658. https://doi.org/10.5465/amr.2007.24351879 Burgess, M. (2020, October 20). A deepfake porn bot is being used to abuse thousands of women. WIRED. https://www.wired.co.uk/article/telegram-deepfakes-deepnude-ai Canadian Women’s Foundation. (2019). The facts about online hate and cyberviolence against women and girls. https://canadianwomen.org/the-facts/online-hate-and-cyberviolence/ Chesney, R., & Citron, D. (2019a). Deep fakes: A looming challenge for privacy, democracy, and national security. California Law Review, 107, 1753-1820. https://doi.org/10.15779/Z38RV0D15J Chesney, R., & Citron, D. (2019b). Deepfakes and the new disinformation war: The coming age of post-truth geopolitics. Foreign Affairs, 98, 147. https://heinonline.org/HOL/LandingPage?handle=hein.journals/fora98&div=18&id=&pa ge= Ciancaglini, V., Gibson, C., Sancho, D., McCarthy, O., Eira, M., Amann, P., & Klayn, A. (2020). Malicious uses and abuses of artificial intelligence. Trend Micro Research. https://documents.trendmicro.com/assets/white_papers/wp-malicious-uses-and-abuses-of- artificial-intelligence.pdf Citron, D. K. (2014). Hate crimes in cyberspace. Harvard University Press. Citron, D. K., & Franks, M. A. (2014). Criminalizing revenge porn. Wake Forest Law Review, 49, 345-391. https://digitalcommons.law.umaryland.edu/fac_pubs/1420/ Cole, S. (2019). This horrifying app undresses a photo of any woman with a single click. Motherboard - VICE. https://www.vice.com/en/article/kzm59x/deepnude-app-creates- fake-nudes-of-any-woman -13- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Davidson, J., Livingstone, S., Jenkins, S., Gekoski, A., Choak, C., Ike, T., & Phillips, K. (2019). Adult online hate, harassment and abuse: A rapid evidence assessment. UK Council for Internet Safety. https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_ data/file/811450/Adult_Online_Harms_Report_2019.pdf Dawson, M., Sutton, D., Zecha, A., Boyd C., Johnson, A., & Mitchell, A. (2020). #Callitfemicide: Understanding sex/gender-related killings of women and girls in Canada, 2020. Canadian Femicide Observatory for Justice and Accountability. https://femicideincanada.ca/callitfemicide2020.pdf Dennehy, R., Meaney, S., Cronin, M., & Arensman, E. (2020). The psychosocial impacts of cybervictimisation and barriers to seeking social support: Young people’s perspectives. Children and Youth Services Review, 111, 104872. https://doi.org/10.1016/j.childyouth.2020.104872 Department of Justice Canada. (2019). JustFacts: Sexual assault. https://www.justice.gc.ca/eng/rp-pr/jr/jf-pf/2019/apr01.html Dodge, A., & Johnstone, E. (2018). Using fake video technology to perpetuate intimate partner abuse: Domestic violence advisory. Without My Consent. https://www.cpedv.org/sites/main/files/webform/deepfake_domestic_violence_advisory.p df Dooley, J. J., Pyżalski, J., & Cross, D. (2009). Cyberbullying versus face-to-face bullying: A theoretical and conceptual review. Zeitschrift für Psychologie/Journal of Psychology, 217(4), 182–188. https://doi.org/10.1027/0044-3409.217.4.182 Dunn, S. (2020, March 3). Women, not politicians, are targeted most often by deepfake videos. Centre for International Governance Innovation. https://www.cigionline.org/articles/women-not-politicians-are-targeted-most-often- deepfake-videos/ Duggans, M. (2017, July 11). The broader context of online harassment. Pew Research Center. https://www.pewresearch.org/internet/2017/07/11/the-broader-context-of-online- harassment/ Eaton, A. A., & McGlynn, C. (2020). The psychology of nonconsensual porn: Understanding and addressing a growing form of sexual violence. Policy Insights from the Behavioral and Brain Sciences, 7(2), 190-197. https://doi.org/10.1177/2372732220941534 -14- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Eckert, S. (2018). Fighting for recognition: Online abuse of women bloggers in Germany, Switzerland, the United Kingdom, and the United States. New Media & Society, 20(4), 1282-1302. https://doi.org/10.1177/1461444816688457 Erlingsson, C., & Brysiewicz, P. (2017). A hands-on guide to doing content analysis. African Journal of Emergency Medicine, 7(3), 93–99. https://doi.org/10.1016/j.afjem.2017.08.001 Galston, W. A. (2020, January 8). Is seeing still believing? The deepfake challenge to truth in politics. Brookings. https://www.brookings.edu/research/is-seeing-still-believing-the- deepfake-challenge-to-truth-in-politics/ Gender, Women and Democracy. (2017). #Notthecost: Stopping violence against women in politics. National Democratic Institute. https://www.ndi.org/sites/default/files/not-the- cost-program-guidance-final.pdf Grigg, D. W. (2010). Cyber-aggression: Definition and concept of cyberbullying. Journal of Psychologists and Counsellors in Schools, 20(2), https://doi.org/10.1375/ajgc.20.2.143 Greenhalgh, T., Thorne, S., & Malterud, K. (2018). Time to challenge the spurious hierarchy of systematic over narrative reviews? European Journal of Clinical Investigation, 48(6), e12931. https://doi.org/10.1111/eci.12931 Gosse, C. (2022). “Don’t take on the responsibility of somebody else’s fu**ed up behavior”: Responding to online abuse in the context of barriers to support. Canadian Journal of Law and Technology, 19(2), 249-275. https://digitalcommons.schulichlaw.dal.ca/cjlt/vol19/iss2/3/ Henry, N., & Flynn, A. (2019). Image-based sexual abuse: Online distribution channels and illicit communities of support. Violence Against Women, 25(16), 1932-1955. https://doi.org/10.1177/1077801219863881 Henry, N., Powell, A., & Flynn, A. (2017). Not just ‘revenge pornography’: Australians’ experiences of image-based abuse: A summary report. RMIT University. https://www.rmit.edu.au/content/dam/rmit/documents/college-of-design-and-social- context/schools/global-urban-and-social-studies/revenge_porn_report_2017.pdf HM Government. (2019). Online harms white paper. https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_ data/file/973939/Online_Harms_White_Paper_V2.pdf Hess, A. (2017, June 14). Why women aren’t welcome on the Internet. Pacific Standard. https://psmag.com/social-justice/women-arent-welcome-internet-72170 -15- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Jaiman, A. (2020, August 19). Deepfakes harms & threat modeling. Towards Data Science. https://towardsdatascience.com/deepfakes-harms-and-threat-modeling-c09cbe0b7883 Jane, E. A. (2014). ‘Back to the kitchen, cunt’: Speaking the unspeakable about online misogyny. Continuum, 28(4), 558-570. https://doi.org/10.1080/10304312.2014.924479 Jane, E. A. (2016). Misogyny online: A short (and brutish) history. Sage. Jeffrey, C., Peltier, C., & Vannest, K. (2020). The effects of an online psychoeducational workshop to decrease anxiety and increase empowerment in victims of trolling and cyberbullying. Journal of Online Learning Research, 6(3), 265-296. https://files.eric.ed.gov/fulltext/EJ1290782.pdf Juntunen, M., & Lehenkari, M. (2021). A narrative literature review process for an academic business research thesis. Studies in Higher Education, 46(2), 330-342. https://doi.org/10.1080/03075079.2019.1630813 Kaur, R., & Garg, S. (2008). Addressing domestic violence against women: An unfinished agenda. Indian Journal of Community Medicine, 33(2), 73–76. https://doi.org/10.4103/0970-0218.40871 Kishore, B. (2011). Understanding harassment at workplace through the traditional theory of violence: A case-based approach. Journal of Management & Public Policy, 3(1). http://jmpp.in/wp-content/uploads/2016/01/Bindhu-Kishore.pdf Kowalski, R. M., & Limber, S. P. (2007). Electronic bullying among middle school students. Journal of Adolescent Health, 41(6), S22-S30. https://doi.org/10.1016/j.jadohealth.2007.08.017 Kreps, S. (2020). The role of technology in online misinformation. Foreign Policy at Brookings. https://www.brookings.edu/wp-content/uploads/2020/06/The-role-of-technology-in- online-misinformation.pdf Langos, C. (2012). Cyberbullying: The challenge to define. Cyberpsychology, Behavior and Social Networking, 15(6), 285–289. https://doi.org/10.1089/cyber.2011.0588 Liggett, R. (2019). Exploring online sextortion. Family & Intimate Partner Violence Quarterly, 11(4), 45-56. Maddocks, S. (2020). ‘A deepfake porn plot intended to silence me’: Exploring continuities between pornographic and ‘political’ deep fakes. Porn Studies, 7(4), 415-423. https://doi.org/10.1080/23268743.2020.1757499 -16- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Mantilla, K. (2013). Gendertrolling: Misogyny adapts to new media. Feminist Studies, 39(2), 563-570. https://www.jstor.org/stable/23719068 Masood, M., Nawaz, M., Malik, K. M., Javed, A., Irtaza, A., & Malik, H. (2022). Deepfakes generation and detection: State-of-the-art, open challenges, countermeasures, and way forward. Applied Intelligence, 53, 3974-4026. https://doi.org/10.1007/s10489-022-03766- z Megarry, J. (2014). Online incivility or sexual harassment? Conceptualising women's experiences in the digital age. Women's Studies International Forum, 47(A), 46-55. https://doi.org/10.1016/j.wsif.2014.07.012 Mercante, A. (2023, January 31). Popular female Twitch streamers targeted in deepfake pornography scandal. Kotaku. https://kotaku.com/deepfake-atrioc-twitch-streamer- apology-legal-action-1850055762 Molloy, S. (2019, January 2). Blackmailers-for-hire ae weaponizing ‘deepfake’ revenge porn. New York Post. https://nypost.com/2019/01/02/blackmailers-for-hire-are-weaponizing- deepfake-revenge-porn/ Morales, C. (2021). Pennsylvania woman accused of using deepfake technology to harass cheerleaders. The New York Times. https://www.nytimes.com/2021/03/14/us/raffaela- spone-victory-vipers-deepfake.html Mort, H. (2022). ‘The images are seared onto my retinas - I felt ashamed’: The poet who was a victim of deepfake porn. The Telegraph. https://www.telegraph.co.uk/women/life/images-seared-onto-retinas-felt-ashamed-poet- victim-deepfake/ Nadim, M., & Fladmoe, A. (2021). Silencing women? Gender and online harassment. Social Science Computer Review, 39(2), 245-258. https://doi.org/10.1177/0894439319865518 Namy, S., Carlson, C., O'Hara, K., Nakuti, J., Bukuluki, P., Lwanyaaga, J., Namakula, S., Nanyunja, B., Wainberg, M. L., Naker, D., & Michau, L. (2017). Towards a feminist understanding of intersecting violence against women and children in the family. Social Science & Medicine, 184, 40–48. https://doi.org/10.1016/j.socscimed.2017.04.042 Nguyen, T. T., Nguyen, Q. V. H., Nguyen, D. T., Nguyen, D. T., Huynh-The, T., Nahavandi, S., Nguyen T. T., Pham, Q.-V., & Nguyen, C. M. (2022). Deep learning for deepfakes creation and detection: A survey. Computer Vision and Image Understanding, 223, 103525. https://arxiv.org/pdf/1909.11573.pdf -17- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 O’Malley, R. L., & Holt, K. M. (2022). Cyber sextortion: An exploratory analysis of different perpetrators engaging in a similar crime. Journal of Interpersonal Violence, 37(1–2), 258–283. https://doi.org/10.1177/0886260520909186 Paris, B., & Donovan, J. (2019). Deepfakes and cheap fakes. Data & Society. https://datasociety.net/wp-content/uploads/2019/09/DS_Deepfakes_Cheap_FakesFinal-1- 1.pdf Patchin, J. W., & Hinduja, S. (2020). Sextortion among adolescents: Results from a national survey of U.S. youth. Sexual Abuse, 32(1), 30-54. https://doi.org/10.1177/1079063218800469 Piner, C. (2016, July 28). Feminist writer Jessica Valenti takes a break from social media after threat against her daughter. Slate. https://slate.com/human-interest/2016/07/feminist- writer-jessica-valenti-takes-a-break-from-social-media-after-threat-against-her- daughter.html Poland, B. (2016). Haters: Harassment, abuse, and violence online. University of Nebraska Press. https://doi.org/10.2307/j.ctt1fq9wdp Popay, J., Roberts, H., Sowden, A., Petticrew, M., Arai, L., Rodgers, M., Britten, N., Roen, K., & Duffy, S. (2006). Guidance on the conduct of narrative synthesis in systematic reviews. A product from the ESRC methods programme. https://www.lancaster.ac.uk/media/lancaster-university/content- assets/documents/fhm/dhr/chir/NSsynthesisguidanceVersion1-April2006.pdf Powell, A., & Henry, N. (2017). Sexual violence in a digital age. Palgrave Macmillan. https://doi.org/10.1057/978-1-137-58047-4 Powell, A., & Henry, N. (2019). Technology-facilitated sexual violence victimization: Results from an online survey of Australian adults. Journal of Interpersonal Violence, 34(17), 3637-3665. https://doi.org/10.1177/0886260516672055 Powell, A., Henry, N., & Flynn, A. (2018). Image-based sexual abuse. In W. S. DeKeseredy & M. Dragiewicz (Eds.), Routledge handbook of critical criminology (pp. 305-315). Routledge. Rivers, N. (2017). Postfeminism(s) and the arrival of the fourth wave: Turning tides. Palgrave Macmillan. https://doi.org/10.1007/978-3-319-59812-3 Royle, S. (2021, January 6). Deepfake porn images still give me nightmares. BBC News. https://www.bbc.com/news/technology-55546372 -18- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Schwartz, O. (2018, November 12). You thought fake news was bad? Deep fakes are where truth goes to die. The Guardian. https://www.theguardian.com/technology/2018/nov/12/deep- fakes-fake-news-truth Silver, L., Huang, C., & Taylor, K. (2019). In emerging economies, smartphone and social media users have broader social networks. Pew Research Center. https://www.pewresearch.org/internet/2019/08/22/in-emerging-economies-smartphone- and-social-media-users-have-broader-social-networks/ Spivak, R. (2018). " Deepfakes": The newest way to commit one of the oldest crimes. Georgetown Law Technology Review, 3, 339. https://georgetownlawtechreview.org/wp-content/uploads/2019/05/3.1-Spivak-pp-339- 400.pdf Sobieraj, S. (2018). Bitch, slut, skank, cunt: Patterned resistance to women’s visibility in digital publics. Information, Communication & Society, 21(11), 1700-1714. https://doi.org/10.1080/1369118X.2017.1348535 Sparks, B. (2021). A snapshot of image-based sexual abuse (IBSA): Narrating a way forward. Sexuality Research and Social Policy, 19(2), 689-704. https://doi.org/10.1007/s13178- 021-00585-8 Strömbäck, J., Tsfati, Y., Boomgaarden, H., Damstra, A., Lindgren, E., Vliegenthart, R., & Lindholm, T. (2020). News media trust and its impact on media use: Toward a framework for future research. Annals of the International Communication Association, 44(2), 139-156. https://doi.org/10.1080/23808985.2020.1755338 The World Bank. (2017). World development report 2017: Governance and the law. https://www.worldbank.org/en/publication/wdr2017 The World Bank. (2019, September 25). Gender-based violence (violence against women and girls). https://www.worldbank.org/en/topic/socialsustainability/brief/violence-against- women-and-girls TRT World. (2021). Deepfakes and cheapfakes: The biggest threat is not what you think. https://www.trtworld.com/magazine/deepfakes-and-cheap-fakes-the-biggest-threat-is-not- what-you-think-43046 United States Department of Homeland Security. (2019). Increasing threats of deep fake identities. https://www.dhs.gov/sites/default/files/publications/increasing_threats_of_deepfake_iden tities_0.pdf -19- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Vaillancourt, T., Faris, R., & Mishna, F. (2017). Cyberbullying in children and youth: Implications for health and clinical practice. The Canadian Journal of Psychiatry, 62(6), 368-373. https://doi.org/10.1177/0706743716684791 van der Nagel, E. (2020). Verifying images: Deepfakes, control, and consent. Porn Studies, 7(4), 424-429. https://doi.org/10.1080/23268743.2020.1741434 Veletsianos, G., Houlden, S., Hodson, J., & Gosse, C. (2018). Women scholars’ experiences with online harassment and abuse: Self-protection, resistance, acceptance, and self-blame. New Media & Society, 20(12), 4689-4708. https://doi.org/10.1177/1461444818781324 Vera-Gray, F. (2017). ‘Talk about a cunt with too much idle time’: Trolling feminist research. Feminist Review, 115(1), 61-78. https://doi.org/10.1057/s41305-017-0038-y Vitak, J., Chadha, K., Steiner, L., & Ashktorab, Z. (2017). Identifying women's experiences with and strategies for mitigating negative effects of online harassment. In Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing (pp. 1231-1245). https://doi.org/10.1145/2998181.2998337 Vogels, E. A. (2021, January 13). The state of online harassment. Pew Research Center. https://www.pewresearch.org/internet/2021/01/13/the-state-of-online-harassment/ Vyas, S., & Jansen, H. A. (2018). Unequal power relations and partner violence against women in Tanzania: A cross-sectional analysis. BMC Women's Health, 18(1), 1-12. https://doi.org/10.1186/s12905-018-0675-0 Watts, L. K., Wagner, J., Velasquez, B., & Behrens, P. I. (2017). Cyberbullying in higher education: A literature review. Computers in Human Behavior, 69, 268-274. https://doi.org/10.1016/j.chb.2016.12.038 Wang, C. (2019, November 1). Deepfakes, revenge porn, and the impact on women. Forbes. https://www.forbes.com/sites/chenxiwang/2019/11/01/deepfakes-revenge-porn-and-the- impact-on-women/ Westerlund, M. (2019). The emergence of deepfake technology: A review. Technology Innovation Management Review, 9(11), 39-52. https://doi.org/10.22215/timreview/1282 Willard, N. E. (2005). Cyberbullying and cyberthreats: Responding to the challenge of online social aggression, threats, and distress. Research Press. Wittes, B., Poplin, C., Jurecic, Q., & Spera, C. (2016). Sextortion: Cybersecurity, teenagers, and remote sexual assault. Brookings. https://www.brookings.edu/research/sextortion- cybersecurity-teenagers-and-remote-sexual-assault/ -20- Journal of Digital Life and Learning 2023, Vol. 3, No. 1, 1-21 DOI: 10.51357/jdll.v3i1.218 Wolak, J., Finkelhor, D., Walsh, W., & Treitman, L. (2018). Sextortion of minors: Characteristics and dynamics. The Journal of Adolescent Health, 62(1), 72–79. https://doi.org/10.1016/j.jadohealth.2017.08.014 World Health Organization. (2014). Violence against women: Intimate partner and sexual violence against women: Intimate partner and sexual violence have serious short-and long-term physical, mental and sexual and reproductive health problems for survivors: Fact sheet (No. WHO/RHR/14.11). World Health Organization. https://apps.who.int/iris/handle/10665/112325 Zhang, Y., & Wildemuth, B. M. (2009). Qualitative analysis of content. In B. M. Wildemuth (Ed.), Applications of social research methods to questions in information and library science (2nd ed.; pp. 318-329). Libraries Unlimited. AUTHORS Dr. Jennifer Laffier is an assistant professor in the Faculty of Education at Ontario Tech University and a licensed Creative Arts Therapist and Psychotherapist with over 29 years of experience in the mental health field. She is the director of the Mental Health in the Digital Age Lab at Ontario Tech University. She researches and teaches human development and mental health in the digital age. Her research topics include wellness during times of crisis, digital wellness, digital self and emotional intelligence, cyber victimisation, digital technologies, psycho-education, creativity, and the arts. Aalyia Rehman graduated from the Life Science program at the University of Toronto in 2018, where she majored in Psychology and Health Studies. She completed her Master of Arts in Education under Dr. Jennifer Laffier and Dr. Janette Hughes's supervision, as well as the Bachelor of Education program at Ontario Tech University. She is currently enrolled in the Doctor of Education program at Ontario Tech University. Her research interests include child and youth mental health, identity development, digital technologies, art-making, and on-and- offline risk-and-resilience processes. Corresponding Author: Dr. Jennifer Laffier [email protected] -21-

Use Quizgecko on...
Browser
Browser