A chapter I co-wrote with the excellent Veli-Matti Karhulahti on Finnish DIY “fuck games” is freshly out in Perspectives on the European Videogame, edited by Víctor Navarro-Remesal and Óliver Pérez-Latorre for Amsterdam University Press. Our chapter, “Finnish Fuck Games: A Lost Historical Footnote” examines the games Strip-tease Ventti, Helttaa Helmaan, Bepa Quest, and Koulu3, all designed in the 1980s and 1990s, and their young male homosocial contexts of creation and use. This is one of the collaborations emerging from the recently finished research project, Sexuality and Play in Media Culture (2017-2021) that I was the PI of.
Category Archives: academic pleasures
Issue 2/2021 of WestEnd is out, with the special theme of “Pornografie. (Un-)Sittlichkeit und Geschlecht” edited by Juliane Rebentisch and Kerstin Stakemeier and including my piece, “Pornokreuzzüge und emotionale Plattformpolitik”. I remain very enthusiastic about my year publishing in a language I don’t speak in any meaningful way. Here, the abstract and whole text in English.
Porn crusades and affective platform politics
Online pornography forms a ubiquitous part of online culture even as its ready and abundant availability continues to fuel social concerns and campaigns aimed at curbing it. Focusing on the recent campaign of US-based journalist, Nicholas Kristof, against a leading video aggregator site, Pornhub, this article examines the logic and politics involved in “the deplatforming of sex”—that is, the expansive removal of nudity and sexual content from online platforms. It argues that Kristof’s campaign, in targeting online payment system providers in particular, represents a shift in anti-pornography activism towards infrastructural interventions aiming to delimit porn sites’ techno-material conditions of operation. As such, it speaks of broader platform politics where regulatory practices specific to the US impact the sexual expression of users on a global scale.
In December 2020, Pornhub, the globally leading porn video aggregator site, suspended access to nine million videos, these amounting to the majority of its content. The action was in response to public attention caused by a The New York Times exposé opinion article by the Pulitzer-winning journalist Nicholas Kristof, titled “Children of Pornhub”. Setting out to reveal the dark side of the platform as one “infested with rape videos”. it dramatically claimed that the site monetizes on “child rapes, revenge pornography, spy cam videos of women showering, racist and misogynist content, and footage of women being asphyxiated in plastic bags” (Kristof 2020). In addition to dwelling on the stories of abused, tortured and trafficked women, Kristof reiterated some well-known problems in the operating principles of Pornhub, a platform long critiqued for building its business model on piracy, having lax moderation practices and responding slowly to complaints on illegal content and requests to remove it (e.g., Auerbach 2014; Grant 2020). When voiced by sex workers, these critiques have had little effect. The case was different with the NYT article.
Starting from Kristof’s lobbying against Pornbub and other platforms trading in commercial sex, this article explores their logic, goals and ramifications within the broader context of “the deplatforming of sex” (Molldrem 2018), namely the increasingly vigilant removal of nudity and sexual content from online platforms. This involves what David M. Halperin (2017, 3) calls “the war on sex”: a cumulative effect of many independent initiatives targeting sex, and especially forms of sex arousing “disapproval on moral, aesthetic, political, or religious grounds” in the United States (see also Halperin 2017, 6; Race 2018, 172–173).
Out with it
Porn video aggregator sites broadly emulate the operating principles of YouTube which has, since 2005, largely defined the principles of online video sharing. With the exception of XVideos and xHamster, the most popular of these (e.g., Pornhub, Redtube, YouPorn) are owned by the same company, MindGeek. Writing on YouTube, Tarleton Gillespie (2010, 352) maps out the notion of platform in four different senses of the term: “computational, something to build upon and innovate from; political, a place from which to speak and be heard; figurative, in that the opportunity is an abstract promise as much as a practical one; and architectural, in that YouTube is designed as an open-armed, egalitarian facilitation of expression”.
In a political and figurative sense, to have a platform means being heard and seen, having the possibility of gaining an audience and potentially impacting culture and society. Conversely, to deplatform means to silence by removing someone’s or something’s access to a channel through which they can be heard and gain an audience. In networked media, deplatforming occurs on diverse levels: by removing user accounts or entire groups (Rogers 2020), by banning content categories (Pilipets and Paasonen 2020) and enforcing such bans through moderation, or by impacting the technical or economic infrastructures necessary for the platform’s operability. The deplatforming of sex in social media operates through in-platform laws such as community standards. In a dramatic example, Tumblr decided to ban nudity and sexual depiction in 2018, these having previously formed a large part of its content (Cho 2018). The platform was popular among sexual and gender nonconforming communities who lost access to networks, resources and archives built over a decade, further adding to their marginalisation (Byron 2019; Molldrem 2019). The third form of deplatforming targeting the infrastructural conditions of an application, site or service – deplatforming in a computational sense – became a topic of debate following Amazon’s decision to remove the Parler app favoured by Trump supporters from its web hosting service in January 2021.
The solutions that Kristof suggested for fixing Pornhub’s problems were partly same to those that sex worker activists had long been calling for: that, in order to curb piratism and illegal content, only verified users should be able to post videos, downloads should be prohibited and content moderation and reporting practices improved. To use Gillespie’s terms, these suggestions cut through Pornuhb as a political, figurative and architectural platform. Pornhub claims to have complied with all these modifications (see Pornhub 2020). Kristof however further suggested cutting the platform’s ties to payment infrastructures: “I don’t see why search engines, banks or credit card companies should bolster a company that monetizes sexual assaults on children or unconscious women”. Visa and Mastercard, alarmed by the negative publicity and looming PR damage, moved quickly to severe their ties with the platform, so that it cannot currently accept credit cards. In 16 April, 2021, Kristof continued his project with another NYT opinion article, “Why Do We Let Corporations Profit from Rape Videos?”, targeting XVideos and calling for both credit card companies and search engines to cut it off.
As Sarah T. Roberts (2019) details in her ethnographic study of commercial content moderation, social media platforms would be rife with materials of torture, both animal and human, were not armies of low-paid employees tasked with weeding it out. For while much of visual content moderation happens through automated, algorithmic means, distinctions pertaining to authenticity and context are hard for machines to make (the format of video posing its specific sets of challenges). The work within the “cesspool” of social media is largely concerned with the brand management of these platforms, seldom comes with sufficient mental health support and entails notable emotional and psychological stress (Roberts 2019, 116–123, 151–154). Kristof (2020) however framed the problem of traumatizing content as specific to Pornhub so that its moderators became both victims of MindGeek and villains facilitating the sexual abuse of children.
The shortage of content moderation resources on porn aggregator sites is an acute concern, yet similar work at Google and Facebook has been discussed as no less “soul-crushing” in making employees “soak up the worst of humanity” (Chen 2014). ISIS beheading videos, documentations of sexual and other violence and footage shot by white extremists on shooting sprees have all been available on mainstream social media before being flagged or removed by commercial content moderators (Gillespie 2018, 9; Parks 2019). Facebook (2021) reports taking action on five million incidents of child nudity and exploitation in the first three months of 2021, catching 98,9% of the content before it was reported by users. As porn video aggregator sites’ principles of operation are to a large extent similar to those of social media platforms, their problems in content moderation are also similar, even as their content policies drastically differ.
The spaces for sexual display and communication have been growing increasingly narrow on social media since the passing of the 2018 “Allow States and Victims to Fight Online Sex Trafficking Act” (FOSTA) and “Stop Enabling Sex Traffickers Act” (SESTA) bills in US senate with overwhelming bipartisan support. As exceptions to Section 230 of the United States Communication Decency Act which protected online services from liability for the content posted by users, FOSTA-SESTA has redefined online platforms as publishers responsible for content aiding sexual solicitation. This has resulted in broad removal of content connected to commercial sex that has nothing to do with trafficking, mainly since no distinction is drawn between consensual and non-consensual sex work (Reynolds 2020): consequently, online advertisements for sexual services have disappeared, as have social media groups and threads for sex workers sharing tips on filtering clients, sexual health resources and managing their careers independently (Blunt and Wolf 2020; Paasonen et al. 2019, 133; Tripp 2020). Since US-based social platforms are globally used, the legislation has broad resonances, also in countries where sex work is legal.
These transformations have impacted content moderation well beyond the realm of commercial sex. As pre-emptive measures, social media companies have tightened content policies since the liabilities of weeding out too little by far overshadow the commercial benefits involved in hosting sexual content – this having always been difficult to monetize as advertisers are unwilling to place their ads next to depictions of nudity and sex (Pilipets and Paasonen 2020). Facebook and Instagram have opted for horizontal content bans pertaining to nudity, sexual display and solicitation, deplatforming sex up to the visibility of female nipples and nude buttocks, users inquiring after each other’s interest in having sex, and the uses of eggplant and peach emojis in a sexual context (for a longer discussion on deplatforming of sex in social media, see Paasonen 2021).
Meanwhile, FOSTA-SESTA is argued to have little impact on curbing trafficking while curbing sex worker’s access to information resources and failing to protect them (Tripp 2020). In her critique, Lura Chamberlain (2019, 2206) defines the law as “deeply flawed” in that it “threats to criminalize significant categories of protected speech have already led to a documented chilling effect on speech due to its gross misunderstanding of the interaction between sex work and sex trafficking.” Long in the planning, FOSTA-SESTA built on a 2017 ban on commercial sex advertising targeting Backpage.com (Goldman 2018). Summing up the impact of FOSTA-SESTA and the closing of down Backpage’s sex advertising, Hacking//Hustling sex worker community report points out that there is no evidence of it having “done anything to prevent sexual labor exploitation. Our research shows that this law has actually put people in more precarious financial situations that actually make individuals more vulnerable to trafficking, as well as decreasing access to previously established channels of communication used to protect sex workers against violence.” (Blunt and Wolf 2020, 35.)
Targeting Backpage in a 2017 NYT opinion piece, Kristof called it “the odious website where most American victims of human trafficking are sold” and argued that SESTA “was crafted exceedingly narrowly to target only those intentionally engaged in trafficking children” and hence, contrary to criticism, has nothing to do with narrowing the freedom of expression online, or with limiting the rights of sex workers. In the light of empirical evidence, this is patently untrue. As in his previous work and later campaigns against Pornhub and XVideos, Kristof’s liberally conflated all commercial sex work with forced and involuntary labour, used the sexual abuse of minors as his affective rhetorical focus and accused tech companies such as Google for being allies of sex traffickers to undermine their critiques of FOSTA-SESTA (e.g., Barnes 2019; see also Kristof 2009). As a rhetorical strategy, the conflation of sex work and trafficking has been highly influential for two decades, cutting through and bringing together the Christian right, abolitionist feminists and governmental actors (Weitzer 2007, 449). This strategy remains knowingly blind to the presence and agency of sex workers as others than victims of abuse, delimiting their possibilities to impact policy, as well as obscuring their different agendas, positions and experiences, both locally and internationally (Bernstein 2019).
Given the impact of US internet governance on users across the globe, initiatives such as FOSTA-SESTA go well beyond regional concerns. This also means that campaigns such as Kristof’s, basically consisting of opinion articles published in one US newspaper and a flow of tweets aiming to impact policy, matter internationally since these policies alter the terms and conditions of online platforms used by billions of people around the world. What may seem – or in fact, be – a moral panic in the US can impact the livelihood of people doing online sex work in Germany, just as it can impact the ways in which social media users can, or cannot, exchange sexual content ranging from sex education resources to historical photographs or titillating selfies, or sexually relate to one another on these platforms.
The association of porn with violence against women has, of course, been key to feminist initiatives that have, since the 1970s, framed pornography as both a symbol and documentation of male violence justifying the sexual objectification, dehumanization and subjugation of women (e.g., Griffin 1981; Dworkin 1989; Kappeler 1986; Long 2012). This line of argumentation has drawn causal connections between porn and sexual violence, as in Robin Morgan’s famous 1974 slogan, “porn is the theory, rape is the practice”. Largely originating from the US, anti-pornography feminism continues to have international influence.
Premised on porn production and consumption being harmful to women both individually and collectively, anti-porn feminism has focused on critiques of patriarchal power relations in the framework of binary gender, so that forms of pornography not including women or made by women for other women, by people not conforming to a gender binary, or not simply fitting the patterns of critique, are either absent or interpreted as offering further proof of patriarchal sexual politics. In its focus on women’s abuse by men, this line of argumentation operates with a deeply hetenormative logic which, while seldom acknowledged, becomes generalised as a framework for sexual fantasies and the work of porn (Thompson 2015; Paasonen et al. 2020, 40–41). Like Kristof’s campaigns, anti-pornography feminism paints a binary universe, both moral and gendered, where porn and sex work lack female agency and help to bolster male hegemony. There is no room for considerations of porn as a site of sexual experimentation or expression, or for sexual desires and fantasies of the unruly, queer and kinky kind. This speaks of the persistent presence of sexual hierarchies of the kind that Gayle Rubin (1989, 281) identified at the early stages of the feminist sex wars as separating “good sex” (heterosexual, married, monogamous, procreative, non-commercial, private, vanilla) from the bad (homosexual, unmarried, promiscuous, non-procreative, commercial, public and kinky).
Since anti-porn feminism’s critique is categorical, it approaches the genre as a singular entity with aligning intentions, aesthetics, politics and economies, firmly placing it in the realm of “bad sex”. As the genre becomes thus homogenised, its inner diversity and fragmentation evaporates from view so that it is impossible to grasp the work or products of contemporary porn – and, consequently, to understand much of what is being discussed (see Paasonen 2011). These critiques also tend to be disinterested in the views of women working in porn, unless they are speaking against the industry, hence excluding their concerns connected to sexual health, income or control over work conditions. The notion of the porn industry, largely coined in the 1980s and 1990s, fails to describe contemporary forms of production involving studios of various sizes, amateurs and semi-amateurs, independent producers and animators aiming to make their products seen on online platforms even as the dominance of video aggregator sites, combined with the invisibility of porn work on social media, means that such visibility is by no means easy to achieve. In her critique of Kristof’s Pornhub campaign, journalist Melissa Gira Grant (2020) points out how,
“For years, porn performers have tried to draw attention to the exploitation at the heart of the tubesite business model—YouTube clones, which now dominate an online porn ecosystem that, not long ago and like much of online media, once offered independent creators more control over their work. Those days are all but over in porn, and the large companies behind websites like Pornhub have drained money out of independent porn, not just by pirating their work but by nearly monopolizing the business. Pornhub’s parent company owns porn-production companies, too, ones that some performers who might otherwise speak out also need to rely on for work. In turn, that has resulted in less work, lower wages, and less control for performers. In monopolies, particularly in industries that operate with little independent oversight and a nonunion labor force, abuse proliferates.”
There is much to critique in Pornhub and MindGeek’s impact on porn work and production culture that has contributed to something of a collapse in the studio system allowing for longer contracts and an emergence of a gig economy of financial precarity while also narrowing down the financial viability of independent producers and distributors (Berg 2021; Paasonen et al. 2019, 44, 59–60). As Grant argues, campaigns for credit card companies and PayPal to halt payments to Pornhub nevertheless do little to amend the situation. Rather, they hurt sex workers and other content producers who depend on the platform for their income. Attacks on Pornhub as a sex trafficking hub are also missing the point in that not only do users post child abuse material on mainstream social media platforms but the majority of reported child sexual abuse material is shared in either the dark web or through encrypted messaging apps like WhatsApp (e.g., Burgess 2021; Kleinman 2021). There are, however, no campaigns to date targeting the Facebook-owned WhatsApp used by two billion people as a child sexual trafficking platform.
A liberal journalist, Kristof is careful to distinguish his critiques of rape videos from arguments of porn being an engine of rape culture, targeting platforms for sharing illegal content instead. In other words, by framing his project as not being about pornography but about rape, he rhetorically detaches it from those aiming to ban pornography in more categorical terms. At the same time, his Pornhub article promotes the efforts of Traffickinghub, a campaign run by Exodus Cry, a religious right organization aiming at “the abolition of the sex trade, including prostitution and porn, by means of the criminal law” (Grant 2020). Kristof’s then appears to have intimate kinship networked anti-pornography initiatives bringing together conservative groups resisting sex education, LGBTQ+ rights, reproductive and abortion rights (Grant 2020). Framing these organizations as anti-trafficking (Weitzer 2007) has helped to neutralized them so that they can receive funding for their diverse actions: arguing to protect the rights of women with anti-trafficking campaigns, they in fact campaign against women’s sexual rights, operating internationally. Meanwhile, Kristof’s journalistic status gives him an aura of objectivity of the kind inaccessible to activists labelled either radical feminist or conservative Christian. With some two million Twitter followers, his political platforms are notable: some of his platform status is evident in the Pornhub article taking up the entire front page of NYT Sunday Review section.
During the Reagan presidency, high-profile radical feminists Andrea Dworkin and Catharine MacKinnon aligned their initiatives with those of Christian conservative coalitions, even as their gendered and sexual politics were fully incompatible (see Vance 1997). A similar alignment is taking place between anti-pornography feminist initiatives and conservative lobbying groups – as well as in Kristof’s alignment with Exodus Cry. Attending to these connections, commentators have been quick to identify Kristof’s attack on Pornhub as a moral crusade (Grant 2000).
His argumentation makes use of visceral examples – such as Backpage advertising “a 13-year-old whose pimp had tattooed his name on her eyelids” (Kristof 2017) – and excerpts from the survivors of abuse. These operate as textual equivalents of anti-pornography feminist slide shows aiming at negative affective responses for a political effect (Gentile 2010, 85–92). Feminist anti-pornography activism has, both historically and within the contemporary, made use of negative affect in arguing for the nefarious impact of porn, associating it with feelings of hurt, sadness, anger, frustration, sorrow, fear and nausea (e.g., Griffin 1981; Dworkin 1989; 2000). This was particularly true with Dworkin whose work is undergoing something of a revival with the publishing of Last Days at Hot Slit (2019), a collection of her writings. Within the cultural context of #MeToo and the fight over reproductive rights in the US, many find her emotional prose, fury at the way things are, and the firmness of her political stance resonant (Paasonen et al. 2020, 46). At the same time, her clarity of argumentation comes with ample simplification and sexual normativity within the framework of binary gender that fits ill with considerations of sexual and gender diversity.
Accounts of negative affect connected to porn, in the variations it has taken from the 1970s to the current day, from feminist texts to Christian fundamentalist ones and to Kristof’s reporting, anchor political argumentation in gut reactions in order to bestow on them a visceral sense of authenticity and acuteness. They operate affectively by putting “the body behind our words” so that words can become “something more than mere words” (Miller 1997, 181). Activism building on the power of feeling (anger, sorrow, disgust) together can be powerful in bridging the personally felt with the collective and the societal (Protevi 2009). At the same time, these forms of affective address work to efface diversity within the aesthetics, sexual routines, bodies, genders, sexualities, economies, politics and ethics connected to porn so as to frame it as singular entity and object assumedly evoking uniform responses. In other words, not only do the cultural objects and practices of pornography become homogenized but so do the presumed ways of experiencing them. All this sets clear limits to how porn can be approached, conceptualized, analysed and known – which, of course, is what these campaigns aim at.
What’s in a word?
The boundaries of porn as a genre have never been set, and they have grown ever more ephemeral in the course of digital and networked production, distribution and consumption involving a plethora of actors, governance practices, financial and political interests. Porn is an umbrella term for practices, aesthetics and economies that may share little similarity with one another across space and time. In order for analyses and critiques of porn to be efficient, these need to be specific, founded in empirical evidence and attuned to the distinctions among the actors and materials addressed.
As we have been communicating through networked means, often unable to connect flesh-to-flesh during the COVID-19 pandemic, the importance of mediated forms of sexual relating for wellbeing has grown strikingly evident. It is crucial not to understand such relating in narrow terms as an extension of extant intimate relationships: it is also a realm of sexual play, experimentation and pleasure involving the (mediated) bodies of virtual strangers through webcams, OnlyFans accounts, porn clips, hookup app profiles, and beyond. As sites for play for some, these are sites for work for others in ways blurring any clear divisions between the two notions (Paasonen 2018, 31). In any case, they are detached from reproductive goals and attuned toward discoveries in what one can sexually enjoy, like and prefer and, consequently, what or who one’s sexual self may be. Such “unpredicted forms of experience” (Warner 2000, 185) can alter one’s understanding of sexuality and desire as “a new sensation, an unusual mood, a previously inconceivable way of relating” comes about (Race 2009, 186). Unexpected incidents happen in encounters with other people, just as they do with mediated images and sounds – porn included. To consider porn in this vein as affording potentially startling and possibly encounters opens up alternative ways of thinking about its affective power and potential.
Within the current cultural conjuncture, it is however also necessary to reconsider what it means to label cultural objects as pornographic to start with: this necessity is pertinent in terms of securing spaces for sexual expression and relating through networked means. People creating sexual media do not necessarily see it as porn even as it can hold great personal importance as a means of exploration and reflection. Sexual depiction and visibility are key to the making and maintenance of gender and sexual nonconforming communities, just as it can be key to self-discovery and social relating (Molldrem 2018). At the same time, social media platforms classify all displays of nudity as offensive and categorically remove them, so as to protect their own brands and the commercial interests of advertisers (Tiidenberg and van der Nagel 2020, 46–47). As sexual content is being zoned to specific sites and as both feminist and Christian right organizations are pushing for closing down these sites’ access to payment systems, it remains crucial to ask whose interests are being served, how, and in whose name.
Carefully contextual analyses of sexual media production, distribution and use are necessary for shifting the foci of public debate so that sexual rights on online platforms are not merely understood in the negative sense as freedoms from (being harassed and abused) but equally as positive freedoms to (express and enjoy sexuality), without the one overweighing or cancelling out the other (Spišák et al. 2021). Such a step also necessitates acknowledging, and working through, the complexities in how people of diverse gender identifications and sexual orientations make use of sexual media and how online platforms – political, figurative, computational and architectural – and their governance shape the ways that sexual sociability can take. Simplified moralistic and ideological takes on what sexual exchanges and bodily displays mean or who produces them do much more harm than good when it comes to the sexual rights of self-expression, pleasure and knowledge.
Auerbach, David. 2014. Vampire Porn: MindGeek is a Cautionary Tale of Consolidating Production and Distribution in a Single, Monopolistic Owner. Slate, 23 October, http://www.slate.com/articles/technology/technology/2014/10/mindgeek_porn_monopoly_its_dominance_is_a_cautionary_tale_for_other_industries.html.
Barnes, Leslie. 2019. False Representations of Sex Workers. AUReporter 49 (3): https://reporter.anu.edu.au/false-representation-sex-workers.
Berg, Heather. 2021. Porn Work: Sex, Labor, and Late Capitalism. Chapel Hill: The University of North Carolina Press.
Bernstein, Elizabeth. 2019. Brokered subjects: Sex, Trafficking, and the Politics of Freedom. Chicago: University of Chicago Press.
Blunt, Danielle and Ariel Wolf. Erased: The Impact of FOSTA-SESTA & The Removal of Backpage. https://hackinghustling.org/erased-the-impact-of-fosta-sesta-2020/
Burgess, Matt. 2021. Police Caught One of the Web’s Most Dangerous Paedophiles. The Everything Went Dark. The Wired, 12 May, https://www.wired.co.uk/article/whatsapp-encryption-child-abuse.
Byron, Paul. 2019. “How could you write your name below that?” The queer life and death of Tumblr. Porn Studies 6 (3): 336–349.
Chamberlain, Lura. 2019. FOSTA: A Hostile Law with a Huma Cost. Fordham Law Review 87 (5): 2171–2211.
Chen, Adrian. 2014. The Laborers Who Keep Dick Pics and Beheadings Out of Your Facebook Feed. Wired, October 23, https://www.wired.com/2014/10/content-moderation/.
Cho, Alexander. 2018. Default publicness: Queer youth of color, social media, and being outed by the machine. New Media & Society 20 (9): 3183–3200.
Dickson, EJ. 2020. ‘OnlyFans Creators and Sex Workers are Getting ‘Purged’ from TikTok. The Rolling Stone. December 17. https://www.rollingstone.com/culture/culture-features/onlyfans-sex-workers-tiktok-purge-banned-1101928/.
Dworkin, Andrea. 1989. Pornography; Men Possessing Women. 2nd edition. New York: E. P. Dutton.
Dworkin, Andrea. 2000. Pornography and Grief (1987). In Drucilla Cornell (ed.), Feminism & Pornography. Oxford: Oxford University Press, 39–44.
Dworkin, Andrea. 2019. Last Days at Hot Slit: The Radical Feminism of Andrea Dworkin. Cambridge, MA: Semiotext(e).
Fabbri, Thomas. 2019. Why is Instagram Deleting the Accounts of Hundreds of Porn Stars? BBC Trending, 24 November. https://www.bbc.co.uk/news/blogs-trending-50222380.
Gentile, Kathy Justice. 2010. Sexing the Look in Popular Visual Culture. Cambridge: Cambridge Scholars Publishing.
Gillespie, Tarleton. 2010. The Politics of “Platforms.” New Media & Society 12 (3): 347–364.
Gillespie, Tarleton. 2018. Custodians of the Internet: Platforms, content moderation, and the hidden decisions that shape social media. New Haven: Yale University Press.
Goldman, Eric. 2019. The Complicated Story of FOSTA and Section 230. First Amendment Law Review 17: 279–293.
Grant, Melissa Gira. 2020. Nick Kristof and the Holy War on Pornhub. The New Republic, December 10. https://newrepublic.com/article/160488/nick-kristof-holy-war-pornhub.
Griffin, Susan. 1981. Pornography and Silence: Culture’s Revenge Against Nature. New York: Harper & Row.
Halperin, David M. 2017. Introduction: The War on Sex. In The War on Sex, ed. David M. Halperin and Trevor Hoppe. Durham, NJ: Duke University Press, 1–61.
Kappeler, Susanne. 1986. The Pornography of Representation. Minneapolis: University of Minnesota Press.
Kleinman, Zoe. 2021. Child Sexual Abuse: Four Held in German-Led Raid on Huge Network. BBC News, 3 May, https://www.bbc.com/news/world-europe-56969414.
Kristof, Nicholas. 2009. If This Isn’t Slavery, What Is? The New York Times, 3 January, https://www.nytimes.com/2009/01/04/opinion/04kristof.html.
Kristof, Nicholas. 2017. Google and Sex Traffickers Like Backpage.com. The New York Times, 7 September, https://www.nytimes.com/2017/09/07/opinion/google-backpagecom-sex-traffickers.html.
Kristof, Nicholas. 2020. Children of Pornhub. The New York Times, 4 December.
Kristof, Nicholas. 2021. Why Do We Let Corporations Profit from Rape Videos? The New York Times, 16 April.
Long, Julia. 2012. Anti-Porn: The Resurgence of Anti-Pornography Feminism. Zed Books Ltd.
Miller, William Ian. 1997. The Anatomy of Disgust. Cambridge, MA: Harvard University Press.
Molldrem, Stephen. 2018. Tumblr’s Decision to Deplatform Sex Will Harm Sexually Marginalized People, Wussy, December 6. https://www.wussymag.com/all/2018/12/6/tumblrs-decision-to-deplatform-sex-will-harm-sexually-marginalized-people.
Paasonen, Susanna. 2011. Carnal Resonance: Affect and Online Pornography. Cambridge, MA: MIT Press.
Paasonen, Susanna. 2018. Many Splendored Things: Thinking Sex and Play. London: Goldsmiths Press.
Paasonen, Susanna, 2021. Intime Abhängigkeiten, fragile Verbindungen, entsexualisierte Plattformen Sexuologie 28 (1–2).
Paasonen, Susanna, Kylie Jarrett and Ben Light. 2019. NSFW: Sex, Humor, and Risk in Social Media. Cambridge, MA: MIT Press.
Paasonen, Susanna, Feona Attwood, Alan McKee, John Mercer and Clarissa Smith. 2020. Objectification: On the Difference Between Sex and Sexism. London: Routledge.
Parks, Lisa. 2019. Dirty Data: Content Moderation, Regulatory Outsourcing, and The Cleaners. Film Quarterly 73 (1): 11–18.
Pilipets, Elena and Susanna Paasonen. 2020. Nipples, Memes, and Algorithmic Failure: NSFW Critique of Tumblr Censorship. New Media & Society. https://journals.sagepub.com/doi/pdf/10.1177/1461444820979280.
Pornhub. 2020. The Last on Our Commitment to Trust and Safety, https://www.pornhub.com/blog/11422.
Protevi, John. 2009. Political Affect: Connecting the Social and the Somatic. Minneapolis: Minnesota University Press.
Race, Kane. 2009. Pleasure Consuming Medicine: The Queer Politics of Drugs. Durham, NC: Duke University Press
Race, Kane. 2018. The Gay Science: Intimate Experiments with the Problem of HIV. New York, NY: Routledge.
Reynolds, Chelsea. 2020. “Craigslist is Nothing More than an Internet Brothel”: Sex Work and Sex Trafficking in US Newspaper Coverage of Craigslist Sex Forums. The Journal of Sex Research, https://doi.org/10.1080/00224499.2020.1786662.
Roberts, Sarah T. 2019. Behind the Screen: Content Moderation in the Shadows of Social Media. New Haven: Yale University Press.
Rogers, Richard, 2020. Deplatforming: Following extreme Internet celebrities to Telegram and alternative social media. European Journal of Communication 35(3): 213–229.
Rubin, Gayle. 1989. Thinking sex. In Carol S. Vance (ed.), Pleasure and Danger: Exploring Female Sexuality. London: Pandora, 267–319.
Spišák, Sanna, Tommi Paalanen, Susanna Paasonen, Elina Pirjatanniemi and Maria Vihlman. 2021. Social Networking Sites’ Gag Order: Commercial Content Moderation’s Adverse Implications for Fundamental Sexual Rights and Wellbeing. Social Media + Society, DOI: 10.1177/20563051211024962.
Thompson, Jay David. 2015. Invisible and everywhere: Heterosexuality in anti-pornography feminism. Sexualities 18 (5–6), 750–764.
Tiidenberg, Katrin and van der Nagel, Emily. 2020. Sex and Social Media. Melbourne: Emerald Publishing.
Tripp, Heidi. 2020. All Sex Workers Deserve Protection: How FOSTA/SESTA Overlooks Consensual Sex Workers in an Attempt to Protect Sex Trafficking Victims. Penn State Law Review 124, 219–246.
Vance, Carole S. 1997, Negotiating Sex and Gender in the Attorney General’s Commission on Pornography. In Roger N. and Micaela di Leonardo (eds.), The Gender/Sexuality reader: Culture, History, Political Economy. New York: Routledge, 440–452.
Warner, Michael. 2000. The Trouble with Normal: Sex, Politics, and the Ethics of Queer Life. Cambridge, MA: Harvard University Press.
Weitzer, Ronald. 2007. The Social Construction of Sex Trafficking: Ideology and Institutionalization of a Moral Crusade. Politics & Society 35 (3): 447–475.
Our collaborative book with Joshua Neves, Aleena Chia and Ravi Sundaram for Meson & University of Minnesota Press’s In Search of Media Series has an approximate due-date for June 2022. TechnoPharmacology examines the close relations of media technologies to pharmaceuticals and pharmacology. It is a modest call to expand media theoretical inquiry by attending to the biological, neurological, and pharmacological dimensions of media and centers on emergent affinities between big data and big pharma. The project has been great fun: it’s an absolute joy to work with these smart people.
My section in the book, titled “Drugs, epidemics, and networked bodies of pleasure”, explores the conflation of online pornography with an addiction of epidemic proportions with the aim of centering pleasure as a matter of gravity in and for critical inquiry. Returning to Derrida’s conceptualization of the pharmakon as both a toxin and a remedy, the cause and the cure, the bad and the good—it considers the ambiguities of pleasure as they come about in encounters with networked media, sexually explicit content, and intoxicating substances.
I am delighted to be one of the three scholars elected for a 2022 Hunt-Simes Visiting Chair in Sexuality Studies position at Sydney Social Sciences and Humanities Advanced Research Centre (SSSHARC), University of Sydney. If in the spring we are living in a world where people fly long, long distances, I’ll have the pleasure of working with Kane Race and the rest of the excellent Sydney team on sexual expression and social media platform governance. Very much honored to be in the same company with Jen Gilbert and Srila Roy,
Our short piece on shadowbanning, Sex in the Shadows of Celebrity, written together with the wonderful Dr Carolina Are, is out on OA with Porn Studies as part of a forthcoming special issue on the deplatforming of sex in social media. Here’s the abstract:
Shadowbanning is a light censorship technique used by social media platforms to limit the reach of potentially objectionable content without deleting it altogether. Such content does not go directly against community standards so that it, or the accounts in question, would be outright removed. Rather, these are borderline cases – often ones involving visual displays of nudity and sex. As the deplatforming of sex in social media has accelerated in the aftermath of the 2018 FOSTA/SESTA legislation, sex workers, strippers and pole dancers in particular have been affected by account deletions and/or shadowbanning, with platforms demoting, instead of promoting, their content. Examining the stakes involved in the shadowbanning of sex, we focus specifically on the double standards at play allowing for ‘sexy’ content posted by or featuring celebrities to thrive while marginalizing or weeding out posts by those affiliated with sex work.
Written together with Feona Attwood, Clarissa Smith, Alan McKee and John Mercer, our article both recapping and elaborating on our argument in the Objectification book that came out last year, Sexual Objects, Sexual Subjects and Certified Freaks: Rethinking “Objectification” is just out today with MAI: Feminism and Visual Culture. It is written with pedagogical purposes in mind so as to be accessible to undergraduate students, and is on open access.
Our article with Mari Lehto, titled ‘I feel the irritation and frustration all over the body’: Affective ambiguities in networked parenting culture is freshly out with The International Journal of Cultural Studies, on open access. The fieldwork was all Mari’s, and here’s the abstract:
This article investigates the affective power of social media by analysing everyday encounters with parenting content among mothers. Drawing on data composed of diaries of social media use and follow-up interviews with six women, we ask how our study participants make sense of their experiences of parenting content and the affective intensities connected to it. Despite the negativity involved in reading and participating in parenting discussions, the participants find themselves wanting to maintain the very connections that irritate them, or even evoke a sense of failure, as these also yield pleasure, joy and recognition. We suggest that the ambiguities addressed in our research data speak of something broader than the specific experiences of the women in question. We argue that they point to the necessity of focusing on, and working through affective ambiguity in social media research in order to gain fuller understanding of the complex appeal of platforms and exchanges.
My book very long in the making, Dependent, Distracted, Bored: Affective Formations in Networked Media, is out April 20 with MIT Press. To mark the occasion, I’m doing an IIPC debate talk the day after summing up some of its central themes and points. Join us April 21, 5:15pm EET, at https://utu.zoom.us/j/67932423692. This is the abstract:
According to a dominant narrative repeated in journalistic and academic accounts for more than a decade, we are addicted to the digital devices, apps, and sites designed to distract us, which drive us to boredom and harm our capacities to focus, relate, remember, and be. Focusing on three affective formations — dependence, distraction, and boredom — as key to understanding both the landscape of contemporary networked media and the concerns connected to it, this talk challenges the dominant narrative and argues for the centrality of accounting for complexity and ambiguity instead. Dependence and agency, distraction and attention, boredom and excitement can be seen as dynamics that enmesh, oscillate, enable, and depend on one another — and, in some instances, cannot be told apart.
The 17th Technology, Knowledge & Society conference, hosted by University of Melbourne and held entirely online, takes place April 8-9, 2021, with the overall theme “Considering Viral Technologies: Pandemic-Driven Opportunities and Challenges”. Very excited about doing a live plenary & garden conversation (8 April 2021 08:00AM CST Chicago // 8 April 2021 16:00PM Finland // 8 April 2021 11:00PM Melbourne) around my soon out book, Dependent, Distracted, Bored: Affective Formations in Networked Media (MITP).
Our article with Jenny Sundén is very freshly out with the Qualitative Research Journal, on open access as part of a forthcoming special issue on Activist methodologies inside and outside of academy, edited by Gabriele Griffin. Titled “We Have Tiny Purses in Our Vaginas!!! #thanksforthat”: Absurdity as a Feminist Method of Intervention, it focuses on the Twitter account, Men Write Women, “Where the women are made up & their anatomy doesn’t matter“. This one virtually wrote itself: hope some of the fun communicates.