Queer profiles and post-surveillance
DOI:
https://doi.org/10.1590/1806-9584-2025v33n1104260Keywords:
Artificial intelligence, queer theory, algorithmic surveillance, user profilingAbstract
This article addresses the hypothetical capacity of algorithmic systems to generate queer user profiles, as a recent modality of the so-called "queer surveillance" and whose antecedents are traceable in Foucauldian reflections on the panopticon and the dispositif of sexuality. We extensively review how the mechanism of user profiling operates since its first prototypes in the late 1970s. We establish what its particularities are with respect to other types of identification. We propose that these systems configure an exponential leap in the technologies of perception of the sexuality of subjects that forces us to rethink central aspects of queer theory such as the Foucauldian sexuality device or the Butlerian heterosexual matrix, in a context of post-surveillance.
Downloads
References
BRUNO, Fernanda. Máquinas de ver, modos de ser: vigilância, tecnologia e subjetividade. Porto Alegre: Editora Sulina, 2013.
BUTLER, Judith. El género en disputa: el feminismo y la subversión de la identidad. México: Paidós, 2014.
COSTA, Flavia. “¿Qué hay de mí en esos datos? Políticas de la materialidad y estrategias críticas en las prácticas bioartísticas en América Latina”. In: TELLO, Andrés Maximiliano (Ed.). Tecnología, política y algoritmos en América Latina. Viña del Mar: Cenaltes Ediciones, 2020. p. 111-130.
CUKIER, Kenneth; MAYER-SCHÖNBERGER, Viktor. Big data, la revolución de los datos masivos. España: Turner Noema, 2013.
CHENEY-LIPPOLD, John. We are data: algorithms and the making of our digital selves. New York: NYU Press, 2017.
CHENEY-LIPPOLD, John. “A new algorithmic identity: soft biopolitics and the modulation of control”. Theory, Culture & Society, Los Angeles, v. 28, n. 6, p. 164-181, 2011.
DA CRUZ, Rui Alexandre P. P.; GARCÍA PEÑALVO, Francisco J.; ROMERO, Luis Alonso. Perfil de usuario, en la senda de la personalización. Informe técnico. Salamanca: Departamento de Informática y Automática, Universidad de Salamanca, 2003. p. 1.
DELEUZE, Gilles. “Post-scriptum sobre las sociedades de control”, Polis, 13, p. 1-6, 2006.
DE VRIES, Katja. “Identity, profiling algorithms and a world of ambient intelligence”. Ethics and Information Technology, v. 12, p. 71-85, 2010.
FOUCAULT, Michel. La historia de la sexualidad: vol. 1. México: Siglo XXI, 2007.
FOUCAULT, Michel. Vigilar y castigar: nacimiento de la prisión. México: Siglo XXI, 2001.
GALIČ, Maša; TIMAN, Tjerk; KOOPS, Bert-Jaap. “Bentham, Deleuze and beyond: an overview of surveillance theories from the panopticon to participation”. Philosophy & Technology, v. 30, p. 9-37, 2017.
GORDON, Diana. “The Electronic Panopticon: A Case Study of the Development of the National Crime Records System”, Politics and Society 15(4), p. 483-511, 1987.
HAGGERTY, Kevin D.; ERICSON, Richard V. “The surveillant assemblage”. The British Journal of Sociology, v. 51, n. 4, p. 605-622, diciembre 2000.
HAN, Byung-Chul. Psicopolítica: neoliberalismo y nuevas técnicas de poder. Barcelona: Herder, 2014.
KAFER, Gary; GRINBERG, Daniel. “Editorial: queer surveillance”. Surveillance & Society, v. 17, n. 5, p. 592-601, 2019.
MATZNER, Tobias. “Beyond data as representation: the performativity of big data in surveillance”. Surveillance & Society, v. 14, n. 2, p. 197-210, 2016.
MORITA, Masahiro; SHINODA, Yoichi. “Information filtering based on user behavior analysis and best match text retrieval”. In: Proceedings of the 17th Annual International ACM-SIGIR Conference on Research and Development in Information Retrieval. Dublin, Ireland, 3-6 July 1994.
MYERS, Brad A.; HOLLAN, Jim; CRUZ, Izabel (Eds.). “Strategic directions in human-computer interaction”. ACM Computing Surveys, v. 28, n. 4, p. 794-809, 1996.
POSTER, Mark. The Mode of Information, Chicago: University of Chicago Press, 1990.
RICH, Elaine. “User modeling via stereotypes”. Cognitive Science, v. 3, p. 329-354, 1979.
ROUVROY, Antoinette; BERNS, Thomas. “Gobernabilidad algorítmica y perspectivas de emancipación: ¿lo dispar como condición de individuación mediante la relación?”. Ecuador Debate, n. 104, agosto de 2018.
SAKAGAMI, Hidekazu; KAMBA, Tomonari. “Computer networks and ISDN systems”. Computer Networks and ISDN Systems, v. 29, p. 1447-1455, 1997.
SHACKEL, Brian. “Man-computer interaction: the contribution of the human sciences”. Ergonomics, v. 12, n. 4, 1969.
SCHRAM, Brian. “Accidental orientations: rethinking queerness in archival times”. Surveillance & Society, v. 17, n. 5, p. 609-610, 2019.
SCOLARI, Carlos. Hacer click: hacia una sociosemiótica de las interacciones digitales. Barcelona: Gedisa, 2004.
U.S. CONGRESS, Office of Technology Assessment, Federal government information technology: Electronic record systems and individual privacy, Washington: Government Printing Office, 1986.
VAN OTTERLO, Martijn. “A machine learning view on profiling”. In: HILDEBRANDT, Mireille; DE VRIES, Katja (Eds.). Privacy, due process and the computational turn: the philosophy of law meets the philosophy of technology. Londres: Routledge, 2013.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Revista Estudos Feministas

This work is licensed under a Creative Commons Attribution 4.0 International License.
Revista Estudos Feministas is under the Creative Commons International 4.0 Attribution License (CC BY 4.0), that allows sharing the work with recognition of authorship and initial publication in this journal.
The license allows:
Sharing (copying and redistributing the material in any support or format) and/or adapting (remixing, transforming, and creating from the material) for any purpose, even if commercial.
The licensor cannot revoke these rights provided the terms of the license are respected. The terms are the following:
Attribution – you should give the appropriate credit, provide a link to the license and indicate if changes were made. This can be done in several ways without suggesting that the licensor has approved of the use.
Without additional restrictions – You cannot apply legal terms or technological measures that prevent others from doing something allowed by the license.