NSW Police want access to Tinder’s sexual assault data. Cybersafety experts explain why it’s a date with disaster.

Rosalie Gillett, Postdoctoral Research Fellow, Centre of Excellence for Automated Decision-Making and Society, Queensland University of Technology, Queensland University of Technology; Kath Albury, Professor of Media and Communication and Associate Investigator, ARC Centre of Excellence for Automated Decision-Making + Society, Swinburne University of Technology, and Zahra Zsuzsanna Stardust, Postdoctoral Research Fellow, Centre of Excellence in Automated Decision-Making and Society, Queensland University of Technology

Dating apps have been under increased scrutiny for their role in facilitating harassment and abuse.

Last year an ABC investigation into Tinder found most users who reported sexual assault offences didn’t receive a response from the platform. Since then, the app has reportedly implemented new features to mitigate abuse and help users feel safe.

In a recent development, New South Wales Police announced they are in conversation with Tinder’s parent company Match Group (which also owns OKCupid, Plenty of Fish and Hinge) regarding a proposal to gain access to a portal of sexual assaults reported on Tinder. The police also suggested using artificial intelligence (AI) to scan users’ conversations for “red flags”.

Tinder already uses automation to monitor users’ instant messages to identify harassment and verify personal photographs. However, increasing surveillance and automated systems doesn’t necessarily make dating apps safer to use.

User safety on dating apps

Research has shown people have differing understandings of “safety” on apps. While many users prefer not to negotiate sexual consent on apps, some do. This can involve disclosure of sexual health (including HIV status) and explicit discussions about sexual tastes and preferences.

If the recent Grindr data breach is anything to go by, there are serious privacy risks whenever users’ sensitive information is collated and archived. As such, some may actually feel less safe if they find out police could be monitoring their chats.

Adding to that, automated features in dating apps (which are supposed to enable identity verification and matching) can actually put certain groups at risk. Trans and non-binary users may be misidentified by automated image and voice recognition systems which are trained to “see” or “hear” gender in binary terms.

Trans people may also be accused of deception if they don’t disclose their trans identity in their profile. And those who do disclose it risk being targeted by transphobic users.

Increasing police surveillance

There’s no evidence to suggest that granting police access to sexual assault reports will increase users’ safety on dating apps, or even help them feel safer. Research has demonstrated users often don’t report harassment and abuse to dating apps or law enforcement.

Consider NSW Police Commissioner Mick Fuller’s misguided “consent app” proposal last month; this is just one of many reasons sexual assault survivors may not want to contact police after an incident. And if police can access personal data, this may deter users from reporting sexual assault.

With high attrition rates, low conviction rates and the prospect of being retraumatised in court, the criminal legal system often fails to deliver justice to sexual assault survivors. Automated referrals to police will only further deny survivors their agency.

Moreover, the proposed partnership with law enforcement sits within a broader project of escalating police surveillance fuelled by platform-verification processes. Tech companies offer police forces a goldmine of data. The needs and experiences of users are rarely the focus of such partnerships.

Match Group and NSW Police have yet to release information about how such a partnership would work and how (or if) users would be notified. Data collected could potentially include usernames, gender, sexuality, identity documents, chat histories, geolocation and sexual health status.

The limits of AI

NSW Police also proposed using AI to scan users’ conversations and identify “red flags” that could indicate potential sexual offenders. This would build on Match Group’s current tools that detect sexual violence in users’ private chats.

While an AI-based system may detect overt abuse, everyday and “ordinary” abuse (which is common in digital dating contexts) may fail to trigger an automated system. Without context, it’s difficult for AI to detect behaviours and language that are harmful to users.

It may detect overt physical threats, but not seemingly innocuous behaviours which are only recognised as abusive by individual users. For instance, repetitive messaging may be welcomed by some, but experienced as harmful by others.

Also, even as automation becomes more sophisticated, users with malicious intent can develop ways to circumvent it.

If data are shared with police, there’s also the risk flawed data on “potential” offenders may be used to train other predictive policing tools.

We know from past research that automated hate-speech detection systems can harbour inherent racial and gender biases (and perpetuate them). At the same time we’ve seen examples of AI trained on prejudicial data making important decisions about people’s lives, such as by giving criminal risk assessment scores that negatively impact marginalised groups.

Dating apps must do a lot more to understand how their users think about safety and harm online. A potential partnership between Tinder and NSW Police takes for granted that the solution to sexual violence simply involves more law enforcement and technological surveillance.

And even so, tech initiatives must always sit alongside well-funded and comprehensive sex education, consent and relationship skill-building, and well-resourced crisis services.


The Conversation was contacted after publication by a Match Group spokesperson who shared the following:

“We recognize we have an important role to play in helping prevent sexual assault and harassment in communities around the world. We are committed to ongoing discussions and collaboration with global partners in law enforcement and with leading sexual assault organizations like RAINN to help make our platforms and communities safer. While members of our safety team are in conversations with police departments and advocacy groups to identify potential collaborative efforts, Match Group and our brands have not agreed to implement the NSW Police proposal.”The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Authors

  • Kath co-leads the Digital inclusion Program at Swinburne's Social Innovation Institute. Her work investigates young people’s practices of digital self-representation, and the role of user-generated media (including social networking platforms) in young people’s formal and informal sexual learning, safety and wellbeing. She is particularly interested in the ways digital media users engage in cultures of care, and the ways that individuals and organisations develop digital literacy and data literacy.

    View all posts
  • Dr Rosalie Gillett is a Postdoctoral Research Fellow in the QUT Node of the Australian Research Council Centre of Excellence for Automated Decision-Making and Society and Research Fellow in the Digital Media Research Centre at QUT. Rosalie’s research expertise includes online gender-based violence; the normalisation of abuse; and platform governance. Her research particularly focuses on women’s experiences of harassment and abuse on dating apps, how toxic subcultures reinforce their social norms online, and digital platforms’ self-regulation practices. She was recently awarded a Facebook Content Governance grant to investigate how Facebook users circumvent content moderation policies.

    View all posts
  • Dr Zahra Stardust is a socio-legal researcher whose work is concerned with intersections between criminal law, sexuality, labour and justice. Her doctoral research Alternative Pornographies, Regulatory Fantasies, Resistance Politics examined the relationships between social movements and law reform. Zahra has fifteen years’ experience working on diverse social justice and human rights issues with community organisations, NGOs and UN bodies. She hosts a podcast called Thinking Justice in which she interviews academics, activists and critical thinkers about the past, the future and the limits of the law.

    View all posts