The group, concerned by the vicious targeting of women online, started a pro-bono project to support users experiencing gendered harassment or abuse. Their website now hosts a detailed reporting form so they can validate and escalate claims of abuse to Twitter.
So while this seems a positive step, historically Twitter hasn’t addressed issues of harassment on its platform well at all. Will this new collaboration change anything?
A growing issue
A quarter of 18- to 24-year-old women surveyed by the Pew Research Center reported having been stalked or sexually harassed online, and a similar report from Demos found terms such as “rape”, “slut” and “whore” were often used in tweets to convey casually misogynistic sentiments, or to threaten and abuse other users.
- a Guardian journalist was harassed after asking Twitter users about tampons
- a Fox News panellist was threatened after rejecting the notion that firearms were a solution to sexual violence
- a former DC Comics editor received a deluge of abuse after criticising a comic book cover that clumsily sexualised teenaged characters
- a feminist critic was forced from her home after criticising the depiction of women in video games
- a grieving daughter was inundated with faked photographs of her father mere hours after his death.
Countless other women deal with online threats and abuse on a daily basis.
The reactions to these women were extreme and irrational. There is no justification for threatening rape, murder and grievous harm, or for the continued collaborative harassment of women online and on social media.
How does Twitter respond?
In response to bomb and rape threats against a number of women in the UK, Twitter was notoriously slow to respond. Twitter’s eventual solution was a button for the Twitter app that allowed users to report abusive tweets to Twitter’s moderation team.
As with any technical solution to a social problem, the button (and Twitter’s ensuing moderation process) is not without flaws.
Some of those were apparent during recent episodes of harassment, where users who tried to help victims by reporting harassing accounts claimed that Twitter refused to moderate tweets that were reported by people other than the victim.
If you see abuse on twitter, and you report it, twitter emails you to tell you they will ignore it because it didn’t happen to you. In the real world, when you see abuse and report it, your observation/testimony is part of the societal feedback loop for correction. Twitter’s approach to abuse reporting is to minimise false reporting rather than solve abuse.
It’s possible that Twitter doesn’t yet understand the extent of online abuse on its platform.
Harassers can use anonymising tools to obscure their identity, they can use “disposable email” services to set up burner accounts on social media to minimise consequences of being blocked.
In some cases, they harass socially, loosely organising across multiple message-boards and social platforms to create what American law professor Danielle Keats Citron terms as “cybermobs”. The behaviour of harassers can be complex, and some of the tools and tactics deployed are intentionally used to make their behaviour difficult to police.
Of course, misogynistic behaviour does not exist in a vacuum. The abusive, harassing, threatening behaviour I’ve highlighted is not exclusive to Twitter.
It’d be cynical to dismiss Twitter’s failings as indifference towards the victims of harassment, threats and online bullying but sadly, the most visible advances Twitter’s made towards combating online abuse are almost always made in response to popular outcry and media attention.
The newly-announced collaboration with WAM is arguably their biggest advance yet, and it’s only a short pilot project, organised by a non-profit organisation.
To quote WAM executive director Jaclyn Friedman:
I don’t think we should have to do this work. I think it’s a scandal that a tiny, under-resourced nonprofit with two staff members is having to do free labour for them.
In an online space where conventional law enforcement is often unable or unwilling to participate, Twitter’s intervention in abusive behaviour is critically important. So when will Twitter step up and take responsibility for the discourses it fosters online?
For now, let’s hope that their collaboration with WAM encourages them to do better.
By Andrew Quodling, PhD Candidate at Queensland University of Technology