We can make Twitter better, but never perfect

take-back-twitterOn a BBC Newsnight discussion last night, panelists were asked if they felt that a watershed moment had been reached in our attitude to misogynist abuse and threats on Twitter.

It certainly feels like a moment: The increasing pervasiveness of social media in UK public life, combined with a resurgent young feminist movement (aided by social media), has led to this tumult of voices. We all seem to agree that something should be done about threats of rape, such as the ones experienced by Labour MP Stella Creasey, but the question is what?

Twitter yesterday released a statement saying they would look into introducing a report abuse button under all tweets. This, the company pointed out, is in fact already available on the Twitter iphone app, and will soon be rolled out on Android and desktop.

But some see a danger in this: a report button, they suggest, could easily be abused. Celebrity users with a large amount of followers could, for example, get their fans to report critics.

Political opponents and human and civil rights activists could easily be marked as abusive by users. I can imagine, for example, that Index on Censorship’s account would repeatedly be reported as abusive by, say, highly co-ordinated pro-Bahraini government tweeters.

What are the alternatives? Writing for the Guardian, Index contributor Jane Fae says that the focus is too much on the potential solutions Twitter can provide and not enough on the existing laws in place. Earlier this year, the Crown Prosecution Service released guidelines on social media user prosecutions. Index greeted these as by and large sensible (interesting to note that these came about because of a feeling that there were too many prosecutions of social media users; though there is a distinction between generally offensive material and specific, direct, threats).

Jane suggests a mechanism for reporting threats to the police, rather than Twitter:

“By all means, let’s have a button – but one that delivers reports of online abuse directly to the local police force, a bit like a security alarm, and not just to Twitter.

This relatively simple solution doesn’t exist yet but I’m currently coding a mock-up of one for Everyday Victim Blaming (EVB), a not-for-profit organisation set up to help stop violence against women. Installed on your computer, the button would let you generate instant email reports, detailing online abuse and asking the police to investigate. It should also copy a report back to EVB or another campaign group. If the police really are swamped, that, in itself, is a thing – and may finally prompt politicians to ditch the soundbite and think a bit deeper on this issue.

Matt Flaherty, the prolific and interesting tweeter and blogger on free speech issues, brings the issue back to Twitter with his idea of a “panic mode” to help tweeters beleaguered by coordinated attacks:

“My solution to this problem would be for Twitter to introduce something like a panic button that would immediately but temporarily place one’s account into a state where all mentions are blocked except for those coming from the followers and following lists. Twitter could also allow the user to open a case and have mentions logged against that case to help Twitter to take action against those who violate the terms. Twitter could also perhaps warn users attempting to put the panicked account into a mention or reply. Perhaps they could ask for confirmation either within the stream or in an email. An alternative would be to actually disallow the creation of the mention or reply. A blocked account is currently not allowed to perform a reply but it can use a mention. The same rules might apply here. The mention would still not be seen.”

It’s an interesting idea, particularly for the kind of abuse Caroline Criado-Perez was subjected to: The campaigner was receiving 50 tweets an hour, suggesting a level of co-ordination. Flaherty’s proposal is a kind of advanced version of Don’t-Feed-The-Trolls, allowing a user to silence the angry mob directed at them.

The Daily Telegraph’s Marta Cooper (formerly of this parish), points out, however, that technological fixes are not the answer to societal problems:

“The offline world is still inhabited by some men who believe that women who voice an opinion must be put in their place. This is simply reflected and amplified online. Twitter did not create the misogynistic monsters that hurled grotesque tweets at Criado-Perez or send the same to other women on a regular basis; online platforms simply make it easier for these individuals to behave in appalling ways, and it is far harder to ignore published text than a hateful comment made by some Neanderthal in a pub.

“As with other societal problems, there is no quick fix – technological or otherwise – to something which has centuries’ worth of roots. It is far more valuable and worthwhile to expose hate for what it is and to debate why it occurs in the first place.”

Martin Belam makes a similar point here.

So what do we do? MPs have called Twitter executives to testify at the House of Commons: there is no doubt Twitter could have handled this better – the action of their New York based manager of journalism Mark S Luckie, blocking critical but polite UK voices, has been particularly unfortunate. No doubt the Twitter representatives will face demands from MPs to do something, and fast.

But we should be careful of quick fix solutions: the police are investigating the threats against Criado-Perez and Creasy; it is likely there will be prosecutions. This may seem frustratingly slow for some, but that is the pace of these things. The instant nature of social media means we often expect instant remedies for problems that are, as Ben Goldacre puts it, a little bit more complicated than that. Collective effort and thought on this should take precedent over the kind of kneejerk thinking about the web that too often dominates (see David Cameron’s “ban all the bad things” speech last week). We want the web to be free and diverse, but we also want the web to be,  as much as possible, a positive experience for people. This is the challenge we face, and our reactions over the issues raised in the past few weeks could have consequences for years to come. Together we can make the web better, but we should be wary of attempts to make it perfect.

Online attacks: What threats do women human rights defenders face?

Human rights activists inevitably face harassment both on and offline, and this effect is magnified by the topics tackled by women’s rights activists. Discussion of deep-seated taboos sparks heated debates across the globe.

The Association for Progressive Communication (APC) recently released a report exploring the kinds of threats that some women’s rights activists face online. APC surveyed 40 activists working across the globe, as part of their ongoing efforts to use technology to battle violence against women online.

The greatest fear for respondents was the possibility of private information being shared online without their consent, while 90 per cent of Spanish-speaking respondents expressed fears over the security of social networking sites. English speaking respondents were also more likely to report harassment online over Spanish speaking ones, but this could do with English being more widely used online.

The report looked at the kinds of threats the activists have faced as well as the kinds of training that activists feel they need:

Analysis of threats faced by surveyed WHRDs  | Association of Progressive Communication (APC)

APC is also working on a more long-term project documenting and fighting against violence against women online — whether that is cyberstalking, surveillance, or manipulated images. It’s important to examine gender-specific obstacles to free speech both on and offline.

More on this story:

Women and free speech

Sara Yasin is an editorial assistant at Index on Censorship