igor_shmel - Fotolia

Twitter tests auto-block feature for accounts at risk of abuse

Latest Twitter feature automatically blocks abusive users, and is intended to help victims regain control of their experience on the platform

Twitter has begun pilot tests of new features and settings designed to protect users from online abuse, with the intention of automatically screening out abusive users and reducing the burden on victims to have to deal with unwelcome or triggering interactions.

The social network’s new Safety Mode is being trialled with selected beta testers, with an emphasis on people of colour, LGBTQ+ people, and women journalists, ahead of an anticipated wider roll-out in the future.

“We want you to enjoy healthy conversations, so this test is one way we’re limiting overwhelming and unwelcome interactions that can interrupt those conversations. Our goal is to better protect the individual on the receiving end of tweets by reducing the prevalence and visibility of harmful remarks,” said Twitter’s senior product manager Jarrod Doherty.

Safety Mode is an automated feature that temporarily blocks abusive accounts for seven days based on behaviours such as using insults or hateful remarks or sending repetitive or uninvited replies or mentions. During that time, they will not be able to follow you, see your tweets or send direct messages.

When switched on, Twitter’s new algorithms will assess the likelihood that a tweet addressed to you by considering both the tweet’s content, and the relationship between you and the replier, taking into account factors such as whether or not the two accounts frequently interact or follow one another, for example.

If you are in the trial group, you will already be able to activate it through Privacy and Safety under Settings. You will also be able to find information about tweets flagged via the feature, and details of the accounts that have been temporarily blocked. Before the seven-day block period ends, you will also receive a notification recapping this information.

Twitter said the system would inevitably make mistakes, so autoblocks can be viewed and retracted at any time via the site’s settings. The firm has also put in place measures to monitor the system’s accuracy and improve it over time.

The new feature was developed alongside a number of listening and feedback sessions with expert partners specialising in online safety, mental health, and human rights – many of them members of Twitter’s existing Trust and Safety Council, who provided feedback to make Safety Mode easier to use, and challenged Twitter’s developers to think through ways in which it might be manipulated and close off loopholes.

“As members of the Trust and Safety Council, we provided feedback on Safety Mode to ensure it entails mitigations that protect counter-speech while also addressing online harassment towards women and journalists,” said a spokesperson for Article 19, a London-based human rights organisation that specialises in digital rights and online equality.

“Safety Mode is another step in the right direction towards making Twitter a safe place to participate in the public conversation without fear of abuse.”

Twitter said it had also committed to the World Wide Web Foundation’s framework to end online gender-based violence, and was participating in discussions to explore new technical means of giving women and other groups more control over their online safety experience.

Read more about online abuse

  • With around a third of cyber pros saying they have personally experienced harassment at work or online, a new initiative is urging organisations to pledge their support to help free the community from the scourge of abuse.
  • Data in a new report appears to show that dangerous stalkerware apps are becoming normalised in younger age groups.
  • Created with the help of survivors, Refuge’s resource site offers guidance on protecting yourself from tech-enabled domestic abuse.

Read more on Security policy and user awareness

CIO
Security
Networking
Data Center
Data Management
Close