Twitter to hide posts from accounts identified as disruptive

Updated

Twitter has launched a new behaviour monitoring system that will hide content from accounts identified as trying to distort or disrupt public conversation on the site.

The site said the new system was designed to deal with "troll-like" behaviour from a minority of users that are often reported as abusive, but in some cases are not in breach of Twitter rules, by instead relegating their posts.

As part of plans to improve the health of discussion on the platform, Twitter said it will use signals such as accounts without a confirmed email address or those that heavily tweet other accounts which do not follow them to spot and demote disruptive content that appears in conversations and search results.

Twitter will demote disruptive content that appears in conversations and search results (Twitter)
Twitter will demote disruptive content that appears in conversations and search results (Twitter)

The firm's vice president for trust and safety, Del Harvey, told the Press Association: "The reason that we're doing this is that when we actually looked at where abuse reports were coming from, the majority of reports come in from those two places - from conversations and search.

"But what we've learned as we've dug into that a little bit more is that there were actually this tiny number of accounts - less than 1% of accounts - that were making up the majority of reports that we were receiving for abuse.

"Within that, some of those were breaking the rules and we take action on those, but there's a lot that aren't and within that grouping there is a number of accounts that are engaged in behaviours that really distort and detract from the experience that people have in those areas.

"We started thinking how could we use a lot of the systems that we've been improving and building up over the past couple of years to improve people's experience really across the board, without putting the burden on the individual to do all the work to clean things up."

Twitter said the majority of the new signals it will use are not visible externally but include behaviours such as one person signing up for multiple accounts simultaneously, or activity that Twitter says might indicate a planned co-ordinated attack on other users.

"These signals will now be considered in how we organise and present content in communal areas like conversation and search", said a blog post by Ms Harvey and Twitter product manager for health David Gasca on the new system.

"Because this content doesn't violate our policies, it will remain on Twitter, and will be available if you click on 'Show more replies' or choose to see everything in your search setting.

"The result is that people contributing to the healthy conversation will be more visible in conversations and search."

The company said early tests of the system had shown a reduction in the number of abuse reports being filed, indicating users were having a "better experience" because of the reorganisation.

Twitter said early tests of the system had shown a reduction in the number of abuse reports being filed (Twitter)
Twitter said early tests of the system had shown a reduction in the number of abuse reports being filed (Twitter)

Earlier this year Twitter announced plans to focus on the health of conversation taking place on the platform, with chief executive Jack Dorsey admitting the site had been misused in the past but was committed to "help increase the collective health, openness, and civility of public conversation, and to hold ourselves publicly accountable towards progress".

Twitter has been regularly criticised for its approach to handling abuse and offensive content, something Mr Dorsey has said it must continue to improve.

Ms Harvey said Twitter was discussing the technology now as a way of being transparent with users.

"We're talking about this system much earlier than we normally would talk about something like this," she said.

"Part of the reason for that is because we want to be open and transparent about what we're doing and where we're focusing our efforts, so that we do get more feedback from people as they have this experience, so we do get more insight from a broader group of people who we get research from, or the more limited targeted feedback that we often get.

"Part of our whole reason for talking about this, even though it's day one of it, is so that we can continue to get that feedback and make sure we are delivering the experience that people want."

Advertisement