A senior Twitter executive has said there is ‘no doubt’ that content on Twitter and other social networks could contribute to radicalisation, while revealing that 1.6 million accounts have been taken down related to terrorism.
Like other platforms such as Facebook and YouTube, Twitter has grappled with a number challenges, aiming to strike the right balance of conversation against a backdrop of hate speech and misinformation, among other issues.
‘I think there is content on Twitter and every platform that contributes to radicalisation, no doubt, but I think we also have a lot of mechanisms and policies in place that we enforce very effectively that combat this,’ said Twitter’s Vijaya Gadde, who overseas the company’s legal, policy, and trust and safety.
According to the executive, 90% of terror-related content taken down is detected by Twitter’s own technologies proactively without any users reporting it. Despite Twitter’s efforts to improve policies, Ms Gadde admitted the social network has ‘hit an inflection point’ following recent attacks such as the shooting at the Christchurch mosque in March, in which 51 people were killed.
However, she stood firm on the social network’s position on violent extremist groups.
‘We have a violent extremist group policy that has banned over 110 violent extremist groups, 90 plus percent of those are white supremacist or white nationalist groups, including the American Nazi Party, the Proud Boys, the KKK,’ she explained at the Code Conference.
‘If you have any affiliation, if you claim any affiliation to those parties, you are not allowed on Twitter, period. You can’t have any accounts – I want to be very clear, that is our policy.’
However, Kayvon Beykpour, Twitter product lead – who spoke alongside Ms Gadde, said that a lot of what people consider abusive on the service does not actually violate its policies, because what one person finds abusive may differ from the next person.
‘One of the things we’ve really had to step up from a product and technology standpoint is proactively de-amplifying content that we don’t think should be amplified,’ he explained.
Twitter recently updated its rules as part of plans to make the site’s processes easier for users to understand, cutting the words in its official guidelines from about 2,500 to fewer than 600.
Source: Read Full Article