14 Words by Email

Friday, 5 April 2013

How to Spot a White Supremacist on Twitter

For as long as the Internet has been commonplace, law enforcement agents have found it helpful in catching criminals (and not just the stupid ones who post video evidence of their crimes on YouTube, either). Extremist groups—such as the various American white-supremacist organizations—also find the Internet useful for connecting with like-minded people.

But it’s intensely time-consuming for police to personally wade through the web’s ever-growing number of white supremacist tweets and other social media postings in search of the relative handful of extremist posts indicating a possible threat.
Maybe there’s an easier way. Two researchers, J. M. Berger and Bill Strathearn, from the International Centre for the Study of Radicalization and Political Violence (ICSR) in London, have developed an algorithm with a high rate of success in identifying extremists on Twitter, by analyzing the relationships between Twitter account holders (as opposed to analyzing the actual posted content).
In a 56-page study (released in pdf form), Berger and Strathearn said:
“It is relatively easy to identify tens of thousands of social media users who have an interest in violent ideologies, but very difficult to figure out which users are worth watching. For students of extremist movements and those working to counter violent extremism online, deciphering the signal amid the noise can prove incredibly daunting.”
{snip}
But there’s a third group: those who don’t merely profess extremist beliefs but are willing to act violently upon them. If you’re in law enforcement, hoping to identify members of the third group before they hurt anybody, how can you do this?
{snip} Berger and Strathearn’s algorithm can make those determinations mathematically.
“Our starting data centered on followers of 12 American white nationalist/white supremacist “seed” accounts on Twitter. We discovered that by limiting our analysis to interactions with this set, the metrics also identified the users who were highly engaged with extremist ideology.”
Those 12 accounts had over 3,500 followers between them (who collectively generated over 340,000 different tweets), yet less than half of those followers publicly self-identified as white supremacists or white nationalists. {snip}
According to Berger and Strathearn:
“By measuring interactions alone—without analyzing user content related to the ideology—we narrowed the starting set down to 100 top-scoring accounts, of which 95 percent overtly self-identified as white nationalist. […] A comparison analysis run on followers of anarchist Twitter accounts suggests the methodology can be used without modification on any number of ideologies.”
The researchers identified three key terms used in their algorithms, which they listed and defined as follows:
  • Influence: A metric measuring a Twitter user’s ability to publicly create and distribute content that is consumed, affirmed and redistributed by other users.
  • Exposure: A metric measuring a Twitter user’s tendency to publicly consume, affirm and redistribute content created by other users.
  • Interactivity: A metric measuring a Twitter user’s combined influence and exposure based on their public activity.
For example: suppose you’re a non-racist person following ex-Klansman David Duke on Twitter, and occasionally sending him a tweet disagreeing with his views. It’s highly unlikely any of Duke’s racist followers will find your comment worth re-tweeting. But a racist Duke follower who sends tweets reinforcing his white power views probably will inspire lots of retweets and conversations in the more bigoted regions of the Twitterverse.
Thus the algorithm focuses on the connections while paying no attention to the content. {snip}
{snip}

0 comMENTS:

Post a Comment

Share

Twitter Delicious Facebook Digg Stumbleupon Favorites More