People don’t commit violence against other groups - or even condone it - spontaneously. First they must be taught to see other people as pests, vermin, aliens, or threats. Malicious leaders often use the same types of rhetoric to do this, in myriad cultures, languages, countries, and historical periods. We call this Dangerous Speech. Violence might be prevented by making it less abundant or less convincing. We work to find the best ways to do this – while protecting freedom of expression.
In June, the Berkman Klein Center for Internet & Society at Harvard University hosted a workshop to discuss how market dynamics, behavioral drivers, laws, and technology contribute to the spread of harmful speech online and inform measures to constrain it. Dangerous Speech Project director Susan Benesch spoke at the workshop, arguing that there should be third-party auditing of platforms' use of algorithms for content removal.
This report on Internet hate speech, hate speech law, and efforts to diminish it synthesizes research and case studies from four countries. It was published by the European project BRICkS Against Hate Speech.
In Charlottesville, Americans watched barriers to Dangerous Speech go down in broad daylight, in the middle of a city, as extremists waved swastikas and chanted hateful slogans. Some people are taking matters into their own hands, reaching out to masses of others to identify and punish marchers in the ‘Unite the Right’ rally, but online shaming often goes too far, reaching into a person’s offline life to inflict punishment
In this essay, Susan Benesch points out that content 'takedown' by Internet companies is not the only solution to harmful speech online. She highlights projects organized by civil society - not governments or platforms - to diminish harmful speech and support its targets. It was published by the Berkman Klein Center for Internet & Society in a collection of essays on harmful speech online.