Manchester - Social-media giants such as
Twitter will have to reveal the scale of cyber bullying in the UK and face being made to pay the cost of dealing with it.
Under the latest guidance by the UK government, technology
companies will be required to publish an annual report on how complaints
are handled, the reported abuse that is pulled down and the extent of
their efforts to moderate bullying or offensive content about children,
women, gay people or religions.
One of the proposals is for “an industry-wide levy so social-media
companies and communication service providers contribute to raise
awareness and counter internet harms,”? according to a statement
published Wednesday that didn’t give further details.
“Behavior that is unacceptable in real life is unacceptable on a computer screen,” Culture Secretary
Karen Bradley said in an email released by her office. "We need an approach to the internet that protects everyone.”
The campaign is part of the government’s wider strategy to force
technology companies to accept greater responsibility for their content.
Amber Rudd has also called on companies to “step up” and assume moral
responsibility for ridding their platforms of terrorist content,
refusing to rule out the prospect of compulsion by fines or legislation.
The UK has been pushing the envelope in terms of how willing it is
to go after Silicon Valley. Efforts to end hate speech and trolling on
social media have intensified in the wake of five terror attacks this
year, yet the desire to regulate tech firms - in ways that are
unprecedented - risks driving them offshore.
On Tuesday, Sharon White, the chief executive of UK media regulator
Ofcom, said she viewed companies like Facebook as news publishers.
Prime Minister Theresa May’s spokesperson, James Slack, later told
reporters that the government was “looking at the role Google and
Facebook play in the news environment” as well as “the roles,
responsibility and legal status of the major internet platforms.”
In May 2016, a number of social-media companies, including Facebook,
Twitter and Google’s YouTube voluntarily committed to trying to take
down illegal content within 24 hours.
But last month the European Commission
called upon the tech firms to do more to block illegal content. Germany
has passed a law requiring hate speech to be removed within 24 hours of
it being flagged, with penalties of up to €50m
for repeated failures to comply.
In September, May went further. At a meeting at the United Nations, she
propose new rules requiring internet companies to take down extremist content within two hours.
SUBSCRIBE FOR FREE UPDATE: Get Fin24's top morning business news and opinions in your inbox.
Read Fin24's top stories trending on Twitter: