This post was originally published on this site
https://i-invdn-com.akamaized.net/trkd-images/LYNXMPEF9U1BV_L.jpgBy Foo Yun Chee
BRUSSELS (Reuters) – Twitter’s (N:) technology tools removed one in two tweets containing abusive content posted in the first half of this year, it said on Thursday, amid calls to the U.S. social media company and its peers to do more to tackle the issue.
Silicon Valley tech giants have in recent months pledged to tighten rules and share more information on abusive content posted on their platforms to avoid more heavy-handed regulatory action on both sides of the Atlantic.
Twitter said in its transparency report that it was investing in proactive technology to reduce the burden on people reporting abusive content to the company.
“More than 50 percent of tweets we remove for abuse are now proactively surfaced using technology, rather than relying on reports to Twitter,” the company said. That compared with 20% a year ago.
The report came a day after Twitter said it would ban political advertising on its platform next month as it and peers including Facebook Inc (O:) also face pressure to stop carrying ads that spread false information that could steer elections.
The report noted a 105% increase in accounts locked or suspended for violating its rules during the six-month period.
There was a 48% rise in accounts reported for potential violation of Twitter’s hateful conduct policies while 115,861 accounts were suspended because of their terrorist content, down 30% from the previous year.
The company also received 67% more legal requests to remove content from 49 countries, 80% of which came from Japan, Russia and Turkey.
Fusion Media or anyone involved with Fusion Media will not accept any liability for loss or damage as a result of reliance on the information including data, quotes, charts and buy/sell signals contained within this website. Please be fully informed regarding the risks and costs associated with trading the financial markets, it is one of the riskiest investment forms possible.