WASHINGTON – Big tech companies, including Google and Facebook owner Meta, on Tuesday said they would team up on a new programme to fight online child sexual abuse and exploitation.
Child victims of abuse online are a hot button issue for regulators, and tech companies are eager to show they are taking adequate measures to protect kids and teenagers.
In the new programme, called Lantern, big tech companies will share signals of activity that violate their policies on child exploitation so that platforms can move more quickly to detect, take down and report problematic content.
Signals could be e-mail addresses, certain hashtags, or keywords that are used to either groom young people into being abused or buy and sell material involving child abuse and exploitation.
“Until now, no consistent procedure existed for companies to collaborate against predatory actors evading detection across services,” said Mr Sean Litton, executive director of the Tech Coalition, which brings together tech companies on the issue.
“Lantern fills this gap and shines a light on cross-platform attempts at online child sexual exploitation and abuse, helping to make the internet safer for kids.”
Other platforms in the Tech Coalition include Snap, Discord and Mega, a privacy-focused platform from New Zealand.
Tech Coalition said that a pilot of the programme saw Meta remove more than 10,000 Facebook profiles, pages and Instagram accounts after data was shared by Mega.
Meta reported the accounts involved to the United States-based National Centre for Missing & Exploited Children, and shared findings with other platforms for their own investigations.
“Predators don’t limit their attempts to harm children to individual platforms,” said Ms Antigone Davis, global head of safety at Meta.
“The technology industry needs to work together to stop predators and protect children on the many apps and websites they use.”
The announcement of Lantern came on the same day that a former Meta senior engineer told a Senate hearing in Washington that top Meta executives, including chief executive Mark Zuckerberg, ignored his warnings that teens were unsafe on the company’s platforms.
Mr Arturo Bejar told lawmakers that in an internal survey of 13- to 15-year-olds on Instagram, 13 per cent of respondents had received unwanted sexual advances on Instagram in the last seven days.
“Meta knows the harm that kids experience on their platform and executives know that their measures fail to address it,” he said. AFP
EMEA Tribune is not involved in this news article, it is taken from our partners and or from the News Agencies. Copyright and Credit go to the News Agencies, email [email protected] Follow our WhatsApp verified Channel