New Delhi, October 26: Google on Thursday said it has developed a prototype that leverages recent advances in Large Language Models, or LLMs, to assist in identifying content abusive at scale. LLMs are a type of artificial intelligence that can generate and understand human language. “Using LLMs, our aim is to be able to rapidly build and train a model in a matter of days — instead of weeks or months — to find specific kinds of abuse on our products,” said Amanda Storey, senior director, trust and safety.

Google is still testing these new techniques, but the prototypes have demonstrated impressive results so far. “It shows promise for a major advance in our effort to proactively protect our users especially from new, and emerging risks,” Storey added. The company, however, did not specify which of its many LLMs it is using to track misinformation. Smartphone Launches in November 2023: From iQOO 12 and Xiaomi 14 To OnePlus Ace 2 Pro, Know Specifications and Other Details of Upcoming Smartphones.

"We’re constantly evolving the tools, policies and techniques we’re using to find content abuse. AI is showing tremendous promise for scaling abuse detection across our platforms,” said Google. Google said it is taking several steps to reduce the threat of misinformation and to promote trustworthy information in generative AI products. Instagram Tests New Feature That Lets Users Update Their Profile Picture With Short or Looping Videos in 'Notes'.

The company has also categorically told developers that all apps, including AI content generators, must comply with its existing developer policies, which prohibit the generation of restricted content like child sexual abuse material (CSAM) and content that enables "deceptive behaviour". To help users find high-quality information about what they see online, Google has also rolled out the "About this image" fact-check tool to English language users globally in Search.

(The above story first appeared on LatestLY on Oct 27, 2023 10:10 AM IST. For more news and updates on politics, world, sports, entertainment and lifestyle, log on to our website latestly.com).