WHO Cautions Against the Usage of AI Chatbots Like ChatGPT, Bard in Healthcare
Carefully examining the risks involved is imperative while using artificial intelligence (AI) tools such as ChatGPT, Bard, and Bert in healthcare, the World Health Organisation (WHO) said on Tuesday.
Geneva, May 16 : Carefully examining the risks involved is imperative while using artificial intelligence (AI) tools such as ChatGPT, Bard, and Bert in healthcare, the World Health Organisation (WHO) said on Tuesday.
While the WHO is enthusiastic about the appropriate use of technologies, including the generated AI tools to support health-care professionals, patients, researchers and scientists, "there is concern that caution that would normally be exercised for any new technology is not being exercised consistently with large language model tools (LLMs)", it said. Apple Introduces New Software for Cognitive, Speech, Vision Accessibility in Its Products To Be Available This Year.
LLMs include ChatGPT, Bard, Bert and others that imitate understanding, processing, and producing human communication. "This includes widespread adherence to key values of transparency, inclusion, public engagement, expert supervision, and rigorous evaluation," the global health body said in a statement.
"It is imperative that the risks be examined carefully when using LLMs to improve access to health information, as a decision-support tool, or even to enhance diagnostic capacity in under-resourced settings to protect people's health and reduce inequity," it added. Google Dark Web Monitoring Tool Expands To All Gmail Users, Now You Can Run Scans To If Your Gmail Address Is On Dark Web.
The WHO said that "precipitous adoption of untested systems could lead to errors by health-care workers, cause harm to patients, erode trust in AI and thereby undermine (or delay) the potential long-term benefits and uses of such technologies".
The WHO's concerns against the AI tools include that data used to train the AI models may be biased, thus generating misleading or inaccurate information which could pose risks to health, equity and inclusiveness.
The LLMs are also likely to generate responses that can appear authoritative and plausible to an end user and these responses may also be completely incorrect or contain serious errors, especially for health-related responses.
Further, the WHO said that AI may not protect sensitive data (including health data), it can misuse data to generate and disseminate highly convincing disinformation in the form of text, audio or video content that is difficult for the public to differentiate from reliable health content.
"WHO proposes that these concerns be addressed, and clear evidence of benefit be measured before their widespread use in routine health care and medicine -- whether by individuals, care providers or health system administrators and policy-makers," the statement said.
(The above story first appeared on LatestLY on May 16, 2023 09:34 PM IST. For more news and updates on politics, world, sports, entertainment and lifestyle, log on to our website latestly.com).