WHO's move for regulation of AI in healthcare highlights risks: Report

The WHO recognises the potential of AI in healthcare, as it could improve existing devices or systems through strengthening clinical trials, improving diagnosis and treatment, and aiding the knowledge and skills of healthcare professionals.

Published On 2023-10-25 04:00 GMT   |   Update On 2023-10-25 09:44 GMT
Advertisement

The World Health Organization's (WHO) recent considerations for the regulation of artificial intelligence (AI) in healthcare, highlights the potential challenges associated with using AI tools in the sector, according to a report on Wednesday.

The WHO regulatory considerations touch on the importance of establishing safety and effectiveness in AI tools, making systems available to those who need them, and fostering dialogue among those who develop and use AI tools.

Advertisement

The WHO recognises the potential of AI in healthcare, as it could improve existing devices or systems through strengthening clinical trials, improving diagnosis and treatment, and aiding the knowledge and skills of healthcare professionals.

Also Read:ChatGPT matches doctors in suggesting likely diagnoses in the emergency medicine department

The report by GlobalData, a data and analytics company, notes that AI technologies are and have been deployed quite quickly, and not always with a full understanding of how they will work in the long run, which could be harmful to healthcare professionals or patients.

"AI has already improved several devices and systems, and there are so many benefits of AI. However, there are risks too with these tools and the rapid adoption of them," said Alexandra Murdoch, Senior Analyst at GlobalData, in a statement.

AI systems in medical or healthcare often have access to personal and medical information, so there should be regulatory frameworks in place to ensure privacy and security. There are a number of other potential challenges with AI in healthcare, such as unethical data collection, cybersecurity risks, and amplifying biases and misinformation.

A recent example of biases in AI tools comes from a study conducted by Stanford University. The study results revealed that some AI chatbots provided responses that perpetuated false medical information about people of colour.

The study ran nine questions through four AI chatbots, including OpenAI’s ChatGPT and Google's Bard. All four of the chatbots used debunked race-based information when asked about kidney and lung function.

"The use of false medical information is deeply concerning and could lead to a number of issues, including misdiagnoses or improper treatment for patients of colour," Murdoch said.

The WHO has released six areas for regulation of AI for health, citing a need to manage the risks of AI amplifying biases in training data. The six areas for regulation are transparency and documentation; risk management; validating data and being clear about the intended use of AI; a commitment to data quality; privacy and data protection; and fostering collaboration.

"With these areas for regulation outlined, governments and regulatory bodies can follow them and hopefully develop some regulations to protect healthcare professionals and patients, and also use AI to its full potential in healthcare," Murdoch said.

Tags:    
Article Source : IANS

Disclaimer: This website is primarily for healthcare professionals. The content here does not replace medical advice and should not be used as medical, diagnostic, endorsement, treatment, or prescription advice. Medical science evolves rapidly, and we strive to keep our information current. If you find any discrepancies, please contact us at corrections@medicaldialogues.in. Read our Correction Policy here. Nothing here should be used as a substitute for medical advice, diagnosis, or treatment. We do not endorse any healthcare advice that contradicts a physician's guidance. Use of this site is subject to our Terms of Use, Privacy Policy, and Advertisement Policy. For more details, read our Full Disclaimer here.

NOTE: Join us in combating medical misinformation. If you encounter a questionable health, medical, or medical education claim, email us at factcheck@medicaldialogues.in for evaluation.

Our comments section is governed by our Comments Policy . By posting comments at Medical Dialogues you automatically agree with our Comments Policy , Terms And Conditions and Privacy Policy .

Similar News