Typos and slang spur AI to discourage seeking medical care

Typos and slang spur AI to discourage seeking medical care

Be prudent to ask for advice on AI at the time of seeing a doctor

Chong Kee Siong / Getty Images

Do you have to see a doctor on your sore throat? AI advice can depend on how you have thoroughly hit your question. When models of artificial intelligence were tested on the simulated writing of potential patients, they were more likely to advise to search for medical care if the writer made typing mistakes, included emotional or uncertain – or women.

“Insidious prejudices can change the tenor and the content of AI advice, which can lead to subtle but important differences” in the way medical resources are distributed, explains Karandeep Singh at the University of California in San Diego, which was not involved in the study.

Abinitha Gourabathina at the Massachusetts Institute of Technology and his colleagues used AI to help create thousands of patient notes in different formats and styles. For example, some messages included additional spaces and typing faults to imitate patients with a limited mastery of English or less ease with seizure. Other notes have used an uncertain language in the style of writers suffering from health anxiety, colorful expressions that have lent a dramatic or emotional tone or non -sexist pronouns.

The researchers then fueled the notes with four large -language models (LLM) commonly used to feed chatbots and told AI to answer questions about the fact that the patient had to manage his condition at home or visit a clinic, and if the patient should receive certain laboratory tests and other medical resources. These AI models included the OPENAI GPT-4, the LLAMA-3-70B and LLAMA-3-8B of META, and the PALMYRA-MED model developed for the health care industry by the IA corporate writer.

The tests have shown that the various format and style changes have made all AI models between 7 and 9 percent more likely to recommend patients to stay at home instead of obtaining medical care. The models were also more likely to recommend that patients remain at home, and follow -up research has shown that they were more likely that human clinicians to change their recommendations for treatments due to gender and language style in messages.

OPENAI and META did not respond to a request for comments. The writer does not “recommend nor support” using LLM – including the company’s palmyra -med model – for clinical decisions or “human -free in the loop”, explains Zayed Yasin at Writer.

Most operational AI tools currently used in electronic health file systems are based on the OPENAI GPT-4O, which has not been specifically studied in this research, explains Singh. But he said that a big point to remember of the study is the need to improve means “to assess and monitor generative AI models” used in the health care industry.

Subjects:

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button