An MIT study finds non-clinical information in patient messages, like typos, extra whitespace, or colorful language, can reduce the accuracy of a large language model deployed to make treatment recommendations. The LLMs were consistently less accurate for female patients, even when all gender markers were removed from the text.
There’s a potentially justifiable use case in training one and evaluating its performance for use in, idk, triaging a mass-casualty event. Similar to the 911 bot they announced the other day.
Also similar to the 911 bot, i expect it’s already being used to justify cuts in necessary staffing so it’s going to be required in every ER to
maintain higher profit marginsjust keep the lights on.