Powered by AI models trained on troves of text pulled from the internet, chatbots such as ChatGPT and Google’s Bard responded to the researchers’ questions with a range of misconceptions and falsehoods about Black patients, sometimes including fabricated, race-based equations, according to the study published Friday in the academic journal Digital Medicine.

Experts worry these systems could cause real-world harms and amplify forms of medical racism that have persisted for generations as more physicians use chatbots for help with daily tasks such as emailing patients or appealing to health insurers.

The report found that all four models tested — ChatGPT and the more advanced GPT-4, both from OpenAI; Google’s Bard, and Anthropic’s Claude — failed when asked to respond to medical questions about kidney function, lung capacity and skin thickness. In some cases, they appeared to reinforce long-held false beliefs about biological differences between Black and white people that experts have spent years trying to eradicate from medical institutions.

    • BURN@lemmy.world
      link
      fedilink
      arrow-up
      9
      ·
      1 year ago

      Because medical research and best practices aren’t always followed in the real world. There’s huge racial and gender equity problems in the healthcare systems that are bigger than any one doctor or hospital.

      AI simply amplifies the bias present in training data. There are sometimes methods to try to mitigate this, but I’m not sure how effective they are these days.

    • Dr. Dabbles@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      1 year ago

      Lots of people still being taught that clack skin is literally thicker, and black people experience less pain.

    • Potatos_are_not_friends@lemmy.world
      link
      fedilink
      arrow-up
      7
      arrow-down
      1
      ·
      1 year ago

      Because a lot of medical research is from a white person perspective and outdated.

      Imagine if they looked at a little girl and went “You are abnormal according to the lungs of the average 35 yo Caucasian”

    • Chetzemoka
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Nurse here: in addition to the very real problem that medical research is still based on norms derived from young and middle-aged white men, those best practices are just where we begin treatment. It’s extraordinarily rare that the best practice “recipe” works perfectly the first time.

      There’s so much we can’t tell about one individual’s genetic, metabolic, or hormonal quirks with any test available to us today. So there’s still a lot of trial and error that requires tweaking the standard formula until we find exactly what works for each individual person based on their unique response to meds and treatments.

      And some of that response includes individual preference. “I don’t like the way this drug makes me feel” is a valid response, and determining how to work with that or around it requires building a trust relationship with a real human being. I can’t picture any future in which AI is robust enough to remove the therapeutic relationship part of the equation.