• grue@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    ·
    6 days ago

    I’d like to know what specific steps are being taken to remove the bias from the training data, then. You cannot just feed the model a big spreadsheet of human decisions up to this point because the current system is itself biased; all you’ll get if you do that is a tool that’s more consistent in applying the same systemic skew.

    • reversedposterior@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      arrow-down
      2
      ·
      6 days ago

      There is an implicit assumption here that models are being ‘trained’, perhaps because LLMs are a hot topic. By models we are usually talking about things like decision trees or regression models or Markov models that put in risk probabilities of various eventualities based on patient characteristics. These things are not designed to mimic human decision makers, they are designed to make as objective a recommendation as possible based on probability and utility and then left down to doctors to use the result in whichever way seems best suited to the context. If you have one liver and 10 patients, it seems prudent to have some sort of calculation as to who is going to have the best likely outcome to decide who to give it to, for example, then just asking one doctor that may be swayed by a bunch of irrelevant factors.

      • grue@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        6 days ago

        There is an implicit assumption here that models are being ‘trained’, perhaps because LLMs are a hot topic. By models we are usually talking about things like decision trees or regression models or Markov models that put in risk probabilities of various eventualities based on patient characteristics.

        [Citation needed]

        If these things were being based on traditional AI techniques instead of neural network techniques, why are they getting implemented now (when, as you say, LLMs are the hot topic) instead of a decade or so ago when that other stuff was in vogue?

        I think the assumption that they’re using training data is a very good one in the absence of evidence to the contrary.