Over just a few months, ChatGPT went from correctly answering a simple math problem 98% of the time to just 2%, study finds. Researchers found wild fluctuations—called drift—in the technology’s abi…::ChatGPT went from answering a simple math correctly 98% of the time to just 2%, over the course of a few months.

  • CaptainAniki@lemmy.flight-crew.org
    link
    fedilink
    English
    arrow-up
    15
    arrow-down
    3
    ·
    edit-2
    1 year ago

    I don’t think it’s that easy. These are vLLMs that feed back on themselves to produce “better” results. These models don’t have single point release cycles. It’s a constantly evolving blob of memory and storage orchestrated across a vast number of disk arrays and cabinets of hardware.

    [e]I am wrong the models are version controlled and do have releases.

    • drspod@lemmy.ml
      link
      fedilink
      English
      arrow-up
      17
      ·
      1 year ago

      That’s not how these LLMs work. There is a training phase which takes a large amount of compute power, and the training generates a model which is a set of weights and could easily be backed up and version-controlled. The model is then used for inference which is a less compute-intensive process and runs on much smaller hardware than the training phase.

      The inference architecture does use feedback mechanisms but the feedback does not modify the model-weights that were generated at training time.

    • agent_flounder@lemmy.one
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      3
      ·
      1 year ago

      Even so, surely they can take snapshots. If they’re that clueless about rudimentary practices of IT operations then it is just a matter of time before an outage wipes everything. I find it hard to believe nobody considered a way to do backups, rollbacks, or any of that.