• Snot Flickerman@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    20
    ·
    edit-2
    3 months ago

    This comes up again and again:

    AI only does a half-ass job, so you need a real human stepping in to “fix it in post.”

    We’re seeing corporations throwing money hand over fist at AI because corporations want it to replace workers.

    We’re burning an extra planets worth of energy for something that still needs human intervention to be usable.

    Maybe, just maybe, we could just pay humans a living fucking wage to do the same work to begin with instead of constantly trying to find more and more ways to just not pay people at all.

    Like, shocker, if you have a fully staffed customer service department, you actually will solve problems for people faster than an AI staffed customer service department.

    The corpos don’t care, they’re not actually interested in solving our problems.

    • just another dev@lemmy.my-box.dev
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      3
      ·
      3 months ago

      That’s a a bit too absolute way to look at it.

      From their point of view the goal isn’t to abolish human involvement, but to minimise the cost. So if they can do the job at the same quality with a quarter of the personnel through AI assistance for less cost, obviously they’re gonna do that.

      At the same time, just because humans having crappy jobs is the current way we solve the problem of people getting money, doesn’t mean we should keep on doing that. Basic income would be a much nicer solution for that, for example. Try to think a bit less conservatively.

      • Snot Flickerman@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        6
        ·
        3 months ago

        but to minimise the cost

        What about the cost to the environment? That cost is just a negative externality to them and you, apparently. Yet I’m the one accused of thinking “conservatively.”

        Burning ten times as many fossil fuels to “minimise the costs” is literally fucking stupid and short-sighted.

        • Elvith Ma'for@feddit.org
          link
          fedilink
          arrow-up
          4
          ·
          3 months ago

          In the end it’s about money. If one had to pay for environmentally damages (e.g. a new tax on $energyUnit, $resourceUnit,…) and you’d not only pay for the resources + some markup for the producing company and just external externalize the “worth” of the damages (read: the taxpayer,…), then it’s cheaper to use these services instead of humans.

  • CameronDev@programming.dev
    link
    fedilink
    arrow-up
    13
    ·
    3 months ago

    This was pretty clear when observing the output of tldrbot. It would just randomly select paragraphs, ignoring surrounding context, and call it a summary.

    • NaN@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      3 months ago

      The bot demonstrated very well what this article is about. I don’t know the internals, but I also can’t image the bot was using the best and most expensive ways of doing analysis.

      It was pretty bad at “getting the point” even when it was obvious, a better system should be able to do so. Sometimes the point is more difficult to discern and there has to be some judgement, you can see this in comments sometimes where people discuss what “the point” was and not just the data. I imagine an AI would have some difficulty determining what is worth summarizing in these situations especially.

    • just another dev@lemmy.my-box.dev
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      3
      ·
      3 months ago

      I’m not sure how long ago that was, but LLM context sizes have grown exponentially in the past year, from 4k tokens to over a hundred k. That doesn’t necessarily affect the quality of the output, although you can’t expect it to summarize what it can’t hold on memory.

  • obbeel@lemmy.eco.br
    link
    fedilink
    arrow-up
    1
    ·
    3 months ago

    Maybe we just need a different type of NLP to work with summarization. I have noticed before LLMs are unlikely to escape their ‘base’ knowledge.

    • ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
      link
      fedilink
      arrow-up
      2
      ·
      3 months ago

      That’s very likely the case. LLMs sucking up all the air in ML research right now, and we shouldn’t be using them as a hammer for every problem.

    • ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
      link
      fedilink
      arrow-up
      6
      ·
      3 months ago

      people doing more work is actually more expensive because human time is by far the biggest cost for most businesses

      A test of AI for Australia’s corporate regulator found that the technology might actually make more work for people, not less.