• Chozo@fedia.io
    link
    fedilink
    arrow-up
    0
    ·
    2 days ago

    Read about how LLMs actually work before you read articles written by people who don’t understand LLMs. The author of this piece is suggesting arguments that imply that LLMs have cognition. “Lying” requires intent, and LLMs have no intention, they only have instructions. The author would have you believe that these LLMs are faulty or unreliable, when in actuality they’re working exactly as they’ve been designed to.

    • thedruid@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      So working as designed means presenting false info?

      Look , no one is ascribing intelligence or intent to the machine. The issue is the machines aren’t very good and are being marketed as awesome. They aren’t

      • Chozo@fedia.io
        link
        fedilink
        arrow-up
        0
        ·
        2 days ago

        So working as designed means presenting false info?

        Yes. It was told to conduct a task. It did so. What part of that seems unintentional to you?

        • thedruid@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          2 days ago

          That’s not completing a task. That’s faking a result for appearance.

          Is that what you’re advocating for ?

          If I ask an llm to tell me the difference between aeolian mode and Dorian mode in the field of music , and it gives me the wrong info, then no it’s not working as intended

          See I chose that example because I know the answer. The llm didn’t. But it gave me an answer. An incorrect one

          I want you to understand this. You’re fighting the wrong battle. The llms do make mistakes. Frequently. So frequently that any human who made the same amount of mistakes wouldn’t keep their job.

          But the investment, the belief in a.i is so engrained for some of us who so want a bright and technically advanced future, that you are now making excuses for it. I get it. I’m not insulting you. We are humans. We do that. There are subjects I am sure you could point at where I do this as well

          But a.i.? No. It’s just wrong so often. It’s not it’s fault. Who knew that when we tried to jump ahead in the tech timeline, that we should have actually invented guardrail tech first?

          Instead we let the cart go before the horses, AGAIN, because we are dumb creatures , and now people are trying to force things that don’t work correctly to somehow be shown to be correct.

          I know. A mouthful. But honestly. A.i. is poorly designed, poorly executed, and poorly used.

          It is hastening the end of man. Because those who have been singing it’s praises are too invested to admit it.

          It simply ain’t ready.

          Edit: changed “would” to “wouldn’t”

          • Chozo@fedia.io
            link
            fedilink
            arrow-up
            0
            ·
            2 days ago

            That’s not completing a task.

            That’s faking a result for appearance.

            That was the task.

            • thedruid@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              2 days ago

              No, the task was To tell me the difference in the two modes.

              It provided incorrect information and passed it off as accurate. It didn’t complete the task

              You know that though. You’re just too invested to admit it. So I will withdraw. Enjoy your day.

    • WanderingThoughts@europe.pub
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 days ago

      as they’ve been designed to

      Well, designed is maybe too strong a term. It’s more like stumbling on something that works and expand from there. It’s all still build on the fundaments of the nonsense generator that was chatGPT 2.

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        0
        ·
        2 days ago

        Given how dramatically LLMs have improved over the past couple of years I think it’s pretty clear at this point that AI trainers do know something of what they’re doing and aren’t just randomly stumbling around.

        • WanderingThoughts@europe.pub
          link
          fedilink
          English
          arrow-up
          0
          ·
          2 days ago

          A lot of the improvement came from finding ways to make it bigger and more efficient. That is running into the inherent limits, so the real work with other models just started.

          • Natanael@infosec.pub
            link
            fedilink
            English
            arrow-up
            0
            ·
            2 days ago

            And from reinforcement learning (specifically, making it repeat tasks where the answer can be computer checked)