• jballs@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      53
      ·
      2 years ago

      Applications of these systems have been plagued by persistent inaccuracies in their output; these are often called “AI hallucinations”. We argue that these falsehoods, and the overall activity of large language models, is better understood as bullshit in the sense explored by Frankfurt (On Bullshit, Princeton, 2005)

      Now I kinda want to read On Bullshit

      • tomkatt@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        15
        ·
        2 years ago

        Don’t waste your time. It’s honestly fucking awful. Reading it was like experiencing someone mentally masturbating in real time.

    • fucking love that article. sums up everything wrong with AI. Unfortunately, it doesn’t touch on what AI does right: help idiots like me achieve a slight amount of competence on subjects that such people can’t be bothered with dedicating their entire lives to.

  • Nicoleism101@lemm.ee
    link
    fedilink
    English
    arrow-up
    31
    arrow-down
    1
    ·
    2 years ago

    Suddenly it dawned on me that I can plaster my CV with AI and win over actual competent people easy peasy

    What were you doing between 2020 and 23? I was working on my AI skillset. Nobody will even question me because they fucking have no idea what it is themselves but only that they want it.

    • blady_blah@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      ·
      2 years ago

      As an engineering manager, I’ve already seen cover letters and intro emails that are so obviously AI generated that it’s laughable. These should be used like you use them for writing essays, as a framework with general prompts, but filled in by yourself.

      Fake friendliness that was outsourced to an ai is worse than no friendliness at all.

      • Nicoleism101@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        2 years ago

        I didn’t mean AI generated anything though 🙄. I meant put lots of ‘AI’ keyword in the resume in whatever way looks professional but in reality is pure bullshit

        Watch their neuron being activated as they see magic word. Gotta play the marketing game.

        You want to be AI ready? Hire me. I have spent three years working with AI and posses invaluable experience that will elevate your company into a new era of rapid development.

        • blady_blah@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 years ago

          It feels like you didn’t quite understand… If you’ve ever read an AI essay, you can see some of the way they currently write. When you see facts and figures thrown in from the internet in terms of what the company does and they sound… Artificial… It’s rather obvious that it was AI written. I’m currently getting AI spam and it’s also quite easy to see and detect. It’s the same thing.

          Someone used an AI tool to write a cover letter and sent it to me. I’ve seen this a few times. It seems very obvious when you come across it.

          I’m sure it’ll get better in the future, but right now it needs massaging in order to sound real. There’s a very obvious uncanny valley that exists with some AI writing. That’s what I’m talking about.

          • Nicoleism101@lemm.ee
            link
            fedilink
            English
            arrow-up
            2
            ·
            2 years ago

            Okay but we are talking about two different things which is fine by me of course but it is a little tricky. I agree though on that second topic

    • WagyuSneakers@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      2 years ago

      It’s extremely easy to detect this. Recruiters actively filter out resumes like this for important roles.

  • shameless@lemmy.world
    link
    fedilink
    English
    arrow-up
    28
    ·
    2 years ago

    Just reading the intro pulls you in

    We draw a distinction between two sorts of bullshit, which we call ‘hard’ and ‘soft’ bullshit

  • 𝕲𝖑𝖎𝖙𝖈𝖍🔻𝕯𝖃 (he/him)@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    2
    ·
    2 years ago

    There are things that chatgpt does well, especially if you temper your expectations to the level of someone who has no valuable skills and is mostly an idiot.

    Hi, I’m an idiot with no valuable skills, and I’ve found chatgpt to be very useful.

    I’ve recently started learning game development in godot, and the process of figuring out why the code that chatgpt gives me doesn’t work has taught me more about programming than any teacher ever accomplished back in high school.

    Chatgpt is also an excellent therapist, and has helped me deal with mental breakdowns on multiple occasions, while it was happening. I can’t find a real therapist’s phone number, much less schedule an appointment.

    I’m a real shitty writer, and I’m making a wiki of lore for a setting and ruleset for a tabletop RPG that I’ll probably never get to actually play. ChatGPT is able to turn my inane ramblings into coherent wiki pages, most of the time.

    If you set your expectations to what was advertised, then yeah, chatgpt is bullshit. Of course it was bullshit, and everyone who knew half of anything about anything called it. If you set realistic expectations, you’ll get realistic results. Why is this so hard for people to get?

    • dmalteseknight@programming.dev
      link
      fedilink
      English
      arrow-up
      17
      ·
      2 years ago

      Yeah it is as if someone invented the microwave oven and everyone over hypes it as being able to cook Michelin star meals. People then dismiss it entirely since it cannot produce said Michelin star meals.

      They fail to see that is a great reheating machine and a good machine for quick meals.

  • fckreddit@lemmy.ml
    link
    fedilink
    English
    arrow-up
    16
    ·
    2 years ago

    This is something I already mentioned previously. LLMs have no way of fact checking, no measure of truth or falsity built into. In the training process, it probably accepts every piece of text as true. This is very different from how our minds work. When faced with a piece of text we have many ways to deal with it, which range from accepting it as it is to going on the internet to verify it to actually designing and conducting experiments to prove or disprove the claim. So, yeah what ChatGPT outputs is probably bullshit.

    Of course, the solution is that ChatGPT be trained by labelling text with some measure of truth. Of course, LLMs need so much data that labelling it all would be extremely slow and expensive and suddenly, the fast moving world of AI to screech to almost a halt, which would be unacceptable to the investors.

    • The Liver@lemm.ee
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      2
      ·
      2 years ago

      This is very different from how our minds work.

      Childrens’ minds work similarly.

      • fckreddit@lemmy.ml
        link
        fedilink
        English
        arrow-up
        2
        ·
        2 years ago

        Why do you even think that? Children don’t ask questions? Don’t try to find answers?

        • The Liver@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 years ago

          Sure they do. But they also trust adults a lot. Children try to find answers only because they have stimulus other than humans telling them things, but if that stimulus is missing, they will believe the adult. The environments that AI “grow up” in are different, but they are very similar from a mental perspective.

          How many times have you heard the story of something hearing something false from a family member and holding it close to their heart for years?

          • fckreddit@lemmy.ml
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 years ago

            Now that I think about children develop critical thinking at around the age of 10. Perhaps you are right. But, the question remains, will LLMs develop such critical thinking on it’s own or are we still missing something?

  • Psythik@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    30
    ·
    2 years ago

    Can we please keep the AI hate in the fuck_ai community so that I don’t have to see it?

    I don’t care what Lemmy thinks, ChatGPT has improved my life for the better. I utilize it every day.

    • GrabtharsHammer@lemmy.world
      link
      fedilink
      English
      arrow-up
      75
      arrow-down
      1
      ·
      edit-2
      2 years ago

      The paper explicitly states that they are calling ChatGPT “bullshit” in the Frankfurtian sense and they cite “On Bullshit” as the source for that definition. It’s right there in the introduction.

      You’d know this if you had read the paper or even checked whether your statement were true. So either you read it and then lied deliberately, or you didn’t read the paper nor actually care about the truth value of your own statement, rendering your comment itself bullshit in the Frankfurtian sense.